{ "best_metric": 1.1950929164886475, "best_model_checkpoint": "autotrain-w77ed-kah7g/checkpoint-3375", "epoch": 27.0, "eval_steps": 500, "global_step": 3375, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.008, "grad_norm": NaN, "learning_rate": 0.0, "loss": 4.0902, "step": 1 }, { "epoch": 0.016, "grad_norm": NaN, "learning_rate": 0.0, "loss": 4.0765, "step": 2 }, { "epoch": 0.024, "grad_norm": Infinity, "learning_rate": 0.0, "loss": 4.1813, "step": 3 }, { "epoch": 0.032, "grad_norm": 102.83165740966797, "learning_rate": 5.330490405117271e-08, "loss": 3.5184, "step": 4 }, { "epoch": 0.04, "grad_norm": 207.04676818847656, "learning_rate": 1.0660980810234542e-07, "loss": 4.7685, "step": 5 }, { "epoch": 0.048, "grad_norm": Infinity, "learning_rate": 1.0660980810234542e-07, "loss": 3.9354, "step": 6 }, { "epoch": 0.056, "grad_norm": NaN, "learning_rate": 1.0660980810234542e-07, "loss": 3.9252, "step": 7 }, { "epoch": 0.064, "grad_norm": 415.4649963378906, "learning_rate": 1.5991471215351813e-07, "loss": 4.0493, "step": 8 }, { "epoch": 0.072, "grad_norm": 252.53280639648438, "learning_rate": 2.1321961620469084e-07, "loss": 4.224, "step": 9 }, { "epoch": 0.08, "grad_norm": 125.05004119873047, "learning_rate": 2.665245202558635e-07, "loss": 4.0366, "step": 10 }, { "epoch": 0.088, "grad_norm": 154.74398803710938, "learning_rate": 3.1982942430703626e-07, "loss": 4.255, "step": 11 }, { "epoch": 0.096, "grad_norm": 105.42467498779297, "learning_rate": 3.7313432835820895e-07, "loss": 4.4188, "step": 12 }, { "epoch": 0.104, "grad_norm": 88.59154510498047, "learning_rate": 4.264392324093817e-07, "loss": 4.2843, "step": 13 }, { "epoch": 0.112, "grad_norm": 127.18827056884766, "learning_rate": 4.797441364605544e-07, "loss": 4.6171, "step": 14 }, { "epoch": 0.12, "grad_norm": 111.16188049316406, "learning_rate": 5.33049040511727e-07, "loss": 4.7212, "step": 15 }, { "epoch": 0.128, "grad_norm": 251.9276580810547, "learning_rate": 5.863539445628999e-07, "loss": 4.0712, "step": 16 }, { "epoch": 0.136, "grad_norm": 52.2146110534668, "learning_rate": 6.396588486140725e-07, "loss": 4.1838, "step": 17 }, { "epoch": 0.144, "grad_norm": 181.1896209716797, "learning_rate": 6.929637526652453e-07, "loss": 4.3838, "step": 18 }, { "epoch": 0.152, "grad_norm": 111.31734466552734, "learning_rate": 7.462686567164179e-07, "loss": 4.9279, "step": 19 }, { "epoch": 0.16, "grad_norm": 178.46221923828125, "learning_rate": 7.995735607675907e-07, "loss": 4.2457, "step": 20 }, { "epoch": 0.168, "grad_norm": 128.06590270996094, "learning_rate": 8.528784648187634e-07, "loss": 4.3477, "step": 21 }, { "epoch": 0.176, "grad_norm": 108.39094543457031, "learning_rate": 9.06183368869936e-07, "loss": 4.1682, "step": 22 }, { "epoch": 0.184, "grad_norm": Infinity, "learning_rate": 9.06183368869936e-07, "loss": 4.1555, "step": 23 }, { "epoch": 0.192, "grad_norm": 229.01416015625, "learning_rate": 9.594882729211088e-07, "loss": 4.4233, "step": 24 }, { "epoch": 0.2, "grad_norm": 145.1098175048828, "learning_rate": 1.0127931769722815e-06, "loss": 4.027, "step": 25 }, { "epoch": 0.208, "grad_norm": 153.5817108154297, "learning_rate": 1.066098081023454e-06, "loss": 3.9316, "step": 26 }, { "epoch": 0.216, "grad_norm": 159.32347106933594, "learning_rate": 1.119402985074627e-06, "loss": 4.3224, "step": 27 }, { "epoch": 0.224, "grad_norm": 136.42689514160156, "learning_rate": 1.1727078891257998e-06, "loss": 3.728, "step": 28 }, { "epoch": 0.232, "grad_norm": Infinity, "learning_rate": 1.1727078891257998e-06, "loss": 4.0202, "step": 29 }, { "epoch": 0.24, "grad_norm": 104.90229797363281, "learning_rate": 1.2260127931769724e-06, "loss": 3.8234, "step": 30 }, { "epoch": 0.248, "grad_norm": 136.44313049316406, "learning_rate": 1.279317697228145e-06, "loss": 4.4017, "step": 31 }, { "epoch": 0.256, "grad_norm": 141.49464416503906, "learning_rate": 1.3326226012793177e-06, "loss": 3.7363, "step": 32 }, { "epoch": 0.264, "grad_norm": 135.90191650390625, "learning_rate": 1.3859275053304905e-06, "loss": 3.7994, "step": 33 }, { "epoch": 0.272, "grad_norm": 103.09834289550781, "learning_rate": 1.4392324093816632e-06, "loss": 3.851, "step": 34 }, { "epoch": 0.28, "grad_norm": 292.1676940917969, "learning_rate": 1.4925373134328358e-06, "loss": 3.7304, "step": 35 }, { "epoch": 0.288, "grad_norm": 891.6026000976562, "learning_rate": 1.5458422174840086e-06, "loss": 4.8845, "step": 36 }, { "epoch": 0.296, "grad_norm": 247.7584228515625, "learning_rate": 1.5991471215351815e-06, "loss": 4.4127, "step": 37 }, { "epoch": 0.304, "grad_norm": 155.49534606933594, "learning_rate": 1.652452025586354e-06, "loss": 3.8562, "step": 38 }, { "epoch": 0.312, "grad_norm": 90.52159118652344, "learning_rate": 1.7057569296375267e-06, "loss": 4.2185, "step": 39 }, { "epoch": 0.32, "grad_norm": 467.27862548828125, "learning_rate": 1.7590618336886994e-06, "loss": 4.1804, "step": 40 }, { "epoch": 0.328, "grad_norm": 443.3922119140625, "learning_rate": 1.812366737739872e-06, "loss": 4.0725, "step": 41 }, { "epoch": 0.336, "grad_norm": 636.8961791992188, "learning_rate": 1.8656716417910446e-06, "loss": 3.9426, "step": 42 }, { "epoch": 0.344, "grad_norm": 177.8821258544922, "learning_rate": 1.9189765458422177e-06, "loss": 3.8715, "step": 43 }, { "epoch": 0.352, "grad_norm": 166.6934814453125, "learning_rate": 1.9722814498933903e-06, "loss": 4.204, "step": 44 }, { "epoch": 0.36, "grad_norm": 123.8388900756836, "learning_rate": 2.025586353944563e-06, "loss": 4.2422, "step": 45 }, { "epoch": 0.368, "grad_norm": 361.8846740722656, "learning_rate": 2.0788912579957356e-06, "loss": 3.7292, "step": 46 }, { "epoch": 0.376, "grad_norm": 147.11236572265625, "learning_rate": 2.132196162046908e-06, "loss": 4.2874, "step": 47 }, { "epoch": 0.384, "grad_norm": 98.98729705810547, "learning_rate": 2.185501066098081e-06, "loss": 3.5887, "step": 48 }, { "epoch": 0.392, "grad_norm": 213.5888671875, "learning_rate": 2.238805970149254e-06, "loss": 4.0783, "step": 49 }, { "epoch": 0.4, "grad_norm": 212.3387908935547, "learning_rate": 2.2921108742004265e-06, "loss": 4.0502, "step": 50 }, { "epoch": 0.408, "grad_norm": 164.93853759765625, "learning_rate": 2.3454157782515996e-06, "loss": 4.0961, "step": 51 }, { "epoch": 0.416, "grad_norm": 501.7349853515625, "learning_rate": 2.398720682302772e-06, "loss": 3.8629, "step": 52 }, { "epoch": 0.424, "grad_norm": 160.78961181640625, "learning_rate": 2.452025586353945e-06, "loss": 4.1261, "step": 53 }, { "epoch": 0.432, "grad_norm": 79.80322265625, "learning_rate": 2.5053304904051175e-06, "loss": 4.4473, "step": 54 }, { "epoch": 0.44, "grad_norm": 367.82684326171875, "learning_rate": 2.55863539445629e-06, "loss": 3.9977, "step": 55 }, { "epoch": 0.448, "grad_norm": 134.03321838378906, "learning_rate": 2.6119402985074627e-06, "loss": 3.7485, "step": 56 }, { "epoch": 0.456, "grad_norm": 156.8968505859375, "learning_rate": 2.6652452025586354e-06, "loss": 4.3997, "step": 57 }, { "epoch": 0.464, "grad_norm": 3208.0517578125, "learning_rate": 2.7185501066098084e-06, "loss": 4.0353, "step": 58 }, { "epoch": 0.472, "grad_norm": 244.50680541992188, "learning_rate": 2.771855010660981e-06, "loss": 4.2937, "step": 59 }, { "epoch": 0.48, "grad_norm": 314.8337097167969, "learning_rate": 2.8251599147121537e-06, "loss": 3.8613, "step": 60 }, { "epoch": 0.488, "grad_norm": 214.97659301757812, "learning_rate": 2.8784648187633263e-06, "loss": 3.6899, "step": 61 }, { "epoch": 0.496, "grad_norm": 146.3199005126953, "learning_rate": 2.931769722814499e-06, "loss": 4.3369, "step": 62 }, { "epoch": 0.504, "grad_norm": 194.36965942382812, "learning_rate": 2.9850746268656716e-06, "loss": 3.6458, "step": 63 }, { "epoch": 0.512, "grad_norm": 263.2576904296875, "learning_rate": 3.038379530916844e-06, "loss": 3.9473, "step": 64 }, { "epoch": 0.52, "grad_norm": 139.71437072753906, "learning_rate": 3.0916844349680173e-06, "loss": 4.036, "step": 65 }, { "epoch": 0.528, "grad_norm": 241.41415405273438, "learning_rate": 3.14498933901919e-06, "loss": 3.7503, "step": 66 }, { "epoch": 0.536, "grad_norm": 484.7990417480469, "learning_rate": 3.198294243070363e-06, "loss": 3.7126, "step": 67 }, { "epoch": 0.544, "grad_norm": 111.4244155883789, "learning_rate": 3.2515991471215356e-06, "loss": 3.76, "step": 68 }, { "epoch": 0.552, "grad_norm": 74.1336669921875, "learning_rate": 3.304904051172708e-06, "loss": 3.9712, "step": 69 }, { "epoch": 0.56, "grad_norm": 773.2742919921875, "learning_rate": 3.358208955223881e-06, "loss": 3.9671, "step": 70 }, { "epoch": 0.568, "grad_norm": 51.486541748046875, "learning_rate": 3.4115138592750535e-06, "loss": 3.2445, "step": 71 }, { "epoch": 0.576, "grad_norm": 178.03555297851562, "learning_rate": 3.464818763326226e-06, "loss": 4.0185, "step": 72 }, { "epoch": 0.584, "grad_norm": 1370.1002197265625, "learning_rate": 3.5181236673773987e-06, "loss": 3.92, "step": 73 }, { "epoch": 0.592, "grad_norm": 128.6123809814453, "learning_rate": 3.5714285714285714e-06, "loss": 3.8237, "step": 74 }, { "epoch": 0.6, "grad_norm": 78.73133087158203, "learning_rate": 3.624733475479744e-06, "loss": 3.6424, "step": 75 }, { "epoch": 0.608, "grad_norm": 99.49358367919922, "learning_rate": 3.6780383795309166e-06, "loss": 3.7831, "step": 76 }, { "epoch": 0.616, "grad_norm": 90.54728698730469, "learning_rate": 3.7313432835820893e-06, "loss": 4.1724, "step": 77 }, { "epoch": 0.624, "grad_norm": 111.95537567138672, "learning_rate": 3.7846481876332627e-06, "loss": 3.8014, "step": 78 }, { "epoch": 0.632, "grad_norm": 239.91896057128906, "learning_rate": 3.837953091684435e-06, "loss": 3.4942, "step": 79 }, { "epoch": 0.64, "grad_norm": 309.67694091796875, "learning_rate": 3.891257995735608e-06, "loss": 3.3581, "step": 80 }, { "epoch": 0.648, "grad_norm": 107.93159484863281, "learning_rate": 3.944562899786781e-06, "loss": 3.9047, "step": 81 }, { "epoch": 0.656, "grad_norm": 226.78793334960938, "learning_rate": 3.997867803837953e-06, "loss": 3.9968, "step": 82 }, { "epoch": 0.664, "grad_norm": 363.2069091796875, "learning_rate": 4.051172707889126e-06, "loss": 3.8009, "step": 83 }, { "epoch": 0.672, "grad_norm": 2541.17724609375, "learning_rate": 4.1044776119402985e-06, "loss": 3.715, "step": 84 }, { "epoch": 0.68, "grad_norm": 111.1706771850586, "learning_rate": 4.157782515991471e-06, "loss": 4.0947, "step": 85 }, { "epoch": 0.688, "grad_norm": 90.63905334472656, "learning_rate": 4.211087420042644e-06, "loss": 4.7212, "step": 86 }, { "epoch": 0.696, "grad_norm": 1199.171875, "learning_rate": 4.264392324093816e-06, "loss": 3.5366, "step": 87 }, { "epoch": 0.704, "grad_norm": 153.15151977539062, "learning_rate": 4.317697228144989e-06, "loss": 3.6041, "step": 88 }, { "epoch": 0.712, "grad_norm": 152.06326293945312, "learning_rate": 4.371002132196162e-06, "loss": 3.9834, "step": 89 }, { "epoch": 0.72, "grad_norm": 102.86729431152344, "learning_rate": 4.424307036247335e-06, "loss": 3.8254, "step": 90 }, { "epoch": 0.728, "grad_norm": 509.9048156738281, "learning_rate": 4.477611940298508e-06, "loss": 3.6889, "step": 91 }, { "epoch": 0.736, "grad_norm": 89.12580108642578, "learning_rate": 4.53091684434968e-06, "loss": 3.8558, "step": 92 }, { "epoch": 0.744, "grad_norm": 453.573486328125, "learning_rate": 4.584221748400853e-06, "loss": 3.7508, "step": 93 }, { "epoch": 0.752, "grad_norm": 90.8260726928711, "learning_rate": 4.6375266524520265e-06, "loss": 3.7198, "step": 94 }, { "epoch": 0.76, "grad_norm": 1374.9129638671875, "learning_rate": 4.690831556503199e-06, "loss": 3.6579, "step": 95 }, { "epoch": 0.768, "grad_norm": 143.34153747558594, "learning_rate": 4.744136460554372e-06, "loss": 4.0542, "step": 96 }, { "epoch": 0.776, "grad_norm": 280.37579345703125, "learning_rate": 4.797441364605544e-06, "loss": 3.5532, "step": 97 }, { "epoch": 0.784, "grad_norm": 200.9027862548828, "learning_rate": 4.850746268656717e-06, "loss": 3.5247, "step": 98 }, { "epoch": 0.792, "grad_norm": 130.72764587402344, "learning_rate": 4.90405117270789e-06, "loss": 3.6846, "step": 99 }, { "epoch": 0.8, "grad_norm": 669.874267578125, "learning_rate": 4.957356076759062e-06, "loss": 4.0704, "step": 100 }, { "epoch": 0.808, "grad_norm": 164.2544708251953, "learning_rate": 5.010660980810235e-06, "loss": 3.7945, "step": 101 }, { "epoch": 0.816, "grad_norm": 202.95809936523438, "learning_rate": 5.0639658848614076e-06, "loss": 4.061, "step": 102 }, { "epoch": 0.824, "grad_norm": 179.996337890625, "learning_rate": 5.11727078891258e-06, "loss": 3.537, "step": 103 }, { "epoch": 0.832, "grad_norm": NaN, "learning_rate": 5.11727078891258e-06, "loss": 5.0408, "step": 104 }, { "epoch": 0.84, "grad_norm": 109.87252044677734, "learning_rate": 5.170575692963753e-06, "loss": 3.9, "step": 105 }, { "epoch": 0.848, "grad_norm": 119.73939514160156, "learning_rate": 5.2238805970149255e-06, "loss": 3.7977, "step": 106 }, { "epoch": 0.856, "grad_norm": 44.383750915527344, "learning_rate": 5.277185501066098e-06, "loss": 3.763, "step": 107 }, { "epoch": 0.864, "grad_norm": 397.5496520996094, "learning_rate": 5.330490405117271e-06, "loss": 3.4719, "step": 108 }, { "epoch": 0.872, "grad_norm": 111.31773376464844, "learning_rate": 5.383795309168443e-06, "loss": 3.5569, "step": 109 }, { "epoch": 0.88, "grad_norm": 78.47952270507812, "learning_rate": 5.437100213219617e-06, "loss": 3.8335, "step": 110 }, { "epoch": 0.888, "grad_norm": 118.01991271972656, "learning_rate": 5.4904051172707895e-06, "loss": 3.4776, "step": 111 }, { "epoch": 0.896, "grad_norm": 178.97244262695312, "learning_rate": 5.543710021321962e-06, "loss": 3.1878, "step": 112 }, { "epoch": 0.904, "grad_norm": 103.04617309570312, "learning_rate": 5.597014925373135e-06, "loss": 3.8646, "step": 113 }, { "epoch": 0.912, "grad_norm": 1371.623046875, "learning_rate": 5.650319829424307e-06, "loss": 3.3474, "step": 114 }, { "epoch": 0.92, "grad_norm": 307.04638671875, "learning_rate": 5.70362473347548e-06, "loss": 3.3268, "step": 115 }, { "epoch": 0.928, "grad_norm": 92.86817932128906, "learning_rate": 5.756929637526653e-06, "loss": 3.3481, "step": 116 }, { "epoch": 0.936, "grad_norm": 160.4913330078125, "learning_rate": 5.810234541577825e-06, "loss": 3.4443, "step": 117 }, { "epoch": 0.944, "grad_norm": 143.83164978027344, "learning_rate": 5.863539445628998e-06, "loss": 3.836, "step": 118 }, { "epoch": 0.952, "grad_norm": 69.20307159423828, "learning_rate": 5.9168443496801705e-06, "loss": 3.5781, "step": 119 }, { "epoch": 0.96, "grad_norm": 364.408447265625, "learning_rate": 5.970149253731343e-06, "loss": 3.4229, "step": 120 }, { "epoch": 0.968, "grad_norm": 586.87060546875, "learning_rate": 6.023454157782516e-06, "loss": 3.3831, "step": 121 }, { "epoch": 0.976, "grad_norm": 1042.12939453125, "learning_rate": 6.076759061833688e-06, "loss": 4.0959, "step": 122 }, { "epoch": 0.984, "grad_norm": 410.02386474609375, "learning_rate": 6.130063965884861e-06, "loss": 3.5827, "step": 123 }, { "epoch": 0.992, "grad_norm": 68.23786163330078, "learning_rate": 6.1833688699360345e-06, "loss": 3.582, "step": 124 }, { "epoch": 1.0, "grad_norm": 172.9263458251953, "learning_rate": 6.236673773987207e-06, "loss": 3.3082, "step": 125 }, { "epoch": 1.0, "eval_loss": 3.5648818016052246, "eval_map": 0.0002, "eval_map_50": 0.0005, "eval_map_75": 0.0001, "eval_map_Coverall": 0.0, "eval_map_Face_Shield": 0.0008, "eval_map_Gloves": 0.0, "eval_map_Goggles": 0.0002, "eval_map_Mask": 0.0, "eval_map_large": 0.0007, "eval_map_medium": 0.0, "eval_map_small": 0.0001, "eval_mar_1": 0.0, "eval_mar_10": 0.0131, "eval_mar_100": 0.0265, "eval_mar_100_Coverall": 0.0, "eval_mar_100_Face_Shield": 0.0765, "eval_mar_100_Gloves": 0.0, "eval_mar_100_Goggles": 0.0562, "eval_mar_100_Mask": 0.0, "eval_mar_large": 0.0858, "eval_mar_medium": 0.0076, "eval_mar_small": 0.0063, "eval_runtime": 1.8338, "eval_samples_per_second": 15.814, "eval_steps_per_second": 1.091, "step": 125 }, { "epoch": 1.008, "grad_norm": 201.54318237304688, "learning_rate": 6.28997867803838e-06, "loss": 3.6304, "step": 126 }, { "epoch": 1.016, "grad_norm": 70.73978424072266, "learning_rate": 6.343283582089552e-06, "loss": 3.053, "step": 127 }, { "epoch": 1.024, "grad_norm": 1042.78857421875, "learning_rate": 6.396588486140726e-06, "loss": 4.0006, "step": 128 }, { "epoch": 1.032, "grad_norm": 1459.15869140625, "learning_rate": 6.4498933901918985e-06, "loss": 3.3635, "step": 129 }, { "epoch": 1.04, "grad_norm": 170.74203491210938, "learning_rate": 6.503198294243071e-06, "loss": 3.3833, "step": 130 }, { "epoch": 1.048, "grad_norm": 313.28314208984375, "learning_rate": 6.556503198294244e-06, "loss": 3.6759, "step": 131 }, { "epoch": 1.056, "grad_norm": 157.8958740234375, "learning_rate": 6.609808102345416e-06, "loss": 4.2157, "step": 132 }, { "epoch": 1.064, "grad_norm": 260.15985107421875, "learning_rate": 6.663113006396589e-06, "loss": 3.0841, "step": 133 }, { "epoch": 1.072, "grad_norm": 87.4892578125, "learning_rate": 6.716417910447762e-06, "loss": 3.622, "step": 134 }, { "epoch": 1.08, "grad_norm": 179.10009765625, "learning_rate": 6.769722814498934e-06, "loss": 3.3819, "step": 135 }, { "epoch": 1.088, "grad_norm": 127.61953735351562, "learning_rate": 6.823027718550107e-06, "loss": 3.4135, "step": 136 }, { "epoch": 1.096, "grad_norm": 275.7341003417969, "learning_rate": 6.8763326226012796e-06, "loss": 3.2269, "step": 137 }, { "epoch": 1.104, "grad_norm": 41.51618957519531, "learning_rate": 6.929637526652452e-06, "loss": 3.4256, "step": 138 }, { "epoch": 1.112, "grad_norm": 72.44122314453125, "learning_rate": 6.982942430703625e-06, "loss": 3.5351, "step": 139 }, { "epoch": 1.12, "grad_norm": 63.155155181884766, "learning_rate": 7.0362473347547975e-06, "loss": 2.8471, "step": 140 }, { "epoch": 1.1280000000000001, "grad_norm": 123.28691101074219, "learning_rate": 7.08955223880597e-06, "loss": 3.351, "step": 141 }, { "epoch": 1.1360000000000001, "grad_norm": 81.33992004394531, "learning_rate": 7.142857142857143e-06, "loss": 3.662, "step": 142 }, { "epoch": 1.144, "grad_norm": 70.97921752929688, "learning_rate": 7.196162046908315e-06, "loss": 3.3122, "step": 143 }, { "epoch": 1.152, "grad_norm": 80.78215789794922, "learning_rate": 7.249466950959488e-06, "loss": 2.9737, "step": 144 }, { "epoch": 1.16, "grad_norm": 553.8621215820312, "learning_rate": 7.302771855010661e-06, "loss": 3.4666, "step": 145 }, { "epoch": 1.168, "grad_norm": 99.74082946777344, "learning_rate": 7.356076759061833e-06, "loss": 3.1747, "step": 146 }, { "epoch": 1.176, "grad_norm": 69.16419219970703, "learning_rate": 7.409381663113006e-06, "loss": 3.3188, "step": 147 }, { "epoch": 1.184, "grad_norm": 64.77079010009766, "learning_rate": 7.4626865671641785e-06, "loss": 3.2032, "step": 148 }, { "epoch": 1.192, "grad_norm": 182.21173095703125, "learning_rate": 7.515991471215353e-06, "loss": 2.9739, "step": 149 }, { "epoch": 1.2, "grad_norm": 37.460670471191406, "learning_rate": 7.5692963752665255e-06, "loss": 3.1697, "step": 150 }, { "epoch": 1.208, "grad_norm": 90.2082748413086, "learning_rate": 7.622601279317698e-06, "loss": 3.1834, "step": 151 }, { "epoch": 1.216, "grad_norm": 47.36703872680664, "learning_rate": 7.67590618336887e-06, "loss": 3.3831, "step": 152 }, { "epoch": 1.224, "grad_norm": 212.1442108154297, "learning_rate": 7.729211087420043e-06, "loss": 2.8811, "step": 153 }, { "epoch": 1.232, "grad_norm": 336.92620849609375, "learning_rate": 7.782515991471216e-06, "loss": 3.0637, "step": 154 }, { "epoch": 1.24, "grad_norm": 78.58368682861328, "learning_rate": 7.835820895522389e-06, "loss": 3.4187, "step": 155 }, { "epoch": 1.248, "grad_norm": 56.04103469848633, "learning_rate": 7.889125799573561e-06, "loss": 3.4535, "step": 156 }, { "epoch": 1.256, "grad_norm": 307.9919128417969, "learning_rate": 7.942430703624734e-06, "loss": 2.6299, "step": 157 }, { "epoch": 1.264, "grad_norm": 85.22039794921875, "learning_rate": 7.995735607675907e-06, "loss": 3.1836, "step": 158 }, { "epoch": 1.272, "grad_norm": 124.40558624267578, "learning_rate": 8.049040511727079e-06, "loss": 3.2279, "step": 159 }, { "epoch": 1.28, "grad_norm": 95.07052612304688, "learning_rate": 8.102345415778252e-06, "loss": 2.9753, "step": 160 }, { "epoch": 1.288, "grad_norm": 44.7347297668457, "learning_rate": 8.155650319829424e-06, "loss": 2.5114, "step": 161 }, { "epoch": 1.296, "grad_norm": 323.22540283203125, "learning_rate": 8.208955223880597e-06, "loss": 3.2924, "step": 162 }, { "epoch": 1.304, "grad_norm": 65.39293670654297, "learning_rate": 8.26226012793177e-06, "loss": 3.1636, "step": 163 }, { "epoch": 1.312, "grad_norm": 123.57412719726562, "learning_rate": 8.315565031982942e-06, "loss": 3.1259, "step": 164 }, { "epoch": 1.32, "grad_norm": 355.5101013183594, "learning_rate": 8.368869936034115e-06, "loss": 3.1166, "step": 165 }, { "epoch": 1.328, "grad_norm": 123.50281524658203, "learning_rate": 8.422174840085288e-06, "loss": 3.1028, "step": 166 }, { "epoch": 1.336, "grad_norm": 125.6513900756836, "learning_rate": 8.47547974413646e-06, "loss": 2.9316, "step": 167 }, { "epoch": 1.3439999999999999, "grad_norm": 553.2716674804688, "learning_rate": 8.528784648187633e-06, "loss": 2.9642, "step": 168 }, { "epoch": 1.3519999999999999, "grad_norm": 81.45542907714844, "learning_rate": 8.582089552238805e-06, "loss": 3.2012, "step": 169 }, { "epoch": 1.3599999999999999, "grad_norm": 41.021705627441406, "learning_rate": 8.635394456289978e-06, "loss": 3.0348, "step": 170 }, { "epoch": 1.3679999999999999, "grad_norm": 168.9619598388672, "learning_rate": 8.68869936034115e-06, "loss": 2.7773, "step": 171 }, { "epoch": 1.376, "grad_norm": 892.9879760742188, "learning_rate": 8.742004264392323e-06, "loss": 2.6601, "step": 172 }, { "epoch": 1.384, "grad_norm": 135.68197631835938, "learning_rate": 8.795309168443498e-06, "loss": 2.6602, "step": 173 }, { "epoch": 1.392, "grad_norm": 128.42799377441406, "learning_rate": 8.84861407249467e-06, "loss": 2.9436, "step": 174 }, { "epoch": 1.4, "grad_norm": 131.6345977783203, "learning_rate": 8.901918976545843e-06, "loss": 2.94, "step": 175 }, { "epoch": 1.408, "grad_norm": 145.66127014160156, "learning_rate": 8.955223880597016e-06, "loss": 2.7112, "step": 176 }, { "epoch": 1.416, "grad_norm": 98.05323791503906, "learning_rate": 9.008528784648188e-06, "loss": 3.8572, "step": 177 }, { "epoch": 1.424, "grad_norm": 89.7735824584961, "learning_rate": 9.06183368869936e-06, "loss": 2.9377, "step": 178 }, { "epoch": 1.432, "grad_norm": 102.77169799804688, "learning_rate": 9.115138592750533e-06, "loss": 2.4665, "step": 179 }, { "epoch": 1.44, "grad_norm": 103.22952270507812, "learning_rate": 9.168443496801706e-06, "loss": 2.7656, "step": 180 }, { "epoch": 1.448, "grad_norm": 56.59215545654297, "learning_rate": 9.221748400852879e-06, "loss": 2.5097, "step": 181 }, { "epoch": 1.456, "grad_norm": 64.85523986816406, "learning_rate": 9.275053304904053e-06, "loss": 2.616, "step": 182 }, { "epoch": 1.464, "grad_norm": 160.47955322265625, "learning_rate": 9.328358208955226e-06, "loss": 2.675, "step": 183 }, { "epoch": 1.472, "grad_norm": 164.58876037597656, "learning_rate": 9.381663113006398e-06, "loss": 2.8629, "step": 184 }, { "epoch": 1.48, "grad_norm": 54.206443786621094, "learning_rate": 9.434968017057571e-06, "loss": 2.8058, "step": 185 }, { "epoch": 1.488, "grad_norm": 319.7212829589844, "learning_rate": 9.488272921108744e-06, "loss": 2.608, "step": 186 }, { "epoch": 1.496, "grad_norm": 62.96849822998047, "learning_rate": 9.541577825159916e-06, "loss": 2.5888, "step": 187 }, { "epoch": 1.504, "grad_norm": 96.88548278808594, "learning_rate": 9.594882729211089e-06, "loss": 2.5048, "step": 188 }, { "epoch": 1.512, "grad_norm": 213.2740936279297, "learning_rate": 9.648187633262261e-06, "loss": 3.051, "step": 189 }, { "epoch": 1.52, "grad_norm": 83.84225463867188, "learning_rate": 9.701492537313434e-06, "loss": 2.328, "step": 190 }, { "epoch": 1.528, "grad_norm": 60.553977966308594, "learning_rate": 9.754797441364607e-06, "loss": 3.0218, "step": 191 }, { "epoch": 1.536, "grad_norm": 37.41594696044922, "learning_rate": 9.80810234541578e-06, "loss": 2.6763, "step": 192 }, { "epoch": 1.544, "grad_norm": 64.42765808105469, "learning_rate": 9.861407249466952e-06, "loss": 2.5087, "step": 193 }, { "epoch": 1.552, "grad_norm": 153.4246368408203, "learning_rate": 9.914712153518125e-06, "loss": 2.9952, "step": 194 }, { "epoch": 1.56, "grad_norm": 137.24520874023438, "learning_rate": 9.968017057569297e-06, "loss": 2.7894, "step": 195 }, { "epoch": 1.568, "grad_norm": 69.07455444335938, "learning_rate": 1.002132196162047e-05, "loss": 2.7672, "step": 196 }, { "epoch": 1.576, "grad_norm": 136.917724609375, "learning_rate": 1.0074626865671643e-05, "loss": 3.1729, "step": 197 }, { "epoch": 1.584, "grad_norm": 80.07211303710938, "learning_rate": 1.0127931769722815e-05, "loss": 2.695, "step": 198 }, { "epoch": 1.592, "grad_norm": 101.99703979492188, "learning_rate": 1.0181236673773988e-05, "loss": 2.4539, "step": 199 }, { "epoch": 1.6, "grad_norm": 166.60708618164062, "learning_rate": 1.023454157782516e-05, "loss": 2.651, "step": 200 }, { "epoch": 1.608, "grad_norm": 112.76644134521484, "learning_rate": 1.0287846481876333e-05, "loss": 2.6673, "step": 201 }, { "epoch": 1.616, "grad_norm": 174.7171173095703, "learning_rate": 1.0341151385927506e-05, "loss": 2.7943, "step": 202 }, { "epoch": 1.624, "grad_norm": 154.41880798339844, "learning_rate": 1.0394456289978678e-05, "loss": 2.7295, "step": 203 }, { "epoch": 1.6320000000000001, "grad_norm": 92.9516830444336, "learning_rate": 1.0447761194029851e-05, "loss": 3.2615, "step": 204 }, { "epoch": 1.6400000000000001, "grad_norm": 32.60859298706055, "learning_rate": 1.0501066098081024e-05, "loss": 2.7533, "step": 205 }, { "epoch": 1.6480000000000001, "grad_norm": 48.73335647583008, "learning_rate": 1.0554371002132196e-05, "loss": 2.6112, "step": 206 }, { "epoch": 1.6560000000000001, "grad_norm": 62.29795837402344, "learning_rate": 1.0607675906183369e-05, "loss": 2.6598, "step": 207 }, { "epoch": 1.6640000000000001, "grad_norm": 47.86444091796875, "learning_rate": 1.0660980810234541e-05, "loss": 2.9107, "step": 208 }, { "epoch": 1.6720000000000002, "grad_norm": 181.4913330078125, "learning_rate": 1.0714285714285714e-05, "loss": 2.9263, "step": 209 }, { "epoch": 1.6800000000000002, "grad_norm": 160.0330352783203, "learning_rate": 1.0767590618336887e-05, "loss": 2.242, "step": 210 }, { "epoch": 1.688, "grad_norm": 116.67021942138672, "learning_rate": 1.082089552238806e-05, "loss": 2.3673, "step": 211 }, { "epoch": 1.696, "grad_norm": 111.9002685546875, "learning_rate": 1.0874200426439234e-05, "loss": 2.3795, "step": 212 }, { "epoch": 1.704, "grad_norm": 71.91146850585938, "learning_rate": 1.0927505330490406e-05, "loss": 2.8601, "step": 213 }, { "epoch": 1.712, "grad_norm": 57.69649124145508, "learning_rate": 1.0980810234541579e-05, "loss": 2.7469, "step": 214 }, { "epoch": 1.72, "grad_norm": 64.64582824707031, "learning_rate": 1.1034115138592752e-05, "loss": 2.1653, "step": 215 }, { "epoch": 1.728, "grad_norm": 153.0703887939453, "learning_rate": 1.1087420042643924e-05, "loss": 2.6883, "step": 216 }, { "epoch": 1.736, "grad_norm": 97.97472381591797, "learning_rate": 1.1140724946695097e-05, "loss": 2.5812, "step": 217 }, { "epoch": 1.744, "grad_norm": 80.96427917480469, "learning_rate": 1.119402985074627e-05, "loss": 2.617, "step": 218 }, { "epoch": 1.752, "grad_norm": 94.6370849609375, "learning_rate": 1.1247334754797442e-05, "loss": 2.98, "step": 219 }, { "epoch": 1.76, "grad_norm": 180.7654266357422, "learning_rate": 1.1300639658848615e-05, "loss": 2.633, "step": 220 }, { "epoch": 1.768, "grad_norm": 70.87854766845703, "learning_rate": 1.1353944562899787e-05, "loss": 2.5007, "step": 221 }, { "epoch": 1.776, "grad_norm": 109.2392578125, "learning_rate": 1.140724946695096e-05, "loss": 2.6486, "step": 222 }, { "epoch": 1.784, "grad_norm": 57.642478942871094, "learning_rate": 1.1460554371002133e-05, "loss": 2.1644, "step": 223 }, { "epoch": 1.792, "grad_norm": 45.05636215209961, "learning_rate": 1.1513859275053305e-05, "loss": 2.5443, "step": 224 }, { "epoch": 1.8, "grad_norm": 81.91133117675781, "learning_rate": 1.1567164179104478e-05, "loss": 2.3528, "step": 225 }, { "epoch": 1.808, "grad_norm": 71.76690673828125, "learning_rate": 1.162046908315565e-05, "loss": 2.2967, "step": 226 }, { "epoch": 1.8159999999999998, "grad_norm": 88.18803405761719, "learning_rate": 1.1673773987206823e-05, "loss": 2.3866, "step": 227 }, { "epoch": 1.8239999999999998, "grad_norm": 78.47636413574219, "learning_rate": 1.1727078891257996e-05, "loss": 2.4913, "step": 228 }, { "epoch": 1.8319999999999999, "grad_norm": 57.78590774536133, "learning_rate": 1.1780383795309168e-05, "loss": 2.9518, "step": 229 }, { "epoch": 1.8399999999999999, "grad_norm": 291.4474792480469, "learning_rate": 1.1833688699360341e-05, "loss": 2.972, "step": 230 }, { "epoch": 1.8479999999999999, "grad_norm": 65.95680236816406, "learning_rate": 1.1886993603411514e-05, "loss": 2.8616, "step": 231 }, { "epoch": 1.8559999999999999, "grad_norm": 33.53388977050781, "learning_rate": 1.1940298507462686e-05, "loss": 3.0519, "step": 232 }, { "epoch": 1.8639999999999999, "grad_norm": 55.065250396728516, "learning_rate": 1.1993603411513859e-05, "loss": 2.2867, "step": 233 }, { "epoch": 1.8719999999999999, "grad_norm": 47.603397369384766, "learning_rate": 1.2046908315565032e-05, "loss": 2.3558, "step": 234 }, { "epoch": 1.88, "grad_norm": 85.05586242675781, "learning_rate": 1.2100213219616204e-05, "loss": 2.1974, "step": 235 }, { "epoch": 1.888, "grad_norm": 156.00799560546875, "learning_rate": 1.2153518123667377e-05, "loss": 2.4138, "step": 236 }, { "epoch": 1.896, "grad_norm": 202.6873779296875, "learning_rate": 1.220682302771855e-05, "loss": 3.0867, "step": 237 }, { "epoch": 1.904, "grad_norm": 62.922096252441406, "learning_rate": 1.2260127931769722e-05, "loss": 2.1231, "step": 238 }, { "epoch": 1.912, "grad_norm": 185.6517333984375, "learning_rate": 1.2313432835820896e-05, "loss": 2.3957, "step": 239 }, { "epoch": 1.92, "grad_norm": 147.08448791503906, "learning_rate": 1.2366737739872069e-05, "loss": 2.5307, "step": 240 }, { "epoch": 1.928, "grad_norm": 93.88536834716797, "learning_rate": 1.2420042643923242e-05, "loss": 2.3475, "step": 241 }, { "epoch": 1.936, "grad_norm": 92.4790267944336, "learning_rate": 1.2473347547974414e-05, "loss": 3.667, "step": 242 }, { "epoch": 1.944, "grad_norm": 181.81109619140625, "learning_rate": 1.2526652452025587e-05, "loss": 3.1625, "step": 243 }, { "epoch": 1.952, "grad_norm": 75.07173156738281, "learning_rate": 1.257995735607676e-05, "loss": 2.1823, "step": 244 }, { "epoch": 1.96, "grad_norm": 143.43711853027344, "learning_rate": 1.2633262260127932e-05, "loss": 2.517, "step": 245 }, { "epoch": 1.968, "grad_norm": 77.44786071777344, "learning_rate": 1.2686567164179105e-05, "loss": 2.6789, "step": 246 }, { "epoch": 1.976, "grad_norm": 75.38809967041016, "learning_rate": 1.2739872068230277e-05, "loss": 2.2935, "step": 247 }, { "epoch": 1.984, "grad_norm": 70.021484375, "learning_rate": 1.2793176972281452e-05, "loss": 2.4581, "step": 248 }, { "epoch": 1.992, "grad_norm": 28.850391387939453, "learning_rate": 1.2846481876332623e-05, "loss": 2.3343, "step": 249 }, { "epoch": 2.0, "grad_norm": 65.98763275146484, "learning_rate": 1.2899786780383797e-05, "loss": 2.2067, "step": 250 }, { "epoch": 2.0, "eval_loss": 2.506380319595337, "eval_map": 0.0022, "eval_map_50": 0.0067, "eval_map_75": 0.0012, "eval_map_Coverall": 0.0081, "eval_map_Face_Shield": 0.0016, "eval_map_Gloves": 0.0001, "eval_map_Goggles": 0.0009, "eval_map_Mask": 0.0002, "eval_map_large": 0.0032, "eval_map_medium": 0.0011, "eval_map_small": 0.0028, "eval_mar_1": 0.0172, "eval_mar_10": 0.057, "eval_mar_100": 0.0806, "eval_mar_100_Coverall": 0.1267, "eval_mar_100_Face_Shield": 0.1235, "eval_mar_100_Gloves": 0.0213, "eval_mar_100_Goggles": 0.1219, "eval_mar_100_Mask": 0.0096, "eval_mar_large": 0.0998, "eval_mar_medium": 0.0631, "eval_mar_small": 0.0544, "eval_runtime": 1.7968, "eval_samples_per_second": 16.14, "eval_steps_per_second": 1.113, "step": 250 }, { "epoch": 2.008, "grad_norm": 72.55103302001953, "learning_rate": 1.2953091684434968e-05, "loss": 2.6435, "step": 251 }, { "epoch": 2.016, "grad_norm": 166.8722381591797, "learning_rate": 1.3006396588486142e-05, "loss": 2.3469, "step": 252 }, { "epoch": 2.024, "grad_norm": 82.16825866699219, "learning_rate": 1.3059701492537313e-05, "loss": 2.6759, "step": 253 }, { "epoch": 2.032, "grad_norm": 77.62806701660156, "learning_rate": 1.3113006396588488e-05, "loss": 2.6061, "step": 254 }, { "epoch": 2.04, "grad_norm": 62.700477600097656, "learning_rate": 1.3166311300639658e-05, "loss": 2.2588, "step": 255 }, { "epoch": 2.048, "grad_norm": 106.48802947998047, "learning_rate": 1.3219616204690833e-05, "loss": 2.299, "step": 256 }, { "epoch": 2.056, "grad_norm": 70.06613159179688, "learning_rate": 1.3272921108742004e-05, "loss": 2.5648, "step": 257 }, { "epoch": 2.064, "grad_norm": 63.72290802001953, "learning_rate": 1.3326226012793178e-05, "loss": 2.0453, "step": 258 }, { "epoch": 2.072, "grad_norm": 122.59272766113281, "learning_rate": 1.3379530916844352e-05, "loss": 2.2345, "step": 259 }, { "epoch": 2.08, "grad_norm": 59.230751037597656, "learning_rate": 1.3432835820895523e-05, "loss": 2.2233, "step": 260 }, { "epoch": 2.088, "grad_norm": 71.48625946044922, "learning_rate": 1.3486140724946698e-05, "loss": 2.4107, "step": 261 }, { "epoch": 2.096, "grad_norm": 60.626312255859375, "learning_rate": 1.3539445628997869e-05, "loss": 2.2728, "step": 262 }, { "epoch": 2.104, "grad_norm": 175.94512939453125, "learning_rate": 1.3592750533049043e-05, "loss": 2.4996, "step": 263 }, { "epoch": 2.112, "grad_norm": 51.979461669921875, "learning_rate": 1.3646055437100214e-05, "loss": 2.5896, "step": 264 }, { "epoch": 2.12, "grad_norm": 32.45353317260742, "learning_rate": 1.3699360341151388e-05, "loss": 2.2236, "step": 265 }, { "epoch": 2.128, "grad_norm": 70.14639282226562, "learning_rate": 1.3752665245202559e-05, "loss": 2.6526, "step": 266 }, { "epoch": 2.136, "grad_norm": 155.2909698486328, "learning_rate": 1.3805970149253733e-05, "loss": 2.3875, "step": 267 }, { "epoch": 2.144, "grad_norm": 100.18324279785156, "learning_rate": 1.3859275053304904e-05, "loss": 1.963, "step": 268 }, { "epoch": 2.152, "grad_norm": 64.66578674316406, "learning_rate": 1.3912579957356079e-05, "loss": 2.1809, "step": 269 }, { "epoch": 2.16, "grad_norm": 99.26841735839844, "learning_rate": 1.396588486140725e-05, "loss": 2.9643, "step": 270 }, { "epoch": 2.168, "grad_norm": 50.207489013671875, "learning_rate": 1.4019189765458424e-05, "loss": 2.4281, "step": 271 }, { "epoch": 2.176, "grad_norm": 48.44723129272461, "learning_rate": 1.4072494669509595e-05, "loss": 2.9968, "step": 272 }, { "epoch": 2.184, "grad_norm": 53.13323974609375, "learning_rate": 1.412579957356077e-05, "loss": 2.8749, "step": 273 }, { "epoch": 2.192, "grad_norm": 76.62045288085938, "learning_rate": 1.417910447761194e-05, "loss": 2.4159, "step": 274 }, { "epoch": 2.2, "grad_norm": 148.3463897705078, "learning_rate": 1.4232409381663114e-05, "loss": 2.2228, "step": 275 }, { "epoch": 2.208, "grad_norm": 80.96959686279297, "learning_rate": 1.4285714285714285e-05, "loss": 1.896, "step": 276 }, { "epoch": 2.216, "grad_norm": 69.7322006225586, "learning_rate": 1.433901918976546e-05, "loss": 2.2475, "step": 277 }, { "epoch": 2.224, "grad_norm": 62.547813415527344, "learning_rate": 1.439232409381663e-05, "loss": 1.8161, "step": 278 }, { "epoch": 2.232, "grad_norm": 75.9747314453125, "learning_rate": 1.4445628997867805e-05, "loss": 2.1932, "step": 279 }, { "epoch": 2.24, "grad_norm": 33.59276580810547, "learning_rate": 1.4498933901918976e-05, "loss": 2.0232, "step": 280 }, { "epoch": 2.248, "grad_norm": 40.5213737487793, "learning_rate": 1.455223880597015e-05, "loss": 2.1488, "step": 281 }, { "epoch": 2.2560000000000002, "grad_norm": 279.9766540527344, "learning_rate": 1.4605543710021321e-05, "loss": 2.0847, "step": 282 }, { "epoch": 2.2640000000000002, "grad_norm": 103.16998291015625, "learning_rate": 1.4658848614072496e-05, "loss": 2.0742, "step": 283 }, { "epoch": 2.2720000000000002, "grad_norm": 70.21025848388672, "learning_rate": 1.4712153518123666e-05, "loss": 2.3612, "step": 284 }, { "epoch": 2.2800000000000002, "grad_norm": 508.66265869140625, "learning_rate": 1.476545842217484e-05, "loss": 2.5057, "step": 285 }, { "epoch": 2.288, "grad_norm": 40.420806884765625, "learning_rate": 1.4818763326226012e-05, "loss": 1.9379, "step": 286 }, { "epoch": 2.296, "grad_norm": 93.10791778564453, "learning_rate": 1.4872068230277186e-05, "loss": 2.5982, "step": 287 }, { "epoch": 2.304, "grad_norm": 148.4358673095703, "learning_rate": 1.4925373134328357e-05, "loss": 2.1824, "step": 288 }, { "epoch": 2.312, "grad_norm": 75.59492492675781, "learning_rate": 1.4978678038379531e-05, "loss": 2.3049, "step": 289 }, { "epoch": 2.32, "grad_norm": 500.5408020019531, "learning_rate": 1.5031982942430706e-05, "loss": 2.16, "step": 290 }, { "epoch": 2.328, "grad_norm": 77.30692291259766, "learning_rate": 1.5085287846481877e-05, "loss": 3.551, "step": 291 }, { "epoch": 2.336, "grad_norm": 38.95351028442383, "learning_rate": 1.5138592750533051e-05, "loss": 2.2078, "step": 292 }, { "epoch": 2.344, "grad_norm": 81.8768539428711, "learning_rate": 1.5191897654584222e-05, "loss": 2.3245, "step": 293 }, { "epoch": 2.352, "grad_norm": 51.94675064086914, "learning_rate": 1.5245202558635396e-05, "loss": 2.3342, "step": 294 }, { "epoch": 2.36, "grad_norm": 35.24325180053711, "learning_rate": 1.529850746268657e-05, "loss": 2.5393, "step": 295 }, { "epoch": 2.368, "grad_norm": 46.08406066894531, "learning_rate": 1.535181236673774e-05, "loss": 2.2287, "step": 296 }, { "epoch": 2.376, "grad_norm": 155.552734375, "learning_rate": 1.5405117270788914e-05, "loss": 2.3717, "step": 297 }, { "epoch": 2.384, "grad_norm": 87.46237182617188, "learning_rate": 1.5458422174840087e-05, "loss": 2.1686, "step": 298 }, { "epoch": 2.392, "grad_norm": 82.69491577148438, "learning_rate": 1.551172707889126e-05, "loss": 2.1367, "step": 299 }, { "epoch": 2.4, "grad_norm": 70.97594451904297, "learning_rate": 1.5565031982942432e-05, "loss": 2.2441, "step": 300 }, { "epoch": 2.408, "grad_norm": 49.11447525024414, "learning_rate": 1.5618336886993605e-05, "loss": 2.1102, "step": 301 }, { "epoch": 2.416, "grad_norm": 63.715213775634766, "learning_rate": 1.5671641791044777e-05, "loss": 1.7923, "step": 302 }, { "epoch": 2.424, "grad_norm": 165.4727783203125, "learning_rate": 1.572494669509595e-05, "loss": 3.1139, "step": 303 }, { "epoch": 2.432, "grad_norm": 76.63203430175781, "learning_rate": 1.5778251599147122e-05, "loss": 2.0153, "step": 304 }, { "epoch": 2.44, "grad_norm": 108.39789581298828, "learning_rate": 1.5831556503198295e-05, "loss": 2.1193, "step": 305 }, { "epoch": 2.448, "grad_norm": 52.993934631347656, "learning_rate": 1.5884861407249468e-05, "loss": 1.8506, "step": 306 }, { "epoch": 2.456, "grad_norm": 304.5983581542969, "learning_rate": 1.593816631130064e-05, "loss": 2.0791, "step": 307 }, { "epoch": 2.464, "grad_norm": 106.17337036132812, "learning_rate": 1.5991471215351813e-05, "loss": 1.9799, "step": 308 }, { "epoch": 2.472, "grad_norm": 224.80665588378906, "learning_rate": 1.6044776119402986e-05, "loss": 1.9202, "step": 309 }, { "epoch": 2.48, "grad_norm": 64.42595672607422, "learning_rate": 1.6098081023454158e-05, "loss": 1.8497, "step": 310 }, { "epoch": 2.488, "grad_norm": 49.51008987426758, "learning_rate": 1.615138592750533e-05, "loss": 2.1886, "step": 311 }, { "epoch": 2.496, "grad_norm": 75.21533966064453, "learning_rate": 1.6204690831556504e-05, "loss": 2.0429, "step": 312 }, { "epoch": 2.504, "grad_norm": 66.07110595703125, "learning_rate": 1.6257995735607676e-05, "loss": 2.0205, "step": 313 }, { "epoch": 2.512, "grad_norm": 111.90451049804688, "learning_rate": 1.631130063965885e-05, "loss": 2.5755, "step": 314 }, { "epoch": 2.52, "grad_norm": 64.3583755493164, "learning_rate": 1.636460554371002e-05, "loss": 2.7214, "step": 315 }, { "epoch": 2.528, "grad_norm": 159.5388641357422, "learning_rate": 1.6417910447761194e-05, "loss": 2.6166, "step": 316 }, { "epoch": 2.536, "grad_norm": 321.57659912109375, "learning_rate": 1.6471215351812367e-05, "loss": 1.7604, "step": 317 }, { "epoch": 2.544, "grad_norm": 54.400848388671875, "learning_rate": 1.652452025586354e-05, "loss": 2.0253, "step": 318 }, { "epoch": 2.552, "grad_norm": 50.30399703979492, "learning_rate": 1.6577825159914712e-05, "loss": 1.9197, "step": 319 }, { "epoch": 2.56, "grad_norm": 62.78925323486328, "learning_rate": 1.6631130063965885e-05, "loss": 2.2627, "step": 320 }, { "epoch": 2.568, "grad_norm": 183.498779296875, "learning_rate": 1.668443496801706e-05, "loss": 2.1099, "step": 321 }, { "epoch": 2.576, "grad_norm": 66.52242279052734, "learning_rate": 1.673773987206823e-05, "loss": 1.492, "step": 322 }, { "epoch": 2.584, "grad_norm": 109.02804565429688, "learning_rate": 1.6791044776119406e-05, "loss": 2.2107, "step": 323 }, { "epoch": 2.592, "grad_norm": 35.20872497558594, "learning_rate": 1.6844349680170575e-05, "loss": 2.513, "step": 324 }, { "epoch": 2.6, "grad_norm": 64.11994171142578, "learning_rate": 1.689765458422175e-05, "loss": 2.3604, "step": 325 }, { "epoch": 2.608, "grad_norm": 74.66423034667969, "learning_rate": 1.695095948827292e-05, "loss": 1.9693, "step": 326 }, { "epoch": 2.616, "grad_norm": 38.35129928588867, "learning_rate": 1.7004264392324096e-05, "loss": 1.7226, "step": 327 }, { "epoch": 2.624, "grad_norm": 74.43780517578125, "learning_rate": 1.7057569296375266e-05, "loss": 1.7298, "step": 328 }, { "epoch": 2.632, "grad_norm": 73.99781799316406, "learning_rate": 1.711087420042644e-05, "loss": 1.7894, "step": 329 }, { "epoch": 2.64, "grad_norm": 116.11531066894531, "learning_rate": 1.716417910447761e-05, "loss": 2.0361, "step": 330 }, { "epoch": 2.648, "grad_norm": 65.52816009521484, "learning_rate": 1.7217484008528787e-05, "loss": 1.8083, "step": 331 }, { "epoch": 2.656, "grad_norm": 42.67534637451172, "learning_rate": 1.7270788912579956e-05, "loss": 1.8664, "step": 332 }, { "epoch": 2.664, "grad_norm": 78.7031478881836, "learning_rate": 1.7324093816631132e-05, "loss": 1.9191, "step": 333 }, { "epoch": 2.672, "grad_norm": 55.599815368652344, "learning_rate": 1.73773987206823e-05, "loss": 1.9363, "step": 334 }, { "epoch": 2.68, "grad_norm": 64.83665466308594, "learning_rate": 1.7430703624733477e-05, "loss": 2.4513, "step": 335 }, { "epoch": 2.6879999999999997, "grad_norm": 85.89076232910156, "learning_rate": 1.7484008528784647e-05, "loss": 2.0027, "step": 336 }, { "epoch": 2.6959999999999997, "grad_norm": 43.06257247924805, "learning_rate": 1.7537313432835823e-05, "loss": 2.1631, "step": 337 }, { "epoch": 2.7039999999999997, "grad_norm": 70.10470581054688, "learning_rate": 1.7590618336886995e-05, "loss": 2.0292, "step": 338 }, { "epoch": 2.7119999999999997, "grad_norm": 94.68647766113281, "learning_rate": 1.7643923240938168e-05, "loss": 2.0141, "step": 339 }, { "epoch": 2.7199999999999998, "grad_norm": 78.19709777832031, "learning_rate": 1.769722814498934e-05, "loss": 1.9514, "step": 340 }, { "epoch": 2.7279999999999998, "grad_norm": 77.90755462646484, "learning_rate": 1.7750533049040513e-05, "loss": 2.8393, "step": 341 }, { "epoch": 2.7359999999999998, "grad_norm": 59.961788177490234, "learning_rate": 1.7803837953091686e-05, "loss": 1.692, "step": 342 }, { "epoch": 2.7439999999999998, "grad_norm": 30.253637313842773, "learning_rate": 1.785714285714286e-05, "loss": 2.2066, "step": 343 }, { "epoch": 2.752, "grad_norm": 53.41312026977539, "learning_rate": 1.791044776119403e-05, "loss": 1.9632, "step": 344 }, { "epoch": 2.76, "grad_norm": 48.56285095214844, "learning_rate": 1.7963752665245204e-05, "loss": 2.3232, "step": 345 }, { "epoch": 2.768, "grad_norm": 104.9471206665039, "learning_rate": 1.8017057569296376e-05, "loss": 2.4298, "step": 346 }, { "epoch": 2.776, "grad_norm": 83.01089477539062, "learning_rate": 1.807036247334755e-05, "loss": 3.0156, "step": 347 }, { "epoch": 2.784, "grad_norm": 48.717777252197266, "learning_rate": 1.812366737739872e-05, "loss": 2.4031, "step": 348 }, { "epoch": 2.792, "grad_norm": 42.17900085449219, "learning_rate": 1.8176972281449894e-05, "loss": 2.1774, "step": 349 }, { "epoch": 2.8, "grad_norm": 67.32838439941406, "learning_rate": 1.8230277185501067e-05, "loss": 2.396, "step": 350 }, { "epoch": 2.808, "grad_norm": 45.85163497924805, "learning_rate": 1.828358208955224e-05, "loss": 2.0439, "step": 351 }, { "epoch": 2.816, "grad_norm": 75.26972198486328, "learning_rate": 1.8336886993603412e-05, "loss": 2.3707, "step": 352 }, { "epoch": 2.824, "grad_norm": 1064.5076904296875, "learning_rate": 1.8390191897654585e-05, "loss": 1.9662, "step": 353 }, { "epoch": 2.832, "grad_norm": 105.02595520019531, "learning_rate": 1.8443496801705757e-05, "loss": 2.0889, "step": 354 }, { "epoch": 2.84, "grad_norm": 220.5780029296875, "learning_rate": 1.849680170575693e-05, "loss": 2.2733, "step": 355 }, { "epoch": 2.848, "grad_norm": 72.6932373046875, "learning_rate": 1.8550106609808106e-05, "loss": 2.0473, "step": 356 }, { "epoch": 2.856, "grad_norm": 168.8657684326172, "learning_rate": 1.8603411513859275e-05, "loss": 2.7461, "step": 357 }, { "epoch": 2.864, "grad_norm": 210.3399658203125, "learning_rate": 1.865671641791045e-05, "loss": 1.688, "step": 358 }, { "epoch": 2.872, "grad_norm": 176.878173828125, "learning_rate": 1.871002132196162e-05, "loss": 1.7865, "step": 359 }, { "epoch": 2.88, "grad_norm": 91.48789978027344, "learning_rate": 1.8763326226012797e-05, "loss": 2.2004, "step": 360 }, { "epoch": 2.888, "grad_norm": 62.26776885986328, "learning_rate": 1.8816631130063966e-05, "loss": 2.0116, "step": 361 }, { "epoch": 2.896, "grad_norm": 52.703922271728516, "learning_rate": 1.8869936034115142e-05, "loss": 2.3233, "step": 362 }, { "epoch": 2.904, "grad_norm": 173.81060791015625, "learning_rate": 1.892324093816631e-05, "loss": 1.8975, "step": 363 }, { "epoch": 2.912, "grad_norm": 61.005435943603516, "learning_rate": 1.8976545842217487e-05, "loss": 2.5226, "step": 364 }, { "epoch": 2.92, "grad_norm": 37.743988037109375, "learning_rate": 1.9029850746268656e-05, "loss": 2.008, "step": 365 }, { "epoch": 2.928, "grad_norm": 116.50171661376953, "learning_rate": 1.9083155650319832e-05, "loss": 2.2233, "step": 366 }, { "epoch": 2.936, "grad_norm": 47.62931442260742, "learning_rate": 1.9136460554371e-05, "loss": 3.421, "step": 367 }, { "epoch": 2.944, "grad_norm": 100.64972686767578, "learning_rate": 1.9189765458422178e-05, "loss": 2.7196, "step": 368 }, { "epoch": 2.952, "grad_norm": 49.90333938598633, "learning_rate": 1.9243070362473347e-05, "loss": 2.0727, "step": 369 }, { "epoch": 2.96, "grad_norm": 71.5282974243164, "learning_rate": 1.9296375266524523e-05, "loss": 2.042, "step": 370 }, { "epoch": 2.968, "grad_norm": 62.33525085449219, "learning_rate": 1.9349680170575692e-05, "loss": 2.1646, "step": 371 }, { "epoch": 2.976, "grad_norm": 94.84945678710938, "learning_rate": 1.9402985074626868e-05, "loss": 3.024, "step": 372 }, { "epoch": 2.984, "grad_norm": 53.67182159423828, "learning_rate": 1.9456289978678037e-05, "loss": 1.8143, "step": 373 }, { "epoch": 2.992, "grad_norm": 159.84988403320312, "learning_rate": 1.9509594882729213e-05, "loss": 2.155, "step": 374 }, { "epoch": 3.0, "grad_norm": 146.0228271484375, "learning_rate": 1.9562899786780383e-05, "loss": 2.262, "step": 375 }, { "epoch": 3.0, "eval_loss": 2.4158928394317627, "eval_map": 0.0294, "eval_map_50": 0.0592, "eval_map_75": 0.0245, "eval_map_Coverall": 0.1308, "eval_map_Face_Shield": 0.0, "eval_map_Gloves": 0.0024, "eval_map_Goggles": 0.0, "eval_map_Mask": 0.0138, "eval_map_large": 0.0284, "eval_map_medium": 0.0064, "eval_map_small": 0.0002, "eval_mar_1": 0.051, "eval_mar_10": 0.1389, "eval_mar_100": 0.1727, "eval_mar_100_Coverall": 0.5867, "eval_mar_100_Face_Shield": 0.0, "eval_mar_100_Gloves": 0.1016, "eval_mar_100_Goggles": 0.0, "eval_mar_100_Mask": 0.175, "eval_mar_large": 0.1628, "eval_mar_medium": 0.0846, "eval_mar_small": 0.0026, "eval_runtime": 1.8223, "eval_samples_per_second": 15.914, "eval_steps_per_second": 1.098, "step": 375 }, { "epoch": 3.008, "grad_norm": 101.75780487060547, "learning_rate": 1.961620469083156e-05, "loss": 2.0698, "step": 376 }, { "epoch": 3.016, "grad_norm": 118.74386596679688, "learning_rate": 1.9669509594882728e-05, "loss": 2.084, "step": 377 }, { "epoch": 3.024, "grad_norm": 60.56581497192383, "learning_rate": 1.9722814498933904e-05, "loss": 2.0854, "step": 378 }, { "epoch": 3.032, "grad_norm": 58.955970764160156, "learning_rate": 1.9776119402985073e-05, "loss": 2.291, "step": 379 }, { "epoch": 3.04, "grad_norm": 73.73709869384766, "learning_rate": 1.982942430703625e-05, "loss": 2.2901, "step": 380 }, { "epoch": 3.048, "grad_norm": 85.03562927246094, "learning_rate": 1.988272921108742e-05, "loss": 2.1454, "step": 381 }, { "epoch": 3.056, "grad_norm": 81.3743667602539, "learning_rate": 1.9936034115138594e-05, "loss": 1.4902, "step": 382 }, { "epoch": 3.064, "grad_norm": 24.32448387145996, "learning_rate": 1.9989339019189764e-05, "loss": 3.5255, "step": 383 }, { "epoch": 3.072, "grad_norm": 44.047061920166016, "learning_rate": 2.004264392324094e-05, "loss": 2.3742, "step": 384 }, { "epoch": 3.08, "grad_norm": 40.302730560302734, "learning_rate": 2.0095948827292112e-05, "loss": 1.9895, "step": 385 }, { "epoch": 3.088, "grad_norm": 43.920650482177734, "learning_rate": 2.0149253731343285e-05, "loss": 2.2819, "step": 386 }, { "epoch": 3.096, "grad_norm": 65.3396224975586, "learning_rate": 2.0202558635394458e-05, "loss": 2.1041, "step": 387 }, { "epoch": 3.104, "grad_norm": 61.35914993286133, "learning_rate": 2.025586353944563e-05, "loss": 2.0276, "step": 388 }, { "epoch": 3.112, "grad_norm": 124.42874908447266, "learning_rate": 2.0309168443496803e-05, "loss": 2.0106, "step": 389 }, { "epoch": 3.12, "grad_norm": 69.69377899169922, "learning_rate": 2.0362473347547976e-05, "loss": 1.8163, "step": 390 }, { "epoch": 3.128, "grad_norm": 44.3976936340332, "learning_rate": 2.0415778251599148e-05, "loss": 2.2435, "step": 391 }, { "epoch": 3.136, "grad_norm": 166.2566680908203, "learning_rate": 2.046908315565032e-05, "loss": 2.4188, "step": 392 }, { "epoch": 3.144, "grad_norm": 120.01055908203125, "learning_rate": 2.0522388059701493e-05, "loss": 2.1133, "step": 393 }, { "epoch": 3.152, "grad_norm": 52.2386360168457, "learning_rate": 2.0575692963752666e-05, "loss": 2.0885, "step": 394 }, { "epoch": 3.16, "grad_norm": 74.59262084960938, "learning_rate": 2.062899786780384e-05, "loss": 2.0558, "step": 395 }, { "epoch": 3.168, "grad_norm": 32.89023971557617, "learning_rate": 2.068230277185501e-05, "loss": 2.4027, "step": 396 }, { "epoch": 3.176, "grad_norm": 75.62948608398438, "learning_rate": 2.0735607675906184e-05, "loss": 2.2644, "step": 397 }, { "epoch": 3.184, "grad_norm": 52.67316818237305, "learning_rate": 2.0788912579957357e-05, "loss": 1.992, "step": 398 }, { "epoch": 3.192, "grad_norm": 40.93146514892578, "learning_rate": 2.084221748400853e-05, "loss": 3.08, "step": 399 }, { "epoch": 3.2, "grad_norm": 32.7237663269043, "learning_rate": 2.0895522388059702e-05, "loss": 1.8428, "step": 400 }, { "epoch": 3.208, "grad_norm": 26.644878387451172, "learning_rate": 2.0948827292110874e-05, "loss": 2.3187, "step": 401 }, { "epoch": 3.216, "grad_norm": 37.30282974243164, "learning_rate": 2.1002132196162047e-05, "loss": 1.455, "step": 402 }, { "epoch": 3.224, "grad_norm": 58.41122055053711, "learning_rate": 2.105543710021322e-05, "loss": 1.7112, "step": 403 }, { "epoch": 3.232, "grad_norm": 52.45249557495117, "learning_rate": 2.1108742004264392e-05, "loss": 2.1047, "step": 404 }, { "epoch": 3.24, "grad_norm": 67.95254516601562, "learning_rate": 2.1162046908315565e-05, "loss": 1.8307, "step": 405 }, { "epoch": 3.248, "grad_norm": 55.19281768798828, "learning_rate": 2.1215351812366738e-05, "loss": 1.6647, "step": 406 }, { "epoch": 3.2560000000000002, "grad_norm": 58.96752166748047, "learning_rate": 2.126865671641791e-05, "loss": 2.3322, "step": 407 }, { "epoch": 3.2640000000000002, "grad_norm": 83.07891845703125, "learning_rate": 2.1321961620469083e-05, "loss": 1.5633, "step": 408 }, { "epoch": 3.2720000000000002, "grad_norm": 60.541072845458984, "learning_rate": 2.1375266524520256e-05, "loss": 1.6818, "step": 409 }, { "epoch": 3.2800000000000002, "grad_norm": 45.087886810302734, "learning_rate": 2.1428571428571428e-05, "loss": 2.0799, "step": 410 }, { "epoch": 3.288, "grad_norm": 42.2776985168457, "learning_rate": 2.14818763326226e-05, "loss": 2.3814, "step": 411 }, { "epoch": 3.296, "grad_norm": 48.80410385131836, "learning_rate": 2.1535181236673773e-05, "loss": 2.2673, "step": 412 }, { "epoch": 3.304, "grad_norm": 40.62537384033203, "learning_rate": 2.158848614072495e-05, "loss": 2.0191, "step": 413 }, { "epoch": 3.312, "grad_norm": 67.03478240966797, "learning_rate": 2.164179104477612e-05, "loss": 1.8516, "step": 414 }, { "epoch": 3.32, "grad_norm": 45.354217529296875, "learning_rate": 2.1695095948827295e-05, "loss": 2.1101, "step": 415 }, { "epoch": 3.328, "grad_norm": 49.36673355102539, "learning_rate": 2.1748400852878467e-05, "loss": 2.0442, "step": 416 }, { "epoch": 3.336, "grad_norm": 72.0628662109375, "learning_rate": 2.180170575692964e-05, "loss": 2.1356, "step": 417 }, { "epoch": 3.344, "grad_norm": 50.46900939941406, "learning_rate": 2.1855010660980813e-05, "loss": 1.7259, "step": 418 }, { "epoch": 3.352, "grad_norm": 58.2097282409668, "learning_rate": 2.1908315565031985e-05, "loss": 1.8808, "step": 419 }, { "epoch": 3.36, "grad_norm": 63.63799285888672, "learning_rate": 2.1961620469083158e-05, "loss": 2.118, "step": 420 }, { "epoch": 3.368, "grad_norm": 104.8059310913086, "learning_rate": 2.201492537313433e-05, "loss": 2.5109, "step": 421 }, { "epoch": 3.376, "grad_norm": 42.35725784301758, "learning_rate": 2.2068230277185503e-05, "loss": 1.973, "step": 422 }, { "epoch": 3.384, "grad_norm": 90.13306427001953, "learning_rate": 2.2121535181236676e-05, "loss": 2.152, "step": 423 }, { "epoch": 3.392, "grad_norm": 52.2588005065918, "learning_rate": 2.217484008528785e-05, "loss": 2.8075, "step": 424 }, { "epoch": 3.4, "grad_norm": 79.99176788330078, "learning_rate": 2.222814498933902e-05, "loss": 1.5668, "step": 425 }, { "epoch": 3.408, "grad_norm": 30.49700164794922, "learning_rate": 2.2281449893390194e-05, "loss": 1.9115, "step": 426 }, { "epoch": 3.416, "grad_norm": 40.82845687866211, "learning_rate": 2.2334754797441366e-05, "loss": 2.1013, "step": 427 }, { "epoch": 3.424, "grad_norm": 32.535037994384766, "learning_rate": 2.238805970149254e-05, "loss": 1.8951, "step": 428 }, { "epoch": 3.432, "grad_norm": 488.69403076171875, "learning_rate": 2.244136460554371e-05, "loss": 2.2242, "step": 429 }, { "epoch": 3.44, "grad_norm": 28.62084197998047, "learning_rate": 2.2494669509594884e-05, "loss": 2.2226, "step": 430 }, { "epoch": 3.448, "grad_norm": 51.504703521728516, "learning_rate": 2.2547974413646057e-05, "loss": 2.1908, "step": 431 }, { "epoch": 3.456, "grad_norm": 265.2933044433594, "learning_rate": 2.260127931769723e-05, "loss": 2.5619, "step": 432 }, { "epoch": 3.464, "grad_norm": 67.91999053955078, "learning_rate": 2.2654584221748402e-05, "loss": 1.9919, "step": 433 }, { "epoch": 3.472, "grad_norm": 71.37107849121094, "learning_rate": 2.2707889125799575e-05, "loss": 1.9814, "step": 434 }, { "epoch": 3.48, "grad_norm": 129.96759033203125, "learning_rate": 2.2761194029850747e-05, "loss": 2.1298, "step": 435 }, { "epoch": 3.488, "grad_norm": 43.98136520385742, "learning_rate": 2.281449893390192e-05, "loss": 1.7564, "step": 436 }, { "epoch": 3.496, "grad_norm": 98.01622772216797, "learning_rate": 2.2867803837953093e-05, "loss": 1.9787, "step": 437 }, { "epoch": 3.504, "grad_norm": 46.8730354309082, "learning_rate": 2.2921108742004265e-05, "loss": 2.3151, "step": 438 }, { "epoch": 3.512, "grad_norm": 57.47915267944336, "learning_rate": 2.2974413646055438e-05, "loss": 2.7678, "step": 439 }, { "epoch": 3.52, "grad_norm": 65.00096893310547, "learning_rate": 2.302771855010661e-05, "loss": 2.0249, "step": 440 }, { "epoch": 3.528, "grad_norm": 44.42772674560547, "learning_rate": 2.3081023454157783e-05, "loss": 1.5484, "step": 441 }, { "epoch": 3.536, "grad_norm": 81.29985046386719, "learning_rate": 2.3134328358208956e-05, "loss": 1.6842, "step": 442 }, { "epoch": 3.544, "grad_norm": 30.51691436767578, "learning_rate": 2.318763326226013e-05, "loss": 2.0289, "step": 443 }, { "epoch": 3.552, "grad_norm": 37.00188064575195, "learning_rate": 2.32409381663113e-05, "loss": 2.3743, "step": 444 }, { "epoch": 3.56, "grad_norm": 90.2144546508789, "learning_rate": 2.3294243070362474e-05, "loss": 1.777, "step": 445 }, { "epoch": 3.568, "grad_norm": 117.61978912353516, "learning_rate": 2.3347547974413646e-05, "loss": 1.6184, "step": 446 }, { "epoch": 3.576, "grad_norm": 348.5857238769531, "learning_rate": 2.3400852878464822e-05, "loss": 1.8289, "step": 447 }, { "epoch": 3.584, "grad_norm": 41.86188507080078, "learning_rate": 2.345415778251599e-05, "loss": 1.6997, "step": 448 }, { "epoch": 3.592, "grad_norm": 37.16120147705078, "learning_rate": 2.3507462686567168e-05, "loss": 1.9718, "step": 449 }, { "epoch": 3.6, "grad_norm": 53.8717155456543, "learning_rate": 2.3560767590618337e-05, "loss": 1.9396, "step": 450 }, { "epoch": 3.608, "grad_norm": 49.46864318847656, "learning_rate": 2.3614072494669513e-05, "loss": 1.9602, "step": 451 }, { "epoch": 3.616, "grad_norm": 68.87065887451172, "learning_rate": 2.3667377398720682e-05, "loss": 2.7811, "step": 452 }, { "epoch": 3.624, "grad_norm": 55.349483489990234, "learning_rate": 2.3720682302771858e-05, "loss": 1.6945, "step": 453 }, { "epoch": 3.632, "grad_norm": 367.5556335449219, "learning_rate": 2.3773987206823027e-05, "loss": 1.5386, "step": 454 }, { "epoch": 3.64, "grad_norm": 69.410400390625, "learning_rate": 2.3827292110874203e-05, "loss": 1.9809, "step": 455 }, { "epoch": 3.648, "grad_norm": 69.57637786865234, "learning_rate": 2.3880597014925373e-05, "loss": 2.2286, "step": 456 }, { "epoch": 3.656, "grad_norm": 121.442626953125, "learning_rate": 2.393390191897655e-05, "loss": 1.9537, "step": 457 }, { "epoch": 3.664, "grad_norm": 36.72260284423828, "learning_rate": 2.3987206823027718e-05, "loss": 2.3149, "step": 458 }, { "epoch": 3.672, "grad_norm": 87.63125610351562, "learning_rate": 2.4040511727078894e-05, "loss": 1.9124, "step": 459 }, { "epoch": 3.68, "grad_norm": 56.9037971496582, "learning_rate": 2.4093816631130063e-05, "loss": 2.0174, "step": 460 }, { "epoch": 3.6879999999999997, "grad_norm": 47.20398712158203, "learning_rate": 2.414712153518124e-05, "loss": 2.4791, "step": 461 }, { "epoch": 3.6959999999999997, "grad_norm": 40.551822662353516, "learning_rate": 2.420042643923241e-05, "loss": 2.3819, "step": 462 }, { "epoch": 3.7039999999999997, "grad_norm": 83.71997833251953, "learning_rate": 2.4253731343283584e-05, "loss": 2.4351, "step": 463 }, { "epoch": 3.7119999999999997, "grad_norm": 40.414546966552734, "learning_rate": 2.4307036247334754e-05, "loss": 1.9891, "step": 464 }, { "epoch": 3.7199999999999998, "grad_norm": 75.16718292236328, "learning_rate": 2.436034115138593e-05, "loss": 2.9922, "step": 465 }, { "epoch": 3.7279999999999998, "grad_norm": 92.5959243774414, "learning_rate": 2.44136460554371e-05, "loss": 1.8834, "step": 466 }, { "epoch": 3.7359999999999998, "grad_norm": 91.39102172851562, "learning_rate": 2.4466950959488275e-05, "loss": 2.2147, "step": 467 }, { "epoch": 3.7439999999999998, "grad_norm": 66.4179916381836, "learning_rate": 2.4520255863539444e-05, "loss": 2.654, "step": 468 }, { "epoch": 3.752, "grad_norm": 272.90264892578125, "learning_rate": 2.457356076759062e-05, "loss": 1.9136, "step": 469 }, { "epoch": 3.76, "grad_norm": 47.681819915771484, "learning_rate": 2.4626865671641793e-05, "loss": 1.9328, "step": 470 }, { "epoch": 3.768, "grad_norm": 163.2669677734375, "learning_rate": 2.4680170575692965e-05, "loss": 2.3275, "step": 471 }, { "epoch": 3.776, "grad_norm": 105.56694030761719, "learning_rate": 2.4733475479744138e-05, "loss": 2.4249, "step": 472 }, { "epoch": 3.784, "grad_norm": 147.8690643310547, "learning_rate": 2.478678038379531e-05, "loss": 1.9867, "step": 473 }, { "epoch": 3.792, "grad_norm": 54.03384017944336, "learning_rate": 2.4840085287846483e-05, "loss": 1.9425, "step": 474 }, { "epoch": 3.8, "grad_norm": 217.9578094482422, "learning_rate": 2.4893390191897656e-05, "loss": 2.1597, "step": 475 }, { "epoch": 3.808, "grad_norm": 355.696533203125, "learning_rate": 2.494669509594883e-05, "loss": 1.7917, "step": 476 }, { "epoch": 3.816, "grad_norm": 95.96643829345703, "learning_rate": 2.5e-05, "loss": 2.0128, "step": 477 }, { "epoch": 3.824, "grad_norm": 140.10513305664062, "learning_rate": 2.5053304904051174e-05, "loss": 2.0886, "step": 478 }, { "epoch": 3.832, "grad_norm": 48.50007629394531, "learning_rate": 2.510660980810235e-05, "loss": 1.8981, "step": 479 }, { "epoch": 3.84, "grad_norm": 79.29582977294922, "learning_rate": 2.515991471215352e-05, "loss": 1.9332, "step": 480 }, { "epoch": 3.848, "grad_norm": 73.7322998046875, "learning_rate": 2.5213219616204692e-05, "loss": 2.446, "step": 481 }, { "epoch": 3.856, "grad_norm": 93.36573028564453, "learning_rate": 2.5266524520255864e-05, "loss": 1.8127, "step": 482 }, { "epoch": 3.864, "grad_norm": 59.118560791015625, "learning_rate": 2.531982942430704e-05, "loss": 2.3931, "step": 483 }, { "epoch": 3.872, "grad_norm": 64.02200317382812, "learning_rate": 2.537313432835821e-05, "loss": 2.0424, "step": 484 }, { "epoch": 3.88, "grad_norm": 60.51904296875, "learning_rate": 2.5426439232409382e-05, "loss": 1.6386, "step": 485 }, { "epoch": 3.888, "grad_norm": 45.91560745239258, "learning_rate": 2.5479744136460555e-05, "loss": 1.8508, "step": 486 }, { "epoch": 3.896, "grad_norm": 100.04412078857422, "learning_rate": 2.553304904051173e-05, "loss": 1.8745, "step": 487 }, { "epoch": 3.904, "grad_norm": 121.10009002685547, "learning_rate": 2.5586353944562904e-05, "loss": 1.8478, "step": 488 }, { "epoch": 3.912, "grad_norm": 487.3673095703125, "learning_rate": 2.5639658848614073e-05, "loss": 1.9153, "step": 489 }, { "epoch": 3.92, "grad_norm": 72.91594696044922, "learning_rate": 2.5692963752665245e-05, "loss": 1.6939, "step": 490 }, { "epoch": 3.928, "grad_norm": 160.2957763671875, "learning_rate": 2.574626865671642e-05, "loss": 1.9313, "step": 491 }, { "epoch": 3.936, "grad_norm": 69.46753692626953, "learning_rate": 2.5799573560767594e-05, "loss": 3.0829, "step": 492 }, { "epoch": 3.944, "grad_norm": 49.235252380371094, "learning_rate": 2.5852878464818763e-05, "loss": 2.7015, "step": 493 }, { "epoch": 3.952, "grad_norm": 30.935379028320312, "learning_rate": 2.5906183368869936e-05, "loss": 2.082, "step": 494 }, { "epoch": 3.96, "grad_norm": 120.84440612792969, "learning_rate": 2.5959488272921112e-05, "loss": 2.1394, "step": 495 }, { "epoch": 3.968, "grad_norm": 200.997802734375, "learning_rate": 2.6012793176972285e-05, "loss": 1.98, "step": 496 }, { "epoch": 3.976, "grad_norm": 131.88218688964844, "learning_rate": 2.6066098081023454e-05, "loss": 2.7392, "step": 497 }, { "epoch": 3.984, "grad_norm": 137.05845642089844, "learning_rate": 2.6119402985074626e-05, "loss": 1.9689, "step": 498 }, { "epoch": 3.992, "grad_norm": 38.30783462524414, "learning_rate": 2.6172707889125802e-05, "loss": 2.4766, "step": 499 }, { "epoch": 4.0, "grad_norm": 51.06193542480469, "learning_rate": 2.6226012793176975e-05, "loss": 1.5907, "step": 500 }, { "epoch": 4.0, "eval_loss": 2.409731149673462, "eval_map": 0.0357, "eval_map_50": 0.067, "eval_map_75": 0.0318, "eval_map_Coverall": 0.1471, "eval_map_Face_Shield": 0.0, "eval_map_Gloves": 0.0197, "eval_map_Goggles": 0.0, "eval_map_Mask": 0.0119, "eval_map_large": 0.0309, "eval_map_medium": 0.0089, "eval_map_small": 0.0318, "eval_mar_1": 0.0539, "eval_mar_10": 0.1606, "eval_mar_100": 0.222, "eval_mar_100_Coverall": 0.6222, "eval_mar_100_Face_Shield": 0.0, "eval_mar_100_Gloves": 0.2361, "eval_mar_100_Goggles": 0.0, "eval_mar_100_Mask": 0.2519, "eval_mar_large": 0.1722, "eval_mar_medium": 0.1285, "eval_mar_small": 0.0755, "eval_runtime": 1.8238, "eval_samples_per_second": 15.901, "eval_steps_per_second": 1.097, "step": 500 }, { "epoch": 4.008, "grad_norm": 78.92124938964844, "learning_rate": 2.6279317697228144e-05, "loss": 2.373, "step": 501 }, { "epoch": 4.016, "grad_norm": 163.3313446044922, "learning_rate": 2.6332622601279317e-05, "loss": 1.6469, "step": 502 }, { "epoch": 4.024, "grad_norm": 64.80167388916016, "learning_rate": 2.6385927505330493e-05, "loss": 1.6118, "step": 503 }, { "epoch": 4.032, "grad_norm": 193.38185119628906, "learning_rate": 2.6439232409381666e-05, "loss": 2.0134, "step": 504 }, { "epoch": 4.04, "grad_norm": 31.162826538085938, "learning_rate": 2.6492537313432835e-05, "loss": 2.1419, "step": 505 }, { "epoch": 4.048, "grad_norm": 48.26814651489258, "learning_rate": 2.6545842217484007e-05, "loss": 2.1823, "step": 506 }, { "epoch": 4.056, "grad_norm": 44.28972625732422, "learning_rate": 2.6599147121535184e-05, "loss": 2.4768, "step": 507 }, { "epoch": 4.064, "grad_norm": 45.70722961425781, "learning_rate": 2.6652452025586356e-05, "loss": 2.3743, "step": 508 }, { "epoch": 4.072, "grad_norm": 74.1947021484375, "learning_rate": 2.6705756929637525e-05, "loss": 2.2563, "step": 509 }, { "epoch": 4.08, "grad_norm": 53.10287857055664, "learning_rate": 2.6759061833688705e-05, "loss": 1.8558, "step": 510 }, { "epoch": 4.088, "grad_norm": 122.7011947631836, "learning_rate": 2.6812366737739874e-05, "loss": 1.7898, "step": 511 }, { "epoch": 4.096, "grad_norm": 81.71632385253906, "learning_rate": 2.6865671641791047e-05, "loss": 1.8906, "step": 512 }, { "epoch": 4.104, "grad_norm": 50.914581298828125, "learning_rate": 2.6918976545842216e-05, "loss": 1.7561, "step": 513 }, { "epoch": 4.112, "grad_norm": 213.11521911621094, "learning_rate": 2.6972281449893395e-05, "loss": 2.2969, "step": 514 }, { "epoch": 4.12, "grad_norm": 64.750732421875, "learning_rate": 2.7025586353944565e-05, "loss": 2.1357, "step": 515 }, { "epoch": 4.128, "grad_norm": 39.834659576416016, "learning_rate": 2.7078891257995737e-05, "loss": 3.2809, "step": 516 }, { "epoch": 4.136, "grad_norm": 230.9078826904297, "learning_rate": 2.7132196162046906e-05, "loss": 2.0321, "step": 517 }, { "epoch": 4.144, "grad_norm": 65.20975494384766, "learning_rate": 2.7185501066098086e-05, "loss": 2.1865, "step": 518 }, { "epoch": 4.152, "grad_norm": 32.31244659423828, "learning_rate": 2.7238805970149255e-05, "loss": 2.4192, "step": 519 }, { "epoch": 4.16, "grad_norm": 69.2395248413086, "learning_rate": 2.7292110874200428e-05, "loss": 2.3756, "step": 520 }, { "epoch": 4.168, "grad_norm": 66.05255126953125, "learning_rate": 2.7345415778251597e-05, "loss": 1.9835, "step": 521 }, { "epoch": 4.176, "grad_norm": 35.04376983642578, "learning_rate": 2.7398720682302776e-05, "loss": 2.0558, "step": 522 }, { "epoch": 4.184, "grad_norm": 46.90531539916992, "learning_rate": 2.7452025586353946e-05, "loss": 2.4171, "step": 523 }, { "epoch": 4.192, "grad_norm": 53.642574310302734, "learning_rate": 2.7505330490405118e-05, "loss": 2.0643, "step": 524 }, { "epoch": 4.2, "grad_norm": 60.9940185546875, "learning_rate": 2.7558635394456287e-05, "loss": 1.8913, "step": 525 }, { "epoch": 4.208, "grad_norm": 167.7407684326172, "learning_rate": 2.7611940298507467e-05, "loss": 1.928, "step": 526 }, { "epoch": 4.216, "grad_norm": 61.571327209472656, "learning_rate": 2.7665245202558636e-05, "loss": 2.3401, "step": 527 }, { "epoch": 4.224, "grad_norm": 39.09475326538086, "learning_rate": 2.771855010660981e-05, "loss": 1.8917, "step": 528 }, { "epoch": 4.232, "grad_norm": 77.27676391601562, "learning_rate": 2.777185501066098e-05, "loss": 1.9161, "step": 529 }, { "epoch": 4.24, "grad_norm": 31.990890502929688, "learning_rate": 2.7825159914712157e-05, "loss": 2.1289, "step": 530 }, { "epoch": 4.248, "grad_norm": 88.22757720947266, "learning_rate": 2.7878464818763327e-05, "loss": 1.8099, "step": 531 }, { "epoch": 4.256, "grad_norm": 33.27095031738281, "learning_rate": 2.79317697228145e-05, "loss": 1.6132, "step": 532 }, { "epoch": 4.264, "grad_norm": 92.78853607177734, "learning_rate": 2.7985074626865672e-05, "loss": 1.7231, "step": 533 }, { "epoch": 4.272, "grad_norm": 204.75274658203125, "learning_rate": 2.8038379530916848e-05, "loss": 1.8921, "step": 534 }, { "epoch": 4.28, "grad_norm": 30.629472732543945, "learning_rate": 2.8091684434968017e-05, "loss": 2.5274, "step": 535 }, { "epoch": 4.288, "grad_norm": 95.90540313720703, "learning_rate": 2.814498933901919e-05, "loss": 2.0575, "step": 536 }, { "epoch": 4.296, "grad_norm": 50.19279098510742, "learning_rate": 2.8198294243070362e-05, "loss": 1.8366, "step": 537 }, { "epoch": 4.304, "grad_norm": 39.78487777709961, "learning_rate": 2.825159914712154e-05, "loss": 1.8261, "step": 538 }, { "epoch": 4.312, "grad_norm": 38.698360443115234, "learning_rate": 2.8304904051172708e-05, "loss": 2.3774, "step": 539 }, { "epoch": 4.32, "grad_norm": 100.81163787841797, "learning_rate": 2.835820895522388e-05, "loss": 1.9214, "step": 540 }, { "epoch": 4.328, "grad_norm": 42.979251861572266, "learning_rate": 2.8411513859275056e-05, "loss": 1.7499, "step": 541 }, { "epoch": 4.336, "grad_norm": 52.07554244995117, "learning_rate": 2.846481876332623e-05, "loss": 1.751, "step": 542 }, { "epoch": 4.344, "grad_norm": 67.0647201538086, "learning_rate": 2.85181236673774e-05, "loss": 2.38, "step": 543 }, { "epoch": 4.352, "grad_norm": 45.40851593017578, "learning_rate": 2.857142857142857e-05, "loss": 2.5535, "step": 544 }, { "epoch": 4.36, "grad_norm": 338.5509948730469, "learning_rate": 2.8624733475479747e-05, "loss": 2.0974, "step": 545 }, { "epoch": 4.368, "grad_norm": 107.16646575927734, "learning_rate": 2.867803837953092e-05, "loss": 1.8053, "step": 546 }, { "epoch": 4.376, "grad_norm": 47.97652816772461, "learning_rate": 2.8731343283582092e-05, "loss": 1.7522, "step": 547 }, { "epoch": 4.384, "grad_norm": 41.47624588012695, "learning_rate": 2.878464818763326e-05, "loss": 1.9397, "step": 548 }, { "epoch": 4.392, "grad_norm": 32.19429016113281, "learning_rate": 2.8837953091684437e-05, "loss": 1.855, "step": 549 }, { "epoch": 4.4, "grad_norm": 40.816749572753906, "learning_rate": 2.889125799573561e-05, "loss": 2.3475, "step": 550 }, { "epoch": 4.408, "grad_norm": 80.6207504272461, "learning_rate": 2.8944562899786783e-05, "loss": 2.1508, "step": 551 }, { "epoch": 4.416, "grad_norm": 46.41642761230469, "learning_rate": 2.8997867803837952e-05, "loss": 2.4884, "step": 552 }, { "epoch": 4.424, "grad_norm": 79.7581787109375, "learning_rate": 2.9051172707889128e-05, "loss": 2.155, "step": 553 }, { "epoch": 4.432, "grad_norm": 108.76033782958984, "learning_rate": 2.91044776119403e-05, "loss": 2.2457, "step": 554 }, { "epoch": 4.44, "grad_norm": 193.62548828125, "learning_rate": 2.9157782515991473e-05, "loss": 1.6664, "step": 555 }, { "epoch": 4.448, "grad_norm": 151.55755615234375, "learning_rate": 2.9211087420042642e-05, "loss": 2.26, "step": 556 }, { "epoch": 4.456, "grad_norm": 43.50374221801758, "learning_rate": 2.926439232409382e-05, "loss": 1.6502, "step": 557 }, { "epoch": 4.464, "grad_norm": 75.4773941040039, "learning_rate": 2.931769722814499e-05, "loss": 2.2028, "step": 558 }, { "epoch": 4.4719999999999995, "grad_norm": 83.87667846679688, "learning_rate": 2.9371002132196164e-05, "loss": 2.1331, "step": 559 }, { "epoch": 4.48, "grad_norm": 59.65290069580078, "learning_rate": 2.9424307036247333e-05, "loss": 2.2233, "step": 560 }, { "epoch": 4.4879999999999995, "grad_norm": 36.0268669128418, "learning_rate": 2.9477611940298512e-05, "loss": 3.4642, "step": 561 }, { "epoch": 4.496, "grad_norm": 75.47502136230469, "learning_rate": 2.953091684434968e-05, "loss": 1.5836, "step": 562 }, { "epoch": 4.504, "grad_norm": 83.73079681396484, "learning_rate": 2.9584221748400854e-05, "loss": 2.0656, "step": 563 }, { "epoch": 4.5120000000000005, "grad_norm": 37.63287353515625, "learning_rate": 2.9637526652452023e-05, "loss": 2.5036, "step": 564 }, { "epoch": 4.52, "grad_norm": 233.36463928222656, "learning_rate": 2.9690831556503203e-05, "loss": 1.9293, "step": 565 }, { "epoch": 4.5280000000000005, "grad_norm": 43.29168701171875, "learning_rate": 2.9744136460554372e-05, "loss": 2.0752, "step": 566 }, { "epoch": 4.536, "grad_norm": 36.14763641357422, "learning_rate": 2.9797441364605545e-05, "loss": 2.551, "step": 567 }, { "epoch": 4.5440000000000005, "grad_norm": 64.3434066772461, "learning_rate": 2.9850746268656714e-05, "loss": 2.6031, "step": 568 }, { "epoch": 4.552, "grad_norm": 41.7719612121582, "learning_rate": 2.9904051172707893e-05, "loss": 2.039, "step": 569 }, { "epoch": 4.5600000000000005, "grad_norm": 47.51850891113281, "learning_rate": 2.9957356076759063e-05, "loss": 1.9371, "step": 570 }, { "epoch": 4.568, "grad_norm": 75.76118469238281, "learning_rate": 3.0010660980810235e-05, "loss": 2.1011, "step": 571 }, { "epoch": 4.576, "grad_norm": 45.60563659667969, "learning_rate": 3.006396588486141e-05, "loss": 1.8765, "step": 572 }, { "epoch": 4.584, "grad_norm": 106.35596466064453, "learning_rate": 3.0117270788912584e-05, "loss": 1.8416, "step": 573 }, { "epoch": 4.592, "grad_norm": 57.2327766418457, "learning_rate": 3.0170575692963753e-05, "loss": 2.0625, "step": 574 }, { "epoch": 4.6, "grad_norm": 94.30432891845703, "learning_rate": 3.0223880597014926e-05, "loss": 1.5145, "step": 575 }, { "epoch": 4.608, "grad_norm": 134.71621704101562, "learning_rate": 3.0277185501066102e-05, "loss": 1.864, "step": 576 }, { "epoch": 4.616, "grad_norm": 37.14790344238281, "learning_rate": 3.0330490405117274e-05, "loss": 1.734, "step": 577 }, { "epoch": 4.624, "grad_norm": 56.77499008178711, "learning_rate": 3.0383795309168444e-05, "loss": 2.0031, "step": 578 }, { "epoch": 4.632, "grad_norm": 105.61327362060547, "learning_rate": 3.0437100213219616e-05, "loss": 1.7981, "step": 579 }, { "epoch": 4.64, "grad_norm": 146.5427703857422, "learning_rate": 3.0490405117270792e-05, "loss": 2.1855, "step": 580 }, { "epoch": 4.648, "grad_norm": 111.44434356689453, "learning_rate": 3.0543710021321965e-05, "loss": 1.8957, "step": 581 }, { "epoch": 4.656, "grad_norm": 57.72542953491211, "learning_rate": 3.059701492537314e-05, "loss": 2.3358, "step": 582 }, { "epoch": 4.664, "grad_norm": 29.861927032470703, "learning_rate": 3.0650319829424303e-05, "loss": 1.9751, "step": 583 }, { "epoch": 4.672, "grad_norm": 76.6773910522461, "learning_rate": 3.070362473347548e-05, "loss": 1.9111, "step": 584 }, { "epoch": 4.68, "grad_norm": 71.11172485351562, "learning_rate": 3.0756929637526656e-05, "loss": 2.1066, "step": 585 }, { "epoch": 4.688, "grad_norm": 64.58554077148438, "learning_rate": 3.081023454157783e-05, "loss": 1.4534, "step": 586 }, { "epoch": 4.696, "grad_norm": 94.48667907714844, "learning_rate": 3.0863539445628994e-05, "loss": 2.0958, "step": 587 }, { "epoch": 4.704, "grad_norm": 81.82279205322266, "learning_rate": 3.0916844349680173e-05, "loss": 2.1789, "step": 588 }, { "epoch": 4.712, "grad_norm": 54.03045654296875, "learning_rate": 3.0970149253731346e-05, "loss": 1.7891, "step": 589 }, { "epoch": 4.72, "grad_norm": 63.78518295288086, "learning_rate": 3.102345415778252e-05, "loss": 2.5812, "step": 590 }, { "epoch": 4.728, "grad_norm": 184.6761474609375, "learning_rate": 3.1076759061833685e-05, "loss": 1.6575, "step": 591 }, { "epoch": 4.736, "grad_norm": 50.69154739379883, "learning_rate": 3.1130063965884864e-05, "loss": 1.8839, "step": 592 }, { "epoch": 4.744, "grad_norm": 43.271305084228516, "learning_rate": 3.1183368869936037e-05, "loss": 1.6055, "step": 593 }, { "epoch": 4.752, "grad_norm": 43.16181182861328, "learning_rate": 3.123667377398721e-05, "loss": 1.6172, "step": 594 }, { "epoch": 4.76, "grad_norm": 66.91848754882812, "learning_rate": 3.128997867803838e-05, "loss": 2.6092, "step": 595 }, { "epoch": 4.768, "grad_norm": 68.38192749023438, "learning_rate": 3.1343283582089554e-05, "loss": 2.1827, "step": 596 }, { "epoch": 4.776, "grad_norm": 41.7916259765625, "learning_rate": 3.139658848614073e-05, "loss": 1.7072, "step": 597 }, { "epoch": 4.784, "grad_norm": 45.17950439453125, "learning_rate": 3.14498933901919e-05, "loss": 1.5859, "step": 598 }, { "epoch": 4.792, "grad_norm": 48.95494842529297, "learning_rate": 3.150319829424307e-05, "loss": 1.6935, "step": 599 }, { "epoch": 4.8, "grad_norm": 96.62737274169922, "learning_rate": 3.1556503198294245e-05, "loss": 1.752, "step": 600 }, { "epoch": 4.808, "grad_norm": 53.38425827026367, "learning_rate": 3.160980810234542e-05, "loss": 2.6325, "step": 601 }, { "epoch": 4.816, "grad_norm": 38.38624954223633, "learning_rate": 3.166311300639659e-05, "loss": 1.6939, "step": 602 }, { "epoch": 4.824, "grad_norm": 110.9060287475586, "learning_rate": 3.171641791044776e-05, "loss": 2.0923, "step": 603 }, { "epoch": 4.832, "grad_norm": 79.89983367919922, "learning_rate": 3.1769722814498935e-05, "loss": 2.1701, "step": 604 }, { "epoch": 4.84, "grad_norm": 46.5838737487793, "learning_rate": 3.182302771855011e-05, "loss": 2.0174, "step": 605 }, { "epoch": 4.848, "grad_norm": 82.09159851074219, "learning_rate": 3.187633262260128e-05, "loss": 1.9414, "step": 606 }, { "epoch": 4.856, "grad_norm": 76.01737213134766, "learning_rate": 3.1929637526652453e-05, "loss": 1.872, "step": 607 }, { "epoch": 4.864, "grad_norm": 150.5392303466797, "learning_rate": 3.1982942430703626e-05, "loss": 1.7877, "step": 608 }, { "epoch": 4.872, "grad_norm": 24.910207748413086, "learning_rate": 3.20362473347548e-05, "loss": 2.5205, "step": 609 }, { "epoch": 4.88, "grad_norm": 71.54961395263672, "learning_rate": 3.208955223880597e-05, "loss": 2.2231, "step": 610 }, { "epoch": 4.888, "grad_norm": 44.92699432373047, "learning_rate": 3.2142857142857144e-05, "loss": 1.8617, "step": 611 }, { "epoch": 4.896, "grad_norm": 114.77074432373047, "learning_rate": 3.2196162046908317e-05, "loss": 2.0493, "step": 612 }, { "epoch": 4.904, "grad_norm": 49.42886734008789, "learning_rate": 3.224946695095949e-05, "loss": 2.4211, "step": 613 }, { "epoch": 4.912, "grad_norm": 47.691307067871094, "learning_rate": 3.230277185501066e-05, "loss": 1.8826, "step": 614 }, { "epoch": 4.92, "grad_norm": 47.12765121459961, "learning_rate": 3.2356076759061834e-05, "loss": 1.3896, "step": 615 }, { "epoch": 4.928, "grad_norm": 40.33041763305664, "learning_rate": 3.240938166311301e-05, "loss": 2.0119, "step": 616 }, { "epoch": 4.936, "grad_norm": 28.012340545654297, "learning_rate": 3.246268656716418e-05, "loss": 1.7023, "step": 617 }, { "epoch": 4.944, "grad_norm": 29.206438064575195, "learning_rate": 3.251599147121535e-05, "loss": 1.7157, "step": 618 }, { "epoch": 4.952, "grad_norm": 56.74425506591797, "learning_rate": 3.2569296375266525e-05, "loss": 2.5133, "step": 619 }, { "epoch": 4.96, "grad_norm": 89.80335235595703, "learning_rate": 3.26226012793177e-05, "loss": 2.2665, "step": 620 }, { "epoch": 4.968, "grad_norm": 41.48575973510742, "learning_rate": 3.267590618336887e-05, "loss": 1.8519, "step": 621 }, { "epoch": 4.976, "grad_norm": 49.826290130615234, "learning_rate": 3.272921108742004e-05, "loss": 2.4788, "step": 622 }, { "epoch": 4.984, "grad_norm": 32.657318115234375, "learning_rate": 3.2782515991471215e-05, "loss": 2.0374, "step": 623 }, { "epoch": 4.992, "grad_norm": 73.67778015136719, "learning_rate": 3.283582089552239e-05, "loss": 2.8276, "step": 624 }, { "epoch": 5.0, "grad_norm": 39.476871490478516, "learning_rate": 3.288912579957356e-05, "loss": 2.0368, "step": 625 }, { "epoch": 5.0, "eval_loss": 2.137108087539673, "eval_map": 0.0544, "eval_map_50": 0.1131, "eval_map_75": 0.0527, "eval_map_Coverall": 0.2262, "eval_map_Face_Shield": 0.0, "eval_map_Gloves": 0.02, "eval_map_Goggles": 0.0, "eval_map_Mask": 0.0257, "eval_map_large": 0.0493, "eval_map_medium": 0.01, "eval_map_small": 0.0103, "eval_mar_1": 0.0734, "eval_mar_10": 0.1708, "eval_mar_100": 0.2096, "eval_mar_100_Coverall": 0.56, "eval_mar_100_Face_Shield": 0.0, "eval_mar_100_Gloves": 0.1902, "eval_mar_100_Goggles": 0.0, "eval_mar_100_Mask": 0.2981, "eval_mar_large": 0.2082, "eval_mar_medium": 0.1277, "eval_mar_small": 0.0427, "eval_runtime": 1.9816, "eval_samples_per_second": 14.635, "eval_steps_per_second": 1.009, "step": 625 }, { "epoch": 5.008, "grad_norm": 77.2724380493164, "learning_rate": 3.294243070362473e-05, "loss": 1.4627, "step": 626 }, { "epoch": 5.016, "grad_norm": 52.474815368652344, "learning_rate": 3.299573560767591e-05, "loss": 1.7037, "step": 627 }, { "epoch": 5.024, "grad_norm": 47.86642837524414, "learning_rate": 3.304904051172708e-05, "loss": 2.1299, "step": 628 }, { "epoch": 5.032, "grad_norm": 102.83904266357422, "learning_rate": 3.310234541577825e-05, "loss": 1.9951, "step": 629 }, { "epoch": 5.04, "grad_norm": 27.309385299682617, "learning_rate": 3.3155650319829424e-05, "loss": 1.8106, "step": 630 }, { "epoch": 5.048, "grad_norm": 136.66476440429688, "learning_rate": 3.32089552238806e-05, "loss": 1.5355, "step": 631 }, { "epoch": 5.056, "grad_norm": 55.03985595703125, "learning_rate": 3.326226012793177e-05, "loss": 1.5354, "step": 632 }, { "epoch": 5.064, "grad_norm": 77.5121841430664, "learning_rate": 3.331556503198294e-05, "loss": 1.9385, "step": 633 }, { "epoch": 5.072, "grad_norm": 65.01448059082031, "learning_rate": 3.336886993603412e-05, "loss": 1.9101, "step": 634 }, { "epoch": 5.08, "grad_norm": 46.9758186340332, "learning_rate": 3.3422174840085294e-05, "loss": 2.1015, "step": 635 }, { "epoch": 5.088, "grad_norm": 67.9593505859375, "learning_rate": 3.347547974413646e-05, "loss": 1.7911, "step": 636 }, { "epoch": 5.096, "grad_norm": 78.51792907714844, "learning_rate": 3.352878464818763e-05, "loss": 2.5986, "step": 637 }, { "epoch": 5.104, "grad_norm": 45.60038375854492, "learning_rate": 3.358208955223881e-05, "loss": 1.6751, "step": 638 }, { "epoch": 5.112, "grad_norm": 49.37856674194336, "learning_rate": 3.3635394456289984e-05, "loss": 1.847, "step": 639 }, { "epoch": 5.12, "grad_norm": 42.260597229003906, "learning_rate": 3.368869936034115e-05, "loss": 1.7399, "step": 640 }, { "epoch": 5.128, "grad_norm": 58.034053802490234, "learning_rate": 3.374200426439232e-05, "loss": 1.9809, "step": 641 }, { "epoch": 5.136, "grad_norm": 257.21783447265625, "learning_rate": 3.37953091684435e-05, "loss": 2.1562, "step": 642 }, { "epoch": 5.144, "grad_norm": 183.8196258544922, "learning_rate": 3.3848614072494675e-05, "loss": 2.1574, "step": 643 }, { "epoch": 5.152, "grad_norm": 33.86476135253906, "learning_rate": 3.390191897654584e-05, "loss": 1.7577, "step": 644 }, { "epoch": 5.16, "grad_norm": 43.74724197387695, "learning_rate": 3.395522388059701e-05, "loss": 1.8622, "step": 645 }, { "epoch": 5.168, "grad_norm": 91.45352172851562, "learning_rate": 3.400852878464819e-05, "loss": 2.0887, "step": 646 }, { "epoch": 5.176, "grad_norm": 95.76984405517578, "learning_rate": 3.4061833688699365e-05, "loss": 1.8492, "step": 647 }, { "epoch": 5.184, "grad_norm": 59.3028678894043, "learning_rate": 3.411513859275053e-05, "loss": 2.1199, "step": 648 }, { "epoch": 5.192, "grad_norm": 95.11681365966797, "learning_rate": 3.4168443496801704e-05, "loss": 2.0028, "step": 649 }, { "epoch": 5.2, "grad_norm": 49.99011993408203, "learning_rate": 3.422174840085288e-05, "loss": 2.1932, "step": 650 }, { "epoch": 5.208, "grad_norm": 62.891597747802734, "learning_rate": 3.4275053304904056e-05, "loss": 1.8263, "step": 651 }, { "epoch": 5.216, "grad_norm": 26.250972747802734, "learning_rate": 3.432835820895522e-05, "loss": 2.4872, "step": 652 }, { "epoch": 5.224, "grad_norm": 155.1195831298828, "learning_rate": 3.4381663113006394e-05, "loss": 1.8548, "step": 653 }, { "epoch": 5.232, "grad_norm": 61.7595100402832, "learning_rate": 3.4434968017057574e-05, "loss": 1.7865, "step": 654 }, { "epoch": 5.24, "grad_norm": 50.275909423828125, "learning_rate": 3.4488272921108746e-05, "loss": 2.2349, "step": 655 }, { "epoch": 5.248, "grad_norm": 82.25180053710938, "learning_rate": 3.454157782515991e-05, "loss": 2.036, "step": 656 }, { "epoch": 5.256, "grad_norm": 27.44087028503418, "learning_rate": 3.4594882729211085e-05, "loss": 1.8274, "step": 657 }, { "epoch": 5.264, "grad_norm": 87.32594299316406, "learning_rate": 3.4648187633262264e-05, "loss": 3.1711, "step": 658 }, { "epoch": 5.272, "grad_norm": 38.50464630126953, "learning_rate": 3.470149253731344e-05, "loss": 1.6409, "step": 659 }, { "epoch": 5.28, "grad_norm": 32.80443572998047, "learning_rate": 3.47547974413646e-05, "loss": 1.6888, "step": 660 }, { "epoch": 5.288, "grad_norm": 53.19151306152344, "learning_rate": 3.4808102345415775e-05, "loss": 2.0697, "step": 661 }, { "epoch": 5.296, "grad_norm": 40.454437255859375, "learning_rate": 3.4861407249466955e-05, "loss": 2.4773, "step": 662 }, { "epoch": 5.304, "grad_norm": 43.74908447265625, "learning_rate": 3.491471215351813e-05, "loss": 1.7664, "step": 663 }, { "epoch": 5.312, "grad_norm": 95.16156768798828, "learning_rate": 3.496801705756929e-05, "loss": 1.7177, "step": 664 }, { "epoch": 5.32, "grad_norm": 103.79588317871094, "learning_rate": 3.502132196162047e-05, "loss": 2.0591, "step": 665 }, { "epoch": 5.328, "grad_norm": 80.74327850341797, "learning_rate": 3.5074626865671645e-05, "loss": 2.1902, "step": 666 }, { "epoch": 5.336, "grad_norm": 137.12521362304688, "learning_rate": 3.512793176972282e-05, "loss": 1.7284, "step": 667 }, { "epoch": 5.344, "grad_norm": 27.114564895629883, "learning_rate": 3.518123667377399e-05, "loss": 2.0691, "step": 668 }, { "epoch": 5.352, "grad_norm": 108.91705322265625, "learning_rate": 3.523454157782516e-05, "loss": 1.6302, "step": 669 }, { "epoch": 5.36, "grad_norm": 31.324949264526367, "learning_rate": 3.5287846481876336e-05, "loss": 1.9884, "step": 670 }, { "epoch": 5.368, "grad_norm": 62.23483657836914, "learning_rate": 3.534115138592751e-05, "loss": 1.774, "step": 671 }, { "epoch": 5.376, "grad_norm": 52.00127410888672, "learning_rate": 3.539445628997868e-05, "loss": 2.0694, "step": 672 }, { "epoch": 5.384, "grad_norm": 24.834827423095703, "learning_rate": 3.5447761194029854e-05, "loss": 2.4418, "step": 673 }, { "epoch": 5.392, "grad_norm": 136.19894409179688, "learning_rate": 3.5501066098081026e-05, "loss": 2.093, "step": 674 }, { "epoch": 5.4, "grad_norm": 76.635498046875, "learning_rate": 3.55543710021322e-05, "loss": 2.1358, "step": 675 }, { "epoch": 5.408, "grad_norm": 36.83510208129883, "learning_rate": 3.560767590618337e-05, "loss": 2.2664, "step": 676 }, { "epoch": 5.416, "grad_norm": 50.86436080932617, "learning_rate": 3.5660980810234544e-05, "loss": 1.856, "step": 677 }, { "epoch": 5.424, "grad_norm": 54.988040924072266, "learning_rate": 3.571428571428572e-05, "loss": 1.9915, "step": 678 }, { "epoch": 5.432, "grad_norm": 336.77508544921875, "learning_rate": 3.576759061833689e-05, "loss": 2.0613, "step": 679 }, { "epoch": 5.44, "grad_norm": 52.69036865234375, "learning_rate": 3.582089552238806e-05, "loss": 2.0339, "step": 680 }, { "epoch": 5.448, "grad_norm": 1328.716064453125, "learning_rate": 3.5874200426439235e-05, "loss": 2.0383, "step": 681 }, { "epoch": 5.456, "grad_norm": 39.94824981689453, "learning_rate": 3.592750533049041e-05, "loss": 2.0516, "step": 682 }, { "epoch": 5.464, "grad_norm": 79.15857696533203, "learning_rate": 3.598081023454158e-05, "loss": 1.9078, "step": 683 }, { "epoch": 5.4719999999999995, "grad_norm": 73.6874008178711, "learning_rate": 3.603411513859275e-05, "loss": 1.5667, "step": 684 }, { "epoch": 5.48, "grad_norm": 134.41835021972656, "learning_rate": 3.6087420042643925e-05, "loss": 2.1766, "step": 685 }, { "epoch": 5.4879999999999995, "grad_norm": 60.275455474853516, "learning_rate": 3.61407249466951e-05, "loss": 2.1971, "step": 686 }, { "epoch": 5.496, "grad_norm": 209.80458068847656, "learning_rate": 3.619402985074627e-05, "loss": 2.4503, "step": 687 }, { "epoch": 5.504, "grad_norm": 101.4869613647461, "learning_rate": 3.624733475479744e-05, "loss": 2.4181, "step": 688 }, { "epoch": 5.5120000000000005, "grad_norm": 137.25277709960938, "learning_rate": 3.6300639658848616e-05, "loss": 2.5375, "step": 689 }, { "epoch": 5.52, "grad_norm": 63.87282943725586, "learning_rate": 3.635394456289979e-05, "loss": 1.7877, "step": 690 }, { "epoch": 5.5280000000000005, "grad_norm": 47.16462707519531, "learning_rate": 3.640724946695096e-05, "loss": 2.0803, "step": 691 }, { "epoch": 5.536, "grad_norm": 144.07559204101562, "learning_rate": 3.6460554371002134e-05, "loss": 1.8735, "step": 692 }, { "epoch": 5.5440000000000005, "grad_norm": 48.085693359375, "learning_rate": 3.6513859275053306e-05, "loss": 2.6871, "step": 693 }, { "epoch": 5.552, "grad_norm": 87.16508483886719, "learning_rate": 3.656716417910448e-05, "loss": 1.482, "step": 694 }, { "epoch": 5.5600000000000005, "grad_norm": 47.54373550415039, "learning_rate": 3.662046908315565e-05, "loss": 2.6355, "step": 695 }, { "epoch": 5.568, "grad_norm": 33.75539779663086, "learning_rate": 3.6673773987206824e-05, "loss": 2.4589, "step": 696 }, { "epoch": 5.576, "grad_norm": 92.1048355102539, "learning_rate": 3.6727078891258e-05, "loss": 2.1943, "step": 697 }, { "epoch": 5.584, "grad_norm": 101.07424926757812, "learning_rate": 3.678038379530917e-05, "loss": 2.0153, "step": 698 }, { "epoch": 5.592, "grad_norm": 160.40020751953125, "learning_rate": 3.683368869936034e-05, "loss": 1.833, "step": 699 }, { "epoch": 5.6, "grad_norm": 66.00830841064453, "learning_rate": 3.6886993603411515e-05, "loss": 1.6738, "step": 700 }, { "epoch": 5.608, "grad_norm": 54.3719596862793, "learning_rate": 3.694029850746269e-05, "loss": 1.783, "step": 701 }, { "epoch": 5.616, "grad_norm": 51.09929656982422, "learning_rate": 3.699360341151386e-05, "loss": 2.1409, "step": 702 }, { "epoch": 5.624, "grad_norm": 47.338844299316406, "learning_rate": 3.704690831556503e-05, "loss": 1.7934, "step": 703 }, { "epoch": 5.632, "grad_norm": 69.06346893310547, "learning_rate": 3.710021321961621e-05, "loss": 1.8595, "step": 704 }, { "epoch": 5.64, "grad_norm": 46.77880096435547, "learning_rate": 3.715351812366738e-05, "loss": 1.8368, "step": 705 }, { "epoch": 5.648, "grad_norm": 49.410316467285156, "learning_rate": 3.720682302771855e-05, "loss": 2.0657, "step": 706 }, { "epoch": 5.656, "grad_norm": 123.15914916992188, "learning_rate": 3.726012793176972e-05, "loss": 2.1033, "step": 707 }, { "epoch": 5.664, "grad_norm": 24.037742614746094, "learning_rate": 3.73134328358209e-05, "loss": 1.6847, "step": 708 }, { "epoch": 5.672, "grad_norm": 43.98957443237305, "learning_rate": 3.736673773987207e-05, "loss": 2.2717, "step": 709 }, { "epoch": 5.68, "grad_norm": 41.1609001159668, "learning_rate": 3.742004264392324e-05, "loss": 1.8317, "step": 710 }, { "epoch": 5.688, "grad_norm": 39.32685852050781, "learning_rate": 3.7473347547974414e-05, "loss": 2.6976, "step": 711 }, { "epoch": 5.696, "grad_norm": 45.331233978271484, "learning_rate": 3.752665245202559e-05, "loss": 1.7802, "step": 712 }, { "epoch": 5.704, "grad_norm": 142.27293395996094, "learning_rate": 3.757995735607676e-05, "loss": 2.356, "step": 713 }, { "epoch": 5.712, "grad_norm": 61.770938873291016, "learning_rate": 3.763326226012793e-05, "loss": 1.7427, "step": 714 }, { "epoch": 5.72, "grad_norm": 58.35750961303711, "learning_rate": 3.7686567164179104e-05, "loss": 1.5548, "step": 715 }, { "epoch": 5.728, "grad_norm": 69.06352996826172, "learning_rate": 3.7739872068230284e-05, "loss": 1.4915, "step": 716 }, { "epoch": 5.736, "grad_norm": 122.13623046875, "learning_rate": 3.779317697228145e-05, "loss": 1.923, "step": 717 }, { "epoch": 5.744, "grad_norm": 44.81269454956055, "learning_rate": 3.784648187633262e-05, "loss": 2.205, "step": 718 }, { "epoch": 5.752, "grad_norm": 39.850563049316406, "learning_rate": 3.7899786780383795e-05, "loss": 1.9967, "step": 719 }, { "epoch": 5.76, "grad_norm": 63.25294876098633, "learning_rate": 3.7953091684434974e-05, "loss": 1.9495, "step": 720 }, { "epoch": 5.768, "grad_norm": 23.52655029296875, "learning_rate": 3.800639658848614e-05, "loss": 2.1998, "step": 721 }, { "epoch": 5.776, "grad_norm": 36.4002799987793, "learning_rate": 3.805970149253731e-05, "loss": 1.8874, "step": 722 }, { "epoch": 5.784, "grad_norm": 51.90754699707031, "learning_rate": 3.8113006396588485e-05, "loss": 2.3515, "step": 723 }, { "epoch": 5.792, "grad_norm": 45.665863037109375, "learning_rate": 3.8166311300639665e-05, "loss": 1.7018, "step": 724 }, { "epoch": 5.8, "grad_norm": 47.398170471191406, "learning_rate": 3.821961620469083e-05, "loss": 1.5501, "step": 725 }, { "epoch": 5.808, "grad_norm": 56.4918212890625, "learning_rate": 3.8272921108742e-05, "loss": 3.5721, "step": 726 }, { "epoch": 5.816, "grad_norm": 35.87611770629883, "learning_rate": 3.8326226012793176e-05, "loss": 1.8283, "step": 727 }, { "epoch": 5.824, "grad_norm": 50.51455307006836, "learning_rate": 3.8379530916844355e-05, "loss": 1.5875, "step": 728 }, { "epoch": 5.832, "grad_norm": 35.856361389160156, "learning_rate": 3.843283582089552e-05, "loss": 2.0093, "step": 729 }, { "epoch": 5.84, "grad_norm": 117.65554809570312, "learning_rate": 3.8486140724946694e-05, "loss": 1.7601, "step": 730 }, { "epoch": 5.848, "grad_norm": 43.680179595947266, "learning_rate": 3.853944562899787e-05, "loss": 2.2472, "step": 731 }, { "epoch": 5.856, "grad_norm": 90.28701782226562, "learning_rate": 3.8592750533049046e-05, "loss": 2.5912, "step": 732 }, { "epoch": 5.864, "grad_norm": 42.63003158569336, "learning_rate": 3.864605543710021e-05, "loss": 1.6904, "step": 733 }, { "epoch": 5.872, "grad_norm": 33.32051467895508, "learning_rate": 3.8699360341151384e-05, "loss": 2.1069, "step": 734 }, { "epoch": 5.88, "grad_norm": 82.60679626464844, "learning_rate": 3.8752665245202564e-05, "loss": 2.1514, "step": 735 }, { "epoch": 5.888, "grad_norm": 62.447357177734375, "learning_rate": 3.8805970149253736e-05, "loss": 2.38, "step": 736 }, { "epoch": 5.896, "grad_norm": 198.05709838867188, "learning_rate": 3.88592750533049e-05, "loss": 2.2958, "step": 737 }, { "epoch": 5.904, "grad_norm": 235.97372436523438, "learning_rate": 3.8912579957356075e-05, "loss": 1.9596, "step": 738 }, { "epoch": 5.912, "grad_norm": 167.79348754882812, "learning_rate": 3.8965884861407254e-05, "loss": 2.033, "step": 739 }, { "epoch": 5.92, "grad_norm": 2165.9716796875, "learning_rate": 3.901918976545843e-05, "loss": 2.0979, "step": 740 }, { "epoch": 5.928, "grad_norm": 51.91420364379883, "learning_rate": 3.90724946695096e-05, "loss": 2.5746, "step": 741 }, { "epoch": 5.936, "grad_norm": 37.00518798828125, "learning_rate": 3.9125799573560765e-05, "loss": 2.1024, "step": 742 }, { "epoch": 5.944, "grad_norm": 97.52835083007812, "learning_rate": 3.9179104477611945e-05, "loss": 2.4316, "step": 743 }, { "epoch": 5.952, "grad_norm": 49.9329833984375, "learning_rate": 3.923240938166312e-05, "loss": 2.7627, "step": 744 }, { "epoch": 5.96, "grad_norm": 525.2008666992188, "learning_rate": 3.928571428571429e-05, "loss": 1.7328, "step": 745 }, { "epoch": 5.968, "grad_norm": 42.05510330200195, "learning_rate": 3.9339019189765456e-05, "loss": 2.035, "step": 746 }, { "epoch": 5.976, "grad_norm": 41.231746673583984, "learning_rate": 3.9392324093816635e-05, "loss": 2.0804, "step": 747 }, { "epoch": 5.984, "grad_norm": 62.54353332519531, "learning_rate": 3.944562899786781e-05, "loss": 2.195, "step": 748 }, { "epoch": 5.992, "grad_norm": 129.8985595703125, "learning_rate": 3.949893390191898e-05, "loss": 2.1179, "step": 749 }, { "epoch": 6.0, "grad_norm": 68.5892562866211, "learning_rate": 3.9552238805970146e-05, "loss": 2.059, "step": 750 }, { "epoch": 6.0, "eval_loss": 2.1225361824035645, "eval_map": 0.0576, "eval_map_50": 0.1422, "eval_map_75": 0.047, "eval_map_Coverall": 0.2292, "eval_map_Face_Shield": 0.0, "eval_map_Gloves": 0.0261, "eval_map_Goggles": 0.0, "eval_map_Mask": 0.0327, "eval_map_large": 0.0575, "eval_map_medium": 0.0123, "eval_map_small": 0.0244, "eval_mar_1": 0.0611, "eval_mar_10": 0.1728, "eval_mar_100": 0.222, "eval_mar_100_Coverall": 0.5556, "eval_mar_100_Face_Shield": 0.0, "eval_mar_100_Gloves": 0.2508, "eval_mar_100_Goggles": 0.0, "eval_mar_100_Mask": 0.3038, "eval_mar_large": 0.1913, "eval_mar_medium": 0.1389, "eval_mar_small": 0.0807, "eval_runtime": 1.8075, "eval_samples_per_second": 16.044, "eval_steps_per_second": 1.107, "step": 750 }, { "epoch": 6.008, "grad_norm": 37.05078125, "learning_rate": 3.9605543710021326e-05, "loss": 1.5481, "step": 751 }, { "epoch": 6.016, "grad_norm": 40.91301727294922, "learning_rate": 3.96588486140725e-05, "loss": 1.7168, "step": 752 }, { "epoch": 6.024, "grad_norm": 51.63536834716797, "learning_rate": 3.971215351812367e-05, "loss": 1.8921, "step": 753 }, { "epoch": 6.032, "grad_norm": 81.63351440429688, "learning_rate": 3.976545842217484e-05, "loss": 2.0167, "step": 754 }, { "epoch": 6.04, "grad_norm": 62.47822189331055, "learning_rate": 3.9818763326226016e-05, "loss": 1.9836, "step": 755 }, { "epoch": 6.048, "grad_norm": 93.62579345703125, "learning_rate": 3.987206823027719e-05, "loss": 2.3467, "step": 756 }, { "epoch": 6.056, "grad_norm": 30.823583602905273, "learning_rate": 3.992537313432836e-05, "loss": 1.7204, "step": 757 }, { "epoch": 6.064, "grad_norm": 93.74984741210938, "learning_rate": 3.997867803837953e-05, "loss": 2.6365, "step": 758 }, { "epoch": 6.072, "grad_norm": 33.29951858520508, "learning_rate": 4.003198294243071e-05, "loss": 1.7181, "step": 759 }, { "epoch": 6.08, "grad_norm": 84.72465515136719, "learning_rate": 4.008528784648188e-05, "loss": 1.7164, "step": 760 }, { "epoch": 6.088, "grad_norm": 38.03929901123047, "learning_rate": 4.013859275053305e-05, "loss": 2.2685, "step": 761 }, { "epoch": 6.096, "grad_norm": 42.80769348144531, "learning_rate": 4.0191897654584225e-05, "loss": 1.9481, "step": 762 }, { "epoch": 6.104, "grad_norm": 33.119468688964844, "learning_rate": 4.02452025586354e-05, "loss": 1.9704, "step": 763 }, { "epoch": 6.112, "grad_norm": 57.66070556640625, "learning_rate": 4.029850746268657e-05, "loss": 2.1374, "step": 764 }, { "epoch": 6.12, "grad_norm": 67.50090789794922, "learning_rate": 4.035181236673774e-05, "loss": 2.2068, "step": 765 }, { "epoch": 6.128, "grad_norm": 59.00775909423828, "learning_rate": 4.0405117270788915e-05, "loss": 1.7895, "step": 766 }, { "epoch": 6.136, "grad_norm": 87.56890106201172, "learning_rate": 4.045842217484009e-05, "loss": 1.4288, "step": 767 }, { "epoch": 6.144, "grad_norm": 66.95140075683594, "learning_rate": 4.051172707889126e-05, "loss": 1.6197, "step": 768 }, { "epoch": 6.152, "grad_norm": 51.41279983520508, "learning_rate": 4.056503198294243e-05, "loss": 1.5392, "step": 769 }, { "epoch": 6.16, "grad_norm": 136.06820678710938, "learning_rate": 4.0618336886993606e-05, "loss": 1.8578, "step": 770 }, { "epoch": 6.168, "grad_norm": 98.99730682373047, "learning_rate": 4.067164179104478e-05, "loss": 2.3759, "step": 771 }, { "epoch": 6.176, "grad_norm": 41.16776657104492, "learning_rate": 4.072494669509595e-05, "loss": 2.2127, "step": 772 }, { "epoch": 6.184, "grad_norm": 109.93490600585938, "learning_rate": 4.0778251599147124e-05, "loss": 1.6997, "step": 773 }, { "epoch": 6.192, "grad_norm": 50.32664489746094, "learning_rate": 4.0831556503198296e-05, "loss": 1.8953, "step": 774 }, { "epoch": 6.2, "grad_norm": 79.4347152709961, "learning_rate": 4.088486140724947e-05, "loss": 1.7544, "step": 775 }, { "epoch": 6.208, "grad_norm": 42.07555389404297, "learning_rate": 4.093816631130064e-05, "loss": 1.6255, "step": 776 }, { "epoch": 6.216, "grad_norm": 39.176536560058594, "learning_rate": 4.0991471215351814e-05, "loss": 1.5542, "step": 777 }, { "epoch": 6.224, "grad_norm": 93.9063720703125, "learning_rate": 4.104477611940299e-05, "loss": 1.9233, "step": 778 }, { "epoch": 6.232, "grad_norm": 70.66390991210938, "learning_rate": 4.109808102345416e-05, "loss": 1.902, "step": 779 }, { "epoch": 6.24, "grad_norm": 29.4874267578125, "learning_rate": 4.115138592750533e-05, "loss": 2.1428, "step": 780 }, { "epoch": 6.248, "grad_norm": 43.75629806518555, "learning_rate": 4.1204690831556505e-05, "loss": 2.9642, "step": 781 }, { "epoch": 6.256, "grad_norm": 33.31045913696289, "learning_rate": 4.125799573560768e-05, "loss": 1.8262, "step": 782 }, { "epoch": 6.264, "grad_norm": 78.85667419433594, "learning_rate": 4.131130063965885e-05, "loss": 2.3252, "step": 783 }, { "epoch": 6.272, "grad_norm": 36.19197463989258, "learning_rate": 4.136460554371002e-05, "loss": 2.6863, "step": 784 }, { "epoch": 6.28, "grad_norm": 468.0359802246094, "learning_rate": 4.1417910447761195e-05, "loss": 1.4422, "step": 785 }, { "epoch": 6.288, "grad_norm": 36.00908279418945, "learning_rate": 4.147121535181237e-05, "loss": 2.2255, "step": 786 }, { "epoch": 6.296, "grad_norm": 27.32050323486328, "learning_rate": 4.152452025586354e-05, "loss": 1.7495, "step": 787 }, { "epoch": 6.304, "grad_norm": 38.79612731933594, "learning_rate": 4.157782515991471e-05, "loss": 1.8373, "step": 788 }, { "epoch": 6.312, "grad_norm": 37.536109924316406, "learning_rate": 4.1631130063965886e-05, "loss": 2.0719, "step": 789 }, { "epoch": 6.32, "grad_norm": 33.2296257019043, "learning_rate": 4.168443496801706e-05, "loss": 2.1114, "step": 790 }, { "epoch": 6.328, "grad_norm": 28.226613998413086, "learning_rate": 4.173773987206823e-05, "loss": 1.7101, "step": 791 }, { "epoch": 6.336, "grad_norm": 26.928985595703125, "learning_rate": 4.1791044776119404e-05, "loss": 2.2318, "step": 792 }, { "epoch": 6.344, "grad_norm": 45.34580612182617, "learning_rate": 4.184434968017058e-05, "loss": 2.1115, "step": 793 }, { "epoch": 6.352, "grad_norm": 70.57929992675781, "learning_rate": 4.189765458422175e-05, "loss": 1.408, "step": 794 }, { "epoch": 6.36, "grad_norm": 31.665882110595703, "learning_rate": 4.195095948827292e-05, "loss": 2.0169, "step": 795 }, { "epoch": 6.368, "grad_norm": 91.7056655883789, "learning_rate": 4.2004264392324094e-05, "loss": 2.1807, "step": 796 }, { "epoch": 6.376, "grad_norm": 138.8243865966797, "learning_rate": 4.2057569296375274e-05, "loss": 1.4644, "step": 797 }, { "epoch": 6.384, "grad_norm": 48.531593322753906, "learning_rate": 4.211087420042644e-05, "loss": 1.4184, "step": 798 }, { "epoch": 6.392, "grad_norm": 48.0137825012207, "learning_rate": 4.216417910447761e-05, "loss": 2.1009, "step": 799 }, { "epoch": 6.4, "grad_norm": 51.559715270996094, "learning_rate": 4.2217484008528785e-05, "loss": 1.8211, "step": 800 }, { "epoch": 6.408, "grad_norm": 39.15354919433594, "learning_rate": 4.2270788912579964e-05, "loss": 2.445, "step": 801 }, { "epoch": 6.416, "grad_norm": 31.623315811157227, "learning_rate": 4.232409381663113e-05, "loss": 1.8177, "step": 802 }, { "epoch": 6.424, "grad_norm": 32.48640060424805, "learning_rate": 4.23773987206823e-05, "loss": 1.6097, "step": 803 }, { "epoch": 6.432, "grad_norm": 32.03213119506836, "learning_rate": 4.2430703624733475e-05, "loss": 2.1942, "step": 804 }, { "epoch": 6.44, "grad_norm": 50.27430725097656, "learning_rate": 4.2484008528784655e-05, "loss": 1.7073, "step": 805 }, { "epoch": 6.448, "grad_norm": 29.197919845581055, "learning_rate": 4.253731343283582e-05, "loss": 1.5677, "step": 806 }, { "epoch": 6.456, "grad_norm": 35.01530075073242, "learning_rate": 4.259061833688699e-05, "loss": 1.6886, "step": 807 }, { "epoch": 6.464, "grad_norm": 50.60932922363281, "learning_rate": 4.2643923240938166e-05, "loss": 1.6094, "step": 808 }, { "epoch": 6.4719999999999995, "grad_norm": 87.7199935913086, "learning_rate": 4.2697228144989345e-05, "loss": 2.2696, "step": 809 }, { "epoch": 6.48, "grad_norm": 59.781768798828125, "learning_rate": 4.275053304904051e-05, "loss": 2.541, "step": 810 }, { "epoch": 6.4879999999999995, "grad_norm": 127.62622833251953, "learning_rate": 4.2803837953091684e-05, "loss": 2.0869, "step": 811 }, { "epoch": 6.496, "grad_norm": 42.9911994934082, "learning_rate": 4.2857142857142856e-05, "loss": 1.8199, "step": 812 }, { "epoch": 6.504, "grad_norm": 50.07044982910156, "learning_rate": 4.2910447761194036e-05, "loss": 1.8854, "step": 813 }, { "epoch": 6.5120000000000005, "grad_norm": 46.923465728759766, "learning_rate": 4.29637526652452e-05, "loss": 2.2493, "step": 814 }, { "epoch": 6.52, "grad_norm": 66.74928283691406, "learning_rate": 4.3017057569296374e-05, "loss": 2.1415, "step": 815 }, { "epoch": 6.5280000000000005, "grad_norm": 90.24658203125, "learning_rate": 4.307036247334755e-05, "loss": 2.73, "step": 816 }, { "epoch": 6.536, "grad_norm": 31.466455459594727, "learning_rate": 4.3123667377398726e-05, "loss": 2.0775, "step": 817 }, { "epoch": 6.5440000000000005, "grad_norm": 23.97210693359375, "learning_rate": 4.31769722814499e-05, "loss": 1.6166, "step": 818 }, { "epoch": 6.552, "grad_norm": 81.92521667480469, "learning_rate": 4.3230277185501065e-05, "loss": 1.5784, "step": 819 }, { "epoch": 6.5600000000000005, "grad_norm": 222.5585174560547, "learning_rate": 4.328358208955224e-05, "loss": 2.3177, "step": 820 }, { "epoch": 6.568, "grad_norm": 74.41816711425781, "learning_rate": 4.333688699360342e-05, "loss": 1.5354, "step": 821 }, { "epoch": 6.576, "grad_norm": 92.18634796142578, "learning_rate": 4.339019189765459e-05, "loss": 1.6136, "step": 822 }, { "epoch": 6.584, "grad_norm": 51.097686767578125, "learning_rate": 4.3443496801705755e-05, "loss": 1.8922, "step": 823 }, { "epoch": 6.592, "grad_norm": 90.13542938232422, "learning_rate": 4.3496801705756935e-05, "loss": 1.8597, "step": 824 }, { "epoch": 6.6, "grad_norm": 262.2388610839844, "learning_rate": 4.355010660980811e-05, "loss": 2.3473, "step": 825 }, { "epoch": 6.608, "grad_norm": 47.430843353271484, "learning_rate": 4.360341151385928e-05, "loss": 2.312, "step": 826 }, { "epoch": 6.616, "grad_norm": 39.90750503540039, "learning_rate": 4.3656716417910446e-05, "loss": 1.6055, "step": 827 }, { "epoch": 6.624, "grad_norm": 31.544815063476562, "learning_rate": 4.3710021321961625e-05, "loss": 1.9967, "step": 828 }, { "epoch": 6.632, "grad_norm": 30.323444366455078, "learning_rate": 4.37633262260128e-05, "loss": 1.8097, "step": 829 }, { "epoch": 6.64, "grad_norm": 31.279512405395508, "learning_rate": 4.381663113006397e-05, "loss": 2.0503, "step": 830 }, { "epoch": 6.648, "grad_norm": 25.876325607299805, "learning_rate": 4.3869936034115136e-05, "loss": 1.9338, "step": 831 }, { "epoch": 6.656, "grad_norm": 33.20934295654297, "learning_rate": 4.3923240938166316e-05, "loss": 1.7652, "step": 832 }, { "epoch": 6.664, "grad_norm": 34.36385726928711, "learning_rate": 4.397654584221749e-05, "loss": 1.3611, "step": 833 }, { "epoch": 6.672, "grad_norm": 98.47650909423828, "learning_rate": 4.402985074626866e-05, "loss": 1.9658, "step": 834 }, { "epoch": 6.68, "grad_norm": 49.317020416259766, "learning_rate": 4.408315565031983e-05, "loss": 2.0824, "step": 835 }, { "epoch": 6.688, "grad_norm": 23.669872283935547, "learning_rate": 4.4136460554371006e-05, "loss": 1.8668, "step": 836 }, { "epoch": 6.696, "grad_norm": 58.06242370605469, "learning_rate": 4.418976545842218e-05, "loss": 1.835, "step": 837 }, { "epoch": 6.704, "grad_norm": 55.26518630981445, "learning_rate": 4.424307036247335e-05, "loss": 1.6597, "step": 838 }, { "epoch": 6.712, "grad_norm": 47.28418731689453, "learning_rate": 4.429637526652452e-05, "loss": 1.5108, "step": 839 }, { "epoch": 6.72, "grad_norm": 67.88128662109375, "learning_rate": 4.43496801705757e-05, "loss": 1.7563, "step": 840 }, { "epoch": 6.728, "grad_norm": 35.79535675048828, "learning_rate": 4.440298507462687e-05, "loss": 1.888, "step": 841 }, { "epoch": 6.736, "grad_norm": 42.82850646972656, "learning_rate": 4.445628997867804e-05, "loss": 1.7288, "step": 842 }, { "epoch": 6.744, "grad_norm": 104.82152557373047, "learning_rate": 4.450959488272921e-05, "loss": 2.3195, "step": 843 }, { "epoch": 6.752, "grad_norm": 55.12379455566406, "learning_rate": 4.456289978678039e-05, "loss": 1.9239, "step": 844 }, { "epoch": 6.76, "grad_norm": 22.72249412536621, "learning_rate": 4.461620469083156e-05, "loss": 1.676, "step": 845 }, { "epoch": 6.768, "grad_norm": 38.31277847290039, "learning_rate": 4.466950959488273e-05, "loss": 1.9985, "step": 846 }, { "epoch": 6.776, "grad_norm": 91.0455093383789, "learning_rate": 4.47228144989339e-05, "loss": 2.6572, "step": 847 }, { "epoch": 6.784, "grad_norm": 17.643020629882812, "learning_rate": 4.477611940298508e-05, "loss": 1.988, "step": 848 }, { "epoch": 6.792, "grad_norm": 30.944318771362305, "learning_rate": 4.482942430703625e-05, "loss": 2.8295, "step": 849 }, { "epoch": 6.8, "grad_norm": 23.366981506347656, "learning_rate": 4.488272921108742e-05, "loss": 1.9935, "step": 850 }, { "epoch": 6.808, "grad_norm": 37.886959075927734, "learning_rate": 4.493603411513859e-05, "loss": 1.7582, "step": 851 }, { "epoch": 6.816, "grad_norm": 40.907012939453125, "learning_rate": 4.498933901918977e-05, "loss": 1.7174, "step": 852 }, { "epoch": 6.824, "grad_norm": 37.028228759765625, "learning_rate": 4.504264392324094e-05, "loss": 1.7586, "step": 853 }, { "epoch": 6.832, "grad_norm": 36.438446044921875, "learning_rate": 4.5095948827292114e-05, "loss": 2.2814, "step": 854 }, { "epoch": 6.84, "grad_norm": 34.76927947998047, "learning_rate": 4.5149253731343286e-05, "loss": 1.9821, "step": 855 }, { "epoch": 6.848, "grad_norm": 52.68210220336914, "learning_rate": 4.520255863539446e-05, "loss": 2.0371, "step": 856 }, { "epoch": 6.856, "grad_norm": 40.66558837890625, "learning_rate": 4.525586353944563e-05, "loss": 1.8678, "step": 857 }, { "epoch": 6.864, "grad_norm": 60.729522705078125, "learning_rate": 4.5309168443496804e-05, "loss": 1.5298, "step": 858 }, { "epoch": 6.872, "grad_norm": 50.63246536254883, "learning_rate": 4.536247334754798e-05, "loss": 2.7484, "step": 859 }, { "epoch": 6.88, "grad_norm": 92.44477844238281, "learning_rate": 4.541577825159915e-05, "loss": 1.7067, "step": 860 }, { "epoch": 6.888, "grad_norm": 34.64543533325195, "learning_rate": 4.546908315565032e-05, "loss": 1.9725, "step": 861 }, { "epoch": 6.896, "grad_norm": 61.68946075439453, "learning_rate": 4.5522388059701495e-05, "loss": 1.8831, "step": 862 }, { "epoch": 6.904, "grad_norm": 207.87744140625, "learning_rate": 4.557569296375267e-05, "loss": 2.3053, "step": 863 }, { "epoch": 6.912, "grad_norm": 62.57316589355469, "learning_rate": 4.562899786780384e-05, "loss": 3.169, "step": 864 }, { "epoch": 6.92, "grad_norm": 37.10798263549805, "learning_rate": 4.568230277185501e-05, "loss": 1.7467, "step": 865 }, { "epoch": 6.928, "grad_norm": 27.26040267944336, "learning_rate": 4.5735607675906185e-05, "loss": 1.8041, "step": 866 }, { "epoch": 6.936, "grad_norm": 45.701866149902344, "learning_rate": 4.578891257995736e-05, "loss": 2.3258, "step": 867 }, { "epoch": 6.944, "grad_norm": 706.0728149414062, "learning_rate": 4.584221748400853e-05, "loss": 1.7101, "step": 868 }, { "epoch": 6.952, "grad_norm": 29.78115463256836, "learning_rate": 4.58955223880597e-05, "loss": 2.0194, "step": 869 }, { "epoch": 6.96, "grad_norm": 29.995431900024414, "learning_rate": 4.5948827292110876e-05, "loss": 1.6078, "step": 870 }, { "epoch": 6.968, "grad_norm": 97.24430847167969, "learning_rate": 4.600213219616205e-05, "loss": 3.1081, "step": 871 }, { "epoch": 6.976, "grad_norm": 22.87664031982422, "learning_rate": 4.605543710021322e-05, "loss": 1.5926, "step": 872 }, { "epoch": 6.984, "grad_norm": 33.584747314453125, "learning_rate": 4.6108742004264394e-05, "loss": 1.7971, "step": 873 }, { "epoch": 6.992, "grad_norm": 62.90135192871094, "learning_rate": 4.6162046908315566e-05, "loss": 1.6081, "step": 874 }, { "epoch": 7.0, "grad_norm": 61.349998474121094, "learning_rate": 4.621535181236674e-05, "loss": 1.5465, "step": 875 }, { "epoch": 7.0, "eval_loss": 1.9815495014190674, "eval_map": 0.0792, "eval_map_50": 0.1682, "eval_map_75": 0.064, "eval_map_Coverall": 0.3072, "eval_map_Face_Shield": 0.0, "eval_map_Gloves": 0.0388, "eval_map_Goggles": 0.0, "eval_map_Mask": 0.0502, "eval_map_large": 0.0873, "eval_map_medium": 0.0221, "eval_map_small": 0.0338, "eval_mar_1": 0.1002, "eval_mar_10": 0.1986, "eval_mar_100": 0.2477, "eval_mar_100_Coverall": 0.6467, "eval_mar_100_Face_Shield": 0.0, "eval_mar_100_Gloves": 0.2689, "eval_mar_100_Goggles": 0.0, "eval_mar_100_Mask": 0.3231, "eval_mar_large": 0.2513, "eval_mar_medium": 0.1495, "eval_mar_small": 0.0752, "eval_runtime": 1.8333, "eval_samples_per_second": 15.818, "eval_steps_per_second": 1.091, "step": 875 }, { "epoch": 7.008, "grad_norm": 33.69021224975586, "learning_rate": 4.626865671641791e-05, "loss": 1.8644, "step": 876 }, { "epoch": 7.016, "grad_norm": 38.12076950073242, "learning_rate": 4.6321961620469084e-05, "loss": 2.0124, "step": 877 }, { "epoch": 7.024, "grad_norm": 36.59566116333008, "learning_rate": 4.637526652452026e-05, "loss": 1.8086, "step": 878 }, { "epoch": 7.032, "grad_norm": 38.74513626098633, "learning_rate": 4.642857142857143e-05, "loss": 1.6351, "step": 879 }, { "epoch": 7.04, "grad_norm": 50.361610412597656, "learning_rate": 4.64818763326226e-05, "loss": 1.8397, "step": 880 }, { "epoch": 7.048, "grad_norm": 29.712587356567383, "learning_rate": 4.6535181236673775e-05, "loss": 1.9608, "step": 881 }, { "epoch": 7.056, "grad_norm": 39.56031036376953, "learning_rate": 4.658848614072495e-05, "loss": 1.7573, "step": 882 }, { "epoch": 7.064, "grad_norm": 42.442264556884766, "learning_rate": 4.664179104477612e-05, "loss": 1.5562, "step": 883 }, { "epoch": 7.072, "grad_norm": 51.172096252441406, "learning_rate": 4.669509594882729e-05, "loss": 2.1067, "step": 884 }, { "epoch": 7.08, "grad_norm": 167.31097412109375, "learning_rate": 4.6748400852878465e-05, "loss": 1.8921, "step": 885 }, { "epoch": 7.088, "grad_norm": 27.224506378173828, "learning_rate": 4.6801705756929645e-05, "loss": 1.7058, "step": 886 }, { "epoch": 7.096, "grad_norm": 41.19367218017578, "learning_rate": 4.685501066098081e-05, "loss": 1.5442, "step": 887 }, { "epoch": 7.104, "grad_norm": 43.358306884765625, "learning_rate": 4.690831556503198e-05, "loss": 1.8916, "step": 888 }, { "epoch": 7.112, "grad_norm": 36.4341926574707, "learning_rate": 4.6961620469083156e-05, "loss": 1.6771, "step": 889 }, { "epoch": 7.12, "grad_norm": 35.266475677490234, "learning_rate": 4.7014925373134335e-05, "loss": 1.5827, "step": 890 }, { "epoch": 7.128, "grad_norm": 36.86835861206055, "learning_rate": 4.706823027718551e-05, "loss": 1.4676, "step": 891 }, { "epoch": 7.136, "grad_norm": 37.48973846435547, "learning_rate": 4.7121535181236674e-05, "loss": 1.5741, "step": 892 }, { "epoch": 7.144, "grad_norm": 36.91202926635742, "learning_rate": 4.7174840085287846e-05, "loss": 1.8157, "step": 893 }, { "epoch": 7.152, "grad_norm": 33.58318328857422, "learning_rate": 4.7228144989339026e-05, "loss": 1.3645, "step": 894 }, { "epoch": 7.16, "grad_norm": 57.403995513916016, "learning_rate": 4.72814498933902e-05, "loss": 1.6969, "step": 895 }, { "epoch": 7.168, "grad_norm": 94.33474731445312, "learning_rate": 4.7334754797441364e-05, "loss": 2.1239, "step": 896 }, { "epoch": 7.176, "grad_norm": 289.90509033203125, "learning_rate": 4.738805970149254e-05, "loss": 2.1125, "step": 897 }, { "epoch": 7.184, "grad_norm": 41.35329055786133, "learning_rate": 4.7441364605543716e-05, "loss": 1.8597, "step": 898 }, { "epoch": 7.192, "grad_norm": 62.537776947021484, "learning_rate": 4.749466950959489e-05, "loss": 1.6432, "step": 899 }, { "epoch": 7.2, "grad_norm": 35.892581939697266, "learning_rate": 4.7547974413646055e-05, "loss": 1.812, "step": 900 }, { "epoch": 7.208, "grad_norm": 455.5152893066406, "learning_rate": 4.760127931769723e-05, "loss": 1.9541, "step": 901 }, { "epoch": 7.216, "grad_norm": 42.381656646728516, "learning_rate": 4.765458422174841e-05, "loss": 2.2873, "step": 902 }, { "epoch": 7.224, "grad_norm": 152.47256469726562, "learning_rate": 4.770788912579958e-05, "loss": 2.4372, "step": 903 }, { "epoch": 7.232, "grad_norm": 57.32434844970703, "learning_rate": 4.7761194029850745e-05, "loss": 2.1535, "step": 904 }, { "epoch": 7.24, "grad_norm": 163.99557495117188, "learning_rate": 4.781449893390192e-05, "loss": 2.0037, "step": 905 }, { "epoch": 7.248, "grad_norm": 64.00248718261719, "learning_rate": 4.78678038379531e-05, "loss": 1.9511, "step": 906 }, { "epoch": 7.256, "grad_norm": 45.002708435058594, "learning_rate": 4.792110874200427e-05, "loss": 2.3328, "step": 907 }, { "epoch": 7.264, "grad_norm": 42.738990783691406, "learning_rate": 4.7974413646055436e-05, "loss": 1.7614, "step": 908 }, { "epoch": 7.272, "grad_norm": 120.0714111328125, "learning_rate": 4.802771855010661e-05, "loss": 2.373, "step": 909 }, { "epoch": 7.28, "grad_norm": 42.95566940307617, "learning_rate": 4.808102345415779e-05, "loss": 1.6421, "step": 910 }, { "epoch": 7.288, "grad_norm": 52.836063385009766, "learning_rate": 4.813432835820896e-05, "loss": 1.2136, "step": 911 }, { "epoch": 7.296, "grad_norm": 62.43743133544922, "learning_rate": 4.8187633262260126e-05, "loss": 2.3433, "step": 912 }, { "epoch": 7.304, "grad_norm": 40.93174362182617, "learning_rate": 4.82409381663113e-05, "loss": 2.0914, "step": 913 }, { "epoch": 7.312, "grad_norm": 27.69137954711914, "learning_rate": 4.829424307036248e-05, "loss": 1.6758, "step": 914 }, { "epoch": 7.32, "grad_norm": 28.375314712524414, "learning_rate": 4.834754797441365e-05, "loss": 1.5638, "step": 915 }, { "epoch": 7.328, "grad_norm": 40.14604568481445, "learning_rate": 4.840085287846482e-05, "loss": 2.2652, "step": 916 }, { "epoch": 7.336, "grad_norm": 498.091552734375, "learning_rate": 4.8454157782515996e-05, "loss": 2.1494, "step": 917 }, { "epoch": 7.344, "grad_norm": 40.53156661987305, "learning_rate": 4.850746268656717e-05, "loss": 1.5417, "step": 918 }, { "epoch": 7.352, "grad_norm": 24.10733985900879, "learning_rate": 4.856076759061834e-05, "loss": 1.5732, "step": 919 }, { "epoch": 7.36, "grad_norm": 97.97477722167969, "learning_rate": 4.861407249466951e-05, "loss": 1.6117, "step": 920 }, { "epoch": 7.368, "grad_norm": 27.428665161132812, "learning_rate": 4.866737739872069e-05, "loss": 1.56, "step": 921 }, { "epoch": 7.376, "grad_norm": 62.29743957519531, "learning_rate": 4.872068230277186e-05, "loss": 1.7247, "step": 922 }, { "epoch": 7.384, "grad_norm": 59.75631332397461, "learning_rate": 4.877398720682303e-05, "loss": 1.779, "step": 923 }, { "epoch": 7.392, "grad_norm": 78.98824310302734, "learning_rate": 4.88272921108742e-05, "loss": 1.9674, "step": 924 }, { "epoch": 7.4, "grad_norm": 79.23694610595703, "learning_rate": 4.888059701492538e-05, "loss": 1.7103, "step": 925 }, { "epoch": 7.408, "grad_norm": 41.866573333740234, "learning_rate": 4.893390191897655e-05, "loss": 1.7494, "step": 926 }, { "epoch": 7.416, "grad_norm": 76.04717254638672, "learning_rate": 4.898720682302772e-05, "loss": 2.2267, "step": 927 }, { "epoch": 7.424, "grad_norm": 46.052223205566406, "learning_rate": 4.904051172707889e-05, "loss": 1.7467, "step": 928 }, { "epoch": 7.432, "grad_norm": 37.799034118652344, "learning_rate": 4.909381663113007e-05, "loss": 1.4052, "step": 929 }, { "epoch": 7.44, "grad_norm": 35.55511474609375, "learning_rate": 4.914712153518124e-05, "loss": 1.648, "step": 930 }, { "epoch": 7.448, "grad_norm": 65.70741271972656, "learning_rate": 4.920042643923241e-05, "loss": 2.0134, "step": 931 }, { "epoch": 7.456, "grad_norm": 54.69572830200195, "learning_rate": 4.9253731343283586e-05, "loss": 2.4284, "step": 932 }, { "epoch": 7.464, "grad_norm": 401.4771423339844, "learning_rate": 4.930703624733476e-05, "loss": 2.2037, "step": 933 }, { "epoch": 7.4719999999999995, "grad_norm": 43.72904586791992, "learning_rate": 4.936034115138593e-05, "loss": 1.9634, "step": 934 }, { "epoch": 7.48, "grad_norm": 43.29294967651367, "learning_rate": 4.9413646055437103e-05, "loss": 2.402, "step": 935 }, { "epoch": 7.4879999999999995, "grad_norm": 92.65357971191406, "learning_rate": 4.9466950959488276e-05, "loss": 2.3332, "step": 936 }, { "epoch": 7.496, "grad_norm": 60.75580978393555, "learning_rate": 4.952025586353945e-05, "loss": 2.1424, "step": 937 }, { "epoch": 7.504, "grad_norm": 29.399410247802734, "learning_rate": 4.957356076759062e-05, "loss": 1.7409, "step": 938 }, { "epoch": 7.5120000000000005, "grad_norm": 30.515026092529297, "learning_rate": 4.9626865671641794e-05, "loss": 1.9916, "step": 939 }, { "epoch": 7.52, "grad_norm": 46.7426872253418, "learning_rate": 4.9680170575692967e-05, "loss": 1.8374, "step": 940 }, { "epoch": 7.5280000000000005, "grad_norm": 37.015926361083984, "learning_rate": 4.973347547974414e-05, "loss": 2.6373, "step": 941 }, { "epoch": 7.536, "grad_norm": 73.2405776977539, "learning_rate": 4.978678038379531e-05, "loss": 2.2161, "step": 942 }, { "epoch": 7.5440000000000005, "grad_norm": 45.67422866821289, "learning_rate": 4.9840085287846485e-05, "loss": 1.5763, "step": 943 }, { "epoch": 7.552, "grad_norm": 29.9785213470459, "learning_rate": 4.989339019189766e-05, "loss": 3.746, "step": 944 }, { "epoch": 7.5600000000000005, "grad_norm": 52.08788299560547, "learning_rate": 4.994669509594883e-05, "loss": 1.5838, "step": 945 }, { "epoch": 7.568, "grad_norm": 26.25680160522461, "learning_rate": 5e-05, "loss": 2.1136, "step": 946 }, { "epoch": 7.576, "grad_norm": 50.23767852783203, "learning_rate": 4.999407372288729e-05, "loss": 2.5049, "step": 947 }, { "epoch": 7.584, "grad_norm": 473.8786315917969, "learning_rate": 4.998814744577457e-05, "loss": 1.9445, "step": 948 }, { "epoch": 7.592, "grad_norm": 41.712005615234375, "learning_rate": 4.9982221168661846e-05, "loss": 1.7159, "step": 949 }, { "epoch": 7.6, "grad_norm": 30.37320899963379, "learning_rate": 4.997629489154913e-05, "loss": 1.6451, "step": 950 }, { "epoch": 7.608, "grad_norm": 41.507164001464844, "learning_rate": 4.997036861443642e-05, "loss": 2.2941, "step": 951 }, { "epoch": 7.616, "grad_norm": 35.56565856933594, "learning_rate": 4.9964442337323696e-05, "loss": 1.7779, "step": 952 }, { "epoch": 7.624, "grad_norm": 67.58474731445312, "learning_rate": 4.995851606021098e-05, "loss": 2.6865, "step": 953 }, { "epoch": 7.632, "grad_norm": 48.98326110839844, "learning_rate": 4.995258978309826e-05, "loss": 1.2896, "step": 954 }, { "epoch": 7.64, "grad_norm": 19.31278419494629, "learning_rate": 4.994666350598554e-05, "loss": 3.1938, "step": 955 }, { "epoch": 7.648, "grad_norm": 86.64693450927734, "learning_rate": 4.9940737228872824e-05, "loss": 1.9738, "step": 956 }, { "epoch": 7.656, "grad_norm": 101.87957000732422, "learning_rate": 4.993481095176011e-05, "loss": 1.7226, "step": 957 }, { "epoch": 7.664, "grad_norm": 36.59895706176758, "learning_rate": 4.992888467464739e-05, "loss": 1.6994, "step": 958 }, { "epoch": 7.672, "grad_norm": 39.776493072509766, "learning_rate": 4.992295839753467e-05, "loss": 1.8373, "step": 959 }, { "epoch": 7.68, "grad_norm": 37.568302154541016, "learning_rate": 4.991703212042195e-05, "loss": 2.2031, "step": 960 }, { "epoch": 7.688, "grad_norm": 38.46575164794922, "learning_rate": 4.991110584330924e-05, "loss": 2.1059, "step": 961 }, { "epoch": 7.696, "grad_norm": 54.49308395385742, "learning_rate": 4.990517956619652e-05, "loss": 1.9824, "step": 962 }, { "epoch": 7.704, "grad_norm": 38.70124053955078, "learning_rate": 4.98992532890838e-05, "loss": 2.1168, "step": 963 }, { "epoch": 7.712, "grad_norm": 71.26351928710938, "learning_rate": 4.989332701197108e-05, "loss": 1.4888, "step": 964 }, { "epoch": 7.72, "grad_norm": 74.41661071777344, "learning_rate": 4.988740073485836e-05, "loss": 1.6061, "step": 965 }, { "epoch": 7.728, "grad_norm": 41.9335823059082, "learning_rate": 4.9881474457745647e-05, "loss": 2.0151, "step": 966 }, { "epoch": 7.736, "grad_norm": 24.74287986755371, "learning_rate": 4.987554818063293e-05, "loss": 2.7264, "step": 967 }, { "epoch": 7.744, "grad_norm": 44.74628448486328, "learning_rate": 4.986962190352021e-05, "loss": 1.3477, "step": 968 }, { "epoch": 7.752, "grad_norm": 74.91075897216797, "learning_rate": 4.986369562640749e-05, "loss": 2.2265, "step": 969 }, { "epoch": 7.76, "grad_norm": 55.067996978759766, "learning_rate": 4.9857769349294775e-05, "loss": 1.8848, "step": 970 }, { "epoch": 7.768, "grad_norm": 46.54352569580078, "learning_rate": 4.985184307218206e-05, "loss": 1.7565, "step": 971 }, { "epoch": 7.776, "grad_norm": 67.09346008300781, "learning_rate": 4.984591679506934e-05, "loss": 1.5941, "step": 972 }, { "epoch": 7.784, "grad_norm": 74.38175964355469, "learning_rate": 4.983999051795662e-05, "loss": 1.9732, "step": 973 }, { "epoch": 7.792, "grad_norm": 50.07896423339844, "learning_rate": 4.9834064240843904e-05, "loss": 2.416, "step": 974 }, { "epoch": 7.8, "grad_norm": 100.56977081298828, "learning_rate": 4.982813796373118e-05, "loss": 3.344, "step": 975 }, { "epoch": 7.808, "grad_norm": 262.9228515625, "learning_rate": 4.982221168661847e-05, "loss": 3.4714, "step": 976 }, { "epoch": 7.816, "grad_norm": 54.9241828918457, "learning_rate": 4.9816285409505754e-05, "loss": 2.8002, "step": 977 }, { "epoch": 7.824, "grad_norm": 153.59994506835938, "learning_rate": 4.981035913239303e-05, "loss": 3.6285, "step": 978 }, { "epoch": 7.832, "grad_norm": 86.25049591064453, "learning_rate": 4.980443285528031e-05, "loss": 4.0569, "step": 979 }, { "epoch": 7.84, "grad_norm": 2191.875, "learning_rate": 4.97985065781676e-05, "loss": 3.7048, "step": 980 }, { "epoch": 7.848, "grad_norm": 375.7640686035156, "learning_rate": 4.979258030105488e-05, "loss": 5.0184, "step": 981 }, { "epoch": 7.856, "grad_norm": 235.3843231201172, "learning_rate": 4.978665402394216e-05, "loss": 4.9398, "step": 982 }, { "epoch": 7.864, "grad_norm": 120.26214599609375, "learning_rate": 4.978072774682944e-05, "loss": 4.6837, "step": 983 }, { "epoch": 7.872, "grad_norm": 97.89712524414062, "learning_rate": 4.9774801469716726e-05, "loss": 5.7838, "step": 984 }, { "epoch": 7.88, "grad_norm": 116.32960510253906, "learning_rate": 4.976887519260401e-05, "loss": 4.3023, "step": 985 }, { "epoch": 7.888, "grad_norm": 86.32508087158203, "learning_rate": 4.976294891549129e-05, "loss": 4.5048, "step": 986 }, { "epoch": 7.896, "grad_norm": 73.55574798583984, "learning_rate": 4.9757022638378576e-05, "loss": 3.9039, "step": 987 }, { "epoch": 7.904, "grad_norm": 135.86123657226562, "learning_rate": 4.9751096361265855e-05, "loss": 2.8894, "step": 988 }, { "epoch": 7.912, "grad_norm": 112.92809295654297, "learning_rate": 4.9745170084153134e-05, "loss": 2.6295, "step": 989 }, { "epoch": 7.92, "grad_norm": 111.18238067626953, "learning_rate": 4.973924380704042e-05, "loss": 3.6904, "step": 990 }, { "epoch": 7.928, "grad_norm": 20.843961715698242, "learning_rate": 4.9733317529927705e-05, "loss": 3.5952, "step": 991 }, { "epoch": 7.936, "grad_norm": 49.67041778564453, "learning_rate": 4.9727391252814984e-05, "loss": 2.988, "step": 992 }, { "epoch": 7.944, "grad_norm": 240.55125427246094, "learning_rate": 4.972146497570226e-05, "loss": 3.2312, "step": 993 }, { "epoch": 7.952, "grad_norm": 73.51168823242188, "learning_rate": 4.971553869858955e-05, "loss": 3.0415, "step": 994 }, { "epoch": 7.96, "grad_norm": 84.29178619384766, "learning_rate": 4.9709612421476834e-05, "loss": 3.2297, "step": 995 }, { "epoch": 7.968, "grad_norm": 71.0030746459961, "learning_rate": 4.970368614436411e-05, "loss": 3.1412, "step": 996 }, { "epoch": 7.976, "grad_norm": 27.907068252563477, "learning_rate": 4.96977598672514e-05, "loss": 3.2442, "step": 997 }, { "epoch": 7.984, "grad_norm": 57.60411071777344, "learning_rate": 4.969183359013868e-05, "loss": 3.2218, "step": 998 }, { "epoch": 7.992, "grad_norm": 34.20361328125, "learning_rate": 4.9685907313025956e-05, "loss": 2.9746, "step": 999 }, { "epoch": 8.0, "grad_norm": 47.95493698120117, "learning_rate": 4.967998103591324e-05, "loss": 2.9868, "step": 1000 }, { "epoch": 8.0, "eval_loss": 3.2292118072509766, "eval_map": 0.0091, "eval_map_50": 0.033, "eval_map_75": 0.0026, "eval_map_Coverall": 0.0443, "eval_map_Face_Shield": 0.0, "eval_map_Gloves": 0.0004, "eval_map_Goggles": 0.0, "eval_map_Mask": 0.001, "eval_map_large": 0.0094, "eval_map_medium": 0.0005, "eval_map_small": 0.0, "eval_mar_1": 0.0137, "eval_mar_10": 0.0586, "eval_mar_100": 0.0699, "eval_mar_100_Coverall": 0.2933, "eval_mar_100_Face_Shield": 0.0, "eval_mar_100_Gloves": 0.0311, "eval_mar_100_Goggles": 0.0, "eval_mar_100_Mask": 0.025, "eval_mar_large": 0.0764, "eval_mar_medium": 0.0166, "eval_mar_small": 0.0, "eval_runtime": 1.8184, "eval_samples_per_second": 15.948, "eval_steps_per_second": 1.1, "step": 1000 }, { "epoch": 8.008, "grad_norm": 42.5607795715332, "learning_rate": 4.967405475880053e-05, "loss": 2.8107, "step": 1001 }, { "epoch": 8.016, "grad_norm": 29.789011001586914, "learning_rate": 4.9668128481687806e-05, "loss": 3.1793, "step": 1002 }, { "epoch": 8.024, "grad_norm": 47.52362823486328, "learning_rate": 4.9662202204575085e-05, "loss": 2.2856, "step": 1003 }, { "epoch": 8.032, "grad_norm": 31.222267150878906, "learning_rate": 4.965627592746237e-05, "loss": 3.2788, "step": 1004 }, { "epoch": 8.04, "grad_norm": 48.489131927490234, "learning_rate": 4.9650349650349656e-05, "loss": 2.8442, "step": 1005 }, { "epoch": 8.048, "grad_norm": 25.12875747680664, "learning_rate": 4.9644423373236935e-05, "loss": 2.8323, "step": 1006 }, { "epoch": 8.056, "grad_norm": 36.55942916870117, "learning_rate": 4.9638497096124214e-05, "loss": 2.7208, "step": 1007 }, { "epoch": 8.064, "grad_norm": 31.779560089111328, "learning_rate": 4.96325708190115e-05, "loss": 2.6691, "step": 1008 }, { "epoch": 8.072, "grad_norm": 53.38471984863281, "learning_rate": 4.962664454189878e-05, "loss": 2.8118, "step": 1009 }, { "epoch": 8.08, "grad_norm": 70.99284362792969, "learning_rate": 4.9620718264786064e-05, "loss": 2.8973, "step": 1010 }, { "epoch": 8.088, "grad_norm": 27.928627014160156, "learning_rate": 4.961479198767335e-05, "loss": 3.1701, "step": 1011 }, { "epoch": 8.096, "grad_norm": 64.58328247070312, "learning_rate": 4.960886571056063e-05, "loss": 2.7139, "step": 1012 }, { "epoch": 8.104, "grad_norm": 61.23844909667969, "learning_rate": 4.960293943344791e-05, "loss": 2.6935, "step": 1013 }, { "epoch": 8.112, "grad_norm": 72.11639404296875, "learning_rate": 4.959701315633519e-05, "loss": 2.4934, "step": 1014 }, { "epoch": 8.12, "grad_norm": 39.307682037353516, "learning_rate": 4.959108687922248e-05, "loss": 2.2639, "step": 1015 }, { "epoch": 8.128, "grad_norm": 47.89452362060547, "learning_rate": 4.958516060210976e-05, "loss": 2.9546, "step": 1016 }, { "epoch": 8.136, "grad_norm": 50.096336364746094, "learning_rate": 4.9579234324997036e-05, "loss": 3.6706, "step": 1017 }, { "epoch": 8.144, "grad_norm": 149.3951873779297, "learning_rate": 4.957330804788432e-05, "loss": 2.8845, "step": 1018 }, { "epoch": 8.152, "grad_norm": 45.21243667602539, "learning_rate": 4.956738177077161e-05, "loss": 3.4497, "step": 1019 }, { "epoch": 8.16, "grad_norm": 66.92747497558594, "learning_rate": 4.9561455493658886e-05, "loss": 3.2798, "step": 1020 }, { "epoch": 8.168, "grad_norm": 34.26417922973633, "learning_rate": 4.955552921654617e-05, "loss": 2.9509, "step": 1021 }, { "epoch": 8.176, "grad_norm": 158.5045166015625, "learning_rate": 4.954960293943345e-05, "loss": 3.6737, "step": 1022 }, { "epoch": 8.184, "grad_norm": 145.74014282226562, "learning_rate": 4.954367666232073e-05, "loss": 2.8516, "step": 1023 }, { "epoch": 8.192, "grad_norm": 29.825603485107422, "learning_rate": 4.9537750385208015e-05, "loss": 2.7941, "step": 1024 }, { "epoch": 8.2, "grad_norm": 29.761350631713867, "learning_rate": 4.95318241080953e-05, "loss": 3.0706, "step": 1025 }, { "epoch": 8.208, "grad_norm": 46.82150650024414, "learning_rate": 4.952589783098258e-05, "loss": 2.4332, "step": 1026 }, { "epoch": 8.216, "grad_norm": 30.840848922729492, "learning_rate": 4.951997155386986e-05, "loss": 3.0066, "step": 1027 }, { "epoch": 8.224, "grad_norm": 33.82938003540039, "learning_rate": 4.951404527675714e-05, "loss": 2.6871, "step": 1028 }, { "epoch": 8.232, "grad_norm": 150.14410400390625, "learning_rate": 4.950811899964443e-05, "loss": 2.6818, "step": 1029 }, { "epoch": 8.24, "grad_norm": 18.35512351989746, "learning_rate": 4.950219272253171e-05, "loss": 3.3984, "step": 1030 }, { "epoch": 8.248, "grad_norm": 44.75056457519531, "learning_rate": 4.949626644541899e-05, "loss": 3.0207, "step": 1031 }, { "epoch": 8.256, "grad_norm": 48.57479476928711, "learning_rate": 4.949034016830627e-05, "loss": 2.7394, "step": 1032 }, { "epoch": 8.264, "grad_norm": 34.821014404296875, "learning_rate": 4.948441389119355e-05, "loss": 2.2969, "step": 1033 }, { "epoch": 8.272, "grad_norm": 58.62574005126953, "learning_rate": 4.947848761408084e-05, "loss": 2.9696, "step": 1034 }, { "epoch": 8.28, "grad_norm": 46.675662994384766, "learning_rate": 4.947256133696812e-05, "loss": 2.7226, "step": 1035 }, { "epoch": 8.288, "grad_norm": 32.828365325927734, "learning_rate": 4.94666350598554e-05, "loss": 2.8801, "step": 1036 }, { "epoch": 8.296, "grad_norm": 37.48468780517578, "learning_rate": 4.946070878274268e-05, "loss": 3.2586, "step": 1037 }, { "epoch": 8.304, "grad_norm": 47.93769454956055, "learning_rate": 4.9454782505629965e-05, "loss": 3.0569, "step": 1038 }, { "epoch": 8.312, "grad_norm": 35.23130798339844, "learning_rate": 4.944885622851725e-05, "loss": 2.7534, "step": 1039 }, { "epoch": 8.32, "grad_norm": 278.1644287109375, "learning_rate": 4.944292995140453e-05, "loss": 3.1417, "step": 1040 }, { "epoch": 8.328, "grad_norm": 48.85346984863281, "learning_rate": 4.943700367429181e-05, "loss": 2.6125, "step": 1041 }, { "epoch": 8.336, "grad_norm": 41.960487365722656, "learning_rate": 4.9431077397179094e-05, "loss": 3.795, "step": 1042 }, { "epoch": 8.344, "grad_norm": 154.69381713867188, "learning_rate": 4.942515112006637e-05, "loss": 3.3757, "step": 1043 }, { "epoch": 8.352, "grad_norm": 408.2906188964844, "learning_rate": 4.941922484295366e-05, "loss": 2.9969, "step": 1044 }, { "epoch": 8.36, "grad_norm": 149.0062255859375, "learning_rate": 4.9413298565840944e-05, "loss": 2.9221, "step": 1045 }, { "epoch": 8.368, "grad_norm": 200.99049377441406, "learning_rate": 4.940737228872822e-05, "loss": 3.0785, "step": 1046 }, { "epoch": 8.376, "grad_norm": 73.29725646972656, "learning_rate": 4.94014460116155e-05, "loss": 5.0678, "step": 1047 }, { "epoch": 8.384, "grad_norm": 167.9386444091797, "learning_rate": 4.939551973450279e-05, "loss": 3.7379, "step": 1048 }, { "epoch": 8.392, "grad_norm": 49.656150817871094, "learning_rate": 4.938959345739007e-05, "loss": 4.1258, "step": 1049 }, { "epoch": 8.4, "grad_norm": 42.180233001708984, "learning_rate": 4.938366718027735e-05, "loss": 3.7723, "step": 1050 }, { "epoch": 8.408, "grad_norm": 54.37676239013672, "learning_rate": 4.937774090316463e-05, "loss": 3.7675, "step": 1051 }, { "epoch": 8.416, "grad_norm": 47.935367584228516, "learning_rate": 4.9371814626051916e-05, "loss": 3.6794, "step": 1052 }, { "epoch": 8.424, "grad_norm": 42.36042785644531, "learning_rate": 4.93658883489392e-05, "loss": 4.0065, "step": 1053 }, { "epoch": 8.432, "grad_norm": 84.34934997558594, "learning_rate": 4.935996207182648e-05, "loss": 4.4485, "step": 1054 }, { "epoch": 8.44, "grad_norm": 26.928926467895508, "learning_rate": 4.9354035794713766e-05, "loss": 3.9971, "step": 1055 }, { "epoch": 8.448, "grad_norm": 33.07783889770508, "learning_rate": 4.9348109517601045e-05, "loss": 4.3781, "step": 1056 }, { "epoch": 8.456, "grad_norm": 33.3063850402832, "learning_rate": 4.9342183240488324e-05, "loss": 3.9141, "step": 1057 }, { "epoch": 8.464, "grad_norm": 43.42816162109375, "learning_rate": 4.933625696337561e-05, "loss": 3.4784, "step": 1058 }, { "epoch": 8.472, "grad_norm": 46.071495056152344, "learning_rate": 4.9330330686262895e-05, "loss": 3.3547, "step": 1059 }, { "epoch": 8.48, "grad_norm": 25.90642738342285, "learning_rate": 4.9324404409150174e-05, "loss": 3.6037, "step": 1060 }, { "epoch": 8.488, "grad_norm": 43.7113151550293, "learning_rate": 4.931847813203745e-05, "loss": 3.1102, "step": 1061 }, { "epoch": 8.496, "grad_norm": 48.85136795043945, "learning_rate": 4.931255185492474e-05, "loss": 3.6676, "step": 1062 }, { "epoch": 8.504, "grad_norm": 34.65818405151367, "learning_rate": 4.9306625577812024e-05, "loss": 3.3748, "step": 1063 }, { "epoch": 8.512, "grad_norm": 87.74609375, "learning_rate": 4.93006993006993e-05, "loss": 3.2628, "step": 1064 }, { "epoch": 8.52, "grad_norm": 115.44638061523438, "learning_rate": 4.929477302358659e-05, "loss": 3.798, "step": 1065 }, { "epoch": 8.528, "grad_norm": 41.00924301147461, "learning_rate": 4.928884674647387e-05, "loss": 2.8635, "step": 1066 }, { "epoch": 8.536, "grad_norm": 139.60618591308594, "learning_rate": 4.9282920469361146e-05, "loss": 3.3367, "step": 1067 }, { "epoch": 8.544, "grad_norm": 122.6593246459961, "learning_rate": 4.927699419224843e-05, "loss": 2.9644, "step": 1068 }, { "epoch": 8.552, "grad_norm": 61.58820724487305, "learning_rate": 4.927106791513572e-05, "loss": 3.3879, "step": 1069 }, { "epoch": 8.56, "grad_norm": 54.31479263305664, "learning_rate": 4.9265141638022996e-05, "loss": 3.5032, "step": 1070 }, { "epoch": 8.568, "grad_norm": 62.13996505737305, "learning_rate": 4.9259215360910275e-05, "loss": 3.1131, "step": 1071 }, { "epoch": 8.576, "grad_norm": 38.663944244384766, "learning_rate": 4.925328908379756e-05, "loss": 3.3565, "step": 1072 }, { "epoch": 8.584, "grad_norm": 22.107650756835938, "learning_rate": 4.9247362806684846e-05, "loss": 2.6637, "step": 1073 }, { "epoch": 8.592, "grad_norm": 26.589954376220703, "learning_rate": 4.9241436529572125e-05, "loss": 2.8478, "step": 1074 }, { "epoch": 8.6, "grad_norm": 31.972509384155273, "learning_rate": 4.9235510252459404e-05, "loss": 3.5803, "step": 1075 }, { "epoch": 8.608, "grad_norm": 315.6596374511719, "learning_rate": 4.922958397534669e-05, "loss": 3.5756, "step": 1076 }, { "epoch": 8.616, "grad_norm": 51.179290771484375, "learning_rate": 4.922365769823397e-05, "loss": 3.1024, "step": 1077 }, { "epoch": 8.624, "grad_norm": 37.12962341308594, "learning_rate": 4.9217731421121254e-05, "loss": 3.1104, "step": 1078 }, { "epoch": 8.632, "grad_norm": 23.98357582092285, "learning_rate": 4.921180514400854e-05, "loss": 3.1725, "step": 1079 }, { "epoch": 8.64, "grad_norm": 24.762699127197266, "learning_rate": 4.920587886689582e-05, "loss": 2.4416, "step": 1080 }, { "epoch": 8.648, "grad_norm": 28.9301815032959, "learning_rate": 4.91999525897831e-05, "loss": 2.9873, "step": 1081 }, { "epoch": 8.656, "grad_norm": 24.406633377075195, "learning_rate": 4.919402631267038e-05, "loss": 2.995, "step": 1082 }, { "epoch": 8.664, "grad_norm": 33.553138732910156, "learning_rate": 4.918810003555767e-05, "loss": 2.4126, "step": 1083 }, { "epoch": 8.672, "grad_norm": 37.78289031982422, "learning_rate": 4.918217375844495e-05, "loss": 3.3753, "step": 1084 }, { "epoch": 8.68, "grad_norm": 17.17049217224121, "learning_rate": 4.9176247481332226e-05, "loss": 3.1146, "step": 1085 }, { "epoch": 8.688, "grad_norm": 20.44700813293457, "learning_rate": 4.917032120421951e-05, "loss": 3.2837, "step": 1086 }, { "epoch": 8.696, "grad_norm": 31.56270408630371, "learning_rate": 4.91643949271068e-05, "loss": 2.4166, "step": 1087 }, { "epoch": 8.704, "grad_norm": 27.877687454223633, "learning_rate": 4.9158468649994076e-05, "loss": 3.2175, "step": 1088 }, { "epoch": 8.712, "grad_norm": 40.39419174194336, "learning_rate": 4.915254237288136e-05, "loss": 2.7863, "step": 1089 }, { "epoch": 8.72, "grad_norm": 37.85723876953125, "learning_rate": 4.914661609576864e-05, "loss": 2.1889, "step": 1090 }, { "epoch": 8.728, "grad_norm": 34.545448303222656, "learning_rate": 4.914068981865592e-05, "loss": 2.3977, "step": 1091 }, { "epoch": 8.736, "grad_norm": 61.84881591796875, "learning_rate": 4.9134763541543205e-05, "loss": 2.2908, "step": 1092 }, { "epoch": 8.744, "grad_norm": 85.13226318359375, "learning_rate": 4.912883726443049e-05, "loss": 2.5815, "step": 1093 }, { "epoch": 8.752, "grad_norm": 31.605422973632812, "learning_rate": 4.912291098731777e-05, "loss": 2.5917, "step": 1094 }, { "epoch": 8.76, "grad_norm": 27.944774627685547, "learning_rate": 4.911698471020505e-05, "loss": 2.1527, "step": 1095 }, { "epoch": 8.768, "grad_norm": 35.781742095947266, "learning_rate": 4.9111058433092333e-05, "loss": 2.6988, "step": 1096 }, { "epoch": 8.776, "grad_norm": 31.73261833190918, "learning_rate": 4.910513215597962e-05, "loss": 2.5439, "step": 1097 }, { "epoch": 8.784, "grad_norm": 28.898679733276367, "learning_rate": 4.90992058788669e-05, "loss": 2.1525, "step": 1098 }, { "epoch": 8.792, "grad_norm": 31.07322883605957, "learning_rate": 4.9093279601754183e-05, "loss": 1.879, "step": 1099 }, { "epoch": 8.8, "grad_norm": 278.30914306640625, "learning_rate": 4.908735332464146e-05, "loss": 1.8071, "step": 1100 }, { "epoch": 8.808, "grad_norm": 27.21160316467285, "learning_rate": 4.908142704752874e-05, "loss": 1.8709, "step": 1101 }, { "epoch": 8.816, "grad_norm": 31.55844497680664, "learning_rate": 4.907550077041603e-05, "loss": 2.2039, "step": 1102 }, { "epoch": 8.824, "grad_norm": 32.46535110473633, "learning_rate": 4.906957449330331e-05, "loss": 1.9989, "step": 1103 }, { "epoch": 8.832, "grad_norm": 20.530012130737305, "learning_rate": 4.906364821619059e-05, "loss": 2.0885, "step": 1104 }, { "epoch": 8.84, "grad_norm": 89.85709381103516, "learning_rate": 4.905772193907787e-05, "loss": 2.0011, "step": 1105 }, { "epoch": 8.848, "grad_norm": 72.84281158447266, "learning_rate": 4.9051795661965156e-05, "loss": 3.2377, "step": 1106 }, { "epoch": 8.856, "grad_norm": 23.68491554260254, "learning_rate": 4.904586938485244e-05, "loss": 1.5971, "step": 1107 }, { "epoch": 8.864, "grad_norm": 29.291461944580078, "learning_rate": 4.903994310773972e-05, "loss": 1.8891, "step": 1108 }, { "epoch": 8.872, "grad_norm": 41.15978240966797, "learning_rate": 4.9034016830627e-05, "loss": 1.9112, "step": 1109 }, { "epoch": 8.88, "grad_norm": 20.400264739990234, "learning_rate": 4.9028090553514284e-05, "loss": 1.5601, "step": 1110 }, { "epoch": 8.888, "grad_norm": 27.135223388671875, "learning_rate": 4.902216427640156e-05, "loss": 1.8611, "step": 1111 }, { "epoch": 8.896, "grad_norm": 20.150590896606445, "learning_rate": 4.901623799928885e-05, "loss": 1.8312, "step": 1112 }, { "epoch": 8.904, "grad_norm": 22.206113815307617, "learning_rate": 4.9010311722176134e-05, "loss": 3.03, "step": 1113 }, { "epoch": 8.912, "grad_norm": 20.418210983276367, "learning_rate": 4.900438544506341e-05, "loss": 1.9435, "step": 1114 }, { "epoch": 8.92, "grad_norm": 25.339454650878906, "learning_rate": 4.899845916795069e-05, "loss": 1.7289, "step": 1115 }, { "epoch": 8.928, "grad_norm": 21.869112014770508, "learning_rate": 4.899253289083798e-05, "loss": 2.2031, "step": 1116 }, { "epoch": 8.936, "grad_norm": 21.26622200012207, "learning_rate": 4.898660661372526e-05, "loss": 1.8066, "step": 1117 }, { "epoch": 8.943999999999999, "grad_norm": 53.025611877441406, "learning_rate": 4.898068033661254e-05, "loss": 1.9761, "step": 1118 }, { "epoch": 8.952, "grad_norm": 48.879547119140625, "learning_rate": 4.897475405949982e-05, "loss": 1.9273, "step": 1119 }, { "epoch": 8.96, "grad_norm": 26.59044075012207, "learning_rate": 4.8968827782387106e-05, "loss": 1.752, "step": 1120 }, { "epoch": 8.968, "grad_norm": 51.61838150024414, "learning_rate": 4.896290150527439e-05, "loss": 2.0518, "step": 1121 }, { "epoch": 8.975999999999999, "grad_norm": 18.886507034301758, "learning_rate": 4.895697522816167e-05, "loss": 2.535, "step": 1122 }, { "epoch": 8.984, "grad_norm": 25.92640495300293, "learning_rate": 4.8951048951048956e-05, "loss": 1.7122, "step": 1123 }, { "epoch": 8.992, "grad_norm": 33.98381805419922, "learning_rate": 4.8945122673936235e-05, "loss": 2.0704, "step": 1124 }, { "epoch": 9.0, "grad_norm": 29.058765411376953, "learning_rate": 4.8939196396823514e-05, "loss": 1.3963, "step": 1125 }, { "epoch": 9.0, "eval_loss": 1.942650318145752, "eval_map": 0.0947, "eval_map_50": 0.2038, "eval_map_75": 0.0788, "eval_map_Coverall": 0.3381, "eval_map_Face_Shield": 0.0172, "eval_map_Gloves": 0.0348, "eval_map_Goggles": 0.0, "eval_map_Mask": 0.0834, "eval_map_large": 0.1074, "eval_map_medium": 0.0383, "eval_map_small": 0.063, "eval_mar_1": 0.1057, "eval_mar_10": 0.2003, "eval_mar_100": 0.2241, "eval_mar_100_Coverall": 0.5844, "eval_mar_100_Face_Shield": 0.0353, "eval_mar_100_Gloves": 0.218, "eval_mar_100_Goggles": 0.0, "eval_mar_100_Mask": 0.2827, "eval_mar_large": 0.2218, "eval_mar_medium": 0.1355, "eval_mar_small": 0.0753, "eval_runtime": 1.8209, "eval_samples_per_second": 15.926, "eval_steps_per_second": 1.098, "step": 1125 }, { "epoch": 9.008, "grad_norm": 27.3416690826416, "learning_rate": 4.89332701197108e-05, "loss": 1.9942, "step": 1126 }, { "epoch": 9.016, "grad_norm": 30.457029342651367, "learning_rate": 4.8927343842598085e-05, "loss": 1.6623, "step": 1127 }, { "epoch": 9.024, "grad_norm": 34.69789505004883, "learning_rate": 4.8921417565485364e-05, "loss": 1.5012, "step": 1128 }, { "epoch": 9.032, "grad_norm": 51.39740753173828, "learning_rate": 4.891549128837264e-05, "loss": 1.9745, "step": 1129 }, { "epoch": 9.04, "grad_norm": 62.16151428222656, "learning_rate": 4.890956501125993e-05, "loss": 2.1052, "step": 1130 }, { "epoch": 9.048, "grad_norm": 19.34756851196289, "learning_rate": 4.8903638734147214e-05, "loss": 1.6059, "step": 1131 }, { "epoch": 9.056, "grad_norm": 30.461828231811523, "learning_rate": 4.889771245703449e-05, "loss": 2.0202, "step": 1132 }, { "epoch": 9.064, "grad_norm": 29.55231475830078, "learning_rate": 4.889178617992178e-05, "loss": 1.7493, "step": 1133 }, { "epoch": 9.072, "grad_norm": 37.2971305847168, "learning_rate": 4.888585990280906e-05, "loss": 2.0189, "step": 1134 }, { "epoch": 9.08, "grad_norm": 21.1416015625, "learning_rate": 4.8879933625696336e-05, "loss": 2.3141, "step": 1135 }, { "epoch": 9.088, "grad_norm": 41.29020690917969, "learning_rate": 4.887400734858362e-05, "loss": 1.884, "step": 1136 }, { "epoch": 9.096, "grad_norm": 47.77928161621094, "learning_rate": 4.886808107147091e-05, "loss": 2.007, "step": 1137 }, { "epoch": 9.104, "grad_norm": 23.063655853271484, "learning_rate": 4.8862154794358186e-05, "loss": 1.6878, "step": 1138 }, { "epoch": 9.112, "grad_norm": 33.03241729736328, "learning_rate": 4.8856228517245465e-05, "loss": 2.0177, "step": 1139 }, { "epoch": 9.12, "grad_norm": 32.84242248535156, "learning_rate": 4.885030224013275e-05, "loss": 1.5434, "step": 1140 }, { "epoch": 9.128, "grad_norm": 44.81193542480469, "learning_rate": 4.8844375963020036e-05, "loss": 1.8546, "step": 1141 }, { "epoch": 9.136, "grad_norm": 67.37677764892578, "learning_rate": 4.8838449685907315e-05, "loss": 1.4021, "step": 1142 }, { "epoch": 9.144, "grad_norm": 46.89931869506836, "learning_rate": 4.8832523408794594e-05, "loss": 2.1383, "step": 1143 }, { "epoch": 9.152, "grad_norm": 23.400989532470703, "learning_rate": 4.882659713168188e-05, "loss": 1.8176, "step": 1144 }, { "epoch": 9.16, "grad_norm": 24.125864028930664, "learning_rate": 4.882067085456916e-05, "loss": 1.5216, "step": 1145 }, { "epoch": 9.168, "grad_norm": 29.558130264282227, "learning_rate": 4.8814744577456444e-05, "loss": 1.4689, "step": 1146 }, { "epoch": 9.176, "grad_norm": 31.405099868774414, "learning_rate": 4.880881830034373e-05, "loss": 1.1641, "step": 1147 }, { "epoch": 9.184, "grad_norm": 46.906158447265625, "learning_rate": 4.880289202323101e-05, "loss": 1.6652, "step": 1148 }, { "epoch": 9.192, "grad_norm": 69.0528335571289, "learning_rate": 4.879696574611829e-05, "loss": 1.8497, "step": 1149 }, { "epoch": 9.2, "grad_norm": 44.209922790527344, "learning_rate": 4.879103946900557e-05, "loss": 2.0038, "step": 1150 }, { "epoch": 9.208, "grad_norm": 20.586637496948242, "learning_rate": 4.878511319189286e-05, "loss": 1.668, "step": 1151 }, { "epoch": 9.216, "grad_norm": 49.33478546142578, "learning_rate": 4.877918691478014e-05, "loss": 1.4624, "step": 1152 }, { "epoch": 9.224, "grad_norm": 37.67305374145508, "learning_rate": 4.8773260637667416e-05, "loss": 1.7127, "step": 1153 }, { "epoch": 9.232, "grad_norm": 28.458240509033203, "learning_rate": 4.87673343605547e-05, "loss": 1.8067, "step": 1154 }, { "epoch": 9.24, "grad_norm": 33.849735260009766, "learning_rate": 4.876140808344199e-05, "loss": 2.0199, "step": 1155 }, { "epoch": 9.248, "grad_norm": 38.45549011230469, "learning_rate": 4.8755481806329266e-05, "loss": 1.7255, "step": 1156 }, { "epoch": 9.256, "grad_norm": 26.327531814575195, "learning_rate": 4.874955552921655e-05, "loss": 1.5886, "step": 1157 }, { "epoch": 9.264, "grad_norm": 80.21159362792969, "learning_rate": 4.874362925210383e-05, "loss": 2.1754, "step": 1158 }, { "epoch": 9.272, "grad_norm": 29.462316513061523, "learning_rate": 4.873770297499111e-05, "loss": 1.4754, "step": 1159 }, { "epoch": 9.28, "grad_norm": 31.830137252807617, "learning_rate": 4.8731776697878395e-05, "loss": 1.3618, "step": 1160 }, { "epoch": 9.288, "grad_norm": 31.731361389160156, "learning_rate": 4.872585042076568e-05, "loss": 1.6011, "step": 1161 }, { "epoch": 9.296, "grad_norm": 41.00416946411133, "learning_rate": 4.871992414365296e-05, "loss": 1.9559, "step": 1162 }, { "epoch": 9.304, "grad_norm": 26.735706329345703, "learning_rate": 4.871399786654024e-05, "loss": 2.0172, "step": 1163 }, { "epoch": 9.312, "grad_norm": 101.67491149902344, "learning_rate": 4.8708071589427524e-05, "loss": 1.8692, "step": 1164 }, { "epoch": 9.32, "grad_norm": 54.67980194091797, "learning_rate": 4.870214531231481e-05, "loss": 1.6955, "step": 1165 }, { "epoch": 9.328, "grad_norm": 26.70917320251465, "learning_rate": 4.869621903520209e-05, "loss": 2.1503, "step": 1166 }, { "epoch": 9.336, "grad_norm": 56.163780212402344, "learning_rate": 4.8690292758089374e-05, "loss": 2.1591, "step": 1167 }, { "epoch": 9.344, "grad_norm": 49.66177749633789, "learning_rate": 4.868436648097665e-05, "loss": 2.6585, "step": 1168 }, { "epoch": 9.352, "grad_norm": 31.236085891723633, "learning_rate": 4.867844020386393e-05, "loss": 1.3607, "step": 1169 }, { "epoch": 9.36, "grad_norm": 23.9475040435791, "learning_rate": 4.867251392675122e-05, "loss": 2.0225, "step": 1170 }, { "epoch": 9.368, "grad_norm": 46.010414123535156, "learning_rate": 4.86665876496385e-05, "loss": 2.1936, "step": 1171 }, { "epoch": 9.376, "grad_norm": 44.66118240356445, "learning_rate": 4.866066137252578e-05, "loss": 1.9986, "step": 1172 }, { "epoch": 9.384, "grad_norm": 70.75537872314453, "learning_rate": 4.865473509541306e-05, "loss": 1.6689, "step": 1173 }, { "epoch": 9.392, "grad_norm": 32.62424850463867, "learning_rate": 4.8648808818300346e-05, "loss": 1.7502, "step": 1174 }, { "epoch": 9.4, "grad_norm": 40.34048843383789, "learning_rate": 4.864288254118763e-05, "loss": 1.6811, "step": 1175 }, { "epoch": 9.408, "grad_norm": 28.89657974243164, "learning_rate": 4.863695626407491e-05, "loss": 1.6454, "step": 1176 }, { "epoch": 9.416, "grad_norm": 46.02423095703125, "learning_rate": 4.863102998696219e-05, "loss": 1.6699, "step": 1177 }, { "epoch": 9.424, "grad_norm": 26.84450912475586, "learning_rate": 4.8625103709849474e-05, "loss": 2.4898, "step": 1178 }, { "epoch": 9.432, "grad_norm": 71.44425964355469, "learning_rate": 4.861917743273675e-05, "loss": 2.0715, "step": 1179 }, { "epoch": 9.44, "grad_norm": 48.26110076904297, "learning_rate": 4.861325115562404e-05, "loss": 2.0676, "step": 1180 }, { "epoch": 9.448, "grad_norm": 43.79471969604492, "learning_rate": 4.8607324878511324e-05, "loss": 1.7295, "step": 1181 }, { "epoch": 9.456, "grad_norm": 63.60765838623047, "learning_rate": 4.86013986013986e-05, "loss": 2.0352, "step": 1182 }, { "epoch": 9.464, "grad_norm": 34.86804962158203, "learning_rate": 4.859547232428588e-05, "loss": 1.7387, "step": 1183 }, { "epoch": 9.472, "grad_norm": 51.83452224731445, "learning_rate": 4.858954604717317e-05, "loss": 2.045, "step": 1184 }, { "epoch": 9.48, "grad_norm": 29.592483520507812, "learning_rate": 4.858361977006045e-05, "loss": 2.0848, "step": 1185 }, { "epoch": 9.488, "grad_norm": 45.02129364013672, "learning_rate": 4.857769349294773e-05, "loss": 1.882, "step": 1186 }, { "epoch": 9.496, "grad_norm": 42.37044143676758, "learning_rate": 4.857176721583501e-05, "loss": 2.0479, "step": 1187 }, { "epoch": 9.504, "grad_norm": 37.53460693359375, "learning_rate": 4.8565840938722296e-05, "loss": 1.8353, "step": 1188 }, { "epoch": 9.512, "grad_norm": 91.59901428222656, "learning_rate": 4.855991466160958e-05, "loss": 1.7995, "step": 1189 }, { "epoch": 9.52, "grad_norm": 37.21684265136719, "learning_rate": 4.855398838449686e-05, "loss": 1.8602, "step": 1190 }, { "epoch": 9.528, "grad_norm": 38.918033599853516, "learning_rate": 4.8548062107384147e-05, "loss": 1.6541, "step": 1191 }, { "epoch": 9.536, "grad_norm": 147.4171142578125, "learning_rate": 4.8542135830271425e-05, "loss": 2.1723, "step": 1192 }, { "epoch": 9.544, "grad_norm": 135.64947509765625, "learning_rate": 4.8536209553158704e-05, "loss": 1.8776, "step": 1193 }, { "epoch": 9.552, "grad_norm": 43.76598358154297, "learning_rate": 4.853028327604599e-05, "loss": 1.9491, "step": 1194 }, { "epoch": 9.56, "grad_norm": 85.4761734008789, "learning_rate": 4.8524356998933275e-05, "loss": 2.0567, "step": 1195 }, { "epoch": 9.568, "grad_norm": 44.868221282958984, "learning_rate": 4.8518430721820554e-05, "loss": 1.7453, "step": 1196 }, { "epoch": 9.576, "grad_norm": 84.5093002319336, "learning_rate": 4.851250444470783e-05, "loss": 1.8132, "step": 1197 }, { "epoch": 9.584, "grad_norm": 75.63383483886719, "learning_rate": 4.850657816759512e-05, "loss": 1.9688, "step": 1198 }, { "epoch": 9.592, "grad_norm": 51.038150787353516, "learning_rate": 4.8500651890482404e-05, "loss": 1.8762, "step": 1199 }, { "epoch": 9.6, "grad_norm": 56.705543518066406, "learning_rate": 4.849472561336968e-05, "loss": 2.0146, "step": 1200 }, { "epoch": 9.608, "grad_norm": 92.64969635009766, "learning_rate": 4.848879933625697e-05, "loss": 2.0169, "step": 1201 }, { "epoch": 9.616, "grad_norm": 67.38916015625, "learning_rate": 4.848287305914425e-05, "loss": 2.4831, "step": 1202 }, { "epoch": 9.624, "grad_norm": 79.04689025878906, "learning_rate": 4.8476946782031526e-05, "loss": 2.9662, "step": 1203 }, { "epoch": 9.632, "grad_norm": 21.784698486328125, "learning_rate": 4.847102050491881e-05, "loss": 2.6603, "step": 1204 }, { "epoch": 9.64, "grad_norm": 81.92082977294922, "learning_rate": 4.84650942278061e-05, "loss": 2.6047, "step": 1205 }, { "epoch": 9.648, "grad_norm": 39.699371337890625, "learning_rate": 4.8459167950693376e-05, "loss": 2.437, "step": 1206 }, { "epoch": 9.656, "grad_norm": 24.474788665771484, "learning_rate": 4.8453241673580655e-05, "loss": 2.2494, "step": 1207 }, { "epoch": 9.664, "grad_norm": 37.207977294921875, "learning_rate": 4.844731539646794e-05, "loss": 1.6084, "step": 1208 }, { "epoch": 9.672, "grad_norm": 32.82332229614258, "learning_rate": 4.8441389119355226e-05, "loss": 2.1491, "step": 1209 }, { "epoch": 9.68, "grad_norm": 188.28176879882812, "learning_rate": 4.8435462842242505e-05, "loss": 2.0149, "step": 1210 }, { "epoch": 9.688, "grad_norm": 82.79517364501953, "learning_rate": 4.842953656512979e-05, "loss": 1.7579, "step": 1211 }, { "epoch": 9.696, "grad_norm": 171.7490234375, "learning_rate": 4.842361028801707e-05, "loss": 2.1284, "step": 1212 }, { "epoch": 9.704, "grad_norm": 55.73076629638672, "learning_rate": 4.8417684010904355e-05, "loss": 2.0029, "step": 1213 }, { "epoch": 9.712, "grad_norm": 53.848934173583984, "learning_rate": 4.8411757733791634e-05, "loss": 1.7167, "step": 1214 }, { "epoch": 9.72, "grad_norm": 31.017879486083984, "learning_rate": 4.840583145667892e-05, "loss": 2.8422, "step": 1215 }, { "epoch": 9.728, "grad_norm": 85.4967041015625, "learning_rate": 4.83999051795662e-05, "loss": 1.912, "step": 1216 }, { "epoch": 9.736, "grad_norm": 33.75318145751953, "learning_rate": 4.839397890245348e-05, "loss": 1.8488, "step": 1217 }, { "epoch": 9.744, "grad_norm": 48.94841003417969, "learning_rate": 4.838805262534076e-05, "loss": 1.9586, "step": 1218 }, { "epoch": 9.752, "grad_norm": 34.455997467041016, "learning_rate": 4.838212634822805e-05, "loss": 2.1138, "step": 1219 }, { "epoch": 9.76, "grad_norm": 125.11497497558594, "learning_rate": 4.837620007111533e-05, "loss": 1.6701, "step": 1220 }, { "epoch": 9.768, "grad_norm": 86.15340423583984, "learning_rate": 4.8370273794002606e-05, "loss": 1.8988, "step": 1221 }, { "epoch": 9.776, "grad_norm": 53.8587760925293, "learning_rate": 4.836434751688989e-05, "loss": 2.1069, "step": 1222 }, { "epoch": 9.784, "grad_norm": 53.30081558227539, "learning_rate": 4.835842123977718e-05, "loss": 2.3663, "step": 1223 }, { "epoch": 9.792, "grad_norm": 52.06148147583008, "learning_rate": 4.8352494962664456e-05, "loss": 1.8507, "step": 1224 }, { "epoch": 9.8, "grad_norm": 48.96346664428711, "learning_rate": 4.834656868555174e-05, "loss": 2.6815, "step": 1225 }, { "epoch": 9.808, "grad_norm": 85.33289337158203, "learning_rate": 4.834064240843902e-05, "loss": 1.8434, "step": 1226 }, { "epoch": 9.816, "grad_norm": 33.9375114440918, "learning_rate": 4.83347161313263e-05, "loss": 2.4215, "step": 1227 }, { "epoch": 9.824, "grad_norm": 47.70186996459961, "learning_rate": 4.8328789854213585e-05, "loss": 1.8517, "step": 1228 }, { "epoch": 9.832, "grad_norm": 57.4036750793457, "learning_rate": 4.832286357710087e-05, "loss": 1.9223, "step": 1229 }, { "epoch": 9.84, "grad_norm": 40.546775817871094, "learning_rate": 4.831693729998815e-05, "loss": 2.5264, "step": 1230 }, { "epoch": 9.848, "grad_norm": 52.271514892578125, "learning_rate": 4.831101102287543e-05, "loss": 1.6619, "step": 1231 }, { "epoch": 9.856, "grad_norm": 36.91080093383789, "learning_rate": 4.8305084745762714e-05, "loss": 1.9346, "step": 1232 }, { "epoch": 9.864, "grad_norm": 40.25997543334961, "learning_rate": 4.829915846865e-05, "loss": 1.6167, "step": 1233 }, { "epoch": 9.872, "grad_norm": 76.33706665039062, "learning_rate": 4.829323219153728e-05, "loss": 3.7105, "step": 1234 }, { "epoch": 9.88, "grad_norm": 23.079967498779297, "learning_rate": 4.8287305914424564e-05, "loss": 1.7504, "step": 1235 }, { "epoch": 9.888, "grad_norm": 22.094175338745117, "learning_rate": 4.828137963731184e-05, "loss": 1.6899, "step": 1236 }, { "epoch": 9.896, "grad_norm": 96.94627380371094, "learning_rate": 4.827545336019912e-05, "loss": 1.6834, "step": 1237 }, { "epoch": 9.904, "grad_norm": 45.34838104248047, "learning_rate": 4.826952708308641e-05, "loss": 1.5336, "step": 1238 }, { "epoch": 9.912, "grad_norm": 24.403783798217773, "learning_rate": 4.826360080597369e-05, "loss": 1.8389, "step": 1239 }, { "epoch": 9.92, "grad_norm": 75.87189483642578, "learning_rate": 4.825767452886097e-05, "loss": 2.1565, "step": 1240 }, { "epoch": 9.928, "grad_norm": 239.28749084472656, "learning_rate": 4.825174825174825e-05, "loss": 2.2178, "step": 1241 }, { "epoch": 9.936, "grad_norm": 47.41549301147461, "learning_rate": 4.8245821974635536e-05, "loss": 2.0473, "step": 1242 }, { "epoch": 9.943999999999999, "grad_norm": 48.69157791137695, "learning_rate": 4.823989569752282e-05, "loss": 2.2453, "step": 1243 }, { "epoch": 9.952, "grad_norm": 34.6706657409668, "learning_rate": 4.82339694204101e-05, "loss": 1.7139, "step": 1244 }, { "epoch": 9.96, "grad_norm": 34.88042068481445, "learning_rate": 4.8228043143297386e-05, "loss": 1.6236, "step": 1245 }, { "epoch": 9.968, "grad_norm": 47.83650207519531, "learning_rate": 4.8222116866184665e-05, "loss": 1.8254, "step": 1246 }, { "epoch": 9.975999999999999, "grad_norm": 42.02348709106445, "learning_rate": 4.821619058907195e-05, "loss": 1.9557, "step": 1247 }, { "epoch": 9.984, "grad_norm": 765.8471069335938, "learning_rate": 4.821026431195923e-05, "loss": 2.2365, "step": 1248 }, { "epoch": 9.992, "grad_norm": 36.63481140136719, "learning_rate": 4.8204338034846515e-05, "loss": 1.5863, "step": 1249 }, { "epoch": 10.0, "grad_norm": 56.880271911621094, "learning_rate": 4.819841175773379e-05, "loss": 1.8243, "step": 1250 }, { "epoch": 10.0, "eval_loss": 2.010802745819092, "eval_map": 0.0866, "eval_map_50": 0.2048, "eval_map_75": 0.0575, "eval_map_Coverall": 0.2133, "eval_map_Face_Shield": 0.0548, "eval_map_Gloves": 0.0567, "eval_map_Goggles": 0.0, "eval_map_Mask": 0.108, "eval_map_large": 0.108, "eval_map_medium": 0.0434, "eval_map_small": 0.0514, "eval_mar_1": 0.1221, "eval_mar_10": 0.23, "eval_mar_100": 0.2527, "eval_mar_100_Coverall": 0.5778, "eval_mar_100_Face_Shield": 0.1824, "eval_mar_100_Gloves": 0.1918, "eval_mar_100_Goggles": 0.0, "eval_mar_100_Mask": 0.3115, "eval_mar_large": 0.2834, "eval_mar_medium": 0.1337, "eval_mar_small": 0.0799, "eval_runtime": 1.8209, "eval_samples_per_second": 15.926, "eval_steps_per_second": 1.098, "step": 1250 }, { "epoch": 10.008, "grad_norm": 29.056364059448242, "learning_rate": 4.819248548062107e-05, "loss": 1.7601, "step": 1251 }, { "epoch": 10.016, "grad_norm": 30.04147720336914, "learning_rate": 4.818655920350836e-05, "loss": 2.0241, "step": 1252 }, { "epoch": 10.024, "grad_norm": 38.6876220703125, "learning_rate": 4.818063292639564e-05, "loss": 2.2045, "step": 1253 }, { "epoch": 10.032, "grad_norm": 37.01424026489258, "learning_rate": 4.817470664928292e-05, "loss": 1.7768, "step": 1254 }, { "epoch": 10.04, "grad_norm": 32.71819305419922, "learning_rate": 4.81687803721702e-05, "loss": 2.2535, "step": 1255 }, { "epoch": 10.048, "grad_norm": 43.834495544433594, "learning_rate": 4.8162854095057487e-05, "loss": 1.6197, "step": 1256 }, { "epoch": 10.056, "grad_norm": 38.50288391113281, "learning_rate": 4.815692781794477e-05, "loss": 1.5773, "step": 1257 }, { "epoch": 10.064, "grad_norm": 19.97999382019043, "learning_rate": 4.815100154083205e-05, "loss": 1.8258, "step": 1258 }, { "epoch": 10.072, "grad_norm": 48.907249450683594, "learning_rate": 4.8145075263719337e-05, "loss": 1.992, "step": 1259 }, { "epoch": 10.08, "grad_norm": 32.518577575683594, "learning_rate": 4.8139148986606615e-05, "loss": 1.6864, "step": 1260 }, { "epoch": 10.088, "grad_norm": 35.5919189453125, "learning_rate": 4.8133222709493894e-05, "loss": 1.7146, "step": 1261 }, { "epoch": 10.096, "grad_norm": 34.99994659423828, "learning_rate": 4.812729643238118e-05, "loss": 1.5309, "step": 1262 }, { "epoch": 10.104, "grad_norm": 40.966156005859375, "learning_rate": 4.8121370155268465e-05, "loss": 1.9239, "step": 1263 }, { "epoch": 10.112, "grad_norm": 57.427764892578125, "learning_rate": 4.8115443878155744e-05, "loss": 2.3837, "step": 1264 }, { "epoch": 10.12, "grad_norm": 47.354862213134766, "learning_rate": 4.810951760104302e-05, "loss": 1.4228, "step": 1265 }, { "epoch": 10.128, "grad_norm": 57.2833137512207, "learning_rate": 4.810359132393031e-05, "loss": 1.9891, "step": 1266 }, { "epoch": 10.136, "grad_norm": 26.429399490356445, "learning_rate": 4.8097665046817594e-05, "loss": 1.53, "step": 1267 }, { "epoch": 10.144, "grad_norm": 25.162538528442383, "learning_rate": 4.809173876970487e-05, "loss": 1.7611, "step": 1268 }, { "epoch": 10.152, "grad_norm": 183.67144775390625, "learning_rate": 4.808581249259216e-05, "loss": 1.945, "step": 1269 }, { "epoch": 10.16, "grad_norm": 22.501575469970703, "learning_rate": 4.807988621547944e-05, "loss": 1.3937, "step": 1270 }, { "epoch": 10.168, "grad_norm": 283.3805236816406, "learning_rate": 4.8073959938366716e-05, "loss": 1.75, "step": 1271 }, { "epoch": 10.176, "grad_norm": 62.49675750732422, "learning_rate": 4.8068033661254e-05, "loss": 2.0249, "step": 1272 }, { "epoch": 10.184, "grad_norm": 122.0244369506836, "learning_rate": 4.806210738414129e-05, "loss": 1.7718, "step": 1273 }, { "epoch": 10.192, "grad_norm": 25.75641632080078, "learning_rate": 4.8056181107028566e-05, "loss": 1.598, "step": 1274 }, { "epoch": 10.2, "grad_norm": 30.83232879638672, "learning_rate": 4.8050254829915845e-05, "loss": 2.5063, "step": 1275 }, { "epoch": 10.208, "grad_norm": 43.39324188232422, "learning_rate": 4.804432855280313e-05, "loss": 1.4026, "step": 1276 }, { "epoch": 10.216, "grad_norm": 565.452880859375, "learning_rate": 4.8038402275690416e-05, "loss": 3.2924, "step": 1277 }, { "epoch": 10.224, "grad_norm": 40.091861724853516, "learning_rate": 4.8032475998577695e-05, "loss": 1.5767, "step": 1278 }, { "epoch": 10.232, "grad_norm": 65.28549194335938, "learning_rate": 4.802654972146498e-05, "loss": 1.7182, "step": 1279 }, { "epoch": 10.24, "grad_norm": 32.430747985839844, "learning_rate": 4.802062344435226e-05, "loss": 1.7359, "step": 1280 }, { "epoch": 10.248, "grad_norm": 30.55211067199707, "learning_rate": 4.8014697167239545e-05, "loss": 2.9295, "step": 1281 }, { "epoch": 10.256, "grad_norm": 31.4814510345459, "learning_rate": 4.8008770890126824e-05, "loss": 1.9716, "step": 1282 }, { "epoch": 10.264, "grad_norm": 21.919506072998047, "learning_rate": 4.800284461301411e-05, "loss": 1.5169, "step": 1283 }, { "epoch": 10.272, "grad_norm": 40.87081527709961, "learning_rate": 4.799691833590139e-05, "loss": 1.395, "step": 1284 }, { "epoch": 10.28, "grad_norm": 43.488372802734375, "learning_rate": 4.799099205878867e-05, "loss": 2.1427, "step": 1285 }, { "epoch": 10.288, "grad_norm": 19.378339767456055, "learning_rate": 4.798506578167595e-05, "loss": 1.8397, "step": 1286 }, { "epoch": 10.296, "grad_norm": 31.60783576965332, "learning_rate": 4.797913950456324e-05, "loss": 1.6352, "step": 1287 }, { "epoch": 10.304, "grad_norm": 40.553646087646484, "learning_rate": 4.797321322745052e-05, "loss": 1.6582, "step": 1288 }, { "epoch": 10.312, "grad_norm": 37.7210693359375, "learning_rate": 4.7967286950337796e-05, "loss": 1.46, "step": 1289 }, { "epoch": 10.32, "grad_norm": 51.02965545654297, "learning_rate": 4.796136067322508e-05, "loss": 1.5537, "step": 1290 }, { "epoch": 10.328, "grad_norm": 39.85866928100586, "learning_rate": 4.795543439611237e-05, "loss": 1.7129, "step": 1291 }, { "epoch": 10.336, "grad_norm": 72.22309875488281, "learning_rate": 4.7949508118999646e-05, "loss": 2.0791, "step": 1292 }, { "epoch": 10.344, "grad_norm": 64.01319885253906, "learning_rate": 4.794358184188693e-05, "loss": 1.8587, "step": 1293 }, { "epoch": 10.352, "grad_norm": 64.42511749267578, "learning_rate": 4.793765556477421e-05, "loss": 2.2719, "step": 1294 }, { "epoch": 10.36, "grad_norm": 51.31149673461914, "learning_rate": 4.793172928766149e-05, "loss": 1.7774, "step": 1295 }, { "epoch": 10.368, "grad_norm": 60.08006286621094, "learning_rate": 4.7925803010548775e-05, "loss": 2.2919, "step": 1296 }, { "epoch": 10.376, "grad_norm": 68.57185363769531, "learning_rate": 4.791987673343606e-05, "loss": 2.5464, "step": 1297 }, { "epoch": 10.384, "grad_norm": 149.6517333984375, "learning_rate": 4.791395045632334e-05, "loss": 2.6686, "step": 1298 }, { "epoch": 10.392, "grad_norm": 28.809980392456055, "learning_rate": 4.790802417921062e-05, "loss": 1.8034, "step": 1299 }, { "epoch": 10.4, "grad_norm": 19.482213973999023, "learning_rate": 4.7902097902097904e-05, "loss": 1.6715, "step": 1300 }, { "epoch": 10.408, "grad_norm": 34.97419738769531, "learning_rate": 4.789617162498519e-05, "loss": 1.3133, "step": 1301 }, { "epoch": 10.416, "grad_norm": 19.34789276123047, "learning_rate": 4.789024534787247e-05, "loss": 1.8393, "step": 1302 }, { "epoch": 10.424, "grad_norm": 39.823760986328125, "learning_rate": 4.7884319070759754e-05, "loss": 1.7573, "step": 1303 }, { "epoch": 10.432, "grad_norm": 97.40597534179688, "learning_rate": 4.787839279364703e-05, "loss": 1.515, "step": 1304 }, { "epoch": 10.44, "grad_norm": 20.364654541015625, "learning_rate": 4.787246651653431e-05, "loss": 1.6941, "step": 1305 }, { "epoch": 10.448, "grad_norm": 17.013036727905273, "learning_rate": 4.78665402394216e-05, "loss": 1.8359, "step": 1306 }, { "epoch": 10.456, "grad_norm": 28.886465072631836, "learning_rate": 4.786061396230888e-05, "loss": 2.0121, "step": 1307 }, { "epoch": 10.464, "grad_norm": 29.34931182861328, "learning_rate": 4.785468768519616e-05, "loss": 1.554, "step": 1308 }, { "epoch": 10.472, "grad_norm": 23.38097381591797, "learning_rate": 4.784876140808344e-05, "loss": 1.4473, "step": 1309 }, { "epoch": 10.48, "grad_norm": 42.35398864746094, "learning_rate": 4.7842835130970726e-05, "loss": 1.6516, "step": 1310 }, { "epoch": 10.488, "grad_norm": 25.179243087768555, "learning_rate": 4.783690885385801e-05, "loss": 1.6643, "step": 1311 }, { "epoch": 10.496, "grad_norm": 26.610172271728516, "learning_rate": 4.783098257674529e-05, "loss": 1.5588, "step": 1312 }, { "epoch": 10.504, "grad_norm": 22.81682014465332, "learning_rate": 4.7825056299632576e-05, "loss": 1.6542, "step": 1313 }, { "epoch": 10.512, "grad_norm": 27.420001983642578, "learning_rate": 4.7819130022519855e-05, "loss": 1.6132, "step": 1314 }, { "epoch": 10.52, "grad_norm": 42.08549118041992, "learning_rate": 4.781320374540714e-05, "loss": 1.606, "step": 1315 }, { "epoch": 10.528, "grad_norm": 33.82231903076172, "learning_rate": 4.780727746829442e-05, "loss": 1.7312, "step": 1316 }, { "epoch": 10.536, "grad_norm": 26.853057861328125, "learning_rate": 4.7801351191181705e-05, "loss": 1.5147, "step": 1317 }, { "epoch": 10.544, "grad_norm": 143.4208526611328, "learning_rate": 4.779542491406899e-05, "loss": 1.2178, "step": 1318 }, { "epoch": 10.552, "grad_norm": 35.75905990600586, "learning_rate": 4.778949863695626e-05, "loss": 1.7449, "step": 1319 }, { "epoch": 10.56, "grad_norm": 24.501861572265625, "learning_rate": 4.778357235984355e-05, "loss": 2.0151, "step": 1320 }, { "epoch": 10.568, "grad_norm": 25.977731704711914, "learning_rate": 4.7777646082730833e-05, "loss": 1.9554, "step": 1321 }, { "epoch": 10.576, "grad_norm": 24.124818801879883, "learning_rate": 4.777171980561811e-05, "loss": 1.9, "step": 1322 }, { "epoch": 10.584, "grad_norm": 36.35528564453125, "learning_rate": 4.776579352850539e-05, "loss": 1.7509, "step": 1323 }, { "epoch": 10.592, "grad_norm": 30.259721755981445, "learning_rate": 4.775986725139268e-05, "loss": 1.8235, "step": 1324 }, { "epoch": 10.6, "grad_norm": 39.28286361694336, "learning_rate": 4.775394097427996e-05, "loss": 1.9062, "step": 1325 }, { "epoch": 10.608, "grad_norm": 28.93393898010254, "learning_rate": 4.774801469716724e-05, "loss": 2.3934, "step": 1326 }, { "epoch": 10.616, "grad_norm": 28.859100341796875, "learning_rate": 4.774208842005453e-05, "loss": 1.8546, "step": 1327 }, { "epoch": 10.624, "grad_norm": 27.210485458374023, "learning_rate": 4.7736162142941805e-05, "loss": 1.6766, "step": 1328 }, { "epoch": 10.632, "grad_norm": 31.99571418762207, "learning_rate": 4.7730235865829084e-05, "loss": 1.7302, "step": 1329 }, { "epoch": 10.64, "grad_norm": 52.51451873779297, "learning_rate": 4.772430958871637e-05, "loss": 2.1534, "step": 1330 }, { "epoch": 10.648, "grad_norm": 50.77283477783203, "learning_rate": 4.7718383311603655e-05, "loss": 1.5119, "step": 1331 }, { "epoch": 10.656, "grad_norm": 27.31285858154297, "learning_rate": 4.7712457034490934e-05, "loss": 1.909, "step": 1332 }, { "epoch": 10.664, "grad_norm": 98.98836517333984, "learning_rate": 4.770653075737821e-05, "loss": 1.6785, "step": 1333 }, { "epoch": 10.672, "grad_norm": 28.903759002685547, "learning_rate": 4.77006044802655e-05, "loss": 1.5157, "step": 1334 }, { "epoch": 10.68, "grad_norm": 51.76710510253906, "learning_rate": 4.7694678203152784e-05, "loss": 1.5494, "step": 1335 }, { "epoch": 10.688, "grad_norm": 56.49921798706055, "learning_rate": 4.768875192604006e-05, "loss": 1.6737, "step": 1336 }, { "epoch": 10.696, "grad_norm": 29.67377471923828, "learning_rate": 4.768282564892735e-05, "loss": 1.7636, "step": 1337 }, { "epoch": 10.704, "grad_norm": 27.22374153137207, "learning_rate": 4.767689937181463e-05, "loss": 1.7136, "step": 1338 }, { "epoch": 10.712, "grad_norm": 24.05215072631836, "learning_rate": 4.7670973094701906e-05, "loss": 1.9508, "step": 1339 }, { "epoch": 10.72, "grad_norm": 22.77069664001465, "learning_rate": 4.766504681758919e-05, "loss": 1.6281, "step": 1340 }, { "epoch": 10.728, "grad_norm": 30.899065017700195, "learning_rate": 4.765912054047648e-05, "loss": 1.763, "step": 1341 }, { "epoch": 10.736, "grad_norm": 30.229389190673828, "learning_rate": 4.7653194263363756e-05, "loss": 2.2466, "step": 1342 }, { "epoch": 10.744, "grad_norm": 26.33673095703125, "learning_rate": 4.7647267986251035e-05, "loss": 1.2131, "step": 1343 }, { "epoch": 10.752, "grad_norm": 70.04798889160156, "learning_rate": 4.764134170913832e-05, "loss": 1.9185, "step": 1344 }, { "epoch": 10.76, "grad_norm": 27.02828598022461, "learning_rate": 4.7635415432025606e-05, "loss": 1.9536, "step": 1345 }, { "epoch": 10.768, "grad_norm": 33.456092834472656, "learning_rate": 4.7629489154912885e-05, "loss": 2.0614, "step": 1346 }, { "epoch": 10.776, "grad_norm": 64.6158218383789, "learning_rate": 4.762356287780017e-05, "loss": 2.2933, "step": 1347 }, { "epoch": 10.784, "grad_norm": 17.003713607788086, "learning_rate": 4.761763660068745e-05, "loss": 1.6144, "step": 1348 }, { "epoch": 10.792, "grad_norm": 24.295452117919922, "learning_rate": 4.7611710323574735e-05, "loss": 2.0844, "step": 1349 }, { "epoch": 10.8, "grad_norm": 32.735687255859375, "learning_rate": 4.7605784046462014e-05, "loss": 2.0155, "step": 1350 }, { "epoch": 10.808, "grad_norm": 92.44884490966797, "learning_rate": 4.75998577693493e-05, "loss": 1.4665, "step": 1351 }, { "epoch": 10.816, "grad_norm": 48.09383773803711, "learning_rate": 4.7593931492236585e-05, "loss": 2.1569, "step": 1352 }, { "epoch": 10.824, "grad_norm": 30.782424926757812, "learning_rate": 4.758800521512386e-05, "loss": 1.9699, "step": 1353 }, { "epoch": 10.832, "grad_norm": 33.10237503051758, "learning_rate": 4.758207893801114e-05, "loss": 2.3016, "step": 1354 }, { "epoch": 10.84, "grad_norm": 30.762378692626953, "learning_rate": 4.757615266089843e-05, "loss": 1.6973, "step": 1355 }, { "epoch": 10.848, "grad_norm": 74.0185546875, "learning_rate": 4.757022638378571e-05, "loss": 1.8295, "step": 1356 }, { "epoch": 10.856, "grad_norm": 74.78443908691406, "learning_rate": 4.7564300106672986e-05, "loss": 1.9734, "step": 1357 }, { "epoch": 10.864, "grad_norm": 100.7544937133789, "learning_rate": 4.755837382956027e-05, "loss": 1.6933, "step": 1358 }, { "epoch": 10.872, "grad_norm": 33.267513275146484, "learning_rate": 4.755244755244756e-05, "loss": 1.699, "step": 1359 }, { "epoch": 10.88, "grad_norm": 35.40420913696289, "learning_rate": 4.7546521275334836e-05, "loss": 1.7901, "step": 1360 }, { "epoch": 10.888, "grad_norm": 71.82662963867188, "learning_rate": 4.754059499822212e-05, "loss": 1.5992, "step": 1361 }, { "epoch": 10.896, "grad_norm": 24.907604217529297, "learning_rate": 4.75346687211094e-05, "loss": 1.4868, "step": 1362 }, { "epoch": 10.904, "grad_norm": 59.716758728027344, "learning_rate": 4.752874244399668e-05, "loss": 1.7167, "step": 1363 }, { "epoch": 10.912, "grad_norm": 65.3953628540039, "learning_rate": 4.7522816166883965e-05, "loss": 2.2165, "step": 1364 }, { "epoch": 10.92, "grad_norm": 55.83665084838867, "learning_rate": 4.751688988977125e-05, "loss": 1.5449, "step": 1365 }, { "epoch": 10.928, "grad_norm": 30.558799743652344, "learning_rate": 4.751096361265853e-05, "loss": 1.514, "step": 1366 }, { "epoch": 10.936, "grad_norm": 23.05535316467285, "learning_rate": 4.750503733554581e-05, "loss": 1.5904, "step": 1367 }, { "epoch": 10.943999999999999, "grad_norm": 24.74164581298828, "learning_rate": 4.7499111058433094e-05, "loss": 1.9917, "step": 1368 }, { "epoch": 10.952, "grad_norm": 37.21934509277344, "learning_rate": 4.749318478132038e-05, "loss": 1.7183, "step": 1369 }, { "epoch": 10.96, "grad_norm": 30.621644973754883, "learning_rate": 4.748725850420766e-05, "loss": 1.741, "step": 1370 }, { "epoch": 10.968, "grad_norm": 24.91676902770996, "learning_rate": 4.7481332227094944e-05, "loss": 1.5695, "step": 1371 }, { "epoch": 10.975999999999999, "grad_norm": 22.150747299194336, "learning_rate": 4.747540594998222e-05, "loss": 2.2226, "step": 1372 }, { "epoch": 10.984, "grad_norm": 35.229217529296875, "learning_rate": 4.74694796728695e-05, "loss": 1.6915, "step": 1373 }, { "epoch": 10.992, "grad_norm": 26.78944206237793, "learning_rate": 4.746355339575679e-05, "loss": 2.2355, "step": 1374 }, { "epoch": 11.0, "grad_norm": 20.02462387084961, "learning_rate": 4.745762711864407e-05, "loss": 1.8699, "step": 1375 }, { "epoch": 11.0, "eval_loss": 1.8328379392623901, "eval_map": 0.128, "eval_map_50": 0.2749, "eval_map_75": 0.0937, "eval_map_Coverall": 0.305, "eval_map_Face_Shield": 0.1074, "eval_map_Gloves": 0.0628, "eval_map_Goggles": 0.0, "eval_map_Mask": 0.1647, "eval_map_large": 0.1611, "eval_map_medium": 0.0638, "eval_map_small": 0.1385, "eval_mar_1": 0.156, "eval_mar_10": 0.3023, "eval_mar_100": 0.326, "eval_mar_100_Coverall": 0.6578, "eval_mar_100_Face_Shield": 0.3294, "eval_mar_100_Gloves": 0.2426, "eval_mar_100_Goggles": 0.0, "eval_mar_100_Mask": 0.4, "eval_mar_large": 0.3952, "eval_mar_medium": 0.1633, "eval_mar_small": 0.2124, "eval_runtime": 1.815, "eval_samples_per_second": 15.978, "eval_steps_per_second": 1.102, "step": 1375 }, { "epoch": 11.008, "grad_norm": 38.922733306884766, "learning_rate": 4.745170084153135e-05, "loss": 1.4984, "step": 1376 }, { "epoch": 11.016, "grad_norm": 22.319351196289062, "learning_rate": 4.744577456441863e-05, "loss": 2.0862, "step": 1377 }, { "epoch": 11.024, "grad_norm": 32.69185256958008, "learning_rate": 4.7439848287305916e-05, "loss": 1.8952, "step": 1378 }, { "epoch": 11.032, "grad_norm": 23.71277618408203, "learning_rate": 4.74339220101932e-05, "loss": 2.0501, "step": 1379 }, { "epoch": 11.04, "grad_norm": 63.00109100341797, "learning_rate": 4.742799573308048e-05, "loss": 1.8383, "step": 1380 }, { "epoch": 11.048, "grad_norm": 52.013214111328125, "learning_rate": 4.7422069455967766e-05, "loss": 2.5321, "step": 1381 }, { "epoch": 11.056, "grad_norm": 128.77964782714844, "learning_rate": 4.7416143178855045e-05, "loss": 2.2165, "step": 1382 }, { "epoch": 11.064, "grad_norm": 22.31953239440918, "learning_rate": 4.741021690174233e-05, "loss": 1.5746, "step": 1383 }, { "epoch": 11.072, "grad_norm": 34.9461669921875, "learning_rate": 4.740429062462961e-05, "loss": 1.7421, "step": 1384 }, { "epoch": 11.08, "grad_norm": 18.143980026245117, "learning_rate": 4.7398364347516895e-05, "loss": 1.5137, "step": 1385 }, { "epoch": 11.088, "grad_norm": 41.04205322265625, "learning_rate": 4.739243807040418e-05, "loss": 1.5847, "step": 1386 }, { "epoch": 11.096, "grad_norm": 60.41270065307617, "learning_rate": 4.738651179329145e-05, "loss": 1.5277, "step": 1387 }, { "epoch": 11.104, "grad_norm": 49.1497917175293, "learning_rate": 4.738058551617874e-05, "loss": 1.5151, "step": 1388 }, { "epoch": 11.112, "grad_norm": 24.962682723999023, "learning_rate": 4.7374659239066023e-05, "loss": 2.4087, "step": 1389 }, { "epoch": 11.12, "grad_norm": 20.85420036315918, "learning_rate": 4.73687329619533e-05, "loss": 1.3841, "step": 1390 }, { "epoch": 11.128, "grad_norm": 23.552030563354492, "learning_rate": 4.736280668484058e-05, "loss": 1.6073, "step": 1391 }, { "epoch": 11.136, "grad_norm": 17.5859375, "learning_rate": 4.735688040772787e-05, "loss": 1.8259, "step": 1392 }, { "epoch": 11.144, "grad_norm": 40.24345779418945, "learning_rate": 4.735095413061515e-05, "loss": 2.1105, "step": 1393 }, { "epoch": 11.152, "grad_norm": 89.42118835449219, "learning_rate": 4.734502785350243e-05, "loss": 1.5387, "step": 1394 }, { "epoch": 11.16, "grad_norm": 18.74957275390625, "learning_rate": 4.733910157638972e-05, "loss": 2.1155, "step": 1395 }, { "epoch": 11.168, "grad_norm": 47.433650970458984, "learning_rate": 4.7333175299276996e-05, "loss": 2.0068, "step": 1396 }, { "epoch": 11.176, "grad_norm": 29.294540405273438, "learning_rate": 4.7327249022164274e-05, "loss": 1.7474, "step": 1397 }, { "epoch": 11.184, "grad_norm": 16.281774520874023, "learning_rate": 4.732132274505156e-05, "loss": 1.922, "step": 1398 }, { "epoch": 11.192, "grad_norm": 21.415607452392578, "learning_rate": 4.7315396467938846e-05, "loss": 1.845, "step": 1399 }, { "epoch": 11.2, "grad_norm": 44.93983840942383, "learning_rate": 4.7309470190826124e-05, "loss": 1.8019, "step": 1400 }, { "epoch": 11.208, "grad_norm": 36.22068786621094, "learning_rate": 4.73035439137134e-05, "loss": 1.724, "step": 1401 }, { "epoch": 11.216, "grad_norm": 63.61293029785156, "learning_rate": 4.729761763660069e-05, "loss": 1.7823, "step": 1402 }, { "epoch": 11.224, "grad_norm": 26.535490036010742, "learning_rate": 4.7291691359487974e-05, "loss": 1.735, "step": 1403 }, { "epoch": 11.232, "grad_norm": 25.46103286743164, "learning_rate": 4.728576508237525e-05, "loss": 1.4448, "step": 1404 }, { "epoch": 11.24, "grad_norm": 25.76384735107422, "learning_rate": 4.727983880526254e-05, "loss": 1.8532, "step": 1405 }, { "epoch": 11.248, "grad_norm": 35.58079147338867, "learning_rate": 4.727391252814982e-05, "loss": 1.5959, "step": 1406 }, { "epoch": 11.256, "grad_norm": 30.731307983398438, "learning_rate": 4.7267986251037096e-05, "loss": 1.6643, "step": 1407 }, { "epoch": 11.264, "grad_norm": 19.23733139038086, "learning_rate": 4.726205997392438e-05, "loss": 1.3306, "step": 1408 }, { "epoch": 11.272, "grad_norm": 35.38924026489258, "learning_rate": 4.725613369681167e-05, "loss": 1.6904, "step": 1409 }, { "epoch": 11.28, "grad_norm": 23.92121124267578, "learning_rate": 4.7250207419698946e-05, "loss": 1.8206, "step": 1410 }, { "epoch": 11.288, "grad_norm": 34.4172477722168, "learning_rate": 4.7244281142586225e-05, "loss": 1.8504, "step": 1411 }, { "epoch": 11.296, "grad_norm": 25.757183074951172, "learning_rate": 4.723835486547351e-05, "loss": 1.7345, "step": 1412 }, { "epoch": 11.304, "grad_norm": 19.934473037719727, "learning_rate": 4.7232428588360796e-05, "loss": 1.5035, "step": 1413 }, { "epoch": 11.312, "grad_norm": 43.32765197753906, "learning_rate": 4.7226502311248075e-05, "loss": 1.755, "step": 1414 }, { "epoch": 11.32, "grad_norm": 28.88433265686035, "learning_rate": 4.722057603413536e-05, "loss": 1.9117, "step": 1415 }, { "epoch": 11.328, "grad_norm": 30.097232818603516, "learning_rate": 4.721464975702264e-05, "loss": 3.1135, "step": 1416 }, { "epoch": 11.336, "grad_norm": 27.389636993408203, "learning_rate": 4.7208723479909925e-05, "loss": 1.8786, "step": 1417 }, { "epoch": 11.344, "grad_norm": 42.27301025390625, "learning_rate": 4.7202797202797204e-05, "loss": 1.9347, "step": 1418 }, { "epoch": 11.352, "grad_norm": 29.855859756469727, "learning_rate": 4.719687092568449e-05, "loss": 1.5926, "step": 1419 }, { "epoch": 11.36, "grad_norm": 26.40216636657715, "learning_rate": 4.7190944648571775e-05, "loss": 1.7521, "step": 1420 }, { "epoch": 11.368, "grad_norm": 34.3143196105957, "learning_rate": 4.718501837145905e-05, "loss": 1.4799, "step": 1421 }, { "epoch": 11.376, "grad_norm": 28.419551849365234, "learning_rate": 4.717909209434633e-05, "loss": 1.5231, "step": 1422 }, { "epoch": 11.384, "grad_norm": 53.59221649169922, "learning_rate": 4.717316581723362e-05, "loss": 1.2808, "step": 1423 }, { "epoch": 11.392, "grad_norm": 40.82282257080078, "learning_rate": 4.71672395401209e-05, "loss": 1.8506, "step": 1424 }, { "epoch": 11.4, "grad_norm": 31.237346649169922, "learning_rate": 4.7161313263008176e-05, "loss": 1.7603, "step": 1425 }, { "epoch": 11.408, "grad_norm": 73.38853454589844, "learning_rate": 4.715538698589546e-05, "loss": 2.2977, "step": 1426 }, { "epoch": 11.416, "grad_norm": 230.39309692382812, "learning_rate": 4.714946070878275e-05, "loss": 1.5965, "step": 1427 }, { "epoch": 11.424, "grad_norm": 34.62883377075195, "learning_rate": 4.7143534431670026e-05, "loss": 1.9345, "step": 1428 }, { "epoch": 11.432, "grad_norm": 104.52456665039062, "learning_rate": 4.713760815455731e-05, "loss": 2.0604, "step": 1429 }, { "epoch": 11.44, "grad_norm": 34.2955207824707, "learning_rate": 4.713168187744459e-05, "loss": 1.6476, "step": 1430 }, { "epoch": 11.448, "grad_norm": 33.437835693359375, "learning_rate": 4.712575560033187e-05, "loss": 1.9149, "step": 1431 }, { "epoch": 11.456, "grad_norm": 58.33608627319336, "learning_rate": 4.7119829323219155e-05, "loss": 1.549, "step": 1432 }, { "epoch": 11.464, "grad_norm": 27.62922477722168, "learning_rate": 4.711390304610644e-05, "loss": 1.7829, "step": 1433 }, { "epoch": 11.472, "grad_norm": 23.446224212646484, "learning_rate": 4.710797676899372e-05, "loss": 1.4135, "step": 1434 }, { "epoch": 11.48, "grad_norm": 46.82141876220703, "learning_rate": 4.7102050491881e-05, "loss": 3.0096, "step": 1435 }, { "epoch": 11.488, "grad_norm": 54.4076042175293, "learning_rate": 4.7096124214768284e-05, "loss": 2.2216, "step": 1436 }, { "epoch": 11.496, "grad_norm": 25.239673614501953, "learning_rate": 4.709019793765557e-05, "loss": 2.1324, "step": 1437 }, { "epoch": 11.504, "grad_norm": 27.954408645629883, "learning_rate": 4.708427166054285e-05, "loss": 1.5452, "step": 1438 }, { "epoch": 11.512, "grad_norm": 50.48488998413086, "learning_rate": 4.7078345383430134e-05, "loss": 1.7151, "step": 1439 }, { "epoch": 11.52, "grad_norm": 61.545066833496094, "learning_rate": 4.707241910631741e-05, "loss": 1.8218, "step": 1440 }, { "epoch": 11.528, "grad_norm": 55.50855255126953, "learning_rate": 4.706649282920469e-05, "loss": 1.9443, "step": 1441 }, { "epoch": 11.536, "grad_norm": 68.71734619140625, "learning_rate": 4.706056655209198e-05, "loss": 1.6279, "step": 1442 }, { "epoch": 11.544, "grad_norm": 34.36857986450195, "learning_rate": 4.705464027497926e-05, "loss": 1.8603, "step": 1443 }, { "epoch": 11.552, "grad_norm": 43.07667541503906, "learning_rate": 4.704871399786654e-05, "loss": 1.6267, "step": 1444 }, { "epoch": 11.56, "grad_norm": 33.61103820800781, "learning_rate": 4.704278772075382e-05, "loss": 2.006, "step": 1445 }, { "epoch": 11.568, "grad_norm": 47.00699234008789, "learning_rate": 4.7036861443641106e-05, "loss": 1.5437, "step": 1446 }, { "epoch": 11.576, "grad_norm": 22.49601173400879, "learning_rate": 4.703093516652839e-05, "loss": 1.5518, "step": 1447 }, { "epoch": 11.584, "grad_norm": 34.60078048706055, "learning_rate": 4.702500888941567e-05, "loss": 1.6503, "step": 1448 }, { "epoch": 11.592, "grad_norm": 26.050758361816406, "learning_rate": 4.7019082612302956e-05, "loss": 1.1971, "step": 1449 }, { "epoch": 11.6, "grad_norm": 15.103487968444824, "learning_rate": 4.7013156335190235e-05, "loss": 2.2326, "step": 1450 }, { "epoch": 11.608, "grad_norm": 31.457305908203125, "learning_rate": 4.700723005807752e-05, "loss": 1.4794, "step": 1451 }, { "epoch": 11.616, "grad_norm": 64.84506225585938, "learning_rate": 4.70013037809648e-05, "loss": 1.662, "step": 1452 }, { "epoch": 11.624, "grad_norm": 28.355775833129883, "learning_rate": 4.6995377503852085e-05, "loss": 1.9482, "step": 1453 }, { "epoch": 11.632, "grad_norm": 21.65627098083496, "learning_rate": 4.698945122673937e-05, "loss": 1.3691, "step": 1454 }, { "epoch": 11.64, "grad_norm": 34.961669921875, "learning_rate": 4.698352494962664e-05, "loss": 2.0228, "step": 1455 }, { "epoch": 11.648, "grad_norm": 43.72503662109375, "learning_rate": 4.697759867251393e-05, "loss": 2.1282, "step": 1456 }, { "epoch": 11.656, "grad_norm": 38.92758560180664, "learning_rate": 4.6971672395401214e-05, "loss": 1.8838, "step": 1457 }, { "epoch": 11.664, "grad_norm": 28.016870498657227, "learning_rate": 4.696574611828849e-05, "loss": 1.7645, "step": 1458 }, { "epoch": 11.672, "grad_norm": 33.08509063720703, "learning_rate": 4.695981984117578e-05, "loss": 1.9473, "step": 1459 }, { "epoch": 11.68, "grad_norm": 26.65176773071289, "learning_rate": 4.695389356406306e-05, "loss": 2.0393, "step": 1460 }, { "epoch": 11.688, "grad_norm": 41.73750305175781, "learning_rate": 4.694796728695034e-05, "loss": 2.0947, "step": 1461 }, { "epoch": 11.696, "grad_norm": 35.560691833496094, "learning_rate": 4.694204100983762e-05, "loss": 1.8692, "step": 1462 }, { "epoch": 11.704, "grad_norm": 38.533634185791016, "learning_rate": 4.693611473272491e-05, "loss": 1.8402, "step": 1463 }, { "epoch": 11.712, "grad_norm": 34.80143737792969, "learning_rate": 4.6930188455612186e-05, "loss": 1.5393, "step": 1464 }, { "epoch": 11.72, "grad_norm": 25.750349044799805, "learning_rate": 4.6924262178499464e-05, "loss": 1.4078, "step": 1465 }, { "epoch": 11.728, "grad_norm": 45.99150848388672, "learning_rate": 4.691833590138675e-05, "loss": 1.3392, "step": 1466 }, { "epoch": 11.736, "grad_norm": 58.25362014770508, "learning_rate": 4.6912409624274036e-05, "loss": 1.4558, "step": 1467 }, { "epoch": 11.744, "grad_norm": 36.59910202026367, "learning_rate": 4.6906483347161314e-05, "loss": 2.0667, "step": 1468 }, { "epoch": 11.752, "grad_norm": 56.64133834838867, "learning_rate": 4.690055707004859e-05, "loss": 1.5946, "step": 1469 }, { "epoch": 11.76, "grad_norm": 35.08295822143555, "learning_rate": 4.689463079293588e-05, "loss": 1.7777, "step": 1470 }, { "epoch": 11.768, "grad_norm": 29.579225540161133, "learning_rate": 4.6888704515823164e-05, "loss": 1.9638, "step": 1471 }, { "epoch": 11.776, "grad_norm": 41.46890640258789, "learning_rate": 4.688277823871044e-05, "loss": 1.9905, "step": 1472 }, { "epoch": 11.784, "grad_norm": 15.033431053161621, "learning_rate": 4.687685196159773e-05, "loss": 1.2584, "step": 1473 }, { "epoch": 11.792, "grad_norm": 29.261693954467773, "learning_rate": 4.687092568448501e-05, "loss": 2.1788, "step": 1474 }, { "epoch": 11.8, "grad_norm": 20.03814125061035, "learning_rate": 4.6864999407372287e-05, "loss": 1.9542, "step": 1475 }, { "epoch": 11.808, "grad_norm": 17.89442253112793, "learning_rate": 4.685907313025957e-05, "loss": 1.6497, "step": 1476 }, { "epoch": 11.816, "grad_norm": 28.667720794677734, "learning_rate": 4.685314685314686e-05, "loss": 2.1949, "step": 1477 }, { "epoch": 11.824, "grad_norm": 37.85551834106445, "learning_rate": 4.6847220576034137e-05, "loss": 2.0333, "step": 1478 }, { "epoch": 11.832, "grad_norm": 23.9453067779541, "learning_rate": 4.6841294298921415e-05, "loss": 1.4069, "step": 1479 }, { "epoch": 11.84, "grad_norm": 21.036468505859375, "learning_rate": 4.68353680218087e-05, "loss": 1.3523, "step": 1480 }, { "epoch": 11.848, "grad_norm": 29.25921058654785, "learning_rate": 4.6829441744695987e-05, "loss": 1.8997, "step": 1481 }, { "epoch": 11.856, "grad_norm": 22.12059211730957, "learning_rate": 4.6823515467583265e-05, "loss": 1.8746, "step": 1482 }, { "epoch": 11.864, "grad_norm": 28.1610050201416, "learning_rate": 4.681758919047055e-05, "loss": 1.5282, "step": 1483 }, { "epoch": 11.872, "grad_norm": 38.0184211730957, "learning_rate": 4.681166291335783e-05, "loss": 1.9151, "step": 1484 }, { "epoch": 11.88, "grad_norm": 23.34464454650879, "learning_rate": 4.6805736636245115e-05, "loss": 1.9122, "step": 1485 }, { "epoch": 11.888, "grad_norm": 21.029211044311523, "learning_rate": 4.6799810359132394e-05, "loss": 2.6057, "step": 1486 }, { "epoch": 11.896, "grad_norm": 38.87861633300781, "learning_rate": 4.679388408201968e-05, "loss": 2.0169, "step": 1487 }, { "epoch": 11.904, "grad_norm": 131.15354919433594, "learning_rate": 4.6787957804906965e-05, "loss": 1.8298, "step": 1488 }, { "epoch": 11.912, "grad_norm": 36.51360321044922, "learning_rate": 4.678203152779424e-05, "loss": 2.2261, "step": 1489 }, { "epoch": 11.92, "grad_norm": 37.596588134765625, "learning_rate": 4.677610525068152e-05, "loss": 2.1062, "step": 1490 }, { "epoch": 11.928, "grad_norm": 38.723575592041016, "learning_rate": 4.677017897356881e-05, "loss": 1.8254, "step": 1491 }, { "epoch": 11.936, "grad_norm": 40.53322219848633, "learning_rate": 4.676425269645609e-05, "loss": 1.9721, "step": 1492 }, { "epoch": 11.943999999999999, "grad_norm": 31.436349868774414, "learning_rate": 4.675832641934337e-05, "loss": 1.5998, "step": 1493 }, { "epoch": 11.952, "grad_norm": 57.93141555786133, "learning_rate": 4.675240014223065e-05, "loss": 1.7692, "step": 1494 }, { "epoch": 11.96, "grad_norm": 38.52320861816406, "learning_rate": 4.674647386511794e-05, "loss": 1.6411, "step": 1495 }, { "epoch": 11.968, "grad_norm": 32.987464904785156, "learning_rate": 4.6740547588005216e-05, "loss": 1.979, "step": 1496 }, { "epoch": 11.975999999999999, "grad_norm": 26.42680549621582, "learning_rate": 4.67346213108925e-05, "loss": 1.413, "step": 1497 }, { "epoch": 11.984, "grad_norm": 25.824148178100586, "learning_rate": 4.672869503377978e-05, "loss": 1.3274, "step": 1498 }, { "epoch": 11.992, "grad_norm": 44.7701416015625, "learning_rate": 4.672276875666706e-05, "loss": 1.8307, "step": 1499 }, { "epoch": 12.0, "grad_norm": 35.91097640991211, "learning_rate": 4.6716842479554345e-05, "loss": 1.6324, "step": 1500 }, { "epoch": 12.0, "eval_loss": 1.8351855278015137, "eval_map": 0.1345, "eval_map_50": 0.2925, "eval_map_75": 0.101, "eval_map_Coverall": 0.3234, "eval_map_Face_Shield": 0.0739, "eval_map_Gloves": 0.0764, "eval_map_Goggles": 0.001, "eval_map_Mask": 0.1976, "eval_map_large": 0.1525, "eval_map_medium": 0.0792, "eval_map_small": 0.1302, "eval_mar_1": 0.13, "eval_mar_10": 0.2842, "eval_mar_100": 0.302, "eval_mar_100_Coverall": 0.5622, "eval_mar_100_Face_Shield": 0.3235, "eval_mar_100_Gloves": 0.2689, "eval_mar_100_Goggles": 0.0094, "eval_mar_100_Mask": 0.3462, "eval_mar_large": 0.3047, "eval_mar_medium": 0.2008, "eval_mar_small": 0.1531, "eval_runtime": 1.8489, "eval_samples_per_second": 15.685, "eval_steps_per_second": 1.082, "step": 1500 }, { "epoch": 12.008, "grad_norm": 27.226354598999023, "learning_rate": 4.671091620244163e-05, "loss": 1.7405, "step": 1501 }, { "epoch": 12.016, "grad_norm": 22.308652877807617, "learning_rate": 4.670498992532891e-05, "loss": 1.5282, "step": 1502 }, { "epoch": 12.024, "grad_norm": 22.088481903076172, "learning_rate": 4.669906364821619e-05, "loss": 1.8797, "step": 1503 }, { "epoch": 12.032, "grad_norm": 31.237796783447266, "learning_rate": 4.6693137371103474e-05, "loss": 2.0314, "step": 1504 }, { "epoch": 12.04, "grad_norm": 53.123741149902344, "learning_rate": 4.668721109399076e-05, "loss": 1.3993, "step": 1505 }, { "epoch": 12.048, "grad_norm": 29.765031814575195, "learning_rate": 4.668128481687804e-05, "loss": 1.8174, "step": 1506 }, { "epoch": 12.056, "grad_norm": 28.092578887939453, "learning_rate": 4.6675358539765324e-05, "loss": 2.0501, "step": 1507 }, { "epoch": 12.064, "grad_norm": 43.4786376953125, "learning_rate": 4.66694322626526e-05, "loss": 1.7637, "step": 1508 }, { "epoch": 12.072, "grad_norm": 41.96384048461914, "learning_rate": 4.666350598553988e-05, "loss": 1.8898, "step": 1509 }, { "epoch": 12.08, "grad_norm": 25.729825973510742, "learning_rate": 4.665757970842717e-05, "loss": 1.4379, "step": 1510 }, { "epoch": 12.088, "grad_norm": 39.97755813598633, "learning_rate": 4.665165343131445e-05, "loss": 1.8138, "step": 1511 }, { "epoch": 12.096, "grad_norm": 114.41862487792969, "learning_rate": 4.664572715420173e-05, "loss": 1.9408, "step": 1512 }, { "epoch": 12.104, "grad_norm": 17.124452590942383, "learning_rate": 4.663980087708901e-05, "loss": 2.3119, "step": 1513 }, { "epoch": 12.112, "grad_norm": 49.711944580078125, "learning_rate": 4.6633874599976296e-05, "loss": 1.6257, "step": 1514 }, { "epoch": 12.12, "grad_norm": 60.98859405517578, "learning_rate": 4.662794832286358e-05, "loss": 1.9921, "step": 1515 }, { "epoch": 12.128, "grad_norm": 47.88844680786133, "learning_rate": 4.662202204575086e-05, "loss": 1.5589, "step": 1516 }, { "epoch": 12.136, "grad_norm": 16.988988876342773, "learning_rate": 4.6616095768638146e-05, "loss": 1.5149, "step": 1517 }, { "epoch": 12.144, "grad_norm": 19.36642074584961, "learning_rate": 4.6610169491525425e-05, "loss": 1.9769, "step": 1518 }, { "epoch": 12.152, "grad_norm": 21.059249877929688, "learning_rate": 4.660424321441271e-05, "loss": 1.2723, "step": 1519 }, { "epoch": 12.16, "grad_norm": 27.51677894592285, "learning_rate": 4.659831693729999e-05, "loss": 1.487, "step": 1520 }, { "epoch": 12.168, "grad_norm": 48.25457763671875, "learning_rate": 4.6592390660187275e-05, "loss": 1.7563, "step": 1521 }, { "epoch": 12.176, "grad_norm": 52.93272399902344, "learning_rate": 4.658646438307456e-05, "loss": 1.4133, "step": 1522 }, { "epoch": 12.184, "grad_norm": 29.427797317504883, "learning_rate": 4.658053810596183e-05, "loss": 2.4814, "step": 1523 }, { "epoch": 12.192, "grad_norm": 53.46786117553711, "learning_rate": 4.657461182884912e-05, "loss": 1.78, "step": 1524 }, { "epoch": 12.2, "grad_norm": 20.99195098876953, "learning_rate": 4.6568685551736404e-05, "loss": 1.9422, "step": 1525 }, { "epoch": 12.208, "grad_norm": 56.36357879638672, "learning_rate": 4.656275927462368e-05, "loss": 1.3087, "step": 1526 }, { "epoch": 12.216, "grad_norm": 221.8643341064453, "learning_rate": 4.655683299751097e-05, "loss": 1.5219, "step": 1527 }, { "epoch": 12.224, "grad_norm": 21.04922866821289, "learning_rate": 4.655090672039825e-05, "loss": 1.3992, "step": 1528 }, { "epoch": 12.232, "grad_norm": 25.74230194091797, "learning_rate": 4.654498044328553e-05, "loss": 1.5453, "step": 1529 }, { "epoch": 12.24, "grad_norm": 40.45294189453125, "learning_rate": 4.653905416617281e-05, "loss": 1.6216, "step": 1530 }, { "epoch": 12.248, "grad_norm": 23.838695526123047, "learning_rate": 4.65331278890601e-05, "loss": 1.7779, "step": 1531 }, { "epoch": 12.256, "grad_norm": 23.772830963134766, "learning_rate": 4.6527201611947376e-05, "loss": 2.0535, "step": 1532 }, { "epoch": 12.264, "grad_norm": 23.327606201171875, "learning_rate": 4.6521275334834655e-05, "loss": 1.7952, "step": 1533 }, { "epoch": 12.272, "grad_norm": 33.60472106933594, "learning_rate": 4.651534905772194e-05, "loss": 1.7209, "step": 1534 }, { "epoch": 12.28, "grad_norm": 18.994247436523438, "learning_rate": 4.6509422780609226e-05, "loss": 1.7432, "step": 1535 }, { "epoch": 12.288, "grad_norm": 36.11886978149414, "learning_rate": 4.6503496503496505e-05, "loss": 2.4103, "step": 1536 }, { "epoch": 12.296, "grad_norm": 25.99385643005371, "learning_rate": 4.649757022638378e-05, "loss": 1.6801, "step": 1537 }, { "epoch": 12.304, "grad_norm": 31.752063751220703, "learning_rate": 4.649164394927107e-05, "loss": 1.5099, "step": 1538 }, { "epoch": 12.312, "grad_norm": 28.587608337402344, "learning_rate": 4.6485717672158355e-05, "loss": 1.8694, "step": 1539 }, { "epoch": 12.32, "grad_norm": 18.627553939819336, "learning_rate": 4.647979139504563e-05, "loss": 1.9356, "step": 1540 }, { "epoch": 12.328, "grad_norm": 20.536746978759766, "learning_rate": 4.647386511793292e-05, "loss": 1.6116, "step": 1541 }, { "epoch": 12.336, "grad_norm": 25.11901092529297, "learning_rate": 4.64679388408202e-05, "loss": 1.6984, "step": 1542 }, { "epoch": 12.344, "grad_norm": 41.711116790771484, "learning_rate": 4.6462012563707477e-05, "loss": 2.145, "step": 1543 }, { "epoch": 12.352, "grad_norm": 20.55744171142578, "learning_rate": 4.645608628659476e-05, "loss": 1.7583, "step": 1544 }, { "epoch": 12.36, "grad_norm": 24.64180564880371, "learning_rate": 4.645016000948205e-05, "loss": 1.3157, "step": 1545 }, { "epoch": 12.368, "grad_norm": 42.80349349975586, "learning_rate": 4.6444233732369327e-05, "loss": 2.1377, "step": 1546 }, { "epoch": 12.376, "grad_norm": 26.619632720947266, "learning_rate": 4.6438307455256605e-05, "loss": 1.9968, "step": 1547 }, { "epoch": 12.384, "grad_norm": 43.8679084777832, "learning_rate": 4.643238117814389e-05, "loss": 1.8773, "step": 1548 }, { "epoch": 12.392, "grad_norm": 22.934555053710938, "learning_rate": 4.642645490103118e-05, "loss": 1.6659, "step": 1549 }, { "epoch": 12.4, "grad_norm": 40.098052978515625, "learning_rate": 4.6420528623918455e-05, "loss": 1.3134, "step": 1550 }, { "epoch": 12.408, "grad_norm": 29.6146183013916, "learning_rate": 4.641460234680574e-05, "loss": 1.5411, "step": 1551 }, { "epoch": 12.416, "grad_norm": 45.58068084716797, "learning_rate": 4.640867606969302e-05, "loss": 1.645, "step": 1552 }, { "epoch": 12.424, "grad_norm": 75.42662811279297, "learning_rate": 4.6402749792580305e-05, "loss": 1.5409, "step": 1553 }, { "epoch": 12.432, "grad_norm": 21.197500228881836, "learning_rate": 4.6396823515467584e-05, "loss": 1.9641, "step": 1554 }, { "epoch": 12.44, "grad_norm": 21.898296356201172, "learning_rate": 4.639089723835487e-05, "loss": 1.4883, "step": 1555 }, { "epoch": 12.448, "grad_norm": 28.35267448425293, "learning_rate": 4.6384970961242155e-05, "loss": 1.7674, "step": 1556 }, { "epoch": 12.456, "grad_norm": 26.689006805419922, "learning_rate": 4.637904468412943e-05, "loss": 1.2792, "step": 1557 }, { "epoch": 12.464, "grad_norm": 39.56782150268555, "learning_rate": 4.637311840701671e-05, "loss": 1.636, "step": 1558 }, { "epoch": 12.472, "grad_norm": 44.36517333984375, "learning_rate": 4.6367192129904e-05, "loss": 1.9893, "step": 1559 }, { "epoch": 12.48, "grad_norm": 25.787261962890625, "learning_rate": 4.636126585279128e-05, "loss": 1.5865, "step": 1560 }, { "epoch": 12.488, "grad_norm": 34.16432571411133, "learning_rate": 4.635533957567856e-05, "loss": 1.357, "step": 1561 }, { "epoch": 12.496, "grad_norm": 31.07647132873535, "learning_rate": 4.634941329856584e-05, "loss": 1.7831, "step": 1562 }, { "epoch": 12.504, "grad_norm": 48.077056884765625, "learning_rate": 4.634348702145313e-05, "loss": 1.8384, "step": 1563 }, { "epoch": 12.512, "grad_norm": 42.24587631225586, "learning_rate": 4.6337560744340406e-05, "loss": 2.7287, "step": 1564 }, { "epoch": 12.52, "grad_norm": 28.521635055541992, "learning_rate": 4.633163446722769e-05, "loss": 2.0976, "step": 1565 }, { "epoch": 12.528, "grad_norm": 20.916156768798828, "learning_rate": 4.632570819011498e-05, "loss": 1.7709, "step": 1566 }, { "epoch": 12.536, "grad_norm": 32.487850189208984, "learning_rate": 4.631978191300225e-05, "loss": 1.4486, "step": 1567 }, { "epoch": 12.544, "grad_norm": 19.1827449798584, "learning_rate": 4.6313855635889535e-05, "loss": 1.1304, "step": 1568 }, { "epoch": 12.552, "grad_norm": 37.56589889526367, "learning_rate": 4.630792935877682e-05, "loss": 1.614, "step": 1569 }, { "epoch": 12.56, "grad_norm": 42.905113220214844, "learning_rate": 4.63020030816641e-05, "loss": 1.5272, "step": 1570 }, { "epoch": 12.568, "grad_norm": 31.424776077270508, "learning_rate": 4.629607680455138e-05, "loss": 1.6834, "step": 1571 }, { "epoch": 12.576, "grad_norm": 21.57159423828125, "learning_rate": 4.6290150527438664e-05, "loss": 1.4642, "step": 1572 }, { "epoch": 12.584, "grad_norm": 27.462669372558594, "learning_rate": 4.628422425032595e-05, "loss": 1.6717, "step": 1573 }, { "epoch": 12.592, "grad_norm": 26.82038116455078, "learning_rate": 4.627829797321323e-05, "loss": 1.5747, "step": 1574 }, { "epoch": 12.6, "grad_norm": 24.364013671875, "learning_rate": 4.6272371696100514e-05, "loss": 1.5027, "step": 1575 }, { "epoch": 12.608, "grad_norm": 34.629390716552734, "learning_rate": 4.626644541898779e-05, "loss": 1.4359, "step": 1576 }, { "epoch": 12.616, "grad_norm": 42.36894989013672, "learning_rate": 4.626051914187507e-05, "loss": 1.7901, "step": 1577 }, { "epoch": 12.624, "grad_norm": 50.863094329833984, "learning_rate": 4.625459286476236e-05, "loss": 1.1754, "step": 1578 }, { "epoch": 12.632, "grad_norm": 59.89326095581055, "learning_rate": 4.624866658764964e-05, "loss": 1.6208, "step": 1579 }, { "epoch": 12.64, "grad_norm": 52.906829833984375, "learning_rate": 4.624274031053692e-05, "loss": 1.4945, "step": 1580 }, { "epoch": 12.648, "grad_norm": 40.15111541748047, "learning_rate": 4.62368140334242e-05, "loss": 1.6712, "step": 1581 }, { "epoch": 12.656, "grad_norm": 34.3464241027832, "learning_rate": 4.6230887756311486e-05, "loss": 1.7316, "step": 1582 }, { "epoch": 12.664, "grad_norm": 61.878116607666016, "learning_rate": 4.622496147919877e-05, "loss": 1.9712, "step": 1583 }, { "epoch": 12.672, "grad_norm": 29.316810607910156, "learning_rate": 4.621903520208605e-05, "loss": 2.3421, "step": 1584 }, { "epoch": 12.68, "grad_norm": 149.0146942138672, "learning_rate": 4.6213108924973336e-05, "loss": 1.7835, "step": 1585 }, { "epoch": 12.688, "grad_norm": 56.08359909057617, "learning_rate": 4.6207182647860615e-05, "loss": 3.5572, "step": 1586 }, { "epoch": 12.696, "grad_norm": 25.63938331604004, "learning_rate": 4.62012563707479e-05, "loss": 1.556, "step": 1587 }, { "epoch": 12.704, "grad_norm": 2389.701416015625, "learning_rate": 4.619533009363518e-05, "loss": 2.0894, "step": 1588 }, { "epoch": 12.712, "grad_norm": 36.0816650390625, "learning_rate": 4.6189403816522465e-05, "loss": 1.7004, "step": 1589 }, { "epoch": 12.72, "grad_norm": 35.39521408081055, "learning_rate": 4.618347753940975e-05, "loss": 1.4824, "step": 1590 }, { "epoch": 12.728, "grad_norm": 38.28635025024414, "learning_rate": 4.617755126229702e-05, "loss": 2.5856, "step": 1591 }, { "epoch": 12.736, "grad_norm": 29.782522201538086, "learning_rate": 4.617162498518431e-05, "loss": 1.6143, "step": 1592 }, { "epoch": 12.744, "grad_norm": 42.31830978393555, "learning_rate": 4.6165698708071594e-05, "loss": 1.9194, "step": 1593 }, { "epoch": 12.752, "grad_norm": 34.30498504638672, "learning_rate": 4.615977243095887e-05, "loss": 1.6522, "step": 1594 }, { "epoch": 12.76, "grad_norm": 113.5435562133789, "learning_rate": 4.615384615384616e-05, "loss": 1.5141, "step": 1595 }, { "epoch": 12.768, "grad_norm": 19.543720245361328, "learning_rate": 4.614791987673344e-05, "loss": 1.9955, "step": 1596 }, { "epoch": 12.776, "grad_norm": 34.13007736206055, "learning_rate": 4.614199359962072e-05, "loss": 1.3976, "step": 1597 }, { "epoch": 12.784, "grad_norm": 37.899593353271484, "learning_rate": 4.6136067322508e-05, "loss": 1.5514, "step": 1598 }, { "epoch": 12.792, "grad_norm": 16.887165069580078, "learning_rate": 4.613014104539529e-05, "loss": 1.8919, "step": 1599 }, { "epoch": 12.8, "grad_norm": 36.99783706665039, "learning_rate": 4.612421476828257e-05, "loss": 1.4164, "step": 1600 }, { "epoch": 12.808, "grad_norm": 38.413124084472656, "learning_rate": 4.6118288491169845e-05, "loss": 1.8875, "step": 1601 }, { "epoch": 12.816, "grad_norm": 28.287303924560547, "learning_rate": 4.611236221405713e-05, "loss": 1.626, "step": 1602 }, { "epoch": 12.824, "grad_norm": 27.68961524963379, "learning_rate": 4.6106435936944416e-05, "loss": 1.5484, "step": 1603 }, { "epoch": 12.832, "grad_norm": 25.803693771362305, "learning_rate": 4.6100509659831695e-05, "loss": 1.7651, "step": 1604 }, { "epoch": 12.84, "grad_norm": 33.92333984375, "learning_rate": 4.6094583382718973e-05, "loss": 1.1862, "step": 1605 }, { "epoch": 12.848, "grad_norm": 28.005685806274414, "learning_rate": 4.608865710560626e-05, "loss": 1.4669, "step": 1606 }, { "epoch": 12.856, "grad_norm": 27.96135139465332, "learning_rate": 4.6082730828493545e-05, "loss": 1.665, "step": 1607 }, { "epoch": 12.864, "grad_norm": 27.39142417907715, "learning_rate": 4.6076804551380823e-05, "loss": 1.8724, "step": 1608 }, { "epoch": 12.872, "grad_norm": 26.95987892150879, "learning_rate": 4.607087827426811e-05, "loss": 1.6783, "step": 1609 }, { "epoch": 12.88, "grad_norm": 39.29452133178711, "learning_rate": 4.606495199715539e-05, "loss": 1.4923, "step": 1610 }, { "epoch": 12.888, "grad_norm": 21.020244598388672, "learning_rate": 4.605902572004267e-05, "loss": 1.6379, "step": 1611 }, { "epoch": 12.896, "grad_norm": 24.242521286010742, "learning_rate": 4.605309944292995e-05, "loss": 1.5247, "step": 1612 }, { "epoch": 12.904, "grad_norm": 42.184261322021484, "learning_rate": 4.604717316581724e-05, "loss": 1.3923, "step": 1613 }, { "epoch": 12.912, "grad_norm": 68.35374450683594, "learning_rate": 4.6041246888704523e-05, "loss": 1.2639, "step": 1614 }, { "epoch": 12.92, "grad_norm": 19.267900466918945, "learning_rate": 4.6035320611591796e-05, "loss": 1.6834, "step": 1615 }, { "epoch": 12.928, "grad_norm": 36.38999938964844, "learning_rate": 4.602939433447908e-05, "loss": 2.5739, "step": 1616 }, { "epoch": 12.936, "grad_norm": 39.58866882324219, "learning_rate": 4.602346805736637e-05, "loss": 1.3716, "step": 1617 }, { "epoch": 12.943999999999999, "grad_norm": 40.763099670410156, "learning_rate": 4.6017541780253646e-05, "loss": 1.6704, "step": 1618 }, { "epoch": 12.952, "grad_norm": 35.84151077270508, "learning_rate": 4.601161550314093e-05, "loss": 1.3837, "step": 1619 }, { "epoch": 12.96, "grad_norm": 44.99966812133789, "learning_rate": 4.600568922602821e-05, "loss": 1.436, "step": 1620 }, { "epoch": 12.968, "grad_norm": 43.48197555541992, "learning_rate": 4.5999762948915496e-05, "loss": 1.3772, "step": 1621 }, { "epoch": 12.975999999999999, "grad_norm": 21.170177459716797, "learning_rate": 4.5993836671802774e-05, "loss": 1.115, "step": 1622 }, { "epoch": 12.984, "grad_norm": 31.49482536315918, "learning_rate": 4.598791039469006e-05, "loss": 1.446, "step": 1623 }, { "epoch": 12.992, "grad_norm": 27.436769485473633, "learning_rate": 4.5981984117577346e-05, "loss": 1.5224, "step": 1624 }, { "epoch": 13.0, "grad_norm": 77.18489074707031, "learning_rate": 4.597605784046462e-05, "loss": 1.6761, "step": 1625 }, { "epoch": 13.0, "eval_loss": 1.6934905052185059, "eval_map": 0.1647, "eval_map_50": 0.3792, "eval_map_75": 0.1064, "eval_map_Coverall": 0.3796, "eval_map_Face_Shield": 0.1365, "eval_map_Gloves": 0.0727, "eval_map_Goggles": 0.0154, "eval_map_Mask": 0.2191, "eval_map_large": 0.2236, "eval_map_medium": 0.0889, "eval_map_small": 0.1281, "eval_mar_1": 0.1665, "eval_mar_10": 0.34, "eval_mar_100": 0.3667, "eval_mar_100_Coverall": 0.5756, "eval_mar_100_Face_Shield": 0.4882, "eval_mar_100_Gloves": 0.3066, "eval_mar_100_Goggles": 0.1094, "eval_mar_100_Mask": 0.3538, "eval_mar_large": 0.5021, "eval_mar_medium": 0.2716, "eval_mar_small": 0.1536, "eval_runtime": 1.8498, "eval_samples_per_second": 15.678, "eval_steps_per_second": 1.081, "step": 1625 }, { "epoch": 13.008, "grad_norm": 21.717180252075195, "learning_rate": 4.59701315633519e-05, "loss": 1.4145, "step": 1626 }, { "epoch": 13.016, "grad_norm": 27.684036254882812, "learning_rate": 4.596420528623919e-05, "loss": 1.4619, "step": 1627 }, { "epoch": 13.024, "grad_norm": 67.22924041748047, "learning_rate": 4.595827900912647e-05, "loss": 1.3231, "step": 1628 }, { "epoch": 13.032, "grad_norm": 24.95175552368164, "learning_rate": 4.595235273201375e-05, "loss": 1.8031, "step": 1629 }, { "epoch": 13.04, "grad_norm": 51.49076461791992, "learning_rate": 4.594642645490103e-05, "loss": 1.4782, "step": 1630 }, { "epoch": 13.048, "grad_norm": 17.89876937866211, "learning_rate": 4.594050017778832e-05, "loss": 1.2797, "step": 1631 }, { "epoch": 13.056, "grad_norm": 33.05839157104492, "learning_rate": 4.5934573900675596e-05, "loss": 1.5725, "step": 1632 }, { "epoch": 13.064, "grad_norm": 22.29979133605957, "learning_rate": 4.592864762356288e-05, "loss": 1.4143, "step": 1633 }, { "epoch": 13.072, "grad_norm": 28.229990005493164, "learning_rate": 4.592272134645017e-05, "loss": 2.0171, "step": 1634 }, { "epoch": 13.08, "grad_norm": 78.04182434082031, "learning_rate": 4.591679506933744e-05, "loss": 1.7292, "step": 1635 }, { "epoch": 13.088, "grad_norm": 34.658424377441406, "learning_rate": 4.5910868792224725e-05, "loss": 1.9409, "step": 1636 }, { "epoch": 13.096, "grad_norm": 16.986936569213867, "learning_rate": 4.590494251511201e-05, "loss": 1.5758, "step": 1637 }, { "epoch": 13.104, "grad_norm": 51.620384216308594, "learning_rate": 4.589901623799929e-05, "loss": 1.5403, "step": 1638 }, { "epoch": 13.112, "grad_norm": 25.237089157104492, "learning_rate": 4.589308996088657e-05, "loss": 1.4927, "step": 1639 }, { "epoch": 13.12, "grad_norm": 30.451337814331055, "learning_rate": 4.5887163683773854e-05, "loss": 1.7916, "step": 1640 }, { "epoch": 13.128, "grad_norm": 35.25887680053711, "learning_rate": 4.588123740666114e-05, "loss": 1.5573, "step": 1641 }, { "epoch": 13.136, "grad_norm": 31.27984046936035, "learning_rate": 4.587531112954842e-05, "loss": 1.7147, "step": 1642 }, { "epoch": 13.144, "grad_norm": 20.055341720581055, "learning_rate": 4.5869384852435704e-05, "loss": 1.42, "step": 1643 }, { "epoch": 13.152, "grad_norm": 28.80607032775879, "learning_rate": 4.586345857532298e-05, "loss": 1.4656, "step": 1644 }, { "epoch": 13.16, "grad_norm": 19.337263107299805, "learning_rate": 4.585753229821026e-05, "loss": 1.8027, "step": 1645 }, { "epoch": 13.168, "grad_norm": 42.11556625366211, "learning_rate": 4.585160602109755e-05, "loss": 1.7431, "step": 1646 }, { "epoch": 13.176, "grad_norm": 22.110952377319336, "learning_rate": 4.584567974398483e-05, "loss": 1.5846, "step": 1647 }, { "epoch": 13.184, "grad_norm": 35.32429122924805, "learning_rate": 4.583975346687212e-05, "loss": 1.5555, "step": 1648 }, { "epoch": 13.192, "grad_norm": 20.512771606445312, "learning_rate": 4.583382718975939e-05, "loss": 1.3919, "step": 1649 }, { "epoch": 13.2, "grad_norm": 62.244102478027344, "learning_rate": 4.5827900912646676e-05, "loss": 1.8189, "step": 1650 }, { "epoch": 13.208, "grad_norm": 99.39779663085938, "learning_rate": 4.582197463553396e-05, "loss": 2.2736, "step": 1651 }, { "epoch": 13.216, "grad_norm": 28.002195358276367, "learning_rate": 4.581604835842124e-05, "loss": 1.3844, "step": 1652 }, { "epoch": 13.224, "grad_norm": 33.76576232910156, "learning_rate": 4.5810122081308526e-05, "loss": 1.3619, "step": 1653 }, { "epoch": 13.232, "grad_norm": 27.770601272583008, "learning_rate": 4.5804195804195805e-05, "loss": 1.5254, "step": 1654 }, { "epoch": 13.24, "grad_norm": 34.6903076171875, "learning_rate": 4.579826952708309e-05, "loss": 1.2064, "step": 1655 }, { "epoch": 13.248, "grad_norm": 19.709407806396484, "learning_rate": 4.579234324997037e-05, "loss": 2.2985, "step": 1656 }, { "epoch": 13.256, "grad_norm": 32.936363220214844, "learning_rate": 4.5786416972857655e-05, "loss": 1.241, "step": 1657 }, { "epoch": 13.264, "grad_norm": 22.20589828491211, "learning_rate": 4.578049069574494e-05, "loss": 1.3074, "step": 1658 }, { "epoch": 13.272, "grad_norm": 21.081743240356445, "learning_rate": 4.577456441863221e-05, "loss": 1.3775, "step": 1659 }, { "epoch": 13.28, "grad_norm": 20.34925651550293, "learning_rate": 4.57686381415195e-05, "loss": 1.4992, "step": 1660 }, { "epoch": 13.288, "grad_norm": 67.80567169189453, "learning_rate": 4.5762711864406784e-05, "loss": 2.5062, "step": 1661 }, { "epoch": 13.296, "grad_norm": 41.40937805175781, "learning_rate": 4.575678558729406e-05, "loss": 1.8361, "step": 1662 }, { "epoch": 13.304, "grad_norm": 29.1641788482666, "learning_rate": 4.575085931018135e-05, "loss": 1.4503, "step": 1663 }, { "epoch": 13.312, "grad_norm": 22.44009780883789, "learning_rate": 4.574493303306863e-05, "loss": 2.0063, "step": 1664 }, { "epoch": 13.32, "grad_norm": 29.98116111755371, "learning_rate": 4.573900675595591e-05, "loss": 1.4075, "step": 1665 }, { "epoch": 13.328, "grad_norm": 48.52291488647461, "learning_rate": 4.573308047884319e-05, "loss": 1.1845, "step": 1666 }, { "epoch": 13.336, "grad_norm": 31.435501098632812, "learning_rate": 4.572715420173048e-05, "loss": 2.04, "step": 1667 }, { "epoch": 13.344, "grad_norm": 28.703960418701172, "learning_rate": 4.572122792461776e-05, "loss": 1.5261, "step": 1668 }, { "epoch": 13.352, "grad_norm": 37.5667839050293, "learning_rate": 4.5715301647505035e-05, "loss": 1.4684, "step": 1669 }, { "epoch": 13.36, "grad_norm": 32.147518157958984, "learning_rate": 4.570937537039232e-05, "loss": 1.4098, "step": 1670 }, { "epoch": 13.368, "grad_norm": 33.321903228759766, "learning_rate": 4.5703449093279606e-05, "loss": 2.3605, "step": 1671 }, { "epoch": 13.376, "grad_norm": 32.043434143066406, "learning_rate": 4.5697522816166885e-05, "loss": 1.2122, "step": 1672 }, { "epoch": 13.384, "grad_norm": 143.82447814941406, "learning_rate": 4.5691596539054164e-05, "loss": 1.9902, "step": 1673 }, { "epoch": 13.392, "grad_norm": 62.8162956237793, "learning_rate": 4.568567026194145e-05, "loss": 1.7079, "step": 1674 }, { "epoch": 13.4, "grad_norm": 34.13608169555664, "learning_rate": 4.5679743984828735e-05, "loss": 1.3795, "step": 1675 }, { "epoch": 13.408, "grad_norm": 41.328399658203125, "learning_rate": 4.5673817707716014e-05, "loss": 1.8782, "step": 1676 }, { "epoch": 13.416, "grad_norm": 41.227561950683594, "learning_rate": 4.56678914306033e-05, "loss": 1.3705, "step": 1677 }, { "epoch": 13.424, "grad_norm": 28.75196647644043, "learning_rate": 4.566196515349058e-05, "loss": 2.9955, "step": 1678 }, { "epoch": 13.432, "grad_norm": 80.16110229492188, "learning_rate": 4.5656038876377864e-05, "loss": 1.6998, "step": 1679 }, { "epoch": 13.44, "grad_norm": 52.45417404174805, "learning_rate": 4.565011259926514e-05, "loss": 1.226, "step": 1680 }, { "epoch": 13.448, "grad_norm": 57.93387985229492, "learning_rate": 4.564418632215243e-05, "loss": 1.2804, "step": 1681 }, { "epoch": 13.456, "grad_norm": 92.38145446777344, "learning_rate": 4.5638260045039714e-05, "loss": 2.3126, "step": 1682 }, { "epoch": 13.464, "grad_norm": 48.41566467285156, "learning_rate": 4.5632333767926986e-05, "loss": 1.6035, "step": 1683 }, { "epoch": 13.472, "grad_norm": 33.47966766357422, "learning_rate": 4.562640749081427e-05, "loss": 1.3263, "step": 1684 }, { "epoch": 13.48, "grad_norm": 23.240663528442383, "learning_rate": 4.562048121370156e-05, "loss": 2.2835, "step": 1685 }, { "epoch": 13.488, "grad_norm": 22.970317840576172, "learning_rate": 4.5614554936588836e-05, "loss": 1.4674, "step": 1686 }, { "epoch": 13.496, "grad_norm": 38.7963752746582, "learning_rate": 4.560862865947612e-05, "loss": 1.7091, "step": 1687 }, { "epoch": 13.504, "grad_norm": 64.58072662353516, "learning_rate": 4.56027023823634e-05, "loss": 1.5092, "step": 1688 }, { "epoch": 13.512, "grad_norm": 142.52166748046875, "learning_rate": 4.5596776105250686e-05, "loss": 1.4378, "step": 1689 }, { "epoch": 13.52, "grad_norm": 30.387540817260742, "learning_rate": 4.5590849828137964e-05, "loss": 2.2145, "step": 1690 }, { "epoch": 13.528, "grad_norm": 41.582984924316406, "learning_rate": 4.558492355102525e-05, "loss": 1.8526, "step": 1691 }, { "epoch": 13.536, "grad_norm": 35.07518005371094, "learning_rate": 4.5578997273912536e-05, "loss": 1.9814, "step": 1692 }, { "epoch": 13.544, "grad_norm": 69.28009796142578, "learning_rate": 4.557307099679981e-05, "loss": 1.9899, "step": 1693 }, { "epoch": 13.552, "grad_norm": 42.565853118896484, "learning_rate": 4.556714471968709e-05, "loss": 2.195, "step": 1694 }, { "epoch": 13.56, "grad_norm": 41.551239013671875, "learning_rate": 4.556121844257438e-05, "loss": 2.0363, "step": 1695 }, { "epoch": 13.568, "grad_norm": 33.15504455566406, "learning_rate": 4.555529216546166e-05, "loss": 3.2234, "step": 1696 }, { "epoch": 13.576, "grad_norm": 16.938539505004883, "learning_rate": 4.554936588834894e-05, "loss": 1.3476, "step": 1697 }, { "epoch": 13.584, "grad_norm": 21.588464736938477, "learning_rate": 4.554343961123622e-05, "loss": 1.6733, "step": 1698 }, { "epoch": 13.592, "grad_norm": 33.104305267333984, "learning_rate": 4.553751333412351e-05, "loss": 2.0388, "step": 1699 }, { "epoch": 13.6, "grad_norm": 61.71878433227539, "learning_rate": 4.5531587057010786e-05, "loss": 1.7039, "step": 1700 }, { "epoch": 13.608, "grad_norm": 34.438846588134766, "learning_rate": 4.552566077989807e-05, "loss": 2.4569, "step": 1701 }, { "epoch": 13.616, "grad_norm": 25.62622833251953, "learning_rate": 4.551973450278536e-05, "loss": 1.9292, "step": 1702 }, { "epoch": 13.624, "grad_norm": 25.998008728027344, "learning_rate": 4.551380822567263e-05, "loss": 2.0446, "step": 1703 }, { "epoch": 13.632, "grad_norm": 43.4490966796875, "learning_rate": 4.5507881948559915e-05, "loss": 1.8027, "step": 1704 }, { "epoch": 13.64, "grad_norm": 24.753604888916016, "learning_rate": 4.55019556714472e-05, "loss": 1.685, "step": 1705 }, { "epoch": 13.648, "grad_norm": 49.496646881103516, "learning_rate": 4.549602939433448e-05, "loss": 1.199, "step": 1706 }, { "epoch": 13.656, "grad_norm": 28.33000373840332, "learning_rate": 4.5490103117221765e-05, "loss": 1.9553, "step": 1707 }, { "epoch": 13.664, "grad_norm": 39.154178619384766, "learning_rate": 4.5484176840109044e-05, "loss": 1.1098, "step": 1708 }, { "epoch": 13.672, "grad_norm": 62.5846061706543, "learning_rate": 4.547825056299633e-05, "loss": 1.4397, "step": 1709 }, { "epoch": 13.68, "grad_norm": 39.2205696105957, "learning_rate": 4.547232428588361e-05, "loss": 1.4322, "step": 1710 }, { "epoch": 13.688, "grad_norm": 49.86079406738281, "learning_rate": 4.5466398008770894e-05, "loss": 2.0078, "step": 1711 }, { "epoch": 13.696, "grad_norm": 30.90574836730957, "learning_rate": 4.546047173165817e-05, "loss": 1.2994, "step": 1712 }, { "epoch": 13.704, "grad_norm": 489.9351806640625, "learning_rate": 4.545454545454546e-05, "loss": 1.4763, "step": 1713 }, { "epoch": 13.712, "grad_norm": 416.6802673339844, "learning_rate": 4.544861917743274e-05, "loss": 1.8598, "step": 1714 }, { "epoch": 13.72, "grad_norm": 49.23097229003906, "learning_rate": 4.544269290032002e-05, "loss": 1.7111, "step": 1715 }, { "epoch": 13.728, "grad_norm": 36.54994583129883, "learning_rate": 4.543676662320731e-05, "loss": 1.8305, "step": 1716 }, { "epoch": 13.736, "grad_norm": 35.975555419921875, "learning_rate": 4.543084034609458e-05, "loss": 1.5993, "step": 1717 }, { "epoch": 13.744, "grad_norm": 65.60619354248047, "learning_rate": 4.5424914068981866e-05, "loss": 1.4389, "step": 1718 }, { "epoch": 13.752, "grad_norm": 29.293804168701172, "learning_rate": 4.541898779186915e-05, "loss": 1.5059, "step": 1719 }, { "epoch": 13.76, "grad_norm": 14.996262550354004, "learning_rate": 4.541306151475643e-05, "loss": 1.1933, "step": 1720 }, { "epoch": 13.768, "grad_norm": 24.301475524902344, "learning_rate": 4.5407135237643716e-05, "loss": 1.7539, "step": 1721 }, { "epoch": 13.776, "grad_norm": 32.79924774169922, "learning_rate": 4.5401208960530995e-05, "loss": 1.7333, "step": 1722 }, { "epoch": 13.784, "grad_norm": 25.376365661621094, "learning_rate": 4.539528268341828e-05, "loss": 1.4001, "step": 1723 }, { "epoch": 13.792, "grad_norm": 39.79709243774414, "learning_rate": 4.538935640630556e-05, "loss": 1.4889, "step": 1724 }, { "epoch": 13.8, "grad_norm": 18.248992919921875, "learning_rate": 4.5383430129192845e-05, "loss": 1.7579, "step": 1725 }, { "epoch": 13.808, "grad_norm": 47.83329391479492, "learning_rate": 4.537750385208013e-05, "loss": 1.5726, "step": 1726 }, { "epoch": 13.816, "grad_norm": 29.98651885986328, "learning_rate": 4.53715775749674e-05, "loss": 1.3734, "step": 1727 }, { "epoch": 13.824, "grad_norm": 29.370431900024414, "learning_rate": 4.536565129785469e-05, "loss": 1.6449, "step": 1728 }, { "epoch": 13.832, "grad_norm": 35.067020416259766, "learning_rate": 4.5359725020741974e-05, "loss": 1.4993, "step": 1729 }, { "epoch": 13.84, "grad_norm": 19.774335861206055, "learning_rate": 4.535379874362925e-05, "loss": 1.9158, "step": 1730 }, { "epoch": 13.848, "grad_norm": 47.205413818359375, "learning_rate": 4.534787246651654e-05, "loss": 1.8135, "step": 1731 }, { "epoch": 13.856, "grad_norm": 123.65296936035156, "learning_rate": 4.534194618940382e-05, "loss": 1.6103, "step": 1732 }, { "epoch": 13.864, "grad_norm": 157.9384002685547, "learning_rate": 4.53360199122911e-05, "loss": 1.6457, "step": 1733 }, { "epoch": 13.872, "grad_norm": 61.38283157348633, "learning_rate": 4.533009363517838e-05, "loss": 1.6389, "step": 1734 }, { "epoch": 13.88, "grad_norm": 30.98291778564453, "learning_rate": 4.532416735806567e-05, "loss": 1.5794, "step": 1735 }, { "epoch": 13.888, "grad_norm": 17.811931610107422, "learning_rate": 4.531824108095295e-05, "loss": 1.4655, "step": 1736 }, { "epoch": 13.896, "grad_norm": 60.750770568847656, "learning_rate": 4.5312314803840225e-05, "loss": 2.3324, "step": 1737 }, { "epoch": 13.904, "grad_norm": 31.155488967895508, "learning_rate": 4.530638852672751e-05, "loss": 1.448, "step": 1738 }, { "epoch": 13.912, "grad_norm": 43.5023307800293, "learning_rate": 4.5300462249614796e-05, "loss": 1.5315, "step": 1739 }, { "epoch": 13.92, "grad_norm": 30.275903701782227, "learning_rate": 4.5294535972502075e-05, "loss": 1.1739, "step": 1740 }, { "epoch": 13.928, "grad_norm": 58.56461715698242, "learning_rate": 4.528860969538936e-05, "loss": 1.8834, "step": 1741 }, { "epoch": 13.936, "grad_norm": 182.21484375, "learning_rate": 4.528268341827664e-05, "loss": 1.203, "step": 1742 }, { "epoch": 13.943999999999999, "grad_norm": 35.337791442871094, "learning_rate": 4.5276757141163925e-05, "loss": 1.8102, "step": 1743 }, { "epoch": 13.952, "grad_norm": 39.74998474121094, "learning_rate": 4.5270830864051204e-05, "loss": 2.4202, "step": 1744 }, { "epoch": 13.96, "grad_norm": 198.41807556152344, "learning_rate": 4.526490458693849e-05, "loss": 1.9665, "step": 1745 }, { "epoch": 13.968, "grad_norm": 51.5546875, "learning_rate": 4.525897830982577e-05, "loss": 1.6168, "step": 1746 }, { "epoch": 13.975999999999999, "grad_norm": 50.02120590209961, "learning_rate": 4.5253052032713054e-05, "loss": 1.5653, "step": 1747 }, { "epoch": 13.984, "grad_norm": 337.9266662597656, "learning_rate": 4.524712575560033e-05, "loss": 1.9993, "step": 1748 }, { "epoch": 13.992, "grad_norm": 31.241771697998047, "learning_rate": 4.524119947848762e-05, "loss": 1.5021, "step": 1749 }, { "epoch": 14.0, "grad_norm": 36.26698303222656, "learning_rate": 4.5235273201374904e-05, "loss": 1.8273, "step": 1750 }, { "epoch": 14.0, "eval_loss": 1.8490545749664307, "eval_map": 0.1261, "eval_map_50": 0.298, "eval_map_75": 0.0814, "eval_map_Coverall": 0.2845, "eval_map_Face_Shield": 0.0992, "eval_map_Gloves": 0.0718, "eval_map_Goggles": 0.0133, "eval_map_Mask": 0.1615, "eval_map_large": 0.1792, "eval_map_medium": 0.0786, "eval_map_small": 0.0667, "eval_mar_1": 0.1548, "eval_mar_10": 0.2715, "eval_mar_100": 0.2976, "eval_mar_100_Coverall": 0.6356, "eval_mar_100_Face_Shield": 0.3412, "eval_mar_100_Gloves": 0.2115, "eval_mar_100_Goggles": 0.05, "eval_mar_100_Mask": 0.25, "eval_mar_large": 0.3997, "eval_mar_medium": 0.168, "eval_mar_small": 0.0749, "eval_runtime": 1.8471, "eval_samples_per_second": 15.701, "eval_steps_per_second": 1.083, "step": 1750 }, { "epoch": 14.008, "grad_norm": 53.042083740234375, "learning_rate": 4.5229346924262176e-05, "loss": 1.6438, "step": 1751 }, { "epoch": 14.016, "grad_norm": 29.740066528320312, "learning_rate": 4.522342064714946e-05, "loss": 1.7166, "step": 1752 }, { "epoch": 14.024, "grad_norm": 33.055355072021484, "learning_rate": 4.521749437003675e-05, "loss": 1.4412, "step": 1753 }, { "epoch": 14.032, "grad_norm": 68.55250549316406, "learning_rate": 4.5211568092924026e-05, "loss": 1.9784, "step": 1754 }, { "epoch": 14.04, "grad_norm": 28.211326599121094, "learning_rate": 4.520564181581131e-05, "loss": 1.3017, "step": 1755 }, { "epoch": 14.048, "grad_norm": 50.17369079589844, "learning_rate": 4.519971553869859e-05, "loss": 1.922, "step": 1756 }, { "epoch": 14.056, "grad_norm": 49.753055572509766, "learning_rate": 4.5193789261585876e-05, "loss": 1.4787, "step": 1757 }, { "epoch": 14.064, "grad_norm": 52.36049270629883, "learning_rate": 4.5187862984473154e-05, "loss": 1.9674, "step": 1758 }, { "epoch": 14.072, "grad_norm": 38.64689254760742, "learning_rate": 4.518193670736044e-05, "loss": 1.7419, "step": 1759 }, { "epoch": 14.08, "grad_norm": 80.44627380371094, "learning_rate": 4.5176010430247726e-05, "loss": 1.5858, "step": 1760 }, { "epoch": 14.088, "grad_norm": 49.463050842285156, "learning_rate": 4.5170084153135e-05, "loss": 1.3744, "step": 1761 }, { "epoch": 14.096, "grad_norm": 269.1121520996094, "learning_rate": 4.516415787602228e-05, "loss": 1.8483, "step": 1762 }, { "epoch": 14.104, "grad_norm": 43.692481994628906, "learning_rate": 4.515823159890957e-05, "loss": 1.8592, "step": 1763 }, { "epoch": 14.112, "grad_norm": 29.935543060302734, "learning_rate": 4.515230532179685e-05, "loss": 1.9192, "step": 1764 }, { "epoch": 14.12, "grad_norm": 35.548770904541016, "learning_rate": 4.514637904468413e-05, "loss": 1.4736, "step": 1765 }, { "epoch": 14.128, "grad_norm": 23.992231369018555, "learning_rate": 4.514045276757141e-05, "loss": 1.9658, "step": 1766 }, { "epoch": 14.136, "grad_norm": 26.941587448120117, "learning_rate": 4.51345264904587e-05, "loss": 2.3447, "step": 1767 }, { "epoch": 14.144, "grad_norm": 21.437965393066406, "learning_rate": 4.5128600213345977e-05, "loss": 1.6559, "step": 1768 }, { "epoch": 14.152, "grad_norm": 51.98960494995117, "learning_rate": 4.512267393623326e-05, "loss": 1.6231, "step": 1769 }, { "epoch": 14.16, "grad_norm": 202.79627990722656, "learning_rate": 4.511674765912055e-05, "loss": 1.5851, "step": 1770 }, { "epoch": 14.168, "grad_norm": 25.243358612060547, "learning_rate": 4.511082138200782e-05, "loss": 1.6732, "step": 1771 }, { "epoch": 14.176, "grad_norm": 60.98804473876953, "learning_rate": 4.5104895104895105e-05, "loss": 1.4383, "step": 1772 }, { "epoch": 14.184, "grad_norm": 35.8972282409668, "learning_rate": 4.509896882778239e-05, "loss": 1.5314, "step": 1773 }, { "epoch": 14.192, "grad_norm": 29.72483253479004, "learning_rate": 4.509304255066967e-05, "loss": 1.9731, "step": 1774 }, { "epoch": 14.2, "grad_norm": 79.49376678466797, "learning_rate": 4.5087116273556955e-05, "loss": 1.9452, "step": 1775 }, { "epoch": 14.208, "grad_norm": 45.94391632080078, "learning_rate": 4.5081189996444234e-05, "loss": 1.5542, "step": 1776 }, { "epoch": 14.216, "grad_norm": 26.48576545715332, "learning_rate": 4.507526371933152e-05, "loss": 1.7145, "step": 1777 }, { "epoch": 14.224, "grad_norm": 38.3747444152832, "learning_rate": 4.50693374422188e-05, "loss": 1.4249, "step": 1778 }, { "epoch": 14.232, "grad_norm": 20.005117416381836, "learning_rate": 4.5063411165106084e-05, "loss": 1.6001, "step": 1779 }, { "epoch": 14.24, "grad_norm": 45.620941162109375, "learning_rate": 4.505748488799336e-05, "loss": 1.5257, "step": 1780 }, { "epoch": 14.248, "grad_norm": 27.929813385009766, "learning_rate": 4.505155861088065e-05, "loss": 1.5295, "step": 1781 }, { "epoch": 14.256, "grad_norm": 60.814273834228516, "learning_rate": 4.504563233376793e-05, "loss": 2.5821, "step": 1782 }, { "epoch": 14.264, "grad_norm": 47.92988586425781, "learning_rate": 4.503970605665521e-05, "loss": 1.4727, "step": 1783 }, { "epoch": 14.272, "grad_norm": 46.44248962402344, "learning_rate": 4.50337797795425e-05, "loss": 2.2146, "step": 1784 }, { "epoch": 14.28, "grad_norm": 24.403779983520508, "learning_rate": 4.502785350242977e-05, "loss": 1.5154, "step": 1785 }, { "epoch": 14.288, "grad_norm": 20.03521728515625, "learning_rate": 4.5021927225317056e-05, "loss": 1.6062, "step": 1786 }, { "epoch": 14.296, "grad_norm": 27.864351272583008, "learning_rate": 4.501600094820434e-05, "loss": 1.3953, "step": 1787 }, { "epoch": 14.304, "grad_norm": 67.47174072265625, "learning_rate": 4.501007467109162e-05, "loss": 1.7846, "step": 1788 }, { "epoch": 14.312, "grad_norm": 33.68482208251953, "learning_rate": 4.5004148393978906e-05, "loss": 1.7669, "step": 1789 }, { "epoch": 14.32, "grad_norm": 29.148542404174805, "learning_rate": 4.4998222116866185e-05, "loss": 2.221, "step": 1790 }, { "epoch": 14.328, "grad_norm": 65.87110900878906, "learning_rate": 4.499229583975347e-05, "loss": 1.4727, "step": 1791 }, { "epoch": 14.336, "grad_norm": 59.82496643066406, "learning_rate": 4.498636956264075e-05, "loss": 1.3523, "step": 1792 }, { "epoch": 14.344, "grad_norm": 31.995038986206055, "learning_rate": 4.4980443285528035e-05, "loss": 1.209, "step": 1793 }, { "epoch": 14.352, "grad_norm": 50.34724807739258, "learning_rate": 4.497451700841532e-05, "loss": 1.5327, "step": 1794 }, { "epoch": 14.36, "grad_norm": 72.68290710449219, "learning_rate": 4.496859073130259e-05, "loss": 1.3016, "step": 1795 }, { "epoch": 14.368, "grad_norm": 28.40871810913086, "learning_rate": 4.496266445418988e-05, "loss": 1.2614, "step": 1796 }, { "epoch": 14.376, "grad_norm": 87.87661743164062, "learning_rate": 4.4956738177077164e-05, "loss": 1.422, "step": 1797 }, { "epoch": 14.384, "grad_norm": 73.81114196777344, "learning_rate": 4.495081189996444e-05, "loss": 1.5191, "step": 1798 }, { "epoch": 14.392, "grad_norm": 49.63905715942383, "learning_rate": 4.494488562285173e-05, "loss": 1.6868, "step": 1799 }, { "epoch": 14.4, "grad_norm": 20.682750701904297, "learning_rate": 4.493895934573901e-05, "loss": 1.636, "step": 1800 }, { "epoch": 14.408, "grad_norm": 58.32742691040039, "learning_rate": 4.493303306862629e-05, "loss": 1.744, "step": 1801 }, { "epoch": 14.416, "grad_norm": 34.77192687988281, "learning_rate": 4.492710679151357e-05, "loss": 1.4713, "step": 1802 }, { "epoch": 14.424, "grad_norm": 95.12814331054688, "learning_rate": 4.492118051440086e-05, "loss": 1.6833, "step": 1803 }, { "epoch": 14.432, "grad_norm": 52.68931198120117, "learning_rate": 4.491525423728814e-05, "loss": 1.7373, "step": 1804 }, { "epoch": 14.44, "grad_norm": 35.0311164855957, "learning_rate": 4.4909327960175415e-05, "loss": 1.8046, "step": 1805 }, { "epoch": 14.448, "grad_norm": 46.11937713623047, "learning_rate": 4.49034016830627e-05, "loss": 1.5642, "step": 1806 }, { "epoch": 14.456, "grad_norm": 51.634586334228516, "learning_rate": 4.4897475405949986e-05, "loss": 2.0236, "step": 1807 }, { "epoch": 14.464, "grad_norm": 38.08592987060547, "learning_rate": 4.4891549128837265e-05, "loss": 1.5823, "step": 1808 }, { "epoch": 14.472, "grad_norm": 26.29191780090332, "learning_rate": 4.488562285172455e-05, "loss": 1.3532, "step": 1809 }, { "epoch": 14.48, "grad_norm": 27.648273468017578, "learning_rate": 4.487969657461183e-05, "loss": 1.0972, "step": 1810 }, { "epoch": 14.488, "grad_norm": 36.09336471557617, "learning_rate": 4.4873770297499115e-05, "loss": 2.187, "step": 1811 }, { "epoch": 14.496, "grad_norm": 66.62216186523438, "learning_rate": 4.4867844020386394e-05, "loss": 1.6689, "step": 1812 }, { "epoch": 14.504, "grad_norm": 25.479869842529297, "learning_rate": 4.486191774327368e-05, "loss": 1.4364, "step": 1813 }, { "epoch": 14.512, "grad_norm": 39.160011291503906, "learning_rate": 4.4855991466160965e-05, "loss": 3.5495, "step": 1814 }, { "epoch": 14.52, "grad_norm": 68.71603393554688, "learning_rate": 4.4850065189048244e-05, "loss": 1.5833, "step": 1815 }, { "epoch": 14.528, "grad_norm": 36.02763366699219, "learning_rate": 4.484413891193552e-05, "loss": 1.545, "step": 1816 }, { "epoch": 14.536, "grad_norm": 27.534786224365234, "learning_rate": 4.483821263482281e-05, "loss": 1.6961, "step": 1817 }, { "epoch": 14.544, "grad_norm": 23.141128540039062, "learning_rate": 4.4832286357710094e-05, "loss": 1.4106, "step": 1818 }, { "epoch": 14.552, "grad_norm": 25.29660415649414, "learning_rate": 4.4826360080597366e-05, "loss": 1.0018, "step": 1819 }, { "epoch": 14.56, "grad_norm": 26.416217803955078, "learning_rate": 4.482043380348465e-05, "loss": 1.6964, "step": 1820 }, { "epoch": 14.568, "grad_norm": 30.136791229248047, "learning_rate": 4.481450752637194e-05, "loss": 1.52, "step": 1821 }, { "epoch": 14.576, "grad_norm": 62.370361328125, "learning_rate": 4.4808581249259216e-05, "loss": 1.4273, "step": 1822 }, { "epoch": 14.584, "grad_norm": 21.542627334594727, "learning_rate": 4.48026549721465e-05, "loss": 1.7432, "step": 1823 }, { "epoch": 14.592, "grad_norm": 26.158292770385742, "learning_rate": 4.479672869503378e-05, "loss": 1.4129, "step": 1824 }, { "epoch": 14.6, "grad_norm": 85.59032440185547, "learning_rate": 4.4790802417921066e-05, "loss": 2.034, "step": 1825 }, { "epoch": 14.608, "grad_norm": 29.820281982421875, "learning_rate": 4.4784876140808345e-05, "loss": 1.5424, "step": 1826 }, { "epoch": 14.616, "grad_norm": 25.69774055480957, "learning_rate": 4.477894986369563e-05, "loss": 1.312, "step": 1827 }, { "epoch": 14.624, "grad_norm": 35.30072784423828, "learning_rate": 4.4773023586582916e-05, "loss": 1.5977, "step": 1828 }, { "epoch": 14.632, "grad_norm": 55.734130859375, "learning_rate": 4.476709730947019e-05, "loss": 1.7031, "step": 1829 }, { "epoch": 14.64, "grad_norm": 54.36968994140625, "learning_rate": 4.4761171032357473e-05, "loss": 1.6876, "step": 1830 }, { "epoch": 14.648, "grad_norm": 31.567584991455078, "learning_rate": 4.475524475524476e-05, "loss": 2.271, "step": 1831 }, { "epoch": 14.656, "grad_norm": 25.633079528808594, "learning_rate": 4.474931847813204e-05, "loss": 1.2444, "step": 1832 }, { "epoch": 14.664, "grad_norm": 73.41236877441406, "learning_rate": 4.4743392201019323e-05, "loss": 1.2333, "step": 1833 }, { "epoch": 14.672, "grad_norm": 27.6829891204834, "learning_rate": 4.47374659239066e-05, "loss": 1.6371, "step": 1834 }, { "epoch": 14.68, "grad_norm": 30.166410446166992, "learning_rate": 4.473153964679389e-05, "loss": 1.6717, "step": 1835 }, { "epoch": 14.688, "grad_norm": 28.887985229492188, "learning_rate": 4.472561336968117e-05, "loss": 1.6721, "step": 1836 }, { "epoch": 14.696, "grad_norm": 640.5750732421875, "learning_rate": 4.471968709256845e-05, "loss": 1.4391, "step": 1837 }, { "epoch": 14.704, "grad_norm": 40.46721267700195, "learning_rate": 4.471376081545574e-05, "loss": 1.6107, "step": 1838 }, { "epoch": 14.712, "grad_norm": 36.94208526611328, "learning_rate": 4.470783453834301e-05, "loss": 2.0385, "step": 1839 }, { "epoch": 14.72, "grad_norm": 25.53841209411621, "learning_rate": 4.4701908261230295e-05, "loss": 1.4823, "step": 1840 }, { "epoch": 14.728, "grad_norm": 66.6404037475586, "learning_rate": 4.469598198411758e-05, "loss": 1.5608, "step": 1841 }, { "epoch": 14.736, "grad_norm": 43.11506271362305, "learning_rate": 4.469005570700486e-05, "loss": 1.4178, "step": 1842 }, { "epoch": 14.744, "grad_norm": 54.44127655029297, "learning_rate": 4.4684129429892145e-05, "loss": 1.6444, "step": 1843 }, { "epoch": 14.752, "grad_norm": 62.57443618774414, "learning_rate": 4.4678203152779424e-05, "loss": 1.8343, "step": 1844 }, { "epoch": 14.76, "grad_norm": 86.24658966064453, "learning_rate": 4.467227687566671e-05, "loss": 1.3513, "step": 1845 }, { "epoch": 14.768, "grad_norm": 20.176685333251953, "learning_rate": 4.466635059855399e-05, "loss": 1.3444, "step": 1846 }, { "epoch": 14.776, "grad_norm": 27.332807540893555, "learning_rate": 4.4660424321441274e-05, "loss": 1.8548, "step": 1847 }, { "epoch": 14.784, "grad_norm": 112.90435028076172, "learning_rate": 4.465449804432856e-05, "loss": 1.5146, "step": 1848 }, { "epoch": 14.792, "grad_norm": 74.28414916992188, "learning_rate": 4.464857176721584e-05, "loss": 1.2075, "step": 1849 }, { "epoch": 14.8, "grad_norm": 32.478736877441406, "learning_rate": 4.464264549010312e-05, "loss": 1.4233, "step": 1850 }, { "epoch": 14.808, "grad_norm": 29.465600967407227, "learning_rate": 4.46367192129904e-05, "loss": 1.6684, "step": 1851 }, { "epoch": 14.816, "grad_norm": 25.489389419555664, "learning_rate": 4.463079293587769e-05, "loss": 1.6497, "step": 1852 }, { "epoch": 14.824, "grad_norm": 31.89231300354004, "learning_rate": 4.462486665876496e-05, "loss": 1.4582, "step": 1853 }, { "epoch": 14.832, "grad_norm": 35.38837432861328, "learning_rate": 4.4618940381652246e-05, "loss": 1.4514, "step": 1854 }, { "epoch": 14.84, "grad_norm": 35.216209411621094, "learning_rate": 4.461301410453953e-05, "loss": 1.5143, "step": 1855 }, { "epoch": 14.848, "grad_norm": 20.43658447265625, "learning_rate": 4.460708782742681e-05, "loss": 1.2997, "step": 1856 }, { "epoch": 14.856, "grad_norm": 17.55248260498047, "learning_rate": 4.4601161550314096e-05, "loss": 1.5131, "step": 1857 }, { "epoch": 14.864, "grad_norm": 49.36525344848633, "learning_rate": 4.4595235273201375e-05, "loss": 1.3015, "step": 1858 }, { "epoch": 14.872, "grad_norm": 33.87916564941406, "learning_rate": 4.458930899608866e-05, "loss": 1.5123, "step": 1859 }, { "epoch": 14.88, "grad_norm": 31.4554500579834, "learning_rate": 4.458338271897594e-05, "loss": 1.4753, "step": 1860 }, { "epoch": 14.888, "grad_norm": 40.92142105102539, "learning_rate": 4.4577456441863225e-05, "loss": 1.6257, "step": 1861 }, { "epoch": 14.896, "grad_norm": 48.70922088623047, "learning_rate": 4.457153016475051e-05, "loss": 1.5835, "step": 1862 }, { "epoch": 14.904, "grad_norm": 42.67210388183594, "learning_rate": 4.456560388763778e-05, "loss": 1.3623, "step": 1863 }, { "epoch": 14.912, "grad_norm": 32.85511016845703, "learning_rate": 4.455967761052507e-05, "loss": 1.3799, "step": 1864 }, { "epoch": 14.92, "grad_norm": 22.943815231323242, "learning_rate": 4.4553751333412354e-05, "loss": 1.3316, "step": 1865 }, { "epoch": 14.928, "grad_norm": 17.598844528198242, "learning_rate": 4.454782505629963e-05, "loss": 1.5703, "step": 1866 }, { "epoch": 14.936, "grad_norm": 22.77312469482422, "learning_rate": 4.454189877918692e-05, "loss": 2.7335, "step": 1867 }, { "epoch": 14.943999999999999, "grad_norm": 35.7796630859375, "learning_rate": 4.45359725020742e-05, "loss": 2.2433, "step": 1868 }, { "epoch": 14.952, "grad_norm": 25.816587448120117, "learning_rate": 4.453004622496148e-05, "loss": 1.3301, "step": 1869 }, { "epoch": 14.96, "grad_norm": 48.960968017578125, "learning_rate": 4.452411994784876e-05, "loss": 1.466, "step": 1870 }, { "epoch": 14.968, "grad_norm": 61.194740295410156, "learning_rate": 4.451819367073605e-05, "loss": 1.7497, "step": 1871 }, { "epoch": 14.975999999999999, "grad_norm": 54.826480865478516, "learning_rate": 4.451226739362333e-05, "loss": 2.5045, "step": 1872 }, { "epoch": 14.984, "grad_norm": 86.11405944824219, "learning_rate": 4.4506341116510605e-05, "loss": 1.649, "step": 1873 }, { "epoch": 14.992, "grad_norm": 39.66370391845703, "learning_rate": 4.450041483939789e-05, "loss": 1.4115, "step": 1874 }, { "epoch": 15.0, "grad_norm": 67.38797760009766, "learning_rate": 4.4494488562285176e-05, "loss": 1.7126, "step": 1875 }, { "epoch": 15.0, "eval_loss": 1.6890997886657715, "eval_map": 0.1975, "eval_map_50": 0.4328, "eval_map_75": 0.1627, "eval_map_Coverall": 0.3875, "eval_map_Face_Shield": 0.1943, "eval_map_Gloves": 0.12, "eval_map_Goggles": 0.0268, "eval_map_Mask": 0.259, "eval_map_large": 0.2535, "eval_map_medium": 0.1145, "eval_map_small": 0.2167, "eval_mar_1": 0.1879, "eval_mar_10": 0.3585, "eval_mar_100": 0.3683, "eval_mar_100_Coverall": 0.6178, "eval_mar_100_Face_Shield": 0.4588, "eval_mar_100_Gloves": 0.2295, "eval_mar_100_Goggles": 0.1719, "eval_mar_100_Mask": 0.3635, "eval_mar_large": 0.4763, "eval_mar_medium": 0.2441, "eval_mar_small": 0.2405, "eval_runtime": 1.8475, "eval_samples_per_second": 15.697, "eval_steps_per_second": 1.083, "step": 1875 }, { "epoch": 15.008, "grad_norm": 25.686914443969727, "learning_rate": 4.4488562285172455e-05, "loss": 3.1202, "step": 1876 }, { "epoch": 15.016, "grad_norm": 23.0462589263916, "learning_rate": 4.448263600805974e-05, "loss": 1.7298, "step": 1877 }, { "epoch": 15.024, "grad_norm": 27.92193603515625, "learning_rate": 4.447670973094702e-05, "loss": 1.9034, "step": 1878 }, { "epoch": 15.032, "grad_norm": 44.40872573852539, "learning_rate": 4.4470783453834305e-05, "loss": 1.6108, "step": 1879 }, { "epoch": 15.04, "grad_norm": 43.592281341552734, "learning_rate": 4.4464857176721584e-05, "loss": 1.8482, "step": 1880 }, { "epoch": 15.048, "grad_norm": 62.99137878417969, "learning_rate": 4.445893089960887e-05, "loss": 1.3592, "step": 1881 }, { "epoch": 15.056, "grad_norm": 24.803234100341797, "learning_rate": 4.4453004622496155e-05, "loss": 1.6178, "step": 1882 }, { "epoch": 15.064, "grad_norm": 26.606266021728516, "learning_rate": 4.4447078345383434e-05, "loss": 1.4507, "step": 1883 }, { "epoch": 15.072, "grad_norm": 46.70610809326172, "learning_rate": 4.444115206827071e-05, "loss": 2.4069, "step": 1884 }, { "epoch": 15.08, "grad_norm": 39.43834686279297, "learning_rate": 4.4435225791158e-05, "loss": 1.5879, "step": 1885 }, { "epoch": 15.088, "grad_norm": 44.25823211669922, "learning_rate": 4.4429299514045284e-05, "loss": 1.8276, "step": 1886 }, { "epoch": 15.096, "grad_norm": 31.80576515197754, "learning_rate": 4.4423373236932556e-05, "loss": 1.3416, "step": 1887 }, { "epoch": 15.104, "grad_norm": 22.149377822875977, "learning_rate": 4.441744695981984e-05, "loss": 1.6585, "step": 1888 }, { "epoch": 15.112, "grad_norm": 33.936851501464844, "learning_rate": 4.441152068270713e-05, "loss": 1.4207, "step": 1889 }, { "epoch": 15.12, "grad_norm": 64.58915710449219, "learning_rate": 4.4405594405594406e-05, "loss": 1.411, "step": 1890 }, { "epoch": 15.128, "grad_norm": 40.99076461791992, "learning_rate": 4.439966812848169e-05, "loss": 1.4774, "step": 1891 }, { "epoch": 15.136, "grad_norm": 50.21870422363281, "learning_rate": 4.439374185136897e-05, "loss": 1.1986, "step": 1892 }, { "epoch": 15.144, "grad_norm": 38.28419876098633, "learning_rate": 4.4387815574256256e-05, "loss": 1.559, "step": 1893 }, { "epoch": 15.152, "grad_norm": 31.914432525634766, "learning_rate": 4.4381889297143535e-05, "loss": 1.9404, "step": 1894 }, { "epoch": 15.16, "grad_norm": 48.44687271118164, "learning_rate": 4.437596302003082e-05, "loss": 1.3079, "step": 1895 }, { "epoch": 15.168, "grad_norm": 33.66891098022461, "learning_rate": 4.4370036742918106e-05, "loss": 1.4191, "step": 1896 }, { "epoch": 15.176, "grad_norm": 55.14872741699219, "learning_rate": 4.436411046580538e-05, "loss": 2.4295, "step": 1897 }, { "epoch": 15.184, "grad_norm": 53.81166076660156, "learning_rate": 4.4358184188692663e-05, "loss": 1.2126, "step": 1898 }, { "epoch": 15.192, "grad_norm": 17.52688980102539, "learning_rate": 4.435225791157995e-05, "loss": 1.7346, "step": 1899 }, { "epoch": 15.2, "grad_norm": 25.520404815673828, "learning_rate": 4.434633163446723e-05, "loss": 1.1458, "step": 1900 }, { "epoch": 15.208, "grad_norm": 59.47291946411133, "learning_rate": 4.4340405357354513e-05, "loss": 1.6412, "step": 1901 }, { "epoch": 15.216, "grad_norm": 26.287275314331055, "learning_rate": 4.433447908024179e-05, "loss": 1.3341, "step": 1902 }, { "epoch": 15.224, "grad_norm": 40.86334991455078, "learning_rate": 4.432855280312908e-05, "loss": 1.8003, "step": 1903 }, { "epoch": 15.232, "grad_norm": 251.53823852539062, "learning_rate": 4.432262652601636e-05, "loss": 1.967, "step": 1904 }, { "epoch": 15.24, "grad_norm": 44.21027374267578, "learning_rate": 4.431670024890364e-05, "loss": 1.385, "step": 1905 }, { "epoch": 15.248, "grad_norm": 23.136873245239258, "learning_rate": 4.431077397179093e-05, "loss": 1.6091, "step": 1906 }, { "epoch": 15.256, "grad_norm": 41.67219161987305, "learning_rate": 4.43048476946782e-05, "loss": 1.7166, "step": 1907 }, { "epoch": 15.264, "grad_norm": 29.261369705200195, "learning_rate": 4.4298921417565486e-05, "loss": 1.8954, "step": 1908 }, { "epoch": 15.272, "grad_norm": 24.788162231445312, "learning_rate": 4.429299514045277e-05, "loss": 1.7702, "step": 1909 }, { "epoch": 15.28, "grad_norm": 25.312992095947266, "learning_rate": 4.428706886334005e-05, "loss": 1.4161, "step": 1910 }, { "epoch": 15.288, "grad_norm": 19.424428939819336, "learning_rate": 4.4281142586227336e-05, "loss": 1.8262, "step": 1911 }, { "epoch": 15.296, "grad_norm": 36.47010040283203, "learning_rate": 4.4275216309114614e-05, "loss": 1.5279, "step": 1912 }, { "epoch": 15.304, "grad_norm": 29.225582122802734, "learning_rate": 4.42692900320019e-05, "loss": 1.6987, "step": 1913 }, { "epoch": 15.312, "grad_norm": 57.94115447998047, "learning_rate": 4.426336375488918e-05, "loss": 1.537, "step": 1914 }, { "epoch": 15.32, "grad_norm": 20.751890182495117, "learning_rate": 4.4257437477776464e-05, "loss": 1.6018, "step": 1915 }, { "epoch": 15.328, "grad_norm": 30.181241989135742, "learning_rate": 4.425151120066375e-05, "loss": 1.6018, "step": 1916 }, { "epoch": 15.336, "grad_norm": 42.5529899597168, "learning_rate": 4.424558492355103e-05, "loss": 1.9172, "step": 1917 }, { "epoch": 15.344, "grad_norm": 40.28279113769531, "learning_rate": 4.423965864643831e-05, "loss": 2.1012, "step": 1918 }, { "epoch": 15.352, "grad_norm": 41.38288116455078, "learning_rate": 4.423373236932559e-05, "loss": 1.5385, "step": 1919 }, { "epoch": 15.36, "grad_norm": 26.581878662109375, "learning_rate": 4.422780609221288e-05, "loss": 1.3132, "step": 1920 }, { "epoch": 15.368, "grad_norm": 19.41523551940918, "learning_rate": 4.422187981510015e-05, "loss": 1.2267, "step": 1921 }, { "epoch": 15.376, "grad_norm": 25.079931259155273, "learning_rate": 4.4215953537987436e-05, "loss": 1.6776, "step": 1922 }, { "epoch": 15.384, "grad_norm": 110.33370208740234, "learning_rate": 4.421002726087472e-05, "loss": 1.1945, "step": 1923 }, { "epoch": 15.392, "grad_norm": 40.574180603027344, "learning_rate": 4.4204100983762e-05, "loss": 1.7044, "step": 1924 }, { "epoch": 15.4, "grad_norm": 44.80343246459961, "learning_rate": 4.4198174706649286e-05, "loss": 1.1946, "step": 1925 }, { "epoch": 15.408, "grad_norm": 17.430564880371094, "learning_rate": 4.4192248429536565e-05, "loss": 1.3113, "step": 1926 }, { "epoch": 15.416, "grad_norm": 24.78534507751465, "learning_rate": 4.418632215242385e-05, "loss": 1.3744, "step": 1927 }, { "epoch": 15.424, "grad_norm": 15.774944305419922, "learning_rate": 4.418039587531113e-05, "loss": 1.2288, "step": 1928 }, { "epoch": 15.432, "grad_norm": 50.53965759277344, "learning_rate": 4.4174469598198415e-05, "loss": 1.6775, "step": 1929 }, { "epoch": 15.44, "grad_norm": 25.2083683013916, "learning_rate": 4.41685433210857e-05, "loss": 1.4881, "step": 1930 }, { "epoch": 15.448, "grad_norm": 31.109689712524414, "learning_rate": 4.416261704397297e-05, "loss": 1.5133, "step": 1931 }, { "epoch": 15.456, "grad_norm": 22.34514045715332, "learning_rate": 4.415669076686026e-05, "loss": 1.7328, "step": 1932 }, { "epoch": 15.464, "grad_norm": 62.980106353759766, "learning_rate": 4.4150764489747544e-05, "loss": 1.2865, "step": 1933 }, { "epoch": 15.472, "grad_norm": 31.398448944091797, "learning_rate": 4.414483821263482e-05, "loss": 1.6067, "step": 1934 }, { "epoch": 15.48, "grad_norm": 88.89078521728516, "learning_rate": 4.413891193552211e-05, "loss": 1.6205, "step": 1935 }, { "epoch": 15.488, "grad_norm": 24.189468383789062, "learning_rate": 4.413298565840939e-05, "loss": 1.3391, "step": 1936 }, { "epoch": 15.496, "grad_norm": 19.74510955810547, "learning_rate": 4.412705938129667e-05, "loss": 1.647, "step": 1937 }, { "epoch": 15.504, "grad_norm": 51.99030303955078, "learning_rate": 4.412113310418395e-05, "loss": 1.5066, "step": 1938 }, { "epoch": 15.512, "grad_norm": 21.22608184814453, "learning_rate": 4.411520682707124e-05, "loss": 1.0993, "step": 1939 }, { "epoch": 15.52, "grad_norm": 22.729108810424805, "learning_rate": 4.410928054995852e-05, "loss": 1.7115, "step": 1940 }, { "epoch": 15.528, "grad_norm": 21.458322525024414, "learning_rate": 4.4103354272845795e-05, "loss": 1.4134, "step": 1941 }, { "epoch": 15.536, "grad_norm": 30.547607421875, "learning_rate": 4.409742799573308e-05, "loss": 1.1637, "step": 1942 }, { "epoch": 15.544, "grad_norm": 34.04389953613281, "learning_rate": 4.4091501718620366e-05, "loss": 1.9673, "step": 1943 }, { "epoch": 15.552, "grad_norm": 31.88866424560547, "learning_rate": 4.4085575441507645e-05, "loss": 1.2609, "step": 1944 }, { "epoch": 15.56, "grad_norm": 23.35139274597168, "learning_rate": 4.407964916439493e-05, "loss": 1.7181, "step": 1945 }, { "epoch": 15.568, "grad_norm": 38.607669830322266, "learning_rate": 4.407372288728221e-05, "loss": 1.7006, "step": 1946 }, { "epoch": 15.576, "grad_norm": 33.47078323364258, "learning_rate": 4.4067796610169495e-05, "loss": 2.0173, "step": 1947 }, { "epoch": 15.584, "grad_norm": 42.59926986694336, "learning_rate": 4.4061870333056774e-05, "loss": 1.4563, "step": 1948 }, { "epoch": 15.592, "grad_norm": 46.33942794799805, "learning_rate": 4.405594405594406e-05, "loss": 2.7839, "step": 1949 }, { "epoch": 15.6, "grad_norm": 38.10004425048828, "learning_rate": 4.4050017778831345e-05, "loss": 1.5783, "step": 1950 }, { "epoch": 15.608, "grad_norm": 25.836345672607422, "learning_rate": 4.4044091501718624e-05, "loss": 1.432, "step": 1951 }, { "epoch": 15.616, "grad_norm": 21.738285064697266, "learning_rate": 4.40381652246059e-05, "loss": 1.4757, "step": 1952 }, { "epoch": 15.624, "grad_norm": 33.34440231323242, "learning_rate": 4.403223894749319e-05, "loss": 2.0626, "step": 1953 }, { "epoch": 15.632, "grad_norm": 22.486804962158203, "learning_rate": 4.4026312670380474e-05, "loss": 1.7391, "step": 1954 }, { "epoch": 15.64, "grad_norm": 62.631343841552734, "learning_rate": 4.402038639326775e-05, "loss": 1.3898, "step": 1955 }, { "epoch": 15.648, "grad_norm": 25.80547332763672, "learning_rate": 4.401446011615503e-05, "loss": 1.8481, "step": 1956 }, { "epoch": 15.656, "grad_norm": 17.407154083251953, "learning_rate": 4.400853383904232e-05, "loss": 1.3496, "step": 1957 }, { "epoch": 15.664, "grad_norm": 30.667022705078125, "learning_rate": 4.4002607561929596e-05, "loss": 1.1685, "step": 1958 }, { "epoch": 15.672, "grad_norm": 25.58618927001953, "learning_rate": 4.399668128481688e-05, "loss": 1.2941, "step": 1959 }, { "epoch": 15.68, "grad_norm": 24.707717895507812, "learning_rate": 4.399075500770416e-05, "loss": 1.7032, "step": 1960 }, { "epoch": 15.688, "grad_norm": 30.580961227416992, "learning_rate": 4.3984828730591446e-05, "loss": 1.2518, "step": 1961 }, { "epoch": 15.696, "grad_norm": 37.068973541259766, "learning_rate": 4.3978902453478725e-05, "loss": 1.2523, "step": 1962 }, { "epoch": 15.704, "grad_norm": 25.708620071411133, "learning_rate": 4.397297617636601e-05, "loss": 1.156, "step": 1963 }, { "epoch": 15.712, "grad_norm": 17.102752685546875, "learning_rate": 4.3967049899253296e-05, "loss": 1.7418, "step": 1964 }, { "epoch": 15.72, "grad_norm": 24.243858337402344, "learning_rate": 4.396112362214057e-05, "loss": 0.9544, "step": 1965 }, { "epoch": 15.728, "grad_norm": 30.596559524536133, "learning_rate": 4.3955197345027854e-05, "loss": 1.8197, "step": 1966 }, { "epoch": 15.736, "grad_norm": 63.7642936706543, "learning_rate": 4.394927106791514e-05, "loss": 1.6218, "step": 1967 }, { "epoch": 15.744, "grad_norm": 1978.3433837890625, "learning_rate": 4.394334479080242e-05, "loss": 1.4668, "step": 1968 }, { "epoch": 15.752, "grad_norm": 27.29568099975586, "learning_rate": 4.3937418513689704e-05, "loss": 1.7982, "step": 1969 }, { "epoch": 15.76, "grad_norm": 50.52635192871094, "learning_rate": 4.393149223657698e-05, "loss": 2.5697, "step": 1970 }, { "epoch": 15.768, "grad_norm": 19.104413986206055, "learning_rate": 4.392556595946427e-05, "loss": 1.8864, "step": 1971 }, { "epoch": 15.776, "grad_norm": 59.78030014038086, "learning_rate": 4.391963968235155e-05, "loss": 1.4449, "step": 1972 }, { "epoch": 15.784, "grad_norm": 32.630149841308594, "learning_rate": 4.391371340523883e-05, "loss": 1.3433, "step": 1973 }, { "epoch": 15.792, "grad_norm": 16.201889038085938, "learning_rate": 4.390778712812612e-05, "loss": 1.7045, "step": 1974 }, { "epoch": 15.8, "grad_norm": 16.907621383666992, "learning_rate": 4.390186085101339e-05, "loss": 1.851, "step": 1975 }, { "epoch": 15.808, "grad_norm": 48.373775482177734, "learning_rate": 4.3895934573900676e-05, "loss": 1.2088, "step": 1976 }, { "epoch": 15.816, "grad_norm": 29.84575653076172, "learning_rate": 4.389000829678796e-05, "loss": 1.1433, "step": 1977 }, { "epoch": 15.824, "grad_norm": 32.379798889160156, "learning_rate": 4.388408201967524e-05, "loss": 1.8312, "step": 1978 }, { "epoch": 15.832, "grad_norm": 32.20342254638672, "learning_rate": 4.3878155742562526e-05, "loss": 1.8578, "step": 1979 }, { "epoch": 15.84, "grad_norm": 17.305063247680664, "learning_rate": 4.3872229465449804e-05, "loss": 1.4944, "step": 1980 }, { "epoch": 15.848, "grad_norm": 28.594655990600586, "learning_rate": 4.386630318833709e-05, "loss": 1.3466, "step": 1981 }, { "epoch": 15.856, "grad_norm": 21.739063262939453, "learning_rate": 4.386037691122437e-05, "loss": 1.3074, "step": 1982 }, { "epoch": 15.864, "grad_norm": 21.003816604614258, "learning_rate": 4.3854450634111654e-05, "loss": 1.3721, "step": 1983 }, { "epoch": 15.872, "grad_norm": 25.25197982788086, "learning_rate": 4.384852435699894e-05, "loss": 1.3036, "step": 1984 }, { "epoch": 15.88, "grad_norm": 43.840030670166016, "learning_rate": 4.384259807988622e-05, "loss": 1.0711, "step": 1985 }, { "epoch": 15.888, "grad_norm": 14.895344734191895, "learning_rate": 4.38366718027735e-05, "loss": 1.2728, "step": 1986 }, { "epoch": 15.896, "grad_norm": 24.9890079498291, "learning_rate": 4.383074552566078e-05, "loss": 1.5063, "step": 1987 }, { "epoch": 15.904, "grad_norm": 32.782718658447266, "learning_rate": 4.382481924854807e-05, "loss": 1.6482, "step": 1988 }, { "epoch": 15.912, "grad_norm": 130.9415740966797, "learning_rate": 4.381889297143535e-05, "loss": 1.6071, "step": 1989 }, { "epoch": 15.92, "grad_norm": 22.459781646728516, "learning_rate": 4.3812966694322627e-05, "loss": 1.1465, "step": 1990 }, { "epoch": 15.928, "grad_norm": 26.918365478515625, "learning_rate": 4.380704041720991e-05, "loss": 1.4981, "step": 1991 }, { "epoch": 15.936, "grad_norm": 30.763465881347656, "learning_rate": 4.380111414009719e-05, "loss": 1.7866, "step": 1992 }, { "epoch": 15.943999999999999, "grad_norm": 52.55703353881836, "learning_rate": 4.3795187862984477e-05, "loss": 1.3971, "step": 1993 }, { "epoch": 15.952, "grad_norm": 27.880775451660156, "learning_rate": 4.3789261585871755e-05, "loss": 1.8314, "step": 1994 }, { "epoch": 15.96, "grad_norm": 20.842294692993164, "learning_rate": 4.378333530875904e-05, "loss": 0.9851, "step": 1995 }, { "epoch": 15.968, "grad_norm": 26.83616828918457, "learning_rate": 4.377740903164632e-05, "loss": 1.2454, "step": 1996 }, { "epoch": 15.975999999999999, "grad_norm": 55.91450500488281, "learning_rate": 4.3771482754533605e-05, "loss": 1.7777, "step": 1997 }, { "epoch": 15.984, "grad_norm": 255.81065368652344, "learning_rate": 4.376555647742089e-05, "loss": 1.3205, "step": 1998 }, { "epoch": 15.992, "grad_norm": 20.133241653442383, "learning_rate": 4.375963020030816e-05, "loss": 1.2665, "step": 1999 }, { "epoch": 16.0, "grad_norm": 53.61433410644531, "learning_rate": 4.375370392319545e-05, "loss": 1.5176, "step": 2000 }, { "epoch": 16.0, "eval_loss": 1.5962547063827515, "eval_map": 0.2266, "eval_map_50": 0.5059, "eval_map_75": 0.1712, "eval_map_Coverall": 0.4193, "eval_map_Face_Shield": 0.2142, "eval_map_Gloves": 0.1105, "eval_map_Goggles": 0.0751, "eval_map_Mask": 0.3139, "eval_map_large": 0.2777, "eval_map_medium": 0.1522, "eval_map_small": 0.1088, "eval_mar_1": 0.1986, "eval_mar_10": 0.3629, "eval_mar_100": 0.3717, "eval_mar_100_Coverall": 0.6178, "eval_mar_100_Face_Shield": 0.3706, "eval_mar_100_Gloves": 0.2557, "eval_mar_100_Goggles": 0.2125, "eval_mar_100_Mask": 0.4019, "eval_mar_large": 0.4593, "eval_mar_medium": 0.2642, "eval_mar_small": 0.1446, "eval_runtime": 1.8348, "eval_samples_per_second": 15.806, "eval_steps_per_second": 1.09, "step": 2000 }, { "epoch": 16.008, "grad_norm": 38.40007781982422, "learning_rate": 4.3747777646082734e-05, "loss": 1.2867, "step": 2001 }, { "epoch": 16.016, "grad_norm": 37.185630798339844, "learning_rate": 4.374185136897001e-05, "loss": 0.9906, "step": 2002 }, { "epoch": 16.024, "grad_norm": 29.725082397460938, "learning_rate": 4.37359250918573e-05, "loss": 1.4149, "step": 2003 }, { "epoch": 16.032, "grad_norm": 38.98472595214844, "learning_rate": 4.372999881474458e-05, "loss": 2.0641, "step": 2004 }, { "epoch": 16.04, "grad_norm": 37.29946517944336, "learning_rate": 4.372407253763186e-05, "loss": 1.4289, "step": 2005 }, { "epoch": 16.048, "grad_norm": 17.174667358398438, "learning_rate": 4.371814626051914e-05, "loss": 1.2796, "step": 2006 }, { "epoch": 16.056, "grad_norm": 29.335222244262695, "learning_rate": 4.371221998340643e-05, "loss": 1.3898, "step": 2007 }, { "epoch": 16.064, "grad_norm": 44.5551643371582, "learning_rate": 4.370629370629371e-05, "loss": 1.4055, "step": 2008 }, { "epoch": 16.072, "grad_norm": 24.46993064880371, "learning_rate": 4.3700367429180985e-05, "loss": 1.2316, "step": 2009 }, { "epoch": 16.08, "grad_norm": 85.44889831542969, "learning_rate": 4.369444115206827e-05, "loss": 1.387, "step": 2010 }, { "epoch": 16.088, "grad_norm": 28.599218368530273, "learning_rate": 4.3688514874955556e-05, "loss": 1.6205, "step": 2011 }, { "epoch": 16.096, "grad_norm": 25.112003326416016, "learning_rate": 4.3682588597842835e-05, "loss": 1.4828, "step": 2012 }, { "epoch": 16.104, "grad_norm": 228.80296325683594, "learning_rate": 4.367666232073012e-05, "loss": 1.3991, "step": 2013 }, { "epoch": 16.112, "grad_norm": 17.30461883544922, "learning_rate": 4.36707360436174e-05, "loss": 1.318, "step": 2014 }, { "epoch": 16.12, "grad_norm": 47.628936767578125, "learning_rate": 4.3664809766504685e-05, "loss": 1.4229, "step": 2015 }, { "epoch": 16.128, "grad_norm": 24.61898422241211, "learning_rate": 4.3658883489391964e-05, "loss": 1.0688, "step": 2016 }, { "epoch": 16.136, "grad_norm": 20.55666160583496, "learning_rate": 4.365295721227925e-05, "loss": 1.3771, "step": 2017 }, { "epoch": 16.144, "grad_norm": 31.31732940673828, "learning_rate": 4.3647030935166535e-05, "loss": 1.5715, "step": 2018 }, { "epoch": 16.152, "grad_norm": 25.083511352539062, "learning_rate": 4.3641104658053814e-05, "loss": 1.3253, "step": 2019 }, { "epoch": 16.16, "grad_norm": 35.94451904296875, "learning_rate": 4.363517838094109e-05, "loss": 1.572, "step": 2020 }, { "epoch": 16.168, "grad_norm": 28.271142959594727, "learning_rate": 4.362925210382838e-05, "loss": 1.4824, "step": 2021 }, { "epoch": 16.176, "grad_norm": 24.08671760559082, "learning_rate": 4.3623325826715664e-05, "loss": 1.2725, "step": 2022 }, { "epoch": 16.184, "grad_norm": 40.52501678466797, "learning_rate": 4.361739954960294e-05, "loss": 1.413, "step": 2023 }, { "epoch": 16.192, "grad_norm": 20.380615234375, "learning_rate": 4.361147327249022e-05, "loss": 1.2535, "step": 2024 }, { "epoch": 16.2, "grad_norm": 30.935346603393555, "learning_rate": 4.360554699537751e-05, "loss": 1.8895, "step": 2025 }, { "epoch": 16.208, "grad_norm": 27.846973419189453, "learning_rate": 4.3599620718264786e-05, "loss": 2.5615, "step": 2026 }, { "epoch": 16.216, "grad_norm": 22.911067962646484, "learning_rate": 4.359369444115207e-05, "loss": 1.1969, "step": 2027 }, { "epoch": 16.224, "grad_norm": 43.5001220703125, "learning_rate": 4.358776816403935e-05, "loss": 1.3717, "step": 2028 }, { "epoch": 16.232, "grad_norm": 30.151283264160156, "learning_rate": 4.3581841886926636e-05, "loss": 1.2379, "step": 2029 }, { "epoch": 16.24, "grad_norm": 30.95677947998047, "learning_rate": 4.3575915609813915e-05, "loss": 2.2117, "step": 2030 }, { "epoch": 16.248, "grad_norm": 96.08143615722656, "learning_rate": 4.35699893327012e-05, "loss": 1.2437, "step": 2031 }, { "epoch": 16.256, "grad_norm": 26.565061569213867, "learning_rate": 4.3564063055588486e-05, "loss": 2.0333, "step": 2032 }, { "epoch": 16.264, "grad_norm": 30.929214477539062, "learning_rate": 4.355813677847576e-05, "loss": 1.5116, "step": 2033 }, { "epoch": 16.272, "grad_norm": 67.59716033935547, "learning_rate": 4.3552210501363044e-05, "loss": 2.0837, "step": 2034 }, { "epoch": 16.28, "grad_norm": 29.007549285888672, "learning_rate": 4.354628422425033e-05, "loss": 1.4435, "step": 2035 }, { "epoch": 16.288, "grad_norm": 34.40584945678711, "learning_rate": 4.354035794713761e-05, "loss": 1.9358, "step": 2036 }, { "epoch": 16.296, "grad_norm": 30.677471160888672, "learning_rate": 4.3534431670024894e-05, "loss": 1.5631, "step": 2037 }, { "epoch": 16.304, "grad_norm": 38.79332733154297, "learning_rate": 4.352850539291217e-05, "loss": 1.6641, "step": 2038 }, { "epoch": 16.312, "grad_norm": 38.85026931762695, "learning_rate": 4.352257911579946e-05, "loss": 1.0433, "step": 2039 }, { "epoch": 16.32, "grad_norm": 24.978519439697266, "learning_rate": 4.351665283868674e-05, "loss": 1.6422, "step": 2040 }, { "epoch": 16.328, "grad_norm": 107.57600402832031, "learning_rate": 4.351072656157402e-05, "loss": 1.5869, "step": 2041 }, { "epoch": 16.336, "grad_norm": 30.92363929748535, "learning_rate": 4.350480028446131e-05, "loss": 1.1887, "step": 2042 }, { "epoch": 16.344, "grad_norm": 43.66545104980469, "learning_rate": 4.349887400734858e-05, "loss": 2.0567, "step": 2043 }, { "epoch": 16.352, "grad_norm": 31.633434295654297, "learning_rate": 4.3492947730235866e-05, "loss": 1.4253, "step": 2044 }, { "epoch": 16.36, "grad_norm": 34.905948638916016, "learning_rate": 4.348702145312315e-05, "loss": 1.6435, "step": 2045 }, { "epoch": 16.368, "grad_norm": 46.377498626708984, "learning_rate": 4.348109517601043e-05, "loss": 1.7274, "step": 2046 }, { "epoch": 16.376, "grad_norm": 25.682077407836914, "learning_rate": 4.3475168898897716e-05, "loss": 1.5903, "step": 2047 }, { "epoch": 16.384, "grad_norm": 23.3165283203125, "learning_rate": 4.3469242621784995e-05, "loss": 1.5741, "step": 2048 }, { "epoch": 16.392, "grad_norm": 19.42941665649414, "learning_rate": 4.346331634467228e-05, "loss": 1.2253, "step": 2049 }, { "epoch": 16.4, "grad_norm": 45.30641555786133, "learning_rate": 4.345739006755956e-05, "loss": 1.9325, "step": 2050 }, { "epoch": 16.408, "grad_norm": 41.374298095703125, "learning_rate": 4.3451463790446845e-05, "loss": 3.3113, "step": 2051 }, { "epoch": 16.416, "grad_norm": 15.300479888916016, "learning_rate": 4.344553751333413e-05, "loss": 1.0658, "step": 2052 }, { "epoch": 16.424, "grad_norm": 26.631807327270508, "learning_rate": 4.343961123622141e-05, "loss": 1.6186, "step": 2053 }, { "epoch": 16.432, "grad_norm": 30.729326248168945, "learning_rate": 4.343368495910869e-05, "loss": 1.3335, "step": 2054 }, { "epoch": 16.44, "grad_norm": 148.95347595214844, "learning_rate": 4.342775868199597e-05, "loss": 1.3004, "step": 2055 }, { "epoch": 16.448, "grad_norm": 47.058250427246094, "learning_rate": 4.342183240488326e-05, "loss": 1.9138, "step": 2056 }, { "epoch": 16.456, "grad_norm": 19.248178482055664, "learning_rate": 4.341590612777054e-05, "loss": 1.1437, "step": 2057 }, { "epoch": 16.464, "grad_norm": 27.720561981201172, "learning_rate": 4.3409979850657817e-05, "loss": 1.4057, "step": 2058 }, { "epoch": 16.472, "grad_norm": 17.378551483154297, "learning_rate": 4.34040535735451e-05, "loss": 1.1715, "step": 2059 }, { "epoch": 16.48, "grad_norm": 22.76700782775879, "learning_rate": 4.339812729643238e-05, "loss": 1.6501, "step": 2060 }, { "epoch": 16.488, "grad_norm": 233.58831787109375, "learning_rate": 4.339220101931967e-05, "loss": 1.6095, "step": 2061 }, { "epoch": 16.496, "grad_norm": 37.139827728271484, "learning_rate": 4.338627474220695e-05, "loss": 1.8315, "step": 2062 }, { "epoch": 16.504, "grad_norm": 27.004165649414062, "learning_rate": 4.338034846509423e-05, "loss": 1.3101, "step": 2063 }, { "epoch": 16.512, "grad_norm": 38.63199996948242, "learning_rate": 4.337442218798151e-05, "loss": 1.5258, "step": 2064 }, { "epoch": 16.52, "grad_norm": 97.12505340576172, "learning_rate": 4.3368495910868795e-05, "loss": 1.3326, "step": 2065 }, { "epoch": 16.528, "grad_norm": 20.869953155517578, "learning_rate": 4.336256963375608e-05, "loss": 1.6518, "step": 2066 }, { "epoch": 16.536, "grad_norm": 26.270177841186523, "learning_rate": 4.335664335664335e-05, "loss": 1.5149, "step": 2067 }, { "epoch": 16.544, "grad_norm": 28.690412521362305, "learning_rate": 4.335071707953064e-05, "loss": 1.5271, "step": 2068 }, { "epoch": 16.552, "grad_norm": 34.565982818603516, "learning_rate": 4.3344790802417924e-05, "loss": 1.7763, "step": 2069 }, { "epoch": 16.56, "grad_norm": 21.548337936401367, "learning_rate": 4.33388645253052e-05, "loss": 1.3328, "step": 2070 }, { "epoch": 16.568, "grad_norm": 86.05451202392578, "learning_rate": 4.333293824819249e-05, "loss": 1.6183, "step": 2071 }, { "epoch": 16.576, "grad_norm": 26.90651512145996, "learning_rate": 4.332701197107977e-05, "loss": 1.2708, "step": 2072 }, { "epoch": 16.584, "grad_norm": 24.91130828857422, "learning_rate": 4.332108569396705e-05, "loss": 1.4639, "step": 2073 }, { "epoch": 16.592, "grad_norm": 33.946128845214844, "learning_rate": 4.331515941685433e-05, "loss": 1.4883, "step": 2074 }, { "epoch": 16.6, "grad_norm": 20.12603187561035, "learning_rate": 4.330923313974162e-05, "loss": 1.7078, "step": 2075 }, { "epoch": 16.608, "grad_norm": 34.692527770996094, "learning_rate": 4.33033068626289e-05, "loss": 1.2535, "step": 2076 }, { "epoch": 16.616, "grad_norm": 19.224218368530273, "learning_rate": 4.3297380585516175e-05, "loss": 1.2364, "step": 2077 }, { "epoch": 16.624, "grad_norm": 20.60236167907715, "learning_rate": 4.329145430840346e-05, "loss": 1.4772, "step": 2078 }, { "epoch": 16.632, "grad_norm": 32.998146057128906, "learning_rate": 4.3285528031290746e-05, "loss": 1.4188, "step": 2079 }, { "epoch": 16.64, "grad_norm": 32.03946304321289, "learning_rate": 4.3279601754178025e-05, "loss": 2.0196, "step": 2080 }, { "epoch": 16.648, "grad_norm": 19.895732879638672, "learning_rate": 4.327367547706531e-05, "loss": 1.6917, "step": 2081 }, { "epoch": 16.656, "grad_norm": 34.56035232543945, "learning_rate": 4.326774919995259e-05, "loss": 2.2177, "step": 2082 }, { "epoch": 16.664, "grad_norm": 47.158668518066406, "learning_rate": 4.3261822922839875e-05, "loss": 1.4709, "step": 2083 }, { "epoch": 16.672, "grad_norm": 61.668582916259766, "learning_rate": 4.3255896645727154e-05, "loss": 1.3215, "step": 2084 }, { "epoch": 16.68, "grad_norm": 21.74842071533203, "learning_rate": 4.324997036861444e-05, "loss": 1.5874, "step": 2085 }, { "epoch": 16.688, "grad_norm": 26.375774383544922, "learning_rate": 4.3244044091501725e-05, "loss": 1.4251, "step": 2086 }, { "epoch": 16.696, "grad_norm": 24.25136947631836, "learning_rate": 4.3238117814389004e-05, "loss": 1.4958, "step": 2087 }, { "epoch": 16.704, "grad_norm": 31.232587814331055, "learning_rate": 4.323219153727628e-05, "loss": 1.3588, "step": 2088 }, { "epoch": 16.712, "grad_norm": 67.05785369873047, "learning_rate": 4.322626526016357e-05, "loss": 1.2387, "step": 2089 }, { "epoch": 16.72, "grad_norm": 24.21996307373047, "learning_rate": 4.3220338983050854e-05, "loss": 1.467, "step": 2090 }, { "epoch": 16.728, "grad_norm": 22.962383270263672, "learning_rate": 4.321441270593813e-05, "loss": 1.6986, "step": 2091 }, { "epoch": 16.736, "grad_norm": 33.63120651245117, "learning_rate": 4.320848642882541e-05, "loss": 1.9645, "step": 2092 }, { "epoch": 16.744, "grad_norm": 21.244253158569336, "learning_rate": 4.32025601517127e-05, "loss": 1.3428, "step": 2093 }, { "epoch": 16.752, "grad_norm": 43.586002349853516, "learning_rate": 4.3196633874599976e-05, "loss": 1.5994, "step": 2094 }, { "epoch": 16.76, "grad_norm": 22.905046463012695, "learning_rate": 4.319070759748726e-05, "loss": 1.4375, "step": 2095 }, { "epoch": 16.768, "grad_norm": 29.215377807617188, "learning_rate": 4.318478132037455e-05, "loss": 1.447, "step": 2096 }, { "epoch": 16.776, "grad_norm": 58.67940139770508, "learning_rate": 4.3178855043261826e-05, "loss": 1.488, "step": 2097 }, { "epoch": 16.784, "grad_norm": 26.913434982299805, "learning_rate": 4.3172928766149105e-05, "loss": 1.7006, "step": 2098 }, { "epoch": 16.792, "grad_norm": 29.621801376342773, "learning_rate": 4.316700248903639e-05, "loss": 1.9878, "step": 2099 }, { "epoch": 16.8, "grad_norm": 108.44657897949219, "learning_rate": 4.3161076211923676e-05, "loss": 1.8011, "step": 2100 }, { "epoch": 16.808, "grad_norm": 38.61625289916992, "learning_rate": 4.315514993481095e-05, "loss": 1.2977, "step": 2101 }, { "epoch": 16.816, "grad_norm": 59.89971923828125, "learning_rate": 4.3149223657698234e-05, "loss": 1.9371, "step": 2102 }, { "epoch": 16.824, "grad_norm": 17.60259437561035, "learning_rate": 4.314329738058552e-05, "loss": 1.3487, "step": 2103 }, { "epoch": 16.832, "grad_norm": 37.802490234375, "learning_rate": 4.31373711034728e-05, "loss": 1.5878, "step": 2104 }, { "epoch": 16.84, "grad_norm": 63.58702087402344, "learning_rate": 4.3131444826360084e-05, "loss": 2.2391, "step": 2105 }, { "epoch": 16.848, "grad_norm": 51.60921096801758, "learning_rate": 4.312551854924736e-05, "loss": 1.5926, "step": 2106 }, { "epoch": 16.856, "grad_norm": 45.96404266357422, "learning_rate": 4.311959227213465e-05, "loss": 1.6054, "step": 2107 }, { "epoch": 16.864, "grad_norm": 36.73582077026367, "learning_rate": 4.311366599502193e-05, "loss": 1.6314, "step": 2108 }, { "epoch": 16.872, "grad_norm": 28.366830825805664, "learning_rate": 4.310773971790921e-05, "loss": 1.3494, "step": 2109 }, { "epoch": 16.88, "grad_norm": 28.050491333007812, "learning_rate": 4.31018134407965e-05, "loss": 1.8586, "step": 2110 }, { "epoch": 16.888, "grad_norm": 15.624736785888672, "learning_rate": 4.309588716368377e-05, "loss": 1.5525, "step": 2111 }, { "epoch": 16.896, "grad_norm": 68.32652282714844, "learning_rate": 4.3089960886571056e-05, "loss": 2.1112, "step": 2112 }, { "epoch": 16.904, "grad_norm": 24.55436897277832, "learning_rate": 4.308403460945834e-05, "loss": 1.4465, "step": 2113 }, { "epoch": 16.912, "grad_norm": 62.43718338012695, "learning_rate": 4.307810833234563e-05, "loss": 1.3372, "step": 2114 }, { "epoch": 16.92, "grad_norm": 21.9360294342041, "learning_rate": 4.3072182055232906e-05, "loss": 1.5905, "step": 2115 }, { "epoch": 16.928, "grad_norm": 28.10933494567871, "learning_rate": 4.3066255778120185e-05, "loss": 1.8852, "step": 2116 }, { "epoch": 16.936, "grad_norm": 35.167266845703125, "learning_rate": 4.306032950100747e-05, "loss": 1.2339, "step": 2117 }, { "epoch": 16.944, "grad_norm": 36.54167938232422, "learning_rate": 4.305440322389475e-05, "loss": 1.3686, "step": 2118 }, { "epoch": 16.951999999999998, "grad_norm": 122.52957153320312, "learning_rate": 4.3048476946782035e-05, "loss": 1.4837, "step": 2119 }, { "epoch": 16.96, "grad_norm": 34.1450080871582, "learning_rate": 4.304255066966932e-05, "loss": 1.3685, "step": 2120 }, { "epoch": 16.968, "grad_norm": 48.543521881103516, "learning_rate": 4.30366243925566e-05, "loss": 1.187, "step": 2121 }, { "epoch": 16.976, "grad_norm": 36.44997024536133, "learning_rate": 4.303069811544388e-05, "loss": 2.5366, "step": 2122 }, { "epoch": 16.984, "grad_norm": 16.675294876098633, "learning_rate": 4.3024771838331163e-05, "loss": 1.2331, "step": 2123 }, { "epoch": 16.992, "grad_norm": 20.951744079589844, "learning_rate": 4.301884556121845e-05, "loss": 1.5101, "step": 2124 }, { "epoch": 17.0, "grad_norm": 43.725677490234375, "learning_rate": 4.301291928410573e-05, "loss": 1.5395, "step": 2125 }, { "epoch": 17.0, "eval_loss": 1.4585342407226562, "eval_map": 0.2477, "eval_map_50": 0.5238, "eval_map_75": 0.2128, "eval_map_Coverall": 0.4945, "eval_map_Face_Shield": 0.2233, "eval_map_Gloves": 0.1363, "eval_map_Goggles": 0.0998, "eval_map_Mask": 0.2847, "eval_map_large": 0.3194, "eval_map_medium": 0.1492, "eval_map_small": 0.1331, "eval_mar_1": 0.2327, "eval_mar_10": 0.4145, "eval_mar_100": 0.4337, "eval_mar_100_Coverall": 0.6822, "eval_mar_100_Face_Shield": 0.4647, "eval_mar_100_Gloves": 0.341, "eval_mar_100_Goggles": 0.3094, "eval_mar_100_Mask": 0.3712, "eval_mar_large": 0.5253, "eval_mar_medium": 0.3201, "eval_mar_small": 0.188, "eval_runtime": 1.8067, "eval_samples_per_second": 16.051, "eval_steps_per_second": 1.107, "step": 2125 }, { "epoch": 17.008, "grad_norm": 24.800798416137695, "learning_rate": 4.300699300699301e-05, "loss": 2.1268, "step": 2126 }, { "epoch": 17.016, "grad_norm": 32.27647018432617, "learning_rate": 4.300106672988029e-05, "loss": 1.8479, "step": 2127 }, { "epoch": 17.024, "grad_norm": 18.148460388183594, "learning_rate": 4.299514045276757e-05, "loss": 1.3869, "step": 2128 }, { "epoch": 17.032, "grad_norm": 20.019872665405273, "learning_rate": 4.298921417565486e-05, "loss": 1.6077, "step": 2129 }, { "epoch": 17.04, "grad_norm": 46.52250671386719, "learning_rate": 4.298328789854214e-05, "loss": 2.6311, "step": 2130 }, { "epoch": 17.048, "grad_norm": 17.772350311279297, "learning_rate": 4.297736162142942e-05, "loss": 1.1354, "step": 2131 }, { "epoch": 17.056, "grad_norm": 69.54328918457031, "learning_rate": 4.29714353443167e-05, "loss": 1.3472, "step": 2132 }, { "epoch": 17.064, "grad_norm": 23.271875381469727, "learning_rate": 4.2965509067203986e-05, "loss": 1.5386, "step": 2133 }, { "epoch": 17.072, "grad_norm": 31.25193977355957, "learning_rate": 4.295958279009127e-05, "loss": 1.7126, "step": 2134 }, { "epoch": 17.08, "grad_norm": 41.56293869018555, "learning_rate": 4.295365651297854e-05, "loss": 1.6727, "step": 2135 }, { "epoch": 17.088, "grad_norm": 23.631532669067383, "learning_rate": 4.294773023586583e-05, "loss": 1.1058, "step": 2136 }, { "epoch": 17.096, "grad_norm": 26.303842544555664, "learning_rate": 4.2941803958753114e-05, "loss": 1.4092, "step": 2137 }, { "epoch": 17.104, "grad_norm": 34.76488494873047, "learning_rate": 4.293587768164039e-05, "loss": 1.4358, "step": 2138 }, { "epoch": 17.112, "grad_norm": 26.860685348510742, "learning_rate": 4.292995140452768e-05, "loss": 1.452, "step": 2139 }, { "epoch": 17.12, "grad_norm": 34.582550048828125, "learning_rate": 4.292402512741496e-05, "loss": 2.0768, "step": 2140 }, { "epoch": 17.128, "grad_norm": 49.8938102722168, "learning_rate": 4.291809885030224e-05, "loss": 1.3767, "step": 2141 }, { "epoch": 17.136, "grad_norm": 35.94752502441406, "learning_rate": 4.291217257318952e-05, "loss": 1.3282, "step": 2142 }, { "epoch": 17.144, "grad_norm": 20.1522274017334, "learning_rate": 4.290624629607681e-05, "loss": 1.6917, "step": 2143 }, { "epoch": 17.152, "grad_norm": 86.34185791015625, "learning_rate": 4.290032001896409e-05, "loss": 1.5667, "step": 2144 }, { "epoch": 17.16, "grad_norm": 24.39583396911621, "learning_rate": 4.2894393741851365e-05, "loss": 1.1124, "step": 2145 }, { "epoch": 17.168, "grad_norm": 35.881595611572266, "learning_rate": 4.288846746473865e-05, "loss": 1.8254, "step": 2146 }, { "epoch": 17.176, "grad_norm": 36.746376037597656, "learning_rate": 4.2882541187625936e-05, "loss": 1.7574, "step": 2147 }, { "epoch": 17.184, "grad_norm": 28.349903106689453, "learning_rate": 4.287661491051322e-05, "loss": 1.6372, "step": 2148 }, { "epoch": 17.192, "grad_norm": 26.735214233398438, "learning_rate": 4.28706886334005e-05, "loss": 1.7507, "step": 2149 }, { "epoch": 17.2, "grad_norm": 26.698808670043945, "learning_rate": 4.286476235628778e-05, "loss": 1.1861, "step": 2150 }, { "epoch": 17.208, "grad_norm": 47.8529167175293, "learning_rate": 4.2858836079175065e-05, "loss": 1.7569, "step": 2151 }, { "epoch": 17.216, "grad_norm": 18.256053924560547, "learning_rate": 4.2852909802062344e-05, "loss": 1.2554, "step": 2152 }, { "epoch": 17.224, "grad_norm": 27.037302017211914, "learning_rate": 4.284698352494963e-05, "loss": 1.1995, "step": 2153 }, { "epoch": 17.232, "grad_norm": 35.3917236328125, "learning_rate": 4.2841057247836915e-05, "loss": 1.4479, "step": 2154 }, { "epoch": 17.24, "grad_norm": 38.28911209106445, "learning_rate": 4.2835130970724194e-05, "loss": 1.3156, "step": 2155 }, { "epoch": 17.248, "grad_norm": 35.092994689941406, "learning_rate": 4.282920469361147e-05, "loss": 1.8353, "step": 2156 }, { "epoch": 17.256, "grad_norm": 22.307960510253906, "learning_rate": 4.282327841649876e-05, "loss": 1.8714, "step": 2157 }, { "epoch": 17.264, "grad_norm": 15.361300468444824, "learning_rate": 4.2817352139386044e-05, "loss": 1.6001, "step": 2158 }, { "epoch": 17.272, "grad_norm": 32.108055114746094, "learning_rate": 4.281142586227332e-05, "loss": 1.0988, "step": 2159 }, { "epoch": 17.28, "grad_norm": 21.050504684448242, "learning_rate": 4.28054995851606e-05, "loss": 1.3521, "step": 2160 }, { "epoch": 17.288, "grad_norm": 63.62972640991211, "learning_rate": 4.279957330804789e-05, "loss": 1.527, "step": 2161 }, { "epoch": 17.296, "grad_norm": 26.873817443847656, "learning_rate": 4.2793647030935166e-05, "loss": 1.3344, "step": 2162 }, { "epoch": 17.304, "grad_norm": 22.871347427368164, "learning_rate": 4.278772075382245e-05, "loss": 1.0957, "step": 2163 }, { "epoch": 17.312, "grad_norm": 28.77234649658203, "learning_rate": 4.278179447670974e-05, "loss": 1.4774, "step": 2164 }, { "epoch": 17.32, "grad_norm": 25.599071502685547, "learning_rate": 4.2775868199597016e-05, "loss": 1.3707, "step": 2165 }, { "epoch": 17.328, "grad_norm": 33.61066436767578, "learning_rate": 4.2769941922484295e-05, "loss": 1.8108, "step": 2166 }, { "epoch": 17.336, "grad_norm": 27.836200714111328, "learning_rate": 4.276401564537158e-05, "loss": 3.2126, "step": 2167 }, { "epoch": 17.344, "grad_norm": 50.12571716308594, "learning_rate": 4.2758089368258866e-05, "loss": 1.8094, "step": 2168 }, { "epoch": 17.352, "grad_norm": 46.128204345703125, "learning_rate": 4.2752163091146145e-05, "loss": 2.057, "step": 2169 }, { "epoch": 17.36, "grad_norm": 18.2650203704834, "learning_rate": 4.2746236814033424e-05, "loss": 1.4629, "step": 2170 }, { "epoch": 17.368, "grad_norm": 31.992319107055664, "learning_rate": 4.274031053692071e-05, "loss": 1.8214, "step": 2171 }, { "epoch": 17.376, "grad_norm": 35.6358757019043, "learning_rate": 4.273438425980799e-05, "loss": 1.6961, "step": 2172 }, { "epoch": 17.384, "grad_norm": 30.179851531982422, "learning_rate": 4.2728457982695274e-05, "loss": 1.275, "step": 2173 }, { "epoch": 17.392, "grad_norm": 25.02046012878418, "learning_rate": 4.272253170558255e-05, "loss": 1.4793, "step": 2174 }, { "epoch": 17.4, "grad_norm": 33.271095275878906, "learning_rate": 4.271660542846984e-05, "loss": 1.3515, "step": 2175 }, { "epoch": 17.408, "grad_norm": 47.06979751586914, "learning_rate": 4.271067915135712e-05, "loss": 1.5405, "step": 2176 }, { "epoch": 17.416, "grad_norm": 31.065031051635742, "learning_rate": 4.27047528742444e-05, "loss": 2.4462, "step": 2177 }, { "epoch": 17.424, "grad_norm": 44.494140625, "learning_rate": 4.269882659713169e-05, "loss": 1.1827, "step": 2178 }, { "epoch": 17.432, "grad_norm": 25.985979080200195, "learning_rate": 4.269290032001897e-05, "loss": 1.5722, "step": 2179 }, { "epoch": 17.44, "grad_norm": 34.57381820678711, "learning_rate": 4.2686974042906246e-05, "loss": 2.022, "step": 2180 }, { "epoch": 17.448, "grad_norm": 28.17375373840332, "learning_rate": 4.268104776579353e-05, "loss": 1.3268, "step": 2181 }, { "epoch": 17.456, "grad_norm": 84.20639038085938, "learning_rate": 4.267512148868082e-05, "loss": 1.3746, "step": 2182 }, { "epoch": 17.464, "grad_norm": 45.12877655029297, "learning_rate": 4.2669195211568096e-05, "loss": 1.0389, "step": 2183 }, { "epoch": 17.472, "grad_norm": 31.363317489624023, "learning_rate": 4.2663268934455375e-05, "loss": 1.1568, "step": 2184 }, { "epoch": 17.48, "grad_norm": 20.530426025390625, "learning_rate": 4.265734265734266e-05, "loss": 1.6632, "step": 2185 }, { "epoch": 17.488, "grad_norm": 41.09867858886719, "learning_rate": 4.265141638022994e-05, "loss": 1.5482, "step": 2186 }, { "epoch": 17.496, "grad_norm": 25.20719337463379, "learning_rate": 4.2645490103117225e-05, "loss": 1.2565, "step": 2187 }, { "epoch": 17.504, "grad_norm": 23.915748596191406, "learning_rate": 4.263956382600451e-05, "loss": 1.459, "step": 2188 }, { "epoch": 17.512, "grad_norm": 20.70015525817871, "learning_rate": 4.263363754889179e-05, "loss": 1.2162, "step": 2189 }, { "epoch": 17.52, "grad_norm": 39.568851470947266, "learning_rate": 4.262771127177907e-05, "loss": 1.4696, "step": 2190 }, { "epoch": 17.528, "grad_norm": 32.31692886352539, "learning_rate": 4.2621784994666354e-05, "loss": 1.4301, "step": 2191 }, { "epoch": 17.536, "grad_norm": 26.64870834350586, "learning_rate": 4.261585871755364e-05, "loss": 1.5475, "step": 2192 }, { "epoch": 17.544, "grad_norm": 21.020191192626953, "learning_rate": 4.260993244044092e-05, "loss": 1.3197, "step": 2193 }, { "epoch": 17.552, "grad_norm": 27.311527252197266, "learning_rate": 4.26040061633282e-05, "loss": 1.3672, "step": 2194 }, { "epoch": 17.56, "grad_norm": 17.765867233276367, "learning_rate": 4.259807988621548e-05, "loss": 1.2853, "step": 2195 }, { "epoch": 17.568, "grad_norm": 26.428682327270508, "learning_rate": 4.259215360910276e-05, "loss": 1.7383, "step": 2196 }, { "epoch": 17.576, "grad_norm": 54.87455368041992, "learning_rate": 4.258622733199005e-05, "loss": 1.3405, "step": 2197 }, { "epoch": 17.584, "grad_norm": 149.69012451171875, "learning_rate": 4.258030105487733e-05, "loss": 1.2216, "step": 2198 }, { "epoch": 17.592, "grad_norm": 44.89847946166992, "learning_rate": 4.257437477776461e-05, "loss": 1.5847, "step": 2199 }, { "epoch": 17.6, "grad_norm": 29.250003814697266, "learning_rate": 4.256844850065189e-05, "loss": 1.136, "step": 2200 }, { "epoch": 17.608, "grad_norm": 34.702945709228516, "learning_rate": 4.2562522223539176e-05, "loss": 1.8406, "step": 2201 }, { "epoch": 17.616, "grad_norm": 34.617759704589844, "learning_rate": 4.255659594642646e-05, "loss": 1.4364, "step": 2202 }, { "epoch": 17.624, "grad_norm": 22.63028907775879, "learning_rate": 4.255066966931374e-05, "loss": 2.7851, "step": 2203 }, { "epoch": 17.632, "grad_norm": 33.062713623046875, "learning_rate": 4.254474339220102e-05, "loss": 1.3198, "step": 2204 }, { "epoch": 17.64, "grad_norm": 21.7491512298584, "learning_rate": 4.2538817115088304e-05, "loss": 2.0047, "step": 2205 }, { "epoch": 17.648, "grad_norm": 179.44090270996094, "learning_rate": 4.253289083797558e-05, "loss": 1.758, "step": 2206 }, { "epoch": 17.656, "grad_norm": 23.145626068115234, "learning_rate": 4.252696456086287e-05, "loss": 1.5695, "step": 2207 }, { "epoch": 17.664, "grad_norm": 45.283729553222656, "learning_rate": 4.252103828375015e-05, "loss": 1.3996, "step": 2208 }, { "epoch": 17.672, "grad_norm": 22.51864242553711, "learning_rate": 4.251511200663743e-05, "loss": 1.3841, "step": 2209 }, { "epoch": 17.68, "grad_norm": 20.7056884765625, "learning_rate": 4.250918572952471e-05, "loss": 1.159, "step": 2210 }, { "epoch": 17.688, "grad_norm": 22.661544799804688, "learning_rate": 4.2503259452412e-05, "loss": 0.7475, "step": 2211 }, { "epoch": 17.696, "grad_norm": 23.39006996154785, "learning_rate": 4.249733317529928e-05, "loss": 1.4989, "step": 2212 }, { "epoch": 17.704, "grad_norm": 21.354496002197266, "learning_rate": 4.249140689818656e-05, "loss": 1.6897, "step": 2213 }, { "epoch": 17.712, "grad_norm": 19.816423416137695, "learning_rate": 4.248548062107384e-05, "loss": 1.059, "step": 2214 }, { "epoch": 17.72, "grad_norm": 23.708576202392578, "learning_rate": 4.2479554343961126e-05, "loss": 1.5622, "step": 2215 }, { "epoch": 17.728, "grad_norm": 18.82630729675293, "learning_rate": 4.247362806684841e-05, "loss": 1.489, "step": 2216 }, { "epoch": 17.736, "grad_norm": 26.255502700805664, "learning_rate": 4.246770178973569e-05, "loss": 1.2292, "step": 2217 }, { "epoch": 17.744, "grad_norm": 22.11045265197754, "learning_rate": 4.246177551262297e-05, "loss": 1.4447, "step": 2218 }, { "epoch": 17.752, "grad_norm": 33.2034912109375, "learning_rate": 4.2455849235510255e-05, "loss": 1.6129, "step": 2219 }, { "epoch": 17.76, "grad_norm": 35.469139099121094, "learning_rate": 4.2449922958397534e-05, "loss": 1.483, "step": 2220 }, { "epoch": 17.768, "grad_norm": 37.05902099609375, "learning_rate": 4.244399668128482e-05, "loss": 1.3012, "step": 2221 }, { "epoch": 17.776, "grad_norm": 101.9736328125, "learning_rate": 4.2438070404172105e-05, "loss": 1.9107, "step": 2222 }, { "epoch": 17.784, "grad_norm": 38.1668815612793, "learning_rate": 4.2432144127059384e-05, "loss": 1.4782, "step": 2223 }, { "epoch": 17.792, "grad_norm": 32.08005905151367, "learning_rate": 4.242621784994666e-05, "loss": 1.2373, "step": 2224 }, { "epoch": 17.8, "grad_norm": 29.378841400146484, "learning_rate": 4.242029157283395e-05, "loss": 1.6488, "step": 2225 }, { "epoch": 17.808, "grad_norm": 20.792621612548828, "learning_rate": 4.2414365295721234e-05, "loss": 1.5153, "step": 2226 }, { "epoch": 17.816, "grad_norm": 119.51431274414062, "learning_rate": 4.240843901860851e-05, "loss": 1.3833, "step": 2227 }, { "epoch": 17.824, "grad_norm": 19.343215942382812, "learning_rate": 4.240251274149579e-05, "loss": 1.5093, "step": 2228 }, { "epoch": 17.832, "grad_norm": 30.590299606323242, "learning_rate": 4.239658646438308e-05, "loss": 1.5342, "step": 2229 }, { "epoch": 17.84, "grad_norm": 23.162479400634766, "learning_rate": 4.2390660187270356e-05, "loss": 0.9841, "step": 2230 }, { "epoch": 17.848, "grad_norm": 23.133075714111328, "learning_rate": 4.238473391015764e-05, "loss": 1.5324, "step": 2231 }, { "epoch": 17.856, "grad_norm": 28.723407745361328, "learning_rate": 4.237880763304493e-05, "loss": 1.2149, "step": 2232 }, { "epoch": 17.864, "grad_norm": 22.823593139648438, "learning_rate": 4.2372881355932206e-05, "loss": 1.079, "step": 2233 }, { "epoch": 17.872, "grad_norm": 21.469806671142578, "learning_rate": 4.2366955078819485e-05, "loss": 1.4341, "step": 2234 }, { "epoch": 17.88, "grad_norm": 16.638103485107422, "learning_rate": 4.236102880170677e-05, "loss": 1.7522, "step": 2235 }, { "epoch": 17.888, "grad_norm": 28.4420108795166, "learning_rate": 4.2355102524594056e-05, "loss": 1.1459, "step": 2236 }, { "epoch": 17.896, "grad_norm": 26.55855941772461, "learning_rate": 4.2349176247481335e-05, "loss": 1.6949, "step": 2237 }, { "epoch": 17.904, "grad_norm": 33.7769660949707, "learning_rate": 4.2343249970368614e-05, "loss": 1.1347, "step": 2238 }, { "epoch": 17.912, "grad_norm": 21.274261474609375, "learning_rate": 4.23373236932559e-05, "loss": 1.4311, "step": 2239 }, { "epoch": 17.92, "grad_norm": 25.297895431518555, "learning_rate": 4.233139741614318e-05, "loss": 1.6303, "step": 2240 }, { "epoch": 17.928, "grad_norm": 20.427043914794922, "learning_rate": 4.2325471139030464e-05, "loss": 1.3501, "step": 2241 }, { "epoch": 17.936, "grad_norm": 21.42218780517578, "learning_rate": 4.231954486191774e-05, "loss": 1.502, "step": 2242 }, { "epoch": 17.944, "grad_norm": 29.085798263549805, "learning_rate": 4.231361858480503e-05, "loss": 1.6615, "step": 2243 }, { "epoch": 17.951999999999998, "grad_norm": 36.92374038696289, "learning_rate": 4.230769230769231e-05, "loss": 1.659, "step": 2244 }, { "epoch": 17.96, "grad_norm": 25.741756439208984, "learning_rate": 4.230176603057959e-05, "loss": 1.6601, "step": 2245 }, { "epoch": 17.968, "grad_norm": 38.836517333984375, "learning_rate": 4.229583975346688e-05, "loss": 2.1117, "step": 2246 }, { "epoch": 17.976, "grad_norm": 36.249813079833984, "learning_rate": 4.228991347635416e-05, "loss": 1.4542, "step": 2247 }, { "epoch": 17.984, "grad_norm": 30.212556838989258, "learning_rate": 4.2283987199241436e-05, "loss": 2.1147, "step": 2248 }, { "epoch": 17.992, "grad_norm": 55.206817626953125, "learning_rate": 4.227806092212872e-05, "loss": 1.6591, "step": 2249 }, { "epoch": 18.0, "grad_norm": 18.572681427001953, "learning_rate": 4.227213464501601e-05, "loss": 1.3943, "step": 2250 }, { "epoch": 18.0, "eval_loss": 1.4264775514602661, "eval_map": 0.2324, "eval_map_50": 0.5319, "eval_map_75": 0.1778, "eval_map_Coverall": 0.4628, "eval_map_Face_Shield": 0.217, "eval_map_Gloves": 0.1436, "eval_map_Goggles": 0.1047, "eval_map_Mask": 0.2337, "eval_map_large": 0.3093, "eval_map_medium": 0.186, "eval_map_small": 0.0917, "eval_mar_1": 0.2324, "eval_mar_10": 0.4182, "eval_mar_100": 0.4311, "eval_mar_100_Coverall": 0.68, "eval_mar_100_Face_Shield": 0.4471, "eval_mar_100_Gloves": 0.3443, "eval_mar_100_Goggles": 0.3094, "eval_mar_100_Mask": 0.375, "eval_mar_large": 0.5806, "eval_mar_medium": 0.357, "eval_mar_small": 0.1366, "eval_runtime": 1.823, "eval_samples_per_second": 15.908, "eval_steps_per_second": 1.097, "step": 2250 }, { "epoch": 18.008, "grad_norm": 29.7591609954834, "learning_rate": 4.2266208367903286e-05, "loss": 1.2338, "step": 2251 }, { "epoch": 18.016, "grad_norm": 35.370182037353516, "learning_rate": 4.2260282090790565e-05, "loss": 1.993, "step": 2252 }, { "epoch": 18.024, "grad_norm": 18.3090877532959, "learning_rate": 4.225435581367785e-05, "loss": 1.1742, "step": 2253 }, { "epoch": 18.032, "grad_norm": 26.33888053894043, "learning_rate": 4.224842953656513e-05, "loss": 1.3852, "step": 2254 }, { "epoch": 18.04, "grad_norm": 32.372806549072266, "learning_rate": 4.2242503259452415e-05, "loss": 1.3283, "step": 2255 }, { "epoch": 18.048, "grad_norm": 21.045835494995117, "learning_rate": 4.22365769823397e-05, "loss": 1.5677, "step": 2256 }, { "epoch": 18.056, "grad_norm": 40.07466125488281, "learning_rate": 4.223065070522698e-05, "loss": 1.9533, "step": 2257 }, { "epoch": 18.064, "grad_norm": 28.336593627929688, "learning_rate": 4.222472442811426e-05, "loss": 1.853, "step": 2258 }, { "epoch": 18.072, "grad_norm": 287.7193603515625, "learning_rate": 4.2218798151001544e-05, "loss": 1.3297, "step": 2259 }, { "epoch": 18.08, "grad_norm": 64.4892349243164, "learning_rate": 4.221287187388883e-05, "loss": 1.9383, "step": 2260 }, { "epoch": 18.088, "grad_norm": 25.78652572631836, "learning_rate": 4.220694559677611e-05, "loss": 1.0613, "step": 2261 }, { "epoch": 18.096, "grad_norm": 28.72054100036621, "learning_rate": 4.220101931966339e-05, "loss": 1.3316, "step": 2262 }, { "epoch": 18.104, "grad_norm": 22.535398483276367, "learning_rate": 4.219509304255067e-05, "loss": 1.179, "step": 2263 }, { "epoch": 18.112, "grad_norm": 23.270919799804688, "learning_rate": 4.218916676543795e-05, "loss": 1.3355, "step": 2264 }, { "epoch": 18.12, "grad_norm": 41.000572204589844, "learning_rate": 4.218324048832524e-05, "loss": 1.9138, "step": 2265 }, { "epoch": 18.128, "grad_norm": 47.91848373413086, "learning_rate": 4.217731421121252e-05, "loss": 1.6281, "step": 2266 }, { "epoch": 18.136, "grad_norm": 31.48707389831543, "learning_rate": 4.21713879340998e-05, "loss": 1.3411, "step": 2267 }, { "epoch": 18.144, "grad_norm": 18.308921813964844, "learning_rate": 4.216546165698708e-05, "loss": 1.2699, "step": 2268 }, { "epoch": 18.152, "grad_norm": 60.5955810546875, "learning_rate": 4.2159535379874366e-05, "loss": 1.0817, "step": 2269 }, { "epoch": 18.16, "grad_norm": 62.27886199951172, "learning_rate": 4.215360910276165e-05, "loss": 1.4952, "step": 2270 }, { "epoch": 18.168, "grad_norm": 73.51675415039062, "learning_rate": 4.214768282564893e-05, "loss": 1.6058, "step": 2271 }, { "epoch": 18.176, "grad_norm": 35.70814895629883, "learning_rate": 4.214175654853621e-05, "loss": 1.5341, "step": 2272 }, { "epoch": 18.184, "grad_norm": 25.938404083251953, "learning_rate": 4.2135830271423495e-05, "loss": 1.6918, "step": 2273 }, { "epoch": 18.192, "grad_norm": 47.669620513916016, "learning_rate": 4.212990399431077e-05, "loss": 1.3035, "step": 2274 }, { "epoch": 18.2, "grad_norm": 33.84437942504883, "learning_rate": 4.212397771719806e-05, "loss": 1.7588, "step": 2275 }, { "epoch": 18.208, "grad_norm": 20.835573196411133, "learning_rate": 4.211805144008534e-05, "loss": 1.3084, "step": 2276 }, { "epoch": 18.216, "grad_norm": 56.863250732421875, "learning_rate": 4.211212516297262e-05, "loss": 2.1238, "step": 2277 }, { "epoch": 18.224, "grad_norm": 18.845977783203125, "learning_rate": 4.21061988858599e-05, "loss": 1.3431, "step": 2278 }, { "epoch": 18.232, "grad_norm": 58.81623458862305, "learning_rate": 4.210027260874719e-05, "loss": 2.2531, "step": 2279 }, { "epoch": 18.24, "grad_norm": 24.608318328857422, "learning_rate": 4.209434633163447e-05, "loss": 1.6009, "step": 2280 }, { "epoch": 18.248, "grad_norm": 26.6842098236084, "learning_rate": 4.208842005452175e-05, "loss": 1.4626, "step": 2281 }, { "epoch": 18.256, "grad_norm": 38.08470916748047, "learning_rate": 4.208249377740903e-05, "loss": 1.2817, "step": 2282 }, { "epoch": 18.264, "grad_norm": 28.230457305908203, "learning_rate": 4.2076567500296317e-05, "loss": 1.4562, "step": 2283 }, { "epoch": 18.272, "grad_norm": 72.37814331054688, "learning_rate": 4.20706412231836e-05, "loss": 1.4313, "step": 2284 }, { "epoch": 18.28, "grad_norm": 38.10063171386719, "learning_rate": 4.206471494607088e-05, "loss": 1.2754, "step": 2285 }, { "epoch": 18.288, "grad_norm": 65.21988677978516, "learning_rate": 4.205878866895816e-05, "loss": 1.3432, "step": 2286 }, { "epoch": 18.296, "grad_norm": 42.43834686279297, "learning_rate": 4.2052862391845445e-05, "loss": 1.6549, "step": 2287 }, { "epoch": 18.304, "grad_norm": 18.6822509765625, "learning_rate": 4.2046936114732724e-05, "loss": 1.4767, "step": 2288 }, { "epoch": 18.312, "grad_norm": 27.674184799194336, "learning_rate": 4.204100983762001e-05, "loss": 1.2014, "step": 2289 }, { "epoch": 18.32, "grad_norm": 28.989704132080078, "learning_rate": 4.2035083560507295e-05, "loss": 1.6161, "step": 2290 }, { "epoch": 18.328, "grad_norm": 47.07798767089844, "learning_rate": 4.2029157283394574e-05, "loss": 1.5238, "step": 2291 }, { "epoch": 18.336, "grad_norm": 30.77895164489746, "learning_rate": 4.202323100628185e-05, "loss": 1.3217, "step": 2292 }, { "epoch": 18.344, "grad_norm": 73.539306640625, "learning_rate": 4.201730472916914e-05, "loss": 1.8631, "step": 2293 }, { "epoch": 18.352, "grad_norm": 106.10955047607422, "learning_rate": 4.2011378452056424e-05, "loss": 1.8327, "step": 2294 }, { "epoch": 18.36, "grad_norm": 29.24458122253418, "learning_rate": 4.20054521749437e-05, "loss": 1.7599, "step": 2295 }, { "epoch": 18.368, "grad_norm": 19.6904354095459, "learning_rate": 4.199952589783098e-05, "loss": 1.4492, "step": 2296 }, { "epoch": 18.376, "grad_norm": 108.13353729248047, "learning_rate": 4.199359962071827e-05, "loss": 1.4329, "step": 2297 }, { "epoch": 18.384, "grad_norm": 44.720943450927734, "learning_rate": 4.1987673343605546e-05, "loss": 1.4055, "step": 2298 }, { "epoch": 18.392, "grad_norm": 32.38014221191406, "learning_rate": 4.198174706649283e-05, "loss": 1.2841, "step": 2299 }, { "epoch": 18.4, "grad_norm": 33.7532958984375, "learning_rate": 4.197582078938012e-05, "loss": 1.4511, "step": 2300 }, { "epoch": 18.408, "grad_norm": 30.203859329223633, "learning_rate": 4.1969894512267396e-05, "loss": 1.813, "step": 2301 }, { "epoch": 18.416, "grad_norm": 64.84244537353516, "learning_rate": 4.1963968235154675e-05, "loss": 1.7627, "step": 2302 }, { "epoch": 18.424, "grad_norm": 29.037250518798828, "learning_rate": 4.195804195804196e-05, "loss": 1.8514, "step": 2303 }, { "epoch": 18.432, "grad_norm": 26.803380966186523, "learning_rate": 4.1952115680929246e-05, "loss": 1.3131, "step": 2304 }, { "epoch": 18.44, "grad_norm": 29.22881507873535, "learning_rate": 4.1946189403816525e-05, "loss": 1.2785, "step": 2305 }, { "epoch": 18.448, "grad_norm": 32.53632354736328, "learning_rate": 4.1940263126703804e-05, "loss": 1.3928, "step": 2306 }, { "epoch": 18.456, "grad_norm": 27.74120330810547, "learning_rate": 4.193433684959109e-05, "loss": 1.1362, "step": 2307 }, { "epoch": 18.464, "grad_norm": 117.2841796875, "learning_rate": 4.192841057247837e-05, "loss": 1.9922, "step": 2308 }, { "epoch": 18.472, "grad_norm": 22.803783416748047, "learning_rate": 4.1922484295365654e-05, "loss": 1.3416, "step": 2309 }, { "epoch": 18.48, "grad_norm": 22.866588592529297, "learning_rate": 4.191655801825294e-05, "loss": 1.0231, "step": 2310 }, { "epoch": 18.488, "grad_norm": 20.233327865600586, "learning_rate": 4.191063174114022e-05, "loss": 1.3212, "step": 2311 }, { "epoch": 18.496, "grad_norm": 28.389318466186523, "learning_rate": 4.19047054640275e-05, "loss": 1.3376, "step": 2312 }, { "epoch": 18.504, "grad_norm": 25.754606246948242, "learning_rate": 4.189877918691478e-05, "loss": 1.9717, "step": 2313 }, { "epoch": 18.512, "grad_norm": 55.276424407958984, "learning_rate": 4.189285290980207e-05, "loss": 1.1378, "step": 2314 }, { "epoch": 18.52, "grad_norm": 32.72603225708008, "learning_rate": 4.188692663268935e-05, "loss": 1.7784, "step": 2315 }, { "epoch": 18.528, "grad_norm": 17.14956283569336, "learning_rate": 4.1881000355576626e-05, "loss": 1.4115, "step": 2316 }, { "epoch": 18.536, "grad_norm": 40.968692779541016, "learning_rate": 4.187507407846391e-05, "loss": 1.3337, "step": 2317 }, { "epoch": 18.544, "grad_norm": 75.3390884399414, "learning_rate": 4.18691478013512e-05, "loss": 1.4447, "step": 2318 }, { "epoch": 18.552, "grad_norm": 21.629457473754883, "learning_rate": 4.1863221524238476e-05, "loss": 1.2882, "step": 2319 }, { "epoch": 18.56, "grad_norm": 27.835100173950195, "learning_rate": 4.1857295247125755e-05, "loss": 1.692, "step": 2320 }, { "epoch": 18.568, "grad_norm": 26.79738426208496, "learning_rate": 4.185136897001304e-05, "loss": 1.4905, "step": 2321 }, { "epoch": 18.576, "grad_norm": 28.10477638244629, "learning_rate": 4.184544269290032e-05, "loss": 1.3505, "step": 2322 }, { "epoch": 18.584, "grad_norm": 35.897789001464844, "learning_rate": 4.1839516415787605e-05, "loss": 2.5511, "step": 2323 }, { "epoch": 18.592, "grad_norm": 403.2672119140625, "learning_rate": 4.183359013867489e-05, "loss": 1.8851, "step": 2324 }, { "epoch": 18.6, "grad_norm": 39.01925277709961, "learning_rate": 4.182766386156217e-05, "loss": 1.2867, "step": 2325 }, { "epoch": 18.608, "grad_norm": 20.21622657775879, "learning_rate": 4.182173758444945e-05, "loss": 1.24, "step": 2326 }, { "epoch": 18.616, "grad_norm": 90.11030578613281, "learning_rate": 4.1815811307336734e-05, "loss": 1.4816, "step": 2327 }, { "epoch": 18.624, "grad_norm": 23.220338821411133, "learning_rate": 4.180988503022402e-05, "loss": 2.7785, "step": 2328 }, { "epoch": 18.632, "grad_norm": 456.8704833984375, "learning_rate": 4.18039587531113e-05, "loss": 1.2821, "step": 2329 }, { "epoch": 18.64, "grad_norm": 32.32100296020508, "learning_rate": 4.179803247599858e-05, "loss": 1.9162, "step": 2330 }, { "epoch": 18.648, "grad_norm": 33.59547805786133, "learning_rate": 4.179210619888586e-05, "loss": 1.6159, "step": 2331 }, { "epoch": 18.656, "grad_norm": 44.37907791137695, "learning_rate": 4.178617992177314e-05, "loss": 1.4894, "step": 2332 }, { "epoch": 18.664, "grad_norm": 45.21086502075195, "learning_rate": 4.178025364466043e-05, "loss": 1.5845, "step": 2333 }, { "epoch": 18.672, "grad_norm": 55.25777816772461, "learning_rate": 4.177432736754771e-05, "loss": 1.4274, "step": 2334 }, { "epoch": 18.68, "grad_norm": 56.7229118347168, "learning_rate": 4.176840109043499e-05, "loss": 2.055, "step": 2335 }, { "epoch": 18.688, "grad_norm": 822.040283203125, "learning_rate": 4.176247481332227e-05, "loss": 1.4347, "step": 2336 }, { "epoch": 18.696, "grad_norm": 24.89990997314453, "learning_rate": 4.1756548536209556e-05, "loss": 1.6016, "step": 2337 }, { "epoch": 18.704, "grad_norm": 27.63473129272461, "learning_rate": 4.175062225909684e-05, "loss": 1.0692, "step": 2338 }, { "epoch": 18.712, "grad_norm": 100.5373306274414, "learning_rate": 4.174469598198412e-05, "loss": 1.6912, "step": 2339 }, { "epoch": 18.72, "grad_norm": 56.686973571777344, "learning_rate": 4.17387697048714e-05, "loss": 1.3153, "step": 2340 }, { "epoch": 18.728, "grad_norm": 53.21240234375, "learning_rate": 4.1732843427758685e-05, "loss": 1.2104, "step": 2341 }, { "epoch": 18.736, "grad_norm": 21.637229919433594, "learning_rate": 4.1726917150645963e-05, "loss": 1.2609, "step": 2342 }, { "epoch": 18.744, "grad_norm": 25.597444534301758, "learning_rate": 4.172099087353325e-05, "loss": 1.5552, "step": 2343 }, { "epoch": 18.752, "grad_norm": 27.257064819335938, "learning_rate": 4.1715064596420535e-05, "loss": 1.8854, "step": 2344 }, { "epoch": 18.76, "grad_norm": 24.20531463623047, "learning_rate": 4.1709138319307813e-05, "loss": 1.2033, "step": 2345 }, { "epoch": 18.768, "grad_norm": 25.600915908813477, "learning_rate": 4.170321204219509e-05, "loss": 1.676, "step": 2346 }, { "epoch": 18.776, "grad_norm": 42.51124572753906, "learning_rate": 4.169728576508238e-05, "loss": 1.7487, "step": 2347 }, { "epoch": 18.784, "grad_norm": 36.96995544433594, "learning_rate": 4.1691359487969663e-05, "loss": 1.2116, "step": 2348 }, { "epoch": 18.792, "grad_norm": 27.649063110351562, "learning_rate": 4.168543321085694e-05, "loss": 1.5294, "step": 2349 }, { "epoch": 18.8, "grad_norm": 32.91782760620117, "learning_rate": 4.167950693374422e-05, "loss": 1.3539, "step": 2350 }, { "epoch": 18.808, "grad_norm": 36.595436096191406, "learning_rate": 4.167358065663151e-05, "loss": 1.8516, "step": 2351 }, { "epoch": 18.816, "grad_norm": 44.14629364013672, "learning_rate": 4.166765437951879e-05, "loss": 1.2119, "step": 2352 }, { "epoch": 18.824, "grad_norm": 24.613065719604492, "learning_rate": 4.166172810240607e-05, "loss": 1.1295, "step": 2353 }, { "epoch": 18.832, "grad_norm": 23.426549911499023, "learning_rate": 4.165580182529335e-05, "loss": 1.1736, "step": 2354 }, { "epoch": 18.84, "grad_norm": 24.659194946289062, "learning_rate": 4.1649875548180635e-05, "loss": 1.6608, "step": 2355 }, { "epoch": 18.848, "grad_norm": 69.96940612792969, "learning_rate": 4.1643949271067914e-05, "loss": 1.4728, "step": 2356 }, { "epoch": 18.856, "grad_norm": 53.91593551635742, "learning_rate": 4.16380229939552e-05, "loss": 1.8428, "step": 2357 }, { "epoch": 18.864, "grad_norm": 72.65718078613281, "learning_rate": 4.1632096716842485e-05, "loss": 1.5611, "step": 2358 }, { "epoch": 18.872, "grad_norm": 35.62009811401367, "learning_rate": 4.1626170439729764e-05, "loss": 1.3276, "step": 2359 }, { "epoch": 18.88, "grad_norm": 108.64946746826172, "learning_rate": 4.162024416261704e-05, "loss": 1.7391, "step": 2360 }, { "epoch": 18.888, "grad_norm": 37.112037658691406, "learning_rate": 4.161431788550433e-05, "loss": 1.2376, "step": 2361 }, { "epoch": 18.896, "grad_norm": 28.991348266601562, "learning_rate": 4.1608391608391614e-05, "loss": 1.1166, "step": 2362 }, { "epoch": 18.904, "grad_norm": 28.606252670288086, "learning_rate": 4.160246533127889e-05, "loss": 1.1754, "step": 2363 }, { "epoch": 18.912, "grad_norm": 40.32387161254883, "learning_rate": 4.159653905416617e-05, "loss": 1.7585, "step": 2364 }, { "epoch": 18.92, "grad_norm": 21.50009536743164, "learning_rate": 4.159061277705346e-05, "loss": 1.6698, "step": 2365 }, { "epoch": 18.928, "grad_norm": 26.60630226135254, "learning_rate": 4.1584686499940736e-05, "loss": 1.2878, "step": 2366 }, { "epoch": 18.936, "grad_norm": 62.0495491027832, "learning_rate": 4.157876022282802e-05, "loss": 1.2173, "step": 2367 }, { "epoch": 18.944, "grad_norm": 41.00098419189453, "learning_rate": 4.157283394571531e-05, "loss": 1.6763, "step": 2368 }, { "epoch": 18.951999999999998, "grad_norm": 52.15534210205078, "learning_rate": 4.1566907668602586e-05, "loss": 1.438, "step": 2369 }, { "epoch": 18.96, "grad_norm": 94.5442886352539, "learning_rate": 4.1560981391489865e-05, "loss": 3.1178, "step": 2370 }, { "epoch": 18.968, "grad_norm": 22.09161376953125, "learning_rate": 4.155505511437715e-05, "loss": 1.5082, "step": 2371 }, { "epoch": 18.976, "grad_norm": 99.81302642822266, "learning_rate": 4.1549128837264436e-05, "loss": 1.0849, "step": 2372 }, { "epoch": 18.984, "grad_norm": 49.11785125732422, "learning_rate": 4.1543202560151715e-05, "loss": 1.4058, "step": 2373 }, { "epoch": 18.992, "grad_norm": 42.27607345581055, "learning_rate": 4.1537276283038994e-05, "loss": 1.7476, "step": 2374 }, { "epoch": 19.0, "grad_norm": 38.434059143066406, "learning_rate": 4.153135000592628e-05, "loss": 1.9512, "step": 2375 }, { "epoch": 19.0, "eval_loss": 1.7378926277160645, "eval_map": 0.1969, "eval_map_50": 0.4308, "eval_map_75": 0.1441, "eval_map_Coverall": 0.4093, "eval_map_Face_Shield": 0.2063, "eval_map_Gloves": 0.0744, "eval_map_Goggles": 0.085, "eval_map_Mask": 0.2099, "eval_map_large": 0.2638, "eval_map_medium": 0.1326, "eval_map_small": 0.0832, "eval_mar_1": 0.1908, "eval_mar_10": 0.3456, "eval_mar_100": 0.3592, "eval_mar_100_Coverall": 0.6511, "eval_mar_100_Face_Shield": 0.4, "eval_mar_100_Gloves": 0.2016, "eval_mar_100_Goggles": 0.2375, "eval_mar_100_Mask": 0.3058, "eval_mar_large": 0.5084, "eval_mar_medium": 0.2491, "eval_mar_small": 0.0851, "eval_runtime": 1.8281, "eval_samples_per_second": 15.864, "eval_steps_per_second": 1.094, "step": 2375 }, { "epoch": 19.008, "grad_norm": 20.413206100463867, "learning_rate": 4.152542372881356e-05, "loss": 1.1703, "step": 2376 }, { "epoch": 19.016, "grad_norm": 14.959589004516602, "learning_rate": 4.1519497451700844e-05, "loss": 1.2502, "step": 2377 }, { "epoch": 19.024, "grad_norm": 31.610586166381836, "learning_rate": 4.151357117458813e-05, "loss": 1.4157, "step": 2378 }, { "epoch": 19.032, "grad_norm": 71.79167175292969, "learning_rate": 4.150764489747541e-05, "loss": 1.8852, "step": 2379 }, { "epoch": 19.04, "grad_norm": 40.21986389160156, "learning_rate": 4.150171862036269e-05, "loss": 1.0954, "step": 2380 }, { "epoch": 19.048, "grad_norm": 32.302520751953125, "learning_rate": 4.149579234324997e-05, "loss": 1.4989, "step": 2381 }, { "epoch": 19.056, "grad_norm": 18.78073501586914, "learning_rate": 4.148986606613726e-05, "loss": 1.257, "step": 2382 }, { "epoch": 19.064, "grad_norm": 162.94412231445312, "learning_rate": 4.148393978902454e-05, "loss": 1.9615, "step": 2383 }, { "epoch": 19.072, "grad_norm": 23.00236701965332, "learning_rate": 4.1478013511911816e-05, "loss": 1.6851, "step": 2384 }, { "epoch": 19.08, "grad_norm": 19.30396270751953, "learning_rate": 4.14720872347991e-05, "loss": 1.3047, "step": 2385 }, { "epoch": 19.088, "grad_norm": 18.477514266967773, "learning_rate": 4.146616095768639e-05, "loss": 1.3192, "step": 2386 }, { "epoch": 19.096, "grad_norm": 28.03845977783203, "learning_rate": 4.1460234680573666e-05, "loss": 1.7812, "step": 2387 }, { "epoch": 19.104, "grad_norm": 24.376686096191406, "learning_rate": 4.1454308403460945e-05, "loss": 1.5238, "step": 2388 }, { "epoch": 19.112, "grad_norm": 34.32772445678711, "learning_rate": 4.144838212634823e-05, "loss": 1.561, "step": 2389 }, { "epoch": 19.12, "grad_norm": 18.859390258789062, "learning_rate": 4.144245584923551e-05, "loss": 1.4274, "step": 2390 }, { "epoch": 19.128, "grad_norm": 22.810129165649414, "learning_rate": 4.1436529572122795e-05, "loss": 1.8813, "step": 2391 }, { "epoch": 19.136, "grad_norm": 53.66380310058594, "learning_rate": 4.143060329501008e-05, "loss": 1.7235, "step": 2392 }, { "epoch": 19.144, "grad_norm": 30.666780471801758, "learning_rate": 4.142467701789736e-05, "loss": 1.5917, "step": 2393 }, { "epoch": 19.152, "grad_norm": 28.496728897094727, "learning_rate": 4.141875074078464e-05, "loss": 1.1216, "step": 2394 }, { "epoch": 19.16, "grad_norm": 32.05178451538086, "learning_rate": 4.1412824463671924e-05, "loss": 1.0794, "step": 2395 }, { "epoch": 19.168, "grad_norm": 233.16261291503906, "learning_rate": 4.140689818655921e-05, "loss": 1.4138, "step": 2396 }, { "epoch": 19.176, "grad_norm": 27.06987953186035, "learning_rate": 4.140097190944649e-05, "loss": 1.4512, "step": 2397 }, { "epoch": 19.184, "grad_norm": 45.44176483154297, "learning_rate": 4.139504563233377e-05, "loss": 1.3956, "step": 2398 }, { "epoch": 19.192, "grad_norm": 77.31614685058594, "learning_rate": 4.138911935522105e-05, "loss": 1.2648, "step": 2399 }, { "epoch": 19.2, "grad_norm": 20.078060150146484, "learning_rate": 4.138319307810833e-05, "loss": 1.5783, "step": 2400 }, { "epoch": 19.208, "grad_norm": 31.034931182861328, "learning_rate": 4.137726680099562e-05, "loss": 1.1872, "step": 2401 }, { "epoch": 19.216, "grad_norm": 31.561941146850586, "learning_rate": 4.13713405238829e-05, "loss": 1.2481, "step": 2402 }, { "epoch": 19.224, "grad_norm": 22.252920150756836, "learning_rate": 4.136541424677018e-05, "loss": 1.6919, "step": 2403 }, { "epoch": 19.232, "grad_norm": 29.850862503051758, "learning_rate": 4.135948796965746e-05, "loss": 1.7471, "step": 2404 }, { "epoch": 19.24, "grad_norm": 15.163207054138184, "learning_rate": 4.1353561692544746e-05, "loss": 1.6319, "step": 2405 }, { "epoch": 19.248, "grad_norm": 15.996826171875, "learning_rate": 4.134763541543203e-05, "loss": 1.1157, "step": 2406 }, { "epoch": 19.256, "grad_norm": 134.99752807617188, "learning_rate": 4.134170913831931e-05, "loss": 1.1763, "step": 2407 }, { "epoch": 19.264, "grad_norm": 48.25074768066406, "learning_rate": 4.133578286120659e-05, "loss": 1.8295, "step": 2408 }, { "epoch": 19.272, "grad_norm": 28.445844650268555, "learning_rate": 4.1329856584093875e-05, "loss": 1.2282, "step": 2409 }, { "epoch": 19.28, "grad_norm": 56.97773742675781, "learning_rate": 4.1323930306981153e-05, "loss": 2.0525, "step": 2410 }, { "epoch": 19.288, "grad_norm": 25.82994270324707, "learning_rate": 4.131800402986844e-05, "loss": 1.3626, "step": 2411 }, { "epoch": 19.296, "grad_norm": 22.36441993713379, "learning_rate": 4.1312077752755725e-05, "loss": 1.6377, "step": 2412 }, { "epoch": 19.304, "grad_norm": 26.230863571166992, "learning_rate": 4.1306151475643003e-05, "loss": 1.332, "step": 2413 }, { "epoch": 19.312, "grad_norm": 28.947345733642578, "learning_rate": 4.130022519853028e-05, "loss": 2.1677, "step": 2414 }, { "epoch": 19.32, "grad_norm": 19.18657684326172, "learning_rate": 4.129429892141757e-05, "loss": 1.3127, "step": 2415 }, { "epoch": 19.328, "grad_norm": 40.12021255493164, "learning_rate": 4.1288372644304853e-05, "loss": 1.1454, "step": 2416 }, { "epoch": 19.336, "grad_norm": 50.44798278808594, "learning_rate": 4.128244636719213e-05, "loss": 1.5424, "step": 2417 }, { "epoch": 19.344, "grad_norm": 39.80916976928711, "learning_rate": 4.127652009007941e-05, "loss": 1.5332, "step": 2418 }, { "epoch": 19.352, "grad_norm": 72.71581268310547, "learning_rate": 4.12705938129667e-05, "loss": 1.2853, "step": 2419 }, { "epoch": 19.36, "grad_norm": 74.18849182128906, "learning_rate": 4.126466753585398e-05, "loss": 1.2563, "step": 2420 }, { "epoch": 19.368, "grad_norm": 22.449613571166992, "learning_rate": 4.125874125874126e-05, "loss": 1.3229, "step": 2421 }, { "epoch": 19.376, "grad_norm": 30.680389404296875, "learning_rate": 4.125281498162854e-05, "loss": 1.3143, "step": 2422 }, { "epoch": 19.384, "grad_norm": 42.63113021850586, "learning_rate": 4.1246888704515826e-05, "loss": 1.3333, "step": 2423 }, { "epoch": 19.392, "grad_norm": 52.116539001464844, "learning_rate": 4.1240962427403104e-05, "loss": 1.4607, "step": 2424 }, { "epoch": 19.4, "grad_norm": 30.43817138671875, "learning_rate": 4.123503615029039e-05, "loss": 1.2604, "step": 2425 }, { "epoch": 19.408, "grad_norm": 33.04831314086914, "learning_rate": 4.1229109873177676e-05, "loss": 1.4485, "step": 2426 }, { "epoch": 19.416, "grad_norm": 41.308414459228516, "learning_rate": 4.1223183596064954e-05, "loss": 1.3255, "step": 2427 }, { "epoch": 19.424, "grad_norm": 23.073318481445312, "learning_rate": 4.121725731895223e-05, "loss": 2.9866, "step": 2428 }, { "epoch": 19.432, "grad_norm": 23.064542770385742, "learning_rate": 4.121133104183952e-05, "loss": 1.1759, "step": 2429 }, { "epoch": 19.44, "grad_norm": 44.07405090332031, "learning_rate": 4.1205404764726804e-05, "loss": 1.0846, "step": 2430 }, { "epoch": 19.448, "grad_norm": 24.466466903686523, "learning_rate": 4.119947848761408e-05, "loss": 1.3708, "step": 2431 }, { "epoch": 19.456, "grad_norm": 70.37220764160156, "learning_rate": 4.119355221050136e-05, "loss": 1.488, "step": 2432 }, { "epoch": 19.464, "grad_norm": 29.74408531188965, "learning_rate": 4.118762593338865e-05, "loss": 1.1974, "step": 2433 }, { "epoch": 19.472, "grad_norm": 20.68267250061035, "learning_rate": 4.1181699656275926e-05, "loss": 1.3351, "step": 2434 }, { "epoch": 19.48, "grad_norm": 20.61431884765625, "learning_rate": 4.117577337916321e-05, "loss": 1.5069, "step": 2435 }, { "epoch": 19.488, "grad_norm": 17.484193801879883, "learning_rate": 4.11698471020505e-05, "loss": 1.1898, "step": 2436 }, { "epoch": 19.496, "grad_norm": 31.91611099243164, "learning_rate": 4.1163920824937776e-05, "loss": 1.5779, "step": 2437 }, { "epoch": 19.504, "grad_norm": 29.957456588745117, "learning_rate": 4.1157994547825055e-05, "loss": 1.332, "step": 2438 }, { "epoch": 19.512, "grad_norm": 74.07955169677734, "learning_rate": 4.115206827071234e-05, "loss": 1.8472, "step": 2439 }, { "epoch": 19.52, "grad_norm": 21.66368293762207, "learning_rate": 4.1146141993599626e-05, "loss": 1.0225, "step": 2440 }, { "epoch": 19.528, "grad_norm": 23.40432357788086, "learning_rate": 4.1140215716486905e-05, "loss": 1.6137, "step": 2441 }, { "epoch": 19.536, "grad_norm": 25.293819427490234, "learning_rate": 4.1134289439374184e-05, "loss": 1.2486, "step": 2442 }, { "epoch": 19.544, "grad_norm": 30.418746948242188, "learning_rate": 4.112836316226147e-05, "loss": 1.4771, "step": 2443 }, { "epoch": 19.552, "grad_norm": 22.838424682617188, "learning_rate": 4.112243688514875e-05, "loss": 1.1027, "step": 2444 }, { "epoch": 19.56, "grad_norm": 32.66651153564453, "learning_rate": 4.1116510608036034e-05, "loss": 1.1436, "step": 2445 }, { "epoch": 19.568, "grad_norm": 22.930906295776367, "learning_rate": 4.111058433092332e-05, "loss": 1.389, "step": 2446 }, { "epoch": 19.576, "grad_norm": 15.102883338928223, "learning_rate": 4.11046580538106e-05, "loss": 1.5818, "step": 2447 }, { "epoch": 19.584, "grad_norm": 20.236783981323242, "learning_rate": 4.109873177669788e-05, "loss": 2.5978, "step": 2448 }, { "epoch": 19.592, "grad_norm": 20.291601181030273, "learning_rate": 4.109280549958516e-05, "loss": 1.0624, "step": 2449 }, { "epoch": 19.6, "grad_norm": 104.00243377685547, "learning_rate": 4.108687922247245e-05, "loss": 1.0694, "step": 2450 }, { "epoch": 19.608, "grad_norm": 44.02994155883789, "learning_rate": 4.108095294535973e-05, "loss": 1.2006, "step": 2451 }, { "epoch": 19.616, "grad_norm": 24.583053588867188, "learning_rate": 4.1075026668247006e-05, "loss": 1.1426, "step": 2452 }, { "epoch": 19.624, "grad_norm": 22.756685256958008, "learning_rate": 4.106910039113429e-05, "loss": 1.3041, "step": 2453 }, { "epoch": 19.632, "grad_norm": 32.487361907958984, "learning_rate": 4.106317411402158e-05, "loss": 1.8962, "step": 2454 }, { "epoch": 19.64, "grad_norm": 29.436626434326172, "learning_rate": 4.1057247836908856e-05, "loss": 1.171, "step": 2455 }, { "epoch": 19.648, "grad_norm": 18.09650993347168, "learning_rate": 4.1051321559796135e-05, "loss": 1.2549, "step": 2456 }, { "epoch": 19.656, "grad_norm": 26.92609405517578, "learning_rate": 4.104539528268342e-05, "loss": 2.0732, "step": 2457 }, { "epoch": 19.664, "grad_norm": 31.58161735534668, "learning_rate": 4.10394690055707e-05, "loss": 1.2498, "step": 2458 }, { "epoch": 19.672, "grad_norm": 20.59027862548828, "learning_rate": 4.1033542728457985e-05, "loss": 1.2551, "step": 2459 }, { "epoch": 19.68, "grad_norm": 37.93081283569336, "learning_rate": 4.102761645134527e-05, "loss": 1.2543, "step": 2460 }, { "epoch": 19.688, "grad_norm": 13.004794120788574, "learning_rate": 4.102169017423255e-05, "loss": 1.538, "step": 2461 }, { "epoch": 19.696, "grad_norm": 43.44667053222656, "learning_rate": 4.101576389711983e-05, "loss": 1.4787, "step": 2462 }, { "epoch": 19.704, "grad_norm": 31.62773895263672, "learning_rate": 4.1009837620007114e-05, "loss": 1.24, "step": 2463 }, { "epoch": 19.712, "grad_norm": 28.274002075195312, "learning_rate": 4.10039113428944e-05, "loss": 1.2174, "step": 2464 }, { "epoch": 19.72, "grad_norm": 24.47705841064453, "learning_rate": 4.099798506578168e-05, "loss": 1.5693, "step": 2465 }, { "epoch": 19.728, "grad_norm": 16.782377243041992, "learning_rate": 4.099205878866896e-05, "loss": 1.5542, "step": 2466 }, { "epoch": 19.736, "grad_norm": 32.540401458740234, "learning_rate": 4.098613251155624e-05, "loss": 0.9879, "step": 2467 }, { "epoch": 19.744, "grad_norm": 14.72663402557373, "learning_rate": 4.098020623444352e-05, "loss": 2.1611, "step": 2468 }, { "epoch": 19.752, "grad_norm": 30.773996353149414, "learning_rate": 4.097427995733081e-05, "loss": 1.8106, "step": 2469 }, { "epoch": 19.76, "grad_norm": 39.09892272949219, "learning_rate": 4.096835368021809e-05, "loss": 1.267, "step": 2470 }, { "epoch": 19.768, "grad_norm": 28.811813354492188, "learning_rate": 4.096242740310537e-05, "loss": 2.268, "step": 2471 }, { "epoch": 19.776, "grad_norm": 30.90420913696289, "learning_rate": 4.095650112599265e-05, "loss": 1.418, "step": 2472 }, { "epoch": 19.784, "grad_norm": 30.37932014465332, "learning_rate": 4.0950574848879936e-05, "loss": 1.5825, "step": 2473 }, { "epoch": 19.792, "grad_norm": 37.009796142578125, "learning_rate": 4.094464857176722e-05, "loss": 1.333, "step": 2474 }, { "epoch": 19.8, "grad_norm": 35.623291015625, "learning_rate": 4.09387222946545e-05, "loss": 1.4215, "step": 2475 }, { "epoch": 19.808, "grad_norm": 23.495460510253906, "learning_rate": 4.093279601754178e-05, "loss": 1.3143, "step": 2476 }, { "epoch": 19.816, "grad_norm": 25.534683227539062, "learning_rate": 4.0926869740429065e-05, "loss": 1.0291, "step": 2477 }, { "epoch": 19.824, "grad_norm": 28.32093048095703, "learning_rate": 4.0920943463316344e-05, "loss": 1.4057, "step": 2478 }, { "epoch": 19.832, "grad_norm": 24.63149070739746, "learning_rate": 4.091501718620363e-05, "loss": 0.9603, "step": 2479 }, { "epoch": 19.84, "grad_norm": 31.958494186401367, "learning_rate": 4.0909090909090915e-05, "loss": 1.3264, "step": 2480 }, { "epoch": 19.848, "grad_norm": 37.06574249267578, "learning_rate": 4.0903164631978194e-05, "loss": 0.9927, "step": 2481 }, { "epoch": 19.856, "grad_norm": 59.93120574951172, "learning_rate": 4.089723835486547e-05, "loss": 1.2948, "step": 2482 }, { "epoch": 19.864, "grad_norm": 31.25486183166504, "learning_rate": 4.089131207775276e-05, "loss": 1.4031, "step": 2483 }, { "epoch": 19.872, "grad_norm": 32.231937408447266, "learning_rate": 4.0885385800640044e-05, "loss": 1.0666, "step": 2484 }, { "epoch": 19.88, "grad_norm": 35.95759963989258, "learning_rate": 4.087945952352732e-05, "loss": 1.5167, "step": 2485 }, { "epoch": 19.888, "grad_norm": 41.074493408203125, "learning_rate": 4.08735332464146e-05, "loss": 1.7671, "step": 2486 }, { "epoch": 19.896, "grad_norm": 71.88697814941406, "learning_rate": 4.086760696930189e-05, "loss": 1.3015, "step": 2487 }, { "epoch": 19.904, "grad_norm": 27.667924880981445, "learning_rate": 4.086168069218917e-05, "loss": 1.0528, "step": 2488 }, { "epoch": 19.912, "grad_norm": 31.870220184326172, "learning_rate": 4.085575441507645e-05, "loss": 1.0674, "step": 2489 }, { "epoch": 19.92, "grad_norm": 18.623470306396484, "learning_rate": 4.084982813796373e-05, "loss": 1.4287, "step": 2490 }, { "epoch": 19.928, "grad_norm": 20.441200256347656, "learning_rate": 4.0843901860851016e-05, "loss": 1.1444, "step": 2491 }, { "epoch": 19.936, "grad_norm": 23.73478126525879, "learning_rate": 4.0837975583738294e-05, "loss": 1.2001, "step": 2492 }, { "epoch": 19.944, "grad_norm": 22.0040283203125, "learning_rate": 4.083204930662558e-05, "loss": 1.0329, "step": 2493 }, { "epoch": 19.951999999999998, "grad_norm": 35.92507553100586, "learning_rate": 4.0826123029512866e-05, "loss": 1.1999, "step": 2494 }, { "epoch": 19.96, "grad_norm": 26.482152938842773, "learning_rate": 4.0820196752400144e-05, "loss": 1.1456, "step": 2495 }, { "epoch": 19.968, "grad_norm": 18.07729721069336, "learning_rate": 4.081427047528742e-05, "loss": 1.0671, "step": 2496 }, { "epoch": 19.976, "grad_norm": 28.79245948791504, "learning_rate": 4.080834419817471e-05, "loss": 1.1169, "step": 2497 }, { "epoch": 19.984, "grad_norm": 29.098400115966797, "learning_rate": 4.0802417921061994e-05, "loss": 1.2481, "step": 2498 }, { "epoch": 19.992, "grad_norm": 42.438655853271484, "learning_rate": 4.079649164394927e-05, "loss": 1.2363, "step": 2499 }, { "epoch": 20.0, "grad_norm": 64.01585388183594, "learning_rate": 4.079056536683655e-05, "loss": 1.7184, "step": 2500 }, { "epoch": 20.0, "eval_loss": 1.4192875623703003, "eval_map": 0.2657, "eval_map_50": 0.5655, "eval_map_75": 0.1847, "eval_map_Coverall": 0.4752, "eval_map_Face_Shield": 0.2071, "eval_map_Gloves": 0.1461, "eval_map_Goggles": 0.1277, "eval_map_Mask": 0.3722, "eval_map_large": 0.3401, "eval_map_medium": 0.1912, "eval_map_small": 0.1722, "eval_mar_1": 0.2526, "eval_mar_10": 0.4532, "eval_mar_100": 0.4704, "eval_mar_100_Coverall": 0.6889, "eval_mar_100_Face_Shield": 0.5, "eval_mar_100_Gloves": 0.3508, "eval_mar_100_Goggles": 0.3375, "eval_mar_100_Mask": 0.475, "eval_mar_large": 0.6364, "eval_mar_medium": 0.3285, "eval_mar_small": 0.2436, "eval_runtime": 1.8337, "eval_samples_per_second": 15.815, "eval_steps_per_second": 1.091, "step": 2500 }, { "epoch": 20.008, "grad_norm": 37.627540588378906, "learning_rate": 4.078463908972384e-05, "loss": 1.6461, "step": 2501 }, { "epoch": 20.016, "grad_norm": 42.90981674194336, "learning_rate": 4.0778712812611117e-05, "loss": 1.1618, "step": 2502 }, { "epoch": 20.024, "grad_norm": 26.250511169433594, "learning_rate": 4.07727865354984e-05, "loss": 1.1672, "step": 2503 }, { "epoch": 20.032, "grad_norm": 26.95331382751465, "learning_rate": 4.076686025838569e-05, "loss": 1.4481, "step": 2504 }, { "epoch": 20.04, "grad_norm": 21.404434204101562, "learning_rate": 4.0760933981272967e-05, "loss": 1.552, "step": 2505 }, { "epoch": 20.048, "grad_norm": 36.758663177490234, "learning_rate": 4.0755007704160245e-05, "loss": 1.1417, "step": 2506 }, { "epoch": 20.056, "grad_norm": 38.03287887573242, "learning_rate": 4.074908142704753e-05, "loss": 1.2208, "step": 2507 }, { "epoch": 20.064, "grad_norm": 24.062870025634766, "learning_rate": 4.0743155149934817e-05, "loss": 1.9864, "step": 2508 }, { "epoch": 20.072, "grad_norm": 19.486209869384766, "learning_rate": 4.0737228872822095e-05, "loss": 1.4933, "step": 2509 }, { "epoch": 20.08, "grad_norm": 23.180917739868164, "learning_rate": 4.0731302595709374e-05, "loss": 1.6424, "step": 2510 }, { "epoch": 20.088, "grad_norm": 21.780820846557617, "learning_rate": 4.072537631859666e-05, "loss": 1.1703, "step": 2511 }, { "epoch": 20.096, "grad_norm": 24.079334259033203, "learning_rate": 4.071945004148394e-05, "loss": 1.242, "step": 2512 }, { "epoch": 20.104, "grad_norm": 56.070796966552734, "learning_rate": 4.0713523764371224e-05, "loss": 1.4824, "step": 2513 }, { "epoch": 20.112, "grad_norm": 34.7274055480957, "learning_rate": 4.070759748725851e-05, "loss": 1.1897, "step": 2514 }, { "epoch": 20.12, "grad_norm": 91.18883514404297, "learning_rate": 4.070167121014579e-05, "loss": 1.5682, "step": 2515 }, { "epoch": 20.128, "grad_norm": 46.80442428588867, "learning_rate": 4.069574493303307e-05, "loss": 1.3074, "step": 2516 }, { "epoch": 20.136, "grad_norm": 31.954002380371094, "learning_rate": 4.068981865592035e-05, "loss": 1.145, "step": 2517 }, { "epoch": 20.144, "grad_norm": 34.628021240234375, "learning_rate": 4.068389237880764e-05, "loss": 1.342, "step": 2518 }, { "epoch": 20.152, "grad_norm": 24.190853118896484, "learning_rate": 4.067796610169492e-05, "loss": 1.3489, "step": 2519 }, { "epoch": 20.16, "grad_norm": 17.61078453063965, "learning_rate": 4.0672039824582196e-05, "loss": 1.3715, "step": 2520 }, { "epoch": 20.168, "grad_norm": 35.388545989990234, "learning_rate": 4.066611354746948e-05, "loss": 1.3214, "step": 2521 }, { "epoch": 20.176, "grad_norm": 39.683223724365234, "learning_rate": 4.066018727035677e-05, "loss": 1.4677, "step": 2522 }, { "epoch": 20.184, "grad_norm": 22.75623893737793, "learning_rate": 4.0654260993244046e-05, "loss": 2.3385, "step": 2523 }, { "epoch": 20.192, "grad_norm": 27.693546295166016, "learning_rate": 4.0648334716131325e-05, "loss": 1.6824, "step": 2524 }, { "epoch": 20.2, "grad_norm": 23.124340057373047, "learning_rate": 4.064240843901861e-05, "loss": 1.6784, "step": 2525 }, { "epoch": 20.208, "grad_norm": 23.489459991455078, "learning_rate": 4.063648216190589e-05, "loss": 1.8195, "step": 2526 }, { "epoch": 20.216, "grad_norm": 47.15576934814453, "learning_rate": 4.0630555884793175e-05, "loss": 1.5001, "step": 2527 }, { "epoch": 20.224, "grad_norm": 27.26197624206543, "learning_rate": 4.062462960768046e-05, "loss": 1.0045, "step": 2528 }, { "epoch": 20.232, "grad_norm": 23.69361686706543, "learning_rate": 4.061870333056774e-05, "loss": 1.5555, "step": 2529 }, { "epoch": 20.24, "grad_norm": 45.0792350769043, "learning_rate": 4.061277705345502e-05, "loss": 1.4248, "step": 2530 }, { "epoch": 20.248, "grad_norm": 20.90306854248047, "learning_rate": 4.0606850776342304e-05, "loss": 1.3885, "step": 2531 }, { "epoch": 20.256, "grad_norm": 23.308473587036133, "learning_rate": 4.060092449922959e-05, "loss": 1.5107, "step": 2532 }, { "epoch": 20.264, "grad_norm": 27.100845336914062, "learning_rate": 4.059499822211687e-05, "loss": 1.105, "step": 2533 }, { "epoch": 20.272, "grad_norm": 31.030126571655273, "learning_rate": 4.058907194500415e-05, "loss": 1.6166, "step": 2534 }, { "epoch": 20.28, "grad_norm": 29.312780380249023, "learning_rate": 4.058314566789143e-05, "loss": 1.2389, "step": 2535 }, { "epoch": 20.288, "grad_norm": 24.383867263793945, "learning_rate": 4.057721939077871e-05, "loss": 0.9749, "step": 2536 }, { "epoch": 20.296, "grad_norm": 21.169700622558594, "learning_rate": 4.0571293113666e-05, "loss": 1.087, "step": 2537 }, { "epoch": 20.304, "grad_norm": 34.28033447265625, "learning_rate": 4.056536683655328e-05, "loss": 1.2056, "step": 2538 }, { "epoch": 20.312, "grad_norm": 26.623764038085938, "learning_rate": 4.055944055944056e-05, "loss": 1.6558, "step": 2539 }, { "epoch": 20.32, "grad_norm": 21.208728790283203, "learning_rate": 4.055351428232784e-05, "loss": 1.147, "step": 2540 }, { "epoch": 20.328, "grad_norm": 29.098743438720703, "learning_rate": 4.0547588005215126e-05, "loss": 1.4783, "step": 2541 }, { "epoch": 20.336, "grad_norm": 47.58907699584961, "learning_rate": 4.054166172810241e-05, "loss": 2.1717, "step": 2542 }, { "epoch": 20.344, "grad_norm": 22.82941436767578, "learning_rate": 4.053573545098969e-05, "loss": 1.1924, "step": 2543 }, { "epoch": 20.352, "grad_norm": 22.91252899169922, "learning_rate": 4.052980917387697e-05, "loss": 1.178, "step": 2544 }, { "epoch": 20.36, "grad_norm": 31.788673400878906, "learning_rate": 4.0523882896764255e-05, "loss": 1.3685, "step": 2545 }, { "epoch": 20.368, "grad_norm": 49.11943435668945, "learning_rate": 4.0517956619651534e-05, "loss": 1.648, "step": 2546 }, { "epoch": 20.376, "grad_norm": 27.470802307128906, "learning_rate": 4.051203034253882e-05, "loss": 1.4161, "step": 2547 }, { "epoch": 20.384, "grad_norm": 24.18962287902832, "learning_rate": 4.0506104065426105e-05, "loss": 2.0781, "step": 2548 }, { "epoch": 20.392, "grad_norm": 54.98915100097656, "learning_rate": 4.0500177788313384e-05, "loss": 1.1485, "step": 2549 }, { "epoch": 20.4, "grad_norm": 15.235400199890137, "learning_rate": 4.049425151120066e-05, "loss": 1.8738, "step": 2550 }, { "epoch": 20.408, "grad_norm": 18.95864486694336, "learning_rate": 4.048832523408795e-05, "loss": 1.2499, "step": 2551 }, { "epoch": 20.416, "grad_norm": 24.488481521606445, "learning_rate": 4.0482398956975234e-05, "loss": 1.7557, "step": 2552 }, { "epoch": 20.424, "grad_norm": 24.4366397857666, "learning_rate": 4.047647267986251e-05, "loss": 1.5369, "step": 2553 }, { "epoch": 20.432, "grad_norm": 44.50730895996094, "learning_rate": 4.047054640274979e-05, "loss": 2.3507, "step": 2554 }, { "epoch": 20.44, "grad_norm": 23.731712341308594, "learning_rate": 4.046462012563708e-05, "loss": 1.2465, "step": 2555 }, { "epoch": 20.448, "grad_norm": 16.945568084716797, "learning_rate": 4.045869384852436e-05, "loss": 1.0566, "step": 2556 }, { "epoch": 20.456, "grad_norm": 21.969959259033203, "learning_rate": 4.045276757141164e-05, "loss": 1.6055, "step": 2557 }, { "epoch": 20.464, "grad_norm": 25.942195892333984, "learning_rate": 4.044684129429893e-05, "loss": 1.0885, "step": 2558 }, { "epoch": 20.472, "grad_norm": 23.933570861816406, "learning_rate": 4.0440915017186206e-05, "loss": 1.0595, "step": 2559 }, { "epoch": 20.48, "grad_norm": 29.958980560302734, "learning_rate": 4.0434988740073485e-05, "loss": 1.0516, "step": 2560 }, { "epoch": 20.488, "grad_norm": 26.022268295288086, "learning_rate": 4.042906246296077e-05, "loss": 1.3383, "step": 2561 }, { "epoch": 20.496, "grad_norm": 24.599365234375, "learning_rate": 4.0423136185848056e-05, "loss": 1.6428, "step": 2562 }, { "epoch": 20.504, "grad_norm": 19.501523971557617, "learning_rate": 4.0417209908735335e-05, "loss": 1.3878, "step": 2563 }, { "epoch": 20.512, "grad_norm": 152.2682342529297, "learning_rate": 4.041128363162261e-05, "loss": 3.0665, "step": 2564 }, { "epoch": 20.52, "grad_norm": 29.03986358642578, "learning_rate": 4.04053573545099e-05, "loss": 1.2262, "step": 2565 }, { "epoch": 20.528, "grad_norm": 27.920915603637695, "learning_rate": 4.0399431077397185e-05, "loss": 1.2413, "step": 2566 }, { "epoch": 20.536, "grad_norm": 21.570877075195312, "learning_rate": 4.039350480028446e-05, "loss": 0.8308, "step": 2567 }, { "epoch": 20.544, "grad_norm": 20.34467315673828, "learning_rate": 4.038757852317174e-05, "loss": 1.509, "step": 2568 }, { "epoch": 20.552, "grad_norm": 29.25185775756836, "learning_rate": 4.038165224605903e-05, "loss": 1.1147, "step": 2569 }, { "epoch": 20.56, "grad_norm": 42.86314010620117, "learning_rate": 4.0375725968946307e-05, "loss": 1.6598, "step": 2570 }, { "epoch": 20.568, "grad_norm": 44.093650817871094, "learning_rate": 4.036979969183359e-05, "loss": 1.5188, "step": 2571 }, { "epoch": 20.576, "grad_norm": 90.62466430664062, "learning_rate": 4.036387341472088e-05, "loss": 1.0377, "step": 2572 }, { "epoch": 20.584, "grad_norm": 33.51825714111328, "learning_rate": 4.035794713760816e-05, "loss": 1.4834, "step": 2573 }, { "epoch": 20.592, "grad_norm": 21.093992233276367, "learning_rate": 4.0352020860495435e-05, "loss": 1.5846, "step": 2574 }, { "epoch": 20.6, "grad_norm": 17.20598030090332, "learning_rate": 4.034609458338272e-05, "loss": 1.3564, "step": 2575 }, { "epoch": 20.608, "grad_norm": 31.105255126953125, "learning_rate": 4.034016830627001e-05, "loss": 1.35, "step": 2576 }, { "epoch": 20.616, "grad_norm": 31.119827270507812, "learning_rate": 4.0334242029157285e-05, "loss": 0.9509, "step": 2577 }, { "epoch": 20.624, "grad_norm": 31.985004425048828, "learning_rate": 4.0328315752044564e-05, "loss": 1.529, "step": 2578 }, { "epoch": 20.632, "grad_norm": 36.78902053833008, "learning_rate": 4.032238947493185e-05, "loss": 0.9151, "step": 2579 }, { "epoch": 20.64, "grad_norm": 27.48998260498047, "learning_rate": 4.031646319781913e-05, "loss": 1.1944, "step": 2580 }, { "epoch": 20.648, "grad_norm": 50.53257751464844, "learning_rate": 4.0310536920706414e-05, "loss": 1.1321, "step": 2581 }, { "epoch": 20.656, "grad_norm": 48.9189567565918, "learning_rate": 4.03046106435937e-05, "loss": 1.429, "step": 2582 }, { "epoch": 20.664, "grad_norm": 19.855899810791016, "learning_rate": 4.029868436648098e-05, "loss": 1.7122, "step": 2583 }, { "epoch": 20.672, "grad_norm": 21.64097785949707, "learning_rate": 4.029275808936826e-05, "loss": 1.3131, "step": 2584 }, { "epoch": 20.68, "grad_norm": 45.821807861328125, "learning_rate": 4.028683181225554e-05, "loss": 1.6173, "step": 2585 }, { "epoch": 20.688, "grad_norm": 44.90935516357422, "learning_rate": 4.028090553514283e-05, "loss": 1.2251, "step": 2586 }, { "epoch": 20.696, "grad_norm": 27.7172908782959, "learning_rate": 4.027497925803011e-05, "loss": 1.1999, "step": 2587 }, { "epoch": 20.704, "grad_norm": 86.55828857421875, "learning_rate": 4.0269052980917386e-05, "loss": 1.0686, "step": 2588 }, { "epoch": 20.712, "grad_norm": 28.63361358642578, "learning_rate": 4.026312670380467e-05, "loss": 1.1201, "step": 2589 }, { "epoch": 20.72, "grad_norm": 23.641441345214844, "learning_rate": 4.025720042669196e-05, "loss": 1.0527, "step": 2590 }, { "epoch": 20.728, "grad_norm": 76.27790832519531, "learning_rate": 4.0251274149579236e-05, "loss": 1.4578, "step": 2591 }, { "epoch": 20.736, "grad_norm": 33.27108383178711, "learning_rate": 4.024534787246652e-05, "loss": 1.3788, "step": 2592 }, { "epoch": 20.744, "grad_norm": 20.26272201538086, "learning_rate": 4.02394215953538e-05, "loss": 0.972, "step": 2593 }, { "epoch": 20.752, "grad_norm": 32.53107452392578, "learning_rate": 4.023349531824108e-05, "loss": 1.2444, "step": 2594 }, { "epoch": 20.76, "grad_norm": 30.716224670410156, "learning_rate": 4.0227569041128365e-05, "loss": 1.4883, "step": 2595 }, { "epoch": 20.768, "grad_norm": 26.915315628051758, "learning_rate": 4.022164276401565e-05, "loss": 1.4159, "step": 2596 }, { "epoch": 20.776, "grad_norm": 37.77144241333008, "learning_rate": 4.021571648690293e-05, "loss": 0.8986, "step": 2597 }, { "epoch": 20.784, "grad_norm": 138.7915496826172, "learning_rate": 4.020979020979021e-05, "loss": 1.135, "step": 2598 }, { "epoch": 20.792, "grad_norm": 32.3948860168457, "learning_rate": 4.0203863932677494e-05, "loss": 1.2836, "step": 2599 }, { "epoch": 20.8, "grad_norm": 237.1825714111328, "learning_rate": 4.019793765556478e-05, "loss": 1.198, "step": 2600 }, { "epoch": 20.808, "grad_norm": 22.197925567626953, "learning_rate": 4.019201137845206e-05, "loss": 1.3802, "step": 2601 }, { "epoch": 20.816, "grad_norm": 79.1783218383789, "learning_rate": 4.018608510133934e-05, "loss": 1.3636, "step": 2602 }, { "epoch": 20.824, "grad_norm": 29.144134521484375, "learning_rate": 4.018015882422662e-05, "loss": 1.0066, "step": 2603 }, { "epoch": 20.832, "grad_norm": 16.44243621826172, "learning_rate": 4.01742325471139e-05, "loss": 1.2233, "step": 2604 }, { "epoch": 20.84, "grad_norm": 47.06879806518555, "learning_rate": 4.016830627000119e-05, "loss": 1.2313, "step": 2605 }, { "epoch": 20.848, "grad_norm": 19.302562713623047, "learning_rate": 4.016237999288847e-05, "loss": 1.2097, "step": 2606 }, { "epoch": 20.856, "grad_norm": 40.96134948730469, "learning_rate": 4.015645371577575e-05, "loss": 2.2987, "step": 2607 }, { "epoch": 20.864, "grad_norm": 35.24637222290039, "learning_rate": 4.015052743866303e-05, "loss": 1.5114, "step": 2608 }, { "epoch": 20.872, "grad_norm": 22.487323760986328, "learning_rate": 4.0144601161550316e-05, "loss": 1.5223, "step": 2609 }, { "epoch": 20.88, "grad_norm": 24.171907424926758, "learning_rate": 4.01386748844376e-05, "loss": 1.423, "step": 2610 }, { "epoch": 20.888, "grad_norm": 44.35542678833008, "learning_rate": 4.013274860732488e-05, "loss": 1.1207, "step": 2611 }, { "epoch": 20.896, "grad_norm": 97.93327331542969, "learning_rate": 4.012682233021216e-05, "loss": 1.4402, "step": 2612 }, { "epoch": 20.904, "grad_norm": 32.91543197631836, "learning_rate": 4.0120896053099445e-05, "loss": 1.5441, "step": 2613 }, { "epoch": 20.912, "grad_norm": 24.051624298095703, "learning_rate": 4.011496977598673e-05, "loss": 2.3859, "step": 2614 }, { "epoch": 20.92, "grad_norm": 48.490318298339844, "learning_rate": 4.010904349887401e-05, "loss": 1.4698, "step": 2615 }, { "epoch": 20.928, "grad_norm": 30.55628204345703, "learning_rate": 4.0103117221761295e-05, "loss": 1.5384, "step": 2616 }, { "epoch": 20.936, "grad_norm": 31.285320281982422, "learning_rate": 4.0097190944648574e-05, "loss": 1.0934, "step": 2617 }, { "epoch": 20.944, "grad_norm": 94.65945434570312, "learning_rate": 4.009126466753585e-05, "loss": 1.3988, "step": 2618 }, { "epoch": 20.951999999999998, "grad_norm": 18.989259719848633, "learning_rate": 4.008533839042314e-05, "loss": 1.225, "step": 2619 }, { "epoch": 20.96, "grad_norm": 21.180938720703125, "learning_rate": 4.0079412113310424e-05, "loss": 1.6309, "step": 2620 }, { "epoch": 20.968, "grad_norm": 19.6774845123291, "learning_rate": 4.00734858361977e-05, "loss": 1.3583, "step": 2621 }, { "epoch": 20.976, "grad_norm": 19.70520782470703, "learning_rate": 4.006755955908498e-05, "loss": 1.344, "step": 2622 }, { "epoch": 20.984, "grad_norm": 37.53491973876953, "learning_rate": 4.006163328197227e-05, "loss": 1.4063, "step": 2623 }, { "epoch": 20.992, "grad_norm": 108.57771301269531, "learning_rate": 4.005570700485955e-05, "loss": 1.5285, "step": 2624 }, { "epoch": 21.0, "grad_norm": 25.70987892150879, "learning_rate": 4.004978072774683e-05, "loss": 1.4366, "step": 2625 }, { "epoch": 21.0, "eval_loss": 1.3553361892700195, "eval_map": 0.24, "eval_map_50": 0.5054, "eval_map_75": 0.1867, "eval_map_Coverall": 0.4809, "eval_map_Face_Shield": 0.1515, "eval_map_Gloves": 0.1316, "eval_map_Goggles": 0.0737, "eval_map_Mask": 0.362, "eval_map_large": 0.4046, "eval_map_medium": 0.1606, "eval_map_small": 0.1116, "eval_mar_1": 0.24, "eval_mar_10": 0.4239, "eval_mar_100": 0.4611, "eval_mar_100_Coverall": 0.6311, "eval_mar_100_Face_Shield": 0.4941, "eval_mar_100_Gloves": 0.3574, "eval_mar_100_Goggles": 0.375, "eval_mar_100_Mask": 0.4481, "eval_mar_large": 0.6626, "eval_mar_medium": 0.3265, "eval_mar_small": 0.2037, "eval_runtime": 1.8169, "eval_samples_per_second": 15.961, "eval_steps_per_second": 1.101, "step": 2625 }, { "epoch": 21.008, "grad_norm": 39.60087203979492, "learning_rate": 4.004385445063412e-05, "loss": 1.3512, "step": 2626 }, { "epoch": 21.016, "grad_norm": 23.45526123046875, "learning_rate": 4.0037928173521396e-05, "loss": 1.6742, "step": 2627 }, { "epoch": 21.024, "grad_norm": 23.22933578491211, "learning_rate": 4.0032001896408675e-05, "loss": 1.8223, "step": 2628 }, { "epoch": 21.032, "grad_norm": 19.11606216430664, "learning_rate": 4.002607561929596e-05, "loss": 1.2951, "step": 2629 }, { "epoch": 21.04, "grad_norm": 31.464279174804688, "learning_rate": 4.0020149342183246e-05, "loss": 1.4596, "step": 2630 }, { "epoch": 21.048, "grad_norm": 27.190872192382812, "learning_rate": 4.0014223065070525e-05, "loss": 1.0118, "step": 2631 }, { "epoch": 21.056, "grad_norm": 19.632431030273438, "learning_rate": 4.0008296787957803e-05, "loss": 1.3762, "step": 2632 }, { "epoch": 21.064, "grad_norm": 42.47008514404297, "learning_rate": 4.000237051084509e-05, "loss": 2.0798, "step": 2633 }, { "epoch": 21.072, "grad_norm": 37.669925689697266, "learning_rate": 3.9996444233732375e-05, "loss": 1.1134, "step": 2634 }, { "epoch": 21.08, "grad_norm": 24.186914443969727, "learning_rate": 3.9990517956619653e-05, "loss": 1.4487, "step": 2635 }, { "epoch": 21.088, "grad_norm": 30.40351104736328, "learning_rate": 3.998459167950693e-05, "loss": 1.1566, "step": 2636 }, { "epoch": 21.096, "grad_norm": 86.84015655517578, "learning_rate": 3.997866540239422e-05, "loss": 0.9877, "step": 2637 }, { "epoch": 21.104, "grad_norm": 26.47760581970215, "learning_rate": 3.99727391252815e-05, "loss": 1.3991, "step": 2638 }, { "epoch": 21.112, "grad_norm": 17.781187057495117, "learning_rate": 3.996681284816878e-05, "loss": 1.6208, "step": 2639 }, { "epoch": 21.12, "grad_norm": 22.58083152770996, "learning_rate": 3.996088657105607e-05, "loss": 1.465, "step": 2640 }, { "epoch": 21.128, "grad_norm": 32.27186965942383, "learning_rate": 3.995496029394335e-05, "loss": 1.3512, "step": 2641 }, { "epoch": 21.136, "grad_norm": 22.10335922241211, "learning_rate": 3.9949034016830626e-05, "loss": 1.1814, "step": 2642 }, { "epoch": 21.144, "grad_norm": 14.655496597290039, "learning_rate": 3.994310773971791e-05, "loss": 1.3375, "step": 2643 }, { "epoch": 21.152, "grad_norm": 54.65118408203125, "learning_rate": 3.99371814626052e-05, "loss": 1.4445, "step": 2644 }, { "epoch": 21.16, "grad_norm": 19.83011245727539, "learning_rate": 3.9931255185492476e-05, "loss": 1.0465, "step": 2645 }, { "epoch": 21.168, "grad_norm": 31.636211395263672, "learning_rate": 3.9925328908379754e-05, "loss": 1.1353, "step": 2646 }, { "epoch": 21.176, "grad_norm": 16.066131591796875, "learning_rate": 3.991940263126704e-05, "loss": 1.167, "step": 2647 }, { "epoch": 21.184, "grad_norm": 35.54179763793945, "learning_rate": 3.9913476354154326e-05, "loss": 2.5066, "step": 2648 }, { "epoch": 21.192, "grad_norm": 19.968761444091797, "learning_rate": 3.9907550077041604e-05, "loss": 1.2668, "step": 2649 }, { "epoch": 21.2, "grad_norm": 63.295860290527344, "learning_rate": 3.990162379992889e-05, "loss": 1.212, "step": 2650 }, { "epoch": 21.208, "grad_norm": 22.635061264038086, "learning_rate": 3.989569752281617e-05, "loss": 1.1558, "step": 2651 }, { "epoch": 21.216, "grad_norm": 34.30824661254883, "learning_rate": 3.988977124570345e-05, "loss": 1.2239, "step": 2652 }, { "epoch": 21.224, "grad_norm": 60.68344497680664, "learning_rate": 3.988384496859073e-05, "loss": 1.4599, "step": 2653 }, { "epoch": 21.232, "grad_norm": 44.81584548950195, "learning_rate": 3.987791869147802e-05, "loss": 1.4452, "step": 2654 }, { "epoch": 21.24, "grad_norm": 31.696664810180664, "learning_rate": 3.98719924143653e-05, "loss": 1.4041, "step": 2655 }, { "epoch": 21.248, "grad_norm": 30.661453247070312, "learning_rate": 3.9866066137252576e-05, "loss": 1.111, "step": 2656 }, { "epoch": 21.256, "grad_norm": 29.867290496826172, "learning_rate": 3.986013986013986e-05, "loss": 1.2067, "step": 2657 }, { "epoch": 21.264, "grad_norm": 16.273590087890625, "learning_rate": 3.985421358302715e-05, "loss": 1.1832, "step": 2658 }, { "epoch": 21.272, "grad_norm": 20.23128890991211, "learning_rate": 3.9848287305914426e-05, "loss": 1.317, "step": 2659 }, { "epoch": 21.28, "grad_norm": 44.2840576171875, "learning_rate": 3.984236102880171e-05, "loss": 1.8394, "step": 2660 }, { "epoch": 21.288, "grad_norm": 19.518503189086914, "learning_rate": 3.983643475168899e-05, "loss": 1.1941, "step": 2661 }, { "epoch": 21.296, "grad_norm": 25.107765197753906, "learning_rate": 3.983050847457627e-05, "loss": 1.3233, "step": 2662 }, { "epoch": 21.304, "grad_norm": 60.41316604614258, "learning_rate": 3.9824582197463555e-05, "loss": 1.8337, "step": 2663 }, { "epoch": 21.312, "grad_norm": 27.068767547607422, "learning_rate": 3.981865592035084e-05, "loss": 1.5234, "step": 2664 }, { "epoch": 21.32, "grad_norm": 40.215755462646484, "learning_rate": 3.981272964323812e-05, "loss": 1.2295, "step": 2665 }, { "epoch": 21.328, "grad_norm": 41.1734733581543, "learning_rate": 3.98068033661254e-05, "loss": 1.4148, "step": 2666 }, { "epoch": 21.336, "grad_norm": 27.23029327392578, "learning_rate": 3.9800877089012684e-05, "loss": 1.7697, "step": 2667 }, { "epoch": 21.344, "grad_norm": 23.568374633789062, "learning_rate": 3.979495081189997e-05, "loss": 1.5705, "step": 2668 }, { "epoch": 21.352, "grad_norm": 28.292146682739258, "learning_rate": 3.978902453478725e-05, "loss": 1.3212, "step": 2669 }, { "epoch": 21.36, "grad_norm": 28.510393142700195, "learning_rate": 3.978309825767453e-05, "loss": 1.2389, "step": 2670 }, { "epoch": 21.368, "grad_norm": 19.442842483520508, "learning_rate": 3.977717198056181e-05, "loss": 1.4982, "step": 2671 }, { "epoch": 21.376, "grad_norm": 33.62866973876953, "learning_rate": 3.977124570344909e-05, "loss": 1.4148, "step": 2672 }, { "epoch": 21.384, "grad_norm": 30.83207893371582, "learning_rate": 3.976531942633638e-05, "loss": 1.5734, "step": 2673 }, { "epoch": 21.392, "grad_norm": 27.49104118347168, "learning_rate": 3.975939314922366e-05, "loss": 1.3052, "step": 2674 }, { "epoch": 21.4, "grad_norm": 38.684844970703125, "learning_rate": 3.975346687211094e-05, "loss": 1.3319, "step": 2675 }, { "epoch": 21.408, "grad_norm": 16.395395278930664, "learning_rate": 3.974754059499822e-05, "loss": 1.4606, "step": 2676 }, { "epoch": 21.416, "grad_norm": 26.636104583740234, "learning_rate": 3.9741614317885506e-05, "loss": 1.4641, "step": 2677 }, { "epoch": 21.424, "grad_norm": 30.742876052856445, "learning_rate": 3.973568804077279e-05, "loss": 1.2083, "step": 2678 }, { "epoch": 21.432, "grad_norm": 36.371734619140625, "learning_rate": 3.972976176366007e-05, "loss": 1.5187, "step": 2679 }, { "epoch": 21.44, "grad_norm": 197.90106201171875, "learning_rate": 3.972383548654735e-05, "loss": 1.4643, "step": 2680 }, { "epoch": 21.448, "grad_norm": 41.90885543823242, "learning_rate": 3.9717909209434635e-05, "loss": 1.4654, "step": 2681 }, { "epoch": 21.456, "grad_norm": 30.094152450561523, "learning_rate": 3.971198293232192e-05, "loss": 1.2679, "step": 2682 }, { "epoch": 21.464, "grad_norm": 27.957592010498047, "learning_rate": 3.97060566552092e-05, "loss": 1.14, "step": 2683 }, { "epoch": 21.472, "grad_norm": 22.47945785522461, "learning_rate": 3.9700130378096485e-05, "loss": 1.5445, "step": 2684 }, { "epoch": 21.48, "grad_norm": 27.502540588378906, "learning_rate": 3.9694204100983764e-05, "loss": 1.1952, "step": 2685 }, { "epoch": 21.488, "grad_norm": 198.27944946289062, "learning_rate": 3.968827782387104e-05, "loss": 3.4094, "step": 2686 }, { "epoch": 21.496, "grad_norm": 19.234830856323242, "learning_rate": 3.968235154675833e-05, "loss": 1.496, "step": 2687 }, { "epoch": 21.504, "grad_norm": 26.02581214904785, "learning_rate": 3.9676425269645614e-05, "loss": 2.7236, "step": 2688 }, { "epoch": 21.512, "grad_norm": 30.565834045410156, "learning_rate": 3.967049899253289e-05, "loss": 1.1905, "step": 2689 }, { "epoch": 21.52, "grad_norm": 16.158432006835938, "learning_rate": 3.966457271542017e-05, "loss": 1.4404, "step": 2690 }, { "epoch": 21.528, "grad_norm": 26.834871292114258, "learning_rate": 3.965864643830746e-05, "loss": 1.4104, "step": 2691 }, { "epoch": 21.536, "grad_norm": 29.825157165527344, "learning_rate": 3.965272016119474e-05, "loss": 1.2657, "step": 2692 }, { "epoch": 21.544, "grad_norm": 27.18134307861328, "learning_rate": 3.964679388408202e-05, "loss": 1.5307, "step": 2693 }, { "epoch": 21.552, "grad_norm": 28.950580596923828, "learning_rate": 3.964086760696931e-05, "loss": 2.049, "step": 2694 }, { "epoch": 21.56, "grad_norm": 36.78469467163086, "learning_rate": 3.9634941329856586e-05, "loss": 1.4184, "step": 2695 }, { "epoch": 21.568, "grad_norm": 41.77628707885742, "learning_rate": 3.9629015052743865e-05, "loss": 1.4331, "step": 2696 }, { "epoch": 21.576, "grad_norm": 45.39119338989258, "learning_rate": 3.962308877563115e-05, "loss": 1.4217, "step": 2697 }, { "epoch": 21.584, "grad_norm": 36.78221893310547, "learning_rate": 3.9617162498518436e-05, "loss": 1.4666, "step": 2698 }, { "epoch": 21.592, "grad_norm": 20.069303512573242, "learning_rate": 3.9611236221405715e-05, "loss": 1.2428, "step": 2699 }, { "epoch": 21.6, "grad_norm": 38.13924026489258, "learning_rate": 3.9605309944292994e-05, "loss": 1.5218, "step": 2700 }, { "epoch": 21.608, "grad_norm": 19.676010131835938, "learning_rate": 3.959938366718028e-05, "loss": 1.2755, "step": 2701 }, { "epoch": 21.616, "grad_norm": 41.854713439941406, "learning_rate": 3.9593457390067565e-05, "loss": 1.1642, "step": 2702 }, { "epoch": 21.624, "grad_norm": 44.905418395996094, "learning_rate": 3.9587531112954844e-05, "loss": 1.404, "step": 2703 }, { "epoch": 21.632, "grad_norm": 20.969552993774414, "learning_rate": 3.958160483584212e-05, "loss": 1.7109, "step": 2704 }, { "epoch": 21.64, "grad_norm": 31.070343017578125, "learning_rate": 3.957567855872941e-05, "loss": 1.0578, "step": 2705 }, { "epoch": 21.648, "grad_norm": 28.878555297851562, "learning_rate": 3.956975228161669e-05, "loss": 1.3976, "step": 2706 }, { "epoch": 21.656, "grad_norm": 39.560001373291016, "learning_rate": 3.956382600450397e-05, "loss": 1.4176, "step": 2707 }, { "epoch": 21.664, "grad_norm": 30.665508270263672, "learning_rate": 3.955789972739126e-05, "loss": 1.6247, "step": 2708 }, { "epoch": 21.672, "grad_norm": 41.428565979003906, "learning_rate": 3.955197345027854e-05, "loss": 2.2239, "step": 2709 }, { "epoch": 21.68, "grad_norm": 39.889991760253906, "learning_rate": 3.9546047173165816e-05, "loss": 1.2352, "step": 2710 }, { "epoch": 21.688, "grad_norm": 34.11921310424805, "learning_rate": 3.95401208960531e-05, "loss": 1.5313, "step": 2711 }, { "epoch": 21.696, "grad_norm": 49.865257263183594, "learning_rate": 3.953419461894039e-05, "loss": 1.5759, "step": 2712 }, { "epoch": 21.704, "grad_norm": 31.193729400634766, "learning_rate": 3.9528268341827666e-05, "loss": 1.9923, "step": 2713 }, { "epoch": 21.712, "grad_norm": 17.92323875427246, "learning_rate": 3.9522342064714944e-05, "loss": 2.1372, "step": 2714 }, { "epoch": 21.72, "grad_norm": 63.10919952392578, "learning_rate": 3.951641578760223e-05, "loss": 1.3986, "step": 2715 }, { "epoch": 21.728, "grad_norm": 19.467992782592773, "learning_rate": 3.9510489510489516e-05, "loss": 1.1972, "step": 2716 }, { "epoch": 21.736, "grad_norm": 32.86369323730469, "learning_rate": 3.9504563233376794e-05, "loss": 1.7137, "step": 2717 }, { "epoch": 21.744, "grad_norm": 27.80095672607422, "learning_rate": 3.949863695626408e-05, "loss": 1.345, "step": 2718 }, { "epoch": 21.752, "grad_norm": 19.132892608642578, "learning_rate": 3.949271067915136e-05, "loss": 1.6307, "step": 2719 }, { "epoch": 21.76, "grad_norm": 72.31013488769531, "learning_rate": 3.948678440203864e-05, "loss": 1.1545, "step": 2720 }, { "epoch": 21.768, "grad_norm": 23.23817253112793, "learning_rate": 3.948085812492592e-05, "loss": 1.2201, "step": 2721 }, { "epoch": 21.776, "grad_norm": 29.311311721801758, "learning_rate": 3.947493184781321e-05, "loss": 1.4119, "step": 2722 }, { "epoch": 21.784, "grad_norm": 48.21745300292969, "learning_rate": 3.946900557070049e-05, "loss": 1.6127, "step": 2723 }, { "epoch": 21.792, "grad_norm": 71.78904724121094, "learning_rate": 3.9463079293587766e-05, "loss": 0.8592, "step": 2724 }, { "epoch": 21.8, "grad_norm": 79.74588775634766, "learning_rate": 3.945715301647505e-05, "loss": 1.3766, "step": 2725 }, { "epoch": 21.808, "grad_norm": 34.06328582763672, "learning_rate": 3.945122673936234e-05, "loss": 1.2387, "step": 2726 }, { "epoch": 21.816, "grad_norm": 25.420427322387695, "learning_rate": 3.9445300462249616e-05, "loss": 1.0882, "step": 2727 }, { "epoch": 21.824, "grad_norm": 43.66666030883789, "learning_rate": 3.94393741851369e-05, "loss": 1.3671, "step": 2728 }, { "epoch": 21.832, "grad_norm": 24.32247543334961, "learning_rate": 3.943344790802418e-05, "loss": 1.411, "step": 2729 }, { "epoch": 21.84, "grad_norm": 51.43497848510742, "learning_rate": 3.942752163091146e-05, "loss": 1.2969, "step": 2730 }, { "epoch": 21.848, "grad_norm": 60.862648010253906, "learning_rate": 3.9421595353798745e-05, "loss": 1.3788, "step": 2731 }, { "epoch": 21.856, "grad_norm": 33.49512481689453, "learning_rate": 3.941566907668603e-05, "loss": 1.334, "step": 2732 }, { "epoch": 21.864, "grad_norm": 35.17912673950195, "learning_rate": 3.940974279957331e-05, "loss": 1.2893, "step": 2733 }, { "epoch": 21.872, "grad_norm": 36.56206130981445, "learning_rate": 3.940381652246059e-05, "loss": 1.2938, "step": 2734 }, { "epoch": 21.88, "grad_norm": 28.592077255249023, "learning_rate": 3.9397890245347874e-05, "loss": 1.5577, "step": 2735 }, { "epoch": 21.888, "grad_norm": 22.352046966552734, "learning_rate": 3.939196396823516e-05, "loss": 0.885, "step": 2736 }, { "epoch": 21.896, "grad_norm": 46.96390914916992, "learning_rate": 3.938603769112244e-05, "loss": 1.1998, "step": 2737 }, { "epoch": 21.904, "grad_norm": 18.91257095336914, "learning_rate": 3.938011141400972e-05, "loss": 1.2439, "step": 2738 }, { "epoch": 21.912, "grad_norm": 30.504070281982422, "learning_rate": 3.9374185136897e-05, "loss": 1.1851, "step": 2739 }, { "epoch": 21.92, "grad_norm": 32.585060119628906, "learning_rate": 3.936825885978428e-05, "loss": 1.1625, "step": 2740 }, { "epoch": 21.928, "grad_norm": 38.61286544799805, "learning_rate": 3.936233258267157e-05, "loss": 1.4109, "step": 2741 }, { "epoch": 21.936, "grad_norm": 33.354557037353516, "learning_rate": 3.935640630555885e-05, "loss": 1.3766, "step": 2742 }, { "epoch": 21.944, "grad_norm": 24.621774673461914, "learning_rate": 3.935048002844613e-05, "loss": 1.3144, "step": 2743 }, { "epoch": 21.951999999999998, "grad_norm": 28.820127487182617, "learning_rate": 3.934455375133341e-05, "loss": 1.8884, "step": 2744 }, { "epoch": 21.96, "grad_norm": 58.2872200012207, "learning_rate": 3.9338627474220696e-05, "loss": 1.2304, "step": 2745 }, { "epoch": 21.968, "grad_norm": 15.862359046936035, "learning_rate": 3.933270119710798e-05, "loss": 1.2339, "step": 2746 }, { "epoch": 21.976, "grad_norm": 16.317705154418945, "learning_rate": 3.932677491999526e-05, "loss": 1.0684, "step": 2747 }, { "epoch": 21.984, "grad_norm": 27.929277420043945, "learning_rate": 3.932084864288254e-05, "loss": 1.7201, "step": 2748 }, { "epoch": 21.992, "grad_norm": 21.721773147583008, "learning_rate": 3.9314922365769825e-05, "loss": 1.08, "step": 2749 }, { "epoch": 22.0, "grad_norm": 29.220314025878906, "learning_rate": 3.930899608865711e-05, "loss": 0.8223, "step": 2750 }, { "epoch": 22.0, "eval_loss": 1.3938605785369873, "eval_map": 0.2523, "eval_map_50": 0.543, "eval_map_75": 0.1845, "eval_map_Coverall": 0.4121, "eval_map_Face_Shield": 0.2024, "eval_map_Gloves": 0.2117, "eval_map_Goggles": 0.0962, "eval_map_Mask": 0.3388, "eval_map_large": 0.3171, "eval_map_medium": 0.1886, "eval_map_small": 0.1035, "eval_mar_1": 0.2439, "eval_mar_10": 0.4202, "eval_mar_100": 0.439, "eval_mar_100_Coverall": 0.6044, "eval_mar_100_Face_Shield": 0.4529, "eval_mar_100_Gloves": 0.382, "eval_mar_100_Goggles": 0.3344, "eval_mar_100_Mask": 0.4212, "eval_mar_large": 0.5295, "eval_mar_medium": 0.3619, "eval_mar_small": 0.1605, "eval_runtime": 1.8189, "eval_samples_per_second": 15.944, "eval_steps_per_second": 1.1, "step": 2750 }, { "epoch": 22.008, "grad_norm": 38.21408462524414, "learning_rate": 3.930306981154439e-05, "loss": 1.3673, "step": 2751 }, { "epoch": 22.016, "grad_norm": 51.28593826293945, "learning_rate": 3.9297143534431675e-05, "loss": 1.284, "step": 2752 }, { "epoch": 22.024, "grad_norm": 31.555654525756836, "learning_rate": 3.9291217257318954e-05, "loss": 1.274, "step": 2753 }, { "epoch": 22.032, "grad_norm": 20.628704071044922, "learning_rate": 3.928529098020623e-05, "loss": 1.514, "step": 2754 }, { "epoch": 22.04, "grad_norm": 31.51971435546875, "learning_rate": 3.927936470309352e-05, "loss": 1.3861, "step": 2755 }, { "epoch": 22.048, "grad_norm": 17.725919723510742, "learning_rate": 3.9273438425980804e-05, "loss": 1.2235, "step": 2756 }, { "epoch": 22.056, "grad_norm": 21.849863052368164, "learning_rate": 3.926751214886808e-05, "loss": 0.9996, "step": 2757 }, { "epoch": 22.064, "grad_norm": 24.808671951293945, "learning_rate": 3.926158587175536e-05, "loss": 1.6838, "step": 2758 }, { "epoch": 22.072, "grad_norm": 29.242586135864258, "learning_rate": 3.925565959464265e-05, "loss": 1.6087, "step": 2759 }, { "epoch": 22.08, "grad_norm": 27.83601188659668, "learning_rate": 3.924973331752993e-05, "loss": 1.2095, "step": 2760 }, { "epoch": 22.088, "grad_norm": 16.047119140625, "learning_rate": 3.924380704041721e-05, "loss": 1.1965, "step": 2761 }, { "epoch": 22.096, "grad_norm": 21.172237396240234, "learning_rate": 3.92378807633045e-05, "loss": 0.9955, "step": 2762 }, { "epoch": 22.104, "grad_norm": 133.13156127929688, "learning_rate": 3.9231954486191776e-05, "loss": 0.9874, "step": 2763 }, { "epoch": 22.112, "grad_norm": 53.79939651489258, "learning_rate": 3.9226028209079055e-05, "loss": 1.3391, "step": 2764 }, { "epoch": 22.12, "grad_norm": 17.469573974609375, "learning_rate": 3.922010193196634e-05, "loss": 1.0794, "step": 2765 }, { "epoch": 22.128, "grad_norm": 28.797527313232422, "learning_rate": 3.9214175654853626e-05, "loss": 1.518, "step": 2766 }, { "epoch": 22.136, "grad_norm": 41.6311149597168, "learning_rate": 3.9208249377740905e-05, "loss": 1.8505, "step": 2767 }, { "epoch": 22.144, "grad_norm": 16.419954299926758, "learning_rate": 3.9202323100628184e-05, "loss": 1.3613, "step": 2768 }, { "epoch": 22.152, "grad_norm": 46.640785217285156, "learning_rate": 3.919639682351547e-05, "loss": 1.0919, "step": 2769 }, { "epoch": 22.16, "grad_norm": 27.324844360351562, "learning_rate": 3.9190470546402755e-05, "loss": 1.4286, "step": 2770 }, { "epoch": 22.168, "grad_norm": 27.70271110534668, "learning_rate": 3.9184544269290034e-05, "loss": 0.947, "step": 2771 }, { "epoch": 22.176, "grad_norm": 48.34519958496094, "learning_rate": 3.917861799217731e-05, "loss": 1.2834, "step": 2772 }, { "epoch": 22.184, "grad_norm": 24.98556137084961, "learning_rate": 3.91726917150646e-05, "loss": 1.4188, "step": 2773 }, { "epoch": 22.192, "grad_norm": 36.93457794189453, "learning_rate": 3.916676543795188e-05, "loss": 1.4228, "step": 2774 }, { "epoch": 22.2, "grad_norm": 29.513986587524414, "learning_rate": 3.916083916083916e-05, "loss": 1.4787, "step": 2775 }, { "epoch": 22.208, "grad_norm": 38.721229553222656, "learning_rate": 3.915491288372645e-05, "loss": 1.2716, "step": 2776 }, { "epoch": 22.216, "grad_norm": 27.84457015991211, "learning_rate": 3.914898660661373e-05, "loss": 1.1598, "step": 2777 }, { "epoch": 22.224, "grad_norm": 41.847713470458984, "learning_rate": 3.9143060329501006e-05, "loss": 0.9646, "step": 2778 }, { "epoch": 22.232, "grad_norm": 19.92191505432129, "learning_rate": 3.913713405238829e-05, "loss": 1.2505, "step": 2779 }, { "epoch": 22.24, "grad_norm": 38.2964973449707, "learning_rate": 3.913120777527558e-05, "loss": 1.3535, "step": 2780 }, { "epoch": 22.248, "grad_norm": 36.2984619140625, "learning_rate": 3.9125281498162856e-05, "loss": 1.6436, "step": 2781 }, { "epoch": 22.256, "grad_norm": 17.080020904541016, "learning_rate": 3.9119355221050134e-05, "loss": 1.2974, "step": 2782 }, { "epoch": 22.264, "grad_norm": 25.71816062927246, "learning_rate": 3.911342894393742e-05, "loss": 1.2016, "step": 2783 }, { "epoch": 22.272, "grad_norm": 15.601407051086426, "learning_rate": 3.9107502666824706e-05, "loss": 1.541, "step": 2784 }, { "epoch": 22.28, "grad_norm": 24.515838623046875, "learning_rate": 3.9101576389711985e-05, "loss": 1.3909, "step": 2785 }, { "epoch": 22.288, "grad_norm": 34.05758285522461, "learning_rate": 3.909565011259927e-05, "loss": 1.159, "step": 2786 }, { "epoch": 22.296, "grad_norm": 21.58578109741211, "learning_rate": 3.908972383548655e-05, "loss": 2.6186, "step": 2787 }, { "epoch": 22.304, "grad_norm": 86.37124633789062, "learning_rate": 3.908379755837383e-05, "loss": 1.2677, "step": 2788 }, { "epoch": 22.312, "grad_norm": 90.23856353759766, "learning_rate": 3.907787128126111e-05, "loss": 1.527, "step": 2789 }, { "epoch": 22.32, "grad_norm": 30.89533233642578, "learning_rate": 3.90719450041484e-05, "loss": 1.1053, "step": 2790 }, { "epoch": 22.328, "grad_norm": 202.981201171875, "learning_rate": 3.906601872703568e-05, "loss": 1.597, "step": 2791 }, { "epoch": 22.336, "grad_norm": 16.223876953125, "learning_rate": 3.9060092449922957e-05, "loss": 2.7315, "step": 2792 }, { "epoch": 22.344, "grad_norm": 16.709556579589844, "learning_rate": 3.905416617281024e-05, "loss": 1.4282, "step": 2793 }, { "epoch": 22.352, "grad_norm": 27.141788482666016, "learning_rate": 3.904823989569753e-05, "loss": 1.2301, "step": 2794 }, { "epoch": 22.36, "grad_norm": 111.17371368408203, "learning_rate": 3.9042313618584807e-05, "loss": 1.2048, "step": 2795 }, { "epoch": 22.368, "grad_norm": 33.69668197631836, "learning_rate": 3.903638734147209e-05, "loss": 1.1068, "step": 2796 }, { "epoch": 22.376, "grad_norm": 23.920629501342773, "learning_rate": 3.903046106435937e-05, "loss": 1.1274, "step": 2797 }, { "epoch": 22.384, "grad_norm": 27.165985107421875, "learning_rate": 3.902453478724665e-05, "loss": 1.3062, "step": 2798 }, { "epoch": 22.392, "grad_norm": 48.603004455566406, "learning_rate": 3.9018608510133935e-05, "loss": 1.4327, "step": 2799 }, { "epoch": 22.4, "grad_norm": 38.71613311767578, "learning_rate": 3.901268223302122e-05, "loss": 1.4795, "step": 2800 }, { "epoch": 22.408, "grad_norm": 229.78619384765625, "learning_rate": 3.90067559559085e-05, "loss": 2.1619, "step": 2801 }, { "epoch": 22.416, "grad_norm": 57.21892547607422, "learning_rate": 3.900082967879578e-05, "loss": 1.1184, "step": 2802 }, { "epoch": 22.424, "grad_norm": 19.788541793823242, "learning_rate": 3.8994903401683064e-05, "loss": 1.3631, "step": 2803 }, { "epoch": 22.432, "grad_norm": 42.61720657348633, "learning_rate": 3.898897712457035e-05, "loss": 0.9926, "step": 2804 }, { "epoch": 22.44, "grad_norm": 15.488974571228027, "learning_rate": 3.898305084745763e-05, "loss": 1.1674, "step": 2805 }, { "epoch": 22.448, "grad_norm": 39.22461700439453, "learning_rate": 3.8977124570344914e-05, "loss": 1.0147, "step": 2806 }, { "epoch": 22.456, "grad_norm": 23.371002197265625, "learning_rate": 3.897119829323219e-05, "loss": 1.1908, "step": 2807 }, { "epoch": 22.464, "grad_norm": 58.50806427001953, "learning_rate": 3.896527201611947e-05, "loss": 1.5014, "step": 2808 }, { "epoch": 22.472, "grad_norm": 31.35214614868164, "learning_rate": 3.895934573900676e-05, "loss": 1.4923, "step": 2809 }, { "epoch": 22.48, "grad_norm": 31.654436111450195, "learning_rate": 3.895341946189404e-05, "loss": 1.5575, "step": 2810 }, { "epoch": 22.488, "grad_norm": 20.025774002075195, "learning_rate": 3.894749318478132e-05, "loss": 1.443, "step": 2811 }, { "epoch": 22.496, "grad_norm": 45.23146057128906, "learning_rate": 3.89415669076686e-05, "loss": 1.2957, "step": 2812 }, { "epoch": 22.504, "grad_norm": 34.19441604614258, "learning_rate": 3.8935640630555886e-05, "loss": 1.3242, "step": 2813 }, { "epoch": 22.512, "grad_norm": 24.03110694885254, "learning_rate": 3.892971435344317e-05, "loss": 1.1983, "step": 2814 }, { "epoch": 22.52, "grad_norm": 20.09260368347168, "learning_rate": 3.892378807633045e-05, "loss": 1.2406, "step": 2815 }, { "epoch": 22.528, "grad_norm": 38.59563446044922, "learning_rate": 3.891786179921773e-05, "loss": 1.6002, "step": 2816 }, { "epoch": 22.536, "grad_norm": 96.85535430908203, "learning_rate": 3.8911935522105015e-05, "loss": 1.4474, "step": 2817 }, { "epoch": 22.544, "grad_norm": 42.78696060180664, "learning_rate": 3.89060092449923e-05, "loss": 1.5754, "step": 2818 }, { "epoch": 22.552, "grad_norm": 27.614818572998047, "learning_rate": 3.890008296787958e-05, "loss": 1.5443, "step": 2819 }, { "epoch": 22.56, "grad_norm": 23.488985061645508, "learning_rate": 3.8894156690766865e-05, "loss": 1.0649, "step": 2820 }, { "epoch": 22.568, "grad_norm": 19.607126235961914, "learning_rate": 3.8888230413654144e-05, "loss": 1.0509, "step": 2821 }, { "epoch": 22.576, "grad_norm": 53.542335510253906, "learning_rate": 3.888230413654142e-05, "loss": 1.6353, "step": 2822 }, { "epoch": 22.584, "grad_norm": 48.34862518310547, "learning_rate": 3.887637785942871e-05, "loss": 1.2991, "step": 2823 }, { "epoch": 22.592, "grad_norm": 22.567964553833008, "learning_rate": 3.8870451582315994e-05, "loss": 1.1862, "step": 2824 }, { "epoch": 22.6, "grad_norm": 23.43340492248535, "learning_rate": 3.886452530520327e-05, "loss": 1.2129, "step": 2825 }, { "epoch": 22.608, "grad_norm": 37.089576721191406, "learning_rate": 3.885859902809055e-05, "loss": 1.884, "step": 2826 }, { "epoch": 22.616, "grad_norm": 30.44918441772461, "learning_rate": 3.885267275097784e-05, "loss": 1.6631, "step": 2827 }, { "epoch": 22.624, "grad_norm": 26.92423439025879, "learning_rate": 3.884674647386512e-05, "loss": 1.2678, "step": 2828 }, { "epoch": 22.632, "grad_norm": 31.814807891845703, "learning_rate": 3.88408201967524e-05, "loss": 1.3683, "step": 2829 }, { "epoch": 22.64, "grad_norm": 28.592758178710938, "learning_rate": 3.883489391963969e-05, "loss": 1.3987, "step": 2830 }, { "epoch": 22.648, "grad_norm": 40.534236907958984, "learning_rate": 3.8828967642526966e-05, "loss": 1.0588, "step": 2831 }, { "epoch": 22.656, "grad_norm": 17.844356536865234, "learning_rate": 3.8823041365414245e-05, "loss": 2.0884, "step": 2832 }, { "epoch": 22.664, "grad_norm": 38.78521728515625, "learning_rate": 3.881711508830153e-05, "loss": 1.9647, "step": 2833 }, { "epoch": 22.672, "grad_norm": 33.05083465576172, "learning_rate": 3.8811188811188816e-05, "loss": 1.9145, "step": 2834 }, { "epoch": 22.68, "grad_norm": 55.858699798583984, "learning_rate": 3.8805262534076095e-05, "loss": 1.1985, "step": 2835 }, { "epoch": 22.688, "grad_norm": 27.268156051635742, "learning_rate": 3.8799336256963374e-05, "loss": 1.3661, "step": 2836 }, { "epoch": 22.696, "grad_norm": 20.8428897857666, "learning_rate": 3.879340997985066e-05, "loss": 1.043, "step": 2837 }, { "epoch": 22.704, "grad_norm": 41.96402359008789, "learning_rate": 3.8787483702737945e-05, "loss": 1.4243, "step": 2838 }, { "epoch": 22.712, "grad_norm": 63.723350524902344, "learning_rate": 3.8781557425625224e-05, "loss": 1.244, "step": 2839 }, { "epoch": 22.72, "grad_norm": 17.094751358032227, "learning_rate": 3.877563114851251e-05, "loss": 1.3117, "step": 2840 }, { "epoch": 22.728, "grad_norm": 22.218685150146484, "learning_rate": 3.876970487139979e-05, "loss": 1.0321, "step": 2841 }, { "epoch": 22.736, "grad_norm": 33.66652297973633, "learning_rate": 3.876377859428707e-05, "loss": 1.2257, "step": 2842 }, { "epoch": 22.744, "grad_norm": 15.31194019317627, "learning_rate": 3.875785231717435e-05, "loss": 1.4024, "step": 2843 }, { "epoch": 22.752, "grad_norm": 21.096946716308594, "learning_rate": 3.875192604006164e-05, "loss": 1.4225, "step": 2844 }, { "epoch": 22.76, "grad_norm": 24.156314849853516, "learning_rate": 3.874599976294892e-05, "loss": 1.2267, "step": 2845 }, { "epoch": 22.768, "grad_norm": 18.890342712402344, "learning_rate": 3.8740073485836196e-05, "loss": 0.9296, "step": 2846 }, { "epoch": 22.776, "grad_norm": 50.182369232177734, "learning_rate": 3.873414720872348e-05, "loss": 1.5171, "step": 2847 }, { "epoch": 22.784, "grad_norm": 33.062591552734375, "learning_rate": 3.872822093161077e-05, "loss": 1.3036, "step": 2848 }, { "epoch": 22.792, "grad_norm": 39.76524353027344, "learning_rate": 3.8722294654498046e-05, "loss": 1.4796, "step": 2849 }, { "epoch": 22.8, "grad_norm": 48.4266471862793, "learning_rate": 3.8716368377385325e-05, "loss": 1.2555, "step": 2850 }, { "epoch": 22.808, "grad_norm": 33.48597717285156, "learning_rate": 3.871044210027261e-05, "loss": 1.2996, "step": 2851 }, { "epoch": 22.816, "grad_norm": 20.16936492919922, "learning_rate": 3.8704515823159896e-05, "loss": 1.247, "step": 2852 }, { "epoch": 22.824, "grad_norm": 38.114810943603516, "learning_rate": 3.8698589546047175e-05, "loss": 1.4016, "step": 2853 }, { "epoch": 22.832, "grad_norm": 30.991193771362305, "learning_rate": 3.869266326893446e-05, "loss": 1.3679, "step": 2854 }, { "epoch": 22.84, "grad_norm": 15.997693061828613, "learning_rate": 3.868673699182174e-05, "loss": 1.1441, "step": 2855 }, { "epoch": 22.848, "grad_norm": 24.716720581054688, "learning_rate": 3.868081071470902e-05, "loss": 0.9624, "step": 2856 }, { "epoch": 22.856, "grad_norm": 23.594268798828125, "learning_rate": 3.8674884437596303e-05, "loss": 1.3873, "step": 2857 }, { "epoch": 22.864, "grad_norm": 63.52898025512695, "learning_rate": 3.866895816048359e-05, "loss": 1.6394, "step": 2858 }, { "epoch": 22.872, "grad_norm": 72.21688079833984, "learning_rate": 3.866303188337087e-05, "loss": 1.1272, "step": 2859 }, { "epoch": 22.88, "grad_norm": 42.73821258544922, "learning_rate": 3.865710560625815e-05, "loss": 1.0099, "step": 2860 }, { "epoch": 22.888, "grad_norm": 32.23469161987305, "learning_rate": 3.865117932914543e-05, "loss": 1.4306, "step": 2861 }, { "epoch": 22.896, "grad_norm": 24.110401153564453, "learning_rate": 3.864525305203272e-05, "loss": 1.1916, "step": 2862 }, { "epoch": 22.904, "grad_norm": 58.12630081176758, "learning_rate": 3.863932677492e-05, "loss": 0.8817, "step": 2863 }, { "epoch": 22.912, "grad_norm": 23.09487533569336, "learning_rate": 3.863340049780728e-05, "loss": 1.4344, "step": 2864 }, { "epoch": 22.92, "grad_norm": 31.101045608520508, "learning_rate": 3.862747422069456e-05, "loss": 2.1963, "step": 2865 }, { "epoch": 22.928, "grad_norm": 23.378170013427734, "learning_rate": 3.862154794358184e-05, "loss": 1.0158, "step": 2866 }, { "epoch": 22.936, "grad_norm": 48.320655822753906, "learning_rate": 3.8615621666469125e-05, "loss": 1.3071, "step": 2867 }, { "epoch": 22.944, "grad_norm": 32.44013214111328, "learning_rate": 3.860969538935641e-05, "loss": 1.3977, "step": 2868 }, { "epoch": 22.951999999999998, "grad_norm": 30.992733001708984, "learning_rate": 3.860376911224369e-05, "loss": 1.1361, "step": 2869 }, { "epoch": 22.96, "grad_norm": 29.880475997924805, "learning_rate": 3.859784283513097e-05, "loss": 1.5825, "step": 2870 }, { "epoch": 22.968, "grad_norm": 13.885811805725098, "learning_rate": 3.8591916558018254e-05, "loss": 1.4175, "step": 2871 }, { "epoch": 22.976, "grad_norm": 27.491891860961914, "learning_rate": 3.858599028090554e-05, "loss": 0.8048, "step": 2872 }, { "epoch": 22.984, "grad_norm": 35.92876434326172, "learning_rate": 3.858006400379282e-05, "loss": 1.4879, "step": 2873 }, { "epoch": 22.992, "grad_norm": 26.507841110229492, "learning_rate": 3.8574137726680104e-05, "loss": 1.6634, "step": 2874 }, { "epoch": 23.0, "grad_norm": 29.369388580322266, "learning_rate": 3.856821144956738e-05, "loss": 1.0102, "step": 2875 }, { "epoch": 23.0, "eval_loss": 1.31263267993927, "eval_map": 0.2505, "eval_map_50": 0.5431, "eval_map_75": 0.1641, "eval_map_Coverall": 0.5115, "eval_map_Face_Shield": 0.1886, "eval_map_Gloves": 0.1854, "eval_map_Goggles": 0.1268, "eval_map_Mask": 0.2401, "eval_map_large": 0.4056, "eval_map_medium": 0.1796, "eval_map_small": 0.1316, "eval_mar_1": 0.2392, "eval_mar_10": 0.4387, "eval_mar_100": 0.4591, "eval_mar_100_Coverall": 0.6867, "eval_mar_100_Face_Shield": 0.5353, "eval_mar_100_Gloves": 0.3836, "eval_mar_100_Goggles": 0.3594, "eval_mar_100_Mask": 0.3308, "eval_mar_large": 0.6877, "eval_mar_medium": 0.3391, "eval_mar_small": 0.175, "eval_runtime": 1.8651, "eval_samples_per_second": 15.548, "eval_steps_per_second": 1.072, "step": 2875 }, { "epoch": 23.008, "grad_norm": 16.440589904785156, "learning_rate": 3.856228517245466e-05, "loss": 1.7441, "step": 2876 }, { "epoch": 23.016, "grad_norm": 15.624666213989258, "learning_rate": 3.855635889534195e-05, "loss": 1.1459, "step": 2877 }, { "epoch": 23.024, "grad_norm": 23.08249282836914, "learning_rate": 3.855043261822923e-05, "loss": 1.2921, "step": 2878 }, { "epoch": 23.032, "grad_norm": 24.541950225830078, "learning_rate": 3.854450634111651e-05, "loss": 1.5641, "step": 2879 }, { "epoch": 23.04, "grad_norm": 61.62240219116211, "learning_rate": 3.853858006400379e-05, "loss": 1.486, "step": 2880 }, { "epoch": 23.048, "grad_norm": 23.787086486816406, "learning_rate": 3.8532653786891076e-05, "loss": 1.0847, "step": 2881 }, { "epoch": 23.056, "grad_norm": 22.499223709106445, "learning_rate": 3.852672750977836e-05, "loss": 0.9732, "step": 2882 }, { "epoch": 23.064, "grad_norm": 29.96513557434082, "learning_rate": 3.852080123266564e-05, "loss": 1.0139, "step": 2883 }, { "epoch": 23.072, "grad_norm": 38.73072052001953, "learning_rate": 3.851487495555292e-05, "loss": 1.1107, "step": 2884 }, { "epoch": 23.08, "grad_norm": 49.88114929199219, "learning_rate": 3.8508948678440205e-05, "loss": 1.229, "step": 2885 }, { "epoch": 23.088, "grad_norm": 45.928428649902344, "learning_rate": 3.850302240132749e-05, "loss": 1.5874, "step": 2886 }, { "epoch": 23.096, "grad_norm": 43.13016128540039, "learning_rate": 3.849709612421477e-05, "loss": 1.362, "step": 2887 }, { "epoch": 23.104, "grad_norm": 23.788341522216797, "learning_rate": 3.8491169847102055e-05, "loss": 1.359, "step": 2888 }, { "epoch": 23.112, "grad_norm": 40.915653228759766, "learning_rate": 3.8485243569989334e-05, "loss": 1.3289, "step": 2889 }, { "epoch": 23.12, "grad_norm": 25.776445388793945, "learning_rate": 3.847931729287661e-05, "loss": 1.0663, "step": 2890 }, { "epoch": 23.128, "grad_norm": 25.72249984741211, "learning_rate": 3.84733910157639e-05, "loss": 1.1912, "step": 2891 }, { "epoch": 23.136, "grad_norm": 40.01628875732422, "learning_rate": 3.8467464738651184e-05, "loss": 1.2806, "step": 2892 }, { "epoch": 23.144, "grad_norm": 22.608640670776367, "learning_rate": 3.846153846153846e-05, "loss": 0.8797, "step": 2893 }, { "epoch": 23.152, "grad_norm": 31.35634994506836, "learning_rate": 3.845561218442574e-05, "loss": 1.4746, "step": 2894 }, { "epoch": 23.16, "grad_norm": 20.97101402282715, "learning_rate": 3.844968590731303e-05, "loss": 1.4433, "step": 2895 }, { "epoch": 23.168, "grad_norm": 38.49714279174805, "learning_rate": 3.844375963020031e-05, "loss": 1.7258, "step": 2896 }, { "epoch": 23.176, "grad_norm": 40.07431411743164, "learning_rate": 3.843783335308759e-05, "loss": 1.4058, "step": 2897 }, { "epoch": 23.184, "grad_norm": 21.62346839904785, "learning_rate": 3.843190707597488e-05, "loss": 1.3758, "step": 2898 }, { "epoch": 23.192, "grad_norm": 32.680084228515625, "learning_rate": 3.8425980798862156e-05, "loss": 1.3866, "step": 2899 }, { "epoch": 23.2, "grad_norm": 25.39842987060547, "learning_rate": 3.8420054521749435e-05, "loss": 1.2061, "step": 2900 }, { "epoch": 23.208, "grad_norm": 28.14711570739746, "learning_rate": 3.841412824463672e-05, "loss": 1.1085, "step": 2901 }, { "epoch": 23.216, "grad_norm": 23.504074096679688, "learning_rate": 3.8408201967524006e-05, "loss": 1.577, "step": 2902 }, { "epoch": 23.224, "grad_norm": 21.14130210876465, "learning_rate": 3.8402275690411285e-05, "loss": 1.2872, "step": 2903 }, { "epoch": 23.232, "grad_norm": 39.94953536987305, "learning_rate": 3.8396349413298564e-05, "loss": 1.0574, "step": 2904 }, { "epoch": 23.24, "grad_norm": 68.6928939819336, "learning_rate": 3.839042313618585e-05, "loss": 1.3737, "step": 2905 }, { "epoch": 23.248, "grad_norm": 25.74102210998535, "learning_rate": 3.8384496859073135e-05, "loss": 1.245, "step": 2906 }, { "epoch": 23.256, "grad_norm": 29.30389404296875, "learning_rate": 3.8378570581960414e-05, "loss": 1.4891, "step": 2907 }, { "epoch": 23.264, "grad_norm": 61.81850814819336, "learning_rate": 3.83726443048477e-05, "loss": 1.3615, "step": 2908 }, { "epoch": 23.272, "grad_norm": 28.373008728027344, "learning_rate": 3.836671802773498e-05, "loss": 1.176, "step": 2909 }, { "epoch": 23.28, "grad_norm": 30.290767669677734, "learning_rate": 3.836079175062226e-05, "loss": 1.1034, "step": 2910 }, { "epoch": 23.288, "grad_norm": 49.972137451171875, "learning_rate": 3.835486547350954e-05, "loss": 2.8556, "step": 2911 }, { "epoch": 23.296, "grad_norm": 34.53862380981445, "learning_rate": 3.834893919639683e-05, "loss": 1.5653, "step": 2912 }, { "epoch": 23.304, "grad_norm": 25.02128028869629, "learning_rate": 3.834301291928411e-05, "loss": 1.8216, "step": 2913 }, { "epoch": 23.312, "grad_norm": 98.5018310546875, "learning_rate": 3.8337086642171386e-05, "loss": 1.4003, "step": 2914 }, { "epoch": 23.32, "grad_norm": 110.62004089355469, "learning_rate": 3.833116036505867e-05, "loss": 1.4798, "step": 2915 }, { "epoch": 23.328, "grad_norm": 19.341957092285156, "learning_rate": 3.832523408794596e-05, "loss": 1.1441, "step": 2916 }, { "epoch": 23.336, "grad_norm": 28.154672622680664, "learning_rate": 3.8319307810833236e-05, "loss": 1.2446, "step": 2917 }, { "epoch": 23.344, "grad_norm": 20.91708755493164, "learning_rate": 3.8313381533720515e-05, "loss": 1.1851, "step": 2918 }, { "epoch": 23.352, "grad_norm": 42.46785354614258, "learning_rate": 3.83074552566078e-05, "loss": 1.3795, "step": 2919 }, { "epoch": 23.36, "grad_norm": 21.542577743530273, "learning_rate": 3.8301528979495086e-05, "loss": 0.8995, "step": 2920 }, { "epoch": 23.368, "grad_norm": 28.511329650878906, "learning_rate": 3.8295602702382365e-05, "loss": 1.2155, "step": 2921 }, { "epoch": 23.376, "grad_norm": 67.95337677001953, "learning_rate": 3.828967642526965e-05, "loss": 1.3968, "step": 2922 }, { "epoch": 23.384, "grad_norm": 36.685813903808594, "learning_rate": 3.828375014815693e-05, "loss": 1.6654, "step": 2923 }, { "epoch": 23.392, "grad_norm": 36.39916229248047, "learning_rate": 3.827782387104421e-05, "loss": 1.5065, "step": 2924 }, { "epoch": 23.4, "grad_norm": 27.57369613647461, "learning_rate": 3.8271897593931493e-05, "loss": 0.7339, "step": 2925 }, { "epoch": 23.408, "grad_norm": 32.02682876586914, "learning_rate": 3.826597131681878e-05, "loss": 1.4941, "step": 2926 }, { "epoch": 23.416, "grad_norm": 36.97500991821289, "learning_rate": 3.826004503970606e-05, "loss": 1.7975, "step": 2927 }, { "epoch": 23.424, "grad_norm": 40.2255859375, "learning_rate": 3.825411876259334e-05, "loss": 1.2091, "step": 2928 }, { "epoch": 23.432, "grad_norm": 16.852052688598633, "learning_rate": 3.824819248548062e-05, "loss": 1.523, "step": 2929 }, { "epoch": 23.44, "grad_norm": 39.015235900878906, "learning_rate": 3.824226620836791e-05, "loss": 1.4231, "step": 2930 }, { "epoch": 23.448, "grad_norm": 20.173372268676758, "learning_rate": 3.823633993125519e-05, "loss": 1.5012, "step": 2931 }, { "epoch": 23.456, "grad_norm": 28.464967727661133, "learning_rate": 3.823041365414247e-05, "loss": 1.1142, "step": 2932 }, { "epoch": 23.464, "grad_norm": 15.097392082214355, "learning_rate": 3.822448737702975e-05, "loss": 1.2648, "step": 2933 }, { "epoch": 23.472, "grad_norm": 25.903902053833008, "learning_rate": 3.821856109991703e-05, "loss": 1.2539, "step": 2934 }, { "epoch": 23.48, "grad_norm": 21.550491333007812, "learning_rate": 3.8212634822804316e-05, "loss": 1.3053, "step": 2935 }, { "epoch": 23.488, "grad_norm": 51.3222541809082, "learning_rate": 3.82067085456916e-05, "loss": 1.5375, "step": 2936 }, { "epoch": 23.496, "grad_norm": 18.7091007232666, "learning_rate": 3.820078226857888e-05, "loss": 1.3568, "step": 2937 }, { "epoch": 23.504, "grad_norm": 21.93474578857422, "learning_rate": 3.819485599146616e-05, "loss": 1.2832, "step": 2938 }, { "epoch": 23.512, "grad_norm": 16.669527053833008, "learning_rate": 3.8188929714353444e-05, "loss": 1.1344, "step": 2939 }, { "epoch": 23.52, "grad_norm": 20.30495834350586, "learning_rate": 3.818300343724073e-05, "loss": 1.228, "step": 2940 }, { "epoch": 23.528, "grad_norm": 26.82176971435547, "learning_rate": 3.817707716012801e-05, "loss": 1.1079, "step": 2941 }, { "epoch": 23.536, "grad_norm": 11.517854690551758, "learning_rate": 3.8171150883015294e-05, "loss": 1.1889, "step": 2942 }, { "epoch": 23.544, "grad_norm": 27.048845291137695, "learning_rate": 3.816522460590257e-05, "loss": 2.365, "step": 2943 }, { "epoch": 23.552, "grad_norm": 22.938518524169922, "learning_rate": 3.815929832878985e-05, "loss": 1.2723, "step": 2944 }, { "epoch": 23.56, "grad_norm": 15.723653793334961, "learning_rate": 3.815337205167714e-05, "loss": 1.0096, "step": 2945 }, { "epoch": 23.568, "grad_norm": 18.82154655456543, "learning_rate": 3.814744577456442e-05, "loss": 1.1123, "step": 2946 }, { "epoch": 23.576, "grad_norm": 23.06734275817871, "learning_rate": 3.81415194974517e-05, "loss": 0.9332, "step": 2947 }, { "epoch": 23.584, "grad_norm": 46.780975341796875, "learning_rate": 3.813559322033898e-05, "loss": 1.5644, "step": 2948 }, { "epoch": 23.592, "grad_norm": 19.577598571777344, "learning_rate": 3.8129666943226266e-05, "loss": 1.5036, "step": 2949 }, { "epoch": 23.6, "grad_norm": 25.5849666595459, "learning_rate": 3.812374066611355e-05, "loss": 1.2142, "step": 2950 }, { "epoch": 23.608, "grad_norm": 22.52874755859375, "learning_rate": 3.811781438900083e-05, "loss": 1.0862, "step": 2951 }, { "epoch": 23.616, "grad_norm": 17.13327407836914, "learning_rate": 3.811188811188811e-05, "loss": 1.0272, "step": 2952 }, { "epoch": 23.624, "grad_norm": 34.55685043334961, "learning_rate": 3.8105961834775395e-05, "loss": 1.1883, "step": 2953 }, { "epoch": 23.632, "grad_norm": 29.078800201416016, "learning_rate": 3.810003555766268e-05, "loss": 1.1541, "step": 2954 }, { "epoch": 23.64, "grad_norm": 19.788406372070312, "learning_rate": 3.809410928054996e-05, "loss": 1.8533, "step": 2955 }, { "epoch": 23.648, "grad_norm": 26.36920166015625, "learning_rate": 3.8088183003437245e-05, "loss": 1.7452, "step": 2956 }, { "epoch": 23.656, "grad_norm": 52.639617919921875, "learning_rate": 3.8082256726324524e-05, "loss": 1.8815, "step": 2957 }, { "epoch": 23.664, "grad_norm": 53.92191696166992, "learning_rate": 3.80763304492118e-05, "loss": 0.9862, "step": 2958 }, { "epoch": 23.672, "grad_norm": 18.91614532470703, "learning_rate": 3.807040417209909e-05, "loss": 1.0011, "step": 2959 }, { "epoch": 23.68, "grad_norm": 22.566720962524414, "learning_rate": 3.8064477894986374e-05, "loss": 1.061, "step": 2960 }, { "epoch": 23.688, "grad_norm": 20.423763275146484, "learning_rate": 3.805855161787365e-05, "loss": 0.9735, "step": 2961 }, { "epoch": 23.696, "grad_norm": 39.14419174194336, "learning_rate": 3.805262534076093e-05, "loss": 1.8662, "step": 2962 }, { "epoch": 23.704, "grad_norm": 28.29563331604004, "learning_rate": 3.804669906364822e-05, "loss": 2.7449, "step": 2963 }, { "epoch": 23.712, "grad_norm": 17.945035934448242, "learning_rate": 3.80407727865355e-05, "loss": 1.4073, "step": 2964 }, { "epoch": 23.72, "grad_norm": 19.966575622558594, "learning_rate": 3.803484650942278e-05, "loss": 1.4709, "step": 2965 }, { "epoch": 23.728, "grad_norm": 16.704633712768555, "learning_rate": 3.802892023231007e-05, "loss": 1.2954, "step": 2966 }, { "epoch": 23.736, "grad_norm": 33.121726989746094, "learning_rate": 3.8022993955197346e-05, "loss": 1.251, "step": 2967 }, { "epoch": 23.744, "grad_norm": 33.3961067199707, "learning_rate": 3.8017067678084625e-05, "loss": 1.3676, "step": 2968 }, { "epoch": 23.752, "grad_norm": 34.652442932128906, "learning_rate": 3.801114140097191e-05, "loss": 1.2222, "step": 2969 }, { "epoch": 23.76, "grad_norm": 36.134857177734375, "learning_rate": 3.8005215123859196e-05, "loss": 2.5594, "step": 2970 }, { "epoch": 23.768, "grad_norm": 29.744741439819336, "learning_rate": 3.7999288846746475e-05, "loss": 1.0264, "step": 2971 }, { "epoch": 23.776, "grad_norm": 36.959922790527344, "learning_rate": 3.7993362569633754e-05, "loss": 0.9454, "step": 2972 }, { "epoch": 23.784, "grad_norm": 24.445127487182617, "learning_rate": 3.798743629252104e-05, "loss": 1.6414, "step": 2973 }, { "epoch": 23.792, "grad_norm": 18.064496994018555, "learning_rate": 3.7981510015408325e-05, "loss": 1.1646, "step": 2974 }, { "epoch": 23.8, "grad_norm": 24.758312225341797, "learning_rate": 3.7975583738295604e-05, "loss": 1.7443, "step": 2975 }, { "epoch": 23.808, "grad_norm": 27.96967124938965, "learning_rate": 3.796965746118289e-05, "loss": 1.3053, "step": 2976 }, { "epoch": 23.816, "grad_norm": 18.18918228149414, "learning_rate": 3.796373118407017e-05, "loss": 1.1282, "step": 2977 }, { "epoch": 23.824, "grad_norm": 35.420623779296875, "learning_rate": 3.795780490695745e-05, "loss": 1.1771, "step": 2978 }, { "epoch": 23.832, "grad_norm": 24.87523651123047, "learning_rate": 3.795187862984473e-05, "loss": 1.1625, "step": 2979 }, { "epoch": 23.84, "grad_norm": 22.21480369567871, "learning_rate": 3.794595235273202e-05, "loss": 1.2498, "step": 2980 }, { "epoch": 23.848, "grad_norm": 27.31482696533203, "learning_rate": 3.79400260756193e-05, "loss": 0.9603, "step": 2981 }, { "epoch": 23.856, "grad_norm": 47.069183349609375, "learning_rate": 3.7934099798506576e-05, "loss": 2.1082, "step": 2982 }, { "epoch": 23.864, "grad_norm": 17.169504165649414, "learning_rate": 3.792817352139386e-05, "loss": 0.9681, "step": 2983 }, { "epoch": 23.872, "grad_norm": 31.6928653717041, "learning_rate": 3.792224724428115e-05, "loss": 0.7693, "step": 2984 }, { "epoch": 23.88, "grad_norm": 26.465246200561523, "learning_rate": 3.7916320967168426e-05, "loss": 1.1809, "step": 2985 }, { "epoch": 23.888, "grad_norm": 29.0456485748291, "learning_rate": 3.7910394690055705e-05, "loss": 1.4536, "step": 2986 }, { "epoch": 23.896, "grad_norm": 17.020843505859375, "learning_rate": 3.790446841294299e-05, "loss": 1.1003, "step": 2987 }, { "epoch": 23.904, "grad_norm": 27.551855087280273, "learning_rate": 3.7898542135830276e-05, "loss": 1.5414, "step": 2988 }, { "epoch": 23.912, "grad_norm": 56.82477569580078, "learning_rate": 3.7892615858717555e-05, "loss": 1.2636, "step": 2989 }, { "epoch": 23.92, "grad_norm": 52.905757904052734, "learning_rate": 3.788668958160484e-05, "loss": 1.2051, "step": 2990 }, { "epoch": 23.928, "grad_norm": 24.163471221923828, "learning_rate": 3.788076330449212e-05, "loss": 1.0562, "step": 2991 }, { "epoch": 23.936, "grad_norm": 15.14578628540039, "learning_rate": 3.78748370273794e-05, "loss": 1.2271, "step": 2992 }, { "epoch": 23.944, "grad_norm": 35.857276916503906, "learning_rate": 3.7868910750266684e-05, "loss": 1.002, "step": 2993 }, { "epoch": 23.951999999999998, "grad_norm": 32.22053527832031, "learning_rate": 3.786298447315397e-05, "loss": 1.5018, "step": 2994 }, { "epoch": 23.96, "grad_norm": 42.54901885986328, "learning_rate": 3.785705819604125e-05, "loss": 1.3036, "step": 2995 }, { "epoch": 23.968, "grad_norm": 12.956634521484375, "learning_rate": 3.785113191892853e-05, "loss": 1.3091, "step": 2996 }, { "epoch": 23.976, "grad_norm": 46.70847702026367, "learning_rate": 3.784520564181581e-05, "loss": 1.0257, "step": 2997 }, { "epoch": 23.984, "grad_norm": 25.70252227783203, "learning_rate": 3.78392793647031e-05, "loss": 1.3344, "step": 2998 }, { "epoch": 23.992, "grad_norm": 28.42646598815918, "learning_rate": 3.783335308759038e-05, "loss": 1.4247, "step": 2999 }, { "epoch": 24.0, "grad_norm": 32.293636322021484, "learning_rate": 3.782742681047766e-05, "loss": 1.3406, "step": 3000 }, { "epoch": 24.0, "eval_loss": 1.2288641929626465, "eval_map": 0.3131, "eval_map_50": 0.6162, "eval_map_75": 0.2296, "eval_map_Coverall": 0.5029, "eval_map_Face_Shield": 0.3311, "eval_map_Gloves": 0.2115, "eval_map_Goggles": 0.1607, "eval_map_Mask": 0.359, "eval_map_large": 0.4126, "eval_map_medium": 0.2938, "eval_map_small": 0.2262, "eval_mar_1": 0.2679, "eval_mar_10": 0.4936, "eval_mar_100": 0.515, "eval_mar_100_Coverall": 0.6844, "eval_mar_100_Face_Shield": 0.6176, "eval_mar_100_Gloves": 0.4066, "eval_mar_100_Goggles": 0.4125, "eval_mar_100_Mask": 0.4538, "eval_mar_large": 0.635, "eval_mar_medium": 0.4791, "eval_mar_small": 0.2952, "eval_runtime": 1.832, "eval_samples_per_second": 15.83, "eval_steps_per_second": 1.092, "step": 3000 }, { "epoch": 24.008, "grad_norm": 41.962223052978516, "learning_rate": 3.782150053336494e-05, "loss": 1.1526, "step": 3001 }, { "epoch": 24.016, "grad_norm": 47.48983383178711, "learning_rate": 3.781557425625222e-05, "loss": 1.1678, "step": 3002 }, { "epoch": 24.024, "grad_norm": 35.768035888671875, "learning_rate": 3.7809647979139506e-05, "loss": 1.3615, "step": 3003 }, { "epoch": 24.032, "grad_norm": 19.335098266601562, "learning_rate": 3.780372170202679e-05, "loss": 1.1318, "step": 3004 }, { "epoch": 24.04, "grad_norm": 76.99103546142578, "learning_rate": 3.779779542491407e-05, "loss": 0.884, "step": 3005 }, { "epoch": 24.048, "grad_norm": 80.20285034179688, "learning_rate": 3.779186914780135e-05, "loss": 2.01, "step": 3006 }, { "epoch": 24.056, "grad_norm": 26.1405029296875, "learning_rate": 3.7785942870688634e-05, "loss": 0.7985, "step": 3007 }, { "epoch": 24.064, "grad_norm": 24.126121520996094, "learning_rate": 3.778001659357592e-05, "loss": 0.9862, "step": 3008 }, { "epoch": 24.072, "grad_norm": 27.607749938964844, "learning_rate": 3.77740903164632e-05, "loss": 1.2673, "step": 3009 }, { "epoch": 24.08, "grad_norm": 19.94792938232422, "learning_rate": 3.7768164039350484e-05, "loss": 1.3789, "step": 3010 }, { "epoch": 24.088, "grad_norm": 25.586353302001953, "learning_rate": 3.776223776223776e-05, "loss": 1.4252, "step": 3011 }, { "epoch": 24.096, "grad_norm": 15.468131065368652, "learning_rate": 3.775631148512504e-05, "loss": 1.5816, "step": 3012 }, { "epoch": 24.104, "grad_norm": 31.882944107055664, "learning_rate": 3.775038520801233e-05, "loss": 1.1438, "step": 3013 }, { "epoch": 24.112, "grad_norm": 22.7651309967041, "learning_rate": 3.774445893089961e-05, "loss": 1.3959, "step": 3014 }, { "epoch": 24.12, "grad_norm": 24.087234497070312, "learning_rate": 3.77385326537869e-05, "loss": 1.0494, "step": 3015 }, { "epoch": 24.128, "grad_norm": 17.358543395996094, "learning_rate": 3.773260637667417e-05, "loss": 1.1908, "step": 3016 }, { "epoch": 24.136, "grad_norm": 22.210227966308594, "learning_rate": 3.7726680099561457e-05, "loss": 1.92, "step": 3017 }, { "epoch": 24.144, "grad_norm": 17.563676834106445, "learning_rate": 3.772075382244874e-05, "loss": 1.0945, "step": 3018 }, { "epoch": 24.152, "grad_norm": 57.3486328125, "learning_rate": 3.771482754533602e-05, "loss": 1.1445, "step": 3019 }, { "epoch": 24.16, "grad_norm": 105.44314575195312, "learning_rate": 3.77089012682233e-05, "loss": 1.2803, "step": 3020 }, { "epoch": 24.168, "grad_norm": 16.329151153564453, "learning_rate": 3.7702974991110585e-05, "loss": 1.2757, "step": 3021 }, { "epoch": 24.176, "grad_norm": 25.686382293701172, "learning_rate": 3.769704871399787e-05, "loss": 1.2932, "step": 3022 }, { "epoch": 24.184, "grad_norm": 36.102508544921875, "learning_rate": 3.769112243688515e-05, "loss": 1.0622, "step": 3023 }, { "epoch": 24.192, "grad_norm": 27.25140953063965, "learning_rate": 3.7685196159772435e-05, "loss": 2.412, "step": 3024 }, { "epoch": 24.2, "grad_norm": 30.804975509643555, "learning_rate": 3.7679269882659714e-05, "loss": 0.8635, "step": 3025 }, { "epoch": 24.208, "grad_norm": 18.42919921875, "learning_rate": 3.767334360554699e-05, "loss": 1.2089, "step": 3026 }, { "epoch": 24.216, "grad_norm": 50.164344787597656, "learning_rate": 3.766741732843428e-05, "loss": 0.9364, "step": 3027 }, { "epoch": 24.224, "grad_norm": 36.63404846191406, "learning_rate": 3.7661491051321564e-05, "loss": 1.3108, "step": 3028 }, { "epoch": 24.232, "grad_norm": 18.04081916809082, "learning_rate": 3.765556477420884e-05, "loss": 1.3844, "step": 3029 }, { "epoch": 24.24, "grad_norm": 22.06305503845215, "learning_rate": 3.764963849709612e-05, "loss": 1.4232, "step": 3030 }, { "epoch": 24.248, "grad_norm": 30.995302200317383, "learning_rate": 3.764371221998341e-05, "loss": 1.6611, "step": 3031 }, { "epoch": 24.256, "grad_norm": 32.44347381591797, "learning_rate": 3.763778594287069e-05, "loss": 0.9958, "step": 3032 }, { "epoch": 24.264, "grad_norm": 77.7869644165039, "learning_rate": 3.763185966575797e-05, "loss": 1.3036, "step": 3033 }, { "epoch": 24.272, "grad_norm": 22.325321197509766, "learning_rate": 3.762593338864526e-05, "loss": 1.3353, "step": 3034 }, { "epoch": 24.28, "grad_norm": 24.561208724975586, "learning_rate": 3.7620007111532536e-05, "loss": 0.9887, "step": 3035 }, { "epoch": 24.288, "grad_norm": 16.87555694580078, "learning_rate": 3.7614080834419815e-05, "loss": 1.1992, "step": 3036 }, { "epoch": 24.296, "grad_norm": 20.71826171875, "learning_rate": 3.76081545573071e-05, "loss": 0.9231, "step": 3037 }, { "epoch": 24.304, "grad_norm": 25.433061599731445, "learning_rate": 3.7602228280194386e-05, "loss": 1.4306, "step": 3038 }, { "epoch": 24.312, "grad_norm": 35.912750244140625, "learning_rate": 3.7596302003081665e-05, "loss": 1.7055, "step": 3039 }, { "epoch": 24.32, "grad_norm": 23.51597785949707, "learning_rate": 3.7590375725968944e-05, "loss": 1.0385, "step": 3040 }, { "epoch": 24.328, "grad_norm": 18.533708572387695, "learning_rate": 3.758444944885623e-05, "loss": 1.3198, "step": 3041 }, { "epoch": 24.336, "grad_norm": 32.726070404052734, "learning_rate": 3.7578523171743515e-05, "loss": 1.7302, "step": 3042 }, { "epoch": 24.344, "grad_norm": 25.64406394958496, "learning_rate": 3.7572596894630794e-05, "loss": 1.011, "step": 3043 }, { "epoch": 24.352, "grad_norm": 17.166152954101562, "learning_rate": 3.756667061751808e-05, "loss": 1.1927, "step": 3044 }, { "epoch": 24.36, "grad_norm": 21.152345657348633, "learning_rate": 3.756074434040536e-05, "loss": 1.4096, "step": 3045 }, { "epoch": 24.368, "grad_norm": 23.661100387573242, "learning_rate": 3.755481806329264e-05, "loss": 1.5548, "step": 3046 }, { "epoch": 24.376, "grad_norm": 30.707683563232422, "learning_rate": 3.754889178617992e-05, "loss": 1.0704, "step": 3047 }, { "epoch": 24.384, "grad_norm": 17.696094512939453, "learning_rate": 3.754296550906721e-05, "loss": 1.2371, "step": 3048 }, { "epoch": 24.392, "grad_norm": 45.63333511352539, "learning_rate": 3.7537039231954494e-05, "loss": 1.9393, "step": 3049 }, { "epoch": 24.4, "grad_norm": 38.04482650756836, "learning_rate": 3.7531112954841766e-05, "loss": 1.3464, "step": 3050 }, { "epoch": 24.408, "grad_norm": 33.32264709472656, "learning_rate": 3.752518667772905e-05, "loss": 1.06, "step": 3051 }, { "epoch": 24.416, "grad_norm": 32.56545639038086, "learning_rate": 3.751926040061634e-05, "loss": 0.8796, "step": 3052 }, { "epoch": 24.424, "grad_norm": 38.02740478515625, "learning_rate": 3.7513334123503616e-05, "loss": 1.3876, "step": 3053 }, { "epoch": 24.432, "grad_norm": 23.03337860107422, "learning_rate": 3.75074078463909e-05, "loss": 1.2904, "step": 3054 }, { "epoch": 24.44, "grad_norm": 28.221582412719727, "learning_rate": 3.750148156927818e-05, "loss": 1.2247, "step": 3055 }, { "epoch": 24.448, "grad_norm": 53.475528717041016, "learning_rate": 3.7495555292165466e-05, "loss": 1.222, "step": 3056 }, { "epoch": 24.456, "grad_norm": 16.56322479248047, "learning_rate": 3.7489629015052745e-05, "loss": 1.2321, "step": 3057 }, { "epoch": 24.464, "grad_norm": 71.57428741455078, "learning_rate": 3.748370273794003e-05, "loss": 1.2654, "step": 3058 }, { "epoch": 24.472, "grad_norm": 21.41347885131836, "learning_rate": 3.747777646082731e-05, "loss": 1.2823, "step": 3059 }, { "epoch": 24.48, "grad_norm": 31.24066925048828, "learning_rate": 3.747185018371459e-05, "loss": 0.9563, "step": 3060 }, { "epoch": 24.488, "grad_norm": 31.75592613220215, "learning_rate": 3.7465923906601874e-05, "loss": 1.4524, "step": 3061 }, { "epoch": 24.496, "grad_norm": 22.820011138916016, "learning_rate": 3.745999762948916e-05, "loss": 1.4802, "step": 3062 }, { "epoch": 24.504, "grad_norm": 31.04153060913086, "learning_rate": 3.745407135237644e-05, "loss": 1.079, "step": 3063 }, { "epoch": 24.512, "grad_norm": 73.52232360839844, "learning_rate": 3.744814507526372e-05, "loss": 1.7256, "step": 3064 }, { "epoch": 24.52, "grad_norm": 29.978164672851562, "learning_rate": 3.7442218798151e-05, "loss": 1.4606, "step": 3065 }, { "epoch": 24.528, "grad_norm": 65.5276107788086, "learning_rate": 3.743629252103829e-05, "loss": 1.0135, "step": 3066 }, { "epoch": 24.536, "grad_norm": 35.906097412109375, "learning_rate": 3.743036624392557e-05, "loss": 1.156, "step": 3067 }, { "epoch": 24.544, "grad_norm": 36.2604866027832, "learning_rate": 3.742443996681285e-05, "loss": 1.1631, "step": 3068 }, { "epoch": 24.552, "grad_norm": 44.13218307495117, "learning_rate": 3.741851368970013e-05, "loss": 1.0851, "step": 3069 }, { "epoch": 24.56, "grad_norm": 16.509998321533203, "learning_rate": 3.741258741258741e-05, "loss": 1.5053, "step": 3070 }, { "epoch": 24.568, "grad_norm": 52.71544647216797, "learning_rate": 3.7406661135474696e-05, "loss": 1.8787, "step": 3071 }, { "epoch": 24.576, "grad_norm": 37.855804443359375, "learning_rate": 3.740073485836198e-05, "loss": 1.3032, "step": 3072 }, { "epoch": 24.584, "grad_norm": 26.789737701416016, "learning_rate": 3.739480858124926e-05, "loss": 2.2018, "step": 3073 }, { "epoch": 24.592, "grad_norm": 32.59318542480469, "learning_rate": 3.738888230413654e-05, "loss": 1.0227, "step": 3074 }, { "epoch": 24.6, "grad_norm": 16.60238265991211, "learning_rate": 3.7382956027023825e-05, "loss": 1.0295, "step": 3075 }, { "epoch": 24.608, "grad_norm": 27.641767501831055, "learning_rate": 3.737702974991111e-05, "loss": 1.5159, "step": 3076 }, { "epoch": 24.616, "grad_norm": 69.77837371826172, "learning_rate": 3.737110347279839e-05, "loss": 1.6373, "step": 3077 }, { "epoch": 24.624, "grad_norm": 18.04350471496582, "learning_rate": 3.7365177195685675e-05, "loss": 1.0066, "step": 3078 }, { "epoch": 24.632, "grad_norm": 66.7489013671875, "learning_rate": 3.735925091857295e-05, "loss": 1.7466, "step": 3079 }, { "epoch": 24.64, "grad_norm": 47.859947204589844, "learning_rate": 3.735332464146024e-05, "loss": 1.3394, "step": 3080 }, { "epoch": 24.648, "grad_norm": 18.879518508911133, "learning_rate": 3.734739836434752e-05, "loss": 1.6872, "step": 3081 }, { "epoch": 24.656, "grad_norm": 35.41045379638672, "learning_rate": 3.73414720872348e-05, "loss": 1.1639, "step": 3082 }, { "epoch": 24.664, "grad_norm": 21.666940689086914, "learning_rate": 3.733554581012209e-05, "loss": 1.198, "step": 3083 }, { "epoch": 24.672, "grad_norm": 17.086214065551758, "learning_rate": 3.732961953300936e-05, "loss": 1.4697, "step": 3084 }, { "epoch": 24.68, "grad_norm": 36.462806701660156, "learning_rate": 3.7323693255896647e-05, "loss": 1.1933, "step": 3085 }, { "epoch": 24.688, "grad_norm": 25.871679306030273, "learning_rate": 3.731776697878393e-05, "loss": 1.1365, "step": 3086 }, { "epoch": 24.696, "grad_norm": 28.96134376525879, "learning_rate": 3.731184070167121e-05, "loss": 1.2454, "step": 3087 }, { "epoch": 24.704, "grad_norm": 27.617366790771484, "learning_rate": 3.73059144245585e-05, "loss": 0.8443, "step": 3088 }, { "epoch": 24.712, "grad_norm": 24.948646545410156, "learning_rate": 3.7299988147445775e-05, "loss": 1.3834, "step": 3089 }, { "epoch": 24.72, "grad_norm": 42.8243522644043, "learning_rate": 3.729406187033306e-05, "loss": 1.4504, "step": 3090 }, { "epoch": 24.728, "grad_norm": 19.63335418701172, "learning_rate": 3.728813559322034e-05, "loss": 1.0835, "step": 3091 }, { "epoch": 24.736, "grad_norm": 19.153512954711914, "learning_rate": 3.7282209316107625e-05, "loss": 1.7922, "step": 3092 }, { "epoch": 24.744, "grad_norm": 24.817922592163086, "learning_rate": 3.7276283038994904e-05, "loss": 1.2291, "step": 3093 }, { "epoch": 24.752, "grad_norm": 34.65597152709961, "learning_rate": 3.727035676188218e-05, "loss": 0.9102, "step": 3094 }, { "epoch": 24.76, "grad_norm": 18.087724685668945, "learning_rate": 3.726443048476947e-05, "loss": 1.4554, "step": 3095 }, { "epoch": 24.768, "grad_norm": 33.15453338623047, "learning_rate": 3.7258504207656754e-05, "loss": 1.4951, "step": 3096 }, { "epoch": 24.776, "grad_norm": 30.800344467163086, "learning_rate": 3.725257793054403e-05, "loss": 0.9341, "step": 3097 }, { "epoch": 24.784, "grad_norm": 26.96466636657715, "learning_rate": 3.724665165343131e-05, "loss": 1.1858, "step": 3098 }, { "epoch": 24.792, "grad_norm": 28.06049919128418, "learning_rate": 3.72407253763186e-05, "loss": 1.7995, "step": 3099 }, { "epoch": 24.8, "grad_norm": 31.812137603759766, "learning_rate": 3.723479909920588e-05, "loss": 1.3245, "step": 3100 }, { "epoch": 24.808, "grad_norm": 62.03565216064453, "learning_rate": 3.722887282209316e-05, "loss": 1.1541, "step": 3101 }, { "epoch": 24.816, "grad_norm": 26.504634857177734, "learning_rate": 3.722294654498045e-05, "loss": 1.2008, "step": 3102 }, { "epoch": 24.824, "grad_norm": 48.27610397338867, "learning_rate": 3.7217020267867726e-05, "loss": 1.4443, "step": 3103 }, { "epoch": 24.832, "grad_norm": 19.289865493774414, "learning_rate": 3.7211093990755005e-05, "loss": 1.2937, "step": 3104 }, { "epoch": 24.84, "grad_norm": 87.18985748291016, "learning_rate": 3.720516771364229e-05, "loss": 1.5412, "step": 3105 }, { "epoch": 24.848, "grad_norm": 32.08565902709961, "learning_rate": 3.7199241436529576e-05, "loss": 0.9842, "step": 3106 }, { "epoch": 24.856, "grad_norm": 18.575836181640625, "learning_rate": 3.7193315159416855e-05, "loss": 1.2936, "step": 3107 }, { "epoch": 24.864, "grad_norm": 44.82548141479492, "learning_rate": 3.7187388882304134e-05, "loss": 1.3456, "step": 3108 }, { "epoch": 24.872, "grad_norm": 66.62661743164062, "learning_rate": 3.718146260519142e-05, "loss": 1.5527, "step": 3109 }, { "epoch": 24.88, "grad_norm": 38.876808166503906, "learning_rate": 3.7175536328078705e-05, "loss": 1.7992, "step": 3110 }, { "epoch": 24.888, "grad_norm": 36.1246337890625, "learning_rate": 3.7169610050965984e-05, "loss": 1.418, "step": 3111 }, { "epoch": 24.896, "grad_norm": 60.69220733642578, "learning_rate": 3.716368377385327e-05, "loss": 1.3053, "step": 3112 }, { "epoch": 24.904, "grad_norm": 21.772403717041016, "learning_rate": 3.715775749674055e-05, "loss": 1.3544, "step": 3113 }, { "epoch": 24.912, "grad_norm": 25.219017028808594, "learning_rate": 3.7151831219627834e-05, "loss": 1.315, "step": 3114 }, { "epoch": 24.92, "grad_norm": 27.35947036743164, "learning_rate": 3.714590494251511e-05, "loss": 1.6915, "step": 3115 }, { "epoch": 24.928, "grad_norm": 26.828704833984375, "learning_rate": 3.71399786654024e-05, "loss": 1.0821, "step": 3116 }, { "epoch": 24.936, "grad_norm": 26.123350143432617, "learning_rate": 3.7134052388289684e-05, "loss": 1.1406, "step": 3117 }, { "epoch": 24.944, "grad_norm": 19.486743927001953, "learning_rate": 3.7128126111176956e-05, "loss": 1.257, "step": 3118 }, { "epoch": 24.951999999999998, "grad_norm": 32.586456298828125, "learning_rate": 3.712219983406424e-05, "loss": 1.0134, "step": 3119 }, { "epoch": 24.96, "grad_norm": 19.615184783935547, "learning_rate": 3.711627355695153e-05, "loss": 0.9412, "step": 3120 }, { "epoch": 24.968, "grad_norm": 13.815975189208984, "learning_rate": 3.7110347279838806e-05, "loss": 1.1812, "step": 3121 }, { "epoch": 24.976, "grad_norm": 22.019092559814453, "learning_rate": 3.710442100272609e-05, "loss": 1.2747, "step": 3122 }, { "epoch": 24.984, "grad_norm": 20.05815315246582, "learning_rate": 3.709849472561337e-05, "loss": 1.0255, "step": 3123 }, { "epoch": 24.992, "grad_norm": 34.92574691772461, "learning_rate": 3.7092568448500656e-05, "loss": 1.0471, "step": 3124 }, { "epoch": 25.0, "grad_norm": 30.123550415039062, "learning_rate": 3.7086642171387935e-05, "loss": 2.8858, "step": 3125 }, { "epoch": 25.0, "eval_loss": 1.3762599229812622, "eval_map": 0.2818, "eval_map_50": 0.5733, "eval_map_75": 0.2004, "eval_map_Coverall": 0.4512, "eval_map_Face_Shield": 0.2803, "eval_map_Gloves": 0.217, "eval_map_Goggles": 0.1323, "eval_map_Mask": 0.3285, "eval_map_large": 0.4054, "eval_map_medium": 0.2138, "eval_map_small": 0.1399, "eval_mar_1": 0.2624, "eval_mar_10": 0.4399, "eval_mar_100": 0.4656, "eval_mar_100_Coverall": 0.6822, "eval_mar_100_Face_Shield": 0.5412, "eval_mar_100_Gloves": 0.359, "eval_mar_100_Goggles": 0.3281, "eval_mar_100_Mask": 0.4173, "eval_mar_large": 0.648, "eval_mar_medium": 0.3527, "eval_mar_small": 0.1779, "eval_runtime": 1.8426, "eval_samples_per_second": 15.739, "eval_steps_per_second": 1.085, "step": 3125 }, { "epoch": 25.008, "grad_norm": 28.957477569580078, "learning_rate": 3.708071589427522e-05, "loss": 1.2362, "step": 3126 }, { "epoch": 25.016, "grad_norm": 30.910661697387695, "learning_rate": 3.70747896171625e-05, "loss": 1.1323, "step": 3127 }, { "epoch": 25.024, "grad_norm": 36.14187240600586, "learning_rate": 3.706886334004978e-05, "loss": 1.0987, "step": 3128 }, { "epoch": 25.032, "grad_norm": 63.813018798828125, "learning_rate": 3.7062937062937064e-05, "loss": 1.1693, "step": 3129 }, { "epoch": 25.04, "grad_norm": 28.971895217895508, "learning_rate": 3.705701078582435e-05, "loss": 1.3676, "step": 3130 }, { "epoch": 25.048, "grad_norm": 16.640731811523438, "learning_rate": 3.705108450871163e-05, "loss": 1.6504, "step": 3131 }, { "epoch": 25.056, "grad_norm": 22.235258102416992, "learning_rate": 3.704515823159891e-05, "loss": 1.1864, "step": 3132 }, { "epoch": 25.064, "grad_norm": 32.79513931274414, "learning_rate": 3.703923195448619e-05, "loss": 1.4304, "step": 3133 }, { "epoch": 25.072, "grad_norm": 47.46139907836914, "learning_rate": 3.703330567737348e-05, "loss": 1.4129, "step": 3134 }, { "epoch": 25.08, "grad_norm": 21.013473510742188, "learning_rate": 3.702737940026076e-05, "loss": 1.2311, "step": 3135 }, { "epoch": 25.088, "grad_norm": 47.717254638671875, "learning_rate": 3.702145312314804e-05, "loss": 1.3946, "step": 3136 }, { "epoch": 25.096, "grad_norm": 23.398029327392578, "learning_rate": 3.701552684603532e-05, "loss": 1.1772, "step": 3137 }, { "epoch": 25.104, "grad_norm": 23.51512336730957, "learning_rate": 3.70096005689226e-05, "loss": 1.1348, "step": 3138 }, { "epoch": 25.112, "grad_norm": 55.455875396728516, "learning_rate": 3.7003674291809886e-05, "loss": 1.235, "step": 3139 }, { "epoch": 25.12, "grad_norm": 68.5770034790039, "learning_rate": 3.699774801469717e-05, "loss": 1.1685, "step": 3140 }, { "epoch": 25.128, "grad_norm": 50.101192474365234, "learning_rate": 3.699182173758445e-05, "loss": 1.5737, "step": 3141 }, { "epoch": 25.136, "grad_norm": 32.46836471557617, "learning_rate": 3.698589546047173e-05, "loss": 1.1428, "step": 3142 }, { "epoch": 25.144, "grad_norm": 30.417165756225586, "learning_rate": 3.6979969183359015e-05, "loss": 1.0493, "step": 3143 }, { "epoch": 25.152, "grad_norm": 43.36977767944336, "learning_rate": 3.69740429062463e-05, "loss": 1.2331, "step": 3144 }, { "epoch": 25.16, "grad_norm": 20.997661590576172, "learning_rate": 3.696811662913358e-05, "loss": 2.9131, "step": 3145 }, { "epoch": 25.168, "grad_norm": 20.477216720581055, "learning_rate": 3.6962190352020865e-05, "loss": 1.3917, "step": 3146 }, { "epoch": 25.176, "grad_norm": 20.890914916992188, "learning_rate": 3.6956264074908143e-05, "loss": 1.2275, "step": 3147 }, { "epoch": 25.184, "grad_norm": 31.91927719116211, "learning_rate": 3.695033779779543e-05, "loss": 1.1049, "step": 3148 }, { "epoch": 25.192, "grad_norm": 79.01759338378906, "learning_rate": 3.694441152068271e-05, "loss": 1.2128, "step": 3149 }, { "epoch": 25.2, "grad_norm": 31.742582321166992, "learning_rate": 3.6938485243569993e-05, "loss": 1.4327, "step": 3150 }, { "epoch": 25.208, "grad_norm": 35.18943405151367, "learning_rate": 3.693255896645728e-05, "loss": 1.3579, "step": 3151 }, { "epoch": 25.216, "grad_norm": 29.036333084106445, "learning_rate": 3.692663268934455e-05, "loss": 1.231, "step": 3152 }, { "epoch": 25.224, "grad_norm": 58.798004150390625, "learning_rate": 3.692070641223184e-05, "loss": 1.059, "step": 3153 }, { "epoch": 25.232, "grad_norm": 20.487407684326172, "learning_rate": 3.691478013511912e-05, "loss": 0.9268, "step": 3154 }, { "epoch": 25.24, "grad_norm": 51.15981674194336, "learning_rate": 3.69088538580064e-05, "loss": 1.1734, "step": 3155 }, { "epoch": 25.248, "grad_norm": 26.142044067382812, "learning_rate": 3.690292758089369e-05, "loss": 1.1478, "step": 3156 }, { "epoch": 25.256, "grad_norm": 16.842071533203125, "learning_rate": 3.6897001303780966e-05, "loss": 1.1912, "step": 3157 }, { "epoch": 25.264, "grad_norm": 20.818449020385742, "learning_rate": 3.689107502666825e-05, "loss": 1.1775, "step": 3158 }, { "epoch": 25.272, "grad_norm": 28.076053619384766, "learning_rate": 3.688514874955553e-05, "loss": 1.651, "step": 3159 }, { "epoch": 25.28, "grad_norm": 20.86532974243164, "learning_rate": 3.6879222472442816e-05, "loss": 1.183, "step": 3160 }, { "epoch": 25.288, "grad_norm": 28.24625587463379, "learning_rate": 3.68732961953301e-05, "loss": 1.1323, "step": 3161 }, { "epoch": 25.296, "grad_norm": 21.66533660888672, "learning_rate": 3.686736991821737e-05, "loss": 1.3895, "step": 3162 }, { "epoch": 25.304, "grad_norm": 31.07552146911621, "learning_rate": 3.686144364110466e-05, "loss": 1.8119, "step": 3163 }, { "epoch": 25.312, "grad_norm": 37.22642517089844, "learning_rate": 3.6855517363991944e-05, "loss": 1.2501, "step": 3164 }, { "epoch": 25.32, "grad_norm": 350.5786437988281, "learning_rate": 3.684959108687922e-05, "loss": 1.8119, "step": 3165 }, { "epoch": 25.328, "grad_norm": 23.472105026245117, "learning_rate": 3.68436648097665e-05, "loss": 1.3021, "step": 3166 }, { "epoch": 25.336, "grad_norm": 20.039304733276367, "learning_rate": 3.683773853265379e-05, "loss": 1.4003, "step": 3167 }, { "epoch": 25.344, "grad_norm": 33.30084228515625, "learning_rate": 3.683181225554107e-05, "loss": 1.1529, "step": 3168 }, { "epoch": 25.352, "grad_norm": 21.25555419921875, "learning_rate": 3.682588597842835e-05, "loss": 1.1157, "step": 3169 }, { "epoch": 25.36, "grad_norm": 31.5017147064209, "learning_rate": 3.681995970131564e-05, "loss": 1.4945, "step": 3170 }, { "epoch": 25.368, "grad_norm": 19.552715301513672, "learning_rate": 3.6814033424202916e-05, "loss": 0.9519, "step": 3171 }, { "epoch": 25.376, "grad_norm": 23.84681510925293, "learning_rate": 3.6808107147090195e-05, "loss": 1.2333, "step": 3172 }, { "epoch": 25.384, "grad_norm": 23.36615753173828, "learning_rate": 3.680218086997748e-05, "loss": 1.2739, "step": 3173 }, { "epoch": 25.392, "grad_norm": 45.63249206542969, "learning_rate": 3.6796254592864766e-05, "loss": 1.982, "step": 3174 }, { "epoch": 25.4, "grad_norm": 71.14925384521484, "learning_rate": 3.6790328315752045e-05, "loss": 1.3889, "step": 3175 }, { "epoch": 25.408, "grad_norm": 32.30643081665039, "learning_rate": 3.6784402038639324e-05, "loss": 2.3553, "step": 3176 }, { "epoch": 25.416, "grad_norm": 60.11336135864258, "learning_rate": 3.677847576152661e-05, "loss": 1.2092, "step": 3177 }, { "epoch": 25.424, "grad_norm": 124.60100555419922, "learning_rate": 3.6772549484413895e-05, "loss": 1.2398, "step": 3178 }, { "epoch": 25.432, "grad_norm": 35.182411193847656, "learning_rate": 3.6766623207301174e-05, "loss": 1.4277, "step": 3179 }, { "epoch": 25.44, "grad_norm": 22.474838256835938, "learning_rate": 3.676069693018846e-05, "loss": 1.2333, "step": 3180 }, { "epoch": 25.448, "grad_norm": 203.8440399169922, "learning_rate": 3.675477065307574e-05, "loss": 1.294, "step": 3181 }, { "epoch": 25.456, "grad_norm": 30.28087615966797, "learning_rate": 3.6748844375963024e-05, "loss": 1.6237, "step": 3182 }, { "epoch": 25.464, "grad_norm": 41.51110076904297, "learning_rate": 3.67429180988503e-05, "loss": 1.5017, "step": 3183 }, { "epoch": 25.472, "grad_norm": 44.75986099243164, "learning_rate": 3.673699182173759e-05, "loss": 1.0406, "step": 3184 }, { "epoch": 25.48, "grad_norm": 56.2927131652832, "learning_rate": 3.6731065544624874e-05, "loss": 1.3397, "step": 3185 }, { "epoch": 25.488, "grad_norm": 30.908926010131836, "learning_rate": 3.6725139267512146e-05, "loss": 1.0427, "step": 3186 }, { "epoch": 25.496, "grad_norm": 31.74258041381836, "learning_rate": 3.671921299039943e-05, "loss": 1.0798, "step": 3187 }, { "epoch": 25.504, "grad_norm": 51.49248123168945, "learning_rate": 3.671328671328672e-05, "loss": 1.2865, "step": 3188 }, { "epoch": 25.512, "grad_norm": 27.948257446289062, "learning_rate": 3.6707360436173996e-05, "loss": 1.1396, "step": 3189 }, { "epoch": 25.52, "grad_norm": 24.75918197631836, "learning_rate": 3.670143415906128e-05, "loss": 1.2651, "step": 3190 }, { "epoch": 25.528, "grad_norm": 29.12664794921875, "learning_rate": 3.669550788194856e-05, "loss": 1.3221, "step": 3191 }, { "epoch": 25.536, "grad_norm": 109.70962524414062, "learning_rate": 3.6689581604835846e-05, "loss": 1.3494, "step": 3192 }, { "epoch": 25.544, "grad_norm": 60.377689361572266, "learning_rate": 3.6683655327723125e-05, "loss": 1.9391, "step": 3193 }, { "epoch": 25.552, "grad_norm": 33.50133514404297, "learning_rate": 3.667772905061041e-05, "loss": 1.303, "step": 3194 }, { "epoch": 25.56, "grad_norm": 38.65882110595703, "learning_rate": 3.6671802773497696e-05, "loss": 1.4244, "step": 3195 }, { "epoch": 25.568, "grad_norm": 43.05598831176758, "learning_rate": 3.666587649638497e-05, "loss": 1.4438, "step": 3196 }, { "epoch": 25.576, "grad_norm": 24.08331298828125, "learning_rate": 3.6659950219272254e-05, "loss": 1.4001, "step": 3197 }, { "epoch": 25.584, "grad_norm": 35.27714920043945, "learning_rate": 3.665402394215954e-05, "loss": 1.7778, "step": 3198 }, { "epoch": 25.592, "grad_norm": 18.216344833374023, "learning_rate": 3.664809766504682e-05, "loss": 1.5372, "step": 3199 }, { "epoch": 25.6, "grad_norm": 24.268991470336914, "learning_rate": 3.66421713879341e-05, "loss": 1.1715, "step": 3200 }, { "epoch": 25.608, "grad_norm": 48.00251770019531, "learning_rate": 3.663624511082138e-05, "loss": 1.242, "step": 3201 }, { "epoch": 25.616, "grad_norm": 65.34463500976562, "learning_rate": 3.663031883370867e-05, "loss": 1.2036, "step": 3202 }, { "epoch": 25.624, "grad_norm": 33.16816711425781, "learning_rate": 3.662439255659595e-05, "loss": 1.4297, "step": 3203 }, { "epoch": 25.632, "grad_norm": 24.739877700805664, "learning_rate": 3.661846627948323e-05, "loss": 1.2836, "step": 3204 }, { "epoch": 25.64, "grad_norm": 75.9173355102539, "learning_rate": 3.661254000237051e-05, "loss": 1.2233, "step": 3205 }, { "epoch": 25.648, "grad_norm": 34.48314666748047, "learning_rate": 3.660661372525779e-05, "loss": 1.0583, "step": 3206 }, { "epoch": 25.656, "grad_norm": 18.160003662109375, "learning_rate": 3.6600687448145076e-05, "loss": 1.1731, "step": 3207 }, { "epoch": 25.664, "grad_norm": 44.74171447753906, "learning_rate": 3.659476117103236e-05, "loss": 1.0333, "step": 3208 }, { "epoch": 25.672, "grad_norm": 46.425662994384766, "learning_rate": 3.658883489391964e-05, "loss": 1.3661, "step": 3209 }, { "epoch": 25.68, "grad_norm": 27.080135345458984, "learning_rate": 3.658290861680692e-05, "loss": 1.5217, "step": 3210 }, { "epoch": 25.688, "grad_norm": 18.375879287719727, "learning_rate": 3.6576982339694205e-05, "loss": 1.9077, "step": 3211 }, { "epoch": 25.696, "grad_norm": 24.1557559967041, "learning_rate": 3.657105606258149e-05, "loss": 1.5138, "step": 3212 }, { "epoch": 25.704, "grad_norm": 68.68781280517578, "learning_rate": 3.656512978546877e-05, "loss": 1.7491, "step": 3213 }, { "epoch": 25.712, "grad_norm": 26.80524444580078, "learning_rate": 3.6559203508356055e-05, "loss": 1.4472, "step": 3214 }, { "epoch": 25.72, "grad_norm": 18.875736236572266, "learning_rate": 3.6553277231243334e-05, "loss": 0.9214, "step": 3215 }, { "epoch": 25.728, "grad_norm": 65.91741943359375, "learning_rate": 3.654735095413062e-05, "loss": 1.6942, "step": 3216 }, { "epoch": 25.736, "grad_norm": 22.414485931396484, "learning_rate": 3.65414246770179e-05, "loss": 1.2239, "step": 3217 }, { "epoch": 25.744, "grad_norm": 20.788148880004883, "learning_rate": 3.6535498399905184e-05, "loss": 1.3389, "step": 3218 }, { "epoch": 25.752, "grad_norm": 25.59235191345215, "learning_rate": 3.652957212279247e-05, "loss": 1.4496, "step": 3219 }, { "epoch": 25.76, "grad_norm": 39.26118087768555, "learning_rate": 3.652364584567974e-05, "loss": 1.0869, "step": 3220 }, { "epoch": 25.768, "grad_norm": 27.46842384338379, "learning_rate": 3.651771956856703e-05, "loss": 1.5098, "step": 3221 }, { "epoch": 25.776, "grad_norm": 77.215087890625, "learning_rate": 3.651179329145431e-05, "loss": 1.1434, "step": 3222 }, { "epoch": 25.784, "grad_norm": 23.728679656982422, "learning_rate": 3.650586701434159e-05, "loss": 1.3229, "step": 3223 }, { "epoch": 25.792, "grad_norm": 18.90982437133789, "learning_rate": 3.649994073722888e-05, "loss": 0.989, "step": 3224 }, { "epoch": 25.8, "grad_norm": 27.442150115966797, "learning_rate": 3.6494014460116156e-05, "loss": 1.5093, "step": 3225 }, { "epoch": 25.808, "grad_norm": 21.461395263671875, "learning_rate": 3.648808818300344e-05, "loss": 1.2256, "step": 3226 }, { "epoch": 25.816, "grad_norm": 25.591537475585938, "learning_rate": 3.648216190589072e-05, "loss": 1.2246, "step": 3227 }, { "epoch": 25.824, "grad_norm": 34.777156829833984, "learning_rate": 3.6476235628778006e-05, "loss": 1.3109, "step": 3228 }, { "epoch": 25.832, "grad_norm": 20.239728927612305, "learning_rate": 3.647030935166529e-05, "loss": 1.2898, "step": 3229 }, { "epoch": 25.84, "grad_norm": 105.67222595214844, "learning_rate": 3.646438307455256e-05, "loss": 1.0765, "step": 3230 }, { "epoch": 25.848, "grad_norm": 23.972280502319336, "learning_rate": 3.645845679743985e-05, "loss": 1.3364, "step": 3231 }, { "epoch": 25.856, "grad_norm": 29.48146629333496, "learning_rate": 3.6452530520327134e-05, "loss": 0.9525, "step": 3232 }, { "epoch": 25.864, "grad_norm": 18.64466667175293, "learning_rate": 3.644660424321441e-05, "loss": 1.9754, "step": 3233 }, { "epoch": 25.872, "grad_norm": 22.94551658630371, "learning_rate": 3.644067796610169e-05, "loss": 1.2602, "step": 3234 }, { "epoch": 25.88, "grad_norm": 55.83826446533203, "learning_rate": 3.643475168898898e-05, "loss": 1.2892, "step": 3235 }, { "epoch": 25.888, "grad_norm": 25.501684188842773, "learning_rate": 3.642882541187626e-05, "loss": 1.4343, "step": 3236 }, { "epoch": 25.896, "grad_norm": 53.4738655090332, "learning_rate": 3.642289913476354e-05, "loss": 2.2207, "step": 3237 }, { "epoch": 25.904, "grad_norm": 27.345853805541992, "learning_rate": 3.641697285765083e-05, "loss": 1.2146, "step": 3238 }, { "epoch": 25.912, "grad_norm": 33.079227447509766, "learning_rate": 3.6411046580538106e-05, "loss": 1.1765, "step": 3239 }, { "epoch": 25.92, "grad_norm": 42.91801452636719, "learning_rate": 3.6405120303425385e-05, "loss": 1.2803, "step": 3240 }, { "epoch": 25.928, "grad_norm": 20.828689575195312, "learning_rate": 3.639919402631267e-05, "loss": 1.026, "step": 3241 }, { "epoch": 25.936, "grad_norm": 54.224037170410156, "learning_rate": 3.6393267749199956e-05, "loss": 1.478, "step": 3242 }, { "epoch": 25.944, "grad_norm": 37.16827392578125, "learning_rate": 3.6387341472087235e-05, "loss": 1.3574, "step": 3243 }, { "epoch": 25.951999999999998, "grad_norm": 26.53773307800293, "learning_rate": 3.6381415194974514e-05, "loss": 1.0896, "step": 3244 }, { "epoch": 25.96, "grad_norm": 73.8629379272461, "learning_rate": 3.63754889178618e-05, "loss": 1.2402, "step": 3245 }, { "epoch": 25.968, "grad_norm": 25.206710815429688, "learning_rate": 3.6369562640749085e-05, "loss": 1.5582, "step": 3246 }, { "epoch": 25.976, "grad_norm": 34.79302215576172, "learning_rate": 3.6363636363636364e-05, "loss": 1.2096, "step": 3247 }, { "epoch": 25.984, "grad_norm": 45.42173385620117, "learning_rate": 3.635771008652365e-05, "loss": 2.2959, "step": 3248 }, { "epoch": 25.992, "grad_norm": 25.658782958984375, "learning_rate": 3.635178380941093e-05, "loss": 1.1321, "step": 3249 }, { "epoch": 26.0, "grad_norm": 24.33058738708496, "learning_rate": 3.6345857532298214e-05, "loss": 1.3142, "step": 3250 }, { "epoch": 26.0, "eval_loss": 1.3648595809936523, "eval_map": 0.2787, "eval_map_50": 0.5539, "eval_map_75": 0.2452, "eval_map_Coverall": 0.4132, "eval_map_Face_Shield": 0.2847, "eval_map_Gloves": 0.1948, "eval_map_Goggles": 0.1384, "eval_map_Mask": 0.3621, "eval_map_large": 0.4071, "eval_map_medium": 0.2015, "eval_map_small": 0.2062, "eval_mar_1": 0.257, "eval_mar_10": 0.4883, "eval_mar_100": 0.5122, "eval_mar_100_Coverall": 0.6711, "eval_mar_100_Face_Shield": 0.6588, "eval_mar_100_Gloves": 0.4098, "eval_mar_100_Goggles": 0.3656, "eval_mar_100_Mask": 0.4558, "eval_mar_large": 0.733, "eval_mar_medium": 0.4026, "eval_mar_small": 0.2471, "eval_runtime": 1.8249, "eval_samples_per_second": 15.892, "eval_steps_per_second": 1.096, "step": 3250 }, { "epoch": 26.008, "grad_norm": 46.665897369384766, "learning_rate": 3.633993125518549e-05, "loss": 1.1431, "step": 3251 }, { "epoch": 26.016, "grad_norm": 22.65245819091797, "learning_rate": 3.633400497807278e-05, "loss": 1.1472, "step": 3252 }, { "epoch": 26.024, "grad_norm": 32.90817642211914, "learning_rate": 3.6328078700960064e-05, "loss": 1.5113, "step": 3253 }, { "epoch": 26.032, "grad_norm": 135.48944091796875, "learning_rate": 3.6322152423847336e-05, "loss": 0.895, "step": 3254 }, { "epoch": 26.04, "grad_norm": 23.17120933532715, "learning_rate": 3.631622614673462e-05, "loss": 0.8933, "step": 3255 }, { "epoch": 26.048, "grad_norm": 28.286724090576172, "learning_rate": 3.631029986962191e-05, "loss": 1.3043, "step": 3256 }, { "epoch": 26.056, "grad_norm": 18.104524612426758, "learning_rate": 3.6304373592509186e-05, "loss": 1.0397, "step": 3257 }, { "epoch": 26.064, "grad_norm": 38.80641174316406, "learning_rate": 3.629844731539647e-05, "loss": 1.737, "step": 3258 }, { "epoch": 26.072, "grad_norm": 24.239063262939453, "learning_rate": 3.629252103828375e-05, "loss": 1.1739, "step": 3259 }, { "epoch": 26.08, "grad_norm": 40.76322555541992, "learning_rate": 3.6286594761171036e-05, "loss": 0.9373, "step": 3260 }, { "epoch": 26.088, "grad_norm": 22.293842315673828, "learning_rate": 3.6280668484058315e-05, "loss": 1.1559, "step": 3261 }, { "epoch": 26.096, "grad_norm": 20.646516799926758, "learning_rate": 3.62747422069456e-05, "loss": 1.3248, "step": 3262 }, { "epoch": 26.104, "grad_norm": 25.08036231994629, "learning_rate": 3.6268815929832886e-05, "loss": 1.3736, "step": 3263 }, { "epoch": 26.112, "grad_norm": 36.108795166015625, "learning_rate": 3.626288965272016e-05, "loss": 1.1009, "step": 3264 }, { "epoch": 26.12, "grad_norm": 19.974342346191406, "learning_rate": 3.6256963375607444e-05, "loss": 1.1421, "step": 3265 }, { "epoch": 26.128, "grad_norm": 22.124711990356445, "learning_rate": 3.625103709849473e-05, "loss": 1.3195, "step": 3266 }, { "epoch": 26.136, "grad_norm": 227.92648315429688, "learning_rate": 3.624511082138201e-05, "loss": 1.5112, "step": 3267 }, { "epoch": 26.144, "grad_norm": 27.48726463317871, "learning_rate": 3.623918454426929e-05, "loss": 1.1297, "step": 3268 }, { "epoch": 26.152, "grad_norm": 26.124248504638672, "learning_rate": 3.623325826715657e-05, "loss": 1.2855, "step": 3269 }, { "epoch": 26.16, "grad_norm": 37.45722961425781, "learning_rate": 3.622733199004386e-05, "loss": 3.1053, "step": 3270 }, { "epoch": 26.168, "grad_norm": 73.5436019897461, "learning_rate": 3.622140571293114e-05, "loss": 1.249, "step": 3271 }, { "epoch": 26.176, "grad_norm": 44.52085876464844, "learning_rate": 3.621547943581842e-05, "loss": 2.2537, "step": 3272 }, { "epoch": 26.184, "grad_norm": 16.543289184570312, "learning_rate": 3.62095531587057e-05, "loss": 1.2766, "step": 3273 }, { "epoch": 26.192, "grad_norm": 55.12327575683594, "learning_rate": 3.620362688159298e-05, "loss": 2.3015, "step": 3274 }, { "epoch": 26.2, "grad_norm": 29.754201889038086, "learning_rate": 3.6197700604480266e-05, "loss": 0.991, "step": 3275 }, { "epoch": 26.208, "grad_norm": 21.085981369018555, "learning_rate": 3.619177432736755e-05, "loss": 1.1806, "step": 3276 }, { "epoch": 26.216, "grad_norm": 20.40858268737793, "learning_rate": 3.618584805025483e-05, "loss": 0.98, "step": 3277 }, { "epoch": 26.224, "grad_norm": 44.230289459228516, "learning_rate": 3.617992177314211e-05, "loss": 1.1926, "step": 3278 }, { "epoch": 26.232, "grad_norm": 32.96345138549805, "learning_rate": 3.6173995496029395e-05, "loss": 1.282, "step": 3279 }, { "epoch": 26.24, "grad_norm": 27.568281173706055, "learning_rate": 3.616806921891668e-05, "loss": 1.2816, "step": 3280 }, { "epoch": 26.248, "grad_norm": 14.382080078125, "learning_rate": 3.616214294180396e-05, "loss": 1.174, "step": 3281 }, { "epoch": 26.256, "grad_norm": 38.44124221801758, "learning_rate": 3.6156216664691245e-05, "loss": 1.3685, "step": 3282 }, { "epoch": 26.264, "grad_norm": 18.640932083129883, "learning_rate": 3.6150290387578524e-05, "loss": 1.2957, "step": 3283 }, { "epoch": 26.272, "grad_norm": 35.868804931640625, "learning_rate": 3.614436411046581e-05, "loss": 1.25, "step": 3284 }, { "epoch": 26.28, "grad_norm": 41.37274932861328, "learning_rate": 3.613843783335309e-05, "loss": 1.2192, "step": 3285 }, { "epoch": 26.288, "grad_norm": 21.608871459960938, "learning_rate": 3.6132511556240374e-05, "loss": 1.1881, "step": 3286 }, { "epoch": 26.296, "grad_norm": 29.16553497314453, "learning_rate": 3.612658527912766e-05, "loss": 1.1657, "step": 3287 }, { "epoch": 26.304, "grad_norm": 28.71891975402832, "learning_rate": 3.612065900201493e-05, "loss": 0.8093, "step": 3288 }, { "epoch": 26.312, "grad_norm": 26.271671295166016, "learning_rate": 3.611473272490222e-05, "loss": 1.184, "step": 3289 }, { "epoch": 26.32, "grad_norm": 22.887840270996094, "learning_rate": 3.61088064477895e-05, "loss": 1.1144, "step": 3290 }, { "epoch": 26.328, "grad_norm": 24.93805503845215, "learning_rate": 3.610288017067678e-05, "loss": 1.4003, "step": 3291 }, { "epoch": 26.336, "grad_norm": 31.87023162841797, "learning_rate": 3.609695389356407e-05, "loss": 1.1921, "step": 3292 }, { "epoch": 26.344, "grad_norm": 23.161277770996094, "learning_rate": 3.6091027616451346e-05, "loss": 0.8542, "step": 3293 }, { "epoch": 26.352, "grad_norm": 49.60969161987305, "learning_rate": 3.608510133933863e-05, "loss": 1.3242, "step": 3294 }, { "epoch": 26.36, "grad_norm": 16.238916397094727, "learning_rate": 3.607917506222591e-05, "loss": 1.3105, "step": 3295 }, { "epoch": 26.368, "grad_norm": 26.767518997192383, "learning_rate": 3.6073248785113196e-05, "loss": 1.615, "step": 3296 }, { "epoch": 26.376, "grad_norm": 14.258577346801758, "learning_rate": 3.606732250800048e-05, "loss": 0.9531, "step": 3297 }, { "epoch": 26.384, "grad_norm": 27.818857192993164, "learning_rate": 3.606139623088775e-05, "loss": 1.259, "step": 3298 }, { "epoch": 26.392, "grad_norm": 23.054006576538086, "learning_rate": 3.605546995377504e-05, "loss": 0.9785, "step": 3299 }, { "epoch": 26.4, "grad_norm": 27.816965103149414, "learning_rate": 3.6049543676662325e-05, "loss": 1.0174, "step": 3300 }, { "epoch": 26.408, "grad_norm": 21.796863555908203, "learning_rate": 3.60436173995496e-05, "loss": 0.9519, "step": 3301 }, { "epoch": 26.416, "grad_norm": 35.416507720947266, "learning_rate": 3.603769112243689e-05, "loss": 1.5244, "step": 3302 }, { "epoch": 26.424, "grad_norm": 25.912607192993164, "learning_rate": 3.603176484532417e-05, "loss": 0.9248, "step": 3303 }, { "epoch": 26.432, "grad_norm": 22.5422420501709, "learning_rate": 3.602583856821145e-05, "loss": 1.4901, "step": 3304 }, { "epoch": 26.44, "grad_norm": 32.50306701660156, "learning_rate": 3.601991229109873e-05, "loss": 1.1912, "step": 3305 }, { "epoch": 26.448, "grad_norm": 30.04547691345215, "learning_rate": 3.601398601398602e-05, "loss": 0.9601, "step": 3306 }, { "epoch": 26.456, "grad_norm": 38.03318786621094, "learning_rate": 3.6008059736873297e-05, "loss": 1.1117, "step": 3307 }, { "epoch": 26.464, "grad_norm": 75.61845397949219, "learning_rate": 3.6002133459760575e-05, "loss": 1.2814, "step": 3308 }, { "epoch": 26.472, "grad_norm": 69.15403747558594, "learning_rate": 3.599620718264786e-05, "loss": 1.3429, "step": 3309 }, { "epoch": 26.48, "grad_norm": 23.642244338989258, "learning_rate": 3.5990280905535147e-05, "loss": 1.5046, "step": 3310 }, { "epoch": 26.488, "grad_norm": 24.074522018432617, "learning_rate": 3.5984354628422425e-05, "loss": 1.9736, "step": 3311 }, { "epoch": 26.496, "grad_norm": 28.519153594970703, "learning_rate": 3.5978428351309704e-05, "loss": 1.6593, "step": 3312 }, { "epoch": 26.504, "grad_norm": 15.44912338256836, "learning_rate": 3.597250207419699e-05, "loss": 1.2227, "step": 3313 }, { "epoch": 26.512, "grad_norm": 16.326139450073242, "learning_rate": 3.5966575797084275e-05, "loss": 1.0394, "step": 3314 }, { "epoch": 26.52, "grad_norm": 24.655155181884766, "learning_rate": 3.5960649519971554e-05, "loss": 0.966, "step": 3315 }, { "epoch": 26.528, "grad_norm": 22.293657302856445, "learning_rate": 3.595472324285884e-05, "loss": 1.425, "step": 3316 }, { "epoch": 26.536, "grad_norm": 31.163673400878906, "learning_rate": 3.594879696574612e-05, "loss": 1.3983, "step": 3317 }, { "epoch": 26.544, "grad_norm": 15.938191413879395, "learning_rate": 3.5942870688633404e-05, "loss": 1.4583, "step": 3318 }, { "epoch": 26.552, "grad_norm": 33.29003143310547, "learning_rate": 3.593694441152068e-05, "loss": 1.3003, "step": 3319 }, { "epoch": 26.56, "grad_norm": 52.03406524658203, "learning_rate": 3.593101813440797e-05, "loss": 1.2079, "step": 3320 }, { "epoch": 26.568, "grad_norm": 49.69149398803711, "learning_rate": 3.5925091857295254e-05, "loss": 1.3312, "step": 3321 }, { "epoch": 26.576, "grad_norm": 21.594200134277344, "learning_rate": 3.5919165580182526e-05, "loss": 1.105, "step": 3322 }, { "epoch": 26.584, "grad_norm": 31.451169967651367, "learning_rate": 3.591323930306981e-05, "loss": 1.0328, "step": 3323 }, { "epoch": 26.592, "grad_norm": 18.00613021850586, "learning_rate": 3.59073130259571e-05, "loss": 1.9226, "step": 3324 }, { "epoch": 26.6, "grad_norm": 26.270599365234375, "learning_rate": 3.5901386748844376e-05, "loss": 1.4107, "step": 3325 }, { "epoch": 26.608, "grad_norm": 38.40221405029297, "learning_rate": 3.589546047173166e-05, "loss": 1.6889, "step": 3326 }, { "epoch": 26.616, "grad_norm": 12.148204803466797, "learning_rate": 3.588953419461894e-05, "loss": 1.3091, "step": 3327 }, { "epoch": 26.624, "grad_norm": 42.618099212646484, "learning_rate": 3.5883607917506226e-05, "loss": 0.9838, "step": 3328 }, { "epoch": 26.632, "grad_norm": 99.96153259277344, "learning_rate": 3.5877681640393505e-05, "loss": 1.7657, "step": 3329 }, { "epoch": 26.64, "grad_norm": 24.259979248046875, "learning_rate": 3.587175536328079e-05, "loss": 1.5646, "step": 3330 }, { "epoch": 26.648, "grad_norm": 30.54053497314453, "learning_rate": 3.5865829086168076e-05, "loss": 1.2456, "step": 3331 }, { "epoch": 26.656, "grad_norm": 47.89191818237305, "learning_rate": 3.585990280905535e-05, "loss": 1.3442, "step": 3332 }, { "epoch": 26.664, "grad_norm": 33.20600128173828, "learning_rate": 3.5853976531942634e-05, "loss": 1.5849, "step": 3333 }, { "epoch": 26.672, "grad_norm": 60.48743438720703, "learning_rate": 3.584805025482992e-05, "loss": 1.0293, "step": 3334 }, { "epoch": 26.68, "grad_norm": 133.77151489257812, "learning_rate": 3.58421239777172e-05, "loss": 1.5126, "step": 3335 }, { "epoch": 26.688, "grad_norm": 57.414146423339844, "learning_rate": 3.5836197700604484e-05, "loss": 1.5528, "step": 3336 }, { "epoch": 26.696, "grad_norm": 16.690824508666992, "learning_rate": 3.583027142349176e-05, "loss": 1.173, "step": 3337 }, { "epoch": 26.704, "grad_norm": 16.60665512084961, "learning_rate": 3.582434514637905e-05, "loss": 1.3039, "step": 3338 }, { "epoch": 26.712, "grad_norm": 17.894304275512695, "learning_rate": 3.581841886926633e-05, "loss": 1.082, "step": 3339 }, { "epoch": 26.72, "grad_norm": 35.53397750854492, "learning_rate": 3.581249259215361e-05, "loss": 1.235, "step": 3340 }, { "epoch": 26.728, "grad_norm": 28.49491310119629, "learning_rate": 3.580656631504089e-05, "loss": 1.387, "step": 3341 }, { "epoch": 26.736, "grad_norm": 29.565162658691406, "learning_rate": 3.580064003792817e-05, "loss": 1.0429, "step": 3342 }, { "epoch": 26.744, "grad_norm": 21.716472625732422, "learning_rate": 3.5794713760815456e-05, "loss": 0.8973, "step": 3343 }, { "epoch": 26.752, "grad_norm": 31.507394790649414, "learning_rate": 3.578878748370274e-05, "loss": 1.4653, "step": 3344 }, { "epoch": 26.76, "grad_norm": 33.533817291259766, "learning_rate": 3.578286120659002e-05, "loss": 0.9853, "step": 3345 }, { "epoch": 26.768, "grad_norm": 30.317611694335938, "learning_rate": 3.57769349294773e-05, "loss": 0.7116, "step": 3346 }, { "epoch": 26.776, "grad_norm": 29.559722900390625, "learning_rate": 3.5771008652364585e-05, "loss": 1.2768, "step": 3347 }, { "epoch": 26.784, "grad_norm": 21.496084213256836, "learning_rate": 3.576508237525187e-05, "loss": 1.5904, "step": 3348 }, { "epoch": 26.792, "grad_norm": 24.086774826049805, "learning_rate": 3.575915609813915e-05, "loss": 1.3454, "step": 3349 }, { "epoch": 26.8, "grad_norm": 24.01566505432129, "learning_rate": 3.5753229821026435e-05, "loss": 1.1888, "step": 3350 }, { "epoch": 26.808, "grad_norm": 37.220985412597656, "learning_rate": 3.5747303543913714e-05, "loss": 1.554, "step": 3351 }, { "epoch": 26.816, "grad_norm": 39.22897720336914, "learning_rate": 3.5741377266801e-05, "loss": 1.2659, "step": 3352 }, { "epoch": 26.824, "grad_norm": 40.721961975097656, "learning_rate": 3.573545098968828e-05, "loss": 1.3019, "step": 3353 }, { "epoch": 26.832, "grad_norm": 21.762388229370117, "learning_rate": 3.5729524712575564e-05, "loss": 1.1094, "step": 3354 }, { "epoch": 26.84, "grad_norm": 25.91400146484375, "learning_rate": 3.572359843546285e-05, "loss": 1.7435, "step": 3355 }, { "epoch": 26.848, "grad_norm": 22.662385940551758, "learning_rate": 3.571767215835012e-05, "loss": 1.8686, "step": 3356 }, { "epoch": 26.856, "grad_norm": 46.761375427246094, "learning_rate": 3.571174588123741e-05, "loss": 1.1016, "step": 3357 }, { "epoch": 26.864, "grad_norm": 24.42765235900879, "learning_rate": 3.570581960412469e-05, "loss": 1.5868, "step": 3358 }, { "epoch": 26.872, "grad_norm": 22.895854949951172, "learning_rate": 3.569989332701197e-05, "loss": 1.3552, "step": 3359 }, { "epoch": 26.88, "grad_norm": 19.478918075561523, "learning_rate": 3.569396704989926e-05, "loss": 0.9047, "step": 3360 }, { "epoch": 26.888, "grad_norm": 21.95102310180664, "learning_rate": 3.5688040772786536e-05, "loss": 1.4716, "step": 3361 }, { "epoch": 26.896, "grad_norm": 37.81241226196289, "learning_rate": 3.568211449567382e-05, "loss": 1.4666, "step": 3362 }, { "epoch": 26.904, "grad_norm": 20.882434844970703, "learning_rate": 3.56761882185611e-05, "loss": 1.3996, "step": 3363 }, { "epoch": 26.912, "grad_norm": 24.274784088134766, "learning_rate": 3.5670261941448386e-05, "loss": 1.4266, "step": 3364 }, { "epoch": 26.92, "grad_norm": 22.462615966796875, "learning_rate": 3.566433566433567e-05, "loss": 0.9111, "step": 3365 }, { "epoch": 26.928, "grad_norm": 138.2371368408203, "learning_rate": 3.5658409387222943e-05, "loss": 0.8341, "step": 3366 }, { "epoch": 26.936, "grad_norm": 30.161386489868164, "learning_rate": 3.565248311011023e-05, "loss": 1.4168, "step": 3367 }, { "epoch": 26.944, "grad_norm": 114.9844970703125, "learning_rate": 3.5646556832997515e-05, "loss": 0.9429, "step": 3368 }, { "epoch": 26.951999999999998, "grad_norm": 45.01750564575195, "learning_rate": 3.5640630555884793e-05, "loss": 1.4817, "step": 3369 }, { "epoch": 26.96, "grad_norm": 26.358243942260742, "learning_rate": 3.563470427877208e-05, "loss": 1.2221, "step": 3370 }, { "epoch": 26.968, "grad_norm": 39.86807632446289, "learning_rate": 3.562877800165936e-05, "loss": 1.1147, "step": 3371 }, { "epoch": 26.976, "grad_norm": 25.74915313720703, "learning_rate": 3.5622851724546643e-05, "loss": 1.3442, "step": 3372 }, { "epoch": 26.984, "grad_norm": 40.955909729003906, "learning_rate": 3.561692544743392e-05, "loss": 1.3502, "step": 3373 }, { "epoch": 26.992, "grad_norm": 17.466022491455078, "learning_rate": 3.561099917032121e-05, "loss": 1.084, "step": 3374 }, { "epoch": 27.0, "grad_norm": 38.84275436401367, "learning_rate": 3.560507289320849e-05, "loss": 2.1081, "step": 3375 }, { "epoch": 27.0, "eval_loss": 1.1950929164886475, "eval_map": 0.3186, "eval_map_50": 0.638, "eval_map_75": 0.2701, "eval_map_Coverall": 0.5105, "eval_map_Face_Shield": 0.2804, "eval_map_Gloves": 0.2836, "eval_map_Goggles": 0.1154, "eval_map_Mask": 0.4031, "eval_map_large": 0.4688, "eval_map_medium": 0.2412, "eval_map_small": 0.1709, "eval_mar_1": 0.2876, "eval_mar_10": 0.4874, "eval_mar_100": 0.4997, "eval_mar_100_Coverall": 0.6889, "eval_mar_100_Face_Shield": 0.5765, "eval_mar_100_Gloves": 0.4246, "eval_mar_100_Goggles": 0.3375, "eval_mar_100_Mask": 0.4712, "eval_mar_large": 0.6428, "eval_mar_medium": 0.4182, "eval_mar_small": 0.2254, "eval_runtime": 1.8356, "eval_samples_per_second": 15.799, "eval_steps_per_second": 1.09, "step": 3375 } ], "logging_steps": 1, "max_steps": 9375, "num_input_tokens_seen": 0, "num_train_epochs": 75, "save_steps": 500, "total_flos": 1.05705530712e+19, "train_batch_size": 8, "trial_name": null, "trial_params": null }