{ "best_metric": 0.23615802824497223, "best_model_checkpoint": "chickens-composite-403232323232-150-epochs-w-transform-metrics-test/checkpoint-123000", "epoch": 150.0, "eval_steps": 500, "global_step": 150000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.03, "grad_norm": 59.60230255126953, "learning_rate": 9.999999013039593e-06, "loss": 2.4437, "step": 30 }, { "epoch": 0.06, "grad_norm": 173.15296936035156, "learning_rate": 9.99999605215876e-06, "loss": 2.1202, "step": 60 }, { "epoch": 0.09, "grad_norm": 55.09599685668945, "learning_rate": 9.99999111735867e-06, "loss": 2.1122, "step": 90 }, { "epoch": 0.12, "grad_norm": 165.06375122070312, "learning_rate": 9.999984208641271e-06, "loss": 1.8063, "step": 120 }, { "epoch": 0.15, "grad_norm": 66.5302505493164, "learning_rate": 9.999975326009292e-06, "loss": 1.9324, "step": 150 }, { "epoch": 0.18, "grad_norm": 47.085636138916016, "learning_rate": 9.999964469466236e-06, "loss": 1.7749, "step": 180 }, { "epoch": 0.21, "grad_norm": 54.95656967163086, "learning_rate": 9.999951639016396e-06, "loss": 1.684, "step": 210 }, { "epoch": 0.24, "grad_norm": 104.04905700683594, "learning_rate": 9.99993683466483e-06, "loss": 1.6048, "step": 240 }, { "epoch": 0.27, "grad_norm": 282.3728332519531, "learning_rate": 9.999920056417385e-06, "loss": 1.526, "step": 270 }, { "epoch": 0.3, "grad_norm": 110.48187255859375, "learning_rate": 9.999901304280686e-06, "loss": 1.3908, "step": 300 }, { "epoch": 0.33, "grad_norm": 113.96318817138672, "learning_rate": 9.999880578262135e-06, "loss": 1.6735, "step": 330 }, { "epoch": 0.36, "grad_norm": 38.64141082763672, "learning_rate": 9.999857878369917e-06, "loss": 1.4272, "step": 360 }, { "epoch": 0.39, "grad_norm": 70.68793487548828, "learning_rate": 9.999833204612988e-06, "loss": 1.5643, "step": 390 }, { "epoch": 0.42, "grad_norm": 62.593807220458984, "learning_rate": 9.999806557001092e-06, "loss": 1.4075, "step": 420 }, { "epoch": 0.45, "grad_norm": 88.04399871826172, "learning_rate": 9.99977793554475e-06, "loss": 1.3844, "step": 450 }, { "epoch": 0.48, "grad_norm": 66.90438842773438, "learning_rate": 9.99974734025526e-06, "loss": 1.3663, "step": 480 }, { "epoch": 0.51, "grad_norm": 49.52287673950195, "learning_rate": 9.9997147711447e-06, "loss": 1.3628, "step": 510 }, { "epoch": 0.54, "grad_norm": 113.30903625488281, "learning_rate": 9.99968022822593e-06, "loss": 1.2915, "step": 540 }, { "epoch": 0.57, "grad_norm": 86.755859375, "learning_rate": 9.999643711512586e-06, "loss": 1.2864, "step": 570 }, { "epoch": 0.6, "grad_norm": 73.81269836425781, "learning_rate": 9.999605221019082e-06, "loss": 1.2369, "step": 600 }, { "epoch": 0.63, "grad_norm": 45.95207595825195, "learning_rate": 9.999564756760616e-06, "loss": 1.1184, "step": 630 }, { "epoch": 0.66, "grad_norm": 78.30489349365234, "learning_rate": 9.99952231875316e-06, "loss": 1.2153, "step": 660 }, { "epoch": 0.69, "grad_norm": 59.20403289794922, "learning_rate": 9.999477907013473e-06, "loss": 1.2404, "step": 690 }, { "epoch": 0.72, "grad_norm": 225.3522491455078, "learning_rate": 9.999431521559081e-06, "loss": 1.2987, "step": 720 }, { "epoch": 0.75, "grad_norm": 43.183326721191406, "learning_rate": 9.999383162408303e-06, "loss": 1.2685, "step": 750 }, { "epoch": 0.78, "grad_norm": 150.42115783691406, "learning_rate": 9.999332829580227e-06, "loss": 1.1827, "step": 780 }, { "epoch": 0.81, "grad_norm": 194.66871643066406, "learning_rate": 9.999280523094724e-06, "loss": 1.4356, "step": 810 }, { "epoch": 0.84, "grad_norm": 122.09674072265625, "learning_rate": 9.999226242972445e-06, "loss": 1.1521, "step": 840 }, { "epoch": 0.87, "grad_norm": 113.1258773803711, "learning_rate": 9.999169989234815e-06, "loss": 1.3035, "step": 870 }, { "epoch": 0.9, "grad_norm": 61.528377532958984, "learning_rate": 9.999111761904046e-06, "loss": 1.23, "step": 900 }, { "epoch": 0.93, "grad_norm": 45.063499450683594, "learning_rate": 9.999051561003124e-06, "loss": 1.2244, "step": 930 }, { "epoch": 0.96, "grad_norm": 347.16650390625, "learning_rate": 9.998989386555815e-06, "loss": 1.2313, "step": 960 }, { "epoch": 0.99, "grad_norm": 58.914268493652344, "learning_rate": 9.998925238586666e-06, "loss": 1.1267, "step": 990 }, { "epoch": 1.0, "eval_loss": 1.2143162488937378, "eval_map": 0.2128, "eval_map_50": 0.3053, "eval_map_75": 0.2462, "eval_map_chicken": 0.1029, "eval_map_duck": 0.0023, "eval_map_large": 0.2574, "eval_map_medium": 0.1247, "eval_map_plant": 0.5333, "eval_map_small": 0.0212, "eval_mar_1": 0.1042, "eval_mar_10": 0.3253, "eval_mar_100": 0.3593, "eval_mar_100_chicken": 0.3376, "eval_mar_100_duck": 0.0036, "eval_mar_100_plant": 0.7365, "eval_mar_large": 0.3725, "eval_mar_medium": 0.3264, "eval_mar_small": 0.0767, "eval_runtime": 12.5953, "eval_samples_per_second": 15.879, "eval_steps_per_second": 1.985, "step": 1000 }, { "epoch": 1.02, "grad_norm": 46.245826721191406, "learning_rate": 9.998859117121e-06, "loss": 1.0944, "step": 1020 }, { "epoch": 1.05, "grad_norm": 54.24517059326172, "learning_rate": 9.998791022184921e-06, "loss": 1.2681, "step": 1050 }, { "epoch": 1.08, "grad_norm": 110.3233642578125, "learning_rate": 9.998720953805312e-06, "loss": 1.1793, "step": 1080 }, { "epoch": 1.11, "grad_norm": 115.63661193847656, "learning_rate": 9.998648912009835e-06, "loss": 1.243, "step": 1110 }, { "epoch": 1.1400000000000001, "grad_norm": 318.80340576171875, "learning_rate": 9.998574896826931e-06, "loss": 1.3515, "step": 1140 }, { "epoch": 1.17, "grad_norm": 57.54459762573242, "learning_rate": 9.99849890828582e-06, "loss": 1.3478, "step": 1170 }, { "epoch": 1.2, "grad_norm": 69.71772003173828, "learning_rate": 9.9984209464165e-06, "loss": 1.2867, "step": 1200 }, { "epoch": 1.23, "grad_norm": 76.68997955322266, "learning_rate": 9.998341011249752e-06, "loss": 1.46, "step": 1230 }, { "epoch": 1.26, "grad_norm": 24.792621612548828, "learning_rate": 9.99825910281713e-06, "loss": 1.1404, "step": 1260 }, { "epoch": 1.29, "grad_norm": 89.17694854736328, "learning_rate": 9.99817522115097e-06, "loss": 1.195, "step": 1290 }, { "epoch": 1.32, "grad_norm": 43.80031967163086, "learning_rate": 9.998089366284392e-06, "loss": 0.9176, "step": 1320 }, { "epoch": 1.35, "grad_norm": 74.01356506347656, "learning_rate": 9.998001538251283e-06, "loss": 1.1058, "step": 1350 }, { "epoch": 1.38, "grad_norm": 40.9770622253418, "learning_rate": 9.997911737086322e-06, "loss": 1.1492, "step": 1380 }, { "epoch": 1.41, "grad_norm": 62.99734878540039, "learning_rate": 9.997819962824958e-06, "loss": 0.9646, "step": 1410 }, { "epoch": 1.44, "grad_norm": 87.90940856933594, "learning_rate": 9.997726215503422e-06, "loss": 1.1021, "step": 1440 }, { "epoch": 1.47, "grad_norm": 103.54161834716797, "learning_rate": 9.997630495158728e-06, "loss": 1.1281, "step": 1470 }, { "epoch": 1.5, "grad_norm": 28.80501365661621, "learning_rate": 9.997532801828659e-06, "loss": 1.1304, "step": 1500 }, { "epoch": 1.53, "grad_norm": 42.98833465576172, "learning_rate": 9.997433135551786e-06, "loss": 1.1578, "step": 1530 }, { "epoch": 1.56, "grad_norm": 23.64927101135254, "learning_rate": 9.997331496367455e-06, "loss": 0.9452, "step": 1560 }, { "epoch": 1.5899999999999999, "grad_norm": 55.565670013427734, "learning_rate": 9.997227884315792e-06, "loss": 0.9862, "step": 1590 }, { "epoch": 1.62, "grad_norm": 228.7602081298828, "learning_rate": 9.9971222994377e-06, "loss": 1.0688, "step": 1620 }, { "epoch": 1.65, "grad_norm": 110.9461441040039, "learning_rate": 9.997014741774866e-06, "loss": 1.0945, "step": 1650 }, { "epoch": 1.6800000000000002, "grad_norm": 92.61849212646484, "learning_rate": 9.996905211369748e-06, "loss": 1.1725, "step": 1680 }, { "epoch": 1.71, "grad_norm": 45.66724395751953, "learning_rate": 9.996793708265586e-06, "loss": 0.9675, "step": 1710 }, { "epoch": 1.74, "grad_norm": 98.07511901855469, "learning_rate": 9.996680232506404e-06, "loss": 1.1953, "step": 1740 }, { "epoch": 1.77, "grad_norm": 94.22811889648438, "learning_rate": 9.996564784137e-06, "loss": 1.0345, "step": 1770 }, { "epoch": 1.8, "grad_norm": 72.64576721191406, "learning_rate": 9.996447363202947e-06, "loss": 1.1105, "step": 1800 }, { "epoch": 1.83, "grad_norm": 60.5106086730957, "learning_rate": 9.996327969750605e-06, "loss": 0.944, "step": 1830 }, { "epoch": 1.8599999999999999, "grad_norm": 30.475387573242188, "learning_rate": 9.996206603827105e-06, "loss": 0.9899, "step": 1860 }, { "epoch": 1.8900000000000001, "grad_norm": 134.0454559326172, "learning_rate": 9.996083265480366e-06, "loss": 0.9667, "step": 1890 }, { "epoch": 1.92, "grad_norm": 29.525602340698242, "learning_rate": 9.995957954759073e-06, "loss": 0.9126, "step": 1920 }, { "epoch": 1.95, "grad_norm": 32.04520797729492, "learning_rate": 9.995830671712701e-06, "loss": 0.9726, "step": 1950 }, { "epoch": 1.98, "grad_norm": 23.49850082397461, "learning_rate": 9.9957014163915e-06, "loss": 1.1423, "step": 1980 }, { "epoch": 2.0, "eval_loss": 1.0582586526870728, "eval_map": 0.2633, "eval_map_50": 0.384, "eval_map_75": 0.2877, "eval_map_chicken": 0.1457, "eval_map_duck": 0.0, "eval_map_large": 0.2767, "eval_map_medium": 0.1816, "eval_map_plant": 0.6442, "eval_map_small": 0.0549, "eval_mar_1": 0.1024, "eval_mar_10": 0.3946, "eval_mar_100": 0.4425, "eval_mar_100_chicken": 0.5795, "eval_mar_100_duck": 0.0, "eval_mar_100_plant": 0.7479, "eval_mar_large": 0.4552, "eval_mar_medium": 0.4032, "eval_mar_small": 0.13, "eval_runtime": 12.3197, "eval_samples_per_second": 16.234, "eval_steps_per_second": 2.029, "step": 2000 }, { "epoch": 2.01, "grad_norm": 59.291404724121094, "learning_rate": 9.995570188846495e-06, "loss": 0.8712, "step": 2010 }, { "epoch": 2.04, "grad_norm": 64.33667755126953, "learning_rate": 9.995436989129495e-06, "loss": 0.9764, "step": 2040 }, { "epoch": 2.07, "grad_norm": 100.77915954589844, "learning_rate": 9.995301817293084e-06, "loss": 1.0456, "step": 2070 }, { "epoch": 2.1, "grad_norm": 133.7511444091797, "learning_rate": 9.995164673390624e-06, "loss": 1.0333, "step": 2100 }, { "epoch": 2.13, "grad_norm": 65.17808532714844, "learning_rate": 9.99502555747626e-06, "loss": 0.9172, "step": 2130 }, { "epoch": 2.16, "grad_norm": 83.52637481689453, "learning_rate": 9.994884469604913e-06, "loss": 0.9227, "step": 2160 }, { "epoch": 2.19, "grad_norm": 69.61637115478516, "learning_rate": 9.99474140983228e-06, "loss": 0.9325, "step": 2190 }, { "epoch": 2.22, "grad_norm": 31.19818115234375, "learning_rate": 9.99459637821484e-06, "loss": 1.0868, "step": 2220 }, { "epoch": 2.25, "grad_norm": 104.18785858154297, "learning_rate": 9.994449374809851e-06, "loss": 1.0583, "step": 2250 }, { "epoch": 2.2800000000000002, "grad_norm": 55.10357666015625, "learning_rate": 9.994300399675342e-06, "loss": 1.1248, "step": 2280 }, { "epoch": 2.31, "grad_norm": 58.71828079223633, "learning_rate": 9.994149452870133e-06, "loss": 0.8934, "step": 2310 }, { "epoch": 2.34, "grad_norm": 186.6011505126953, "learning_rate": 9.993996534453812e-06, "loss": 1.0213, "step": 2340 }, { "epoch": 2.37, "grad_norm": 137.37557983398438, "learning_rate": 9.993841644486747e-06, "loss": 0.9938, "step": 2370 }, { "epoch": 2.4, "grad_norm": 44.16426467895508, "learning_rate": 9.99368478303009e-06, "loss": 0.8664, "step": 2400 }, { "epoch": 2.43, "grad_norm": 208.6126708984375, "learning_rate": 9.993525950145761e-06, "loss": 1.0559, "step": 2430 }, { "epoch": 2.46, "grad_norm": 64.49674224853516, "learning_rate": 9.993365145896473e-06, "loss": 0.9284, "step": 2460 }, { "epoch": 2.49, "grad_norm": 85.07061004638672, "learning_rate": 9.993202370345705e-06, "loss": 0.9254, "step": 2490 }, { "epoch": 2.52, "grad_norm": 104.88370513916016, "learning_rate": 9.993037623557716e-06, "loss": 1.0071, "step": 2520 }, { "epoch": 2.55, "grad_norm": 58.13627624511719, "learning_rate": 9.992870905597549e-06, "loss": 0.8822, "step": 2550 }, { "epoch": 2.58, "grad_norm": 39.32762908935547, "learning_rate": 9.99270221653102e-06, "loss": 0.9391, "step": 2580 }, { "epoch": 2.61, "grad_norm": 64.0282974243164, "learning_rate": 9.992531556424726e-06, "loss": 0.9914, "step": 2610 }, { "epoch": 2.64, "grad_norm": 51.778141021728516, "learning_rate": 9.99235892534604e-06, "loss": 0.9049, "step": 2640 }, { "epoch": 2.67, "grad_norm": 29.94393539428711, "learning_rate": 9.992184323363112e-06, "loss": 0.9853, "step": 2670 }, { "epoch": 2.7, "grad_norm": 26.890628814697266, "learning_rate": 9.992007750544876e-06, "loss": 0.8851, "step": 2700 }, { "epoch": 2.73, "grad_norm": 55.287662506103516, "learning_rate": 9.991829206961038e-06, "loss": 1.0641, "step": 2730 }, { "epoch": 2.76, "grad_norm": 67.6606674194336, "learning_rate": 9.991648692682083e-06, "loss": 0.9312, "step": 2760 }, { "epoch": 2.79, "grad_norm": 85.00022888183594, "learning_rate": 9.991466207779279e-06, "loss": 0.782, "step": 2790 }, { "epoch": 2.82, "grad_norm": 98.71064758300781, "learning_rate": 9.991281752324664e-06, "loss": 0.9597, "step": 2820 }, { "epoch": 2.85, "grad_norm": 92.63137817382812, "learning_rate": 9.991095326391061e-06, "loss": 0.937, "step": 2850 }, { "epoch": 2.88, "grad_norm": 44.04415512084961, "learning_rate": 9.990906930052065e-06, "loss": 1.0176, "step": 2880 }, { "epoch": 2.91, "grad_norm": 165.56471252441406, "learning_rate": 9.990716563382055e-06, "loss": 0.8975, "step": 2910 }, { "epoch": 2.94, "grad_norm": 224.167236328125, "learning_rate": 9.990524226456182e-06, "loss": 1.1667, "step": 2940 }, { "epoch": 2.9699999999999998, "grad_norm": 76.42039489746094, "learning_rate": 9.990329919350382e-06, "loss": 1.0253, "step": 2970 }, { "epoch": 3.0, "grad_norm": 35.52342224121094, "learning_rate": 9.990133642141359e-06, "loss": 0.9168, "step": 3000 }, { "epoch": 3.0, "eval_loss": 0.9335120916366577, "eval_map": 0.3008, "eval_map_50": 0.4485, "eval_map_75": 0.341, "eval_map_chicken": 0.2437, "eval_map_duck": 0.0, "eval_map_large": 0.3019, "eval_map_medium": 0.2511, "eval_map_plant": 0.6588, "eval_map_small": 0.0694, "eval_mar_1": 0.1107, "eval_mar_10": 0.4104, "eval_mar_100": 0.4208, "eval_mar_100_chicken": 0.534, "eval_mar_100_duck": 0.0, "eval_mar_100_plant": 0.7284, "eval_mar_large": 0.4023, "eval_mar_medium": 0.3965, "eval_mar_small": 0.1154, "eval_runtime": 13.12, "eval_samples_per_second": 15.244, "eval_steps_per_second": 1.905, "step": 3000 }, { "epoch": 3.03, "grad_norm": 40.63835906982422, "learning_rate": 9.989935394906602e-06, "loss": 0.8388, "step": 3030 }, { "epoch": 3.06, "grad_norm": 162.6512451171875, "learning_rate": 9.989735177724378e-06, "loss": 0.8675, "step": 3060 }, { "epoch": 3.09, "grad_norm": 371.4659118652344, "learning_rate": 9.989532990673729e-06, "loss": 1.07, "step": 3090 }, { "epoch": 3.12, "grad_norm": 49.209171295166016, "learning_rate": 9.989328833834472e-06, "loss": 0.9865, "step": 3120 }, { "epoch": 3.15, "grad_norm": 66.55098724365234, "learning_rate": 9.98912270728721e-06, "loss": 0.8362, "step": 3150 }, { "epoch": 3.18, "grad_norm": 46.931678771972656, "learning_rate": 9.988914611113311e-06, "loss": 0.7478, "step": 3180 }, { "epoch": 3.21, "grad_norm": 0.017829274758696556, "learning_rate": 9.988704545394936e-06, "loss": 0.9651, "step": 3210 }, { "epoch": 3.24, "grad_norm": 70.68403625488281, "learning_rate": 9.988492510215011e-06, "loss": 1.0492, "step": 3240 }, { "epoch": 3.27, "grad_norm": 222.65328979492188, "learning_rate": 9.988278505657247e-06, "loss": 0.845, "step": 3270 }, { "epoch": 3.3, "grad_norm": 41.1923713684082, "learning_rate": 9.988062531806127e-06, "loss": 0.9929, "step": 3300 }, { "epoch": 3.33, "grad_norm": 29.829824447631836, "learning_rate": 9.987844588746916e-06, "loss": 0.8909, "step": 3330 }, { "epoch": 3.36, "grad_norm": 80.94469451904297, "learning_rate": 9.987624676565652e-06, "loss": 1.0083, "step": 3360 }, { "epoch": 3.39, "grad_norm": 47.34257507324219, "learning_rate": 9.987402795349154e-06, "loss": 0.8306, "step": 3390 }, { "epoch": 3.42, "grad_norm": 66.8982162475586, "learning_rate": 9.987178945185019e-06, "loss": 0.924, "step": 3420 }, { "epoch": 3.45, "grad_norm": 41.35850524902344, "learning_rate": 9.98695312616162e-06, "loss": 0.9452, "step": 3450 }, { "epoch": 3.48, "grad_norm": 34.22974395751953, "learning_rate": 9.986725338368103e-06, "loss": 0.855, "step": 3480 }, { "epoch": 3.51, "grad_norm": 0.014032869599759579, "learning_rate": 9.986495581894396e-06, "loss": 0.8999, "step": 3510 }, { "epoch": 3.54, "grad_norm": 43.12446594238281, "learning_rate": 9.986263856831204e-06, "loss": 0.9319, "step": 3540 }, { "epoch": 3.57, "grad_norm": 129.39205932617188, "learning_rate": 9.986030163270011e-06, "loss": 0.8143, "step": 3570 }, { "epoch": 3.6, "grad_norm": 202.19500732421875, "learning_rate": 9.98579450130307e-06, "loss": 0.9833, "step": 3600 }, { "epoch": 3.63, "grad_norm": 68.46711730957031, "learning_rate": 9.98555687102342e-06, "loss": 0.923, "step": 3630 }, { "epoch": 3.66, "grad_norm": 68.88642883300781, "learning_rate": 9.985317272524876e-06, "loss": 0.9257, "step": 3660 }, { "epoch": 3.69, "grad_norm": 34.893184661865234, "learning_rate": 9.985075705902024e-06, "loss": 0.9542, "step": 3690 }, { "epoch": 3.7199999999999998, "grad_norm": 182.26841735839844, "learning_rate": 9.98483217125023e-06, "loss": 0.7825, "step": 3720 }, { "epoch": 3.75, "grad_norm": 31.906808853149414, "learning_rate": 9.984586668665641e-06, "loss": 0.8616, "step": 3750 }, { "epoch": 3.7800000000000002, "grad_norm": 114.5857925415039, "learning_rate": 9.984339198245175e-06, "loss": 0.9343, "step": 3780 }, { "epoch": 3.81, "grad_norm": 47.947818756103516, "learning_rate": 9.984089760086531e-06, "loss": 0.9681, "step": 3810 }, { "epoch": 3.84, "grad_norm": 98.68987274169922, "learning_rate": 9.983838354288181e-06, "loss": 1.0136, "step": 3840 }, { "epoch": 3.87, "grad_norm": 80.91329956054688, "learning_rate": 9.98358498094938e-06, "loss": 0.7334, "step": 3870 }, { "epoch": 3.9, "grad_norm": 88.02969360351562, "learning_rate": 9.98332964017015e-06, "loss": 0.8853, "step": 3900 }, { "epoch": 3.93, "grad_norm": 93.37176513671875, "learning_rate": 9.9830723320513e-06, "loss": 1.0608, "step": 3930 }, { "epoch": 3.96, "grad_norm": 78.63481903076172, "learning_rate": 9.982813056694411e-06, "loss": 0.7254, "step": 3960 }, { "epoch": 3.99, "grad_norm": 84.25390625, "learning_rate": 9.98255181420184e-06, "loss": 0.8732, "step": 3990 }, { "epoch": 4.0, "eval_loss": 0.8629758954048157, "eval_map": 0.3268, "eval_map_50": 0.4634, "eval_map_75": 0.3815, "eval_map_chicken": 0.2845, "eval_map_duck": 0.0, "eval_map_large": 0.3552, "eval_map_medium": 0.2883, "eval_map_plant": 0.6958, "eval_map_small": 0.0599, "eval_mar_1": 0.1214, "eval_mar_10": 0.4573, "eval_mar_100": 0.49, "eval_mar_100_chicken": 0.7165, "eval_mar_100_duck": 0.0, "eval_mar_100_plant": 0.7536, "eval_mar_large": 0.5136, "eval_mar_medium": 0.4629, "eval_mar_small": 0.1374, "eval_runtime": 12.3779, "eval_samples_per_second": 16.158, "eval_steps_per_second": 2.02, "step": 4000 }, { "epoch": 4.02, "grad_norm": 93.95651245117188, "learning_rate": 9.982288604676719e-06, "loss": 0.8475, "step": 4020 }, { "epoch": 4.05, "grad_norm": 116.90863800048828, "learning_rate": 9.982023428222963e-06, "loss": 0.9066, "step": 4050 }, { "epoch": 4.08, "grad_norm": 50.993064880371094, "learning_rate": 9.981756284945256e-06, "loss": 0.8973, "step": 4080 }, { "epoch": 4.11, "grad_norm": 63.041656494140625, "learning_rate": 9.981487174949065e-06, "loss": 0.7599, "step": 4110 }, { "epoch": 4.14, "grad_norm": 70.02943420410156, "learning_rate": 9.98121609834063e-06, "loss": 0.6971, "step": 4140 }, { "epoch": 4.17, "grad_norm": 101.08905029296875, "learning_rate": 9.980943055226964e-06, "loss": 0.7628, "step": 4170 }, { "epoch": 4.2, "grad_norm": 65.84382629394531, "learning_rate": 9.980668045715864e-06, "loss": 0.8121, "step": 4200 }, { "epoch": 4.23, "grad_norm": 125.61593627929688, "learning_rate": 9.980391069915897e-06, "loss": 0.7737, "step": 4230 }, { "epoch": 4.26, "grad_norm": 69.36189270019531, "learning_rate": 9.98011212793641e-06, "loss": 0.7853, "step": 4260 }, { "epoch": 4.29, "grad_norm": 50.37016677856445, "learning_rate": 9.979831219887526e-06, "loss": 0.7914, "step": 4290 }, { "epoch": 4.32, "grad_norm": 58.28656005859375, "learning_rate": 9.979548345880142e-06, "loss": 0.8652, "step": 4320 }, { "epoch": 4.35, "grad_norm": 0.49502643942832947, "learning_rate": 9.97926350602593e-06, "loss": 0.8194, "step": 4350 }, { "epoch": 4.38, "grad_norm": 41.404151916503906, "learning_rate": 9.978976700437341e-06, "loss": 1.1021, "step": 4380 }, { "epoch": 4.41, "grad_norm": 35.13054275512695, "learning_rate": 9.978687929227606e-06, "loss": 0.7529, "step": 4410 }, { "epoch": 4.44, "grad_norm": 63.76109313964844, "learning_rate": 9.978397192510722e-06, "loss": 0.8391, "step": 4440 }, { "epoch": 4.47, "grad_norm": 40.27415084838867, "learning_rate": 9.978104490401468e-06, "loss": 0.8856, "step": 4470 }, { "epoch": 4.5, "grad_norm": 66.84982299804688, "learning_rate": 9.9778098230154e-06, "loss": 0.7131, "step": 4500 }, { "epoch": 4.53, "grad_norm": 38.403099060058594, "learning_rate": 9.977513190468848e-06, "loss": 0.8658, "step": 4530 }, { "epoch": 4.5600000000000005, "grad_norm": 180.49864196777344, "learning_rate": 9.977214592878917e-06, "loss": 0.8249, "step": 4560 }, { "epoch": 4.59, "grad_norm": 49.14180374145508, "learning_rate": 9.976914030363488e-06, "loss": 0.738, "step": 4590 }, { "epoch": 4.62, "grad_norm": 48.01270294189453, "learning_rate": 9.976611503041218e-06, "loss": 0.8032, "step": 4620 }, { "epoch": 4.65, "grad_norm": 68.6293716430664, "learning_rate": 9.976307011031542e-06, "loss": 1.0592, "step": 4650 }, { "epoch": 4.68, "grad_norm": 56.11207962036133, "learning_rate": 9.976000554454668e-06, "loss": 0.855, "step": 4680 }, { "epoch": 4.71, "grad_norm": 63.52250671386719, "learning_rate": 9.975692133431579e-06, "loss": 0.7616, "step": 4710 }, { "epoch": 4.74, "grad_norm": 38.014957427978516, "learning_rate": 9.975381748084035e-06, "loss": 0.9962, "step": 4740 }, { "epoch": 4.77, "grad_norm": 287.0184326171875, "learning_rate": 9.975069398534574e-06, "loss": 0.7554, "step": 4770 }, { "epoch": 4.8, "grad_norm": 68.91691589355469, "learning_rate": 9.974755084906503e-06, "loss": 0.831, "step": 4800 }, { "epoch": 4.83, "grad_norm": 40.459022521972656, "learning_rate": 9.974438807323907e-06, "loss": 0.8852, "step": 4830 }, { "epoch": 4.86, "grad_norm": 40.74941635131836, "learning_rate": 9.974120565911653e-06, "loss": 0.6567, "step": 4860 }, { "epoch": 4.89, "grad_norm": 26.657499313354492, "learning_rate": 9.973800360795372e-06, "loss": 0.8009, "step": 4890 }, { "epoch": 4.92, "grad_norm": 101.05900573730469, "learning_rate": 9.97347819210148e-06, "loss": 0.7003, "step": 4920 }, { "epoch": 4.95, "grad_norm": 57.03788757324219, "learning_rate": 9.973154059957162e-06, "loss": 0.8258, "step": 4950 }, { "epoch": 4.98, "grad_norm": 83.70702362060547, "learning_rate": 9.972827964490382e-06, "loss": 0.7036, "step": 4980 }, { "epoch": 5.0, "eval_loss": 0.7594767808914185, "eval_map": 0.3359, "eval_map_50": 0.4642, "eval_map_75": 0.3931, "eval_map_chicken": 0.2899, "eval_map_duck": 0.0, "eval_map_large": 0.3764, "eval_map_medium": 0.2912, "eval_map_plant": 0.7178, "eval_map_small": 0.0739, "eval_mar_1": 0.1255, "eval_mar_10": 0.4737, "eval_mar_100": 0.5037, "eval_mar_100_chicken": 0.7404, "eval_mar_100_duck": 0.0, "eval_mar_100_plant": 0.7707, "eval_mar_large": 0.5219, "eval_mar_medium": 0.4693, "eval_mar_small": 0.1679, "eval_runtime": 12.1395, "eval_samples_per_second": 16.475, "eval_steps_per_second": 2.059, "step": 5000 }, { "epoch": 5.01, "grad_norm": 64.3818130493164, "learning_rate": 9.972499905829874e-06, "loss": 0.7193, "step": 5010 }, { "epoch": 5.04, "grad_norm": 33.02555465698242, "learning_rate": 9.972169884105155e-06, "loss": 0.7268, "step": 5040 }, { "epoch": 5.07, "grad_norm": 96.5431900024414, "learning_rate": 9.971837899446505e-06, "loss": 0.8703, "step": 5070 }, { "epoch": 5.1, "grad_norm": 48.66056823730469, "learning_rate": 9.971503951984996e-06, "loss": 0.7031, "step": 5100 }, { "epoch": 5.13, "grad_norm": 54.251007080078125, "learning_rate": 9.971168041852456e-06, "loss": 0.8166, "step": 5130 }, { "epoch": 5.16, "grad_norm": 36.01968002319336, "learning_rate": 9.970830169181504e-06, "loss": 0.7903, "step": 5160 }, { "epoch": 5.19, "grad_norm": 39.61149215698242, "learning_rate": 9.970490334105525e-06, "loss": 0.8338, "step": 5190 }, { "epoch": 5.22, "grad_norm": 22.683454513549805, "learning_rate": 9.970148536758678e-06, "loss": 0.6444, "step": 5220 }, { "epoch": 5.25, "grad_norm": 44.20682907104492, "learning_rate": 9.9698047772759e-06, "loss": 0.8153, "step": 5250 }, { "epoch": 5.28, "grad_norm": 44.122798919677734, "learning_rate": 9.969459055792903e-06, "loss": 0.7408, "step": 5280 }, { "epoch": 5.31, "grad_norm": 53.325653076171875, "learning_rate": 9.969111372446171e-06, "loss": 0.8432, "step": 5310 }, { "epoch": 5.34, "grad_norm": 46.1613655090332, "learning_rate": 9.968761727372965e-06, "loss": 0.8252, "step": 5340 }, { "epoch": 5.37, "grad_norm": 64.1738510131836, "learning_rate": 9.968410120711321e-06, "loss": 0.7511, "step": 5370 }, { "epoch": 5.4, "grad_norm": 94.20152282714844, "learning_rate": 9.968056552600043e-06, "loss": 0.686, "step": 5400 }, { "epoch": 5.43, "grad_norm": 44.39435577392578, "learning_rate": 9.967701023178717e-06, "loss": 0.7301, "step": 5430 }, { "epoch": 5.46, "grad_norm": 45.8010368347168, "learning_rate": 9.967343532587701e-06, "loss": 0.6625, "step": 5460 }, { "epoch": 5.49, "grad_norm": 71.59171295166016, "learning_rate": 9.966984080968128e-06, "loss": 1.0089, "step": 5490 }, { "epoch": 5.52, "grad_norm": 60.024173736572266, "learning_rate": 9.966622668461899e-06, "loss": 0.9506, "step": 5520 }, { "epoch": 5.55, "grad_norm": 60.47734069824219, "learning_rate": 9.966259295211698e-06, "loss": 0.7548, "step": 5550 }, { "epoch": 5.58, "grad_norm": 153.8086700439453, "learning_rate": 9.965893961360977e-06, "loss": 0.8064, "step": 5580 }, { "epoch": 5.61, "grad_norm": 45.45552444458008, "learning_rate": 9.965526667053964e-06, "loss": 0.6659, "step": 5610 }, { "epoch": 5.64, "grad_norm": 55.69154739379883, "learning_rate": 9.965157412435663e-06, "loss": 0.8157, "step": 5640 }, { "epoch": 5.67, "grad_norm": 36.18803787231445, "learning_rate": 9.964786197651848e-06, "loss": 1.0301, "step": 5670 }, { "epoch": 5.7, "grad_norm": 61.78434371948242, "learning_rate": 9.964413022849069e-06, "loss": 0.7108, "step": 5700 }, { "epoch": 5.73, "grad_norm": 58.902870178222656, "learning_rate": 9.96403788817465e-06, "loss": 1.0545, "step": 5730 }, { "epoch": 5.76, "grad_norm": 97.89854431152344, "learning_rate": 9.963660793776689e-06, "loss": 0.8038, "step": 5760 }, { "epoch": 5.79, "grad_norm": 37.48524475097656, "learning_rate": 9.963281739804054e-06, "loss": 0.8685, "step": 5790 }, { "epoch": 5.82, "grad_norm": 96.37757110595703, "learning_rate": 9.96290072640639e-06, "loss": 0.6926, "step": 5820 }, { "epoch": 5.85, "grad_norm": 34.71822738647461, "learning_rate": 9.96251775373412e-06, "loss": 0.7277, "step": 5850 }, { "epoch": 5.88, "grad_norm": 62.3222541809082, "learning_rate": 9.96213282193843e-06, "loss": 0.7629, "step": 5880 }, { "epoch": 5.91, "grad_norm": 68.81719970703125, "learning_rate": 9.961745931171288e-06, "loss": 0.9502, "step": 5910 }, { "epoch": 5.9399999999999995, "grad_norm": 75.99760437011719, "learning_rate": 9.96135708158543e-06, "loss": 0.6638, "step": 5940 }, { "epoch": 5.97, "grad_norm": 58.565250396728516, "learning_rate": 9.96096627333437e-06, "loss": 0.7397, "step": 5970 }, { "epoch": 6.0, "grad_norm": 70.04283905029297, "learning_rate": 9.960573506572391e-06, "loss": 0.979, "step": 6000 }, { "epoch": 6.0, "eval_loss": 0.7068777680397034, "eval_map": 0.3728, "eval_map_50": 0.5214, "eval_map_75": 0.4438, "eval_map_chicken": 0.4058, "eval_map_duck": 0.0, "eval_map_large": 0.4012, "eval_map_medium": 0.3332, "eval_map_plant": 0.7125, "eval_map_small": 0.0768, "eval_mar_1": 0.1296, "eval_mar_10": 0.4833, "eval_mar_100": 0.4882, "eval_mar_100_chicken": 0.7012, "eval_mar_100_duck": 0.0, "eval_mar_100_plant": 0.7633, "eval_mar_large": 0.5065, "eval_mar_medium": 0.4582, "eval_mar_small": 0.1172, "eval_runtime": 12.6646, "eval_samples_per_second": 15.792, "eval_steps_per_second": 1.974, "step": 6000 }, { "epoch": 6.03, "grad_norm": 74.1376724243164, "learning_rate": 9.96017878145455e-06, "loss": 0.8877, "step": 6030 }, { "epoch": 6.06, "grad_norm": 133.20069885253906, "learning_rate": 9.959782098136683e-06, "loss": 0.7131, "step": 6060 }, { "epoch": 6.09, "grad_norm": 109.30853271484375, "learning_rate": 9.959383456775392e-06, "loss": 0.7067, "step": 6090 }, { "epoch": 6.12, "grad_norm": 49.84526824951172, "learning_rate": 9.958982857528053e-06, "loss": 0.7361, "step": 6120 }, { "epoch": 6.15, "grad_norm": 45.34609603881836, "learning_rate": 9.958580300552816e-06, "loss": 0.7969, "step": 6150 }, { "epoch": 6.18, "grad_norm": 46.07512664794922, "learning_rate": 9.958175786008605e-06, "loss": 0.9145, "step": 6180 }, { "epoch": 6.21, "grad_norm": 122.06670379638672, "learning_rate": 9.957769314055117e-06, "loss": 0.657, "step": 6210 }, { "epoch": 6.24, "grad_norm": 86.56077575683594, "learning_rate": 9.957360884852819e-06, "loss": 0.7242, "step": 6240 }, { "epoch": 6.27, "grad_norm": 136.27645874023438, "learning_rate": 9.956950498562954e-06, "loss": 0.6895, "step": 6270 }, { "epoch": 6.3, "grad_norm": 56.636924743652344, "learning_rate": 9.956538155347534e-06, "loss": 0.6644, "step": 6300 }, { "epoch": 6.33, "grad_norm": 36.84674835205078, "learning_rate": 9.956123855369346e-06, "loss": 0.8481, "step": 6330 }, { "epoch": 6.36, "grad_norm": 233.78173828125, "learning_rate": 9.955707598791952e-06, "loss": 0.7379, "step": 6360 }, { "epoch": 6.39, "grad_norm": 28.2652587890625, "learning_rate": 9.955289385779681e-06, "loss": 0.8178, "step": 6390 }, { "epoch": 6.42, "grad_norm": 89.63736724853516, "learning_rate": 9.954869216497636e-06, "loss": 0.7235, "step": 6420 }, { "epoch": 6.45, "grad_norm": 213.698486328125, "learning_rate": 9.954447091111695e-06, "loss": 0.6498, "step": 6450 }, { "epoch": 6.48, "grad_norm": 64.32272338867188, "learning_rate": 9.954023009788505e-06, "loss": 0.6848, "step": 6480 }, { "epoch": 6.51, "grad_norm": 0.002004330512136221, "learning_rate": 9.953596972695487e-06, "loss": 0.6877, "step": 6510 }, { "epoch": 6.54, "grad_norm": 69.6749038696289, "learning_rate": 9.953168980000836e-06, "loss": 0.8507, "step": 6540 }, { "epoch": 6.57, "grad_norm": 43.652671813964844, "learning_rate": 9.952739031873513e-06, "loss": 0.691, "step": 6570 }, { "epoch": 6.6, "grad_norm": 31.398818969726562, "learning_rate": 9.952307128483257e-06, "loss": 0.7495, "step": 6600 }, { "epoch": 6.63, "grad_norm": 77.4420394897461, "learning_rate": 9.951873270000576e-06, "loss": 0.6521, "step": 6630 }, { "epoch": 6.66, "grad_norm": 96.93862915039062, "learning_rate": 9.951437456596751e-06, "loss": 0.6123, "step": 6660 }, { "epoch": 6.6899999999999995, "grad_norm": 29.853755950927734, "learning_rate": 9.950999688443833e-06, "loss": 0.6852, "step": 6690 }, { "epoch": 6.72, "grad_norm": 24.771831512451172, "learning_rate": 9.950559965714647e-06, "loss": 0.6546, "step": 6720 }, { "epoch": 6.75, "grad_norm": 156.53634643554688, "learning_rate": 9.95011828858279e-06, "loss": 0.7306, "step": 6750 }, { "epoch": 6.78, "grad_norm": 37.48930358886719, "learning_rate": 9.949674657222624e-06, "loss": 0.6974, "step": 6780 }, { "epoch": 6.8100000000000005, "grad_norm": 128.7861328125, "learning_rate": 9.949229071809294e-06, "loss": 0.6849, "step": 6810 }, { "epoch": 6.84, "grad_norm": 32.026039123535156, "learning_rate": 9.948781532518706e-06, "loss": 0.7232, "step": 6840 }, { "epoch": 6.87, "grad_norm": 59.086326599121094, "learning_rate": 9.948332039527541e-06, "loss": 0.6455, "step": 6870 }, { "epoch": 6.9, "grad_norm": 48.41843795776367, "learning_rate": 9.947880593013256e-06, "loss": 0.7087, "step": 6900 }, { "epoch": 6.93, "grad_norm": 80.14769744873047, "learning_rate": 9.94742719315407e-06, "loss": 0.9707, "step": 6930 }, { "epoch": 6.96, "grad_norm": 0.0015557038132101297, "learning_rate": 9.946971840128982e-06, "loss": 0.535, "step": 6960 }, { "epoch": 6.99, "grad_norm": 123.63168334960938, "learning_rate": 9.946514534117755e-06, "loss": 0.7254, "step": 6990 }, { "epoch": 7.0, "eval_loss": 0.6566085815429688, "eval_map": 0.3939, "eval_map_50": 0.5385, "eval_map_75": 0.4648, "eval_map_chicken": 0.4487, "eval_map_duck": 0.0, "eval_map_large": 0.4192, "eval_map_medium": 0.3571, "eval_map_plant": 0.7331, "eval_map_small": 0.0613, "eval_mar_1": 0.1369, "eval_mar_10": 0.4969, "eval_mar_100": 0.5026, "eval_mar_100_chicken": 0.7254, "eval_mar_100_duck": 0.0, "eval_mar_100_plant": 0.7825, "eval_mar_large": 0.5184, "eval_mar_medium": 0.4737, "eval_mar_small": 0.153, "eval_runtime": 12.4857, "eval_samples_per_second": 16.018, "eval_steps_per_second": 2.002, "step": 7000 }, { "epoch": 7.02, "grad_norm": 65.63529205322266, "learning_rate": 9.946055275300929e-06, "loss": 0.7078, "step": 7020 }, { "epoch": 7.05, "grad_norm": 24.253419876098633, "learning_rate": 9.94559406385981e-06, "loss": 0.7273, "step": 7050 }, { "epoch": 7.08, "grad_norm": 81.4408950805664, "learning_rate": 9.945130899976477e-06, "loss": 0.6802, "step": 7080 }, { "epoch": 7.11, "grad_norm": 46.03237533569336, "learning_rate": 9.944665783833782e-06, "loss": 0.7239, "step": 7110 }, { "epoch": 7.14, "grad_norm": 59.89938735961914, "learning_rate": 9.944198715615343e-06, "loss": 0.6658, "step": 7140 }, { "epoch": 7.17, "grad_norm": 56.362300872802734, "learning_rate": 9.943729695505552e-06, "loss": 0.6944, "step": 7170 }, { "epoch": 7.2, "grad_norm": 32.118507385253906, "learning_rate": 9.94325872368957e-06, "loss": 0.6362, "step": 7200 }, { "epoch": 7.23, "grad_norm": 92.2921142578125, "learning_rate": 9.942785800353332e-06, "loss": 0.6866, "step": 7230 }, { "epoch": 7.26, "grad_norm": 50.029296875, "learning_rate": 9.942310925683538e-06, "loss": 0.7707, "step": 7260 }, { "epoch": 7.29, "grad_norm": 0.001584346522577107, "learning_rate": 9.94183409986766e-06, "loss": 0.7759, "step": 7290 }, { "epoch": 7.32, "grad_norm": 104.0340805053711, "learning_rate": 9.941355323093944e-06, "loss": 0.6975, "step": 7320 }, { "epoch": 7.35, "grad_norm": 56.41421890258789, "learning_rate": 9.940874595551403e-06, "loss": 0.6543, "step": 7350 }, { "epoch": 7.38, "grad_norm": 33.60560607910156, "learning_rate": 9.94039191742982e-06, "loss": 0.612, "step": 7380 }, { "epoch": 7.41, "grad_norm": 81.51013946533203, "learning_rate": 9.939907288919749e-06, "loss": 0.5793, "step": 7410 }, { "epoch": 7.44, "grad_norm": 248.1967010498047, "learning_rate": 9.939420710212511e-06, "loss": 0.5627, "step": 7440 }, { "epoch": 7.47, "grad_norm": 24.327590942382812, "learning_rate": 9.938932181500206e-06, "loss": 0.6923, "step": 7470 }, { "epoch": 7.5, "grad_norm": 34.124855041503906, "learning_rate": 9.938441702975689e-06, "loss": 0.7954, "step": 7500 }, { "epoch": 7.53, "grad_norm": 39.90071105957031, "learning_rate": 9.9379492748326e-06, "loss": 0.6371, "step": 7530 }, { "epoch": 7.5600000000000005, "grad_norm": 94.69256591796875, "learning_rate": 9.937454897265338e-06, "loss": 0.6343, "step": 7560 }, { "epoch": 7.59, "grad_norm": 35.00869369506836, "learning_rate": 9.936958570469077e-06, "loss": 0.7949, "step": 7590 }, { "epoch": 7.62, "grad_norm": 63.846656799316406, "learning_rate": 9.93646029463976e-06, "loss": 0.6559, "step": 7620 }, { "epoch": 7.65, "grad_norm": 51.685035705566406, "learning_rate": 9.935960069974096e-06, "loss": 0.6802, "step": 7650 }, { "epoch": 7.68, "grad_norm": 211.3177032470703, "learning_rate": 9.935457896669568e-06, "loss": 0.6446, "step": 7680 }, { "epoch": 7.71, "grad_norm": 105.8477783203125, "learning_rate": 9.934953774924425e-06, "loss": 0.6016, "step": 7710 }, { "epoch": 7.74, "grad_norm": 109.14087677001953, "learning_rate": 9.934447704937684e-06, "loss": 0.6017, "step": 7740 }, { "epoch": 7.77, "grad_norm": 82.0282974243164, "learning_rate": 9.933939686909137e-06, "loss": 0.754, "step": 7770 }, { "epoch": 7.8, "grad_norm": 102.84513854980469, "learning_rate": 9.93342972103934e-06, "loss": 0.7222, "step": 7800 }, { "epoch": 7.83, "grad_norm": 83.25312805175781, "learning_rate": 9.93291780752962e-06, "loss": 0.6955, "step": 7830 }, { "epoch": 7.86, "grad_norm": 74.26616668701172, "learning_rate": 9.932403946582071e-06, "loss": 0.7081, "step": 7860 }, { "epoch": 7.89, "grad_norm": 67.56425476074219, "learning_rate": 9.931888138399562e-06, "loss": 0.6676, "step": 7890 }, { "epoch": 7.92, "grad_norm": 79.110595703125, "learning_rate": 9.931370383185717e-06, "loss": 0.6087, "step": 7920 }, { "epoch": 7.95, "grad_norm": 55.531272888183594, "learning_rate": 9.930850681144946e-06, "loss": 0.6725, "step": 7950 }, { "epoch": 7.98, "grad_norm": 53.5663948059082, "learning_rate": 9.930329032482412e-06, "loss": 0.6696, "step": 7980 }, { "epoch": 8.0, "eval_loss": 0.6275544166564941, "eval_map": 0.4213, "eval_map_50": 0.5797, "eval_map_75": 0.5037, "eval_map_chicken": 0.5311, "eval_map_duck": 0.0, "eval_map_large": 0.4437, "eval_map_medium": 0.3791, "eval_map_plant": 0.7327, "eval_map_small": 0.0573, "eval_mar_1": 0.1378, "eval_mar_10": 0.4964, "eval_mar_100": 0.4996, "eval_mar_100_chicken": 0.7203, "eval_mar_100_duck": 0.0, "eval_mar_100_plant": 0.7786, "eval_mar_large": 0.5195, "eval_mar_medium": 0.467, "eval_mar_small": 0.1033, "eval_runtime": 12.2441, "eval_samples_per_second": 16.334, "eval_steps_per_second": 2.042, "step": 8000 }, { "epoch": 8.01, "grad_norm": 55.83026123046875, "learning_rate": 9.929805437404061e-06, "loss": 0.7421, "step": 8010 }, { "epoch": 8.04, "grad_norm": 45.21809005737305, "learning_rate": 9.929279896116595e-06, "loss": 0.7358, "step": 8040 }, { "epoch": 8.07, "grad_norm": 38.08073043823242, "learning_rate": 9.92875240882749e-06, "loss": 0.5985, "step": 8070 }, { "epoch": 8.1, "grad_norm": 50.15561294555664, "learning_rate": 9.928222975744992e-06, "loss": 0.8388, "step": 8100 }, { "epoch": 8.13, "grad_norm": 314.2345275878906, "learning_rate": 9.927691597078109e-06, "loss": 0.8124, "step": 8130 }, { "epoch": 8.16, "grad_norm": 62.64537048339844, "learning_rate": 9.927158273036624e-06, "loss": 0.8133, "step": 8160 }, { "epoch": 8.19, "grad_norm": 30.060697555541992, "learning_rate": 9.926623003831085e-06, "loss": 0.7182, "step": 8190 }, { "epoch": 8.22, "grad_norm": 47.56553649902344, "learning_rate": 9.926085789672806e-06, "loss": 0.6734, "step": 8220 }, { "epoch": 8.25, "grad_norm": 87.50015258789062, "learning_rate": 9.92554663077387e-06, "loss": 0.7023, "step": 8250 }, { "epoch": 8.28, "grad_norm": 55.88727951049805, "learning_rate": 9.925005527347132e-06, "loss": 0.7398, "step": 8280 }, { "epoch": 8.31, "grad_norm": 55.43495559692383, "learning_rate": 9.924462479606207e-06, "loss": 0.5943, "step": 8310 }, { "epoch": 8.34, "grad_norm": 43.17909240722656, "learning_rate": 9.923917487765484e-06, "loss": 0.6705, "step": 8340 }, { "epoch": 8.37, "grad_norm": 33.53215789794922, "learning_rate": 9.923370552040117e-06, "loss": 0.8007, "step": 8370 }, { "epoch": 8.4, "grad_norm": 293.8423767089844, "learning_rate": 9.922821672646028e-06, "loss": 0.7027, "step": 8400 }, { "epoch": 8.43, "grad_norm": 62.3926887512207, "learning_rate": 9.922270849799903e-06, "loss": 0.645, "step": 8430 }, { "epoch": 8.46, "grad_norm": 47.301029205322266, "learning_rate": 9.921718083719203e-06, "loss": 0.6276, "step": 8460 }, { "epoch": 8.49, "grad_norm": 62.63656234741211, "learning_rate": 9.921163374622147e-06, "loss": 0.6349, "step": 8490 }, { "epoch": 8.52, "grad_norm": 73.08213806152344, "learning_rate": 9.920606722727726e-06, "loss": 0.7825, "step": 8520 }, { "epoch": 8.55, "grad_norm": 36.310306549072266, "learning_rate": 9.920048128255699e-06, "loss": 0.621, "step": 8550 }, { "epoch": 8.58, "grad_norm": 42.77394485473633, "learning_rate": 9.919487591426591e-06, "loss": 0.7919, "step": 8580 }, { "epoch": 8.61, "grad_norm": 44.64291000366211, "learning_rate": 9.918925112461688e-06, "loss": 0.726, "step": 8610 }, { "epoch": 8.64, "grad_norm": 43.97916793823242, "learning_rate": 9.918360691583056e-06, "loss": 0.6377, "step": 8640 }, { "epoch": 8.67, "grad_norm": 137.5167236328125, "learning_rate": 9.917794329013511e-06, "loss": 0.677, "step": 8670 }, { "epoch": 8.7, "grad_norm": 38.42242431640625, "learning_rate": 9.91722602497665e-06, "loss": 0.5589, "step": 8700 }, { "epoch": 8.73, "grad_norm": 36.903079986572266, "learning_rate": 9.916655779696826e-06, "loss": 0.5643, "step": 8730 }, { "epoch": 8.76, "grad_norm": 72.70779418945312, "learning_rate": 9.916083593399167e-06, "loss": 0.621, "step": 8760 }, { "epoch": 8.79, "grad_norm": 125.7283706665039, "learning_rate": 9.91550946630956e-06, "loss": 0.6044, "step": 8790 }, { "epoch": 8.82, "grad_norm": 35.659515380859375, "learning_rate": 9.914933398654663e-06, "loss": 0.666, "step": 8820 }, { "epoch": 8.85, "grad_norm": 66.64339447021484, "learning_rate": 9.914355390661897e-06, "loss": 0.6251, "step": 8850 }, { "epoch": 8.88, "grad_norm": 60.74051284790039, "learning_rate": 9.913775442559451e-06, "loss": 0.6205, "step": 8880 }, { "epoch": 8.91, "grad_norm": 455.9437561035156, "learning_rate": 9.91319355457628e-06, "loss": 0.9078, "step": 8910 }, { "epoch": 8.94, "grad_norm": 116.52461242675781, "learning_rate": 9.912609726942104e-06, "loss": 0.6109, "step": 8940 }, { "epoch": 8.97, "grad_norm": 99.65922546386719, "learning_rate": 9.912023959887408e-06, "loss": 0.7021, "step": 8970 }, { "epoch": 9.0, "grad_norm": 57.34294509887695, "learning_rate": 9.911436253643445e-06, "loss": 0.6583, "step": 9000 }, { "epoch": 9.0, "eval_loss": 0.6020973920822144, "eval_map": 0.4331, "eval_map_50": 0.5873, "eval_map_75": 0.5108, "eval_map_chicken": 0.5463, "eval_map_duck": 0.0, "eval_map_large": 0.4557, "eval_map_medium": 0.3986, "eval_map_plant": 0.7528, "eval_map_small": 0.0796, "eval_mar_1": 0.1397, "eval_mar_10": 0.5046, "eval_mar_100": 0.5091, "eval_mar_100_chicken": 0.7264, "eval_mar_100_duck": 0.0, "eval_mar_100_plant": 0.8009, "eval_mar_large": 0.5262, "eval_mar_medium": 0.478, "eval_mar_small": 0.1539, "eval_runtime": 12.4759, "eval_samples_per_second": 16.031, "eval_steps_per_second": 2.004, "step": 9000 }, { "epoch": 9.03, "grad_norm": 49.462806701660156, "learning_rate": 9.910846608442229e-06, "loss": 0.6088, "step": 9030 }, { "epoch": 9.06, "grad_norm": 40.3636589050293, "learning_rate": 9.910255024516546e-06, "loss": 0.5836, "step": 9060 }, { "epoch": 9.09, "grad_norm": 45.10319519042969, "learning_rate": 9.909661502099943e-06, "loss": 0.62, "step": 9090 }, { "epoch": 9.12, "grad_norm": 60.286251068115234, "learning_rate": 9.909066041426733e-06, "loss": 0.8129, "step": 9120 }, { "epoch": 9.15, "grad_norm": 78.95413208007812, "learning_rate": 9.908468642731996e-06, "loss": 0.6351, "step": 9150 }, { "epoch": 9.18, "grad_norm": 35.29153060913086, "learning_rate": 9.907869306251571e-06, "loss": 0.7036, "step": 9180 }, { "epoch": 9.21, "grad_norm": 47.9434700012207, "learning_rate": 9.907268032222072e-06, "loss": 0.7898, "step": 9210 }, { "epoch": 9.24, "grad_norm": 46.56057357788086, "learning_rate": 9.906664820880869e-06, "loss": 0.7279, "step": 9240 }, { "epoch": 9.27, "grad_norm": 630.6410522460938, "learning_rate": 9.9060596724661e-06, "loss": 0.5917, "step": 9270 }, { "epoch": 9.3, "grad_norm": 29.160694122314453, "learning_rate": 9.90545258721667e-06, "loss": 0.585, "step": 9300 }, { "epoch": 9.33, "grad_norm": 111.9803237915039, "learning_rate": 9.904843565372249e-06, "loss": 0.7176, "step": 9330 }, { "epoch": 9.36, "grad_norm": 77.09111022949219, "learning_rate": 9.904232607173262e-06, "loss": 0.6292, "step": 9360 }, { "epoch": 9.39, "grad_norm": 104.63125610351562, "learning_rate": 9.903619712860912e-06, "loss": 0.7052, "step": 9390 }, { "epoch": 9.42, "grad_norm": 78.857177734375, "learning_rate": 9.903004882677157e-06, "loss": 0.5715, "step": 9420 }, { "epoch": 9.45, "grad_norm": 48.32848358154297, "learning_rate": 9.902388116864723e-06, "loss": 0.6449, "step": 9450 }, { "epoch": 9.48, "grad_norm": 45.87548065185547, "learning_rate": 9.9017694156671e-06, "loss": 0.64, "step": 9480 }, { "epoch": 9.51, "grad_norm": 389.7173767089844, "learning_rate": 9.90114877932854e-06, "loss": 0.6348, "step": 9510 }, { "epoch": 9.54, "grad_norm": 36.60053634643555, "learning_rate": 9.900526208094061e-06, "loss": 0.6961, "step": 9540 }, { "epoch": 9.57, "grad_norm": 88.65057373046875, "learning_rate": 9.899901702209445e-06, "loss": 0.6904, "step": 9570 }, { "epoch": 9.6, "grad_norm": 62.89002990722656, "learning_rate": 9.899275261921236e-06, "loss": 0.8458, "step": 9600 }, { "epoch": 9.63, "grad_norm": 53.33452224731445, "learning_rate": 9.898646887476742e-06, "loss": 0.6378, "step": 9630 }, { "epoch": 9.66, "grad_norm": 77.60774230957031, "learning_rate": 9.898016579124039e-06, "loss": 0.6957, "step": 9660 }, { "epoch": 9.69, "grad_norm": 36.759971618652344, "learning_rate": 9.897384337111956e-06, "loss": 0.6269, "step": 9690 }, { "epoch": 9.72, "grad_norm": 48.42854309082031, "learning_rate": 9.8967501616901e-06, "loss": 0.5723, "step": 9720 }, { "epoch": 9.75, "grad_norm": 61.33018493652344, "learning_rate": 9.89611405310883e-06, "loss": 0.6114, "step": 9750 }, { "epoch": 9.78, "grad_norm": 98.9421157836914, "learning_rate": 9.895476011619269e-06, "loss": 0.6784, "step": 9780 }, { "epoch": 9.81, "grad_norm": 36.58964538574219, "learning_rate": 9.89483603747331e-06, "loss": 0.8626, "step": 9810 }, { "epoch": 9.84, "grad_norm": 52.97993087768555, "learning_rate": 9.894194130923602e-06, "loss": 0.752, "step": 9840 }, { "epoch": 9.87, "grad_norm": 54.635459899902344, "learning_rate": 9.89355029222356e-06, "loss": 0.648, "step": 9870 }, { "epoch": 9.9, "grad_norm": 59.22683334350586, "learning_rate": 9.89290452162736e-06, "loss": 0.6415, "step": 9900 }, { "epoch": 9.93, "grad_norm": 70.11124420166016, "learning_rate": 9.892256819389947e-06, "loss": 0.665, "step": 9930 }, { "epoch": 9.96, "grad_norm": 120.08607482910156, "learning_rate": 9.891607185767018e-06, "loss": 0.6032, "step": 9960 }, { "epoch": 9.99, "grad_norm": 48.52140808105469, "learning_rate": 9.890955621015039e-06, "loss": 0.5876, "step": 9990 }, { "epoch": 10.0, "eval_loss": 0.5795682668685913, "eval_map": 0.4447, "eval_map_50": 0.6032, "eval_map_75": 0.5274, "eval_map_chicken": 0.5865, "eval_map_duck": 0.0, "eval_map_large": 0.4701, "eval_map_medium": 0.4084, "eval_map_plant": 0.7475, "eval_map_small": 0.1098, "eval_mar_1": 0.1426, "eval_mar_10": 0.5064, "eval_mar_100": 0.5112, "eval_mar_100_chicken": 0.7376, "eval_mar_100_duck": 0.0, "eval_mar_100_plant": 0.7961, "eval_mar_large": 0.5263, "eval_mar_medium": 0.4843, "eval_mar_small": 0.1833, "eval_runtime": 12.7335, "eval_samples_per_second": 15.707, "eval_steps_per_second": 1.963, "step": 10000 }, { "epoch": 10.02, "grad_norm": 59.877235412597656, "learning_rate": 9.89030212539124e-06, "loss": 0.6838, "step": 10020 }, { "epoch": 10.05, "grad_norm": 66.97042846679688, "learning_rate": 9.88964669915361e-06, "loss": 0.6275, "step": 10050 }, { "epoch": 10.08, "grad_norm": 0.12585313618183136, "learning_rate": 9.8889893425609e-06, "loss": 0.656, "step": 10080 }, { "epoch": 10.11, "grad_norm": 161.67050170898438, "learning_rate": 9.888330055872623e-06, "loss": 0.5953, "step": 10110 }, { "epoch": 10.14, "grad_norm": 110.2093734741211, "learning_rate": 9.887668839349057e-06, "loss": 0.6314, "step": 10140 }, { "epoch": 10.17, "grad_norm": 43.92682647705078, "learning_rate": 9.88700569325124e-06, "loss": 0.6476, "step": 10170 }, { "epoch": 10.2, "grad_norm": 70.8399429321289, "learning_rate": 9.886340617840968e-06, "loss": 0.6184, "step": 10200 }, { "epoch": 10.23, "grad_norm": 28.79657745361328, "learning_rate": 9.885673613380806e-06, "loss": 0.7678, "step": 10230 }, { "epoch": 10.26, "grad_norm": 81.58057403564453, "learning_rate": 9.885004680134075e-06, "loss": 0.6515, "step": 10260 }, { "epoch": 10.29, "grad_norm": 45.42242431640625, "learning_rate": 9.884333818364861e-06, "loss": 0.7245, "step": 10290 }, { "epoch": 10.32, "grad_norm": 58.49011993408203, "learning_rate": 9.883661028338009e-06, "loss": 0.5625, "step": 10320 }, { "epoch": 10.35, "grad_norm": 572.979736328125, "learning_rate": 9.882986310319124e-06, "loss": 0.625, "step": 10350 }, { "epoch": 10.38, "grad_norm": 76.84716796875, "learning_rate": 9.882309664574576e-06, "loss": 0.5692, "step": 10380 }, { "epoch": 10.41, "grad_norm": 28.76850700378418, "learning_rate": 9.881631091371492e-06, "loss": 0.7298, "step": 10410 }, { "epoch": 10.44, "grad_norm": 47.94281768798828, "learning_rate": 9.880950590977764e-06, "loss": 0.626, "step": 10440 }, { "epoch": 10.47, "grad_norm": 50.27431869506836, "learning_rate": 9.880268163662043e-06, "loss": 0.61, "step": 10470 }, { "epoch": 10.5, "grad_norm": 390.40252685546875, "learning_rate": 9.879583809693737e-06, "loss": 0.5357, "step": 10500 }, { "epoch": 10.53, "grad_norm": 79.29381561279297, "learning_rate": 9.878897529343023e-06, "loss": 0.6091, "step": 10530 }, { "epoch": 10.56, "grad_norm": 92.53865051269531, "learning_rate": 9.87820932288083e-06, "loss": 0.6803, "step": 10560 }, { "epoch": 10.59, "grad_norm": 231.27809143066406, "learning_rate": 9.877519190578852e-06, "loss": 0.6286, "step": 10590 }, { "epoch": 10.62, "grad_norm": 70.90177917480469, "learning_rate": 9.876827132709545e-06, "loss": 0.5554, "step": 10620 }, { "epoch": 10.65, "grad_norm": 217.32301330566406, "learning_rate": 9.876133149546117e-06, "loss": 0.5283, "step": 10650 }, { "epoch": 10.68, "grad_norm": 92.49636840820312, "learning_rate": 9.875437241362546e-06, "loss": 0.7217, "step": 10680 }, { "epoch": 10.71, "grad_norm": 67.02864074707031, "learning_rate": 9.874739408433565e-06, "loss": 0.8025, "step": 10710 }, { "epoch": 10.74, "grad_norm": 41.669464111328125, "learning_rate": 9.874039651034665e-06, "loss": 0.6051, "step": 10740 }, { "epoch": 10.77, "grad_norm": 25.317764282226562, "learning_rate": 9.873337969442102e-06, "loss": 0.5883, "step": 10770 }, { "epoch": 10.8, "grad_norm": 0.0009413706720806658, "learning_rate": 9.872634363932887e-06, "loss": 0.6054, "step": 10800 }, { "epoch": 10.83, "grad_norm": 36.305171966552734, "learning_rate": 9.871928834784793e-06, "loss": 0.5818, "step": 10830 }, { "epoch": 10.86, "grad_norm": 101.89096069335938, "learning_rate": 9.87122138227635e-06, "loss": 0.7958, "step": 10860 }, { "epoch": 10.89, "grad_norm": 32.61777114868164, "learning_rate": 9.870512006686852e-06, "loss": 0.665, "step": 10890 }, { "epoch": 10.92, "grad_norm": 0.0003835726820398122, "learning_rate": 9.869800708296347e-06, "loss": 0.5564, "step": 10920 }, { "epoch": 10.95, "grad_norm": 80.3359146118164, "learning_rate": 9.869087487385644e-06, "loss": 0.7522, "step": 10950 }, { "epoch": 10.98, "grad_norm": 52.274864196777344, "learning_rate": 9.868372344236314e-06, "loss": 0.4736, "step": 10980 }, { "epoch": 11.0, "eval_loss": 0.5644980669021606, "eval_map": 0.4457, "eval_map_50": 0.6013, "eval_map_75": 0.5331, "eval_map_chicken": 0.5734, "eval_map_duck": 0.0, "eval_map_large": 0.4769, "eval_map_medium": 0.4073, "eval_map_plant": 0.7638, "eval_map_small": 0.1092, "eval_mar_1": 0.1397, "eval_mar_10": 0.5073, "eval_mar_100": 0.5115, "eval_mar_100_chicken": 0.7258, "eval_mar_100_duck": 0.0, "eval_mar_100_plant": 0.8087, "eval_mar_large": 0.5278, "eval_mar_medium": 0.4811, "eval_mar_small": 0.2014, "eval_runtime": 13.2656, "eval_samples_per_second": 15.077, "eval_steps_per_second": 1.885, "step": 11000 }, { "epoch": 11.01, "grad_norm": 83.56844329833984, "learning_rate": 9.867655279130684e-06, "loss": 0.5334, "step": 11010 }, { "epoch": 11.04, "grad_norm": 73.52943420410156, "learning_rate": 9.866936292351837e-06, "loss": 0.6666, "step": 11040 }, { "epoch": 11.07, "grad_norm": 78.49867248535156, "learning_rate": 9.86621538418362e-06, "loss": 0.6505, "step": 11070 }, { "epoch": 11.1, "grad_norm": 30.247465133666992, "learning_rate": 9.865492554910634e-06, "loss": 0.6362, "step": 11100 }, { "epoch": 11.13, "grad_norm": 107.76499938964844, "learning_rate": 9.864767804818242e-06, "loss": 0.5868, "step": 11130 }, { "epoch": 11.16, "grad_norm": 45.03202438354492, "learning_rate": 9.864041134192563e-06, "loss": 0.6364, "step": 11160 }, { "epoch": 11.19, "grad_norm": 235.00357055664062, "learning_rate": 9.863312543320479e-06, "loss": 0.7198, "step": 11190 }, { "epoch": 11.22, "grad_norm": 37.47989273071289, "learning_rate": 9.862582032489621e-06, "loss": 1.2176, "step": 11220 }, { "epoch": 11.25, "grad_norm": 45.66279220581055, "learning_rate": 9.861849601988384e-06, "loss": 0.6242, "step": 11250 }, { "epoch": 11.28, "grad_norm": 47.580230712890625, "learning_rate": 9.861115252105922e-06, "loss": 0.646, "step": 11280 }, { "epoch": 11.31, "grad_norm": 24.262310028076172, "learning_rate": 9.860378983132144e-06, "loss": 0.6862, "step": 11310 }, { "epoch": 11.34, "grad_norm": 55.1291389465332, "learning_rate": 9.859640795357716e-06, "loss": 0.5503, "step": 11340 }, { "epoch": 11.37, "grad_norm": 40.002655029296875, "learning_rate": 9.858900689074065e-06, "loss": 0.684, "step": 11370 }, { "epoch": 11.4, "grad_norm": 74.77377319335938, "learning_rate": 9.85815866457337e-06, "loss": 0.5737, "step": 11400 }, { "epoch": 11.43, "grad_norm": 77.88835144042969, "learning_rate": 9.857414722148574e-06, "loss": 0.7299, "step": 11430 }, { "epoch": 11.46, "grad_norm": 73.92407989501953, "learning_rate": 9.856668862093372e-06, "loss": 0.5717, "step": 11460 }, { "epoch": 11.49, "grad_norm": 31.301790237426758, "learning_rate": 9.85592108470222e-06, "loss": 0.6349, "step": 11490 }, { "epoch": 11.52, "grad_norm": 35.23759841918945, "learning_rate": 9.855171390270325e-06, "loss": 0.6165, "step": 11520 }, { "epoch": 11.55, "grad_norm": 33.284358978271484, "learning_rate": 9.854419779093656e-06, "loss": 0.7139, "step": 11550 }, { "epoch": 11.58, "grad_norm": 93.20217895507812, "learning_rate": 9.853666251468938e-06, "loss": 0.5642, "step": 11580 }, { "epoch": 11.61, "grad_norm": 69.21809387207031, "learning_rate": 9.85291080769365e-06, "loss": 0.5721, "step": 11610 }, { "epoch": 11.64, "grad_norm": 49.745338439941406, "learning_rate": 9.852153448066031e-06, "loss": 0.6338, "step": 11640 }, { "epoch": 11.67, "grad_norm": 111.23006439208984, "learning_rate": 9.851394172885075e-06, "loss": 0.5749, "step": 11670 }, { "epoch": 11.7, "grad_norm": 51.23693084716797, "learning_rate": 9.85063298245053e-06, "loss": 0.5806, "step": 11700 }, { "epoch": 11.73, "grad_norm": 56.1314697265625, "learning_rate": 9.849869877062903e-06, "loss": 0.6313, "step": 11730 }, { "epoch": 11.76, "grad_norm": 155.66836547851562, "learning_rate": 9.849104857023455e-06, "loss": 0.5683, "step": 11760 }, { "epoch": 11.79, "grad_norm": 88.0706787109375, "learning_rate": 9.848337922634205e-06, "loss": 0.5988, "step": 11790 }, { "epoch": 11.82, "grad_norm": 74.23865509033203, "learning_rate": 9.847569074197927e-06, "loss": 0.4886, "step": 11820 }, { "epoch": 11.85, "grad_norm": 47.310272216796875, "learning_rate": 9.846798312018147e-06, "loss": 0.6158, "step": 11850 }, { "epoch": 11.88, "grad_norm": 42.320274353027344, "learning_rate": 9.846025636399152e-06, "loss": 0.5583, "step": 11880 }, { "epoch": 11.91, "grad_norm": 48.930152893066406, "learning_rate": 9.845251047645984e-06, "loss": 0.5315, "step": 11910 }, { "epoch": 11.94, "grad_norm": 73.43531799316406, "learning_rate": 9.844474546064436e-06, "loss": 0.7535, "step": 11940 }, { "epoch": 11.97, "grad_norm": 76.36287689208984, "learning_rate": 9.843696131961058e-06, "loss": 0.554, "step": 11970 }, { "epoch": 12.0, "grad_norm": 68.55497741699219, "learning_rate": 9.842915805643156e-06, "loss": 0.6139, "step": 12000 }, { "epoch": 12.0, "eval_loss": 0.5576649308204651, "eval_map": 0.4428, "eval_map_50": 0.5984, "eval_map_75": 0.5244, "eval_map_chicken": 0.5648, "eval_map_duck": 0.0, "eval_map_large": 0.4745, "eval_map_medium": 0.4043, "eval_map_plant": 0.7636, "eval_map_small": 0.0509, "eval_mar_1": 0.1392, "eval_mar_10": 0.5128, "eval_mar_100": 0.5177, "eval_mar_100_chicken": 0.7408, "eval_mar_100_duck": 0.0, "eval_mar_100_plant": 0.8124, "eval_mar_large": 0.5341, "eval_mar_medium": 0.4876, "eval_mar_small": 0.1709, "eval_runtime": 12.5775, "eval_samples_per_second": 15.901, "eval_steps_per_second": 1.988, "step": 12000 }, { "epoch": 12.03, "grad_norm": 48.19761657714844, "learning_rate": 9.842133567418793e-06, "loss": 0.6227, "step": 12030 }, { "epoch": 12.06, "grad_norm": 39.90418243408203, "learning_rate": 9.84134941759678e-06, "loss": 0.6214, "step": 12060 }, { "epoch": 12.09, "grad_norm": 72.14737701416016, "learning_rate": 9.84056335648669e-06, "loss": 0.8288, "step": 12090 }, { "epoch": 12.12, "grad_norm": 75.5544204711914, "learning_rate": 9.839775384398846e-06, "loss": 0.5997, "step": 12120 }, { "epoch": 12.15, "grad_norm": 61.25050735473633, "learning_rate": 9.838985501644329e-06, "loss": 0.6265, "step": 12150 }, { "epoch": 12.18, "grad_norm": 0.0004860181943513453, "learning_rate": 9.838193708534969e-06, "loss": 0.61, "step": 12180 }, { "epoch": 12.21, "grad_norm": 50.9096794128418, "learning_rate": 9.837400005383355e-06, "loss": 0.6748, "step": 12210 }, { "epoch": 12.24, "grad_norm": 76.47835540771484, "learning_rate": 9.836604392502829e-06, "loss": 0.6796, "step": 12240 }, { "epoch": 12.27, "grad_norm": 97.92862701416016, "learning_rate": 9.835806870207487e-06, "loss": 0.6005, "step": 12270 }, { "epoch": 12.3, "grad_norm": 55.349552154541016, "learning_rate": 9.835007438812177e-06, "loss": 0.5917, "step": 12300 }, { "epoch": 12.33, "grad_norm": 82.7293701171875, "learning_rate": 9.834206098632499e-06, "loss": 0.6243, "step": 12330 }, { "epoch": 12.36, "grad_norm": 57.17816162109375, "learning_rate": 9.833402849984815e-06, "loss": 0.5736, "step": 12360 }, { "epoch": 12.39, "grad_norm": 44.93244934082031, "learning_rate": 9.832597693186233e-06, "loss": 0.6185, "step": 12390 }, { "epoch": 12.42, "grad_norm": 42.12499237060547, "learning_rate": 9.831790628554613e-06, "loss": 0.6321, "step": 12420 }, { "epoch": 12.45, "grad_norm": 44.744537353515625, "learning_rate": 9.830981656408575e-06, "loss": 0.6367, "step": 12450 }, { "epoch": 12.48, "grad_norm": 136.20523071289062, "learning_rate": 9.830170777067486e-06, "loss": 0.6312, "step": 12480 }, { "epoch": 12.51, "grad_norm": 59.62192153930664, "learning_rate": 9.82935799085147e-06, "loss": 0.5583, "step": 12510 }, { "epoch": 12.54, "grad_norm": 61.16498947143555, "learning_rate": 9.828543298081401e-06, "loss": 0.5336, "step": 12540 }, { "epoch": 12.57, "grad_norm": 0.001350203063338995, "learning_rate": 9.827726699078907e-06, "loss": 0.6137, "step": 12570 }, { "epoch": 12.6, "grad_norm": 36.39663314819336, "learning_rate": 9.82690819416637e-06, "loss": 0.5071, "step": 12600 }, { "epoch": 12.63, "grad_norm": 31.63799285888672, "learning_rate": 9.82608778366692e-06, "loss": 0.5705, "step": 12630 }, { "epoch": 12.66, "grad_norm": 36.55143737792969, "learning_rate": 9.825265467904446e-06, "loss": 0.5428, "step": 12660 }, { "epoch": 12.69, "grad_norm": 139.78842163085938, "learning_rate": 9.82444124720358e-06, "loss": 0.5458, "step": 12690 }, { "epoch": 12.72, "grad_norm": 73.02035522460938, "learning_rate": 9.823615121889716e-06, "loss": 0.5915, "step": 12720 }, { "epoch": 12.75, "grad_norm": 32.08891296386719, "learning_rate": 9.822787092288991e-06, "loss": 0.5189, "step": 12750 }, { "epoch": 12.78, "grad_norm": 49.90654754638672, "learning_rate": 9.821957158728302e-06, "loss": 0.602, "step": 12780 }, { "epoch": 12.81, "grad_norm": 23.52266502380371, "learning_rate": 9.82112532153529e-06, "loss": 0.6163, "step": 12810 }, { "epoch": 12.84, "grad_norm": 41.693634033203125, "learning_rate": 9.820291581038354e-06, "loss": 0.5427, "step": 12840 }, { "epoch": 12.87, "grad_norm": 159.3751220703125, "learning_rate": 9.819455937566642e-06, "loss": 0.6797, "step": 12870 }, { "epoch": 12.9, "grad_norm": 43.2912712097168, "learning_rate": 9.81861839145005e-06, "loss": 0.7264, "step": 12900 }, { "epoch": 12.93, "grad_norm": 86.80297088623047, "learning_rate": 9.817778943019228e-06, "loss": 0.7134, "step": 12930 }, { "epoch": 12.96, "grad_norm": 23.1685791015625, "learning_rate": 9.81693759260558e-06, "loss": 0.6061, "step": 12960 }, { "epoch": 12.99, "grad_norm": 38.79433059692383, "learning_rate": 9.816094340541256e-06, "loss": 0.7356, "step": 12990 }, { "epoch": 13.0, "eval_loss": 0.5288656949996948, "eval_map": 0.4651, "eval_map_50": 0.6128, "eval_map_75": 0.5378, "eval_map_chicken": 0.6242, "eval_map_duck": 0.0, "eval_map_large": 0.489, "eval_map_medium": 0.4393, "eval_map_plant": 0.7711, "eval_map_small": 0.048, "eval_mar_1": 0.146, "eval_mar_10": 0.5238, "eval_mar_100": 0.5272, "eval_mar_100_chicken": 0.762, "eval_mar_100_duck": 0.0, "eval_mar_100_plant": 0.8197, "eval_mar_large": 0.5375, "eval_mar_medium": 0.5075, "eval_mar_small": 0.1102, "eval_runtime": 12.6426, "eval_samples_per_second": 15.82, "eval_steps_per_second": 1.977, "step": 13000 }, { "epoch": 13.02, "grad_norm": 102.14237976074219, "learning_rate": 9.815249187159158e-06, "loss": 0.8396, "step": 13020 }, { "epoch": 13.05, "grad_norm": 56.234466552734375, "learning_rate": 9.814402132792939e-06, "loss": 0.5583, "step": 13050 }, { "epoch": 13.08, "grad_norm": 55.06719207763672, "learning_rate": 9.813553177777005e-06, "loss": 0.5653, "step": 13080 }, { "epoch": 13.11, "grad_norm": 66.1046371459961, "learning_rate": 9.812702322446506e-06, "loss": 0.5604, "step": 13110 }, { "epoch": 13.14, "grad_norm": 54.93820571899414, "learning_rate": 9.811849567137351e-06, "loss": 0.5529, "step": 13140 }, { "epoch": 13.17, "grad_norm": 42.904747009277344, "learning_rate": 9.81099491218619e-06, "loss": 0.6219, "step": 13170 }, { "epoch": 13.2, "grad_norm": 36.545841217041016, "learning_rate": 9.81013835793043e-06, "loss": 0.5358, "step": 13200 }, { "epoch": 13.23, "grad_norm": 41.60871124267578, "learning_rate": 9.809279904708224e-06, "loss": 0.6728, "step": 13230 }, { "epoch": 13.26, "grad_norm": 172.75672912597656, "learning_rate": 9.808419552858477e-06, "loss": 0.6414, "step": 13260 }, { "epoch": 13.29, "grad_norm": 54.869163513183594, "learning_rate": 9.80755730272084e-06, "loss": 0.6593, "step": 13290 }, { "epoch": 13.32, "grad_norm": 36.91462707519531, "learning_rate": 9.806693154635719e-06, "loss": 0.6588, "step": 13320 }, { "epoch": 13.35, "grad_norm": 73.8152084350586, "learning_rate": 9.80582710894426e-06, "loss": 0.7069, "step": 13350 }, { "epoch": 13.38, "grad_norm": 41.862327575683594, "learning_rate": 9.80495916598837e-06, "loss": 0.5875, "step": 13380 }, { "epoch": 13.41, "grad_norm": 50.769920349121094, "learning_rate": 9.804089326110697e-06, "loss": 0.5792, "step": 13410 }, { "epoch": 13.44, "grad_norm": 28.787174224853516, "learning_rate": 9.803217589654642e-06, "loss": 0.5875, "step": 13440 }, { "epoch": 13.47, "grad_norm": 131.36802673339844, "learning_rate": 9.802343956964348e-06, "loss": 0.6202, "step": 13470 }, { "epoch": 13.5, "grad_norm": 403.9072570800781, "learning_rate": 9.801468428384716e-06, "loss": 0.5827, "step": 13500 }, { "epoch": 13.53, "grad_norm": 39.42959213256836, "learning_rate": 9.800591004261388e-06, "loss": 0.6477, "step": 13530 }, { "epoch": 13.56, "grad_norm": 31.025798797607422, "learning_rate": 9.79971168494076e-06, "loss": 0.5893, "step": 13560 }, { "epoch": 13.59, "grad_norm": 59.21128463745117, "learning_rate": 9.79883047076997e-06, "loss": 0.5496, "step": 13590 }, { "epoch": 13.62, "grad_norm": 37.99839401245117, "learning_rate": 9.797947362096909e-06, "loss": 0.5657, "step": 13620 }, { "epoch": 13.65, "grad_norm": 38.36595916748047, "learning_rate": 9.797062359270215e-06, "loss": 0.6566, "step": 13650 }, { "epoch": 13.68, "grad_norm": 64.33842468261719, "learning_rate": 9.796175462639273e-06, "loss": 0.6589, "step": 13680 }, { "epoch": 13.71, "grad_norm": 29.045379638671875, "learning_rate": 9.795286672554214e-06, "loss": 0.6338, "step": 13710 }, { "epoch": 13.74, "grad_norm": 61.64324951171875, "learning_rate": 9.79439598936592e-06, "loss": 0.6565, "step": 13740 }, { "epoch": 13.77, "grad_norm": 24.872180938720703, "learning_rate": 9.793503413426016e-06, "loss": 0.6189, "step": 13770 }, { "epoch": 13.8, "grad_norm": 60.10523986816406, "learning_rate": 9.79260894508688e-06, "loss": 0.5277, "step": 13800 }, { "epoch": 13.83, "grad_norm": 147.0440216064453, "learning_rate": 9.791712584701634e-06, "loss": 0.5012, "step": 13830 }, { "epoch": 13.86, "grad_norm": 54.5830078125, "learning_rate": 9.790814332624144e-06, "loss": 0.6118, "step": 13860 }, { "epoch": 13.89, "grad_norm": 41.7625732421875, "learning_rate": 9.789914189209028e-06, "loss": 0.7177, "step": 13890 }, { "epoch": 13.92, "grad_norm": 111.00346374511719, "learning_rate": 9.789012154811648e-06, "loss": 0.5315, "step": 13920 }, { "epoch": 13.95, "grad_norm": 50.253334045410156, "learning_rate": 9.788108229788111e-06, "loss": 0.568, "step": 13950 }, { "epoch": 13.98, "grad_norm": 50.6452751159668, "learning_rate": 9.787202414495275e-06, "loss": 0.5112, "step": 13980 }, { "epoch": 14.0, "eval_loss": 0.5339823961257935, "eval_map": 0.4658, "eval_map_50": 0.6239, "eval_map_75": 0.5556, "eval_map_chicken": 0.6346, "eval_map_duck": 0.0, "eval_map_large": 0.4849, "eval_map_medium": 0.4309, "eval_map_plant": 0.7627, "eval_map_small": 0.0889, "eval_mar_1": 0.1444, "eval_mar_10": 0.5135, "eval_mar_100": 0.5164, "eval_mar_100_chicken": 0.7404, "eval_mar_100_duck": 0.0, "eval_mar_100_plant": 0.8087, "eval_mar_large": 0.5316, "eval_mar_medium": 0.4892, "eval_mar_small": 0.1865, "eval_runtime": 12.9887, "eval_samples_per_second": 15.398, "eval_steps_per_second": 1.925, "step": 14000 }, { "epoch": 14.01, "grad_norm": 40.188045501708984, "learning_rate": 9.786294709290741e-06, "loss": 0.5569, "step": 14010 }, { "epoch": 14.04, "grad_norm": 65.31824493408203, "learning_rate": 9.785385114532858e-06, "loss": 0.5591, "step": 14040 }, { "epoch": 14.07, "grad_norm": 58.77595901489258, "learning_rate": 9.784473630580713e-06, "loss": 0.5326, "step": 14070 }, { "epoch": 14.1, "grad_norm": 132.1082000732422, "learning_rate": 9.783560257794153e-06, "loss": 0.5881, "step": 14100 }, { "epoch": 14.13, "grad_norm": 58.373966217041016, "learning_rate": 9.78264499653376e-06, "loss": 0.6041, "step": 14130 }, { "epoch": 14.16, "grad_norm": 32.050018310546875, "learning_rate": 9.781727847160865e-06, "loss": 0.6352, "step": 14160 }, { "epoch": 14.19, "grad_norm": 36.78704071044922, "learning_rate": 9.780808810037543e-06, "loss": 0.5708, "step": 14190 }, { "epoch": 14.22, "grad_norm": 129.79832458496094, "learning_rate": 9.779887885526616e-06, "loss": 0.6765, "step": 14220 }, { "epoch": 14.25, "grad_norm": 27.62218475341797, "learning_rate": 9.778965073991652e-06, "loss": 0.5053, "step": 14250 }, { "epoch": 14.28, "grad_norm": 48.13038635253906, "learning_rate": 9.77804037579696e-06, "loss": 0.5117, "step": 14280 }, { "epoch": 14.31, "grad_norm": 57.49237060546875, "learning_rate": 9.777113791307597e-06, "loss": 0.5111, "step": 14310 }, { "epoch": 14.34, "grad_norm": 80.8060531616211, "learning_rate": 9.776185320889364e-06, "loss": 0.5437, "step": 14340 }, { "epoch": 14.37, "grad_norm": 34.34523010253906, "learning_rate": 9.775254964908807e-06, "loss": 0.5568, "step": 14370 }, { "epoch": 14.4, "grad_norm": 60.69856643676758, "learning_rate": 9.774322723733216e-06, "loss": 0.6386, "step": 14400 }, { "epoch": 14.43, "grad_norm": 123.79764556884766, "learning_rate": 9.773388597730623e-06, "loss": 0.5815, "step": 14430 }, { "epoch": 14.46, "grad_norm": 68.0003433227539, "learning_rate": 9.772452587269808e-06, "loss": 0.5635, "step": 14460 }, { "epoch": 14.49, "grad_norm": 51.943111419677734, "learning_rate": 9.771514692720293e-06, "loss": 0.5949, "step": 14490 }, { "epoch": 14.52, "grad_norm": 106.07717895507812, "learning_rate": 9.770574914452343e-06, "loss": 0.6606, "step": 14520 }, { "epoch": 14.55, "grad_norm": 114.58092498779297, "learning_rate": 9.769633252836969e-06, "loss": 0.5675, "step": 14550 }, { "epoch": 14.58, "grad_norm": 34.003578186035156, "learning_rate": 9.768689708245921e-06, "loss": 0.4975, "step": 14580 }, { "epoch": 14.61, "grad_norm": 41.251251220703125, "learning_rate": 9.767744281051702e-06, "loss": 0.5784, "step": 14610 }, { "epoch": 14.64, "grad_norm": 52.497955322265625, "learning_rate": 9.766796971627543e-06, "loss": 0.6444, "step": 14640 }, { "epoch": 14.67, "grad_norm": 38.4123420715332, "learning_rate": 9.765847780347433e-06, "loss": 0.6797, "step": 14670 }, { "epoch": 14.7, "grad_norm": 53.31391143798828, "learning_rate": 9.764896707586095e-06, "loss": 0.5731, "step": 14700 }, { "epoch": 14.73, "grad_norm": 100.5085678100586, "learning_rate": 9.763943753719e-06, "loss": 0.5691, "step": 14730 }, { "epoch": 14.76, "grad_norm": 0.003614983754232526, "learning_rate": 9.762988919122354e-06, "loss": 0.5472, "step": 14760 }, { "epoch": 14.79, "grad_norm": 55.724876403808594, "learning_rate": 9.762032204173116e-06, "loss": 0.6518, "step": 14790 }, { "epoch": 14.82, "grad_norm": 98.21144104003906, "learning_rate": 9.761073609248981e-06, "loss": 0.5593, "step": 14820 }, { "epoch": 14.85, "grad_norm": 32.798728942871094, "learning_rate": 9.760113134728383e-06, "loss": 0.6774, "step": 14850 }, { "epoch": 14.88, "grad_norm": 70.28900909423828, "learning_rate": 9.759150780990508e-06, "loss": 0.6491, "step": 14880 }, { "epoch": 14.91, "grad_norm": 29.958711624145508, "learning_rate": 9.758186548415274e-06, "loss": 0.4866, "step": 14910 }, { "epoch": 14.94, "grad_norm": 35.03977584838867, "learning_rate": 9.757220437383345e-06, "loss": 0.5871, "step": 14940 }, { "epoch": 14.97, "grad_norm": 44.48859405517578, "learning_rate": 9.756252448276128e-06, "loss": 0.5787, "step": 14970 }, { "epoch": 15.0, "grad_norm": 79.79589080810547, "learning_rate": 9.755282581475769e-06, "loss": 0.5302, "step": 15000 }, { "epoch": 15.0, "eval_loss": 0.5103353261947632, "eval_map": 0.478, "eval_map_50": 0.6328, "eval_map_75": 0.5738, "eval_map_chicken": 0.6635, "eval_map_duck": 0.0, "eval_map_large": 0.4967, "eval_map_medium": 0.4478, "eval_map_plant": 0.7705, "eval_map_small": 0.0727, "eval_mar_1": 0.1478, "eval_mar_10": 0.5216, "eval_mar_100": 0.5244, "eval_mar_100_chicken": 0.7571, "eval_mar_100_duck": 0.0, "eval_mar_100_plant": 0.816, "eval_mar_large": 0.54, "eval_mar_medium": 0.498, "eval_mar_small": 0.2081, "eval_runtime": 12.5941, "eval_samples_per_second": 15.88, "eval_steps_per_second": 1.985, "step": 15000 }, { "epoch": 15.03, "grad_norm": 38.29458236694336, "learning_rate": 9.754310837365155e-06, "loss": 0.5312, "step": 15030 }, { "epoch": 15.06, "grad_norm": 33.89739227294922, "learning_rate": 9.753337216327917e-06, "loss": 0.6456, "step": 15060 }, { "epoch": 15.09, "grad_norm": 179.07749938964844, "learning_rate": 9.752361718748425e-06, "loss": 0.6335, "step": 15090 }, { "epoch": 15.12, "grad_norm": 57.291316986083984, "learning_rate": 9.751384345011787e-06, "loss": 0.7183, "step": 15120 }, { "epoch": 15.15, "grad_norm": 63.326148986816406, "learning_rate": 9.750405095503859e-06, "loss": 0.5498, "step": 15150 }, { "epoch": 15.18, "grad_norm": 0.0006223198142834008, "learning_rate": 9.749423970611232e-06, "loss": 0.6992, "step": 15180 }, { "epoch": 15.21, "grad_norm": 57.37306594848633, "learning_rate": 9.748440970721236e-06, "loss": 0.525, "step": 15210 }, { "epoch": 15.24, "grad_norm": 34.62443161010742, "learning_rate": 9.747456096221946e-06, "loss": 0.4952, "step": 15240 }, { "epoch": 15.27, "grad_norm": 36.6563720703125, "learning_rate": 9.746469347502174e-06, "loss": 0.6904, "step": 15270 }, { "epoch": 15.3, "grad_norm": 76.6176986694336, "learning_rate": 9.745480724951473e-06, "loss": 0.5112, "step": 15300 }, { "epoch": 15.33, "grad_norm": 98.65264129638672, "learning_rate": 9.744490228960137e-06, "loss": 0.573, "step": 15330 }, { "epoch": 15.36, "grad_norm": 45.8120231628418, "learning_rate": 9.743497859919196e-06, "loss": 0.5831, "step": 15360 }, { "epoch": 15.39, "grad_norm": 32.223304748535156, "learning_rate": 9.742503618220422e-06, "loss": 0.6909, "step": 15390 }, { "epoch": 15.42, "grad_norm": 65.93331909179688, "learning_rate": 9.741507504256327e-06, "loss": 0.5817, "step": 15420 }, { "epoch": 15.45, "grad_norm": 216.44024658203125, "learning_rate": 9.74050951842016e-06, "loss": 0.5545, "step": 15450 }, { "epoch": 15.48, "grad_norm": 69.15869140625, "learning_rate": 9.739509661105912e-06, "loss": 0.5848, "step": 15480 }, { "epoch": 15.51, "grad_norm": 72.28836822509766, "learning_rate": 9.738507932708308e-06, "loss": 0.4893, "step": 15510 }, { "epoch": 15.54, "grad_norm": 40.42341995239258, "learning_rate": 9.737504333622814e-06, "loss": 0.5974, "step": 15540 }, { "epoch": 15.57, "grad_norm": 0.00021373255003709346, "learning_rate": 9.736498864245638e-06, "loss": 0.5161, "step": 15570 }, { "epoch": 15.6, "grad_norm": 0.00014961558918002993, "learning_rate": 9.735491524973723e-06, "loss": 0.4971, "step": 15600 }, { "epoch": 15.63, "grad_norm": 24.08332633972168, "learning_rate": 9.734482316204747e-06, "loss": 0.5554, "step": 15630 }, { "epoch": 15.66, "grad_norm": 66.76925659179688, "learning_rate": 9.733471238337136e-06, "loss": 0.5903, "step": 15660 }, { "epoch": 15.69, "grad_norm": 0.00042738320189528167, "learning_rate": 9.73245829177004e-06, "loss": 0.745, "step": 15690 }, { "epoch": 15.72, "grad_norm": 44.16368865966797, "learning_rate": 9.73144347690336e-06, "loss": 0.5715, "step": 15720 }, { "epoch": 15.75, "grad_norm": 46.65747833251953, "learning_rate": 9.730426794137727e-06, "loss": 0.6211, "step": 15750 }, { "epoch": 15.78, "grad_norm": 32.402347564697266, "learning_rate": 9.729408243874511e-06, "loss": 0.5683, "step": 15780 }, { "epoch": 15.81, "grad_norm": 735.0786743164062, "learning_rate": 9.72838782651582e-06, "loss": 0.5091, "step": 15810 }, { "epoch": 15.84, "grad_norm": 32.880088806152344, "learning_rate": 9.727365542464498e-06, "loss": 0.5963, "step": 15840 }, { "epoch": 15.87, "grad_norm": 55.05681228637695, "learning_rate": 9.726341392124127e-06, "loss": 0.6181, "step": 15870 }, { "epoch": 15.9, "grad_norm": 71.13121795654297, "learning_rate": 9.725315375899025e-06, "loss": 0.684, "step": 15900 }, { "epoch": 15.93, "grad_norm": 53.01565933227539, "learning_rate": 9.724287494194247e-06, "loss": 0.5957, "step": 15930 }, { "epoch": 15.96, "grad_norm": 80.8268051147461, "learning_rate": 9.723257747415584e-06, "loss": 0.4883, "step": 15960 }, { "epoch": 15.99, "grad_norm": 90.13495635986328, "learning_rate": 9.722226135969565e-06, "loss": 0.4855, "step": 15990 }, { "epoch": 16.0, "eval_loss": 0.5183222889900208, "eval_map": 0.5038, "eval_map_50": 0.6727, "eval_map_75": 0.6016, "eval_map_chicken": 0.6608, "eval_map_duck": 0.0921, "eval_map_large": 0.5257, "eval_map_medium": 0.4726, "eval_map_plant": 0.7587, "eval_map_small": 0.1285, "eval_mar_1": 0.1738, "eval_mar_10": 0.5454, "eval_mar_100": 0.5486, "eval_mar_100_chicken": 0.7408, "eval_mar_100_duck": 0.0928, "eval_mar_100_plant": 0.8121, "eval_mar_large": 0.5649, "eval_mar_medium": 0.5213, "eval_mar_small": 0.2061, "eval_runtime": 12.504, "eval_samples_per_second": 15.995, "eval_steps_per_second": 1.999, "step": 16000 }, { "epoch": 16.02, "grad_norm": 0.0009297202923335135, "learning_rate": 9.721192660263454e-06, "loss": 0.5069, "step": 16020 }, { "epoch": 16.05, "grad_norm": 63.773170471191406, "learning_rate": 9.72015732070525e-06, "loss": 0.5985, "step": 16050 }, { "epoch": 16.08, "grad_norm": 53.72353744506836, "learning_rate": 9.719120117703688e-06, "loss": 0.6894, "step": 16080 }, { "epoch": 16.11, "grad_norm": 54.251434326171875, "learning_rate": 9.71808105166824e-06, "loss": 0.5258, "step": 16110 }, { "epoch": 16.14, "grad_norm": 30.138662338256836, "learning_rate": 9.717040123009111e-06, "loss": 0.5151, "step": 16140 }, { "epoch": 16.17, "grad_norm": 70.02063751220703, "learning_rate": 9.715997332137248e-06, "loss": 0.5644, "step": 16170 }, { "epoch": 16.2, "grad_norm": 71.34861755371094, "learning_rate": 9.714952679464324e-06, "loss": 0.5944, "step": 16200 }, { "epoch": 16.23, "grad_norm": 51.521331787109375, "learning_rate": 9.713906165402751e-06, "loss": 0.5509, "step": 16230 }, { "epoch": 16.26, "grad_norm": 39.90167999267578, "learning_rate": 9.71285779036568e-06, "loss": 0.5157, "step": 16260 }, { "epoch": 16.29, "grad_norm": 33.25541687011719, "learning_rate": 9.71180755476699e-06, "loss": 0.5068, "step": 16290 }, { "epoch": 16.32, "grad_norm": 71.91514587402344, "learning_rate": 9.710755459021297e-06, "loss": 0.5613, "step": 16320 }, { "epoch": 16.35, "grad_norm": 194.60205078125, "learning_rate": 9.709701503543954e-06, "loss": 0.5071, "step": 16350 }, { "epoch": 16.38, "grad_norm": 41.33855438232422, "learning_rate": 9.708645688751043e-06, "loss": 0.7308, "step": 16380 }, { "epoch": 16.41, "grad_norm": 58.11330032348633, "learning_rate": 9.707588015059387e-06, "loss": 0.6237, "step": 16410 }, { "epoch": 16.44, "grad_norm": 41.07044219970703, "learning_rate": 9.706528482886535e-06, "loss": 0.754, "step": 16440 }, { "epoch": 16.47, "grad_norm": 65.4854736328125, "learning_rate": 9.705467092650775e-06, "loss": 0.6239, "step": 16470 }, { "epoch": 16.5, "grad_norm": 79.28030395507812, "learning_rate": 9.704403844771128e-06, "loss": 0.5604, "step": 16500 }, { "epoch": 16.53, "grad_norm": 44.80597686767578, "learning_rate": 9.703338739667347e-06, "loss": 0.522, "step": 16530 }, { "epoch": 16.56, "grad_norm": 31.29863739013672, "learning_rate": 9.702271777759915e-06, "loss": 0.8049, "step": 16560 }, { "epoch": 16.59, "grad_norm": 25.976999282836914, "learning_rate": 9.701202959470057e-06, "loss": 0.4953, "step": 16590 }, { "epoch": 16.62, "grad_norm": 44.01707458496094, "learning_rate": 9.700132285219724e-06, "loss": 0.5966, "step": 16620 }, { "epoch": 16.65, "grad_norm": 37.909908294677734, "learning_rate": 9.699059755431599e-06, "loss": 0.6412, "step": 16650 }, { "epoch": 16.68, "grad_norm": 39.32249069213867, "learning_rate": 9.697985370529101e-06, "loss": 0.5036, "step": 16680 }, { "epoch": 16.71, "grad_norm": 170.40440368652344, "learning_rate": 9.696909130936382e-06, "loss": 0.5593, "step": 16710 }, { "epoch": 16.74, "grad_norm": 52.815711975097656, "learning_rate": 9.695831037078323e-06, "loss": 0.5083, "step": 16740 }, { "epoch": 16.77, "grad_norm": 18.56916618347168, "learning_rate": 9.694751089380536e-06, "loss": 0.4565, "step": 16770 }, { "epoch": 16.8, "grad_norm": 60.10877990722656, "learning_rate": 9.693669288269371e-06, "loss": 0.6328, "step": 16800 }, { "epoch": 16.83, "grad_norm": 93.7027359008789, "learning_rate": 9.692585634171906e-06, "loss": 0.6052, "step": 16830 }, { "epoch": 16.86, "grad_norm": 121.53091430664062, "learning_rate": 9.691500127515945e-06, "loss": 0.4741, "step": 16860 }, { "epoch": 16.89, "grad_norm": 43.75077438354492, "learning_rate": 9.690412768730036e-06, "loss": 0.5054, "step": 16890 }, { "epoch": 16.92, "grad_norm": 132.06875610351562, "learning_rate": 9.689323558243446e-06, "loss": 0.6834, "step": 16920 }, { "epoch": 16.95, "grad_norm": 223.55857849121094, "learning_rate": 9.688232496486179e-06, "loss": 0.6113, "step": 16950 }, { "epoch": 16.98, "grad_norm": 78.26505279541016, "learning_rate": 9.687139583888971e-06, "loss": 0.4891, "step": 16980 }, { "epoch": 17.0, "eval_loss": 0.4886009991168976, "eval_map": 0.6311, "eval_map_50": 0.8473, "eval_map_75": 0.7764, "eval_map_chicken": 0.676, "eval_map_duck": 0.4514, "eval_map_large": 0.6078, "eval_map_medium": 0.612, "eval_map_plant": 0.7659, "eval_map_small": 0.1735, "eval_mar_1": 0.2544, "eval_mar_10": 0.6734, "eval_mar_100": 0.6769, "eval_mar_100_chicken": 0.7467, "eval_mar_100_duck": 0.4758, "eval_mar_100_plant": 0.8082, "eval_mar_large": 0.6446, "eval_mar_medium": 0.6651, "eval_mar_small": 0.2801, "eval_runtime": 12.9642, "eval_samples_per_second": 15.427, "eval_steps_per_second": 1.928, "step": 17000 }, { "epoch": 17.01, "grad_norm": 51.47096633911133, "learning_rate": 9.686044820883284e-06, "loss": 0.5248, "step": 17010 }, { "epoch": 17.04, "grad_norm": 72.83051300048828, "learning_rate": 9.684948207901315e-06, "loss": 0.3986, "step": 17040 }, { "epoch": 17.07, "grad_norm": 38.38227844238281, "learning_rate": 9.683849745375991e-06, "loss": 0.5109, "step": 17070 }, { "epoch": 17.1, "grad_norm": 109.11122131347656, "learning_rate": 9.682749433740963e-06, "loss": 0.5248, "step": 17100 }, { "epoch": 17.13, "grad_norm": 48.949764251708984, "learning_rate": 9.681647273430618e-06, "loss": 0.6528, "step": 17130 }, { "epoch": 17.16, "grad_norm": 47.363861083984375, "learning_rate": 9.680543264880075e-06, "loss": 0.5199, "step": 17160 }, { "epoch": 17.19, "grad_norm": 71.96512603759766, "learning_rate": 9.679437408525175e-06, "loss": 0.679, "step": 17190 }, { "epoch": 17.22, "grad_norm": 208.92071533203125, "learning_rate": 9.678329704802495e-06, "loss": 0.6403, "step": 17220 }, { "epoch": 17.25, "grad_norm": 82.74077606201172, "learning_rate": 9.677220154149338e-06, "loss": 0.5556, "step": 17250 }, { "epoch": 17.28, "grad_norm": 88.45305633544922, "learning_rate": 9.676108757003735e-06, "loss": 0.6756, "step": 17280 }, { "epoch": 17.31, "grad_norm": 31.44303321838379, "learning_rate": 9.674995513804452e-06, "loss": 0.5609, "step": 17310 }, { "epoch": 17.34, "grad_norm": 68.57437133789062, "learning_rate": 9.673880424990978e-06, "loss": 0.5998, "step": 17340 }, { "epoch": 17.37, "grad_norm": 79.24790954589844, "learning_rate": 9.672763491003531e-06, "loss": 0.5587, "step": 17370 }, { "epoch": 17.4, "grad_norm": 39.926334381103516, "learning_rate": 9.671644712283061e-06, "loss": 0.5118, "step": 17400 }, { "epoch": 17.43, "grad_norm": 42.94596862792969, "learning_rate": 9.670524089271242e-06, "loss": 0.5593, "step": 17430 }, { "epoch": 17.46, "grad_norm": 75.36443328857422, "learning_rate": 9.669401622410481e-06, "loss": 0.6962, "step": 17460 }, { "epoch": 17.49, "grad_norm": 53.405181884765625, "learning_rate": 9.668277312143908e-06, "loss": 0.5197, "step": 17490 }, { "epoch": 17.52, "grad_norm": 35.594825744628906, "learning_rate": 9.667151158915382e-06, "loss": 0.511, "step": 17520 }, { "epoch": 17.55, "grad_norm": 37.150482177734375, "learning_rate": 9.666023163169493e-06, "loss": 0.5272, "step": 17550 }, { "epoch": 17.58, "grad_norm": 58.162933349609375, "learning_rate": 9.664893325351556e-06, "loss": 0.4985, "step": 17580 }, { "epoch": 17.61, "grad_norm": 55.11952209472656, "learning_rate": 9.66376164590761e-06, "loss": 0.4945, "step": 17610 }, { "epoch": 17.64, "grad_norm": 32.13090515136719, "learning_rate": 9.662628125284426e-06, "loss": 0.4764, "step": 17640 }, { "epoch": 17.67, "grad_norm": 79.5674057006836, "learning_rate": 9.661492763929502e-06, "loss": 0.5488, "step": 17670 }, { "epoch": 17.7, "grad_norm": 75.03531646728516, "learning_rate": 9.660355562291055e-06, "loss": 0.4872, "step": 17700 }, { "epoch": 17.73, "grad_norm": 52.89229202270508, "learning_rate": 9.65921652081804e-06, "loss": 0.5207, "step": 17730 }, { "epoch": 17.76, "grad_norm": 39.937767028808594, "learning_rate": 9.65807563996013e-06, "loss": 0.4775, "step": 17760 }, { "epoch": 17.79, "grad_norm": 28.996788024902344, "learning_rate": 9.656932920167727e-06, "loss": 0.5071, "step": 17790 }, { "epoch": 17.82, "grad_norm": 92.10223388671875, "learning_rate": 9.65578836189196e-06, "loss": 0.4371, "step": 17820 }, { "epoch": 17.85, "grad_norm": 51.300636291503906, "learning_rate": 9.65464196558468e-06, "loss": 0.5363, "step": 17850 }, { "epoch": 17.88, "grad_norm": 61.75230026245117, "learning_rate": 9.653493731698467e-06, "loss": 0.5209, "step": 17880 }, { "epoch": 17.91, "grad_norm": 39.956085205078125, "learning_rate": 9.652343660686626e-06, "loss": 0.5377, "step": 17910 }, { "epoch": 17.94, "grad_norm": 45.845909118652344, "learning_rate": 9.651191753003187e-06, "loss": 0.5157, "step": 17940 }, { "epoch": 17.97, "grad_norm": 106.36433410644531, "learning_rate": 9.650038009102905e-06, "loss": 0.6505, "step": 17970 }, { "epoch": 18.0, "grad_norm": 55.83625793457031, "learning_rate": 9.648882429441258e-06, "loss": 0.5348, "step": 18000 }, { "epoch": 18.0, "eval_loss": 0.4467867314815521, "eval_map": 0.6909, "eval_map_50": 0.8959, "eval_map_75": 0.8328, "eval_map_chicken": 0.7055, "eval_map_duck": 0.586, "eval_map_large": 0.6633, "eval_map_medium": 0.6815, "eval_map_plant": 0.7811, "eval_map_small": 0.1362, "eval_mar_1": 0.2864, "eval_mar_10": 0.733, "eval_mar_100": 0.7378, "eval_mar_100_chicken": 0.7714, "eval_mar_100_duck": 0.6165, "eval_mar_100_plant": 0.8256, "eval_mar_large": 0.701, "eval_mar_medium": 0.7358, "eval_mar_small": 0.2592, "eval_runtime": 12.5637, "eval_samples_per_second": 15.919, "eval_steps_per_second": 1.99, "step": 18000 }, { "epoch": 18.03, "grad_norm": 76.63514709472656, "learning_rate": 9.647725014474452e-06, "loss": 0.5323, "step": 18030 }, { "epoch": 18.06, "grad_norm": 45.10977554321289, "learning_rate": 9.646565764659418e-06, "loss": 0.5052, "step": 18060 }, { "epoch": 18.09, "grad_norm": 81.5689697265625, "learning_rate": 9.645404680453805e-06, "loss": 0.6138, "step": 18090 }, { "epoch": 18.12, "grad_norm": 84.59944152832031, "learning_rate": 9.644241762315995e-06, "loss": 0.4884, "step": 18120 }, { "epoch": 18.15, "grad_norm": 105.48139953613281, "learning_rate": 9.643077010705088e-06, "loss": 0.5076, "step": 18150 }, { "epoch": 18.18, "grad_norm": 29.73552703857422, "learning_rate": 9.641910426080909e-06, "loss": 0.5001, "step": 18180 }, { "epoch": 18.21, "grad_norm": 65.31449127197266, "learning_rate": 9.640742008904006e-06, "loss": 0.5032, "step": 18210 }, { "epoch": 18.24, "grad_norm": 41.04438018798828, "learning_rate": 9.639571759635655e-06, "loss": 0.5289, "step": 18240 }, { "epoch": 18.27, "grad_norm": 45.612186431884766, "learning_rate": 9.63839967873785e-06, "loss": 0.5836, "step": 18270 }, { "epoch": 18.3, "grad_norm": 66.51431274414062, "learning_rate": 9.637225766673309e-06, "loss": 0.5875, "step": 18300 }, { "epoch": 18.33, "grad_norm": 46.82362747192383, "learning_rate": 9.636050023905473e-06, "loss": 0.5855, "step": 18330 }, { "epoch": 18.36, "grad_norm": 46.37489318847656, "learning_rate": 9.634872450898511e-06, "loss": 0.6096, "step": 18360 }, { "epoch": 18.39, "grad_norm": 115.26277160644531, "learning_rate": 9.633693048117306e-06, "loss": 0.4496, "step": 18390 }, { "epoch": 18.42, "grad_norm": 52.2386360168457, "learning_rate": 9.63251181602747e-06, "loss": 0.6467, "step": 18420 }, { "epoch": 18.45, "grad_norm": 44.465492248535156, "learning_rate": 9.631328755095334e-06, "loss": 0.7413, "step": 18450 }, { "epoch": 18.48, "grad_norm": 104.21498107910156, "learning_rate": 9.630143865787951e-06, "loss": 0.5893, "step": 18480 }, { "epoch": 18.51, "grad_norm": 127.77880859375, "learning_rate": 9.628957148573099e-06, "loss": 0.569, "step": 18510 }, { "epoch": 18.54, "grad_norm": 33.737037658691406, "learning_rate": 9.62776860391927e-06, "loss": 0.5874, "step": 18540 }, { "epoch": 18.57, "grad_norm": 57.467079162597656, "learning_rate": 9.626578232295689e-06, "loss": 0.5086, "step": 18570 }, { "epoch": 18.6, "grad_norm": 39.28734588623047, "learning_rate": 9.62538603417229e-06, "loss": 0.5893, "step": 18600 }, { "epoch": 18.63, "grad_norm": 53.36603546142578, "learning_rate": 9.624192010019739e-06, "loss": 0.5544, "step": 18630 }, { "epoch": 18.66, "grad_norm": 89.0500259399414, "learning_rate": 9.622996160309415e-06, "loss": 0.4616, "step": 18660 }, { "epoch": 18.69, "grad_norm": 35.472660064697266, "learning_rate": 9.62179848551342e-06, "loss": 0.5538, "step": 18690 }, { "epoch": 18.72, "grad_norm": 34.150299072265625, "learning_rate": 9.620598986104578e-06, "loss": 0.5944, "step": 18720 }, { "epoch": 18.75, "grad_norm": 128.7418975830078, "learning_rate": 9.619397662556434e-06, "loss": 0.5314, "step": 18750 }, { "epoch": 18.78, "grad_norm": 44.098426818847656, "learning_rate": 9.61819451534325e-06, "loss": 0.5272, "step": 18780 }, { "epoch": 18.81, "grad_norm": 66.03379821777344, "learning_rate": 9.616989544940008e-06, "loss": 0.5391, "step": 18810 }, { "epoch": 18.84, "grad_norm": 55.25798034667969, "learning_rate": 9.615782751822413e-06, "loss": 0.5648, "step": 18840 }, { "epoch": 18.87, "grad_norm": 0.0006358227110467851, "learning_rate": 9.61457413646689e-06, "loss": 0.5133, "step": 18870 }, { "epoch": 18.9, "grad_norm": 34.636749267578125, "learning_rate": 9.613363699350575e-06, "loss": 0.5354, "step": 18900 }, { "epoch": 18.93, "grad_norm": 76.81031036376953, "learning_rate": 9.612151440951334e-06, "loss": 0.55, "step": 18930 }, { "epoch": 18.96, "grad_norm": 80.6995620727539, "learning_rate": 9.610937361747747e-06, "loss": 0.532, "step": 18960 }, { "epoch": 18.99, "grad_norm": 53.454673767089844, "learning_rate": 9.609721462219115e-06, "loss": 0.5136, "step": 18990 }, { "epoch": 19.0, "eval_loss": 0.4364337921142578, "eval_map": 0.6945, "eval_map_50": 0.9219, "eval_map_75": 0.8428, "eval_map_chicken": 0.6837, "eval_map_duck": 0.6198, "eval_map_large": 0.73, "eval_map_medium": 0.6706, "eval_map_plant": 0.78, "eval_map_small": 0.1052, "eval_mar_1": 0.2905, "eval_mar_10": 0.7425, "eval_mar_100": 0.7467, "eval_mar_100_chicken": 0.7541, "eval_mar_100_duck": 0.6603, "eval_mar_100_plant": 0.8256, "eval_mar_large": 0.7721, "eval_mar_medium": 0.7333, "eval_mar_small": 0.254, "eval_runtime": 12.4637, "eval_samples_per_second": 16.047, "eval_steps_per_second": 2.006, "step": 19000 }, { "epoch": 19.02, "grad_norm": 78.36320495605469, "learning_rate": 9.60850374284545e-06, "loss": 0.4078, "step": 19020 }, { "epoch": 19.05, "grad_norm": 64.65117645263672, "learning_rate": 9.607284204107493e-06, "loss": 0.516, "step": 19050 }, { "epoch": 19.08, "grad_norm": 48.126094818115234, "learning_rate": 9.606062846486698e-06, "loss": 0.5063, "step": 19080 }, { "epoch": 19.11, "grad_norm": 50.75263214111328, "learning_rate": 9.604839670465236e-06, "loss": 0.5381, "step": 19110 }, { "epoch": 19.14, "grad_norm": 111.42890930175781, "learning_rate": 9.603614676526e-06, "loss": 0.4769, "step": 19140 }, { "epoch": 19.17, "grad_norm": 122.21208953857422, "learning_rate": 9.602387865152597e-06, "loss": 0.5639, "step": 19170 }, { "epoch": 19.2, "grad_norm": 68.34737396240234, "learning_rate": 9.601159236829353e-06, "loss": 0.5134, "step": 19200 }, { "epoch": 19.23, "grad_norm": 92.75222778320312, "learning_rate": 9.59992879204131e-06, "loss": 0.6263, "step": 19230 }, { "epoch": 19.26, "grad_norm": 65.23822784423828, "learning_rate": 9.598696531274227e-06, "loss": 0.447, "step": 19260 }, { "epoch": 19.29, "grad_norm": 85.12554168701172, "learning_rate": 9.597462455014585e-06, "loss": 0.608, "step": 19290 }, { "epoch": 19.32, "grad_norm": 37.38597869873047, "learning_rate": 9.596226563749575e-06, "loss": 0.7476, "step": 19320 }, { "epoch": 19.35, "grad_norm": 65.00273132324219, "learning_rate": 9.594988857967107e-06, "loss": 0.621, "step": 19350 }, { "epoch": 19.38, "grad_norm": 81.0069351196289, "learning_rate": 9.593749338155809e-06, "loss": 0.5607, "step": 19380 }, { "epoch": 19.41, "grad_norm": 55.96475601196289, "learning_rate": 9.592508004805023e-06, "loss": 0.622, "step": 19410 }, { "epoch": 19.44, "grad_norm": 54.13630676269531, "learning_rate": 9.591264858404809e-06, "loss": 0.542, "step": 19440 }, { "epoch": 19.47, "grad_norm": 43.575904846191406, "learning_rate": 9.59001989944594e-06, "loss": 0.4854, "step": 19470 }, { "epoch": 19.5, "grad_norm": 60.43488311767578, "learning_rate": 9.588773128419907e-06, "loss": 0.4844, "step": 19500 }, { "epoch": 19.53, "grad_norm": 45.04825973510742, "learning_rate": 9.587524545818914e-06, "loss": 0.6144, "step": 19530 }, { "epoch": 19.56, "grad_norm": 75.83597564697266, "learning_rate": 9.586274152135883e-06, "loss": 0.7366, "step": 19560 }, { "epoch": 19.59, "grad_norm": 90.47258758544922, "learning_rate": 9.58502194786445e-06, "loss": 0.5199, "step": 19590 }, { "epoch": 19.62, "grad_norm": 50.503944396972656, "learning_rate": 9.583767933498964e-06, "loss": 0.5882, "step": 19620 }, { "epoch": 19.65, "grad_norm": 44.3234748840332, "learning_rate": 9.58251210953449e-06, "loss": 0.4884, "step": 19650 }, { "epoch": 19.68, "grad_norm": 102.5786361694336, "learning_rate": 9.58125447646681e-06, "loss": 0.4806, "step": 19680 }, { "epoch": 19.71, "grad_norm": 78.55628967285156, "learning_rate": 9.579995034792415e-06, "loss": 0.4967, "step": 19710 }, { "epoch": 19.74, "grad_norm": 72.11976623535156, "learning_rate": 9.578733785008513e-06, "loss": 0.5233, "step": 19740 }, { "epoch": 19.77, "grad_norm": 59.13156509399414, "learning_rate": 9.577470727613025e-06, "loss": 0.5231, "step": 19770 }, { "epoch": 19.8, "grad_norm": 57.39673614501953, "learning_rate": 9.576205863104588e-06, "loss": 0.4209, "step": 19800 }, { "epoch": 19.83, "grad_norm": 67.8314208984375, "learning_rate": 9.574939191982549e-06, "loss": 0.6049, "step": 19830 }, { "epoch": 19.86, "grad_norm": 61.0339469909668, "learning_rate": 9.573670714746973e-06, "loss": 0.4383, "step": 19860 }, { "epoch": 19.89, "grad_norm": 73.04076385498047, "learning_rate": 9.572400431898627e-06, "loss": 0.4826, "step": 19890 }, { "epoch": 19.92, "grad_norm": 59.65742111206055, "learning_rate": 9.571128343939006e-06, "loss": 0.557, "step": 19920 }, { "epoch": 19.95, "grad_norm": 37.39786148071289, "learning_rate": 9.569854451370306e-06, "loss": 0.5631, "step": 19950 }, { "epoch": 19.98, "grad_norm": 46.49284744262695, "learning_rate": 9.568578754695443e-06, "loss": 0.7763, "step": 19980 }, { "epoch": 20.0, "eval_loss": 0.4124803841114044, "eval_map": 0.7082, "eval_map_50": 0.9235, "eval_map_75": 0.8439, "eval_map_chicken": 0.6954, "eval_map_duck": 0.6381, "eval_map_large": 0.729, "eval_map_medium": 0.6876, "eval_map_plant": 0.791, "eval_map_small": 0.115, "eval_mar_1": 0.2988, "eval_mar_10": 0.7549, "eval_mar_100": 0.7599, "eval_mar_100_chicken": 0.763, "eval_mar_100_duck": 0.6794, "eval_mar_100_plant": 0.8374, "eval_mar_large": 0.7776, "eval_mar_medium": 0.7483, "eval_mar_small": 0.2393, "eval_runtime": 13.0933, "eval_samples_per_second": 15.275, "eval_steps_per_second": 1.909, "step": 20000 }, { "epoch": 20.01, "grad_norm": 45.134273529052734, "learning_rate": 9.567301254418038e-06, "loss": 0.5728, "step": 20010 }, { "epoch": 20.04, "grad_norm": 104.69942474365234, "learning_rate": 9.566021951042432e-06, "loss": 0.6058, "step": 20040 }, { "epoch": 20.07, "grad_norm": 26.176664352416992, "learning_rate": 9.56474084507367e-06, "loss": 0.5354, "step": 20070 }, { "epoch": 20.1, "grad_norm": 222.71099853515625, "learning_rate": 9.563457937017514e-06, "loss": 0.5721, "step": 20100 }, { "epoch": 20.13, "grad_norm": 53.14629364013672, "learning_rate": 9.562173227380437e-06, "loss": 0.5897, "step": 20130 }, { "epoch": 20.16, "grad_norm": 126.2553939819336, "learning_rate": 9.56088671666962e-06, "loss": 0.5845, "step": 20160 }, { "epoch": 20.19, "grad_norm": 62.81528091430664, "learning_rate": 9.559598405392959e-06, "loss": 0.56, "step": 20190 }, { "epoch": 20.22, "grad_norm": 45.002830505371094, "learning_rate": 9.558308294059055e-06, "loss": 0.5533, "step": 20220 }, { "epoch": 20.25, "grad_norm": 62.43132019042969, "learning_rate": 9.557016383177226e-06, "loss": 0.5172, "step": 20250 }, { "epoch": 20.28, "grad_norm": 30.56139373779297, "learning_rate": 9.555722673257502e-06, "loss": 0.4833, "step": 20280 }, { "epoch": 20.31, "grad_norm": 77.70307159423828, "learning_rate": 9.55442716481061e-06, "loss": 0.4649, "step": 20310 }, { "epoch": 20.34, "grad_norm": 47.1893424987793, "learning_rate": 9.553129858348006e-06, "loss": 0.4674, "step": 20340 }, { "epoch": 20.37, "grad_norm": 45.644752502441406, "learning_rate": 9.55183075438184e-06, "loss": 0.5632, "step": 20370 }, { "epoch": 20.4, "grad_norm": 48.205726623535156, "learning_rate": 9.550529853424979e-06, "loss": 0.4268, "step": 20400 }, { "epoch": 20.43, "grad_norm": 31.547035217285156, "learning_rate": 9.549227155991e-06, "loss": 0.5553, "step": 20430 }, { "epoch": 20.46, "grad_norm": 39.21930694580078, "learning_rate": 9.547922662594183e-06, "loss": 0.5107, "step": 20460 }, { "epoch": 20.49, "grad_norm": 38.481021881103516, "learning_rate": 9.546616373749525e-06, "loss": 0.4495, "step": 20490 }, { "epoch": 20.52, "grad_norm": 41.75307846069336, "learning_rate": 9.545308289972727e-06, "loss": 0.5474, "step": 20520 }, { "epoch": 20.55, "grad_norm": 75.5237045288086, "learning_rate": 9.543998411780202e-06, "loss": 0.4847, "step": 20550 }, { "epoch": 20.58, "grad_norm": 49.65382766723633, "learning_rate": 9.542686739689064e-06, "loss": 0.644, "step": 20580 }, { "epoch": 20.61, "grad_norm": 58.422584533691406, "learning_rate": 9.541373274217145e-06, "loss": 0.5081, "step": 20610 }, { "epoch": 20.64, "grad_norm": 60.62272644042969, "learning_rate": 9.54005801588298e-06, "loss": 0.4621, "step": 20640 }, { "epoch": 20.67, "grad_norm": 37.92499923706055, "learning_rate": 9.538740965205809e-06, "loss": 0.53, "step": 20670 }, { "epoch": 20.7, "grad_norm": 95.88284301757812, "learning_rate": 9.537422122705585e-06, "loss": 0.4625, "step": 20700 }, { "epoch": 20.73, "grad_norm": 60.4619255065918, "learning_rate": 9.536101488902966e-06, "loss": 0.4765, "step": 20730 }, { "epoch": 20.76, "grad_norm": 386.47149658203125, "learning_rate": 9.534779064319318e-06, "loss": 0.5805, "step": 20760 }, { "epoch": 20.79, "grad_norm": 41.91823959350586, "learning_rate": 9.533454849476712e-06, "loss": 0.4908, "step": 20790 }, { "epoch": 20.82, "grad_norm": 39.06875228881836, "learning_rate": 9.532128844897928e-06, "loss": 0.4295, "step": 20820 }, { "epoch": 20.85, "grad_norm": 61.87434768676758, "learning_rate": 9.530801051106449e-06, "loss": 0.5672, "step": 20850 }, { "epoch": 20.88, "grad_norm": 279.25927734375, "learning_rate": 9.529471468626472e-06, "loss": 0.6674, "step": 20880 }, { "epoch": 20.91, "grad_norm": 100.80004119873047, "learning_rate": 9.52814009798289e-06, "loss": 0.5927, "step": 20910 }, { "epoch": 20.94, "grad_norm": 46.709293365478516, "learning_rate": 9.52680693970131e-06, "loss": 0.498, "step": 20940 }, { "epoch": 20.97, "grad_norm": 67.21540832519531, "learning_rate": 9.52547199430804e-06, "loss": 0.4749, "step": 20970 }, { "epoch": 21.0, "grad_norm": 24.060779571533203, "learning_rate": 9.524135262330098e-06, "loss": 0.531, "step": 21000 }, { "epoch": 21.0, "eval_loss": 0.418172687292099, "eval_map": 0.7035, "eval_map_50": 0.9164, "eval_map_75": 0.8414, "eval_map_chicken": 0.6977, "eval_map_duck": 0.6212, "eval_map_large": 0.7277, "eval_map_medium": 0.6748, "eval_map_plant": 0.7917, "eval_map_small": 0.1179, "eval_mar_1": 0.297, "eval_mar_10": 0.7502, "eval_mar_100": 0.7537, "eval_mar_100_chicken": 0.7618, "eval_mar_100_duck": 0.6639, "eval_mar_100_plant": 0.8354, "eval_mar_large": 0.7777, "eval_mar_medium": 0.7369, "eval_mar_small": 0.2354, "eval_runtime": 12.6282, "eval_samples_per_second": 15.838, "eval_steps_per_second": 1.98, "step": 21000 }, { "epoch": 21.03, "grad_norm": 42.197547912597656, "learning_rate": 9.522796744295202e-06, "loss": 0.4398, "step": 21030 }, { "epoch": 21.06, "grad_norm": 65.19886016845703, "learning_rate": 9.52145644073178e-06, "loss": 0.4483, "step": 21060 }, { "epoch": 21.09, "grad_norm": 153.0055694580078, "learning_rate": 9.52011435216896e-06, "loss": 0.6826, "step": 21090 }, { "epoch": 21.12, "grad_norm": 70.12749481201172, "learning_rate": 9.51877047913658e-06, "loss": 0.5087, "step": 21120 }, { "epoch": 21.15, "grad_norm": 40.320350646972656, "learning_rate": 9.517424822165175e-06, "loss": 0.5014, "step": 21150 }, { "epoch": 21.18, "grad_norm": 72.41119384765625, "learning_rate": 9.516077381785995e-06, "loss": 0.5015, "step": 21180 }, { "epoch": 21.21, "grad_norm": 33.972877502441406, "learning_rate": 9.514728158530983e-06, "loss": 0.4939, "step": 21210 }, { "epoch": 21.24, "grad_norm": 68.28968811035156, "learning_rate": 9.513377152932796e-06, "loss": 0.4447, "step": 21240 }, { "epoch": 21.27, "grad_norm": 70.25806427001953, "learning_rate": 9.512024365524788e-06, "loss": 0.466, "step": 21270 }, { "epoch": 21.3, "grad_norm": 59.76402282714844, "learning_rate": 9.510669796841014e-06, "loss": 0.5843, "step": 21300 }, { "epoch": 21.33, "grad_norm": 52.857879638671875, "learning_rate": 9.509313447416241e-06, "loss": 0.5008, "step": 21330 }, { "epoch": 21.36, "grad_norm": 67.54583740234375, "learning_rate": 9.507955317785935e-06, "loss": 0.4397, "step": 21360 }, { "epoch": 21.39, "grad_norm": 37.361236572265625, "learning_rate": 9.50659540848626e-06, "loss": 0.4764, "step": 21390 }, { "epoch": 21.42, "grad_norm": 44.237422943115234, "learning_rate": 9.505233720054086e-06, "loss": 0.6823, "step": 21420 }, { "epoch": 21.45, "grad_norm": 49.40839767456055, "learning_rate": 9.503870253026992e-06, "loss": 0.4639, "step": 21450 }, { "epoch": 21.48, "grad_norm": 42.29047775268555, "learning_rate": 9.502505007943248e-06, "loss": 0.4887, "step": 21480 }, { "epoch": 21.51, "grad_norm": 45.91970443725586, "learning_rate": 9.501137985341832e-06, "loss": 0.4622, "step": 21510 }, { "epoch": 21.54, "grad_norm": 49.44613265991211, "learning_rate": 9.499769185762425e-06, "loss": 0.467, "step": 21540 }, { "epoch": 21.57, "grad_norm": 55.7149658203125, "learning_rate": 9.498398609745405e-06, "loss": 0.6501, "step": 21570 }, { "epoch": 21.6, "grad_norm": 61.80874252319336, "learning_rate": 9.497026257831856e-06, "loss": 0.6241, "step": 21600 }, { "epoch": 21.63, "grad_norm": 452.8546142578125, "learning_rate": 9.49565213056356e-06, "loss": 0.5597, "step": 21630 }, { "epoch": 21.66, "grad_norm": 85.00440216064453, "learning_rate": 9.494276228482998e-06, "loss": 0.474, "step": 21660 }, { "epoch": 21.69, "grad_norm": 34.22127914428711, "learning_rate": 9.492898552133358e-06, "loss": 0.5381, "step": 21690 }, { "epoch": 21.72, "grad_norm": 146.12246704101562, "learning_rate": 9.491519102058523e-06, "loss": 0.5707, "step": 21720 }, { "epoch": 21.75, "grad_norm": 58.734153747558594, "learning_rate": 9.490137878803078e-06, "loss": 0.47, "step": 21750 }, { "epoch": 21.78, "grad_norm": 33.87949752807617, "learning_rate": 9.48875488291231e-06, "loss": 0.5199, "step": 21780 }, { "epoch": 21.81, "grad_norm": 59.80483627319336, "learning_rate": 9.487370114932201e-06, "loss": 0.6429, "step": 21810 }, { "epoch": 21.84, "grad_norm": 48.99253463745117, "learning_rate": 9.48598357540944e-06, "loss": 0.4652, "step": 21840 }, { "epoch": 21.87, "grad_norm": 0.4198257625102997, "learning_rate": 9.484595264891403e-06, "loss": 0.5143, "step": 21870 }, { "epoch": 21.9, "grad_norm": 45.922611236572266, "learning_rate": 9.48320518392618e-06, "loss": 0.4599, "step": 21900 }, { "epoch": 21.93, "grad_norm": 78.56241607666016, "learning_rate": 9.48181333306255e-06, "loss": 0.5004, "step": 21930 }, { "epoch": 21.96, "grad_norm": 85.15208435058594, "learning_rate": 9.480419712849996e-06, "loss": 0.5271, "step": 21960 }, { "epoch": 21.99, "grad_norm": 42.14804458618164, "learning_rate": 9.479024323838694e-06, "loss": 0.5738, "step": 21990 }, { "epoch": 22.0, "eval_loss": 0.4123690724372864, "eval_map": 0.704, "eval_map_50": 0.924, "eval_map_75": 0.852, "eval_map_chicken": 0.6921, "eval_map_duck": 0.6329, "eval_map_large": 0.7077, "eval_map_medium": 0.6842, "eval_map_plant": 0.787, "eval_map_small": 0.1622, "eval_mar_1": 0.2951, "eval_mar_10": 0.7477, "eval_mar_100": 0.7512, "eval_mar_100_chicken": 0.7501, "eval_mar_100_duck": 0.6696, "eval_mar_100_plant": 0.8338, "eval_mar_large": 0.7539, "eval_mar_medium": 0.7373, "eval_mar_small": 0.2837, "eval_runtime": 12.5896, "eval_samples_per_second": 15.886, "eval_steps_per_second": 1.986, "step": 22000 }, { "epoch": 22.02, "grad_norm": 47.1718864440918, "learning_rate": 9.477627166579523e-06, "loss": 0.539, "step": 22020 }, { "epoch": 22.05, "grad_norm": 104.59381103515625, "learning_rate": 9.476228241624059e-06, "loss": 0.4189, "step": 22050 }, { "epoch": 22.08, "grad_norm": 1335.8126220703125, "learning_rate": 9.474827549524574e-06, "loss": 0.5653, "step": 22080 }, { "epoch": 22.11, "grad_norm": 50.901676177978516, "learning_rate": 9.473425090834041e-06, "loss": 0.4772, "step": 22110 }, { "epoch": 22.14, "grad_norm": 74.63755798339844, "learning_rate": 9.472020866106128e-06, "loss": 0.5645, "step": 22140 }, { "epoch": 22.17, "grad_norm": 42.569313049316406, "learning_rate": 9.4706148758952e-06, "loss": 0.4804, "step": 22170 }, { "epoch": 22.2, "grad_norm": 60.32633972167969, "learning_rate": 9.46920712075632e-06, "loss": 0.6011, "step": 22200 }, { "epoch": 22.23, "grad_norm": 81.35128784179688, "learning_rate": 9.467797601245246e-06, "loss": 0.5266, "step": 22230 }, { "epoch": 22.26, "grad_norm": 44.596336364746094, "learning_rate": 9.466386317918436e-06, "loss": 0.5043, "step": 22260 }, { "epoch": 22.29, "grad_norm": 51.18402862548828, "learning_rate": 9.464973271333042e-06, "loss": 0.5534, "step": 22290 }, { "epoch": 22.32, "grad_norm": 39.9325065612793, "learning_rate": 9.463558462046912e-06, "loss": 0.5066, "step": 22320 }, { "epoch": 22.35, "grad_norm": 38.61850357055664, "learning_rate": 9.46214189061859e-06, "loss": 0.4721, "step": 22350 }, { "epoch": 22.38, "grad_norm": 50.930301666259766, "learning_rate": 9.460723557607317e-06, "loss": 0.5231, "step": 22380 }, { "epoch": 22.41, "grad_norm": 38.76648712158203, "learning_rate": 9.459303463573027e-06, "loss": 0.4129, "step": 22410 }, { "epoch": 22.44, "grad_norm": 38.96125793457031, "learning_rate": 9.457881609076352e-06, "loss": 0.4546, "step": 22440 }, { "epoch": 22.47, "grad_norm": 53.50430679321289, "learning_rate": 9.456457994678616e-06, "loss": 0.4035, "step": 22470 }, { "epoch": 22.5, "grad_norm": 59.15469741821289, "learning_rate": 9.45503262094184e-06, "loss": 0.4956, "step": 22500 }, { "epoch": 22.53, "grad_norm": 28.79969024658203, "learning_rate": 9.453605488428741e-06, "loss": 0.4878, "step": 22530 }, { "epoch": 22.56, "grad_norm": 71.59202575683594, "learning_rate": 9.452176597702724e-06, "loss": 0.5711, "step": 22560 }, { "epoch": 22.59, "grad_norm": 83.72734069824219, "learning_rate": 9.450745949327897e-06, "loss": 0.4144, "step": 22590 }, { "epoch": 22.62, "grad_norm": 147.85125732421875, "learning_rate": 9.449313543869056e-06, "loss": 0.4192, "step": 22620 }, { "epoch": 22.65, "grad_norm": 37.386112213134766, "learning_rate": 9.447879381891691e-06, "loss": 0.4645, "step": 22650 }, { "epoch": 22.68, "grad_norm": 42.89706802368164, "learning_rate": 9.446443463961986e-06, "loss": 0.5191, "step": 22680 }, { "epoch": 22.71, "grad_norm": 50.6859016418457, "learning_rate": 9.44500579064682e-06, "loss": 0.4183, "step": 22710 }, { "epoch": 22.74, "grad_norm": 53.751888275146484, "learning_rate": 9.443566362513763e-06, "loss": 0.5052, "step": 22740 }, { "epoch": 22.77, "grad_norm": 59.30571365356445, "learning_rate": 9.44212518013108e-06, "loss": 0.5687, "step": 22770 }, { "epoch": 22.8, "grad_norm": 33.73162078857422, "learning_rate": 9.440682244067724e-06, "loss": 0.4787, "step": 22800 }, { "epoch": 22.83, "grad_norm": 109.20848846435547, "learning_rate": 9.439237554893344e-06, "loss": 0.4545, "step": 22830 }, { "epoch": 22.86, "grad_norm": 30.56178092956543, "learning_rate": 9.437791113178283e-06, "loss": 0.5608, "step": 22860 }, { "epoch": 22.89, "grad_norm": 42.11906051635742, "learning_rate": 9.436342919493571e-06, "loss": 0.4726, "step": 22890 }, { "epoch": 22.92, "grad_norm": 39.664161682128906, "learning_rate": 9.434892974410932e-06, "loss": 0.4497, "step": 22920 }, { "epoch": 22.95, "grad_norm": 50.95674514770508, "learning_rate": 9.433441278502784e-06, "loss": 0.4276, "step": 22950 }, { "epoch": 22.98, "grad_norm": 31.154773712158203, "learning_rate": 9.431987832342228e-06, "loss": 0.4659, "step": 22980 }, { "epoch": 23.0, "eval_loss": 0.38811659812927246, "eval_map": 0.7229, "eval_map_50": 0.9359, "eval_map_75": 0.8721, "eval_map_chicken": 0.7215, "eval_map_duck": 0.6537, "eval_map_large": 0.7482, "eval_map_medium": 0.6982, "eval_map_plant": 0.7935, "eval_map_small": 0.1295, "eval_mar_1": 0.3033, "eval_mar_10": 0.7637, "eval_mar_100": 0.7699, "eval_mar_100_chicken": 0.7793, "eval_mar_100_duck": 0.6887, "eval_mar_100_plant": 0.8418, "eval_mar_large": 0.7945, "eval_mar_medium": 0.7558, "eval_mar_small": 0.3028, "eval_runtime": 12.87, "eval_samples_per_second": 15.54, "eval_steps_per_second": 1.943, "step": 23000 }, { "epoch": 23.01, "grad_norm": 80.10482788085938, "learning_rate": 9.430532636503067e-06, "loss": 0.5563, "step": 23010 }, { "epoch": 23.04, "grad_norm": 21.9283447265625, "learning_rate": 9.429075691559788e-06, "loss": 0.5358, "step": 23040 }, { "epoch": 23.07, "grad_norm": 28.550405502319336, "learning_rate": 9.427616998087568e-06, "loss": 0.6146, "step": 23070 }, { "epoch": 23.1, "grad_norm": 100.07318878173828, "learning_rate": 9.426156556662276e-06, "loss": 0.4679, "step": 23100 }, { "epoch": 23.13, "grad_norm": 129.5098114013672, "learning_rate": 9.424694367860475e-06, "loss": 0.5126, "step": 23130 }, { "epoch": 23.16, "grad_norm": 51.278106689453125, "learning_rate": 9.423230432259409e-06, "loss": 0.4638, "step": 23160 }, { "epoch": 23.19, "grad_norm": 50.4608268737793, "learning_rate": 9.421764750437019e-06, "loss": 0.5146, "step": 23190 }, { "epoch": 23.22, "grad_norm": 73.81622314453125, "learning_rate": 9.420297322971934e-06, "loss": 0.4617, "step": 23220 }, { "epoch": 23.25, "grad_norm": 34.30580139160156, "learning_rate": 9.418828150443469e-06, "loss": 0.4124, "step": 23250 }, { "epoch": 23.28, "grad_norm": 67.1881103515625, "learning_rate": 9.41735723343163e-06, "loss": 0.5615, "step": 23280 }, { "epoch": 23.31, "grad_norm": 91.84432220458984, "learning_rate": 9.415884572517113e-06, "loss": 0.5808, "step": 23310 }, { "epoch": 23.34, "grad_norm": 300.571044921875, "learning_rate": 9.414410168281303e-06, "loss": 0.5066, "step": 23340 }, { "epoch": 23.37, "grad_norm": 45.92863845825195, "learning_rate": 9.412934021306267e-06, "loss": 0.4242, "step": 23370 }, { "epoch": 23.4, "grad_norm": 125.1700439453125, "learning_rate": 9.411456132174768e-06, "loss": 0.539, "step": 23400 }, { "epoch": 23.43, "grad_norm": 137.93832397460938, "learning_rate": 9.40997650147025e-06, "loss": 0.5048, "step": 23430 }, { "epoch": 23.46, "grad_norm": 49.78171920776367, "learning_rate": 9.408495129776851e-06, "loss": 0.5404, "step": 23460 }, { "epoch": 23.49, "grad_norm": 4.611354961525649e-05, "learning_rate": 9.407012017679393e-06, "loss": 0.4783, "step": 23490 }, { "epoch": 23.52, "grad_norm": 32.581390380859375, "learning_rate": 9.405527165763384e-06, "loss": 0.4441, "step": 23520 }, { "epoch": 23.55, "grad_norm": 35.50736999511719, "learning_rate": 9.404040574615018e-06, "loss": 0.8102, "step": 23550 }, { "epoch": 23.58, "grad_norm": 77.08589935302734, "learning_rate": 9.402552244821181e-06, "loss": 0.6079, "step": 23580 }, { "epoch": 23.61, "grad_norm": 53.91266632080078, "learning_rate": 9.401062176969442e-06, "loss": 0.4761, "step": 23610 }, { "epoch": 23.64, "grad_norm": 68.92446899414062, "learning_rate": 9.399570371648052e-06, "loss": 0.4542, "step": 23640 }, { "epoch": 23.67, "grad_norm": 33.291717529296875, "learning_rate": 9.398076829445958e-06, "loss": 0.4851, "step": 23670 }, { "epoch": 23.7, "grad_norm": 53.06465148925781, "learning_rate": 9.396581550952781e-06, "loss": 0.4596, "step": 23700 }, { "epoch": 23.73, "grad_norm": 100.1142807006836, "learning_rate": 9.395084536758838e-06, "loss": 0.4328, "step": 23730 }, { "epoch": 23.76, "grad_norm": 30.885377883911133, "learning_rate": 9.393585787455125e-06, "loss": 0.4499, "step": 23760 }, { "epoch": 23.79, "grad_norm": 779.153076171875, "learning_rate": 9.392085303633322e-06, "loss": 0.5216, "step": 23790 }, { "epoch": 23.82, "grad_norm": 58.899295806884766, "learning_rate": 9.3905830858858e-06, "loss": 0.49, "step": 23820 }, { "epoch": 23.85, "grad_norm": 100.65608978271484, "learning_rate": 9.38907913480561e-06, "loss": 0.482, "step": 23850 }, { "epoch": 23.88, "grad_norm": 67.3552017211914, "learning_rate": 9.387573450986485e-06, "loss": 0.4977, "step": 23880 }, { "epoch": 23.91, "grad_norm": 44.643035888671875, "learning_rate": 9.386066035022849e-06, "loss": 0.5198, "step": 23910 }, { "epoch": 23.94, "grad_norm": 68.50333404541016, "learning_rate": 9.384556887509802e-06, "loss": 0.4707, "step": 23940 }, { "epoch": 23.97, "grad_norm": 77.25891876220703, "learning_rate": 9.383046009043134e-06, "loss": 0.5368, "step": 23970 }, { "epoch": 24.0, "grad_norm": 37.26203918457031, "learning_rate": 9.381533400219319e-06, "loss": 0.476, "step": 24000 }, { "epoch": 24.0, "eval_loss": 0.40412285923957825, "eval_map": 0.6974, "eval_map_50": 0.9329, "eval_map_75": 0.8277, "eval_map_chicken": 0.6745, "eval_map_duck": 0.6167, "eval_map_large": 0.7309, "eval_map_medium": 0.6708, "eval_map_plant": 0.8008, "eval_map_small": 0.1692, "eval_mar_1": 0.2946, "eval_mar_10": 0.7399, "eval_mar_100": 0.7446, "eval_mar_100_chicken": 0.7354, "eval_mar_100_duck": 0.6562, "eval_mar_100_plant": 0.8422, "eval_mar_large": 0.7784, "eval_mar_medium": 0.7272, "eval_mar_small": 0.2606, "eval_runtime": 12.5645, "eval_samples_per_second": 15.918, "eval_steps_per_second": 1.99, "step": 24000 }, { "epoch": 24.03, "grad_norm": 61.880393981933594, "learning_rate": 9.380019061635506e-06, "loss": 0.5, "step": 24030 }, { "epoch": 24.06, "grad_norm": 48.862823486328125, "learning_rate": 9.378502993889533e-06, "loss": 0.4851, "step": 24060 }, { "epoch": 24.09, "grad_norm": 0.00019921764032915235, "learning_rate": 9.376985197579919e-06, "loss": 0.4333, "step": 24090 }, { "epoch": 24.12, "grad_norm": 40.57404708862305, "learning_rate": 9.37546567330587e-06, "loss": 0.5593, "step": 24120 }, { "epoch": 24.15, "grad_norm": 47.584529876708984, "learning_rate": 9.373944421667264e-06, "loss": 0.4892, "step": 24150 }, { "epoch": 24.18, "grad_norm": 56.86312484741211, "learning_rate": 9.372421443264672e-06, "loss": 0.4791, "step": 24180 }, { "epoch": 24.21, "grad_norm": 49.94560241699219, "learning_rate": 9.37089673869934e-06, "loss": 0.5511, "step": 24210 }, { "epoch": 24.24, "grad_norm": 36.18058395385742, "learning_rate": 9.369370308573198e-06, "loss": 0.4231, "step": 24240 }, { "epoch": 24.27, "grad_norm": 42.51197814941406, "learning_rate": 9.367842153488853e-06, "loss": 0.4333, "step": 24270 }, { "epoch": 24.3, "grad_norm": 30.495460510253906, "learning_rate": 9.366312274049602e-06, "loss": 0.4398, "step": 24300 }, { "epoch": 24.33, "grad_norm": 46.86597442626953, "learning_rate": 9.364780670859412e-06, "loss": 0.4444, "step": 24330 }, { "epoch": 24.36, "grad_norm": 46.085304260253906, "learning_rate": 9.363247344522939e-06, "loss": 0.6074, "step": 24360 }, { "epoch": 24.39, "grad_norm": 40.743377685546875, "learning_rate": 9.361712295645515e-06, "loss": 0.4598, "step": 24390 }, { "epoch": 24.42, "grad_norm": 58.11893081665039, "learning_rate": 9.360175524833153e-06, "loss": 0.526, "step": 24420 }, { "epoch": 24.45, "grad_norm": 61.323055267333984, "learning_rate": 9.358637032692546e-06, "loss": 0.579, "step": 24450 }, { "epoch": 24.48, "grad_norm": 58.84001541137695, "learning_rate": 9.357096819831065e-06, "loss": 0.4251, "step": 24480 }, { "epoch": 24.51, "grad_norm": 86.30671691894531, "learning_rate": 9.355554886856762e-06, "loss": 0.5249, "step": 24510 }, { "epoch": 24.54, "grad_norm": 0.0010670045157894492, "learning_rate": 9.35401123437837e-06, "loss": 0.3991, "step": 24540 }, { "epoch": 24.57, "grad_norm": 57.24353790283203, "learning_rate": 9.352465863005295e-06, "loss": 0.4569, "step": 24570 }, { "epoch": 24.6, "grad_norm": 71.466796875, "learning_rate": 9.35091877334763e-06, "loss": 0.4682, "step": 24600 }, { "epoch": 24.63, "grad_norm": 37.96323776245117, "learning_rate": 9.349369966016135e-06, "loss": 0.4854, "step": 24630 }, { "epoch": 24.66, "grad_norm": 45.79912567138672, "learning_rate": 9.347819441622261e-06, "loss": 0.5178, "step": 24660 }, { "epoch": 24.69, "grad_norm": 60.6125602722168, "learning_rate": 9.346267200778127e-06, "loss": 0.6245, "step": 24690 }, { "epoch": 24.72, "grad_norm": 62.02178955078125, "learning_rate": 9.344713244096533e-06, "loss": 0.5198, "step": 24720 }, { "epoch": 24.75, "grad_norm": 0.00032934959745034575, "learning_rate": 9.343157572190957e-06, "loss": 0.5642, "step": 24750 }, { "epoch": 24.78, "grad_norm": 135.9851837158203, "learning_rate": 9.341600185675555e-06, "loss": 0.5526, "step": 24780 }, { "epoch": 24.81, "grad_norm": 34.97378921508789, "learning_rate": 9.340041085165157e-06, "loss": 0.4974, "step": 24810 }, { "epoch": 24.84, "grad_norm": 352.3426513671875, "learning_rate": 9.33848027127527e-06, "loss": 0.5693, "step": 24840 }, { "epoch": 24.87, "grad_norm": 55.76923751831055, "learning_rate": 9.336917744622083e-06, "loss": 0.4378, "step": 24870 }, { "epoch": 24.9, "grad_norm": 88.11799621582031, "learning_rate": 9.33535350582245e-06, "loss": 0.4014, "step": 24900 }, { "epoch": 24.93, "grad_norm": 65.13556671142578, "learning_rate": 9.333787555493915e-06, "loss": 0.4685, "step": 24930 }, { "epoch": 24.96, "grad_norm": 39.27876663208008, "learning_rate": 9.332219894254686e-06, "loss": 0.3969, "step": 24960 }, { "epoch": 24.99, "grad_norm": 55.397918701171875, "learning_rate": 9.330650522723653e-06, "loss": 0.4144, "step": 24990 }, { "epoch": 25.0, "eval_loss": 0.3696702718734741, "eval_map": 0.7384, "eval_map_50": 0.9476, "eval_map_75": 0.8756, "eval_map_chicken": 0.7223, "eval_map_duck": 0.6905, "eval_map_large": 0.7618, "eval_map_medium": 0.7133, "eval_map_plant": 0.8024, "eval_map_small": 0.1688, "eval_mar_1": 0.3101, "eval_mar_10": 0.7763, "eval_mar_100": 0.7809, "eval_mar_100_chicken": 0.773, "eval_mar_100_duck": 0.7237, "eval_mar_100_plant": 0.846, "eval_mar_large": 0.8109, "eval_mar_medium": 0.7666, "eval_mar_small": 0.2761, "eval_runtime": 12.4777, "eval_samples_per_second": 16.029, "eval_steps_per_second": 2.004, "step": 25000 }, { "epoch": 25.02, "grad_norm": 49.776458740234375, "learning_rate": 9.329079441520377e-06, "loss": 0.4893, "step": 25020 }, { "epoch": 25.05, "grad_norm": 83.90098571777344, "learning_rate": 9.327506651265096e-06, "loss": 0.4632, "step": 25050 }, { "epoch": 25.08, "grad_norm": 46.5289421081543, "learning_rate": 9.325932152578726e-06, "loss": 0.5062, "step": 25080 }, { "epoch": 25.11, "grad_norm": 58.45944595336914, "learning_rate": 9.324355946082848e-06, "loss": 0.4177, "step": 25110 }, { "epoch": 25.14, "grad_norm": 75.60499572753906, "learning_rate": 9.322778032399728e-06, "loss": 0.3802, "step": 25140 }, { "epoch": 25.17, "grad_norm": 86.79434967041016, "learning_rate": 9.321198412152303e-06, "loss": 0.4937, "step": 25170 }, { "epoch": 25.2, "grad_norm": 66.01306915283203, "learning_rate": 9.319617085964177e-06, "loss": 0.6052, "step": 25200 }, { "epoch": 25.23, "grad_norm": 82.29691314697266, "learning_rate": 9.318034054459637e-06, "loss": 0.5417, "step": 25230 }, { "epoch": 25.26, "grad_norm": 60.890052795410156, "learning_rate": 9.316449318263635e-06, "loss": 0.5056, "step": 25260 }, { "epoch": 25.29, "grad_norm": 46.09884262084961, "learning_rate": 9.314862878001802e-06, "loss": 0.4614, "step": 25290 }, { "epoch": 25.32, "grad_norm": 27.451629638671875, "learning_rate": 9.31327473430044e-06, "loss": 0.4634, "step": 25320 }, { "epoch": 25.35, "grad_norm": 65.26049041748047, "learning_rate": 9.31168488778652e-06, "loss": 0.478, "step": 25350 }, { "epoch": 25.38, "grad_norm": 43.6810417175293, "learning_rate": 9.31009333908769e-06, "loss": 0.5279, "step": 25380 }, { "epoch": 25.41, "grad_norm": 148.8280792236328, "learning_rate": 9.308500088832271e-06, "loss": 0.4633, "step": 25410 }, { "epoch": 25.44, "grad_norm": 242.63394165039062, "learning_rate": 9.30690513764925e-06, "loss": 0.5544, "step": 25440 }, { "epoch": 25.47, "grad_norm": 88.82725524902344, "learning_rate": 9.305308486168288e-06, "loss": 0.3768, "step": 25470 }, { "epoch": 25.5, "grad_norm": 47.45256805419922, "learning_rate": 9.30371013501972e-06, "loss": 0.4609, "step": 25500 }, { "epoch": 25.53, "grad_norm": 58.19525909423828, "learning_rate": 9.302110084834545e-06, "loss": 0.4608, "step": 25530 }, { "epoch": 25.56, "grad_norm": 40.35570526123047, "learning_rate": 9.300508336244443e-06, "loss": 0.521, "step": 25560 }, { "epoch": 25.59, "grad_norm": 77.77928924560547, "learning_rate": 9.298904889881757e-06, "loss": 0.5415, "step": 25590 }, { "epoch": 25.62, "grad_norm": 60.30169677734375, "learning_rate": 9.297299746379503e-06, "loss": 0.5102, "step": 25620 }, { "epoch": 25.65, "grad_norm": 27.128982543945312, "learning_rate": 9.295692906371362e-06, "loss": 0.463, "step": 25650 }, { "epoch": 25.68, "grad_norm": 42.933475494384766, "learning_rate": 9.294084370491695e-06, "loss": 0.4297, "step": 25680 }, { "epoch": 25.71, "grad_norm": 43.94452667236328, "learning_rate": 9.292474139375523e-06, "loss": 0.4484, "step": 25710 }, { "epoch": 25.74, "grad_norm": 85.80694580078125, "learning_rate": 9.29086221365854e-06, "loss": 0.4167, "step": 25740 }, { "epoch": 25.77, "grad_norm": 28.034530639648438, "learning_rate": 9.28924859397711e-06, "loss": 0.552, "step": 25770 }, { "epoch": 25.8, "grad_norm": 70.52547454833984, "learning_rate": 9.287633280968263e-06, "loss": 0.4515, "step": 25800 }, { "epoch": 25.83, "grad_norm": 153.11341857910156, "learning_rate": 9.2860162752697e-06, "loss": 0.5199, "step": 25830 }, { "epoch": 25.86, "grad_norm": 90.3126220703125, "learning_rate": 9.284397577519788e-06, "loss": 0.5371, "step": 25860 }, { "epoch": 25.89, "grad_norm": 51.882530212402344, "learning_rate": 9.282777188357563e-06, "loss": 0.5151, "step": 25890 }, { "epoch": 25.92, "grad_norm": 34.07367706298828, "learning_rate": 9.281155108422732e-06, "loss": 0.4622, "step": 25920 }, { "epoch": 25.95, "grad_norm": 5.449441960081458e-05, "learning_rate": 9.279531338355666e-06, "loss": 0.4429, "step": 25950 }, { "epoch": 25.98, "grad_norm": 65.40760803222656, "learning_rate": 9.277905878797401e-06, "loss": 0.6979, "step": 25980 }, { "epoch": 26.0, "eval_loss": 0.3698991537094116, "eval_map": 0.7347, "eval_map_50": 0.9514, "eval_map_75": 0.8711, "eval_map_chicken": 0.7269, "eval_map_duck": 0.6696, "eval_map_large": 0.73, "eval_map_medium": 0.7181, "eval_map_plant": 0.8076, "eval_map_small": 0.1661, "eval_mar_1": 0.3042, "eval_mar_10": 0.7751, "eval_mar_100": 0.7798, "eval_mar_100_chicken": 0.7813, "eval_mar_100_duck": 0.7113, "eval_mar_100_plant": 0.8469, "eval_mar_large": 0.7747, "eval_mar_medium": 0.7739, "eval_mar_small": 0.274, "eval_runtime": 13.0459, "eval_samples_per_second": 15.331, "eval_steps_per_second": 1.916, "step": 26000 }, { "epoch": 26.01, "grad_norm": 45.000553131103516, "learning_rate": 9.276278730389642e-06, "loss": 0.4107, "step": 26010 }, { "epoch": 26.04, "grad_norm": 73.11897277832031, "learning_rate": 9.274649893774768e-06, "loss": 0.5214, "step": 26040 }, { "epoch": 26.07, "grad_norm": 58.98204803466797, "learning_rate": 9.27301936959581e-06, "loss": 0.5104, "step": 26070 }, { "epoch": 26.1, "grad_norm": 81.72382354736328, "learning_rate": 9.271387158496477e-06, "loss": 0.6775, "step": 26100 }, { "epoch": 26.13, "grad_norm": 124.47724151611328, "learning_rate": 9.269753261121139e-06, "loss": 0.4383, "step": 26130 }, { "epoch": 26.16, "grad_norm": 92.1301498413086, "learning_rate": 9.268117678114833e-06, "loss": 0.4322, "step": 26160 }, { "epoch": 26.19, "grad_norm": 55.20294952392578, "learning_rate": 9.266480410123264e-06, "loss": 0.5318, "step": 26190 }, { "epoch": 26.22, "grad_norm": 135.23394775390625, "learning_rate": 9.264841457792795e-06, "loss": 0.4637, "step": 26220 }, { "epoch": 26.25, "grad_norm": 41.55864334106445, "learning_rate": 9.263200821770462e-06, "loss": 0.5264, "step": 26250 }, { "epoch": 26.28, "grad_norm": 55.87533187866211, "learning_rate": 9.26155850270396e-06, "loss": 0.5097, "step": 26280 }, { "epoch": 26.31, "grad_norm": 61.20616912841797, "learning_rate": 9.259914501241651e-06, "loss": 0.5378, "step": 26310 }, { "epoch": 26.34, "grad_norm": 48.59476089477539, "learning_rate": 9.25826881803256e-06, "loss": 0.4542, "step": 26340 }, { "epoch": 26.37, "grad_norm": 0.0002657757722772658, "learning_rate": 9.25662145372638e-06, "loss": 0.4202, "step": 26370 }, { "epoch": 26.4, "grad_norm": 36.71871566772461, "learning_rate": 9.25497240897346e-06, "loss": 0.4198, "step": 26400 }, { "epoch": 26.43, "grad_norm": 75.85392761230469, "learning_rate": 9.25332168442482e-06, "loss": 0.5523, "step": 26430 }, { "epoch": 26.46, "grad_norm": 616.0466918945312, "learning_rate": 9.251669280732137e-06, "loss": 0.4992, "step": 26460 }, { "epoch": 26.49, "grad_norm": 73.85807800292969, "learning_rate": 9.250015198547757e-06, "loss": 0.52, "step": 26490 }, { "epoch": 26.52, "grad_norm": 165.83163452148438, "learning_rate": 9.248359438524683e-06, "loss": 0.4366, "step": 26520 }, { "epoch": 26.55, "grad_norm": 73.1701431274414, "learning_rate": 9.246702001316584e-06, "loss": 0.454, "step": 26550 }, { "epoch": 26.58, "grad_norm": 37.664306640625, "learning_rate": 9.245042887577789e-06, "loss": 0.5571, "step": 26580 }, { "epoch": 26.61, "grad_norm": 61.83356857299805, "learning_rate": 9.243382097963292e-06, "loss": 0.4622, "step": 26610 }, { "epoch": 26.64, "grad_norm": 37.221405029296875, "learning_rate": 9.241719633128743e-06, "loss": 0.4803, "step": 26640 }, { "epoch": 26.67, "grad_norm": 72.06493377685547, "learning_rate": 9.24005549373046e-06, "loss": 0.6757, "step": 26670 }, { "epoch": 26.7, "grad_norm": 0.00010413204290671274, "learning_rate": 9.238389680425417e-06, "loss": 0.4552, "step": 26700 }, { "epoch": 26.73, "grad_norm": 105.67713928222656, "learning_rate": 9.236722193871252e-06, "loss": 0.4581, "step": 26730 }, { "epoch": 26.76, "grad_norm": 44.367271423339844, "learning_rate": 9.235053034726261e-06, "loss": 0.4469, "step": 26760 }, { "epoch": 26.79, "grad_norm": 35.41849136352539, "learning_rate": 9.233382203649402e-06, "loss": 0.4044, "step": 26790 }, { "epoch": 26.82, "grad_norm": 33.20879364013672, "learning_rate": 9.231709701300293e-06, "loss": 0.4832, "step": 26820 }, { "epoch": 26.85, "grad_norm": 57.84268569946289, "learning_rate": 9.230035528339212e-06, "loss": 0.4299, "step": 26850 }, { "epoch": 26.88, "grad_norm": 62.57261657714844, "learning_rate": 9.228359685427095e-06, "loss": 0.467, "step": 26880 }, { "epoch": 26.91, "grad_norm": 74.15518188476562, "learning_rate": 9.226682173225537e-06, "loss": 0.4967, "step": 26910 }, { "epoch": 26.94, "grad_norm": 74.8164291381836, "learning_rate": 9.225002992396797e-06, "loss": 0.4975, "step": 26940 }, { "epoch": 26.97, "grad_norm": 31.033750534057617, "learning_rate": 9.223322143603786e-06, "loss": 0.4925, "step": 26970 }, { "epoch": 27.0, "grad_norm": 0.00033921428257599473, "learning_rate": 9.221639627510076e-06, "loss": 0.5105, "step": 27000 }, { "epoch": 27.0, "eval_loss": 0.367226243019104, "eval_map": 0.7439, "eval_map_50": 0.9498, "eval_map_75": 0.8953, "eval_map_chicken": 0.7272, "eval_map_duck": 0.7013, "eval_map_large": 0.7565, "eval_map_medium": 0.7226, "eval_map_plant": 0.8031, "eval_map_small": 0.1973, "eval_mar_1": 0.3107, "eval_mar_10": 0.7836, "eval_mar_100": 0.7882, "eval_mar_100_chicken": 0.7801, "eval_mar_100_duck": 0.7397, "eval_mar_100_plant": 0.845, "eval_mar_large": 0.7989, "eval_mar_medium": 0.7824, "eval_mar_small": 0.3234, "eval_runtime": 12.4473, "eval_samples_per_second": 16.068, "eval_steps_per_second": 2.008, "step": 27000 }, { "epoch": 27.03, "grad_norm": 34.45372772216797, "learning_rate": 9.2199554447799e-06, "loss": 0.5367, "step": 27030 }, { "epoch": 27.06, "grad_norm": 115.48345184326172, "learning_rate": 9.218269596078145e-06, "loss": 0.4506, "step": 27060 }, { "epoch": 27.09, "grad_norm": 54.996482849121094, "learning_rate": 9.216582082070359e-06, "loss": 0.4722, "step": 27090 }, { "epoch": 27.12, "grad_norm": 72.86119842529297, "learning_rate": 9.214892903422745e-06, "loss": 0.4052, "step": 27120 }, { "epoch": 27.15, "grad_norm": 51.30979537963867, "learning_rate": 9.213202060802162e-06, "loss": 0.5695, "step": 27150 }, { "epoch": 27.18, "grad_norm": 90.46855163574219, "learning_rate": 9.21150955487613e-06, "loss": 0.3896, "step": 27180 }, { "epoch": 27.21, "grad_norm": 31.51398468017578, "learning_rate": 9.209815386312824e-06, "loss": 0.4929, "step": 27210 }, { "epoch": 27.24, "grad_norm": 28.97296142578125, "learning_rate": 9.208119555781074e-06, "loss": 0.4199, "step": 27240 }, { "epoch": 27.27, "grad_norm": 0.0006703261169604957, "learning_rate": 9.206422063950368e-06, "loss": 0.394, "step": 27270 }, { "epoch": 27.3, "grad_norm": 56.840579986572266, "learning_rate": 9.204722911490847e-06, "loss": 0.4917, "step": 27300 }, { "epoch": 27.33, "grad_norm": 40.93439865112305, "learning_rate": 9.20302209907331e-06, "loss": 0.5197, "step": 27330 }, { "epoch": 27.36, "grad_norm": 52.597618103027344, "learning_rate": 9.201319627369211e-06, "loss": 0.5255, "step": 27360 }, { "epoch": 27.39, "grad_norm": 122.02558135986328, "learning_rate": 9.19961549705066e-06, "loss": 0.5003, "step": 27390 }, { "epoch": 27.42, "grad_norm": 0.014937914907932281, "learning_rate": 9.19790970879042e-06, "loss": 0.4537, "step": 27420 }, { "epoch": 27.45, "grad_norm": 111.03565979003906, "learning_rate": 9.196202263261908e-06, "loss": 0.3738, "step": 27450 }, { "epoch": 27.48, "grad_norm": 32.892765045166016, "learning_rate": 9.1944931611392e-06, "loss": 0.3662, "step": 27480 }, { "epoch": 27.51, "grad_norm": 70.02935028076172, "learning_rate": 9.192782403097018e-06, "loss": 0.5076, "step": 27510 }, { "epoch": 27.54, "grad_norm": 59.52092361450195, "learning_rate": 9.191069989810743e-06, "loss": 0.4531, "step": 27540 }, { "epoch": 27.57, "grad_norm": 74.52774810791016, "learning_rate": 9.189355921956414e-06, "loss": 0.5125, "step": 27570 }, { "epoch": 27.6, "grad_norm": 68.9722900390625, "learning_rate": 9.18764020021071e-06, "loss": 0.4814, "step": 27600 }, { "epoch": 27.63, "grad_norm": 55.58739471435547, "learning_rate": 9.185922825250975e-06, "loss": 0.4055, "step": 27630 }, { "epoch": 27.66, "grad_norm": 41.6824951171875, "learning_rate": 9.1842037977552e-06, "loss": 0.4753, "step": 27660 }, { "epoch": 27.69, "grad_norm": 77.91014862060547, "learning_rate": 9.182483118402032e-06, "loss": 0.4359, "step": 27690 }, { "epoch": 27.72, "grad_norm": 38.52894592285156, "learning_rate": 9.180760787870766e-06, "loss": 0.4065, "step": 27720 }, { "epoch": 27.75, "grad_norm": 63.405025482177734, "learning_rate": 9.179036806841352e-06, "loss": 0.5039, "step": 27750 }, { "epoch": 27.78, "grad_norm": 47.525516510009766, "learning_rate": 9.17731117599439e-06, "loss": 0.4278, "step": 27780 }, { "epoch": 27.81, "grad_norm": 71.29277801513672, "learning_rate": 9.17558389601113e-06, "loss": 0.4665, "step": 27810 }, { "epoch": 27.84, "grad_norm": 133.62213134765625, "learning_rate": 9.173854967573479e-06, "loss": 0.4774, "step": 27840 }, { "epoch": 27.87, "grad_norm": 44.14418411254883, "learning_rate": 9.172124391363986e-06, "loss": 0.4872, "step": 27870 }, { "epoch": 27.9, "grad_norm": 84.29558563232422, "learning_rate": 9.170392168065858e-06, "loss": 0.4098, "step": 27900 }, { "epoch": 27.93, "grad_norm": 32.416412353515625, "learning_rate": 9.168658298362947e-06, "loss": 0.4149, "step": 27930 }, { "epoch": 27.96, "grad_norm": 43.40243911743164, "learning_rate": 9.166922782939759e-06, "loss": 0.4891, "step": 27960 }, { "epoch": 27.99, "grad_norm": 57.54479217529297, "learning_rate": 9.165185622481447e-06, "loss": 0.4635, "step": 27990 }, { "epoch": 28.0, "eval_loss": 0.3587941825389862, "eval_map": 0.7511, "eval_map_50": 0.9547, "eval_map_75": 0.8886, "eval_map_chicken": 0.7263, "eval_map_duck": 0.7119, "eval_map_large": 0.7617, "eval_map_medium": 0.7298, "eval_map_plant": 0.815, "eval_map_small": 0.246, "eval_mar_1": 0.3085, "eval_mar_10": 0.7913, "eval_mar_100": 0.7971, "eval_mar_100_chicken": 0.7821, "eval_mar_100_duck": 0.7521, "eval_mar_100_plant": 0.8571, "eval_mar_large": 0.8112, "eval_mar_medium": 0.7844, "eval_mar_small": 0.3761, "eval_runtime": 12.5746, "eval_samples_per_second": 15.905, "eval_steps_per_second": 1.988, "step": 28000 }, { "epoch": 28.02, "grad_norm": 236.17120361328125, "learning_rate": 9.163446817673817e-06, "loss": 0.4908, "step": 28020 }, { "epoch": 28.05, "grad_norm": 39.86211395263672, "learning_rate": 9.161706369203319e-06, "loss": 0.482, "step": 28050 }, { "epoch": 28.08, "grad_norm": 29.793560028076172, "learning_rate": 9.159964277757054e-06, "loss": 0.4143, "step": 28080 }, { "epoch": 28.11, "grad_norm": 73.4781723022461, "learning_rate": 9.158220544022773e-06, "loss": 0.5275, "step": 28110 }, { "epoch": 28.14, "grad_norm": 31.252424240112305, "learning_rate": 9.156475168688876e-06, "loss": 0.4921, "step": 28140 }, { "epoch": 28.17, "grad_norm": 40.534385681152344, "learning_rate": 9.154728152444408e-06, "loss": 0.4292, "step": 28170 }, { "epoch": 28.2, "grad_norm": 50.19342803955078, "learning_rate": 9.152979495979064e-06, "loss": 0.4407, "step": 28200 }, { "epoch": 28.23, "grad_norm": 90.49031829833984, "learning_rate": 9.151229199983185e-06, "loss": 0.4921, "step": 28230 }, { "epoch": 28.26, "grad_norm": 53.73204803466797, "learning_rate": 9.149477265147762e-06, "loss": 0.4671, "step": 28260 }, { "epoch": 28.29, "grad_norm": 75.43426513671875, "learning_rate": 9.147723692164428e-06, "loss": 0.5038, "step": 28290 }, { "epoch": 28.32, "grad_norm": 53.25422286987305, "learning_rate": 9.145968481725466e-06, "loss": 0.4633, "step": 28320 }, { "epoch": 28.35, "grad_norm": 44.48306655883789, "learning_rate": 9.14421163452381e-06, "loss": 0.4647, "step": 28350 }, { "epoch": 28.38, "grad_norm": 27.28011131286621, "learning_rate": 9.142453151253033e-06, "loss": 0.4377, "step": 28380 }, { "epoch": 28.41, "grad_norm": 154.42599487304688, "learning_rate": 9.140693032607353e-06, "loss": 0.5738, "step": 28410 }, { "epoch": 28.44, "grad_norm": 56.63320541381836, "learning_rate": 9.13893127928164e-06, "loss": 0.5639, "step": 28440 }, { "epoch": 28.47, "grad_norm": 68.56517791748047, "learning_rate": 9.137167891971407e-06, "loss": 0.4477, "step": 28470 }, { "epoch": 28.5, "grad_norm": 158.8979949951172, "learning_rate": 9.13540287137281e-06, "loss": 0.4359, "step": 28500 }, { "epoch": 28.53, "grad_norm": 32.45115661621094, "learning_rate": 9.133636218182652e-06, "loss": 0.5162, "step": 28530 }, { "epoch": 28.56, "grad_norm": 48.753231048583984, "learning_rate": 9.131867933098379e-06, "loss": 0.5837, "step": 28560 }, { "epoch": 28.59, "grad_norm": 68.9449691772461, "learning_rate": 9.13009801681808e-06, "loss": 0.3924, "step": 28590 }, { "epoch": 28.62, "grad_norm": 44.372074127197266, "learning_rate": 9.128326470040495e-06, "loss": 0.5523, "step": 28620 }, { "epoch": 28.65, "grad_norm": 42.07314682006836, "learning_rate": 9.126553293465e-06, "loss": 0.434, "step": 28650 }, { "epoch": 28.68, "grad_norm": 0.0005659362068399787, "learning_rate": 9.124778487791615e-06, "loss": 0.5093, "step": 28680 }, { "epoch": 28.71, "grad_norm": 39.70720291137695, "learning_rate": 9.123002053721006e-06, "loss": 0.6169, "step": 28710 }, { "epoch": 28.74, "grad_norm": 113.64903259277344, "learning_rate": 9.121223991954484e-06, "loss": 0.4518, "step": 28740 }, { "epoch": 28.77, "grad_norm": 62.51133728027344, "learning_rate": 9.119444303193997e-06, "loss": 0.3704, "step": 28770 }, { "epoch": 28.8, "grad_norm": 68.49651336669922, "learning_rate": 9.117662988142138e-06, "loss": 0.6072, "step": 28800 }, { "epoch": 28.83, "grad_norm": 228.11502075195312, "learning_rate": 9.115880047502142e-06, "loss": 0.648, "step": 28830 }, { "epoch": 28.86, "grad_norm": 34.14138412475586, "learning_rate": 9.114095481977887e-06, "loss": 0.5198, "step": 28860 }, { "epoch": 28.89, "grad_norm": 99.3452377319336, "learning_rate": 9.112309292273891e-06, "loss": 0.4729, "step": 28890 }, { "epoch": 28.92, "grad_norm": 57.397457122802734, "learning_rate": 9.110521479095314e-06, "loss": 0.4897, "step": 28920 }, { "epoch": 28.95, "grad_norm": 232.47564697265625, "learning_rate": 9.108732043147952e-06, "loss": 0.5144, "step": 28950 }, { "epoch": 28.98, "grad_norm": 59.73653030395508, "learning_rate": 9.106940985138251e-06, "loss": 0.5447, "step": 28980 }, { "epoch": 29.0, "eval_loss": 0.4269033670425415, "eval_map": 0.7002, "eval_map_50": 0.9507, "eval_map_75": 0.8595, "eval_map_chicken": 0.6529, "eval_map_duck": 0.6597, "eval_map_large": 0.7176, "eval_map_medium": 0.6811, "eval_map_plant": 0.7882, "eval_map_small": 0.2218, "eval_mar_1": 0.2901, "eval_mar_10": 0.7457, "eval_mar_100": 0.75, "eval_mar_100_chicken": 0.7137, "eval_mar_100_duck": 0.7052, "eval_mar_100_plant": 0.8313, "eval_mar_large": 0.7706, "eval_mar_medium": 0.7405, "eval_mar_small": 0.3412, "eval_runtime": 13.092, "eval_samples_per_second": 15.276, "eval_steps_per_second": 1.91, "step": 29000 }, { "epoch": 29.01, "grad_norm": 44.138431549072266, "learning_rate": 9.10514830577329e-06, "loss": 0.468, "step": 29010 }, { "epoch": 29.04, "grad_norm": 70.59234619140625, "learning_rate": 9.10335400576079e-06, "loss": 0.4689, "step": 29040 }, { "epoch": 29.07, "grad_norm": 58.2227783203125, "learning_rate": 9.101558085809114e-06, "loss": 0.4376, "step": 29070 }, { "epoch": 29.1, "grad_norm": 128.89015197753906, "learning_rate": 9.099760546627262e-06, "loss": 0.5232, "step": 29100 }, { "epoch": 29.13, "grad_norm": 3.6361216189106926e-05, "learning_rate": 9.097961388924875e-06, "loss": 0.4994, "step": 29130 }, { "epoch": 29.16, "grad_norm": 42.61015701293945, "learning_rate": 9.096160613412228e-06, "loss": 0.8201, "step": 29160 }, { "epoch": 29.19, "grad_norm": 79.19149780273438, "learning_rate": 9.094358220800243e-06, "loss": 0.4591, "step": 29190 }, { "epoch": 29.22, "grad_norm": 47.540748596191406, "learning_rate": 9.092554211800476e-06, "loss": 0.4979, "step": 29220 }, { "epoch": 29.25, "grad_norm": 116.0777359008789, "learning_rate": 9.090748587125118e-06, "loss": 0.4871, "step": 29250 }, { "epoch": 29.28, "grad_norm": 8.70522198965773e-05, "learning_rate": 9.088941347487004e-06, "loss": 0.4816, "step": 29280 }, { "epoch": 29.31, "grad_norm": 25.99987030029297, "learning_rate": 9.087132493599601e-06, "loss": 0.4042, "step": 29310 }, { "epoch": 29.34, "grad_norm": 38.19371795654297, "learning_rate": 9.085322026177017e-06, "loss": 0.5335, "step": 29340 }, { "epoch": 29.37, "grad_norm": 34.12751007080078, "learning_rate": 9.083509945933996e-06, "loss": 0.4884, "step": 29370 }, { "epoch": 29.4, "grad_norm": 54.2411003112793, "learning_rate": 9.08169625358592e-06, "loss": 0.3789, "step": 29400 }, { "epoch": 29.43, "grad_norm": 51.79536819458008, "learning_rate": 9.079880949848804e-06, "loss": 0.5517, "step": 29430 }, { "epoch": 29.46, "grad_norm": 48.688514709472656, "learning_rate": 9.078064035439303e-06, "loss": 0.4943, "step": 29460 }, { "epoch": 29.49, "grad_norm": 97.03939819335938, "learning_rate": 9.076245511074704e-06, "loss": 0.3495, "step": 29490 }, { "epoch": 29.52, "grad_norm": 107.6244888305664, "learning_rate": 9.074425377472932e-06, "loss": 0.5039, "step": 29520 }, { "epoch": 29.55, "grad_norm": 32.6119384765625, "learning_rate": 9.072603635352548e-06, "loss": 0.4587, "step": 29550 }, { "epoch": 29.58, "grad_norm": 41.78489303588867, "learning_rate": 9.070780285432746e-06, "loss": 0.3789, "step": 29580 }, { "epoch": 29.61, "grad_norm": 0.00019949959823861718, "learning_rate": 9.068955328433355e-06, "loss": 0.5173, "step": 29610 }, { "epoch": 29.64, "grad_norm": 82.86524200439453, "learning_rate": 9.067128765074842e-06, "loss": 0.4599, "step": 29640 }, { "epoch": 29.67, "grad_norm": 73.73448944091797, "learning_rate": 9.065300596078304e-06, "loss": 0.5617, "step": 29670 }, { "epoch": 29.7, "grad_norm": 76.7964096069336, "learning_rate": 9.06347082216547e-06, "loss": 0.4187, "step": 29700 }, { "epoch": 29.73, "grad_norm": 25.016586303710938, "learning_rate": 9.06163944405871e-06, "loss": 0.4355, "step": 29730 }, { "epoch": 29.76, "grad_norm": 125.9419174194336, "learning_rate": 9.059806462481022e-06, "loss": 0.4034, "step": 29760 }, { "epoch": 29.79, "grad_norm": 51.12861633300781, "learning_rate": 9.057971878156036e-06, "loss": 0.3815, "step": 29790 }, { "epoch": 29.82, "grad_norm": 30.524913787841797, "learning_rate": 9.056135691808019e-06, "loss": 0.4404, "step": 29820 }, { "epoch": 29.85, "grad_norm": 32.9085578918457, "learning_rate": 9.054297904161868e-06, "loss": 0.6188, "step": 29850 }, { "epoch": 29.88, "grad_norm": 66.11001586914062, "learning_rate": 9.052458515943112e-06, "loss": 0.4827, "step": 29880 }, { "epoch": 29.91, "grad_norm": 52.857879638671875, "learning_rate": 9.050617527877911e-06, "loss": 0.5276, "step": 29910 }, { "epoch": 29.94, "grad_norm": 46.174007415771484, "learning_rate": 9.048774940693062e-06, "loss": 0.4573, "step": 29940 }, { "epoch": 29.97, "grad_norm": 72.82232666015625, "learning_rate": 9.046930755115986e-06, "loss": 0.6887, "step": 29970 }, { "epoch": 30.0, "grad_norm": 63.441287994384766, "learning_rate": 9.045084971874738e-06, "loss": 0.3582, "step": 30000 }, { "epoch": 30.0, "eval_loss": 0.35819748044013977, "eval_map": 0.7525, "eval_map_50": 0.9477, "eval_map_75": 0.8899, "eval_map_chicken": 0.7305, "eval_map_duck": 0.7022, "eval_map_large": 0.7843, "eval_map_medium": 0.7345, "eval_map_plant": 0.8248, "eval_map_small": 0.1913, "eval_mar_1": 0.3133, "eval_mar_10": 0.791, "eval_mar_100": 0.7963, "eval_mar_100_chicken": 0.7865, "eval_mar_100_duck": 0.7356, "eval_mar_100_plant": 0.8668, "eval_mar_large": 0.8261, "eval_mar_medium": 0.785, "eval_mar_small": 0.3302, "eval_runtime": 12.7509, "eval_samples_per_second": 15.685, "eval_steps_per_second": 1.961, "step": 30000 }, { "epoch": 30.03, "grad_norm": 51.757469177246094, "learning_rate": 9.043237591698006e-06, "loss": 0.4329, "step": 30030 }, { "epoch": 30.06, "grad_norm": 227.31675720214844, "learning_rate": 9.041388615315102e-06, "loss": 0.566, "step": 30060 }, { "epoch": 30.09, "grad_norm": 32.643550872802734, "learning_rate": 9.03953804345598e-06, "loss": 0.4602, "step": 30090 }, { "epoch": 30.12, "grad_norm": 67.18914031982422, "learning_rate": 9.037685876851211e-06, "loss": 0.5095, "step": 30120 }, { "epoch": 30.15, "grad_norm": 97.85009765625, "learning_rate": 9.035832116232002e-06, "loss": 0.4364, "step": 30150 }, { "epoch": 30.18, "grad_norm": 49.86127471923828, "learning_rate": 9.033976762330189e-06, "loss": 0.4547, "step": 30180 }, { "epoch": 30.21, "grad_norm": 763.9534912109375, "learning_rate": 9.032119815878237e-06, "loss": 0.4889, "step": 30210 }, { "epoch": 30.24, "grad_norm": 79.0152587890625, "learning_rate": 9.030261277609235e-06, "loss": 0.5753, "step": 30240 }, { "epoch": 30.27, "grad_norm": 71.2388687133789, "learning_rate": 9.028401148256911e-06, "loss": 0.4247, "step": 30270 }, { "epoch": 30.3, "grad_norm": 65.65898132324219, "learning_rate": 9.026539428555609e-06, "loss": 0.4554, "step": 30300 }, { "epoch": 30.33, "grad_norm": 48.988216400146484, "learning_rate": 9.024676119240312e-06, "loss": 0.5226, "step": 30330 }, { "epoch": 30.36, "grad_norm": 49.586158752441406, "learning_rate": 9.022811221046618e-06, "loss": 0.4947, "step": 30360 }, { "epoch": 30.39, "grad_norm": 49.75828170776367, "learning_rate": 9.020944734710767e-06, "loss": 0.4833, "step": 30390 }, { "epoch": 30.42, "grad_norm": 76.17473602294922, "learning_rate": 9.01907666096961e-06, "loss": 0.4402, "step": 30420 }, { "epoch": 30.45, "grad_norm": 36.23179244995117, "learning_rate": 9.017207000560639e-06, "loss": 0.4724, "step": 30450 }, { "epoch": 30.48, "grad_norm": 59.29749298095703, "learning_rate": 9.015335754221964e-06, "loss": 0.4982, "step": 30480 }, { "epoch": 30.51, "grad_norm": 82.9033432006836, "learning_rate": 9.013462922692324e-06, "loss": 0.3044, "step": 30510 }, { "epoch": 30.54, "grad_norm": 53.0378303527832, "learning_rate": 9.011588506711085e-06, "loss": 0.46, "step": 30540 }, { "epoch": 30.57, "grad_norm": 46.807655334472656, "learning_rate": 9.00971250701823e-06, "loss": 0.4521, "step": 30570 }, { "epoch": 30.6, "grad_norm": 53.08588409423828, "learning_rate": 9.007834924354384e-06, "loss": 0.4823, "step": 30600 }, { "epoch": 30.63, "grad_norm": 50.80430221557617, "learning_rate": 9.005955759460779e-06, "loss": 0.4341, "step": 30630 }, { "epoch": 30.66, "grad_norm": 61.088809967041016, "learning_rate": 9.004075013079284e-06, "loss": 0.4077, "step": 30660 }, { "epoch": 30.69, "grad_norm": 61.00802993774414, "learning_rate": 9.002192685952385e-06, "loss": 0.4194, "step": 30690 }, { "epoch": 30.72, "grad_norm": 42.223358154296875, "learning_rate": 9.000308778823196e-06, "loss": 0.503, "step": 30720 }, { "epoch": 30.75, "grad_norm": 51.337669372558594, "learning_rate": 8.998423292435455e-06, "loss": 0.4518, "step": 30750 }, { "epoch": 30.78, "grad_norm": 53.67788314819336, "learning_rate": 8.996536227533519e-06, "loss": 0.4601, "step": 30780 }, { "epoch": 30.81, "grad_norm": 50.15966796875, "learning_rate": 8.994647584862374e-06, "loss": 0.4321, "step": 30810 }, { "epoch": 30.84, "grad_norm": 127.74644470214844, "learning_rate": 8.992757365167625e-06, "loss": 0.4367, "step": 30840 }, { "epoch": 30.87, "grad_norm": 36.4520149230957, "learning_rate": 8.990865569195502e-06, "loss": 0.5023, "step": 30870 }, { "epoch": 30.9, "grad_norm": 67.84706115722656, "learning_rate": 8.988972197692857e-06, "loss": 0.3925, "step": 30900 }, { "epoch": 30.93, "grad_norm": 67.85069274902344, "learning_rate": 8.987077251407159e-06, "loss": 0.4208, "step": 30930 }, { "epoch": 30.96, "grad_norm": 34.42999267578125, "learning_rate": 8.985180731086505e-06, "loss": 0.4051, "step": 30960 }, { "epoch": 30.99, "grad_norm": 42.03893280029297, "learning_rate": 8.983282637479613e-06, "loss": 0.5428, "step": 30990 }, { "epoch": 31.0, "eval_loss": 0.3838879466056824, "eval_map": 0.7306, "eval_map_50": 0.9549, "eval_map_75": 0.8861, "eval_map_chicken": 0.7048, "eval_map_duck": 0.6953, "eval_map_large": 0.7562, "eval_map_medium": 0.707, "eval_map_plant": 0.7918, "eval_map_small": 0.157, "eval_mar_1": 0.3038, "eval_mar_10": 0.772, "eval_mar_100": 0.7767, "eval_mar_100_chicken": 0.766, "eval_mar_100_duck": 0.7356, "eval_mar_100_plant": 0.8285, "eval_mar_large": 0.8003, "eval_mar_medium": 0.7621, "eval_mar_small": 0.291, "eval_runtime": 12.6246, "eval_samples_per_second": 15.842, "eval_steps_per_second": 1.98, "step": 31000 }, { "epoch": 31.02, "grad_norm": 0.001372757600620389, "learning_rate": 8.98138297133582e-06, "loss": 0.4556, "step": 31020 }, { "epoch": 31.05, "grad_norm": 64.68170166015625, "learning_rate": 8.97948173340508e-06, "loss": 0.4508, "step": 31050 }, { "epoch": 31.08, "grad_norm": 67.12383270263672, "learning_rate": 8.977578924437976e-06, "loss": 0.3878, "step": 31080 }, { "epoch": 31.11, "grad_norm": 61.60666275024414, "learning_rate": 8.975674545185704e-06, "loss": 0.3965, "step": 31110 }, { "epoch": 31.14, "grad_norm": 37.45740509033203, "learning_rate": 8.973768596400085e-06, "loss": 0.4422, "step": 31140 }, { "epoch": 31.17, "grad_norm": 48.09686279296875, "learning_rate": 8.971861078833558e-06, "loss": 0.4324, "step": 31170 }, { "epoch": 31.2, "grad_norm": 114.8328628540039, "learning_rate": 8.969951993239177e-06, "loss": 0.41, "step": 31200 }, { "epoch": 31.23, "grad_norm": 224.55416870117188, "learning_rate": 8.968041340370622e-06, "loss": 0.4437, "step": 31230 }, { "epoch": 31.26, "grad_norm": 41.15201187133789, "learning_rate": 8.966129120982188e-06, "loss": 0.4884, "step": 31260 }, { "epoch": 31.29, "grad_norm": 66.560791015625, "learning_rate": 8.964215335828788e-06, "loss": 0.451, "step": 31290 }, { "epoch": 31.32, "grad_norm": 54.20158767700195, "learning_rate": 8.962299985665955e-06, "loss": 0.5068, "step": 31320 }, { "epoch": 31.35, "grad_norm": 64.05374145507812, "learning_rate": 8.960383071249837e-06, "loss": 0.495, "step": 31350 }, { "epoch": 31.38, "grad_norm": 131.33737182617188, "learning_rate": 8.958464593337202e-06, "loss": 0.4397, "step": 31380 }, { "epoch": 31.41, "grad_norm": 56.03692626953125, "learning_rate": 8.956544552685437e-06, "loss": 0.4038, "step": 31410 }, { "epoch": 31.44, "grad_norm": 62.55326461791992, "learning_rate": 8.954622950052543e-06, "loss": 0.4467, "step": 31440 }, { "epoch": 31.47, "grad_norm": 59.9977912902832, "learning_rate": 8.952699786197137e-06, "loss": 0.4858, "step": 31470 }, { "epoch": 31.5, "grad_norm": 63.01356887817383, "learning_rate": 8.950775061878453e-06, "loss": 0.4061, "step": 31500 }, { "epoch": 31.53, "grad_norm": 157.830810546875, "learning_rate": 8.948848777856342e-06, "loss": 0.4295, "step": 31530 }, { "epoch": 31.56, "grad_norm": 0.00015298137441277504, "learning_rate": 8.946920934891274e-06, "loss": 0.4347, "step": 31560 }, { "epoch": 31.59, "grad_norm": 53.64508056640625, "learning_rate": 8.944991533744327e-06, "loss": 0.4837, "step": 31590 }, { "epoch": 31.62, "grad_norm": 45.04652404785156, "learning_rate": 8.943060575177197e-06, "loss": 0.5713, "step": 31620 }, { "epoch": 31.65, "grad_norm": 51.152130126953125, "learning_rate": 8.9411280599522e-06, "loss": 0.4605, "step": 31650 }, { "epoch": 31.68, "grad_norm": 64.16661834716797, "learning_rate": 8.939193988832261e-06, "loss": 0.4604, "step": 31680 }, { "epoch": 31.71, "grad_norm": 56.395233154296875, "learning_rate": 8.937258362580918e-06, "loss": 0.4298, "step": 31710 }, { "epoch": 31.74, "grad_norm": 70.98343658447266, "learning_rate": 8.93532118196233e-06, "loss": 0.4391, "step": 31740 }, { "epoch": 31.77, "grad_norm": 111.21575164794922, "learning_rate": 8.93338244774126e-06, "loss": 0.4608, "step": 31770 }, { "epoch": 31.8, "grad_norm": 72.5532455444336, "learning_rate": 8.931442160683094e-06, "loss": 0.4862, "step": 31800 }, { "epoch": 31.83, "grad_norm": 82.3051986694336, "learning_rate": 8.929500321553825e-06, "loss": 0.453, "step": 31830 }, { "epoch": 31.86, "grad_norm": 101.05518341064453, "learning_rate": 8.92755693112006e-06, "loss": 0.4684, "step": 31860 }, { "epoch": 31.89, "grad_norm": 37.76423263549805, "learning_rate": 8.925611990149021e-06, "loss": 0.4367, "step": 31890 }, { "epoch": 31.92, "grad_norm": 39.29276657104492, "learning_rate": 8.923665499408535e-06, "loss": 0.5263, "step": 31920 }, { "epoch": 31.95, "grad_norm": 43.977272033691406, "learning_rate": 8.921717459667052e-06, "loss": 0.3961, "step": 31950 }, { "epoch": 31.98, "grad_norm": 40.99714660644531, "learning_rate": 8.91976787169362e-06, "loss": 0.4257, "step": 31980 }, { "epoch": 32.0, "eval_loss": 0.3435327410697937, "eval_map": 0.7691, "eval_map_50": 0.9579, "eval_map_75": 0.8979, "eval_map_chicken": 0.7551, "eval_map_duck": 0.732, "eval_map_large": 0.7886, "eval_map_medium": 0.7552, "eval_map_plant": 0.8203, "eval_map_small": 0.1462, "eval_mar_1": 0.3172, "eval_mar_10": 0.806, "eval_mar_100": 0.8112, "eval_mar_100_chicken": 0.8058, "eval_mar_100_duck": 0.7675, "eval_mar_100_plant": 0.8601, "eval_mar_large": 0.8352, "eval_mar_medium": 0.8032, "eval_mar_small": 0.2958, "eval_runtime": 13.1551, "eval_samples_per_second": 15.203, "eval_steps_per_second": 1.9, "step": 32000 }, { "epoch": 32.01, "grad_norm": 81.77295684814453, "learning_rate": 8.917816736257912e-06, "loss": 0.4151, "step": 32010 }, { "epoch": 32.04, "grad_norm": 37.240962982177734, "learning_rate": 8.915864054130203e-06, "loss": 0.4776, "step": 32040 }, { "epoch": 32.07, "grad_norm": 41.45079803466797, "learning_rate": 8.91390982608138e-06, "loss": 0.3466, "step": 32070 }, { "epoch": 32.1, "grad_norm": 835.5145263671875, "learning_rate": 8.911954052882941e-06, "loss": 0.5107, "step": 32100 }, { "epoch": 32.13, "grad_norm": 103.15101623535156, "learning_rate": 8.909996735306996e-06, "loss": 0.4561, "step": 32130 }, { "epoch": 32.16, "grad_norm": 48.267539978027344, "learning_rate": 8.908037874126263e-06, "loss": 0.4667, "step": 32160 }, { "epoch": 32.19, "grad_norm": 113.19718170166016, "learning_rate": 8.906077470114068e-06, "loss": 0.3889, "step": 32190 }, { "epoch": 32.22, "grad_norm": 82.04329681396484, "learning_rate": 8.904115524044349e-06, "loss": 0.4498, "step": 32220 }, { "epoch": 32.25, "grad_norm": 99.47281646728516, "learning_rate": 8.902152036691649e-06, "loss": 0.4555, "step": 32250 }, { "epoch": 32.28, "grad_norm": 42.69618606567383, "learning_rate": 8.900187008831124e-06, "loss": 0.5245, "step": 32280 }, { "epoch": 32.31, "grad_norm": 0.00021078799909446388, "learning_rate": 8.898220441238534e-06, "loss": 0.4057, "step": 32310 }, { "epoch": 32.34, "grad_norm": 41.272804260253906, "learning_rate": 8.896252334690251e-06, "loss": 0.4531, "step": 32340 }, { "epoch": 32.37, "grad_norm": 71.31124114990234, "learning_rate": 8.894282689963252e-06, "loss": 0.4513, "step": 32370 }, { "epoch": 32.4, "grad_norm": 76.99893188476562, "learning_rate": 8.892311507835118e-06, "loss": 0.4338, "step": 32400 }, { "epoch": 32.43, "grad_norm": 125.3895263671875, "learning_rate": 8.890338789084043e-06, "loss": 0.4519, "step": 32430 }, { "epoch": 32.46, "grad_norm": 74.70609283447266, "learning_rate": 8.888364534488828e-06, "loss": 0.4475, "step": 32460 }, { "epoch": 32.49, "grad_norm": 36.95541000366211, "learning_rate": 8.886388744828872e-06, "loss": 0.5857, "step": 32490 }, { "epoch": 32.52, "grad_norm": 50.42551040649414, "learning_rate": 8.88441142088419e-06, "loss": 0.4205, "step": 32520 }, { "epoch": 32.55, "grad_norm": 44.49029541015625, "learning_rate": 8.882432563435394e-06, "loss": 0.3738, "step": 32550 }, { "epoch": 32.58, "grad_norm": 80.62117767333984, "learning_rate": 8.880452173263708e-06, "loss": 0.4299, "step": 32580 }, { "epoch": 32.61, "grad_norm": 37.055240631103516, "learning_rate": 8.878470251150959e-06, "loss": 0.4826, "step": 32610 }, { "epoch": 32.64, "grad_norm": 70.06743621826172, "learning_rate": 8.87648679787958e-06, "loss": 0.4241, "step": 32640 }, { "epoch": 32.67, "grad_norm": 56.593482971191406, "learning_rate": 8.874501814232603e-06, "loss": 0.3435, "step": 32670 }, { "epoch": 32.7, "grad_norm": 41.06038284301758, "learning_rate": 8.872515300993669e-06, "loss": 0.4029, "step": 32700 }, { "epoch": 32.73, "grad_norm": 36.221553802490234, "learning_rate": 8.870527258947025e-06, "loss": 0.4044, "step": 32730 }, { "epoch": 32.76, "grad_norm": 32.324520111083984, "learning_rate": 8.868537688877516e-06, "loss": 0.437, "step": 32760 }, { "epoch": 32.79, "grad_norm": 36.65751647949219, "learning_rate": 8.866546591570593e-06, "loss": 0.4263, "step": 32790 }, { "epoch": 32.82, "grad_norm": 256.79693603515625, "learning_rate": 8.86455396781231e-06, "loss": 0.474, "step": 32820 }, { "epoch": 32.85, "grad_norm": 89.47574615478516, "learning_rate": 8.862559818389322e-06, "loss": 0.5592, "step": 32850 }, { "epoch": 32.88, "grad_norm": 32.67885208129883, "learning_rate": 8.860564144088891e-06, "loss": 0.4579, "step": 32880 }, { "epoch": 32.91, "grad_norm": 996.2119750976562, "learning_rate": 8.858566945698874e-06, "loss": 0.4808, "step": 32910 }, { "epoch": 32.94, "grad_norm": 35.17996597290039, "learning_rate": 8.856568224007736e-06, "loss": 0.4624, "step": 32940 }, { "epoch": 32.97, "grad_norm": 70.29095458984375, "learning_rate": 8.854567979804538e-06, "loss": 0.3996, "step": 32970 }, { "epoch": 33.0, "grad_norm": 44.21821212768555, "learning_rate": 8.852566213878947e-06, "loss": 0.4555, "step": 33000 }, { "epoch": 33.0, "eval_loss": 0.3404276371002197, "eval_map": 0.7594, "eval_map_50": 0.9571, "eval_map_75": 0.9004, "eval_map_chicken": 0.7524, "eval_map_duck": 0.7026, "eval_map_large": 0.7696, "eval_map_medium": 0.7488, "eval_map_plant": 0.8232, "eval_map_small": 0.1775, "eval_mar_1": 0.3116, "eval_mar_10": 0.7999, "eval_mar_100": 0.8043, "eval_mar_100_chicken": 0.8014, "eval_mar_100_duck": 0.7495, "eval_mar_100_plant": 0.8622, "eval_mar_large": 0.8178, "eval_mar_medium": 0.7983, "eval_mar_small": 0.2972, "eval_runtime": 12.616, "eval_samples_per_second": 15.853, "eval_steps_per_second": 1.982, "step": 33000 }, { "epoch": 33.03, "grad_norm": 118.58769989013672, "learning_rate": 8.850562927021227e-06, "loss": 0.4165, "step": 33030 }, { "epoch": 33.06, "grad_norm": 78.10658264160156, "learning_rate": 8.848558120022246e-06, "loss": 0.4404, "step": 33060 }, { "epoch": 33.09, "grad_norm": 83.59980773925781, "learning_rate": 8.846551793673467e-06, "loss": 0.4487, "step": 33090 }, { "epoch": 33.12, "grad_norm": 84.0053939819336, "learning_rate": 8.844543948766958e-06, "loss": 0.3364, "step": 33120 }, { "epoch": 33.15, "grad_norm": 51.9716911315918, "learning_rate": 8.842534586095383e-06, "loss": 0.4759, "step": 33150 }, { "epoch": 33.18, "grad_norm": 41.72578430175781, "learning_rate": 8.84052370645201e-06, "loss": 0.4151, "step": 33180 }, { "epoch": 33.21, "grad_norm": 77.96997833251953, "learning_rate": 8.838511310630697e-06, "loss": 0.4128, "step": 33210 }, { "epoch": 33.24, "grad_norm": 63.047119140625, "learning_rate": 8.83649739942591e-06, "loss": 0.4656, "step": 33240 }, { "epoch": 33.27, "grad_norm": 41.162742614746094, "learning_rate": 8.834481973632709e-06, "loss": 0.4672, "step": 33270 }, { "epoch": 33.3, "grad_norm": 53.497711181640625, "learning_rate": 8.83246503404675e-06, "loss": 0.3522, "step": 33300 }, { "epoch": 33.33, "grad_norm": 73.18824768066406, "learning_rate": 8.83044658146429e-06, "loss": 0.4252, "step": 33330 }, { "epoch": 33.36, "grad_norm": 68.3832778930664, "learning_rate": 8.828426616682184e-06, "loss": 0.4495, "step": 33360 }, { "epoch": 33.39, "grad_norm": 44.43447494506836, "learning_rate": 8.826405140497878e-06, "loss": 0.4298, "step": 33390 }, { "epoch": 33.42, "grad_norm": 90.11807250976562, "learning_rate": 8.824382153709423e-06, "loss": 0.4423, "step": 33420 }, { "epoch": 33.45, "grad_norm": 30.87198257446289, "learning_rate": 8.82235765711546e-06, "loss": 0.5179, "step": 33450 }, { "epoch": 33.48, "grad_norm": 112.96526336669922, "learning_rate": 8.820331651515226e-06, "loss": 0.4033, "step": 33480 }, { "epoch": 33.51, "grad_norm": 31.3351993560791, "learning_rate": 8.81830413770856e-06, "loss": 0.6195, "step": 33510 }, { "epoch": 33.54, "grad_norm": 33.98473358154297, "learning_rate": 8.816275116495891e-06, "loss": 0.3446, "step": 33540 }, { "epoch": 33.57, "grad_norm": 61.6156005859375, "learning_rate": 8.814244588678245e-06, "loss": 0.5035, "step": 33570 }, { "epoch": 33.6, "grad_norm": 60.01884460449219, "learning_rate": 8.81221255505724e-06, "loss": 0.6521, "step": 33600 }, { "epoch": 33.63, "grad_norm": 26.35115623474121, "learning_rate": 8.810179016435092e-06, "loss": 0.4559, "step": 33630 }, { "epoch": 33.66, "grad_norm": 41.63239288330078, "learning_rate": 8.808143973614612e-06, "loss": 0.4143, "step": 33660 }, { "epoch": 33.69, "grad_norm": 38.926082611083984, "learning_rate": 8.806107427399198e-06, "loss": 0.4741, "step": 33690 }, { "epoch": 33.72, "grad_norm": 77.3026123046875, "learning_rate": 8.80406937859285e-06, "loss": 0.4062, "step": 33720 }, { "epoch": 33.75, "grad_norm": 76.44752502441406, "learning_rate": 8.802029828000157e-06, "loss": 0.5975, "step": 33750 }, { "epoch": 33.78, "grad_norm": 53.88447189331055, "learning_rate": 8.799988776426298e-06, "loss": 0.406, "step": 33780 }, { "epoch": 33.81, "grad_norm": 4.4060001528123394e-05, "learning_rate": 8.797946224677051e-06, "loss": 0.3486, "step": 33810 }, { "epoch": 33.84, "grad_norm": 39.02163314819336, "learning_rate": 8.795902173558784e-06, "loss": 0.4459, "step": 33840 }, { "epoch": 33.87, "grad_norm": 43.097015380859375, "learning_rate": 8.793856623878453e-06, "loss": 0.4589, "step": 33870 }, { "epoch": 33.9, "grad_norm": 64.11221313476562, "learning_rate": 8.791809576443611e-06, "loss": 0.4663, "step": 33900 }, { "epoch": 33.93, "grad_norm": 47.584049224853516, "learning_rate": 8.789761032062398e-06, "loss": 0.527, "step": 33930 }, { "epoch": 33.96, "grad_norm": 85.9797592163086, "learning_rate": 8.787710991543547e-06, "loss": 0.5751, "step": 33960 }, { "epoch": 33.99, "grad_norm": 0.000363035942427814, "learning_rate": 8.785659455696384e-06, "loss": 0.4041, "step": 33990 }, { "epoch": 34.0, "eval_loss": 0.3375667631626129, "eval_map": 0.7636, "eval_map_50": 0.9597, "eval_map_75": 0.8954, "eval_map_chicken": 0.754, "eval_map_duck": 0.7136, "eval_map_large": 0.7815, "eval_map_medium": 0.7438, "eval_map_plant": 0.8233, "eval_map_small": 0.2114, "eval_mar_1": 0.3155, "eval_mar_10": 0.8009, "eval_mar_100": 0.8066, "eval_mar_100_chicken": 0.8028, "eval_mar_100_duck": 0.7562, "eval_mar_100_plant": 0.8608, "eval_mar_large": 0.8226, "eval_mar_medium": 0.7991, "eval_mar_small": 0.335, "eval_runtime": 12.8057, "eval_samples_per_second": 15.618, "eval_steps_per_second": 1.952, "step": 34000 }, { "epoch": 34.02, "grad_norm": 203.5898895263672, "learning_rate": 8.78360642533082e-06, "loss": 0.4899, "step": 34020 }, { "epoch": 34.05, "grad_norm": 49.63528060913086, "learning_rate": 8.78155190125736e-06, "loss": 0.5247, "step": 34050 }, { "epoch": 34.08, "grad_norm": 174.47315979003906, "learning_rate": 8.779495884287099e-06, "loss": 0.4719, "step": 34080 }, { "epoch": 34.11, "grad_norm": 39.221900939941406, "learning_rate": 8.777438375231717e-06, "loss": 0.3763, "step": 34110 }, { "epoch": 34.14, "grad_norm": 59.68625259399414, "learning_rate": 8.775379374903487e-06, "loss": 0.4002, "step": 34140 }, { "epoch": 34.17, "grad_norm": 44.460899353027344, "learning_rate": 8.773318884115273e-06, "loss": 0.3588, "step": 34170 }, { "epoch": 34.2, "grad_norm": 49.74613952636719, "learning_rate": 8.77125690368052e-06, "loss": 0.5705, "step": 34200 }, { "epoch": 34.23, "grad_norm": 31.476102828979492, "learning_rate": 8.769193434413266e-06, "loss": 0.3744, "step": 34230 }, { "epoch": 34.26, "grad_norm": 54.418121337890625, "learning_rate": 8.767128477128138e-06, "loss": 0.4134, "step": 34260 }, { "epoch": 34.29, "grad_norm": 28.05330467224121, "learning_rate": 8.765062032640346e-06, "loss": 0.5155, "step": 34290 }, { "epoch": 34.32, "grad_norm": 68.76082611083984, "learning_rate": 8.76299410176569e-06, "loss": 0.4405, "step": 34320 }, { "epoch": 34.35, "grad_norm": 44.06267166137695, "learning_rate": 8.760924685320558e-06, "loss": 0.3821, "step": 34350 }, { "epoch": 34.38, "grad_norm": 57.69425582885742, "learning_rate": 8.758853784121921e-06, "loss": 0.4244, "step": 34380 }, { "epoch": 34.41, "grad_norm": 44.65262222290039, "learning_rate": 8.75678139898734e-06, "loss": 0.4364, "step": 34410 }, { "epoch": 34.44, "grad_norm": 35.44918441772461, "learning_rate": 8.754707530734958e-06, "loss": 0.531, "step": 34440 }, { "epoch": 34.47, "grad_norm": 79.0035629272461, "learning_rate": 8.752632180183504e-06, "loss": 0.4243, "step": 34470 }, { "epoch": 34.5, "grad_norm": 54.96945571899414, "learning_rate": 8.750555348152299e-06, "loss": 0.6707, "step": 34500 }, { "epoch": 34.53, "grad_norm": 126.0721664428711, "learning_rate": 8.748477035461237e-06, "loss": 0.4323, "step": 34530 }, { "epoch": 34.56, "grad_norm": 55.744773864746094, "learning_rate": 8.74639724293081e-06, "loss": 0.4942, "step": 34560 }, { "epoch": 34.59, "grad_norm": 101.10917663574219, "learning_rate": 8.744315971382078e-06, "loss": 0.5231, "step": 34590 }, { "epoch": 34.62, "grad_norm": 104.34576416015625, "learning_rate": 8.7422332216367e-06, "loss": 0.4857, "step": 34620 }, { "epoch": 34.65, "grad_norm": 36.7725715637207, "learning_rate": 8.740148994516912e-06, "loss": 0.4883, "step": 34650 }, { "epoch": 34.68, "grad_norm": 43.39622116088867, "learning_rate": 8.738063290845536e-06, "loss": 0.4842, "step": 34680 }, { "epoch": 34.71, "grad_norm": 223.58309936523438, "learning_rate": 8.73597611144597e-06, "loss": 0.4639, "step": 34710 }, { "epoch": 34.74, "grad_norm": 71.21248626708984, "learning_rate": 8.733887457142202e-06, "loss": 0.3761, "step": 34740 }, { "epoch": 34.77, "grad_norm": 32.202754974365234, "learning_rate": 8.7317973287588e-06, "loss": 0.3596, "step": 34770 }, { "epoch": 34.8, "grad_norm": 79.22689819335938, "learning_rate": 8.729705727120911e-06, "loss": 0.3913, "step": 34800 }, { "epoch": 34.83, "grad_norm": 76.49684143066406, "learning_rate": 8.72761265305427e-06, "loss": 0.4527, "step": 34830 }, { "epoch": 34.86, "grad_norm": 34.226829528808594, "learning_rate": 8.725518107385188e-06, "loss": 0.3719, "step": 34860 }, { "epoch": 34.89, "grad_norm": 34.813053131103516, "learning_rate": 8.723422090940556e-06, "loss": 0.4554, "step": 34890 }, { "epoch": 34.92, "grad_norm": 34.034942626953125, "learning_rate": 8.721324604547851e-06, "loss": 0.3905, "step": 34920 }, { "epoch": 34.95, "grad_norm": 62.88547897338867, "learning_rate": 8.719225649035126e-06, "loss": 0.4834, "step": 34950 }, { "epoch": 34.98, "grad_norm": 35.46959686279297, "learning_rate": 8.717125225231018e-06, "loss": 0.4214, "step": 34980 }, { "epoch": 35.0, "eval_loss": 0.3321477472782135, "eval_map": 0.763, "eval_map_50": 0.9569, "eval_map_75": 0.9045, "eval_map_chicken": 0.7431, "eval_map_duck": 0.7167, "eval_map_large": 0.7754, "eval_map_medium": 0.7405, "eval_map_plant": 0.8291, "eval_map_small": 0.2104, "eval_mar_1": 0.3153, "eval_mar_10": 0.8026, "eval_mar_100": 0.8058, "eval_mar_100_chicken": 0.8006, "eval_mar_100_duck": 0.7521, "eval_mar_100_plant": 0.8648, "eval_mar_large": 0.8216, "eval_mar_medium": 0.7932, "eval_mar_small": 0.3039, "eval_runtime": 13.0634, "eval_samples_per_second": 15.31, "eval_steps_per_second": 1.914, "step": 35000 }, { "epoch": 35.01, "grad_norm": 37.040225982666016, "learning_rate": 8.715023333964737e-06, "loss": 0.4665, "step": 35010 }, { "epoch": 35.04, "grad_norm": 53.93110656738281, "learning_rate": 8.712919976066078e-06, "loss": 0.3998, "step": 35040 }, { "epoch": 35.07, "grad_norm": 41.54487991333008, "learning_rate": 8.710815152365416e-06, "loss": 0.4589, "step": 35070 }, { "epoch": 35.1, "grad_norm": 27.006441116333008, "learning_rate": 8.708708863693696e-06, "loss": 0.5815, "step": 35100 }, { "epoch": 35.13, "grad_norm": 67.9949951171875, "learning_rate": 8.706601110882456e-06, "loss": 0.447, "step": 35130 }, { "epoch": 35.16, "grad_norm": 31.97315216064453, "learning_rate": 8.704491894763794e-06, "loss": 0.3925, "step": 35160 }, { "epoch": 35.19, "grad_norm": 39.46290588378906, "learning_rate": 8.702381216170404e-06, "loss": 0.4678, "step": 35190 }, { "epoch": 35.22, "grad_norm": 68.44788360595703, "learning_rate": 8.700269075935542e-06, "loss": 0.3545, "step": 35220 }, { "epoch": 35.25, "grad_norm": 47.177188873291016, "learning_rate": 8.69815547489305e-06, "loss": 0.3771, "step": 35250 }, { "epoch": 35.28, "grad_norm": 29.046361923217773, "learning_rate": 8.696040413877344e-06, "loss": 0.3998, "step": 35280 }, { "epoch": 35.31, "grad_norm": 42.96931457519531, "learning_rate": 8.693923893723415e-06, "loss": 0.494, "step": 35310 }, { "epoch": 35.34, "grad_norm": 61.23044967651367, "learning_rate": 8.691805915266836e-06, "loss": 0.3537, "step": 35340 }, { "epoch": 35.37, "grad_norm": 220.34339904785156, "learning_rate": 8.689686479343747e-06, "loss": 0.4225, "step": 35370 }, { "epoch": 35.4, "grad_norm": 55.82189178466797, "learning_rate": 8.68756558679087e-06, "loss": 0.4882, "step": 35400 }, { "epoch": 35.43, "grad_norm": 48.61567687988281, "learning_rate": 8.6854432384455e-06, "loss": 0.4586, "step": 35430 }, { "epoch": 35.46, "grad_norm": 44.0639533996582, "learning_rate": 8.683319435145503e-06, "loss": 0.5185, "step": 35460 }, { "epoch": 35.49, "grad_norm": 86.15058898925781, "learning_rate": 8.681194177729328e-06, "loss": 0.3701, "step": 35490 }, { "epoch": 35.52, "grad_norm": 42.26691436767578, "learning_rate": 8.679067467035989e-06, "loss": 0.4154, "step": 35520 }, { "epoch": 35.55, "grad_norm": 64.69525909423828, "learning_rate": 8.67693930390508e-06, "loss": 0.336, "step": 35550 }, { "epoch": 35.58, "grad_norm": 168.6240997314453, "learning_rate": 8.674809689176765e-06, "loss": 0.4152, "step": 35580 }, { "epoch": 35.61, "grad_norm": 46.845272064208984, "learning_rate": 8.672678623691783e-06, "loss": 0.4795, "step": 35610 }, { "epoch": 35.64, "grad_norm": 106.617431640625, "learning_rate": 8.670546108291443e-06, "loss": 0.3504, "step": 35640 }, { "epoch": 35.67, "grad_norm": 54.91093063354492, "learning_rate": 8.66841214381763e-06, "loss": 0.3896, "step": 35670 }, { "epoch": 35.7, "grad_norm": 45.59276580810547, "learning_rate": 8.666276731112802e-06, "loss": 0.4636, "step": 35700 }, { "epoch": 35.73, "grad_norm": 111.13752746582031, "learning_rate": 8.664139871019979e-06, "loss": 0.3779, "step": 35730 }, { "epoch": 35.76, "grad_norm": 65.15214538574219, "learning_rate": 8.662001564382768e-06, "loss": 0.3865, "step": 35760 }, { "epoch": 35.79, "grad_norm": 48.30458450317383, "learning_rate": 8.65986181204533e-06, "loss": 0.3603, "step": 35790 }, { "epoch": 35.82, "grad_norm": 33.515533447265625, "learning_rate": 8.657720614852412e-06, "loss": 0.4712, "step": 35820 }, { "epoch": 35.85, "grad_norm": 90.68647003173828, "learning_rate": 8.655577973649322e-06, "loss": 0.4151, "step": 35850 }, { "epoch": 35.88, "grad_norm": 68.15245056152344, "learning_rate": 8.65343388928194e-06, "loss": 0.4721, "step": 35880 }, { "epoch": 35.91, "grad_norm": 40.42775344848633, "learning_rate": 8.651288362596719e-06, "loss": 0.4174, "step": 35910 }, { "epoch": 35.94, "grad_norm": 73.37506103515625, "learning_rate": 8.649141394440677e-06, "loss": 0.3791, "step": 35940 }, { "epoch": 35.97, "grad_norm": 61.76493835449219, "learning_rate": 8.646992985661404e-06, "loss": 0.3961, "step": 35970 }, { "epoch": 36.0, "grad_norm": 75.14311218261719, "learning_rate": 8.644843137107058e-06, "loss": 0.4448, "step": 36000 }, { "epoch": 36.0, "eval_loss": 0.3370722830295563, "eval_map": 0.7682, "eval_map_50": 0.9618, "eval_map_75": 0.9098, "eval_map_chicken": 0.755, "eval_map_duck": 0.7253, "eval_map_large": 0.7862, "eval_map_medium": 0.7433, "eval_map_plant": 0.8242, "eval_map_small": 0.2417, "eval_mar_1": 0.3155, "eval_mar_10": 0.8044, "eval_mar_100": 0.8084, "eval_mar_100_chicken": 0.799, "eval_mar_100_duck": 0.7613, "eval_mar_100_plant": 0.8648, "eval_mar_large": 0.8259, "eval_mar_medium": 0.7944, "eval_mar_small": 0.33, "eval_runtime": 12.6209, "eval_samples_per_second": 15.847, "eval_steps_per_second": 1.981, "step": 36000 }, { "epoch": 36.03, "grad_norm": 69.73445892333984, "learning_rate": 8.642691849626364e-06, "loss": 0.4043, "step": 36030 }, { "epoch": 36.06, "grad_norm": 88.3404769897461, "learning_rate": 8.640539124068617e-06, "loss": 0.4332, "step": 36060 }, { "epoch": 36.09, "grad_norm": 77.94326782226562, "learning_rate": 8.638384961283678e-06, "loss": 0.4643, "step": 36090 }, { "epoch": 36.12, "grad_norm": 41.09723663330078, "learning_rate": 8.636229362121979e-06, "loss": 0.4723, "step": 36120 }, { "epoch": 36.15, "grad_norm": 75.74026489257812, "learning_rate": 8.634072327434515e-06, "loss": 0.4457, "step": 36150 }, { "epoch": 36.18, "grad_norm": 31.598827362060547, "learning_rate": 8.631913858072846e-06, "loss": 0.4706, "step": 36180 }, { "epoch": 36.21, "grad_norm": 96.36372375488281, "learning_rate": 8.629753954889108e-06, "loss": 0.4404, "step": 36210 }, { "epoch": 36.24, "grad_norm": 60.21940994262695, "learning_rate": 8.627592618735989e-06, "loss": 0.5303, "step": 36240 }, { "epoch": 36.27, "grad_norm": 43.889434814453125, "learning_rate": 8.625429850466756e-06, "loss": 0.5078, "step": 36270 }, { "epoch": 36.3, "grad_norm": 123.41419219970703, "learning_rate": 8.623265650935233e-06, "loss": 0.4359, "step": 36300 }, { "epoch": 36.33, "grad_norm": 154.6739501953125, "learning_rate": 8.621100020995813e-06, "loss": 0.4714, "step": 36330 }, { "epoch": 36.36, "grad_norm": 25.10385513305664, "learning_rate": 8.618932961503452e-06, "loss": 0.3829, "step": 36360 }, { "epoch": 36.39, "grad_norm": 87.33155059814453, "learning_rate": 8.616764473313671e-06, "loss": 0.3511, "step": 36390 }, { "epoch": 36.42, "grad_norm": 56.08916091918945, "learning_rate": 8.614594557282553e-06, "loss": 0.3614, "step": 36420 }, { "epoch": 36.45, "grad_norm": 98.41001892089844, "learning_rate": 8.612423214266749e-06, "loss": 0.4684, "step": 36450 }, { "epoch": 36.48, "grad_norm": 56.808162689208984, "learning_rate": 8.610250445123472e-06, "loss": 0.471, "step": 36480 }, { "epoch": 36.51, "grad_norm": 69.67655181884766, "learning_rate": 8.608076250710491e-06, "loss": 0.3452, "step": 36510 }, { "epoch": 36.54, "grad_norm": 22.386394500732422, "learning_rate": 8.605900631886148e-06, "loss": 0.4477, "step": 36540 }, { "epoch": 36.57, "grad_norm": 9.051935921888798e-05, "learning_rate": 8.603723589509342e-06, "loss": 0.4329, "step": 36570 }, { "epoch": 36.6, "grad_norm": 97.88927459716797, "learning_rate": 8.601545124439535e-06, "loss": 0.4204, "step": 36600 }, { "epoch": 36.63, "grad_norm": 40.67652130126953, "learning_rate": 8.599365237536751e-06, "loss": 0.3989, "step": 36630 }, { "epoch": 36.66, "grad_norm": 39.09264373779297, "learning_rate": 8.597183929661573e-06, "loss": 0.3991, "step": 36660 }, { "epoch": 36.69, "grad_norm": 31.05003547668457, "learning_rate": 8.595001201675149e-06, "loss": 0.41, "step": 36690 }, { "epoch": 36.72, "grad_norm": 66.82089233398438, "learning_rate": 8.592817054439184e-06, "loss": 0.4605, "step": 36720 }, { "epoch": 36.75, "grad_norm": 59.56946563720703, "learning_rate": 8.590631488815945e-06, "loss": 0.4303, "step": 36750 }, { "epoch": 36.78, "grad_norm": 46.799842834472656, "learning_rate": 8.588444505668259e-06, "loss": 0.5847, "step": 36780 }, { "epoch": 36.81, "grad_norm": 51.57160568237305, "learning_rate": 8.586256105859512e-06, "loss": 0.4351, "step": 36810 }, { "epoch": 36.84, "grad_norm": 93.68397521972656, "learning_rate": 8.584066290253649e-06, "loss": 0.3875, "step": 36840 }, { "epoch": 36.87, "grad_norm": 107.35001373291016, "learning_rate": 8.581875059715177e-06, "loss": 0.5004, "step": 36870 }, { "epoch": 36.9, "grad_norm": 47.054317474365234, "learning_rate": 8.579682415109156e-06, "loss": 0.37, "step": 36900 }, { "epoch": 36.93, "grad_norm": 64.80899810791016, "learning_rate": 8.577488357301209e-06, "loss": 0.4148, "step": 36930 }, { "epoch": 36.96, "grad_norm": 74.77336883544922, "learning_rate": 8.575292887157515e-06, "loss": 0.3825, "step": 36960 }, { "epoch": 36.99, "grad_norm": 67.75738525390625, "learning_rate": 8.573096005544812e-06, "loss": 0.4959, "step": 36990 }, { "epoch": 37.0, "eval_loss": 0.34534919261932373, "eval_map": 0.7548, "eval_map_50": 0.9564, "eval_map_75": 0.9069, "eval_map_chicken": 0.7553, "eval_map_duck": 0.701, "eval_map_large": 0.7937, "eval_map_medium": 0.7334, "eval_map_plant": 0.8082, "eval_map_small": 0.1768, "eval_mar_1": 0.3097, "eval_mar_10": 0.7944, "eval_mar_100": 0.7977, "eval_mar_100_chicken": 0.8038, "eval_mar_100_duck": 0.7371, "eval_mar_100_plant": 0.8521, "eval_mar_large": 0.8376, "eval_mar_medium": 0.7796, "eval_mar_small": 0.3251, "eval_runtime": 12.5755, "eval_samples_per_second": 15.904, "eval_steps_per_second": 1.988, "step": 37000 }, { "epoch": 37.02, "grad_norm": 65.19686889648438, "learning_rate": 8.570897713330392e-06, "loss": 0.4297, "step": 37020 }, { "epoch": 37.05, "grad_norm": 39.84136962890625, "learning_rate": 8.568698011382108e-06, "loss": 0.4558, "step": 37050 }, { "epoch": 37.08, "grad_norm": 69.7646484375, "learning_rate": 8.566496900568364e-06, "loss": 0.4416, "step": 37080 }, { "epoch": 37.11, "grad_norm": 98.10159301757812, "learning_rate": 8.564294381758128e-06, "loss": 0.4227, "step": 37110 }, { "epoch": 37.14, "grad_norm": 91.87381744384766, "learning_rate": 8.562090455820918e-06, "loss": 0.4419, "step": 37140 }, { "epoch": 37.17, "grad_norm": 62.08787536621094, "learning_rate": 8.559885123626806e-06, "loss": 0.528, "step": 37170 }, { "epoch": 37.2, "grad_norm": 46.64909362792969, "learning_rate": 8.557678386046429e-06, "loss": 0.4421, "step": 37200 }, { "epoch": 37.23, "grad_norm": 9.826561290537938e-05, "learning_rate": 8.555470243950963e-06, "loss": 0.3844, "step": 37230 }, { "epoch": 37.26, "grad_norm": 63.26577377319336, "learning_rate": 8.553260698212156e-06, "loss": 0.412, "step": 37260 }, { "epoch": 37.29, "grad_norm": 22.5672607421875, "learning_rate": 8.551049749702298e-06, "loss": 0.3715, "step": 37290 }, { "epoch": 37.32, "grad_norm": 31.38385581970215, "learning_rate": 8.548837399294235e-06, "loss": 0.4593, "step": 37320 }, { "epoch": 37.35, "grad_norm": 113.40389251708984, "learning_rate": 8.54662364786137e-06, "loss": 0.3992, "step": 37350 }, { "epoch": 37.38, "grad_norm": 62.104736328125, "learning_rate": 8.544408496277657e-06, "loss": 0.4821, "step": 37380 }, { "epoch": 37.41, "grad_norm": 50.064781188964844, "learning_rate": 8.5421919454176e-06, "loss": 0.4494, "step": 37410 }, { "epoch": 37.44, "grad_norm": 73.9268798828125, "learning_rate": 8.539973996156265e-06, "loss": 0.3912, "step": 37440 }, { "epoch": 37.47, "grad_norm": 56.994911193847656, "learning_rate": 8.537754649369256e-06, "loss": 0.4302, "step": 37470 }, { "epoch": 37.5, "grad_norm": 65.491455078125, "learning_rate": 8.535533905932739e-06, "loss": 0.4437, "step": 37500 }, { "epoch": 37.53, "grad_norm": 65.94867706298828, "learning_rate": 8.533311766723428e-06, "loss": 0.4821, "step": 37530 }, { "epoch": 37.56, "grad_norm": 46.408424377441406, "learning_rate": 8.531088232618587e-06, "loss": 0.4004, "step": 37560 }, { "epoch": 37.59, "grad_norm": 27.698884963989258, "learning_rate": 8.528863304496035e-06, "loss": 0.451, "step": 37590 }, { "epoch": 37.62, "grad_norm": 45.09851837158203, "learning_rate": 8.526636983234135e-06, "loss": 0.5114, "step": 37620 }, { "epoch": 37.65, "grad_norm": 206.72683715820312, "learning_rate": 8.524409269711808e-06, "loss": 0.4018, "step": 37650 }, { "epoch": 37.68, "grad_norm": 9.657008922658861e-05, "learning_rate": 8.522180164808515e-06, "loss": 0.4012, "step": 37680 }, { "epoch": 37.71, "grad_norm": 34.6037483215332, "learning_rate": 8.519949669404275e-06, "loss": 0.4603, "step": 37710 }, { "epoch": 37.74, "grad_norm": 61.857547760009766, "learning_rate": 8.51771778437965e-06, "loss": 0.3825, "step": 37740 }, { "epoch": 37.77, "grad_norm": 53.05194091796875, "learning_rate": 8.515484510615753e-06, "loss": 0.4395, "step": 37770 }, { "epoch": 37.8, "grad_norm": 71.83316802978516, "learning_rate": 8.513249848994248e-06, "loss": 0.58, "step": 37800 }, { "epoch": 37.83, "grad_norm": 40.781715393066406, "learning_rate": 8.511013800397338e-06, "loss": 0.4963, "step": 37830 }, { "epoch": 37.86, "grad_norm": 34.28175735473633, "learning_rate": 8.508776365707788e-06, "loss": 0.4001, "step": 37860 }, { "epoch": 37.89, "grad_norm": 91.32330322265625, "learning_rate": 8.506537545808894e-06, "loss": 0.3742, "step": 37890 }, { "epoch": 37.92, "grad_norm": 67.69334411621094, "learning_rate": 8.504297341584509e-06, "loss": 0.4763, "step": 37920 }, { "epoch": 37.95, "grad_norm": 36.653648376464844, "learning_rate": 8.502055753919033e-06, "loss": 0.4514, "step": 37950 }, { "epoch": 37.98, "grad_norm": 51.113059997558594, "learning_rate": 8.499812783697406e-06, "loss": 0.4388, "step": 37980 }, { "epoch": 38.0, "eval_loss": 0.33895206451416016, "eval_map": 0.7615, "eval_map_50": 0.961, "eval_map_75": 0.9044, "eval_map_chicken": 0.7523, "eval_map_duck": 0.7146, "eval_map_large": 0.7754, "eval_map_medium": 0.7481, "eval_map_plant": 0.8175, "eval_map_small": 0.1391, "eval_mar_1": 0.311, "eval_mar_10": 0.7998, "eval_mar_100": 0.8038, "eval_mar_100_chicken": 0.8012, "eval_mar_100_duck": 0.7531, "eval_mar_100_plant": 0.8572, "eval_mar_large": 0.8243, "eval_mar_medium": 0.7954, "eval_mar_small": 0.277, "eval_runtime": 12.6502, "eval_samples_per_second": 15.81, "eval_steps_per_second": 1.976, "step": 38000 }, { "epoch": 38.01, "grad_norm": 38.08417510986328, "learning_rate": 8.497568431805118e-06, "loss": 0.4389, "step": 38010 }, { "epoch": 38.04, "grad_norm": 79.51295471191406, "learning_rate": 8.495322699128206e-06, "loss": 0.4639, "step": 38040 }, { "epoch": 38.07, "grad_norm": 159.63449096679688, "learning_rate": 8.493075586553245e-06, "loss": 0.437, "step": 38070 }, { "epoch": 38.1, "grad_norm": 35.68727111816406, "learning_rate": 8.490827094967364e-06, "loss": 0.4853, "step": 38100 }, { "epoch": 38.13, "grad_norm": 123.4286880493164, "learning_rate": 8.48857722525823e-06, "loss": 0.4015, "step": 38130 }, { "epoch": 38.16, "grad_norm": 30.126785278320312, "learning_rate": 8.486325978314054e-06, "loss": 0.4259, "step": 38160 }, { "epoch": 38.19, "grad_norm": 0.00011374925816198811, "learning_rate": 8.484073355023597e-06, "loss": 0.4481, "step": 38190 }, { "epoch": 38.22, "grad_norm": 133.56454467773438, "learning_rate": 8.481819356276155e-06, "loss": 0.4119, "step": 38220 }, { "epoch": 38.25, "grad_norm": 32.053890228271484, "learning_rate": 8.479563982961572e-06, "loss": 0.4633, "step": 38250 }, { "epoch": 38.28, "grad_norm": 54.11283493041992, "learning_rate": 8.477307235970235e-06, "loss": 0.4266, "step": 38280 }, { "epoch": 38.31, "grad_norm": 23.816856384277344, "learning_rate": 8.475049116193071e-06, "loss": 0.4301, "step": 38310 }, { "epoch": 38.34, "grad_norm": 86.16854858398438, "learning_rate": 8.472789624521552e-06, "loss": 0.4648, "step": 38340 }, { "epoch": 38.37, "grad_norm": 39.290340423583984, "learning_rate": 8.470528761847684e-06, "loss": 0.3554, "step": 38370 }, { "epoch": 38.4, "grad_norm": 60.67293930053711, "learning_rate": 8.468266529064025e-06, "loss": 0.4811, "step": 38400 }, { "epoch": 38.43, "grad_norm": 61.69773864746094, "learning_rate": 8.466002927063668e-06, "loss": 0.6815, "step": 38430 }, { "epoch": 38.46, "grad_norm": 1158.8529052734375, "learning_rate": 8.463737956740246e-06, "loss": 0.3994, "step": 38460 }, { "epoch": 38.49, "grad_norm": 88.92562103271484, "learning_rate": 8.461471618987933e-06, "loss": 0.4261, "step": 38490 }, { "epoch": 38.52, "grad_norm": 28.30469512939453, "learning_rate": 8.459203914701444e-06, "loss": 0.4386, "step": 38520 }, { "epoch": 38.55, "grad_norm": 91.72919464111328, "learning_rate": 8.456934844776033e-06, "loss": 0.4093, "step": 38550 }, { "epoch": 38.58, "grad_norm": 50.374908447265625, "learning_rate": 8.454664410107492e-06, "loss": 0.4569, "step": 38580 }, { "epoch": 38.61, "grad_norm": 120.91517639160156, "learning_rate": 8.452392611592154e-06, "loss": 0.4262, "step": 38610 }, { "epoch": 38.64, "grad_norm": 33.99506759643555, "learning_rate": 8.450119450126889e-06, "loss": 0.3695, "step": 38640 }, { "epoch": 38.67, "grad_norm": 85.88146209716797, "learning_rate": 8.447844926609103e-06, "loss": 0.439, "step": 38670 }, { "epoch": 38.7, "grad_norm": 42.35884094238281, "learning_rate": 8.445569041936743e-06, "loss": 0.4367, "step": 38700 }, { "epoch": 38.73, "grad_norm": 142.43844604492188, "learning_rate": 8.443291797008294e-06, "loss": 0.5513, "step": 38730 }, { "epoch": 38.76, "grad_norm": 36.95314025878906, "learning_rate": 8.441013192722774e-06, "loss": 0.4594, "step": 38760 }, { "epoch": 38.79, "grad_norm": 58.61412811279297, "learning_rate": 8.438733229979741e-06, "loss": 0.4564, "step": 38790 }, { "epoch": 38.82, "grad_norm": 30.98107147216797, "learning_rate": 8.436451909679286e-06, "loss": 0.4818, "step": 38820 }, { "epoch": 38.85, "grad_norm": 46.01520538330078, "learning_rate": 8.434169232722043e-06, "loss": 0.4242, "step": 38850 }, { "epoch": 38.88, "grad_norm": 81.40532684326172, "learning_rate": 8.431885200009172e-06, "loss": 0.4456, "step": 38880 }, { "epoch": 38.91, "grad_norm": 75.6695327758789, "learning_rate": 8.429599812442373e-06, "loss": 0.4029, "step": 38910 }, { "epoch": 38.94, "grad_norm": 49.779197692871094, "learning_rate": 8.427313070923885e-06, "loss": 0.3858, "step": 38940 }, { "epoch": 38.97, "grad_norm": 53.02988052368164, "learning_rate": 8.425024976356474e-06, "loss": 0.3948, "step": 38970 }, { "epoch": 39.0, "grad_norm": 142.9577178955078, "learning_rate": 8.422735529643445e-06, "loss": 0.364, "step": 39000 }, { "epoch": 39.0, "eval_loss": 0.33007195591926575, "eval_map": 0.7624, "eval_map_50": 0.95, "eval_map_75": 0.8996, "eval_map_chicken": 0.7593, "eval_map_duck": 0.7028, "eval_map_large": 0.7724, "eval_map_medium": 0.758, "eval_map_plant": 0.825, "eval_map_small": 0.1968, "eval_mar_1": 0.3122, "eval_mar_10": 0.8012, "eval_mar_100": 0.8046, "eval_mar_100_chicken": 0.8119, "eval_mar_100_duck": 0.7371, "eval_mar_100_plant": 0.8649, "eval_mar_large": 0.8187, "eval_mar_medium": 0.8034, "eval_mar_small": 0.2741, "eval_runtime": 12.5623, "eval_samples_per_second": 15.921, "eval_steps_per_second": 1.99, "step": 39000 }, { "epoch": 39.03, "grad_norm": 94.47872924804688, "learning_rate": 8.420444731688632e-06, "loss": 0.3773, "step": 39030 }, { "epoch": 39.06, "grad_norm": 65.96117401123047, "learning_rate": 8.418152583396411e-06, "loss": 0.3493, "step": 39060 }, { "epoch": 39.09, "grad_norm": 26.68914794921875, "learning_rate": 8.415859085671683e-06, "loss": 0.3534, "step": 39090 }, { "epoch": 39.12, "grad_norm": 73.50113677978516, "learning_rate": 8.413564239419883e-06, "loss": 0.3905, "step": 39120 }, { "epoch": 39.15, "grad_norm": 142.21966552734375, "learning_rate": 8.411268045546984e-06, "loss": 0.3297, "step": 39150 }, { "epoch": 39.18, "grad_norm": 88.85661315917969, "learning_rate": 8.40897050495948e-06, "loss": 0.3861, "step": 39180 }, { "epoch": 39.21, "grad_norm": 27.810489654541016, "learning_rate": 8.406671618564413e-06, "loss": 0.4618, "step": 39210 }, { "epoch": 39.24, "grad_norm": 74.93070983886719, "learning_rate": 8.404371387269341e-06, "loss": 0.4358, "step": 39240 }, { "epoch": 39.27, "grad_norm": 50.285221099853516, "learning_rate": 8.40206981198236e-06, "loss": 0.4857, "step": 39270 }, { "epoch": 39.3, "grad_norm": 51.265289306640625, "learning_rate": 8.399766893612096e-06, "loss": 0.4275, "step": 39300 }, { "epoch": 39.33, "grad_norm": 51.744590759277344, "learning_rate": 8.397462633067705e-06, "loss": 0.3593, "step": 39330 }, { "epoch": 39.36, "grad_norm": 74.507080078125, "learning_rate": 8.39515703125887e-06, "loss": 0.4101, "step": 39360 }, { "epoch": 39.39, "grad_norm": 49.12809371948242, "learning_rate": 8.39285008909581e-06, "loss": 0.5324, "step": 39390 }, { "epoch": 39.42, "grad_norm": 38.36215591430664, "learning_rate": 8.390541807489266e-06, "loss": 0.4706, "step": 39420 }, { "epoch": 39.45, "grad_norm": 117.183837890625, "learning_rate": 8.388232187350513e-06, "loss": 0.3956, "step": 39450 }, { "epoch": 39.48, "grad_norm": 60.95120620727539, "learning_rate": 8.385921229591351e-06, "loss": 0.3848, "step": 39480 }, { "epoch": 39.51, "grad_norm": 38.170291900634766, "learning_rate": 8.383608935124109e-06, "loss": 0.3989, "step": 39510 }, { "epoch": 39.54, "grad_norm": 76.24112701416016, "learning_rate": 8.381295304861647e-06, "loss": 0.5061, "step": 39540 }, { "epoch": 39.57, "grad_norm": 0.0001243163860635832, "learning_rate": 8.378980339717348e-06, "loss": 0.4224, "step": 39570 }, { "epoch": 39.6, "grad_norm": 44.589778900146484, "learning_rate": 8.376664040605122e-06, "loss": 0.3577, "step": 39600 }, { "epoch": 39.63, "grad_norm": 231.2082061767578, "learning_rate": 8.37434640843941e-06, "loss": 0.481, "step": 39630 }, { "epoch": 39.66, "grad_norm": 44.966102600097656, "learning_rate": 8.372027444135176e-06, "loss": 0.4074, "step": 39660 }, { "epoch": 39.69, "grad_norm": 30.87686538696289, "learning_rate": 8.36970714860791e-06, "loss": 0.5014, "step": 39690 }, { "epoch": 39.72, "grad_norm": 68.49152374267578, "learning_rate": 8.367385522773625e-06, "loss": 0.3855, "step": 39720 }, { "epoch": 39.75, "grad_norm": 40.756473541259766, "learning_rate": 8.365062567548868e-06, "loss": 0.4893, "step": 39750 }, { "epoch": 39.78, "grad_norm": 119.0246810913086, "learning_rate": 8.3627382838507e-06, "loss": 0.3654, "step": 39780 }, { "epoch": 39.81, "grad_norm": 79.95480346679688, "learning_rate": 8.360412672596712e-06, "loss": 0.4747, "step": 39810 }, { "epoch": 39.84, "grad_norm": 77.18094635009766, "learning_rate": 8.358085734705021e-06, "loss": 0.4504, "step": 39840 }, { "epoch": 39.87, "grad_norm": 55.03760528564453, "learning_rate": 8.355757471094263e-06, "loss": 0.3282, "step": 39870 }, { "epoch": 39.9, "grad_norm": 54.48746109008789, "learning_rate": 8.353427882683601e-06, "loss": 0.419, "step": 39900 }, { "epoch": 39.93, "grad_norm": 197.91424560546875, "learning_rate": 8.351096970392718e-06, "loss": 0.4564, "step": 39930 }, { "epoch": 39.96, "grad_norm": 36.39276885986328, "learning_rate": 8.348764735141823e-06, "loss": 0.4882, "step": 39960 }, { "epoch": 39.99, "grad_norm": 89.71102142333984, "learning_rate": 8.346431177851645e-06, "loss": 0.4423, "step": 39990 }, { "epoch": 40.0, "eval_loss": 0.32653236389160156, "eval_map": 0.7682, "eval_map_50": 0.951, "eval_map_75": 0.8942, "eval_map_chicken": 0.7557, "eval_map_duck": 0.7204, "eval_map_large": 0.7604, "eval_map_medium": 0.7628, "eval_map_plant": 0.8285, "eval_map_small": 0.1885, "eval_mar_1": 0.3178, "eval_mar_10": 0.8099, "eval_mar_100": 0.8139, "eval_mar_100_chicken": 0.8115, "eval_mar_100_duck": 0.7577, "eval_mar_100_plant": 0.8726, "eval_mar_large": 0.8128, "eval_mar_medium": 0.8098, "eval_mar_small": 0.3056, "eval_runtime": 12.4291, "eval_samples_per_second": 16.091, "eval_steps_per_second": 2.011, "step": 40000 }, { "epoch": 40.02, "grad_norm": 25.755708694458008, "learning_rate": 8.344096299443434e-06, "loss": 0.4005, "step": 40020 }, { "epoch": 40.05, "grad_norm": 75.59601593017578, "learning_rate": 8.341760100838967e-06, "loss": 0.4776, "step": 40050 }, { "epoch": 40.08, "grad_norm": 63.419795989990234, "learning_rate": 8.339422582960533e-06, "loss": 0.4845, "step": 40080 }, { "epoch": 40.11, "grad_norm": 44.362159729003906, "learning_rate": 8.33708374673095e-06, "loss": 0.4088, "step": 40110 }, { "epoch": 40.14, "grad_norm": 21.852285385131836, "learning_rate": 8.334743593073553e-06, "loss": 0.3457, "step": 40140 }, { "epoch": 40.17, "grad_norm": 122.38126373291016, "learning_rate": 8.332402122912198e-06, "loss": 0.3245, "step": 40170 }, { "epoch": 40.2, "grad_norm": 110.66270446777344, "learning_rate": 8.33005933717126e-06, "loss": 0.4426, "step": 40200 }, { "epoch": 40.23, "grad_norm": 110.58285522460938, "learning_rate": 8.327715236775634e-06, "loss": 0.4074, "step": 40230 }, { "epoch": 40.26, "grad_norm": 42.71240234375, "learning_rate": 8.325369822650731e-06, "loss": 0.4001, "step": 40260 }, { "epoch": 40.29, "grad_norm": 79.57636260986328, "learning_rate": 8.323023095722486e-06, "loss": 0.4598, "step": 40290 }, { "epoch": 40.32, "grad_norm": 94.83200073242188, "learning_rate": 8.320675056917353e-06, "loss": 0.3877, "step": 40320 }, { "epoch": 40.35, "grad_norm": 54.34486770629883, "learning_rate": 8.318325707162293e-06, "loss": 0.4375, "step": 40350 }, { "epoch": 40.38, "grad_norm": 25.992382049560547, "learning_rate": 8.315975047384798e-06, "loss": 0.438, "step": 40380 }, { "epoch": 40.41, "grad_norm": 39.75422286987305, "learning_rate": 8.313623078512869e-06, "loss": 0.422, "step": 40410 }, { "epoch": 40.44, "grad_norm": 32.47829055786133, "learning_rate": 8.311269801475026e-06, "loss": 0.3745, "step": 40440 }, { "epoch": 40.47, "grad_norm": 43.39604949951172, "learning_rate": 8.308915217200305e-06, "loss": 0.453, "step": 40470 }, { "epoch": 40.5, "grad_norm": 55.71866226196289, "learning_rate": 8.30655932661826e-06, "loss": 0.3991, "step": 40500 }, { "epoch": 40.53, "grad_norm": 40.85260009765625, "learning_rate": 8.304202130658959e-06, "loss": 0.3759, "step": 40530 }, { "epoch": 40.56, "grad_norm": 439.5714416503906, "learning_rate": 8.301843630252986e-06, "loss": 0.4052, "step": 40560 }, { "epoch": 40.59, "grad_norm": 70.4522705078125, "learning_rate": 8.299483826331436e-06, "loss": 0.3757, "step": 40590 }, { "epoch": 40.62, "grad_norm": 74.4576416015625, "learning_rate": 8.297122719825928e-06, "loss": 0.4548, "step": 40620 }, { "epoch": 40.65, "grad_norm": 86.23299407958984, "learning_rate": 8.294760311668586e-06, "loss": 0.448, "step": 40650 }, { "epoch": 40.68, "grad_norm": 68.10324096679688, "learning_rate": 8.29239660279205e-06, "loss": 0.3547, "step": 40680 }, { "epoch": 40.71, "grad_norm": 47.49440383911133, "learning_rate": 8.29003159412948e-06, "loss": 0.4335, "step": 40710 }, { "epoch": 40.74, "grad_norm": 31.3197021484375, "learning_rate": 8.287665286614538e-06, "loss": 0.5129, "step": 40740 }, { "epoch": 40.77, "grad_norm": 37.6073112487793, "learning_rate": 8.285297681181408e-06, "loss": 0.5047, "step": 40770 }, { "epoch": 40.8, "grad_norm": 88.99430847167969, "learning_rate": 8.282928778764783e-06, "loss": 0.4331, "step": 40800 }, { "epoch": 40.83, "grad_norm": 92.79273223876953, "learning_rate": 8.280558580299868e-06, "loss": 0.4649, "step": 40830 }, { "epoch": 40.86, "grad_norm": 56.76251983642578, "learning_rate": 8.278187086722378e-06, "loss": 0.4794, "step": 40860 }, { "epoch": 40.89, "grad_norm": 35.74705123901367, "learning_rate": 8.275814298968544e-06, "loss": 0.3511, "step": 40890 }, { "epoch": 40.92, "grad_norm": 172.78829956054688, "learning_rate": 8.273440217975103e-06, "loss": 0.4637, "step": 40920 }, { "epoch": 40.95, "grad_norm": 60.8847541809082, "learning_rate": 8.271064844679306e-06, "loss": 0.4191, "step": 40950 }, { "epoch": 40.98, "grad_norm": 68.42402648925781, "learning_rate": 8.268688180018911e-06, "loss": 0.3772, "step": 40980 }, { "epoch": 41.0, "eval_loss": 0.3484590947628021, "eval_map": 0.7493, "eval_map_50": 0.9639, "eval_map_75": 0.9004, "eval_map_chicken": 0.7131, "eval_map_duck": 0.7034, "eval_map_large": 0.7674, "eval_map_medium": 0.7341, "eval_map_plant": 0.8314, "eval_map_small": 0.2278, "eval_mar_1": 0.3049, "eval_mar_10": 0.7884, "eval_mar_100": 0.7944, "eval_mar_100_chicken": 0.7694, "eval_mar_100_duck": 0.7443, "eval_mar_100_plant": 0.8694, "eval_mar_large": 0.8144, "eval_mar_medium": 0.7842, "eval_mar_small": 0.3943, "eval_runtime": 12.5066, "eval_samples_per_second": 15.992, "eval_steps_per_second": 1.999, "step": 41000 }, { "epoch": 41.01, "grad_norm": 103.41185760498047, "learning_rate": 8.266310224932191e-06, "loss": 0.5867, "step": 41010 }, { "epoch": 41.04, "grad_norm": 76.7298355102539, "learning_rate": 8.26393098035792e-06, "loss": 0.472, "step": 41040 }, { "epoch": 41.07, "grad_norm": 38.80534744262695, "learning_rate": 8.261550447235389e-06, "loss": 0.5264, "step": 41070 }, { "epoch": 41.1, "grad_norm": 29.254039764404297, "learning_rate": 8.259168626504395e-06, "loss": 0.3973, "step": 41100 }, { "epoch": 41.13, "grad_norm": 0.00020122693968005478, "learning_rate": 8.256785519105242e-06, "loss": 0.3753, "step": 41130 }, { "epoch": 41.16, "grad_norm": 745.515869140625, "learning_rate": 8.254401125978744e-06, "loss": 0.4534, "step": 41160 }, { "epoch": 41.19, "grad_norm": 28.32933235168457, "learning_rate": 8.25201544806622e-06, "loss": 0.4323, "step": 41190 }, { "epoch": 41.22, "grad_norm": 46.06074142456055, "learning_rate": 8.2496284863095e-06, "loss": 0.4211, "step": 41220 }, { "epoch": 41.25, "grad_norm": 40.98131561279297, "learning_rate": 8.247240241650918e-06, "loss": 0.4724, "step": 41250 }, { "epoch": 41.28, "grad_norm": 53.212589263916016, "learning_rate": 8.244850715033316e-06, "loss": 0.4456, "step": 41280 }, { "epoch": 41.31, "grad_norm": 34.923927307128906, "learning_rate": 8.24245990740004e-06, "loss": 0.4539, "step": 41310 }, { "epoch": 41.34, "grad_norm": 56.931209564208984, "learning_rate": 8.240067819694941e-06, "loss": 0.4187, "step": 41340 }, { "epoch": 41.37, "grad_norm": 43.071319580078125, "learning_rate": 8.23767445286238e-06, "loss": 0.421, "step": 41370 }, { "epoch": 41.4, "grad_norm": 328.1631164550781, "learning_rate": 8.235279807847223e-06, "loss": 0.4807, "step": 41400 }, { "epoch": 41.43, "grad_norm": 100.90284729003906, "learning_rate": 8.232883885594831e-06, "loss": 0.4413, "step": 41430 }, { "epoch": 41.46, "grad_norm": 55.4809684753418, "learning_rate": 8.230486687051082e-06, "loss": 0.4367, "step": 41460 }, { "epoch": 41.49, "grad_norm": 34.61279296875, "learning_rate": 8.22808821316235e-06, "loss": 0.3828, "step": 41490 }, { "epoch": 41.52, "grad_norm": 40.92034149169922, "learning_rate": 8.225688464875514e-06, "loss": 0.3656, "step": 41520 }, { "epoch": 41.55, "grad_norm": 177.9812469482422, "learning_rate": 8.223287443137957e-06, "loss": 0.3576, "step": 41550 }, { "epoch": 41.58, "grad_norm": 52.98564147949219, "learning_rate": 8.220885148897566e-06, "loss": 0.3686, "step": 41580 }, { "epoch": 41.61, "grad_norm": 41.121150970458984, "learning_rate": 8.218481583102725e-06, "loss": 0.4402, "step": 41610 }, { "epoch": 41.64, "grad_norm": 145.01028442382812, "learning_rate": 8.216076746702327e-06, "loss": 0.4646, "step": 41640 }, { "epoch": 41.67, "grad_norm": 37.7918815612793, "learning_rate": 8.213670640645762e-06, "loss": 0.4163, "step": 41670 }, { "epoch": 41.7, "grad_norm": 39.085819244384766, "learning_rate": 8.211263265882923e-06, "loss": 0.4381, "step": 41700 }, { "epoch": 41.73, "grad_norm": 67.86912536621094, "learning_rate": 8.208854623364202e-06, "loss": 0.3821, "step": 41730 }, { "epoch": 41.76, "grad_norm": 72.80220794677734, "learning_rate": 8.206444714040496e-06, "loss": 0.424, "step": 41760 }, { "epoch": 41.79, "grad_norm": 30.086559295654297, "learning_rate": 8.204033538863196e-06, "loss": 0.3653, "step": 41790 }, { "epoch": 41.82, "grad_norm": 49.32461929321289, "learning_rate": 8.201621098784198e-06, "loss": 0.4309, "step": 41820 }, { "epoch": 41.85, "grad_norm": 101.68267822265625, "learning_rate": 8.199207394755892e-06, "loss": 0.4322, "step": 41850 }, { "epoch": 41.88, "grad_norm": 79.99913024902344, "learning_rate": 8.196792427731175e-06, "loss": 0.4556, "step": 41880 }, { "epoch": 41.91, "grad_norm": 44.01887512207031, "learning_rate": 8.194376198663434e-06, "loss": 0.3961, "step": 41910 }, { "epoch": 41.94, "grad_norm": 86.05049896240234, "learning_rate": 8.191958708506557e-06, "loss": 0.4451, "step": 41940 }, { "epoch": 41.97, "grad_norm": 48.83458709716797, "learning_rate": 8.189539958214934e-06, "loss": 0.495, "step": 41970 }, { "epoch": 42.0, "grad_norm": 74.62286376953125, "learning_rate": 8.18711994874345e-06, "loss": 0.4682, "step": 42000 }, { "epoch": 42.0, "eval_loss": 0.3569546043872833, "eval_map": 0.7437, "eval_map_50": 0.9572, "eval_map_75": 0.894, "eval_map_chicken": 0.7434, "eval_map_duck": 0.6911, "eval_map_large": 0.7549, "eval_map_medium": 0.7235, "eval_map_plant": 0.7965, "eval_map_small": 0.2024, "eval_mar_1": 0.3045, "eval_mar_10": 0.782, "eval_mar_100": 0.7869, "eval_mar_100_chicken": 0.794, "eval_mar_100_duck": 0.7299, "eval_mar_100_plant": 0.837, "eval_mar_large": 0.8017, "eval_mar_medium": 0.7728, "eval_mar_small": 0.331, "eval_runtime": 12.5278, "eval_samples_per_second": 15.964, "eval_steps_per_second": 1.996, "step": 42000 }, { "epoch": 42.03, "grad_norm": 40.18670654296875, "learning_rate": 8.184698681047482e-06, "loss": 0.4256, "step": 42030 }, { "epoch": 42.06, "grad_norm": 476.9875183105469, "learning_rate": 8.182276156082911e-06, "loss": 0.6544, "step": 42060 }, { "epoch": 42.09, "grad_norm": 54.456851959228516, "learning_rate": 8.179852374806112e-06, "loss": 0.4082, "step": 42090 }, { "epoch": 42.12, "grad_norm": 57.975555419921875, "learning_rate": 8.177427338173955e-06, "loss": 0.4026, "step": 42120 }, { "epoch": 42.15, "grad_norm": 50.068336486816406, "learning_rate": 8.175001047143804e-06, "loss": 0.4383, "step": 42150 }, { "epoch": 42.18, "grad_norm": 29.130226135253906, "learning_rate": 8.172573502673522e-06, "loss": 0.4062, "step": 42180 }, { "epoch": 42.21, "grad_norm": 0.0001481762737967074, "learning_rate": 8.170144705721465e-06, "loss": 0.4663, "step": 42210 }, { "epoch": 42.24, "grad_norm": 36.15610885620117, "learning_rate": 8.167714657246486e-06, "loss": 0.4034, "step": 42240 }, { "epoch": 42.27, "grad_norm": 54.05368423461914, "learning_rate": 8.165283358207924e-06, "loss": 0.4761, "step": 42270 }, { "epoch": 42.3, "grad_norm": 137.5204620361328, "learning_rate": 8.162850809565623e-06, "loss": 0.3735, "step": 42300 }, { "epoch": 42.33, "grad_norm": 35.466365814208984, "learning_rate": 8.160417012279911e-06, "loss": 0.4592, "step": 42330 }, { "epoch": 42.36, "grad_norm": 48.81232833862305, "learning_rate": 8.157981967311614e-06, "loss": 0.3437, "step": 42360 }, { "epoch": 42.39, "grad_norm": 101.01595306396484, "learning_rate": 8.155545675622049e-06, "loss": 0.4076, "step": 42390 }, { "epoch": 42.42, "grad_norm": 64.87741088867188, "learning_rate": 8.153108138173027e-06, "loss": 0.4312, "step": 42420 }, { "epoch": 42.45, "grad_norm": 60.82405471801758, "learning_rate": 8.150669355926848e-06, "loss": 0.4462, "step": 42450 }, { "epoch": 42.48, "grad_norm": 44.688941955566406, "learning_rate": 8.1482293298463e-06, "loss": 0.4399, "step": 42480 }, { "epoch": 42.51, "grad_norm": 25.950233459472656, "learning_rate": 8.145788060894675e-06, "loss": 0.4016, "step": 42510 }, { "epoch": 42.54, "grad_norm": 46.57575225830078, "learning_rate": 8.143345550035742e-06, "loss": 0.4055, "step": 42540 }, { "epoch": 42.57, "grad_norm": 185.34420776367188, "learning_rate": 8.140901798233766e-06, "loss": 0.4156, "step": 42570 }, { "epoch": 42.6, "grad_norm": 61.05455780029297, "learning_rate": 8.138456806453503e-06, "loss": 0.4581, "step": 42600 }, { "epoch": 42.63, "grad_norm": 63.83736801147461, "learning_rate": 8.136010575660197e-06, "loss": 0.3334, "step": 42630 }, { "epoch": 42.66, "grad_norm": 50.89091110229492, "learning_rate": 8.133563106819579e-06, "loss": 0.3688, "step": 42660 }, { "epoch": 42.69, "grad_norm": 106.39220428466797, "learning_rate": 8.131114400897874e-06, "loss": 0.4347, "step": 42690 }, { "epoch": 42.72, "grad_norm": 56.02287292480469, "learning_rate": 8.12866445886179e-06, "loss": 0.3927, "step": 42720 }, { "epoch": 42.75, "grad_norm": 77.4242935180664, "learning_rate": 8.126213281678527e-06, "loss": 0.5651, "step": 42750 }, { "epoch": 42.78, "grad_norm": 0.00195989478379488, "learning_rate": 8.12376087031577e-06, "loss": 0.4147, "step": 42780 }, { "epoch": 42.81, "grad_norm": 35.393592834472656, "learning_rate": 8.12130722574169e-06, "loss": 0.3458, "step": 42810 }, { "epoch": 42.84, "grad_norm": 5.639511437038891e-05, "learning_rate": 8.118852348924951e-06, "loss": 0.4061, "step": 42840 }, { "epoch": 42.87, "grad_norm": 43.78005599975586, "learning_rate": 8.116396240834699e-06, "loss": 0.3799, "step": 42870 }, { "epoch": 42.9, "grad_norm": 91.38209533691406, "learning_rate": 8.113938902440563e-06, "loss": 0.3982, "step": 42900 }, { "epoch": 42.93, "grad_norm": 0.012147296220064163, "learning_rate": 8.111480334712664e-06, "loss": 0.3665, "step": 42930 }, { "epoch": 42.96, "grad_norm": 54.34785461425781, "learning_rate": 8.109020538621607e-06, "loss": 0.4785, "step": 42960 }, { "epoch": 42.99, "grad_norm": 4.695102325058542e-05, "learning_rate": 8.106559515138477e-06, "loss": 0.4829, "step": 42990 }, { "epoch": 43.0, "eval_loss": 0.32948189973831177, "eval_map": 0.7733, "eval_map_50": 0.9559, "eval_map_75": 0.9061, "eval_map_chicken": 0.77, "eval_map_duck": 0.7311, "eval_map_large": 0.7853, "eval_map_medium": 0.7635, "eval_map_plant": 0.8189, "eval_map_small": 0.2047, "eval_mar_1": 0.3174, "eval_mar_10": 0.8091, "eval_mar_100": 0.8123, "eval_mar_100_chicken": 0.8153, "eval_mar_100_duck": 0.767, "eval_mar_100_plant": 0.8546, "eval_mar_large": 0.8238, "eval_mar_medium": 0.8077, "eval_mar_small": 0.2933, "eval_runtime": 12.5772, "eval_samples_per_second": 15.902, "eval_steps_per_second": 1.988, "step": 43000 }, { "epoch": 43.02, "grad_norm": 74.98933410644531, "learning_rate": 8.10409726523485e-06, "loss": 0.4011, "step": 43020 }, { "epoch": 43.05, "grad_norm": 0.00030158780282363296, "learning_rate": 8.101633789882781e-06, "loss": 0.46, "step": 43050 }, { "epoch": 43.08, "grad_norm": 92.26671600341797, "learning_rate": 8.099169090054812e-06, "loss": 0.2728, "step": 43080 }, { "epoch": 43.11, "grad_norm": 36.43602752685547, "learning_rate": 8.09670316672397e-06, "loss": 0.4859, "step": 43110 }, { "epoch": 43.14, "grad_norm": 30.34481430053711, "learning_rate": 8.094236020863758e-06, "loss": 0.3584, "step": 43140 }, { "epoch": 43.17, "grad_norm": 182.4502716064453, "learning_rate": 8.091767653448169e-06, "loss": 0.3301, "step": 43170 }, { "epoch": 43.2, "grad_norm": 58.02518844604492, "learning_rate": 8.089298065451673e-06, "loss": 0.3116, "step": 43200 }, { "epoch": 43.23, "grad_norm": 78.188720703125, "learning_rate": 8.086827257849225e-06, "loss": 0.4066, "step": 43230 }, { "epoch": 43.26, "grad_norm": 81.51823425292969, "learning_rate": 8.084355231616266e-06, "loss": 0.4155, "step": 43260 }, { "epoch": 43.29, "grad_norm": 56.928958892822266, "learning_rate": 8.081881987728704e-06, "loss": 0.4085, "step": 43290 }, { "epoch": 43.32, "grad_norm": 48.266929626464844, "learning_rate": 8.079407527162944e-06, "loss": 0.436, "step": 43320 }, { "epoch": 43.35, "grad_norm": 56.23191452026367, "learning_rate": 8.076931850895858e-06, "loss": 0.4143, "step": 43350 }, { "epoch": 43.38, "grad_norm": 140.52964782714844, "learning_rate": 8.074454959904807e-06, "loss": 0.4709, "step": 43380 }, { "epoch": 43.41, "grad_norm": 51.77798843383789, "learning_rate": 8.071976855167629e-06, "loss": 0.4903, "step": 43410 }, { "epoch": 43.44, "grad_norm": 33.46908187866211, "learning_rate": 8.069497537662638e-06, "loss": 0.4061, "step": 43440 }, { "epoch": 43.47, "grad_norm": 59.45002365112305, "learning_rate": 8.067017008368632e-06, "loss": 0.528, "step": 43470 }, { "epoch": 43.5, "grad_norm": 37.091312408447266, "learning_rate": 8.064535268264883e-06, "loss": 0.4671, "step": 43500 }, { "epoch": 43.53, "grad_norm": 55.32894515991211, "learning_rate": 8.062052318331142e-06, "loss": 0.4443, "step": 43530 }, { "epoch": 43.56, "grad_norm": 60.84852600097656, "learning_rate": 8.05956815954764e-06, "loss": 0.3684, "step": 43560 }, { "epoch": 43.59, "grad_norm": 47.42966842651367, "learning_rate": 8.057082792895084e-06, "loss": 0.4407, "step": 43590 }, { "epoch": 43.62, "grad_norm": 38.62168884277344, "learning_rate": 8.054596219354655e-06, "loss": 0.4764, "step": 43620 }, { "epoch": 43.65, "grad_norm": 50.650264739990234, "learning_rate": 8.052108439908014e-06, "loss": 0.3778, "step": 43650 }, { "epoch": 43.68, "grad_norm": 26.091550827026367, "learning_rate": 8.049619455537296e-06, "loss": 0.3718, "step": 43680 }, { "epoch": 43.71, "grad_norm": 150.9571990966797, "learning_rate": 8.047129267225116e-06, "loss": 0.4199, "step": 43710 }, { "epoch": 43.74, "grad_norm": 64.30305480957031, "learning_rate": 8.044637875954556e-06, "loss": 0.4189, "step": 43740 }, { "epoch": 43.77, "grad_norm": 41.272544860839844, "learning_rate": 8.042145282709181e-06, "loss": 0.5038, "step": 43770 }, { "epoch": 43.8, "grad_norm": 57.9762077331543, "learning_rate": 8.039651488473028e-06, "loss": 0.335, "step": 43800 }, { "epoch": 43.83, "grad_norm": 49.523006439208984, "learning_rate": 8.037156494230605e-06, "loss": 0.3555, "step": 43830 }, { "epoch": 43.86, "grad_norm": 78.95838165283203, "learning_rate": 8.034660300966898e-06, "loss": 0.43, "step": 43860 }, { "epoch": 43.89, "grad_norm": 41.18893051147461, "learning_rate": 8.032162909667363e-06, "loss": 0.5009, "step": 43890 }, { "epoch": 43.92, "grad_norm": 36.89976501464844, "learning_rate": 8.029664321317932e-06, "loss": 0.4179, "step": 43920 }, { "epoch": 43.95, "grad_norm": 28.92719078063965, "learning_rate": 8.027164536905008e-06, "loss": 0.4252, "step": 43950 }, { "epoch": 43.98, "grad_norm": 157.78529357910156, "learning_rate": 8.024663557415466e-06, "loss": 0.4646, "step": 43980 }, { "epoch": 44.0, "eval_loss": 0.3219194710254669, "eval_map": 0.7697, "eval_map_50": 0.9622, "eval_map_75": 0.9075, "eval_map_chicken": 0.7647, "eval_map_duck": 0.7122, "eval_map_large": 0.7799, "eval_map_medium": 0.756, "eval_map_plant": 0.8321, "eval_map_small": 0.2313, "eval_mar_1": 0.3135, "eval_mar_10": 0.808, "eval_mar_100": 0.8135, "eval_mar_100_chicken": 0.8129, "eval_mar_100_duck": 0.7577, "eval_mar_100_plant": 0.87, "eval_mar_large": 0.8236, "eval_mar_medium": 0.8049, "eval_mar_small": 0.367, "eval_runtime": 12.632, "eval_samples_per_second": 15.833, "eval_steps_per_second": 1.979, "step": 44000 }, { "epoch": 44.01, "grad_norm": 27.94495391845703, "learning_rate": 8.022161383836652e-06, "loss": 0.5338, "step": 44010 }, { "epoch": 44.04, "grad_norm": 126.91854858398438, "learning_rate": 8.019658017156384e-06, "loss": 0.4592, "step": 44040 }, { "epoch": 44.07, "grad_norm": 0.00016747893823776394, "learning_rate": 8.017153458362957e-06, "loss": 0.3887, "step": 44070 }, { "epoch": 44.1, "grad_norm": 58.71892166137695, "learning_rate": 8.014647708445124e-06, "loss": 0.4663, "step": 44100 }, { "epoch": 44.13, "grad_norm": 28.824718475341797, "learning_rate": 8.01214076839212e-06, "loss": 0.379, "step": 44130 }, { "epoch": 44.16, "grad_norm": 62.654964447021484, "learning_rate": 8.009632639193643e-06, "loss": 0.4724, "step": 44160 }, { "epoch": 44.19, "grad_norm": 45.86119842529297, "learning_rate": 8.007123321839865e-06, "loss": 0.5198, "step": 44190 }, { "epoch": 44.22, "grad_norm": 76.88160705566406, "learning_rate": 8.004612817321419e-06, "loss": 0.4986, "step": 44220 }, { "epoch": 44.25, "grad_norm": 103.96524047851562, "learning_rate": 8.002101126629422e-06, "loss": 0.4346, "step": 44250 }, { "epoch": 44.28, "grad_norm": 41.1450309753418, "learning_rate": 7.999588250755442e-06, "loss": 0.4142, "step": 44280 }, { "epoch": 44.31, "grad_norm": 84.7293701171875, "learning_rate": 7.997074190691523e-06, "loss": 0.379, "step": 44310 }, { "epoch": 44.34, "grad_norm": 65.0225830078125, "learning_rate": 7.99455894743018e-06, "loss": 0.466, "step": 44340 }, { "epoch": 44.37, "grad_norm": 28.5335693359375, "learning_rate": 7.99204252196439e-06, "loss": 0.3464, "step": 44370 }, { "epoch": 44.4, "grad_norm": 63.27149200439453, "learning_rate": 7.989524915287595e-06, "loss": 0.4464, "step": 44400 }, { "epoch": 44.43, "grad_norm": 35.89200210571289, "learning_rate": 7.98700612839371e-06, "loss": 0.3678, "step": 44430 }, { "epoch": 44.46, "grad_norm": 25.133045196533203, "learning_rate": 7.98448616227711e-06, "loss": 0.3885, "step": 44460 }, { "epoch": 44.49, "grad_norm": 57.70933151245117, "learning_rate": 7.981965017932638e-06, "loss": 0.3956, "step": 44490 }, { "epoch": 44.52, "grad_norm": 95.73916625976562, "learning_rate": 7.979442696355601e-06, "loss": 0.4414, "step": 44520 }, { "epoch": 44.55, "grad_norm": 42.46126174926758, "learning_rate": 7.976919198541775e-06, "loss": 0.4482, "step": 44550 }, { "epoch": 44.58, "grad_norm": 55.41289138793945, "learning_rate": 7.974394525487395e-06, "loss": 0.4022, "step": 44580 }, { "epoch": 44.61, "grad_norm": 46.46371841430664, "learning_rate": 7.97186867818916e-06, "loss": 0.4747, "step": 44610 }, { "epoch": 44.64, "grad_norm": 55.15605926513672, "learning_rate": 7.969341657644236e-06, "loss": 0.4395, "step": 44640 }, { "epoch": 44.67, "grad_norm": 101.15933990478516, "learning_rate": 7.966813464850252e-06, "loss": 0.4519, "step": 44670 }, { "epoch": 44.7, "grad_norm": 53.96013641357422, "learning_rate": 7.964284100805297e-06, "loss": 0.4577, "step": 44700 }, { "epoch": 44.73, "grad_norm": 36.329044342041016, "learning_rate": 7.961753566507924e-06, "loss": 0.4282, "step": 44730 }, { "epoch": 44.76, "grad_norm": 0.00020565191516652703, "learning_rate": 7.959221862957149e-06, "loss": 0.4663, "step": 44760 }, { "epoch": 44.79, "grad_norm": 44.466915130615234, "learning_rate": 7.956688991152446e-06, "loss": 0.3944, "step": 44790 }, { "epoch": 44.82, "grad_norm": 77.13082122802734, "learning_rate": 7.954154952093754e-06, "loss": 0.402, "step": 44820 }, { "epoch": 44.85, "grad_norm": 105.04215240478516, "learning_rate": 7.951619746781474e-06, "loss": 0.4156, "step": 44850 }, { "epoch": 44.88, "grad_norm": 41.99253845214844, "learning_rate": 7.94908337621646e-06, "loss": 0.3968, "step": 44880 }, { "epoch": 44.91, "grad_norm": 35.881343841552734, "learning_rate": 7.946545841400035e-06, "loss": 0.5002, "step": 44910 }, { "epoch": 44.94, "grad_norm": 96.42914581298828, "learning_rate": 7.944007143333976e-06, "loss": 0.513, "step": 44940 }, { "epoch": 44.97, "grad_norm": 64.91358184814453, "learning_rate": 7.941467283020521e-06, "loss": 0.5819, "step": 44970 }, { "epoch": 45.0, "grad_norm": 61.94123840332031, "learning_rate": 7.938926261462366e-06, "loss": 0.3236, "step": 45000 }, { "epoch": 45.0, "eval_loss": 0.3212050497531891, "eval_map": 0.7713, "eval_map_50": 0.961, "eval_map_75": 0.9025, "eval_map_chicken": 0.7576, "eval_map_duck": 0.7256, "eval_map_large": 0.7637, "eval_map_medium": 0.7623, "eval_map_plant": 0.8307, "eval_map_small": 0.2236, "eval_mar_1": 0.3156, "eval_mar_10": 0.8108, "eval_mar_100": 0.8148, "eval_mar_100_chicken": 0.8139, "eval_mar_100_duck": 0.7624, "eval_mar_100_plant": 0.8683, "eval_mar_large": 0.8092, "eval_mar_medium": 0.8115, "eval_mar_small": 0.3744, "eval_runtime": 12.393, "eval_samples_per_second": 16.138, "eval_steps_per_second": 2.017, "step": 45000 }, { "epoch": 45.03, "grad_norm": 27.931364059448242, "learning_rate": 7.936384079662666e-06, "loss": 0.4163, "step": 45030 }, { "epoch": 45.06, "grad_norm": 53.77098846435547, "learning_rate": 7.933840738625035e-06, "loss": 0.4056, "step": 45060 }, { "epoch": 45.09, "grad_norm": 49.3598518371582, "learning_rate": 7.931296239353546e-06, "loss": 0.3818, "step": 45090 }, { "epoch": 45.12, "grad_norm": 136.22975158691406, "learning_rate": 7.928750582852722e-06, "loss": 0.3416, "step": 45120 }, { "epoch": 45.15, "grad_norm": 73.1864242553711, "learning_rate": 7.926203770127552e-06, "loss": 0.4555, "step": 45150 }, { "epoch": 45.18, "grad_norm": 44.06271743774414, "learning_rate": 7.923655802183475e-06, "loss": 0.3897, "step": 45180 }, { "epoch": 45.21, "grad_norm": 38.79288864135742, "learning_rate": 7.921106680026388e-06, "loss": 0.4634, "step": 45210 }, { "epoch": 45.24, "grad_norm": 43.428627014160156, "learning_rate": 7.918556404662645e-06, "loss": 0.2609, "step": 45240 }, { "epoch": 45.27, "grad_norm": 34.630699157714844, "learning_rate": 7.916004977099054e-06, "loss": 0.3955, "step": 45270 }, { "epoch": 45.3, "grad_norm": 235.07066345214844, "learning_rate": 7.913452398342882e-06, "loss": 0.3664, "step": 45300 }, { "epoch": 45.33, "grad_norm": 83.82819366455078, "learning_rate": 7.91089866940184e-06, "loss": 0.3493, "step": 45330 }, { "epoch": 45.36, "grad_norm": 146.09983825683594, "learning_rate": 7.908343791284104e-06, "loss": 0.5683, "step": 45360 }, { "epoch": 45.39, "grad_norm": 41.229976654052734, "learning_rate": 7.9057877649983e-06, "loss": 0.5386, "step": 45390 }, { "epoch": 45.42, "grad_norm": 64.58255004882812, "learning_rate": 7.903230591553504e-06, "loss": 0.3957, "step": 45420 }, { "epoch": 45.45, "grad_norm": 69.7751693725586, "learning_rate": 7.900672271959247e-06, "loss": 0.4009, "step": 45450 }, { "epoch": 45.48, "grad_norm": 192.54736328125, "learning_rate": 7.898112807225517e-06, "loss": 0.3402, "step": 45480 }, { "epoch": 45.51, "grad_norm": 60.41750717163086, "learning_rate": 7.895552198362748e-06, "loss": 0.417, "step": 45510 }, { "epoch": 45.54, "grad_norm": 41.733829498291016, "learning_rate": 7.892990446381828e-06, "loss": 0.3494, "step": 45540 }, { "epoch": 45.57, "grad_norm": 77.05609130859375, "learning_rate": 7.890427552294093e-06, "loss": 0.3312, "step": 45570 }, { "epoch": 45.6, "grad_norm": 53.536766052246094, "learning_rate": 7.887863517111337e-06, "loss": 0.4106, "step": 45600 }, { "epoch": 45.63, "grad_norm": 55.88993835449219, "learning_rate": 7.885298341845803e-06, "loss": 0.5595, "step": 45630 }, { "epoch": 45.66, "grad_norm": 94.33480072021484, "learning_rate": 7.882732027510174e-06, "loss": 0.3075, "step": 45660 }, { "epoch": 45.69, "grad_norm": 115.30720520019531, "learning_rate": 7.880164575117596e-06, "loss": 0.4247, "step": 45690 }, { "epoch": 45.72, "grad_norm": 50.447059631347656, "learning_rate": 7.877595985681656e-06, "loss": 0.3619, "step": 45720 }, { "epoch": 45.75, "grad_norm": 56.96340560913086, "learning_rate": 7.875026260216395e-06, "loss": 0.4715, "step": 45750 }, { "epoch": 45.78, "grad_norm": 60.88690185546875, "learning_rate": 7.872455399736295e-06, "loss": 0.4422, "step": 45780 }, { "epoch": 45.81, "grad_norm": 98.71990966796875, "learning_rate": 7.869883405256296e-06, "loss": 0.3358, "step": 45810 }, { "epoch": 45.84, "grad_norm": 58.58422088623047, "learning_rate": 7.867310277791778e-06, "loss": 0.3485, "step": 45840 }, { "epoch": 45.87, "grad_norm": 88.91502380371094, "learning_rate": 7.864736018358571e-06, "loss": 0.3808, "step": 45870 }, { "epoch": 45.9, "grad_norm": 24.32729148864746, "learning_rate": 7.862160627972956e-06, "loss": 0.4027, "step": 45900 }, { "epoch": 45.93, "grad_norm": 86.62168884277344, "learning_rate": 7.85958410765165e-06, "loss": 0.3886, "step": 45930 }, { "epoch": 45.96, "grad_norm": 71.1141128540039, "learning_rate": 7.857006458411826e-06, "loss": 0.4332, "step": 45960 }, { "epoch": 45.99, "grad_norm": 110.92717742919922, "learning_rate": 7.8544276812711e-06, "loss": 0.3759, "step": 45990 }, { "epoch": 46.0, "eval_loss": 0.3280904293060303, "eval_map": 0.756, "eval_map_50": 0.9604, "eval_map_75": 0.8943, "eval_map_chicken": 0.7556, "eval_map_duck": 0.6797, "eval_map_large": 0.7606, "eval_map_medium": 0.7466, "eval_map_plant": 0.8326, "eval_map_small": 0.2171, "eval_mar_1": 0.3069, "eval_mar_10": 0.7945, "eval_mar_100": 0.8001, "eval_mar_100_chicken": 0.8042, "eval_mar_100_duck": 0.7237, "eval_mar_100_plant": 0.8725, "eval_mar_large": 0.8076, "eval_mar_medium": 0.7971, "eval_mar_small": 0.3131, "eval_runtime": 12.9953, "eval_samples_per_second": 15.39, "eval_steps_per_second": 1.924, "step": 46000 }, { "epoch": 46.02, "grad_norm": 134.90220642089844, "learning_rate": 7.851847777247528e-06, "loss": 0.3943, "step": 46020 }, { "epoch": 46.05, "grad_norm": 43.66093826293945, "learning_rate": 7.849266747359619e-06, "loss": 0.426, "step": 46050 }, { "epoch": 46.08, "grad_norm": 221.4602508544922, "learning_rate": 7.846684592626324e-06, "loss": 0.498, "step": 46080 }, { "epoch": 46.11, "grad_norm": 89.09764099121094, "learning_rate": 7.844101314067031e-06, "loss": 0.4038, "step": 46110 }, { "epoch": 46.14, "grad_norm": 66.09234619140625, "learning_rate": 7.841516912701585e-06, "loss": 0.3859, "step": 46140 }, { "epoch": 46.17, "grad_norm": 110.86426544189453, "learning_rate": 7.83893138955026e-06, "loss": 0.4571, "step": 46170 }, { "epoch": 46.2, "grad_norm": 100.20338439941406, "learning_rate": 7.836344745633785e-06, "loss": 0.3651, "step": 46200 }, { "epoch": 46.23, "grad_norm": 108.07897186279297, "learning_rate": 7.833756981973321e-06, "loss": 0.4147, "step": 46230 }, { "epoch": 46.26, "grad_norm": 87.84929656982422, "learning_rate": 7.831168099590478e-06, "loss": 0.4641, "step": 46260 }, { "epoch": 46.29, "grad_norm": 74.36924743652344, "learning_rate": 7.828578099507308e-06, "loss": 0.481, "step": 46290 }, { "epoch": 46.32, "grad_norm": 85.95771789550781, "learning_rate": 7.8259869827463e-06, "loss": 0.3666, "step": 46320 }, { "epoch": 46.35, "grad_norm": 32.19763946533203, "learning_rate": 7.823394750330386e-06, "loss": 0.4093, "step": 46350 }, { "epoch": 46.38, "grad_norm": 0.0002737130271270871, "learning_rate": 7.82080140328294e-06, "loss": 0.4967, "step": 46380 }, { "epoch": 46.41, "grad_norm": 41.20289993286133, "learning_rate": 7.81820694262777e-06, "loss": 0.3939, "step": 46410 }, { "epoch": 46.44, "grad_norm": 0.0012339740060269833, "learning_rate": 7.815611369389134e-06, "loss": 0.3576, "step": 46440 }, { "epoch": 46.47, "grad_norm": 96.70281982421875, "learning_rate": 7.813014684591718e-06, "loss": 0.4455, "step": 46470 }, { "epoch": 46.5, "grad_norm": 43.737152099609375, "learning_rate": 7.810416889260653e-06, "loss": 0.3322, "step": 46500 }, { "epoch": 46.53, "grad_norm": 72.09654998779297, "learning_rate": 7.80781798442151e-06, "loss": 0.3494, "step": 46530 }, { "epoch": 46.56, "grad_norm": 84.11396789550781, "learning_rate": 7.805217971100295e-06, "loss": 0.3562, "step": 46560 }, { "epoch": 46.59, "grad_norm": 28.084447860717773, "learning_rate": 7.80261685032345e-06, "loss": 0.4313, "step": 46590 }, { "epoch": 46.62, "grad_norm": 101.02734375, "learning_rate": 7.800014623117858e-06, "loss": 0.4529, "step": 46620 }, { "epoch": 46.65, "grad_norm": 93.55826568603516, "learning_rate": 7.797411290510836e-06, "loss": 0.3587, "step": 46650 }, { "epoch": 46.68, "grad_norm": 36.82004928588867, "learning_rate": 7.794806853530139e-06, "loss": 0.338, "step": 46680 }, { "epoch": 46.71, "grad_norm": 0.0009502794127911329, "learning_rate": 7.792201313203957e-06, "loss": 0.4283, "step": 46710 }, { "epoch": 46.74, "grad_norm": 91.48057556152344, "learning_rate": 7.789594670560917e-06, "loss": 0.4142, "step": 46740 }, { "epoch": 46.77, "grad_norm": 117.46319580078125, "learning_rate": 7.786986926630079e-06, "loss": 0.4367, "step": 46770 }, { "epoch": 46.8, "grad_norm": 62.912784576416016, "learning_rate": 7.78437808244094e-06, "loss": 0.4494, "step": 46800 }, { "epoch": 46.83, "grad_norm": 41.32659912109375, "learning_rate": 7.781768139023431e-06, "loss": 0.3784, "step": 46830 }, { "epoch": 46.86, "grad_norm": 52.3610954284668, "learning_rate": 7.779157097407914e-06, "loss": 0.5707, "step": 46860 }, { "epoch": 46.89, "grad_norm": 61.48744583129883, "learning_rate": 7.776544958625189e-06, "loss": 0.4237, "step": 46890 }, { "epoch": 46.92, "grad_norm": 34.570987701416016, "learning_rate": 7.773931723706487e-06, "loss": 0.4496, "step": 46920 }, { "epoch": 46.95, "grad_norm": 33.0290412902832, "learning_rate": 7.771317393683471e-06, "loss": 0.41, "step": 46950 }, { "epoch": 46.98, "grad_norm": 0.00018745580746326596, "learning_rate": 7.768701969588237e-06, "loss": 0.3627, "step": 46980 }, { "epoch": 47.0, "eval_loss": 0.3180742859840393, "eval_map": 0.7674, "eval_map_50": 0.9546, "eval_map_75": 0.9086, "eval_map_chicken": 0.7595, "eval_map_duck": 0.7131, "eval_map_large": 0.7736, "eval_map_medium": 0.7583, "eval_map_plant": 0.8297, "eval_map_small": 0.2258, "eval_mar_1": 0.3161, "eval_mar_10": 0.8091, "eval_mar_100": 0.8149, "eval_mar_100_chicken": 0.8161, "eval_mar_100_duck": 0.7577, "eval_mar_100_plant": 0.8709, "eval_mar_large": 0.8225, "eval_mar_medium": 0.8108, "eval_mar_small": 0.3537, "eval_runtime": 12.6491, "eval_samples_per_second": 15.811, "eval_steps_per_second": 1.976, "step": 47000 }, { "epoch": 47.01, "grad_norm": 99.37277221679688, "learning_rate": 7.766085452453312e-06, "loss": 0.5268, "step": 47010 }, { "epoch": 47.04, "grad_norm": 35.703369140625, "learning_rate": 7.763467843311658e-06, "loss": 0.5442, "step": 47040 }, { "epoch": 47.07, "grad_norm": 40.959739685058594, "learning_rate": 7.760849143196664e-06, "loss": 0.3296, "step": 47070 }, { "epoch": 47.1, "grad_norm": 42.68363952636719, "learning_rate": 7.758229353142153e-06, "loss": 0.5427, "step": 47100 }, { "epoch": 47.13, "grad_norm": 86.46572875976562, "learning_rate": 7.755608474182372e-06, "loss": 0.4284, "step": 47130 }, { "epoch": 47.16, "grad_norm": 96.36883544921875, "learning_rate": 7.752986507352009e-06, "loss": 0.3792, "step": 47160 }, { "epoch": 47.19, "grad_norm": 70.56456756591797, "learning_rate": 7.75036345368617e-06, "loss": 0.3728, "step": 47190 }, { "epoch": 47.22, "grad_norm": 5.6936936744023114e-05, "learning_rate": 7.747739314220398e-06, "loss": 0.3817, "step": 47220 }, { "epoch": 47.25, "grad_norm": 41.59785079956055, "learning_rate": 7.74511408999066e-06, "loss": 0.4257, "step": 47250 }, { "epoch": 47.28, "grad_norm": 55.534244537353516, "learning_rate": 7.742487782033352e-06, "loss": 0.4528, "step": 47280 }, { "epoch": 47.31, "grad_norm": 94.58809661865234, "learning_rate": 7.739860391385303e-06, "loss": 0.4372, "step": 47310 }, { "epoch": 47.34, "grad_norm": 52.42337417602539, "learning_rate": 7.737231919083761e-06, "loss": 0.378, "step": 47340 }, { "epoch": 47.37, "grad_norm": 67.88436126708984, "learning_rate": 7.734602366166406e-06, "loss": 0.3503, "step": 47370 }, { "epoch": 47.4, "grad_norm": 38.84900665283203, "learning_rate": 7.731971733671347e-06, "loss": 0.3742, "step": 47400 }, { "epoch": 47.43, "grad_norm": 51.55014419555664, "learning_rate": 7.729340022637111e-06, "loss": 0.3587, "step": 47430 }, { "epoch": 47.46, "grad_norm": 65.75831604003906, "learning_rate": 7.72670723410266e-06, "loss": 0.4628, "step": 47460 }, { "epoch": 47.49, "grad_norm": 81.84741973876953, "learning_rate": 7.724073369107376e-06, "loss": 0.3408, "step": 47490 }, { "epoch": 47.52, "grad_norm": 51.145233154296875, "learning_rate": 7.721438428691065e-06, "loss": 0.3766, "step": 47520 }, { "epoch": 47.55, "grad_norm": 43.707401275634766, "learning_rate": 7.718802413893963e-06, "loss": 0.4806, "step": 47550 }, { "epoch": 47.58, "grad_norm": 68.40950012207031, "learning_rate": 7.716165325756727e-06, "loss": 0.3734, "step": 47580 }, { "epoch": 47.61, "grad_norm": 30.974340438842773, "learning_rate": 7.713527165320432e-06, "loss": 0.3683, "step": 47610 }, { "epoch": 47.64, "grad_norm": 120.51797485351562, "learning_rate": 7.71088793362659e-06, "loss": 0.3734, "step": 47640 }, { "epoch": 47.67, "grad_norm": 62.80084991455078, "learning_rate": 7.708247631717122e-06, "loss": 0.4061, "step": 47670 }, { "epoch": 47.7, "grad_norm": 49.4071159362793, "learning_rate": 7.70560626063438e-06, "loss": 0.3617, "step": 47700 }, { "epoch": 47.73, "grad_norm": 81.13092041015625, "learning_rate": 7.702963821421135e-06, "loss": 0.4147, "step": 47730 }, { "epoch": 47.76, "grad_norm": 42.69097900390625, "learning_rate": 7.70032031512058e-06, "loss": 0.3717, "step": 47760 }, { "epoch": 47.79, "grad_norm": 80.27191925048828, "learning_rate": 7.69767574277633e-06, "loss": 0.4091, "step": 47790 }, { "epoch": 47.82, "grad_norm": 47.82072067260742, "learning_rate": 7.695030105432417e-06, "loss": 0.3542, "step": 47820 }, { "epoch": 47.85, "grad_norm": 25.40854263305664, "learning_rate": 7.692383404133302e-06, "loss": 0.3485, "step": 47850 }, { "epoch": 47.88, "grad_norm": 68.59925079345703, "learning_rate": 7.689735639923857e-06, "loss": 0.3767, "step": 47880 }, { "epoch": 47.91, "grad_norm": 51.71120071411133, "learning_rate": 7.687086813849378e-06, "loss": 0.3727, "step": 47910 }, { "epoch": 47.94, "grad_norm": 64.91962432861328, "learning_rate": 7.684436926955584e-06, "loss": 0.4365, "step": 47940 }, { "epoch": 47.97, "grad_norm": 1.568206243973691e-05, "learning_rate": 7.681785980288601e-06, "loss": 0.2885, "step": 47970 }, { "epoch": 48.0, "grad_norm": 53.20429229736328, "learning_rate": 7.679133974894984e-06, "loss": 0.4146, "step": 48000 }, { "epoch": 48.0, "eval_loss": 0.30724024772644043, "eval_map": 0.7807, "eval_map_50": 0.9641, "eval_map_75": 0.9209, "eval_map_chicken": 0.7758, "eval_map_duck": 0.7334, "eval_map_large": 0.8023, "eval_map_medium": 0.7662, "eval_map_plant": 0.833, "eval_map_small": 0.2599, "eval_mar_1": 0.3178, "eval_mar_10": 0.8212, "eval_mar_100": 0.8257, "eval_mar_100_chicken": 0.8249, "eval_mar_100_duck": 0.7758, "eval_mar_100_plant": 0.8763, "eval_mar_large": 0.8447, "eval_mar_medium": 0.8184, "eval_mar_small": 0.3962, "eval_runtime": 12.6622, "eval_samples_per_second": 15.795, "eval_steps_per_second": 1.974, "step": 48000 }, { "epoch": 48.03, "grad_norm": 153.2679901123047, "learning_rate": 7.676480911821705e-06, "loss": 0.3487, "step": 48030 }, { "epoch": 48.06, "grad_norm": 28.18404769897461, "learning_rate": 7.673826792116146e-06, "loss": 0.3477, "step": 48060 }, { "epoch": 48.09, "grad_norm": 43.390174865722656, "learning_rate": 7.671171616826117e-06, "loss": 0.4749, "step": 48090 }, { "epoch": 48.12, "grad_norm": 74.94082641601562, "learning_rate": 7.668515386999837e-06, "loss": 0.3794, "step": 48120 }, { "epoch": 48.15, "grad_norm": 46.8407096862793, "learning_rate": 7.665858103685944e-06, "loss": 0.3173, "step": 48150 }, { "epoch": 48.18, "grad_norm": 29.187829971313477, "learning_rate": 7.663199767933489e-06, "loss": 0.3721, "step": 48180 }, { "epoch": 48.21, "grad_norm": 47.895633697509766, "learning_rate": 7.660540380791944e-06, "loss": 0.3952, "step": 48210 }, { "epoch": 48.24, "grad_norm": 105.55213165283203, "learning_rate": 7.65787994331119e-06, "loss": 0.4589, "step": 48240 }, { "epoch": 48.27, "grad_norm": 51.42699432373047, "learning_rate": 7.65521845654153e-06, "loss": 0.4827, "step": 48270 }, { "epoch": 48.3, "grad_norm": 64.23674011230469, "learning_rate": 7.652555921533671e-06, "loss": 0.4042, "step": 48300 }, { "epoch": 48.33, "grad_norm": 74.43312072753906, "learning_rate": 7.649892339338743e-06, "loss": 0.3602, "step": 48330 }, { "epoch": 48.36, "grad_norm": 52.224735260009766, "learning_rate": 7.647227711008288e-06, "loss": 0.4304, "step": 48360 }, { "epoch": 48.39, "grad_norm": 92.09090423583984, "learning_rate": 7.644562037594254e-06, "loss": 0.3545, "step": 48390 }, { "epoch": 48.42, "grad_norm": 89.98446655273438, "learning_rate": 7.641895320149008e-06, "loss": 0.4613, "step": 48420 }, { "epoch": 48.45, "grad_norm": 47.786643981933594, "learning_rate": 7.639227559725333e-06, "loss": 0.4466, "step": 48450 }, { "epoch": 48.48, "grad_norm": 87.10670471191406, "learning_rate": 7.636558757376413e-06, "loss": 0.4509, "step": 48480 }, { "epoch": 48.51, "grad_norm": 31.337554931640625, "learning_rate": 7.63388891415585e-06, "loss": 0.3475, "step": 48510 }, { "epoch": 48.54, "grad_norm": 26.714500427246094, "learning_rate": 7.631218031117658e-06, "loss": 0.3576, "step": 48540 }, { "epoch": 48.57, "grad_norm": 5.906462683924474e-05, "learning_rate": 7.628546109316257e-06, "loss": 0.3586, "step": 48570 }, { "epoch": 48.6, "grad_norm": 36.910396575927734, "learning_rate": 7.6258731498064796e-06, "loss": 0.397, "step": 48600 }, { "epoch": 48.63, "grad_norm": 3.0802613764535636e-05, "learning_rate": 7.623199153643569e-06, "loss": 0.3359, "step": 48630 }, { "epoch": 48.66, "grad_norm": 108.86724090576172, "learning_rate": 7.620524121883175e-06, "loss": 0.3972, "step": 48660 }, { "epoch": 48.69, "grad_norm": 77.31330871582031, "learning_rate": 7.617848055581361e-06, "loss": 0.3975, "step": 48690 }, { "epoch": 48.72, "grad_norm": 60.56514358520508, "learning_rate": 7.615170955794592e-06, "loss": 0.4315, "step": 48720 }, { "epoch": 48.75, "grad_norm": 49.69939041137695, "learning_rate": 7.612492823579744e-06, "loss": 0.3491, "step": 48750 }, { "epoch": 48.78, "grad_norm": 104.49805450439453, "learning_rate": 7.609813659994107e-06, "loss": 0.3565, "step": 48780 }, { "epoch": 48.81, "grad_norm": 174.05540466308594, "learning_rate": 7.607133466095365e-06, "loss": 0.4205, "step": 48810 }, { "epoch": 48.84, "grad_norm": 47.94099044799805, "learning_rate": 7.604452242941622e-06, "loss": 0.4575, "step": 48840 }, { "epoch": 48.87, "grad_norm": 70.54931640625, "learning_rate": 7.60176999159138e-06, "loss": 0.4239, "step": 48870 }, { "epoch": 48.9, "grad_norm": 25.106689453125, "learning_rate": 7.5990867131035474e-06, "loss": 0.4637, "step": 48900 }, { "epoch": 48.93, "grad_norm": 85.69908905029297, "learning_rate": 7.596402408537444e-06, "loss": 0.3464, "step": 48930 }, { "epoch": 48.96, "grad_norm": 66.3030776977539, "learning_rate": 7.593717078952788e-06, "loss": 0.4193, "step": 48960 }, { "epoch": 48.99, "grad_norm": 33.1015625, "learning_rate": 7.5910307254097075e-06, "loss": 0.4113, "step": 48990 }, { "epoch": 49.0, "eval_loss": 0.3451578915119171, "eval_map": 0.74, "eval_map_50": 0.9649, "eval_map_75": 0.8877, "eval_map_chicken": 0.7012, "eval_map_duck": 0.6958, "eval_map_large": 0.7372, "eval_map_medium": 0.7318, "eval_map_plant": 0.8229, "eval_map_small": 0.258, "eval_mar_1": 0.3014, "eval_mar_10": 0.7885, "eval_mar_100": 0.793, "eval_mar_100_chicken": 0.7678, "eval_mar_100_duck": 0.7443, "eval_mar_100_plant": 0.867, "eval_mar_large": 0.7922, "eval_mar_medium": 0.784, "eval_mar_small": 0.374, "eval_runtime": 13.0357, "eval_samples_per_second": 15.342, "eval_steps_per_second": 1.918, "step": 49000 }, { "epoch": 49.02, "grad_norm": 87.51300048828125, "learning_rate": 7.588343348968728e-06, "loss": 0.4199, "step": 49020 }, { "epoch": 49.05, "grad_norm": 28.917705535888672, "learning_rate": 7.585654950690786e-06, "loss": 0.387, "step": 49050 }, { "epoch": 49.08, "grad_norm": 58.57880783081055, "learning_rate": 7.582965531637221e-06, "loss": 0.4029, "step": 49080 }, { "epoch": 49.11, "grad_norm": 59.211692810058594, "learning_rate": 7.580275092869766e-06, "loss": 0.3611, "step": 49110 }, { "epoch": 49.14, "grad_norm": 39.918216705322266, "learning_rate": 7.577583635450572e-06, "loss": 0.3691, "step": 49140 }, { "epoch": 49.17, "grad_norm": 76.52993774414062, "learning_rate": 7.574891160442179e-06, "loss": 0.3143, "step": 49170 }, { "epoch": 49.2, "grad_norm": 106.82344818115234, "learning_rate": 7.572197668907533e-06, "loss": 0.4744, "step": 49200 }, { "epoch": 49.23, "grad_norm": 63.35402297973633, "learning_rate": 7.569503161909984e-06, "loss": 0.3968, "step": 49230 }, { "epoch": 49.26, "grad_norm": 41.599246978759766, "learning_rate": 7.566807640513278e-06, "loss": 0.4295, "step": 49260 }, { "epoch": 49.29, "grad_norm": 48.082027435302734, "learning_rate": 7.564111105781568e-06, "loss": 0.4474, "step": 49290 }, { "epoch": 49.32, "grad_norm": 29.262224197387695, "learning_rate": 7.561413558779401e-06, "loss": 0.4461, "step": 49320 }, { "epoch": 49.35, "grad_norm": 84.26061248779297, "learning_rate": 7.5587150005717256e-06, "loss": 0.3918, "step": 49350 }, { "epoch": 49.38, "grad_norm": 28.13824462890625, "learning_rate": 7.55601543222389e-06, "loss": 0.4673, "step": 49380 }, { "epoch": 49.41, "grad_norm": 76.19805145263672, "learning_rate": 7.553314854801641e-06, "loss": 0.4348, "step": 49410 }, { "epoch": 49.44, "grad_norm": 39.55094909667969, "learning_rate": 7.550613269371124e-06, "loss": 0.369, "step": 49440 }, { "epoch": 49.47, "grad_norm": 52.18972396850586, "learning_rate": 7.547910676998883e-06, "loss": 0.4158, "step": 49470 }, { "epoch": 49.5, "grad_norm": 96.14642333984375, "learning_rate": 7.545207078751858e-06, "loss": 0.4497, "step": 49500 }, { "epoch": 49.53, "grad_norm": 34.0578498840332, "learning_rate": 7.542502475697385e-06, "loss": 0.4281, "step": 49530 }, { "epoch": 49.56, "grad_norm": 106.44810485839844, "learning_rate": 7.5397968689032e-06, "loss": 0.466, "step": 49560 }, { "epoch": 49.59, "grad_norm": 39.057220458984375, "learning_rate": 7.537090259437436e-06, "loss": 0.4244, "step": 49590 }, { "epoch": 49.62, "grad_norm": 32.037296295166016, "learning_rate": 7.534382648368617e-06, "loss": 0.453, "step": 49620 }, { "epoch": 49.65, "grad_norm": 67.40744018554688, "learning_rate": 7.531674036765662e-06, "loss": 0.477, "step": 49650 }, { "epoch": 49.68, "grad_norm": 36.34019470214844, "learning_rate": 7.528964425697895e-06, "loss": 0.3506, "step": 49680 }, { "epoch": 49.71, "grad_norm": 108.23668670654297, "learning_rate": 7.526253816235023e-06, "loss": 0.5657, "step": 49710 }, { "epoch": 49.74, "grad_norm": 73.4874496459961, "learning_rate": 7.523542209447152e-06, "loss": 0.3404, "step": 49740 }, { "epoch": 49.77, "grad_norm": 22.286128997802734, "learning_rate": 7.520829606404781e-06, "loss": 0.3387, "step": 49770 }, { "epoch": 49.8, "grad_norm": 33.04328536987305, "learning_rate": 7.518116008178805e-06, "loss": 0.3878, "step": 49800 }, { "epoch": 49.83, "grad_norm": 85.5217514038086, "learning_rate": 7.515401415840509e-06, "loss": 0.3538, "step": 49830 }, { "epoch": 49.86, "grad_norm": 26.87479019165039, "learning_rate": 7.512685830461568e-06, "loss": 0.4252, "step": 49860 }, { "epoch": 49.89, "grad_norm": 70.82630920410156, "learning_rate": 7.509969253114056e-06, "loss": 0.3783, "step": 49890 }, { "epoch": 49.92, "grad_norm": 88.36101531982422, "learning_rate": 7.507251684870433e-06, "loss": 0.362, "step": 49920 }, { "epoch": 49.95, "grad_norm": 59.363792419433594, "learning_rate": 7.5045331268035505e-06, "loss": 0.388, "step": 49950 }, { "epoch": 49.98, "grad_norm": 57.58979034423828, "learning_rate": 7.501813579986655e-06, "loss": 0.4146, "step": 49980 }, { "epoch": 50.0, "eval_loss": 0.30914121866226196, "eval_map": 0.7822, "eval_map_50": 0.9588, "eval_map_75": 0.9016, "eval_map_chicken": 0.772, "eval_map_duck": 0.7344, "eval_map_large": 0.7786, "eval_map_medium": 0.7753, "eval_map_plant": 0.8404, "eval_map_small": 0.2296, "eval_mar_1": 0.3203, "eval_mar_10": 0.82, "eval_mar_100": 0.8255, "eval_mar_100_chicken": 0.8284, "eval_mar_100_duck": 0.7716, "eval_mar_100_plant": 0.8766, "eval_mar_large": 0.8262, "eval_mar_medium": 0.8178, "eval_mar_small": 0.3581, "eval_runtime": 12.2733, "eval_samples_per_second": 16.296, "eval_steps_per_second": 2.037, "step": 50000 }, { "epoch": 50.01, "grad_norm": 40.59294891357422, "learning_rate": 7.49909304549338e-06, "loss": 0.4619, "step": 50010 }, { "epoch": 50.04, "grad_norm": 0.00033072647056542337, "learning_rate": 7.496371524397747e-06, "loss": 0.4724, "step": 50040 }, { "epoch": 50.07, "grad_norm": 31.850860595703125, "learning_rate": 7.493649017774172e-06, "loss": 0.3991, "step": 50070 }, { "epoch": 50.1, "grad_norm": 88.05603790283203, "learning_rate": 7.490925526697455e-06, "loss": 0.3711, "step": 50100 }, { "epoch": 50.13, "grad_norm": 84.62833404541016, "learning_rate": 7.488201052242791e-06, "loss": 0.3675, "step": 50130 }, { "epoch": 50.16, "grad_norm": 91.15155792236328, "learning_rate": 7.485475595485756e-06, "loss": 0.3566, "step": 50160 }, { "epoch": 50.19, "grad_norm": 41.35752868652344, "learning_rate": 7.482749157502317e-06, "loss": 0.3916, "step": 50190 }, { "epoch": 50.22, "grad_norm": 38.0494499206543, "learning_rate": 7.480021739368831e-06, "loss": 0.4252, "step": 50220 }, { "epoch": 50.25, "grad_norm": 753.4015502929688, "learning_rate": 7.477293342162038e-06, "loss": 0.3869, "step": 50250 }, { "epoch": 50.28, "grad_norm": 49.08755111694336, "learning_rate": 7.474563966959068e-06, "loss": 0.3506, "step": 50280 }, { "epoch": 50.31, "grad_norm": 135.30055236816406, "learning_rate": 7.471833614837431e-06, "loss": 0.3876, "step": 50310 }, { "epoch": 50.34, "grad_norm": 53.00099563598633, "learning_rate": 7.469102286875029e-06, "loss": 0.4631, "step": 50340 }, { "epoch": 50.37, "grad_norm": 73.98764038085938, "learning_rate": 7.46636998415015e-06, "loss": 0.35, "step": 50370 }, { "epoch": 50.4, "grad_norm": 49.32426071166992, "learning_rate": 7.463636707741458e-06, "loss": 0.4051, "step": 50400 }, { "epoch": 50.43, "grad_norm": 80.86725616455078, "learning_rate": 7.460902458728012e-06, "loss": 0.4064, "step": 50430 }, { "epoch": 50.46, "grad_norm": 54.13125228881836, "learning_rate": 7.458167238189249e-06, "loss": 0.3722, "step": 50460 }, { "epoch": 50.49, "grad_norm": 90.4913101196289, "learning_rate": 7.455431047204988e-06, "loss": 0.3217, "step": 50490 }, { "epoch": 50.52, "grad_norm": 92.44465637207031, "learning_rate": 7.452693886855438e-06, "loss": 0.3993, "step": 50520 }, { "epoch": 50.55, "grad_norm": 51.794986724853516, "learning_rate": 7.449955758221184e-06, "loss": 0.4271, "step": 50550 }, { "epoch": 50.58, "grad_norm": 31.278583526611328, "learning_rate": 7.447216662383196e-06, "loss": 0.3627, "step": 50580 }, { "epoch": 50.61, "grad_norm": 55.533809661865234, "learning_rate": 7.444476600422827e-06, "loss": 0.3269, "step": 50610 }, { "epoch": 50.64, "grad_norm": 30.269277572631836, "learning_rate": 7.4417355734218085e-06, "loss": 0.4043, "step": 50640 }, { "epoch": 50.67, "grad_norm": 82.75839233398438, "learning_rate": 7.438993582462255e-06, "loss": 0.3347, "step": 50670 }, { "epoch": 50.7, "grad_norm": 40.582733154296875, "learning_rate": 7.436250628626662e-06, "loss": 0.3469, "step": 50700 }, { "epoch": 50.73, "grad_norm": 24.801115036010742, "learning_rate": 7.433506712997903e-06, "loss": 0.4653, "step": 50730 }, { "epoch": 50.76, "grad_norm": 63.53864669799805, "learning_rate": 7.430761836659235e-06, "loss": 0.4771, "step": 50760 }, { "epoch": 50.79, "grad_norm": 49.86505889892578, "learning_rate": 7.428016000694287e-06, "loss": 0.4687, "step": 50790 }, { "epoch": 50.82, "grad_norm": 66.61412048339844, "learning_rate": 7.425269206187076e-06, "loss": 0.362, "step": 50820 }, { "epoch": 50.85, "grad_norm": 36.05592727661133, "learning_rate": 7.42252145422199e-06, "loss": 0.408, "step": 50850 }, { "epoch": 50.88, "grad_norm": 41.92254638671875, "learning_rate": 7.4197727458837995e-06, "loss": 0.4124, "step": 50880 }, { "epoch": 50.91, "grad_norm": 42.371673583984375, "learning_rate": 7.417023082257653e-06, "loss": 0.3391, "step": 50910 }, { "epoch": 50.94, "grad_norm": 2.2105425159679726e-05, "learning_rate": 7.414272464429068e-06, "loss": 0.4436, "step": 50940 }, { "epoch": 50.97, "grad_norm": 49.759368896484375, "learning_rate": 7.411520893483952e-06, "loss": 0.4299, "step": 50970 }, { "epoch": 51.0, "grad_norm": 178.89427185058594, "learning_rate": 7.408768370508577e-06, "loss": 0.3777, "step": 51000 }, { "epoch": 51.0, "eval_loss": 0.31677013635635376, "eval_map": 0.7772, "eval_map_50": 0.9559, "eval_map_75": 0.9072, "eval_map_chicken": 0.7685, "eval_map_duck": 0.7287, "eval_map_large": 0.7844, "eval_map_medium": 0.7684, "eval_map_plant": 0.8343, "eval_map_small": 0.1974, "eval_mar_1": 0.3203, "eval_mar_10": 0.8163, "eval_mar_100": 0.8209, "eval_mar_100_chicken": 0.8183, "eval_mar_100_duck": 0.7696, "eval_mar_100_plant": 0.8747, "eval_mar_large": 0.8346, "eval_mar_medium": 0.816, "eval_mar_small": 0.2859, "eval_runtime": 12.5905, "eval_samples_per_second": 15.885, "eval_steps_per_second": 1.986, "step": 51000 }, { "epoch": 51.03, "grad_norm": 112.15713500976562, "learning_rate": 7.406014896589597e-06, "loss": 0.4532, "step": 51030 }, { "epoch": 51.06, "grad_norm": 125.4044418334961, "learning_rate": 7.40326047281404e-06, "loss": 0.4626, "step": 51060 }, { "epoch": 51.09, "grad_norm": 56.685691833496094, "learning_rate": 7.400505100269307e-06, "loss": 0.3872, "step": 51090 }, { "epoch": 51.12, "grad_norm": 47.959293365478516, "learning_rate": 7.397748780043179e-06, "loss": 0.428, "step": 51120 }, { "epoch": 51.15, "grad_norm": 45.2544059753418, "learning_rate": 7.394991513223806e-06, "loss": 0.4722, "step": 51150 }, { "epoch": 51.18, "grad_norm": 34.166996002197266, "learning_rate": 7.392233300899712e-06, "loss": 0.438, "step": 51180 }, { "epoch": 51.21, "grad_norm": 74.33020782470703, "learning_rate": 7.3894741441597965e-06, "loss": 0.3785, "step": 51210 }, { "epoch": 51.24, "grad_norm": 47.328861236572266, "learning_rate": 7.386714044093331e-06, "loss": 0.4709, "step": 51240 }, { "epoch": 51.27, "grad_norm": 38.92949676513672, "learning_rate": 7.38395300178996e-06, "loss": 0.4007, "step": 51270 }, { "epoch": 51.3, "grad_norm": 36.61409378051758, "learning_rate": 7.381191018339697e-06, "loss": 0.3344, "step": 51300 }, { "epoch": 51.33, "grad_norm": 181.11544799804688, "learning_rate": 7.3784280948329315e-06, "loss": 0.39, "step": 51330 }, { "epoch": 51.36, "grad_norm": 58.06565856933594, "learning_rate": 7.375664232360421e-06, "loss": 0.4032, "step": 51360 }, { "epoch": 51.39, "grad_norm": 64.25209045410156, "learning_rate": 7.372899432013294e-06, "loss": 0.3751, "step": 51390 }, { "epoch": 51.42, "grad_norm": 59.21888732910156, "learning_rate": 7.370133694883052e-06, "loss": 0.4723, "step": 51420 }, { "epoch": 51.45, "grad_norm": 43.54585266113281, "learning_rate": 7.3673670220615615e-06, "loss": 0.3143, "step": 51450 }, { "epoch": 51.48, "grad_norm": 39.96760177612305, "learning_rate": 7.364599414641064e-06, "loss": 0.3449, "step": 51480 }, { "epoch": 51.51, "grad_norm": 44.0994873046875, "learning_rate": 7.361830873714165e-06, "loss": 0.3896, "step": 51510 }, { "epoch": 51.54, "grad_norm": 44.684547424316406, "learning_rate": 7.359061400373841e-06, "loss": 0.3971, "step": 51540 }, { "epoch": 51.57, "grad_norm": 60.59151840209961, "learning_rate": 7.356290995713436e-06, "loss": 0.3111, "step": 51570 }, { "epoch": 51.6, "grad_norm": 83.9949951171875, "learning_rate": 7.353519660826665e-06, "loss": 0.4359, "step": 51600 }, { "epoch": 51.63, "grad_norm": 64.61292266845703, "learning_rate": 7.350747396807601e-06, "loss": 0.3131, "step": 51630 }, { "epoch": 51.66, "grad_norm": 62.04275131225586, "learning_rate": 7.3479742047506955e-06, "loss": 0.3814, "step": 51660 }, { "epoch": 51.69, "grad_norm": 58.457698822021484, "learning_rate": 7.345200085750758e-06, "loss": 0.3951, "step": 51690 }, { "epoch": 51.72, "grad_norm": 41.96393966674805, "learning_rate": 7.342425040902967e-06, "loss": 0.4349, "step": 51720 }, { "epoch": 51.75, "grad_norm": 73.87506866455078, "learning_rate": 7.3396490713028674e-06, "loss": 0.3584, "step": 51750 }, { "epoch": 51.78, "grad_norm": 71.14085388183594, "learning_rate": 7.336872178046368e-06, "loss": 0.4355, "step": 51780 }, { "epoch": 51.81, "grad_norm": 62.70282745361328, "learning_rate": 7.334094362229741e-06, "loss": 0.4586, "step": 51810 }, { "epoch": 51.84, "grad_norm": 96.15145874023438, "learning_rate": 7.331315624949624e-06, "loss": 0.3704, "step": 51840 }, { "epoch": 51.87, "grad_norm": 130.48281860351562, "learning_rate": 7.32853596730302e-06, "loss": 0.3627, "step": 51870 }, { "epoch": 51.9, "grad_norm": 86.07929229736328, "learning_rate": 7.325755390387293e-06, "loss": 0.3912, "step": 51900 }, { "epoch": 51.93, "grad_norm": 29.82171058654785, "learning_rate": 7.32297389530017e-06, "loss": 0.4131, "step": 51930 }, { "epoch": 51.96, "grad_norm": 87.52727508544922, "learning_rate": 7.320191483139742e-06, "loss": 0.3911, "step": 51960 }, { "epoch": 51.99, "grad_norm": 229.53866577148438, "learning_rate": 7.317408155004462e-06, "loss": 0.3417, "step": 51990 }, { "epoch": 52.0, "eval_loss": 0.31345629692077637, "eval_map": 0.7745, "eval_map_50": 0.958, "eval_map_75": 0.9068, "eval_map_chicken": 0.7776, "eval_map_duck": 0.7113, "eval_map_large": 0.804, "eval_map_medium": 0.7548, "eval_map_plant": 0.8347, "eval_map_small": 0.2364, "eval_mar_1": 0.3164, "eval_mar_10": 0.8137, "eval_mar_100": 0.8183, "eval_mar_100_chicken": 0.828, "eval_mar_100_duck": 0.7515, "eval_mar_100_plant": 0.8754, "eval_mar_large": 0.8433, "eval_mar_medium": 0.8071, "eval_mar_small": 0.3284, "eval_runtime": 13.0387, "eval_samples_per_second": 15.339, "eval_steps_per_second": 1.917, "step": 52000 }, { "epoch": 52.02, "grad_norm": 155.61407470703125, "learning_rate": 7.314623911993143e-06, "loss": 0.3891, "step": 52020 }, { "epoch": 52.05, "grad_norm": 29.58070182800293, "learning_rate": 7.31183875520496e-06, "loss": 0.3628, "step": 52050 }, { "epoch": 52.08, "grad_norm": 98.71990203857422, "learning_rate": 7.309052685739448e-06, "loss": 0.3694, "step": 52080 }, { "epoch": 52.11, "grad_norm": 99.89366149902344, "learning_rate": 7.306265704696505e-06, "loss": 0.4626, "step": 52110 }, { "epoch": 52.14, "grad_norm": 33.60049819946289, "learning_rate": 7.303477813176385e-06, "loss": 0.3561, "step": 52140 }, { "epoch": 52.17, "grad_norm": 38.70956802368164, "learning_rate": 7.300689012279706e-06, "loss": 0.3767, "step": 52170 }, { "epoch": 52.2, "grad_norm": 152.68310546875, "learning_rate": 7.297899303107441e-06, "loss": 0.3753, "step": 52200 }, { "epoch": 52.23, "grad_norm": 47.59980010986328, "learning_rate": 7.295108686760921e-06, "loss": 0.3657, "step": 52230 }, { "epoch": 52.26, "grad_norm": 48.4316291809082, "learning_rate": 7.29231716434184e-06, "loss": 0.382, "step": 52260 }, { "epoch": 52.29, "grad_norm": 47.350948333740234, "learning_rate": 7.289524736952245e-06, "loss": 0.3713, "step": 52290 }, { "epoch": 52.32, "grad_norm": 57.174957275390625, "learning_rate": 7.286731405694544e-06, "loss": 0.3639, "step": 52320 }, { "epoch": 52.35, "grad_norm": 73.67964935302734, "learning_rate": 7.283937171671498e-06, "loss": 0.3803, "step": 52350 }, { "epoch": 52.38, "grad_norm": 41.79163360595703, "learning_rate": 7.281142035986227e-06, "loss": 0.3643, "step": 52380 }, { "epoch": 52.41, "grad_norm": 55.32611846923828, "learning_rate": 7.2783459997422075e-06, "loss": 0.3725, "step": 52410 }, { "epoch": 52.44, "grad_norm": 116.00684356689453, "learning_rate": 7.275549064043269e-06, "loss": 0.3334, "step": 52440 }, { "epoch": 52.47, "grad_norm": 41.66463851928711, "learning_rate": 7.272751229993598e-06, "loss": 0.4096, "step": 52470 }, { "epoch": 52.5, "grad_norm": 52.13127517700195, "learning_rate": 7.269952498697734e-06, "loss": 0.3695, "step": 52500 }, { "epoch": 52.53, "grad_norm": 44.19420623779297, "learning_rate": 7.267152871260573e-06, "loss": 0.4407, "step": 52530 }, { "epoch": 52.56, "grad_norm": 281.878173828125, "learning_rate": 7.264352348787364e-06, "loss": 0.4397, "step": 52560 }, { "epoch": 52.59, "grad_norm": 81.37568664550781, "learning_rate": 7.261550932383707e-06, "loss": 0.4534, "step": 52590 }, { "epoch": 52.62, "grad_norm": 42.85062026977539, "learning_rate": 7.258748623155558e-06, "loss": 0.3291, "step": 52620 }, { "epoch": 52.65, "grad_norm": 80.00833892822266, "learning_rate": 7.2559454222092265e-06, "loss": 0.366, "step": 52650 }, { "epoch": 52.68, "grad_norm": 82.74235534667969, "learning_rate": 7.253141330651367e-06, "loss": 0.642, "step": 52680 }, { "epoch": 52.71, "grad_norm": 37.78058624267578, "learning_rate": 7.250336349588995e-06, "loss": 0.4131, "step": 52710 }, { "epoch": 52.74, "grad_norm": 118.4503173828125, "learning_rate": 7.247530480129469e-06, "loss": 0.3458, "step": 52740 }, { "epoch": 52.77, "grad_norm": 77.77617645263672, "learning_rate": 7.244723723380504e-06, "loss": 0.3451, "step": 52770 }, { "epoch": 52.8, "grad_norm": 106.86019897460938, "learning_rate": 7.241916080450163e-06, "loss": 0.327, "step": 52800 }, { "epoch": 52.83, "grad_norm": 41.23399353027344, "learning_rate": 7.239107552446858e-06, "loss": 0.3868, "step": 52830 }, { "epoch": 52.86, "grad_norm": 435.7214660644531, "learning_rate": 7.236298140479352e-06, "loss": 0.3449, "step": 52860 }, { "epoch": 52.89, "grad_norm": 39.18952178955078, "learning_rate": 7.233487845656755e-06, "loss": 0.3582, "step": 52890 }, { "epoch": 52.92, "grad_norm": 51.28075408935547, "learning_rate": 7.23067666908853e-06, "loss": 0.4283, "step": 52920 }, { "epoch": 52.95, "grad_norm": 43.4094352722168, "learning_rate": 7.227864611884483e-06, "loss": 0.3736, "step": 52950 }, { "epoch": 52.98, "grad_norm": 51.48238754272461, "learning_rate": 7.225051675154768e-06, "loss": 0.4088, "step": 52980 }, { "epoch": 53.0, "eval_loss": 0.3144761323928833, "eval_map": 0.7689, "eval_map_50": 0.9569, "eval_map_75": 0.9183, "eval_map_chicken": 0.7593, "eval_map_duck": 0.7116, "eval_map_large": 0.7631, "eval_map_medium": 0.7593, "eval_map_plant": 0.8358, "eval_map_small": 0.2712, "eval_mar_1": 0.3134, "eval_mar_10": 0.8108, "eval_mar_100": 0.8147, "eval_mar_100_chicken": 0.8117, "eval_mar_100_duck": 0.7557, "eval_mar_100_plant": 0.8767, "eval_mar_large": 0.8093, "eval_mar_medium": 0.8093, "eval_mar_small": 0.3583, "eval_runtime": 12.5734, "eval_samples_per_second": 15.907, "eval_steps_per_second": 1.988, "step": 53000 }, { "epoch": 53.01, "grad_norm": 55.66585922241211, "learning_rate": 7.222237860009892e-06, "loss": 0.3748, "step": 53010 }, { "epoch": 53.04, "grad_norm": 117.66275024414062, "learning_rate": 7.219423167560701e-06, "loss": 0.3111, "step": 53040 }, { "epoch": 53.07, "grad_norm": 138.09678649902344, "learning_rate": 7.216607598918392e-06, "loss": 0.4399, "step": 53070 }, { "epoch": 53.1, "grad_norm": 37.152889251708984, "learning_rate": 7.21379115519451e-06, "loss": 0.4433, "step": 53100 }, { "epoch": 53.13, "grad_norm": 55.792076110839844, "learning_rate": 7.210973837500936e-06, "loss": 0.3808, "step": 53130 }, { "epoch": 53.16, "grad_norm": 55.7307243347168, "learning_rate": 7.208155646949908e-06, "loss": 0.3803, "step": 53160 }, { "epoch": 53.19, "grad_norm": 107.5602798461914, "learning_rate": 7.205336584653999e-06, "loss": 0.3533, "step": 53190 }, { "epoch": 53.22, "grad_norm": 44.32465744018555, "learning_rate": 7.202516651726135e-06, "loss": 0.4278, "step": 53220 }, { "epoch": 53.25, "grad_norm": 5.8813850046135485e-05, "learning_rate": 7.199695849279576e-06, "loss": 0.3573, "step": 53250 }, { "epoch": 53.28, "grad_norm": 62.56598663330078, "learning_rate": 7.196874178427933e-06, "loss": 0.4694, "step": 53280 }, { "epoch": 53.31, "grad_norm": 45.35331344604492, "learning_rate": 7.194051640285156e-06, "loss": 0.4736, "step": 53310 }, { "epoch": 53.34, "grad_norm": 50.7871208190918, "learning_rate": 7.191228235965539e-06, "loss": 0.3773, "step": 53340 }, { "epoch": 53.37, "grad_norm": 69.21955871582031, "learning_rate": 7.1884039665837165e-06, "loss": 0.427, "step": 53370 }, { "epoch": 53.4, "grad_norm": 61.52739715576172, "learning_rate": 7.185578833254665e-06, "loss": 0.4133, "step": 53400 }, { "epoch": 53.43, "grad_norm": 270.9831848144531, "learning_rate": 7.182752837093704e-06, "loss": 0.4358, "step": 53430 }, { "epoch": 53.46, "grad_norm": 51.44628143310547, "learning_rate": 7.1799259792164914e-06, "loss": 0.4053, "step": 53460 }, { "epoch": 53.49, "grad_norm": 55.884986877441406, "learning_rate": 7.177098260739024e-06, "loss": 0.3463, "step": 53490 }, { "epoch": 53.52, "grad_norm": 30.99408721923828, "learning_rate": 7.1742696827776415e-06, "loss": 0.49, "step": 53520 }, { "epoch": 53.55, "grad_norm": 85.52278137207031, "learning_rate": 7.171440246449024e-06, "loss": 0.4012, "step": 53550 }, { "epoch": 53.58, "grad_norm": 66.05082702636719, "learning_rate": 7.168609952870185e-06, "loss": 0.3938, "step": 53580 }, { "epoch": 53.61, "grad_norm": 0.0005565531319007277, "learning_rate": 7.165778803158481e-06, "loss": 0.5095, "step": 53610 }, { "epoch": 53.64, "grad_norm": 50.68171691894531, "learning_rate": 7.162946798431605e-06, "loss": 0.4264, "step": 53640 }, { "epoch": 53.67, "grad_norm": 134.61260986328125, "learning_rate": 7.160113939807587e-06, "loss": 0.4, "step": 53670 }, { "epoch": 53.7, "grad_norm": 36.757266998291016, "learning_rate": 7.157280228404796e-06, "loss": 0.3682, "step": 53700 }, { "epoch": 53.73, "grad_norm": 51.727447509765625, "learning_rate": 7.154445665341933e-06, "loss": 0.4057, "step": 53730 }, { "epoch": 53.76, "grad_norm": 92.77101135253906, "learning_rate": 7.151610251738045e-06, "loss": 0.3698, "step": 53760 }, { "epoch": 53.79, "grad_norm": 189.5593719482422, "learning_rate": 7.148773988712503e-06, "loss": 0.3646, "step": 53790 }, { "epoch": 53.82, "grad_norm": 33.99925994873047, "learning_rate": 7.145936877385019e-06, "loss": 0.407, "step": 53820 }, { "epoch": 53.85, "grad_norm": 66.19991302490234, "learning_rate": 7.143098918875643e-06, "loss": 0.4137, "step": 53850 }, { "epoch": 53.88, "grad_norm": 48.91218566894531, "learning_rate": 7.1402601143047514e-06, "loss": 0.3482, "step": 53880 }, { "epoch": 53.91, "grad_norm": 8.500175317749381e-05, "learning_rate": 7.1374204647930636e-06, "loss": 0.3469, "step": 53910 }, { "epoch": 53.94, "grad_norm": 135.4546356201172, "learning_rate": 7.134579971461627e-06, "loss": 0.3192, "step": 53940 }, { "epoch": 53.97, "grad_norm": 0.4192346930503845, "learning_rate": 7.131738635431822e-06, "loss": 0.4042, "step": 53970 }, { "epoch": 54.0, "grad_norm": 25.636377334594727, "learning_rate": 7.128896457825364e-06, "loss": 0.4384, "step": 54000 }, { "epoch": 54.0, "eval_loss": 0.29734450578689575, "eval_map": 0.7848, "eval_map_50": 0.9618, "eval_map_75": 0.9157, "eval_map_chicken": 0.768, "eval_map_duck": 0.7407, "eval_map_large": 0.7875, "eval_map_medium": 0.7746, "eval_map_plant": 0.8456, "eval_map_small": 0.2523, "eval_mar_1": 0.3201, "eval_mar_10": 0.8243, "eval_mar_100": 0.8284, "eval_mar_100_chicken": 0.8203, "eval_mar_100_duck": 0.782, "eval_mar_100_plant": 0.8828, "eval_mar_large": 0.8343, "eval_mar_medium": 0.8227, "eval_mar_small": 0.3324, "eval_runtime": 12.4291, "eval_samples_per_second": 16.091, "eval_steps_per_second": 2.011, "step": 54000 }, { "epoch": 54.03, "grad_norm": 34.161983489990234, "learning_rate": 7.1260534397643e-06, "loss": 0.4435, "step": 54030 }, { "epoch": 54.06, "grad_norm": 84.72724914550781, "learning_rate": 7.1232095823710064e-06, "loss": 0.3551, "step": 54060 }, { "epoch": 54.09, "grad_norm": 54.239036560058594, "learning_rate": 7.120364886768197e-06, "loss": 0.3622, "step": 54090 }, { "epoch": 54.12, "grad_norm": 198.5832977294922, "learning_rate": 7.11751935407891e-06, "loss": 0.3543, "step": 54120 }, { "epoch": 54.15, "grad_norm": 244.6286163330078, "learning_rate": 7.114672985426516e-06, "loss": 0.3766, "step": 54150 }, { "epoch": 54.18, "grad_norm": 5.054960274719633e-05, "learning_rate": 7.111825781934719e-06, "loss": 0.318, "step": 54180 }, { "epoch": 54.21, "grad_norm": 111.01826477050781, "learning_rate": 7.108977744727548e-06, "loss": 0.4159, "step": 54210 }, { "epoch": 54.24, "grad_norm": 37.02896499633789, "learning_rate": 7.106128874929364e-06, "loss": 0.4096, "step": 54240 }, { "epoch": 54.27, "grad_norm": 76.27146911621094, "learning_rate": 7.103279173664851e-06, "loss": 0.3938, "step": 54270 }, { "epoch": 54.3, "grad_norm": 129.3892364501953, "learning_rate": 7.100428642059033e-06, "loss": 0.4023, "step": 54300 }, { "epoch": 54.33, "grad_norm": 37.95668029785156, "learning_rate": 7.097577281237249e-06, "loss": 0.3675, "step": 54330 }, { "epoch": 54.36, "grad_norm": 25.897445678710938, "learning_rate": 7.094725092325177e-06, "loss": 0.3911, "step": 54360 }, { "epoch": 54.39, "grad_norm": 101.5472183227539, "learning_rate": 7.09187207644881e-06, "loss": 0.3755, "step": 54390 }, { "epoch": 54.42, "grad_norm": 60.73103713989258, "learning_rate": 7.089018234734476e-06, "loss": 0.3892, "step": 54420 }, { "epoch": 54.45, "grad_norm": 0.005314315669238567, "learning_rate": 7.086163568308828e-06, "loss": 0.367, "step": 54450 }, { "epoch": 54.48, "grad_norm": 46.79784393310547, "learning_rate": 7.08330807829884e-06, "loss": 0.3281, "step": 54480 }, { "epoch": 54.51, "grad_norm": 71.17256927490234, "learning_rate": 7.080451765831817e-06, "loss": 0.4027, "step": 54510 }, { "epoch": 54.54, "grad_norm": 109.13282012939453, "learning_rate": 7.077594632035385e-06, "loss": 0.3779, "step": 54540 }, { "epoch": 54.57, "grad_norm": 83.65421295166016, "learning_rate": 7.074736678037495e-06, "loss": 0.3283, "step": 54570 }, { "epoch": 54.6, "grad_norm": 22.043659210205078, "learning_rate": 7.071877904966422e-06, "loss": 0.4124, "step": 54600 }, { "epoch": 54.63, "grad_norm": 73.91861724853516, "learning_rate": 7.0690183139507625e-06, "loss": 0.3414, "step": 54630 }, { "epoch": 54.66, "grad_norm": 27.246055603027344, "learning_rate": 7.066157906119442e-06, "loss": 0.5438, "step": 54660 }, { "epoch": 54.69, "grad_norm": 58.45350646972656, "learning_rate": 7.063296682601701e-06, "loss": 0.4172, "step": 54690 }, { "epoch": 54.72, "grad_norm": 47.31928253173828, "learning_rate": 7.060434644527105e-06, "loss": 0.3323, "step": 54720 }, { "epoch": 54.75, "grad_norm": 42.68720245361328, "learning_rate": 7.057571793025545e-06, "loss": 0.3715, "step": 54750 }, { "epoch": 54.78, "grad_norm": 50.11931228637695, "learning_rate": 7.054708129227225e-06, "loss": 0.3465, "step": 54780 }, { "epoch": 54.81, "grad_norm": 35.947086334228516, "learning_rate": 7.051843654262675e-06, "loss": 0.3743, "step": 54810 }, { "epoch": 54.84, "grad_norm": 44.884727478027344, "learning_rate": 7.048978369262747e-06, "loss": 0.3801, "step": 54840 }, { "epoch": 54.87, "grad_norm": 91.31758880615234, "learning_rate": 7.046112275358607e-06, "loss": 0.4392, "step": 54870 }, { "epoch": 54.9, "grad_norm": 38.04963302612305, "learning_rate": 7.043245373681746e-06, "loss": 0.376, "step": 54900 }, { "epoch": 54.93, "grad_norm": 105.13047790527344, "learning_rate": 7.0403776653639685e-06, "loss": 0.3105, "step": 54930 }, { "epoch": 54.96, "grad_norm": 47.148502349853516, "learning_rate": 7.037509151537404e-06, "loss": 0.4145, "step": 54960 }, { "epoch": 54.99, "grad_norm": 64.38042449951172, "learning_rate": 7.0346398333344945e-06, "loss": 0.3848, "step": 54990 }, { "epoch": 55.0, "eval_loss": 0.3070791959762573, "eval_map": 0.7806, "eval_map_50": 0.9572, "eval_map_75": 0.911, "eval_map_chicken": 0.7773, "eval_map_duck": 0.7303, "eval_map_large": 0.7916, "eval_map_medium": 0.7711, "eval_map_plant": 0.834, "eval_map_small": 0.2747, "eval_mar_1": 0.3205, "eval_mar_10": 0.8198, "eval_mar_100": 0.8234, "eval_mar_100_chicken": 0.8288, "eval_mar_100_duck": 0.767, "eval_mar_100_plant": 0.8744, "eval_mar_large": 0.8349, "eval_mar_medium": 0.8158, "eval_mar_small": 0.3651, "eval_runtime": 12.8057, "eval_samples_per_second": 15.618, "eval_steps_per_second": 1.952, "step": 55000 }, { "epoch": 55.02, "grad_norm": 36.89799880981445, "learning_rate": 7.031769711887999e-06, "loss": 0.4116, "step": 55020 }, { "epoch": 55.05, "grad_norm": 342.57373046875, "learning_rate": 7.028898788331e-06, "loss": 0.4041, "step": 55050 }, { "epoch": 55.08, "grad_norm": 42.59968948364258, "learning_rate": 7.026027063796891e-06, "loss": 0.4116, "step": 55080 }, { "epoch": 55.11, "grad_norm": 2.9487389838322997e-05, "learning_rate": 7.023154539419383e-06, "loss": 0.4753, "step": 55110 }, { "epoch": 55.14, "grad_norm": 33.64665603637695, "learning_rate": 7.020281216332503e-06, "loss": 0.3921, "step": 55140 }, { "epoch": 55.17, "grad_norm": 31.871530532836914, "learning_rate": 7.017407095670594e-06, "loss": 0.4233, "step": 55170 }, { "epoch": 55.2, "grad_norm": 79.17060852050781, "learning_rate": 7.014532178568314e-06, "loss": 0.3651, "step": 55200 }, { "epoch": 55.23, "grad_norm": 5.0003582146018744e-05, "learning_rate": 7.011656466160633e-06, "loss": 0.3776, "step": 55230 }, { "epoch": 55.26, "grad_norm": 41.38606643676758, "learning_rate": 7.008779959582838e-06, "loss": 0.4627, "step": 55260 }, { "epoch": 55.29, "grad_norm": 93.16791534423828, "learning_rate": 7.005902659970528e-06, "loss": 0.3671, "step": 55290 }, { "epoch": 55.32, "grad_norm": 62.189720153808594, "learning_rate": 7.003024568459614e-06, "loss": 0.3306, "step": 55320 }, { "epoch": 55.35, "grad_norm": 59.500274658203125, "learning_rate": 7.0001456861863236e-06, "loss": 0.4498, "step": 55350 }, { "epoch": 55.38, "grad_norm": 68.85523223876953, "learning_rate": 6.997266014287193e-06, "loss": 0.3574, "step": 55380 }, { "epoch": 55.41, "grad_norm": 135.1203155517578, "learning_rate": 6.994385553899069e-06, "loss": 0.3605, "step": 55410 }, { "epoch": 55.44, "grad_norm": 89.37511444091797, "learning_rate": 6.991504306159115e-06, "loss": 0.3774, "step": 55440 }, { "epoch": 55.47, "grad_norm": 63.217769622802734, "learning_rate": 6.988622272204799e-06, "loss": 0.3799, "step": 55470 }, { "epoch": 55.5, "grad_norm": 41.939453125, "learning_rate": 6.985739453173903e-06, "loss": 0.392, "step": 55500 }, { "epoch": 55.53, "grad_norm": 41.55017852783203, "learning_rate": 6.982855850204521e-06, "loss": 0.3599, "step": 55530 }, { "epoch": 55.56, "grad_norm": 72.655517578125, "learning_rate": 6.9799714644350504e-06, "loss": 0.4146, "step": 55560 }, { "epoch": 55.59, "grad_norm": 75.96229553222656, "learning_rate": 6.9770862970042015e-06, "loss": 0.3561, "step": 55590 }, { "epoch": 55.62, "grad_norm": 55.06479263305664, "learning_rate": 6.974200349050996e-06, "loss": 0.3671, "step": 55620 }, { "epoch": 55.65, "grad_norm": 80.4039077758789, "learning_rate": 6.971313621714756e-06, "loss": 0.3667, "step": 55650 }, { "epoch": 55.68, "grad_norm": 60.08681869506836, "learning_rate": 6.968426116135118e-06, "loss": 0.5581, "step": 55680 }, { "epoch": 55.71, "grad_norm": 54.782081604003906, "learning_rate": 6.965537833452024e-06, "loss": 0.3902, "step": 55710 }, { "epoch": 55.74, "grad_norm": 101.82967376708984, "learning_rate": 6.96264877480572e-06, "loss": 0.438, "step": 55740 }, { "epoch": 55.77, "grad_norm": 53.0418815612793, "learning_rate": 6.959758941336762e-06, "loss": 0.4143, "step": 55770 }, { "epoch": 55.8, "grad_norm": 71.62274932861328, "learning_rate": 6.9568683341860135e-06, "loss": 0.3987, "step": 55800 }, { "epoch": 55.83, "grad_norm": 39.667816162109375, "learning_rate": 6.953976954494636e-06, "loss": 0.4203, "step": 55830 }, { "epoch": 55.86, "grad_norm": 35.82624435424805, "learning_rate": 6.9510848034041e-06, "loss": 0.5617, "step": 55860 }, { "epoch": 55.89, "grad_norm": 93.48584747314453, "learning_rate": 6.948191882056185e-06, "loss": 0.4278, "step": 55890 }, { "epoch": 55.92, "grad_norm": 54.34282302856445, "learning_rate": 6.945298191592967e-06, "loss": 0.3758, "step": 55920 }, { "epoch": 55.95, "grad_norm": 51.37112808227539, "learning_rate": 6.942403733156832e-06, "loss": 0.3773, "step": 55950 }, { "epoch": 55.98, "grad_norm": 68.00861358642578, "learning_rate": 6.939508507890464e-06, "loss": 0.4163, "step": 55980 }, { "epoch": 56.0, "eval_loss": 0.30549460649490356, "eval_map": 0.7775, "eval_map_50": 0.9568, "eval_map_75": 0.9054, "eval_map_chicken": 0.7653, "eval_map_duck": 0.7328, "eval_map_large": 0.784, "eval_map_medium": 0.7647, "eval_map_plant": 0.8343, "eval_map_small": 0.2209, "eval_mar_1": 0.3201, "eval_mar_10": 0.8194, "eval_mar_100": 0.8221, "eval_mar_100_chicken": 0.8181, "eval_mar_100_duck": 0.7789, "eval_mar_100_plant": 0.8694, "eval_mar_large": 0.8257, "eval_mar_medium": 0.8135, "eval_mar_small": 0.3228, "eval_runtime": 12.5207, "eval_samples_per_second": 15.974, "eval_steps_per_second": 1.997, "step": 56000 }, { "epoch": 56.01, "grad_norm": 51.65520477294922, "learning_rate": 6.936612516936852e-06, "loss": 0.6152, "step": 56010 }, { "epoch": 56.04, "grad_norm": 45.4213752746582, "learning_rate": 6.93371576143929e-06, "loss": 0.3318, "step": 56040 }, { "epoch": 56.07, "grad_norm": 65.28303527832031, "learning_rate": 6.930818242541368e-06, "loss": 0.3804, "step": 56070 }, { "epoch": 56.1, "grad_norm": 57.1995735168457, "learning_rate": 6.927919961386984e-06, "loss": 0.4154, "step": 56100 }, { "epoch": 56.13, "grad_norm": 81.52042388916016, "learning_rate": 6.925020919120331e-06, "loss": 0.3695, "step": 56130 }, { "epoch": 56.16, "grad_norm": 102.73677062988281, "learning_rate": 6.922121116885905e-06, "loss": 0.5348, "step": 56160 }, { "epoch": 56.19, "grad_norm": 50.5412483215332, "learning_rate": 6.919220555828502e-06, "loss": 0.4171, "step": 56190 }, { "epoch": 56.22, "grad_norm": 37.32706069946289, "learning_rate": 6.916319237093219e-06, "loss": 0.3777, "step": 56220 }, { "epoch": 56.25, "grad_norm": 32.61982727050781, "learning_rate": 6.913417161825449e-06, "loss": 0.5203, "step": 56250 }, { "epoch": 56.28, "grad_norm": 45.73463439941406, "learning_rate": 6.910514331170888e-06, "loss": 0.3738, "step": 56280 }, { "epoch": 56.31, "grad_norm": 68.09418487548828, "learning_rate": 6.907610746275524e-06, "loss": 0.3466, "step": 56310 }, { "epoch": 56.34, "grad_norm": 60.77250671386719, "learning_rate": 6.904706408285649e-06, "loss": 0.4238, "step": 56340 }, { "epoch": 56.37, "grad_norm": 53.293357849121094, "learning_rate": 6.901801318347848e-06, "loss": 0.383, "step": 56370 }, { "epoch": 56.4, "grad_norm": 61.03451156616211, "learning_rate": 6.898895477609007e-06, "loss": 0.3354, "step": 56400 }, { "epoch": 56.43, "grad_norm": 71.47589111328125, "learning_rate": 6.895988887216303e-06, "loss": 0.3403, "step": 56430 }, { "epoch": 56.46, "grad_norm": 70.13953399658203, "learning_rate": 6.893081548317212e-06, "loss": 0.4063, "step": 56460 }, { "epoch": 56.49, "grad_norm": 39.62968826293945, "learning_rate": 6.890173462059508e-06, "loss": 0.3358, "step": 56490 }, { "epoch": 56.52, "grad_norm": 74.25177001953125, "learning_rate": 6.887264629591254e-06, "loss": 0.4035, "step": 56520 }, { "epoch": 56.55, "grad_norm": 53.739994049072266, "learning_rate": 6.884355052060814e-06, "loss": 0.3683, "step": 56550 }, { "epoch": 56.58, "grad_norm": 107.16183471679688, "learning_rate": 6.881444730616842e-06, "loss": 0.4181, "step": 56580 }, { "epoch": 56.61, "grad_norm": 70.58218383789062, "learning_rate": 6.878533666408286e-06, "loss": 0.3746, "step": 56610 }, { "epoch": 56.64, "grad_norm": 60.49913787841797, "learning_rate": 6.875621860584389e-06, "loss": 0.3856, "step": 56640 }, { "epoch": 56.67, "grad_norm": 70.70442962646484, "learning_rate": 6.872709314294685e-06, "loss": 0.3744, "step": 56670 }, { "epoch": 56.7, "grad_norm": 73.89639282226562, "learning_rate": 6.869796028689002e-06, "loss": 0.397, "step": 56700 }, { "epoch": 56.73, "grad_norm": 34.148963928222656, "learning_rate": 6.86688200491746e-06, "loss": 0.3521, "step": 56730 }, { "epoch": 56.76, "grad_norm": 58.93877029418945, "learning_rate": 6.863967244130467e-06, "loss": 0.3942, "step": 56760 }, { "epoch": 56.79, "grad_norm": 95.49622344970703, "learning_rate": 6.861051747478727e-06, "loss": 0.3285, "step": 56790 }, { "epoch": 56.82, "grad_norm": 56.70945739746094, "learning_rate": 6.858135516113226e-06, "loss": 0.5101, "step": 56820 }, { "epoch": 56.85, "grad_norm": 64.5464859008789, "learning_rate": 6.8552185511852555e-06, "loss": 0.3301, "step": 56850 }, { "epoch": 56.88, "grad_norm": 46.058467864990234, "learning_rate": 6.852300853846381e-06, "loss": 0.3665, "step": 56880 }, { "epoch": 56.91, "grad_norm": 92.30994415283203, "learning_rate": 6.849382425248463e-06, "loss": 0.3702, "step": 56910 }, { "epoch": 56.94, "grad_norm": 81.6889877319336, "learning_rate": 6.846463266543653e-06, "loss": 0.4488, "step": 56940 }, { "epoch": 56.97, "grad_norm": 51.89170455932617, "learning_rate": 6.8435433788843865e-06, "loss": 0.4419, "step": 56970 }, { "epoch": 57.0, "grad_norm": 92.50395965576172, "learning_rate": 6.840622763423391e-06, "loss": 0.4013, "step": 57000 }, { "epoch": 57.0, "eval_loss": 0.3097820580005646, "eval_map": 0.7777, "eval_map_50": 0.9568, "eval_map_75": 0.9094, "eval_map_chicken": 0.7674, "eval_map_duck": 0.7292, "eval_map_large": 0.8032, "eval_map_medium": 0.764, "eval_map_plant": 0.8365, "eval_map_small": 0.2633, "eval_mar_1": 0.3181, "eval_mar_10": 0.8165, "eval_mar_100": 0.8202, "eval_mar_100_chicken": 0.8219, "eval_mar_100_duck": 0.7696, "eval_mar_100_plant": 0.8691, "eval_mar_large": 0.8402, "eval_mar_medium": 0.8118, "eval_mar_small": 0.3683, "eval_runtime": 12.6369, "eval_samples_per_second": 15.827, "eval_steps_per_second": 1.978, "step": 57000 }, { "epoch": 57.03, "grad_norm": 60.87810134887695, "learning_rate": 6.837701421313677e-06, "loss": 0.2917, "step": 57030 }, { "epoch": 57.06, "grad_norm": 45.24128341674805, "learning_rate": 6.8347793537085474e-06, "loss": 0.3407, "step": 57060 }, { "epoch": 57.09, "grad_norm": 188.6120147705078, "learning_rate": 6.831856561761585e-06, "loss": 0.3216, "step": 57090 }, { "epoch": 57.12, "grad_norm": 43.14576721191406, "learning_rate": 6.8289330466266635e-06, "loss": 0.4227, "step": 57120 }, { "epoch": 57.15, "grad_norm": 425.6170959472656, "learning_rate": 6.82600880945794e-06, "loss": 0.3233, "step": 57150 }, { "epoch": 57.18, "grad_norm": 112.72496795654297, "learning_rate": 6.823083851409857e-06, "loss": 0.3508, "step": 57180 }, { "epoch": 57.21, "grad_norm": 58.21677017211914, "learning_rate": 6.820158173637142e-06, "loss": 0.4323, "step": 57210 }, { "epoch": 57.24, "grad_norm": 54.072723388671875, "learning_rate": 6.817231777294804e-06, "loss": 0.3747, "step": 57240 }, { "epoch": 57.27, "grad_norm": 43.41655731201172, "learning_rate": 6.814304663538142e-06, "loss": 0.3902, "step": 57270 }, { "epoch": 57.3, "grad_norm": 46.376014709472656, "learning_rate": 6.811376833522729e-06, "loss": 0.5015, "step": 57300 }, { "epoch": 57.33, "grad_norm": 3.091455437242985e-05, "learning_rate": 6.808448288404431e-06, "loss": 0.3823, "step": 57330 }, { "epoch": 57.36, "grad_norm": 62.625770568847656, "learning_rate": 6.805519029339388e-06, "loss": 0.4116, "step": 57360 }, { "epoch": 57.39, "grad_norm": 73.46949005126953, "learning_rate": 6.802589057484027e-06, "loss": 0.4429, "step": 57390 }, { "epoch": 57.42, "grad_norm": 59.37082290649414, "learning_rate": 6.799658373995054e-06, "loss": 0.3908, "step": 57420 }, { "epoch": 57.45, "grad_norm": 159.38839721679688, "learning_rate": 6.796726980029454e-06, "loss": 0.3935, "step": 57450 }, { "epoch": 57.48, "grad_norm": 57.59449768066406, "learning_rate": 6.793794876744499e-06, "loss": 0.5182, "step": 57480 }, { "epoch": 57.51, "grad_norm": 45.67026138305664, "learning_rate": 6.790862065297733e-06, "loss": 0.2938, "step": 57510 }, { "epoch": 57.54, "grad_norm": 57.891056060791016, "learning_rate": 6.787928546846987e-06, "loss": 0.4114, "step": 57540 }, { "epoch": 57.57, "grad_norm": 81.49954986572266, "learning_rate": 6.784994322550367e-06, "loss": 0.3312, "step": 57570 }, { "epoch": 57.6, "grad_norm": 0.00010706239118007943, "learning_rate": 6.782059393566254e-06, "loss": 0.3852, "step": 57600 }, { "epoch": 57.63, "grad_norm": 52.82946014404297, "learning_rate": 6.779123761053317e-06, "loss": 0.4542, "step": 57630 }, { "epoch": 57.66, "grad_norm": 53.64664840698242, "learning_rate": 6.776187426170494e-06, "loss": 0.4206, "step": 57660 }, { "epoch": 57.69, "grad_norm": 44.826663970947266, "learning_rate": 6.773250390077006e-06, "loss": 0.3956, "step": 57690 }, { "epoch": 57.72, "grad_norm": 41.27812576293945, "learning_rate": 6.770312653932346e-06, "loss": 0.3965, "step": 57720 }, { "epoch": 57.75, "grad_norm": 91.8028335571289, "learning_rate": 6.767374218896286e-06, "loss": 0.4234, "step": 57750 }, { "epoch": 57.78, "grad_norm": 41.62392044067383, "learning_rate": 6.764435086128876e-06, "loss": 0.3865, "step": 57780 }, { "epoch": 57.81, "grad_norm": 44.93864822387695, "learning_rate": 6.761495256790434e-06, "loss": 0.4008, "step": 57810 }, { "epoch": 57.84, "grad_norm": 57.96244812011719, "learning_rate": 6.758554732041564e-06, "loss": 0.4098, "step": 57840 }, { "epoch": 57.87, "grad_norm": 43.79016876220703, "learning_rate": 6.755613513043136e-06, "loss": 0.3682, "step": 57870 }, { "epoch": 57.9, "grad_norm": 85.55265045166016, "learning_rate": 6.752671600956295e-06, "loss": 0.3815, "step": 57900 }, { "epoch": 57.93, "grad_norm": 53.19672393798828, "learning_rate": 6.749728996942465e-06, "loss": 0.4494, "step": 57930 }, { "epoch": 57.96, "grad_norm": 50.64846420288086, "learning_rate": 6.7467857021633354e-06, "loss": 0.4535, "step": 57960 }, { "epoch": 57.99, "grad_norm": 71.36148071289062, "learning_rate": 6.743841717780876e-06, "loss": 0.2877, "step": 57990 }, { "epoch": 58.0, "eval_loss": 0.2941030263900757, "eval_map": 0.7872, "eval_map_50": 0.9582, "eval_map_75": 0.9177, "eval_map_chicken": 0.7804, "eval_map_duck": 0.7313, "eval_map_large": 0.7936, "eval_map_medium": 0.781, "eval_map_plant": 0.8499, "eval_map_small": 0.2115, "eval_mar_1": 0.3202, "eval_mar_10": 0.8214, "eval_mar_100": 0.8249, "eval_mar_100_chicken": 0.8294, "eval_mar_100_duck": 0.7629, "eval_mar_100_plant": 0.8825, "eval_mar_large": 0.8291, "eval_mar_medium": 0.8268, "eval_mar_small": 0.2915, "eval_runtime": 13.352, "eval_samples_per_second": 14.979, "eval_steps_per_second": 1.872, "step": 58000 }, { "epoch": 58.02, "grad_norm": 147.58258056640625, "learning_rate": 6.740897044957322e-06, "loss": 0.2887, "step": 58020 }, { "epoch": 58.05, "grad_norm": 79.18315887451172, "learning_rate": 6.737951684855185e-06, "loss": 0.4548, "step": 58050 }, { "epoch": 58.08, "grad_norm": 122.64994812011719, "learning_rate": 6.7350056386372485e-06, "loss": 0.3761, "step": 58080 }, { "epoch": 58.11, "grad_norm": 67.95075225830078, "learning_rate": 6.7320589074665606e-06, "loss": 0.3314, "step": 58110 }, { "epoch": 58.14, "grad_norm": 85.00120544433594, "learning_rate": 6.72911149250645e-06, "loss": 0.4154, "step": 58140 }, { "epoch": 58.17, "grad_norm": 81.92926788330078, "learning_rate": 6.726163394920503e-06, "loss": 0.3533, "step": 58170 }, { "epoch": 58.2, "grad_norm": 54.335296630859375, "learning_rate": 6.723214615872585e-06, "loss": 0.4626, "step": 58200 }, { "epoch": 58.23, "grad_norm": 40.590736389160156, "learning_rate": 6.720265156526828e-06, "loss": 0.3128, "step": 58230 }, { "epoch": 58.26, "grad_norm": 43.47439193725586, "learning_rate": 6.717315018047631e-06, "loss": 0.3152, "step": 58260 }, { "epoch": 58.29, "grad_norm": 36.22964859008789, "learning_rate": 6.7143642015996626e-06, "loss": 0.3795, "step": 58290 }, { "epoch": 58.32, "grad_norm": 37.55474090576172, "learning_rate": 6.711412708347857e-06, "loss": 0.3402, "step": 58320 }, { "epoch": 58.35, "grad_norm": 78.66886138916016, "learning_rate": 6.708460539457418e-06, "loss": 0.4194, "step": 58350 }, { "epoch": 58.38, "grad_norm": 32.450218200683594, "learning_rate": 6.7055076960938135e-06, "loss": 0.3704, "step": 58380 }, { "epoch": 58.41, "grad_norm": 62.966922760009766, "learning_rate": 6.702554179422782e-06, "loss": 0.3653, "step": 58410 }, { "epoch": 58.44, "grad_norm": 77.44872283935547, "learning_rate": 6.699599990610324e-06, "loss": 0.3517, "step": 58440 }, { "epoch": 58.47, "grad_norm": 28.800491333007812, "learning_rate": 6.696645130822704e-06, "loss": 0.4089, "step": 58470 }, { "epoch": 58.5, "grad_norm": 56.03002166748047, "learning_rate": 6.693689601226458e-06, "loss": 0.5056, "step": 58500 }, { "epoch": 58.53, "grad_norm": 71.70792388916016, "learning_rate": 6.69073340298838e-06, "loss": 0.4396, "step": 58530 }, { "epoch": 58.56, "grad_norm": 111.28275299072266, "learning_rate": 6.68777653727553e-06, "loss": 0.4036, "step": 58560 }, { "epoch": 58.59, "grad_norm": 42.994808197021484, "learning_rate": 6.684819005255232e-06, "loss": 0.3804, "step": 58590 }, { "epoch": 58.62, "grad_norm": 97.53770446777344, "learning_rate": 6.681860808095074e-06, "loss": 0.3825, "step": 58620 }, { "epoch": 58.65, "grad_norm": 66.64350891113281, "learning_rate": 6.6789019469629034e-06, "loss": 0.4022, "step": 58650 }, { "epoch": 58.68, "grad_norm": 61.24783706665039, "learning_rate": 6.675942423026834e-06, "loss": 0.3848, "step": 58680 }, { "epoch": 58.71, "grad_norm": 0.00017483126430306584, "learning_rate": 6.672982237455238e-06, "loss": 0.3253, "step": 58710 }, { "epoch": 58.74, "grad_norm": 32.77238082885742, "learning_rate": 6.6700213914167485e-06, "loss": 0.4047, "step": 58740 }, { "epoch": 58.77, "grad_norm": 50.055423736572266, "learning_rate": 6.667059886080263e-06, "loss": 0.3855, "step": 58770 }, { "epoch": 58.8, "grad_norm": 60.623783111572266, "learning_rate": 6.664097722614934e-06, "loss": 0.3346, "step": 58800 }, { "epoch": 58.83, "grad_norm": 57.326080322265625, "learning_rate": 6.6611349021901795e-06, "loss": 0.3361, "step": 58830 }, { "epoch": 58.86, "grad_norm": 22.644861221313477, "learning_rate": 6.658171425975673e-06, "loss": 0.3256, "step": 58860 }, { "epoch": 58.89, "grad_norm": 64.05980682373047, "learning_rate": 6.655207295141346e-06, "loss": 0.3055, "step": 58890 }, { "epoch": 58.92, "grad_norm": 49.62561798095703, "learning_rate": 6.652242510857395e-06, "loss": 0.4137, "step": 58920 }, { "epoch": 58.95, "grad_norm": 36.36127471923828, "learning_rate": 6.649277074294265e-06, "loss": 0.4067, "step": 58950 }, { "epoch": 58.98, "grad_norm": 56.88331985473633, "learning_rate": 6.6463109866226675e-06, "loss": 0.4487, "step": 58980 }, { "epoch": 59.0, "eval_loss": 0.28298962116241455, "eval_map": 0.7979, "eval_map_50": 0.9649, "eval_map_75": 0.9219, "eval_map_chicken": 0.7902, "eval_map_duck": 0.7438, "eval_map_large": 0.8174, "eval_map_medium": 0.7887, "eval_map_plant": 0.8597, "eval_map_small": 0.2545, "eval_mar_1": 0.322, "eval_mar_10": 0.835, "eval_mar_100": 0.8389, "eval_mar_100_chicken": 0.8378, "eval_mar_100_duck": 0.7881, "eval_mar_100_plant": 0.8907, "eval_mar_large": 0.8519, "eval_mar_medium": 0.8374, "eval_mar_small": 0.377, "eval_runtime": 12.6863, "eval_samples_per_second": 15.765, "eval_steps_per_second": 1.971, "step": 59000 }, { "epoch": 59.01, "grad_norm": 81.05548858642578, "learning_rate": 6.643344249013562e-06, "loss": 0.3652, "step": 59010 }, { "epoch": 59.04, "grad_norm": 85.30073547363281, "learning_rate": 6.640376862638176e-06, "loss": 0.3521, "step": 59040 }, { "epoch": 59.07, "grad_norm": 36.425621032714844, "learning_rate": 6.637408828667982e-06, "loss": 0.3594, "step": 59070 }, { "epoch": 59.1, "grad_norm": 54.088558197021484, "learning_rate": 6.634440148274712e-06, "loss": 0.3741, "step": 59100 }, { "epoch": 59.13, "grad_norm": 75.97935485839844, "learning_rate": 6.6314708226303596e-06, "loss": 0.3263, "step": 59130 }, { "epoch": 59.16, "grad_norm": 39.11371994018555, "learning_rate": 6.6285008529071615e-06, "loss": 0.3666, "step": 59160 }, { "epoch": 59.19, "grad_norm": 4.700033969129436e-05, "learning_rate": 6.6255302402776175e-06, "loss": 0.4572, "step": 59190 }, { "epoch": 59.22, "grad_norm": 83.22734832763672, "learning_rate": 6.622558985914478e-06, "loss": 0.3972, "step": 59220 }, { "epoch": 59.25, "grad_norm": 75.07606506347656, "learning_rate": 6.619587090990748e-06, "loss": 0.3198, "step": 59250 }, { "epoch": 59.28, "grad_norm": 37.45214080810547, "learning_rate": 6.616614556679684e-06, "loss": 0.3759, "step": 59280 }, { "epoch": 59.31, "grad_norm": 38.65396499633789, "learning_rate": 6.613641384154794e-06, "loss": 0.3193, "step": 59310 }, { "epoch": 59.34, "grad_norm": 42.6458625793457, "learning_rate": 6.610667574589841e-06, "loss": 0.2892, "step": 59340 }, { "epoch": 59.37, "grad_norm": 42.6876106262207, "learning_rate": 6.6076931291588375e-06, "loss": 0.3288, "step": 59370 }, { "epoch": 59.4, "grad_norm": 81.7554702758789, "learning_rate": 6.604718049036047e-06, "loss": 0.3404, "step": 59400 }, { "epoch": 59.43, "grad_norm": 54.78104019165039, "learning_rate": 6.601742335395987e-06, "loss": 0.3661, "step": 59430 }, { "epoch": 59.46, "grad_norm": 53.65159225463867, "learning_rate": 6.598765989413419e-06, "loss": 0.36, "step": 59460 }, { "epoch": 59.49, "grad_norm": 89.16064453125, "learning_rate": 6.595789012263356e-06, "loss": 0.3828, "step": 59490 }, { "epoch": 59.52, "grad_norm": 33.220741271972656, "learning_rate": 6.592811405121064e-06, "loss": 0.3598, "step": 59520 }, { "epoch": 59.55, "grad_norm": 43.04564666748047, "learning_rate": 6.589833169162055e-06, "loss": 0.3434, "step": 59550 }, { "epoch": 59.58, "grad_norm": 34.29422378540039, "learning_rate": 6.5868543055620895e-06, "loss": 0.358, "step": 59580 }, { "epoch": 59.61, "grad_norm": 60.96402359008789, "learning_rate": 6.583874815497174e-06, "loss": 0.3691, "step": 59610 }, { "epoch": 59.64, "grad_norm": 72.82115936279297, "learning_rate": 6.580894700143565e-06, "loss": 0.3459, "step": 59640 }, { "epoch": 59.67, "grad_norm": 31.78813362121582, "learning_rate": 6.577913960677766e-06, "loss": 0.5157, "step": 59670 }, { "epoch": 59.7, "grad_norm": 104.75907897949219, "learning_rate": 6.574932598276524e-06, "loss": 0.4806, "step": 59700 }, { "epoch": 59.73, "grad_norm": 84.29928588867188, "learning_rate": 6.571950614116835e-06, "loss": 0.4194, "step": 59730 }, { "epoch": 59.76, "grad_norm": 41.669647216796875, "learning_rate": 6.568968009375938e-06, "loss": 0.4333, "step": 59760 }, { "epoch": 59.79, "grad_norm": 0.00011452862963778898, "learning_rate": 6.5659847852313184e-06, "loss": 0.387, "step": 59790 }, { "epoch": 59.82, "grad_norm": 70.15641021728516, "learning_rate": 6.5630009428607065e-06, "loss": 0.3311, "step": 59820 }, { "epoch": 59.85, "grad_norm": 142.23939514160156, "learning_rate": 6.5600164834420754e-06, "loss": 0.4122, "step": 59850 }, { "epoch": 59.88, "grad_norm": 50.906524658203125, "learning_rate": 6.557031408153642e-06, "loss": 0.363, "step": 59880 }, { "epoch": 59.91, "grad_norm": 57.844200134277344, "learning_rate": 6.554045718173867e-06, "loss": 0.3613, "step": 59910 }, { "epoch": 59.94, "grad_norm": 37.68030548095703, "learning_rate": 6.551059414681455e-06, "loss": 0.3711, "step": 59940 }, { "epoch": 59.97, "grad_norm": 120.58763885498047, "learning_rate": 6.54807249885535e-06, "loss": 0.4175, "step": 59970 }, { "epoch": 60.0, "grad_norm": 82.78518676757812, "learning_rate": 6.545084971874738e-06, "loss": 0.3549, "step": 60000 }, { "epoch": 60.0, "eval_loss": 0.30069810152053833, "eval_map": 0.7947, "eval_map_50": 0.9604, "eval_map_75": 0.9171, "eval_map_chicken": 0.7931, "eval_map_duck": 0.7457, "eval_map_large": 0.8257, "eval_map_medium": 0.7754, "eval_map_plant": 0.8452, "eval_map_small": 0.3052, "eval_mar_1": 0.3241, "eval_mar_10": 0.8311, "eval_mar_100": 0.8339, "eval_mar_100_chicken": 0.8414, "eval_mar_100_duck": 0.783, "eval_mar_100_plant": 0.8771, "eval_mar_large": 0.8606, "eval_mar_medium": 0.8184, "eval_mar_small": 0.3913, "eval_runtime": 12.5515, "eval_samples_per_second": 15.934, "eval_steps_per_second": 1.992, "step": 60000 }, { "epoch": 60.03, "grad_norm": 128.6958465576172, "learning_rate": 6.542096834919049e-06, "loss": 0.4114, "step": 60030 }, { "epoch": 60.06, "grad_norm": 71.9616928100586, "learning_rate": 6.539108089167953e-06, "loss": 0.3334, "step": 60060 }, { "epoch": 60.09, "grad_norm": 49.02715301513672, "learning_rate": 6.536118735801356e-06, "loss": 0.3491, "step": 60090 }, { "epoch": 60.12, "grad_norm": 43.89875030517578, "learning_rate": 6.533128775999411e-06, "loss": 0.374, "step": 60120 }, { "epoch": 60.15, "grad_norm": 618.9513549804688, "learning_rate": 6.530138210942505e-06, "loss": 0.3766, "step": 60150 }, { "epoch": 60.18, "grad_norm": 32.276615142822266, "learning_rate": 6.527147041811266e-06, "loss": 0.3742, "step": 60180 }, { "epoch": 60.21, "grad_norm": 70.07975006103516, "learning_rate": 6.52415526978656e-06, "loss": 0.4379, "step": 60210 }, { "epoch": 60.24, "grad_norm": 55.220096588134766, "learning_rate": 6.521162896049491e-06, "loss": 0.5058, "step": 60240 }, { "epoch": 60.27, "grad_norm": 56.88646697998047, "learning_rate": 6.5181699217814025e-06, "loss": 0.371, "step": 60270 }, { "epoch": 60.3, "grad_norm": 106.04113006591797, "learning_rate": 6.5151763481638705e-06, "loss": 0.5114, "step": 60300 }, { "epoch": 60.33, "grad_norm": 52.668174743652344, "learning_rate": 6.512182176378713e-06, "loss": 0.3793, "step": 60330 }, { "epoch": 60.36, "grad_norm": 35.85966110229492, "learning_rate": 6.509187407607981e-06, "loss": 0.3913, "step": 60360 }, { "epoch": 60.39, "grad_norm": 93.56563568115234, "learning_rate": 6.50619204303396e-06, "loss": 0.3633, "step": 60390 }, { "epoch": 60.42, "grad_norm": 39.62114715576172, "learning_rate": 6.503196083839175e-06, "loss": 0.4034, "step": 60420 }, { "epoch": 60.45, "grad_norm": 51.10441207885742, "learning_rate": 6.500199531206381e-06, "loss": 0.3686, "step": 60450 }, { "epoch": 60.48, "grad_norm": 38.688201904296875, "learning_rate": 6.497202386318573e-06, "loss": 0.3679, "step": 60480 }, { "epoch": 60.51, "grad_norm": 40.257938385009766, "learning_rate": 6.494204650358973e-06, "loss": 0.3247, "step": 60510 }, { "epoch": 60.54, "grad_norm": 102.19364929199219, "learning_rate": 6.491206324511039e-06, "loss": 0.4342, "step": 60540 }, { "epoch": 60.57, "grad_norm": 87.02021026611328, "learning_rate": 6.488207409958466e-06, "loss": 0.3607, "step": 60570 }, { "epoch": 60.6, "grad_norm": 40.293087005615234, "learning_rate": 6.485207907885175e-06, "loss": 0.3868, "step": 60600 }, { "epoch": 60.63, "grad_norm": 95.04915618896484, "learning_rate": 6.482207819475324e-06, "loss": 0.3522, "step": 60630 }, { "epoch": 60.66, "grad_norm": 101.03605651855469, "learning_rate": 6.4792071459133e-06, "loss": 0.3246, "step": 60660 }, { "epoch": 60.69, "grad_norm": 115.00274658203125, "learning_rate": 6.47620588838372e-06, "loss": 0.3186, "step": 60690 }, { "epoch": 60.72, "grad_norm": 35.55888748168945, "learning_rate": 6.473204048071433e-06, "loss": 0.3288, "step": 60720 }, { "epoch": 60.75, "grad_norm": 62.44702911376953, "learning_rate": 6.47020162616152e-06, "loss": 0.4302, "step": 60750 }, { "epoch": 60.78, "grad_norm": 72.7789077758789, "learning_rate": 6.467198623839288e-06, "loss": 0.3485, "step": 60780 }, { "epoch": 60.81, "grad_norm": 77.27767181396484, "learning_rate": 6.464195042290277e-06, "loss": 0.435, "step": 60810 }, { "epoch": 60.84, "grad_norm": 39.44632339477539, "learning_rate": 6.4611908827002504e-06, "loss": 0.3658, "step": 60840 }, { "epoch": 60.87, "grad_norm": 124.9926986694336, "learning_rate": 6.458186146255203e-06, "loss": 0.3403, "step": 60870 }, { "epoch": 60.9, "grad_norm": 43.701534271240234, "learning_rate": 6.455180834141359e-06, "loss": 0.3596, "step": 60900 }, { "epoch": 60.93, "grad_norm": 69.94612884521484, "learning_rate": 6.452174947545169e-06, "loss": 0.4145, "step": 60930 }, { "epoch": 60.96, "grad_norm": 53.247982025146484, "learning_rate": 6.449168487653305e-06, "loss": 0.3422, "step": 60960 }, { "epoch": 60.99, "grad_norm": 41.41682815551758, "learning_rate": 6.446161455652674e-06, "loss": 0.3725, "step": 60990 }, { "epoch": 61.0, "eval_loss": 0.3179991543292999, "eval_map": 0.7636, "eval_map_50": 0.9581, "eval_map_75": 0.9079, "eval_map_chicken": 0.7511, "eval_map_duck": 0.701, "eval_map_large": 0.7981, "eval_map_medium": 0.7514, "eval_map_plant": 0.8388, "eval_map_small": 0.2728, "eval_mar_1": 0.3094, "eval_mar_10": 0.8045, "eval_mar_100": 0.8077, "eval_mar_100_chicken": 0.8074, "eval_mar_100_duck": 0.7433, "eval_mar_100_plant": 0.8723, "eval_mar_large": 0.8388, "eval_mar_medium": 0.798, "eval_mar_small": 0.3622, "eval_runtime": 12.9541, "eval_samples_per_second": 15.439, "eval_steps_per_second": 1.93, "step": 61000 }, { "epoch": 61.02, "grad_norm": 56.95610046386719, "learning_rate": 6.443153852730404e-06, "loss": 0.321, "step": 61020 }, { "epoch": 61.05, "grad_norm": 93.86061096191406, "learning_rate": 6.440145680073847e-06, "loss": 0.3504, "step": 61050 }, { "epoch": 61.08, "grad_norm": 54.353023529052734, "learning_rate": 6.437136938870583e-06, "loss": 0.5213, "step": 61080 }, { "epoch": 61.11, "grad_norm": 42.24439239501953, "learning_rate": 6.434127630308415e-06, "loss": 0.3086, "step": 61110 }, { "epoch": 61.14, "grad_norm": 50.75392150878906, "learning_rate": 6.431117755575371e-06, "loss": 0.3897, "step": 61140 }, { "epoch": 61.17, "grad_norm": 61.51840591430664, "learning_rate": 6.428107315859702e-06, "loss": 0.3362, "step": 61170 }, { "epoch": 61.2, "grad_norm": 43.39870071411133, "learning_rate": 6.425096312349881e-06, "loss": 0.3145, "step": 61200 }, { "epoch": 61.23, "grad_norm": 33.54622268676758, "learning_rate": 6.422084746234605e-06, "loss": 0.3452, "step": 61230 }, { "epoch": 61.26, "grad_norm": 64.30875396728516, "learning_rate": 6.419072618702794e-06, "loss": 0.3717, "step": 61260 }, { "epoch": 61.29, "grad_norm": 69.30331420898438, "learning_rate": 6.416059930943586e-06, "loss": 0.3548, "step": 61290 }, { "epoch": 61.32, "grad_norm": 43.84052276611328, "learning_rate": 6.413046684146343e-06, "loss": 0.3571, "step": 61320 }, { "epoch": 61.35, "grad_norm": 91.97123718261719, "learning_rate": 6.410032879500647e-06, "loss": 0.332, "step": 61350 }, { "epoch": 61.38, "grad_norm": 65.26335906982422, "learning_rate": 6.407018518196303e-06, "loss": 0.4153, "step": 61380 }, { "epoch": 61.41, "grad_norm": 71.75984191894531, "learning_rate": 6.40400360142333e-06, "loss": 0.4211, "step": 61410 }, { "epoch": 61.44, "grad_norm": 121.30509185791016, "learning_rate": 6.400988130371969e-06, "loss": 0.4871, "step": 61440 }, { "epoch": 61.47, "grad_norm": 66.27267456054688, "learning_rate": 6.3979721062326815e-06, "loss": 0.4211, "step": 61470 }, { "epoch": 61.5, "grad_norm": 77.6305923461914, "learning_rate": 6.3949555301961474e-06, "loss": 0.3449, "step": 61500 }, { "epoch": 61.53, "grad_norm": 98.90530395507812, "learning_rate": 6.391938403453262e-06, "loss": 0.3891, "step": 61530 }, { "epoch": 61.56, "grad_norm": 93.91175079345703, "learning_rate": 6.388920727195138e-06, "loss": 0.4165, "step": 61560 }, { "epoch": 61.59, "grad_norm": 30.55654525756836, "learning_rate": 6.385902502613106e-06, "loss": 0.4572, "step": 61590 }, { "epoch": 61.62, "grad_norm": 149.6206512451172, "learning_rate": 6.382883730898717e-06, "loss": 0.4169, "step": 61620 }, { "epoch": 61.65, "grad_norm": 39.769859313964844, "learning_rate": 6.3798644132437304e-06, "loss": 0.4352, "step": 61650 }, { "epoch": 61.68, "grad_norm": 43.55778503417969, "learning_rate": 6.376844550840126e-06, "loss": 0.3872, "step": 61680 }, { "epoch": 61.71, "grad_norm": 126.61575317382812, "learning_rate": 6.373824144880099e-06, "loss": 0.4351, "step": 61710 }, { "epoch": 61.74, "grad_norm": 71.03457641601562, "learning_rate": 6.3708031965560545e-06, "loss": 0.4362, "step": 61740 }, { "epoch": 61.77, "grad_norm": 49.45713424682617, "learning_rate": 6.36778170706062e-06, "loss": 0.4067, "step": 61770 }, { "epoch": 61.8, "grad_norm": 52.46287155151367, "learning_rate": 6.364759677586627e-06, "loss": 0.3256, "step": 61800 }, { "epoch": 61.83, "grad_norm": 43.19639587402344, "learning_rate": 6.361737109327128e-06, "loss": 0.3683, "step": 61830 }, { "epoch": 61.86, "grad_norm": 44.9078254699707, "learning_rate": 6.3587140034753836e-06, "loss": 0.4427, "step": 61860 }, { "epoch": 61.89, "grad_norm": 327.8224182128906, "learning_rate": 6.355690361224869e-06, "loss": 0.4683, "step": 61890 }, { "epoch": 61.92, "grad_norm": 48.17255401611328, "learning_rate": 6.352666183769269e-06, "loss": 0.3777, "step": 61920 }, { "epoch": 61.95, "grad_norm": 44.22959899902344, "learning_rate": 6.349641472302484e-06, "loss": 0.3495, "step": 61950 }, { "epoch": 61.98, "grad_norm": 38.38023376464844, "learning_rate": 6.3466162280186164e-06, "loss": 0.3535, "step": 61980 }, { "epoch": 62.0, "eval_loss": 0.3056127429008484, "eval_map": 0.7781, "eval_map_50": 0.962, "eval_map_75": 0.9129, "eval_map_chicken": 0.7639, "eval_map_duck": 0.7345, "eval_map_large": 0.8043, "eval_map_medium": 0.7632, "eval_map_plant": 0.8359, "eval_map_small": 0.2116, "eval_mar_1": 0.3189, "eval_mar_10": 0.8165, "eval_mar_100": 0.8191, "eval_mar_100_chicken": 0.8173, "eval_mar_100_duck": 0.7686, "eval_mar_100_plant": 0.8713, "eval_mar_large": 0.8377, "eval_mar_medium": 0.8074, "eval_mar_small": 0.3281, "eval_runtime": 12.4478, "eval_samples_per_second": 16.067, "eval_steps_per_second": 2.008, "step": 62000 }, { "epoch": 62.01, "grad_norm": 34.9708251953125, "learning_rate": 6.343590452111991e-06, "loss": 0.3499, "step": 62010 }, { "epoch": 62.04, "grad_norm": 54.371543884277344, "learning_rate": 6.340564145777131e-06, "loss": 0.3425, "step": 62040 }, { "epoch": 62.07, "grad_norm": 62.7802619934082, "learning_rate": 6.337537310208779e-06, "loss": 0.3647, "step": 62070 }, { "epoch": 62.1, "grad_norm": 45.26636505126953, "learning_rate": 6.334509946601879e-06, "loss": 0.4043, "step": 62100 }, { "epoch": 62.13, "grad_norm": 50.49568557739258, "learning_rate": 6.3314820561515854e-06, "loss": 0.407, "step": 62130 }, { "epoch": 62.16, "grad_norm": 52.69154739379883, "learning_rate": 6.328453640053264e-06, "loss": 0.3495, "step": 62160 }, { "epoch": 62.19, "grad_norm": 49.83322525024414, "learning_rate": 6.325424699502482e-06, "loss": 0.3737, "step": 62190 }, { "epoch": 62.22, "grad_norm": 0.00019446555234026164, "learning_rate": 6.322395235695022e-06, "loss": 0.4224, "step": 62220 }, { "epoch": 62.25, "grad_norm": 37.01228332519531, "learning_rate": 6.3193652498268656e-06, "loss": 0.3742, "step": 62250 }, { "epoch": 62.28, "grad_norm": 0.00040778995025902987, "learning_rate": 6.316334743094201e-06, "loss": 0.3515, "step": 62280 }, { "epoch": 62.31, "grad_norm": 55.15784454345703, "learning_rate": 6.313303716693428e-06, "loss": 0.3745, "step": 62310 }, { "epoch": 62.34, "grad_norm": 50.539920806884766, "learning_rate": 6.310272171821145e-06, "loss": 0.4317, "step": 62340 }, { "epoch": 62.37, "grad_norm": 31.564929962158203, "learning_rate": 6.307240109674162e-06, "loss": 0.3579, "step": 62370 }, { "epoch": 62.4, "grad_norm": 59.159034729003906, "learning_rate": 6.304207531449486e-06, "loss": 0.3815, "step": 62400 }, { "epoch": 62.43, "grad_norm": 56.772708892822266, "learning_rate": 6.301174438344329e-06, "loss": 0.359, "step": 62430 }, { "epoch": 62.46, "grad_norm": 91.64718627929688, "learning_rate": 6.298140831556112e-06, "loss": 0.3422, "step": 62460 }, { "epoch": 62.49, "grad_norm": 28.45844841003418, "learning_rate": 6.2951067122824515e-06, "loss": 0.4481, "step": 62490 }, { "epoch": 62.52, "grad_norm": 67.59664916992188, "learning_rate": 6.292072081721173e-06, "loss": 0.464, "step": 62520 }, { "epoch": 62.55, "grad_norm": 70.10758972167969, "learning_rate": 6.2890369410703e-06, "loss": 0.41, "step": 62550 }, { "epoch": 62.58, "grad_norm": 36.02994155883789, "learning_rate": 6.286001291528056e-06, "loss": 0.4122, "step": 62580 }, { "epoch": 62.61, "grad_norm": 101.02970123291016, "learning_rate": 6.282965134292869e-06, "loss": 0.4027, "step": 62610 }, { "epoch": 62.64, "grad_norm": 93.14354705810547, "learning_rate": 6.279928470563365e-06, "loss": 0.4265, "step": 62640 }, { "epoch": 62.67, "grad_norm": 66.82553100585938, "learning_rate": 6.2768913015383696e-06, "loss": 0.4231, "step": 62670 }, { "epoch": 62.7, "grad_norm": 53.57185363769531, "learning_rate": 6.273853628416911e-06, "loss": 0.359, "step": 62700 }, { "epoch": 62.73, "grad_norm": 62.21583938598633, "learning_rate": 6.2708154523982155e-06, "loss": 0.3488, "step": 62730 }, { "epoch": 62.76, "grad_norm": 43.83544158935547, "learning_rate": 6.267776774681703e-06, "loss": 0.3676, "step": 62760 }, { "epoch": 62.79, "grad_norm": 350.2054748535156, "learning_rate": 6.264737596466999e-06, "loss": 0.3521, "step": 62790 }, { "epoch": 62.82, "grad_norm": 124.99991607666016, "learning_rate": 6.261697918953922e-06, "loss": 0.3081, "step": 62820 }, { "epoch": 62.85, "grad_norm": 58.331974029541016, "learning_rate": 6.258657743342486e-06, "loss": 0.3903, "step": 62850 }, { "epoch": 62.88, "grad_norm": 2277.24658203125, "learning_rate": 6.255617070832908e-06, "loss": 0.3953, "step": 62880 }, { "epoch": 62.91, "grad_norm": 99.1843032836914, "learning_rate": 6.252575902625595e-06, "loss": 0.3926, "step": 62910 }, { "epoch": 62.94, "grad_norm": 39.525760650634766, "learning_rate": 6.249534239921154e-06, "loss": 0.3685, "step": 62940 }, { "epoch": 62.97, "grad_norm": 44.85007858276367, "learning_rate": 6.2464920839203805e-06, "loss": 0.3887, "step": 62970 }, { "epoch": 63.0, "grad_norm": 0.00039311478030867875, "learning_rate": 6.243449435824276e-06, "loss": 0.4038, "step": 63000 }, { "epoch": 63.0, "eval_loss": 0.2947438359260559, "eval_map": 0.7899, "eval_map_50": 0.9584, "eval_map_75": 0.9148, "eval_map_chicken": 0.7778, "eval_map_duck": 0.7416, "eval_map_large": 0.8087, "eval_map_medium": 0.7853, "eval_map_plant": 0.8502, "eval_map_small": 0.2826, "eval_mar_1": 0.3239, "eval_mar_10": 0.8263, "eval_mar_100": 0.8289, "eval_mar_100_chicken": 0.8282, "eval_mar_100_duck": 0.7784, "eval_mar_100_plant": 0.8802, "eval_mar_large": 0.8401, "eval_mar_medium": 0.8273, "eval_mar_small": 0.3443, "eval_runtime": 12.5183, "eval_samples_per_second": 15.977, "eval_steps_per_second": 1.997, "step": 63000 }, { "epoch": 63.03, "grad_norm": 62.1936149597168, "learning_rate": 6.240406296834024e-06, "loss": 0.3575, "step": 63030 }, { "epoch": 63.06, "grad_norm": 56.482051849365234, "learning_rate": 6.237362668151013e-06, "loss": 0.4357, "step": 63060 }, { "epoch": 63.09, "grad_norm": 144.40240478515625, "learning_rate": 6.234318550976815e-06, "loss": 0.399, "step": 63090 }, { "epoch": 63.12, "grad_norm": 109.55199432373047, "learning_rate": 6.231273946513201e-06, "loss": 0.4317, "step": 63120 }, { "epoch": 63.15, "grad_norm": 80.31204986572266, "learning_rate": 6.228228855962133e-06, "loss": 0.3278, "step": 63150 }, { "epoch": 63.18, "grad_norm": 69.71114349365234, "learning_rate": 6.225183280525763e-06, "loss": 0.4067, "step": 63180 }, { "epoch": 63.21, "grad_norm": 50.825042724609375, "learning_rate": 6.222137221406439e-06, "loss": 0.36, "step": 63210 }, { "epoch": 63.24, "grad_norm": 77.85767364501953, "learning_rate": 6.219090679806694e-06, "loss": 0.3514, "step": 63240 }, { "epoch": 63.27, "grad_norm": 41.4321403503418, "learning_rate": 6.216043656929254e-06, "loss": 0.3463, "step": 63270 }, { "epoch": 63.3, "grad_norm": 50.52762985229492, "learning_rate": 6.212996153977038e-06, "loss": 0.3476, "step": 63300 }, { "epoch": 63.33, "grad_norm": 74.62356567382812, "learning_rate": 6.20994817215315e-06, "loss": 0.3563, "step": 63330 }, { "epoch": 63.36, "grad_norm": 179.68203735351562, "learning_rate": 6.206899712660887e-06, "loss": 0.4224, "step": 63360 }, { "epoch": 63.39, "grad_norm": 106.73099517822266, "learning_rate": 6.2038507767037295e-06, "loss": 0.3725, "step": 63390 }, { "epoch": 63.42, "grad_norm": 38.56673049926758, "learning_rate": 6.2008013654853505e-06, "loss": 0.3198, "step": 63420 }, { "epoch": 63.45, "grad_norm": 110.48749542236328, "learning_rate": 6.1977514802096105e-06, "loss": 0.3392, "step": 63450 }, { "epoch": 63.48, "grad_norm": 131.63723754882812, "learning_rate": 6.1947011220805535e-06, "loss": 0.3849, "step": 63480 }, { "epoch": 63.51, "grad_norm": 51.075538635253906, "learning_rate": 6.1916502923024145e-06, "loss": 0.3727, "step": 63510 }, { "epoch": 63.54, "grad_norm": 34.31461715698242, "learning_rate": 6.188598992079613e-06, "loss": 0.3435, "step": 63540 }, { "epoch": 63.57, "grad_norm": 30.406259536743164, "learning_rate": 6.1855472226167525e-06, "loss": 0.3291, "step": 63570 }, { "epoch": 63.6, "grad_norm": 93.5394515991211, "learning_rate": 6.182494985118625e-06, "loss": 0.419, "step": 63600 }, { "epoch": 63.63, "grad_norm": 261.0888977050781, "learning_rate": 6.179442280790202e-06, "loss": 0.3793, "step": 63630 }, { "epoch": 63.66, "grad_norm": 23.71967124938965, "learning_rate": 6.176389110836647e-06, "loss": 0.3695, "step": 63660 }, { "epoch": 63.69, "grad_norm": 49.00981521606445, "learning_rate": 6.173335476463303e-06, "loss": 0.3761, "step": 63690 }, { "epoch": 63.72, "grad_norm": 57.05886459350586, "learning_rate": 6.170281378875692e-06, "loss": 0.3398, "step": 63720 }, { "epoch": 63.75, "grad_norm": 59.40810012817383, "learning_rate": 6.1672268192795285e-06, "loss": 0.4344, "step": 63750 }, { "epoch": 63.78, "grad_norm": 170.8481903076172, "learning_rate": 6.1641717988807006e-06, "loss": 0.4192, "step": 63780 }, { "epoch": 63.81, "grad_norm": 88.9516830444336, "learning_rate": 6.161116318885283e-06, "loss": 0.4423, "step": 63810 }, { "epoch": 63.84, "grad_norm": 90.96357727050781, "learning_rate": 6.158060380499533e-06, "loss": 0.3521, "step": 63840 }, { "epoch": 63.87, "grad_norm": 42.6907958984375, "learning_rate": 6.155003984929883e-06, "loss": 0.4106, "step": 63870 }, { "epoch": 63.9, "grad_norm": 54.36552047729492, "learning_rate": 6.151947133382954e-06, "loss": 0.378, "step": 63900 }, { "epoch": 63.93, "grad_norm": 59.22226333618164, "learning_rate": 6.148889827065538e-06, "loss": 0.3698, "step": 63930 }, { "epoch": 63.96, "grad_norm": 40.34580993652344, "learning_rate": 6.145832067184614e-06, "loss": 0.4277, "step": 63960 }, { "epoch": 63.99, "grad_norm": 61.29197692871094, "learning_rate": 6.142773854947336e-06, "loss": 0.4424, "step": 63990 }, { "epoch": 64.0, "eval_loss": 0.2922332286834717, "eval_map": 0.7939, "eval_map_50": 0.9606, "eval_map_75": 0.9113, "eval_map_chicken": 0.7892, "eval_map_duck": 0.742, "eval_map_large": 0.817, "eval_map_medium": 0.7862, "eval_map_plant": 0.8505, "eval_map_small": 0.241, "eval_mar_1": 0.3225, "eval_mar_10": 0.8305, "eval_mar_100": 0.8342, "eval_mar_100_chicken": 0.8402, "eval_mar_100_duck": 0.7804, "eval_mar_100_plant": 0.882, "eval_mar_large": 0.8462, "eval_mar_medium": 0.8289, "eval_mar_small": 0.3485, "eval_runtime": 13.0908, "eval_samples_per_second": 15.278, "eval_steps_per_second": 1.91, "step": 64000 }, { "epoch": 64.02, "grad_norm": 39.95719909667969, "learning_rate": 6.139715191561038e-06, "loss": 0.3836, "step": 64020 }, { "epoch": 64.05, "grad_norm": 56.288917541503906, "learning_rate": 6.136656078233233e-06, "loss": 0.3536, "step": 64050 }, { "epoch": 64.08, "grad_norm": 0.00022501962666865438, "learning_rate": 6.133596516171609e-06, "loss": 0.3307, "step": 64080 }, { "epoch": 64.11, "grad_norm": 62.51616287231445, "learning_rate": 6.130536506584032e-06, "loss": 0.3316, "step": 64110 }, { "epoch": 64.14, "grad_norm": 0.0001461364736314863, "learning_rate": 6.127476050678548e-06, "loss": 0.3975, "step": 64140 }, { "epoch": 64.17, "grad_norm": 108.1631851196289, "learning_rate": 6.124415149663374e-06, "loss": 0.3978, "step": 64170 }, { "epoch": 64.2, "grad_norm": 109.93218231201172, "learning_rate": 6.121353804746907e-06, "loss": 0.4449, "step": 64200 }, { "epoch": 64.23, "grad_norm": 79.91670989990234, "learning_rate": 6.118292017137716e-06, "loss": 0.4376, "step": 64230 }, { "epoch": 64.26, "grad_norm": 45.16354751586914, "learning_rate": 6.1152297880445476e-06, "loss": 0.3307, "step": 64260 }, { "epoch": 64.29, "grad_norm": 37.17070770263672, "learning_rate": 6.112167118676321e-06, "loss": 0.4138, "step": 64290 }, { "epoch": 64.32, "grad_norm": 78.70944213867188, "learning_rate": 6.109104010242127e-06, "loss": 0.3771, "step": 64320 }, { "epoch": 64.35, "grad_norm": 30.487407684326172, "learning_rate": 6.106040463951237e-06, "loss": 0.4005, "step": 64350 }, { "epoch": 64.38, "grad_norm": 99.32769775390625, "learning_rate": 6.102976481013086e-06, "loss": 0.3188, "step": 64380 }, { "epoch": 64.41, "grad_norm": 51.03107833862305, "learning_rate": 6.0999120626372895e-06, "loss": 0.38, "step": 64410 }, { "epoch": 64.44, "grad_norm": 38.63597106933594, "learning_rate": 6.09684721003363e-06, "loss": 0.3917, "step": 64440 }, { "epoch": 64.47, "grad_norm": 49.081336975097656, "learning_rate": 6.093781924412063e-06, "loss": 0.3512, "step": 64470 }, { "epoch": 64.5, "grad_norm": 95.63629913330078, "learning_rate": 6.090716206982714e-06, "loss": 0.2973, "step": 64500 }, { "epoch": 64.53, "grad_norm": 129.19769287109375, "learning_rate": 6.087650058955879e-06, "loss": 0.473, "step": 64530 }, { "epoch": 64.56, "grad_norm": 23.769739151000977, "learning_rate": 6.084583481542028e-06, "loss": 0.3918, "step": 64560 }, { "epoch": 64.59, "grad_norm": 42.32480239868164, "learning_rate": 6.081516475951793e-06, "loss": 0.3934, "step": 64590 }, { "epoch": 64.62, "grad_norm": 44.50368118286133, "learning_rate": 6.078449043395982e-06, "loss": 0.3453, "step": 64620 }, { "epoch": 64.65, "grad_norm": 46.044837951660156, "learning_rate": 6.075381185085568e-06, "loss": 0.4066, "step": 64650 }, { "epoch": 64.68, "grad_norm": 36.061580657958984, "learning_rate": 6.072312902231692e-06, "loss": 0.3259, "step": 64680 }, { "epoch": 64.71, "grad_norm": 31.90561866760254, "learning_rate": 6.069244196045666e-06, "loss": 0.3355, "step": 64710 }, { "epoch": 64.74, "grad_norm": 32.77406692504883, "learning_rate": 6.066175067738964e-06, "loss": 0.4256, "step": 64740 }, { "epoch": 64.77, "grad_norm": 78.5828857421875, "learning_rate": 6.06310551852323e-06, "loss": 0.4878, "step": 64770 }, { "epoch": 64.8, "grad_norm": 49.945648193359375, "learning_rate": 6.060035549610275e-06, "loss": 0.4303, "step": 64800 }, { "epoch": 64.83, "grad_norm": 54.64009475708008, "learning_rate": 6.056965162212072e-06, "loss": 0.383, "step": 64830 }, { "epoch": 64.86, "grad_norm": 117.1504135131836, "learning_rate": 6.053894357540761e-06, "loss": 0.3607, "step": 64860 }, { "epoch": 64.89, "grad_norm": 66.44549560546875, "learning_rate": 6.050823136808649e-06, "loss": 0.3582, "step": 64890 }, { "epoch": 64.92, "grad_norm": 52.47871017456055, "learning_rate": 6.047751501228203e-06, "loss": 0.3886, "step": 64920 }, { "epoch": 64.95, "grad_norm": 66.86099243164062, "learning_rate": 6.044679452012059e-06, "loss": 0.3436, "step": 64950 }, { "epoch": 64.98, "grad_norm": 28.176462173461914, "learning_rate": 6.041606990373012e-06, "loss": 0.3878, "step": 64980 }, { "epoch": 65.0, "eval_loss": 0.2872055172920227, "eval_map": 0.799, "eval_map_50": 0.9605, "eval_map_75": 0.9221, "eval_map_chicken": 0.7892, "eval_map_duck": 0.7498, "eval_map_large": 0.8179, "eval_map_medium": 0.7922, "eval_map_plant": 0.858, "eval_map_small": 0.2855, "eval_mar_1": 0.3216, "eval_mar_10": 0.8324, "eval_mar_100": 0.8358, "eval_mar_100_chicken": 0.837, "eval_mar_100_duck": 0.7809, "eval_mar_100_plant": 0.8895, "eval_mar_large": 0.8544, "eval_mar_medium": 0.8297, "eval_mar_small": 0.3819, "eval_runtime": 12.4257, "eval_samples_per_second": 16.096, "eval_steps_per_second": 2.012, "step": 65000 }, { "epoch": 65.01, "grad_norm": 66.0304183959961, "learning_rate": 6.038534117524021e-06, "loss": 0.3262, "step": 65010 }, { "epoch": 65.04, "grad_norm": 37.1042366027832, "learning_rate": 6.0354608346782075e-06, "loss": 0.4285, "step": 65040 }, { "epoch": 65.07, "grad_norm": 75.76040649414062, "learning_rate": 6.032387143048853e-06, "loss": 0.3664, "step": 65070 }, { "epoch": 65.1, "grad_norm": 80.7643814086914, "learning_rate": 6.029313043849407e-06, "loss": 0.382, "step": 65100 }, { "epoch": 65.13, "grad_norm": 77.18086242675781, "learning_rate": 6.026238538293472e-06, "loss": 0.4351, "step": 65130 }, { "epoch": 65.16, "grad_norm": 33.3852653503418, "learning_rate": 6.023163627594813e-06, "loss": 0.3835, "step": 65160 }, { "epoch": 65.19, "grad_norm": 83.08126068115234, "learning_rate": 6.02008831296736e-06, "loss": 0.3949, "step": 65190 }, { "epoch": 65.22, "grad_norm": 37.40925216674805, "learning_rate": 6.0170125956251935e-06, "loss": 0.3575, "step": 65220 }, { "epoch": 65.25, "grad_norm": 25.961978912353516, "learning_rate": 6.013936476782563e-06, "loss": 0.3558, "step": 65250 }, { "epoch": 65.28, "grad_norm": 50.64202117919922, "learning_rate": 6.010859957653869e-06, "loss": 0.3856, "step": 65280 }, { "epoch": 65.31, "grad_norm": 76.81957244873047, "learning_rate": 6.00778303945367e-06, "loss": 0.3707, "step": 65310 }, { "epoch": 65.34, "grad_norm": 37.44461441040039, "learning_rate": 6.00470572339669e-06, "loss": 0.35, "step": 65340 }, { "epoch": 65.37, "grad_norm": 121.81327819824219, "learning_rate": 6.0016280106978e-06, "loss": 0.4134, "step": 65370 }, { "epoch": 65.4, "grad_norm": 56.49502944946289, "learning_rate": 5.9985499025720354e-06, "loss": 0.3477, "step": 65400 }, { "epoch": 65.43, "grad_norm": 50.142250061035156, "learning_rate": 5.995471400234584e-06, "loss": 0.3217, "step": 65430 }, { "epoch": 65.46, "grad_norm": 140.8568572998047, "learning_rate": 5.992392504900786e-06, "loss": 0.3874, "step": 65460 }, { "epoch": 65.49, "grad_norm": 73.68807220458984, "learning_rate": 5.989313217786146e-06, "loss": 0.363, "step": 65490 }, { "epoch": 65.52, "grad_norm": 37.37092208862305, "learning_rate": 5.986233540106315e-06, "loss": 0.373, "step": 65520 }, { "epoch": 65.55, "grad_norm": 51.58872985839844, "learning_rate": 5.9831534730771e-06, "loss": 0.35, "step": 65550 }, { "epoch": 65.58, "grad_norm": 31.185863494873047, "learning_rate": 5.9800730179144665e-06, "loss": 0.385, "step": 65580 }, { "epoch": 65.61, "grad_norm": 115.38666534423828, "learning_rate": 5.9769921758345254e-06, "loss": 0.4984, "step": 65610 }, { "epoch": 65.64, "grad_norm": 72.06417083740234, "learning_rate": 5.973910948053545e-06, "loss": 0.407, "step": 65640 }, { "epoch": 65.67, "grad_norm": 41.132957458496094, "learning_rate": 5.970829335787946e-06, "loss": 0.4233, "step": 65670 }, { "epoch": 65.7, "grad_norm": 146.32679748535156, "learning_rate": 5.967747340254303e-06, "loss": 0.3458, "step": 65700 }, { "epoch": 65.73, "grad_norm": 74.85420227050781, "learning_rate": 5.9646649626693335e-06, "loss": 0.382, "step": 65730 }, { "epoch": 65.76, "grad_norm": 50.69607162475586, "learning_rate": 5.961582204249915e-06, "loss": 0.3443, "step": 65760 }, { "epoch": 65.79, "grad_norm": 68.29427337646484, "learning_rate": 5.958499066213071e-06, "loss": 0.4402, "step": 65790 }, { "epoch": 65.82, "grad_norm": 62.62144470214844, "learning_rate": 5.955415549775975e-06, "loss": 0.3739, "step": 65820 }, { "epoch": 65.85, "grad_norm": 42.782569885253906, "learning_rate": 5.952331656155951e-06, "loss": 0.3339, "step": 65850 }, { "epoch": 65.88, "grad_norm": 58.887149810791016, "learning_rate": 5.949247386570471e-06, "loss": 0.3679, "step": 65880 }, { "epoch": 65.91, "grad_norm": 53.57959747314453, "learning_rate": 5.9461627422371545e-06, "loss": 0.329, "step": 65910 }, { "epoch": 65.94, "grad_norm": 52.70747375488281, "learning_rate": 5.9430777243737744e-06, "loss": 0.3507, "step": 65940 }, { "epoch": 65.97, "grad_norm": 38.439453125, "learning_rate": 5.939992334198242e-06, "loss": 0.3848, "step": 65970 }, { "epoch": 66.0, "grad_norm": 70.37049102783203, "learning_rate": 5.936906572928625e-06, "loss": 0.3628, "step": 66000 }, { "epoch": 66.0, "eval_loss": 0.29752978682518005, "eval_map": 0.7879, "eval_map_50": 0.9592, "eval_map_75": 0.9135, "eval_map_chicken": 0.791, "eval_map_duck": 0.7272, "eval_map_large": 0.8108, "eval_map_medium": 0.7765, "eval_map_plant": 0.8456, "eval_map_small": 0.2664, "eval_mar_1": 0.3237, "eval_mar_10": 0.8232, "eval_mar_100": 0.8273, "eval_mar_100_chicken": 0.8364, "eval_mar_100_duck": 0.7655, "eval_mar_100_plant": 0.8799, "eval_mar_large": 0.8456, "eval_mar_medium": 0.8191, "eval_mar_small": 0.3577, "eval_runtime": 12.8054, "eval_samples_per_second": 15.618, "eval_steps_per_second": 1.952, "step": 66000 }, { "epoch": 66.03, "grad_norm": 63.069908142089844, "learning_rate": 5.933820441783129e-06, "loss": 0.4469, "step": 66030 }, { "epoch": 66.06, "grad_norm": 83.55052185058594, "learning_rate": 5.930733941980111e-06, "loss": 0.3434, "step": 66060 }, { "epoch": 66.09, "grad_norm": 41.397090911865234, "learning_rate": 5.927647074738074e-06, "loss": 0.3668, "step": 66090 }, { "epoch": 66.12, "grad_norm": 79.25060272216797, "learning_rate": 5.924559841275661e-06, "loss": 0.3606, "step": 66120 }, { "epoch": 66.15, "grad_norm": 78.85431671142578, "learning_rate": 5.9214722428116675e-06, "loss": 0.4146, "step": 66150 }, { "epoch": 66.18, "grad_norm": 35.486934661865234, "learning_rate": 5.918384280565025e-06, "loss": 0.3288, "step": 66180 }, { "epoch": 66.21, "grad_norm": 59.71826934814453, "learning_rate": 5.915295955754812e-06, "loss": 0.3579, "step": 66210 }, { "epoch": 66.24, "grad_norm": 58.44197463989258, "learning_rate": 5.912207269600252e-06, "loss": 0.4133, "step": 66240 }, { "epoch": 66.27, "grad_norm": 75.54934692382812, "learning_rate": 5.9091182233207075e-06, "loss": 0.3331, "step": 66270 }, { "epoch": 66.3, "grad_norm": 62.90567398071289, "learning_rate": 5.906028818135687e-06, "loss": 0.3793, "step": 66300 }, { "epoch": 66.33, "grad_norm": 58.18625259399414, "learning_rate": 5.902939055264838e-06, "loss": 0.3904, "step": 66330 }, { "epoch": 66.36, "grad_norm": 113.20549774169922, "learning_rate": 5.89984893592795e-06, "loss": 0.3408, "step": 66360 }, { "epoch": 66.39, "grad_norm": 49.623043060302734, "learning_rate": 5.8967584613449525e-06, "loss": 0.3601, "step": 66390 }, { "epoch": 66.42, "grad_norm": 25.99820327758789, "learning_rate": 5.893667632735915e-06, "loss": 0.3732, "step": 66420 }, { "epoch": 66.45, "grad_norm": 62.764652252197266, "learning_rate": 5.89057645132105e-06, "loss": 0.4785, "step": 66450 }, { "epoch": 66.48, "grad_norm": 34.843502044677734, "learning_rate": 5.887484918320708e-06, "loss": 0.4353, "step": 66480 }, { "epoch": 66.51, "grad_norm": 72.27490234375, "learning_rate": 5.884393034955373e-06, "loss": 0.3543, "step": 66510 }, { "epoch": 66.54, "grad_norm": 56.0538444519043, "learning_rate": 5.881300802445675e-06, "loss": 0.3473, "step": 66540 }, { "epoch": 66.57, "grad_norm": 137.8314208984375, "learning_rate": 5.878208222012377e-06, "loss": 0.4599, "step": 66570 }, { "epoch": 66.6, "grad_norm": 61.8646240234375, "learning_rate": 5.8751152948763815e-06, "loss": 0.3729, "step": 66600 }, { "epoch": 66.63, "grad_norm": 31.587324142456055, "learning_rate": 5.872022022258726e-06, "loss": 0.3218, "step": 66630 }, { "epoch": 66.66, "grad_norm": 87.37081146240234, "learning_rate": 5.868928405380585e-06, "loss": 0.4447, "step": 66660 }, { "epoch": 66.69, "grad_norm": 84.91444396972656, "learning_rate": 5.865834445463273e-06, "loss": 0.3105, "step": 66690 }, { "epoch": 66.72, "grad_norm": 93.88651275634766, "learning_rate": 5.8627401437282334e-06, "loss": 0.3545, "step": 66720 }, { "epoch": 66.75, "grad_norm": 88.76152038574219, "learning_rate": 5.859645501397048e-06, "loss": 0.3613, "step": 66750 }, { "epoch": 66.78, "grad_norm": 86.98887634277344, "learning_rate": 5.856550519691433e-06, "loss": 0.3469, "step": 66780 }, { "epoch": 66.81, "grad_norm": 47.237857818603516, "learning_rate": 5.853455199833238e-06, "loss": 0.4362, "step": 66810 }, { "epoch": 66.84, "grad_norm": 87.42179870605469, "learning_rate": 5.850359543044446e-06, "loss": 0.3914, "step": 66840 }, { "epoch": 66.87, "grad_norm": 76.63813018798828, "learning_rate": 5.847263550547174e-06, "loss": 0.3684, "step": 66870 }, { "epoch": 66.9, "grad_norm": 112.3596420288086, "learning_rate": 5.844167223563669e-06, "loss": 0.3092, "step": 66900 }, { "epoch": 66.93, "grad_norm": 106.34989166259766, "learning_rate": 5.841070563316316e-06, "loss": 0.4097, "step": 66930 }, { "epoch": 66.96, "grad_norm": 59.01970672607422, "learning_rate": 5.837973571027621e-06, "loss": 0.3169, "step": 66960 }, { "epoch": 66.99, "grad_norm": 49.54807662963867, "learning_rate": 5.834876247920233e-06, "loss": 0.3657, "step": 66990 }, { "epoch": 67.0, "eval_loss": 0.27137479186058044, "eval_map": 0.8096, "eval_map_50": 0.9656, "eval_map_75": 0.9242, "eval_map_chicken": 0.8102, "eval_map_duck": 0.7621, "eval_map_large": 0.8191, "eval_map_medium": 0.799, "eval_map_plant": 0.8566, "eval_map_small": 0.2913, "eval_mar_1": 0.3297, "eval_mar_10": 0.8438, "eval_mar_100": 0.8486, "eval_mar_100_chicken": 0.8543, "eval_mar_100_duck": 0.8036, "eval_mar_100_plant": 0.8879, "eval_mar_large": 0.8561, "eval_mar_medium": 0.8418, "eval_mar_small": 0.4078, "eval_runtime": 12.9811, "eval_samples_per_second": 15.407, "eval_steps_per_second": 1.926, "step": 67000 }, { "epoch": 67.02, "grad_norm": 51.23720169067383, "learning_rate": 5.8317785952169245e-06, "loss": 0.4034, "step": 67020 }, { "epoch": 67.05, "grad_norm": 37.725894927978516, "learning_rate": 5.828680614140599e-06, "loss": 0.3372, "step": 67050 }, { "epoch": 67.08, "grad_norm": 36.62677764892578, "learning_rate": 5.82558230591429e-06, "loss": 0.3361, "step": 67080 }, { "epoch": 67.11, "grad_norm": 0.00012143549247412011, "learning_rate": 5.822483671761164e-06, "loss": 0.35, "step": 67110 }, { "epoch": 67.14, "grad_norm": 47.88471984863281, "learning_rate": 5.819384712904508e-06, "loss": 0.3816, "step": 67140 }, { "epoch": 67.17, "grad_norm": 64.09245300292969, "learning_rate": 5.816285430567743e-06, "loss": 0.4509, "step": 67170 }, { "epoch": 67.2, "grad_norm": 65.91133880615234, "learning_rate": 5.813185825974419e-06, "loss": 0.3435, "step": 67200 }, { "epoch": 67.23, "grad_norm": 40.91411209106445, "learning_rate": 5.810085900348209e-06, "loss": 0.3517, "step": 67230 }, { "epoch": 67.26, "grad_norm": 0.00036215735599398613, "learning_rate": 5.806985654912915e-06, "loss": 0.4363, "step": 67260 }, { "epoch": 67.29, "grad_norm": 26.14224624633789, "learning_rate": 5.803885090892464e-06, "loss": 0.2819, "step": 67290 }, { "epoch": 67.32, "grad_norm": 93.51468658447266, "learning_rate": 5.80078420951091e-06, "loss": 0.3572, "step": 67320 }, { "epoch": 67.35, "grad_norm": 776.4280395507812, "learning_rate": 5.797683011992432e-06, "loss": 0.6607, "step": 67350 }, { "epoch": 67.38, "grad_norm": 28.597293853759766, "learning_rate": 5.794581499561335e-06, "loss": 0.3935, "step": 67380 }, { "epoch": 67.41, "grad_norm": 41.74774169921875, "learning_rate": 5.791479673442044e-06, "loss": 0.3732, "step": 67410 }, { "epoch": 67.44, "grad_norm": 83.00664520263672, "learning_rate": 5.7883775348591146e-06, "loss": 0.3489, "step": 67440 }, { "epoch": 67.47, "grad_norm": 91.33666229248047, "learning_rate": 5.785275085037218e-06, "loss": 0.3392, "step": 67470 }, { "epoch": 67.5, "grad_norm": 0.00021223670046310872, "learning_rate": 5.782172325201155e-06, "loss": 0.4089, "step": 67500 }, { "epoch": 67.53, "grad_norm": 52.48198318481445, "learning_rate": 5.779069256575846e-06, "loss": 0.4223, "step": 67530 }, { "epoch": 67.56, "grad_norm": 43.82588195800781, "learning_rate": 5.77596588038633e-06, "loss": 0.4061, "step": 67560 }, { "epoch": 67.59, "grad_norm": 143.0668487548828, "learning_rate": 5.772862197857776e-06, "loss": 0.3519, "step": 67590 }, { "epoch": 67.62, "grad_norm": 84.51201629638672, "learning_rate": 5.769758210215466e-06, "loss": 0.3864, "step": 67620 }, { "epoch": 67.65, "grad_norm": 79.5323486328125, "learning_rate": 5.766653918684803e-06, "loss": 0.4039, "step": 67650 }, { "epoch": 67.68, "grad_norm": 39.521217346191406, "learning_rate": 5.763549324491317e-06, "loss": 0.3916, "step": 67680 }, { "epoch": 67.71, "grad_norm": 74.71115112304688, "learning_rate": 5.760444428860648e-06, "loss": 0.3364, "step": 67710 }, { "epoch": 67.74, "grad_norm": 44.37779235839844, "learning_rate": 5.757339233018563e-06, "loss": 0.4218, "step": 67740 }, { "epoch": 67.77, "grad_norm": 69.97649383544922, "learning_rate": 5.754233738190942e-06, "loss": 0.361, "step": 67770 }, { "epoch": 67.8, "grad_norm": 55.37105941772461, "learning_rate": 5.751127945603786e-06, "loss": 0.336, "step": 67800 }, { "epoch": 67.83, "grad_norm": 65.24737548828125, "learning_rate": 5.748021856483212e-06, "loss": 0.4023, "step": 67830 }, { "epoch": 67.86, "grad_norm": 32.11103439331055, "learning_rate": 5.744915472055457e-06, "loss": 0.3742, "step": 67860 }, { "epoch": 67.89, "grad_norm": 64.1340560913086, "learning_rate": 5.7418087935468706e-06, "loss": 0.3577, "step": 67890 }, { "epoch": 67.92, "grad_norm": 57.18985366821289, "learning_rate": 5.7387018221839195e-06, "loss": 0.4863, "step": 67920 }, { "epoch": 67.95, "grad_norm": 59.48006820678711, "learning_rate": 5.735594559193187e-06, "loss": 0.405, "step": 67950 }, { "epoch": 67.98, "grad_norm": 53.340641021728516, "learning_rate": 5.7324870058013736e-06, "loss": 0.3254, "step": 67980 }, { "epoch": 68.0, "eval_loss": 0.29342982172966003, "eval_map": 0.793, "eval_map_50": 0.9615, "eval_map_75": 0.919, "eval_map_chicken": 0.7734, "eval_map_duck": 0.7553, "eval_map_large": 0.8032, "eval_map_medium": 0.7855, "eval_map_plant": 0.8502, "eval_map_small": 0.2837, "eval_mar_1": 0.3232, "eval_mar_10": 0.8291, "eval_mar_100": 0.8325, "eval_mar_100_chicken": 0.8199, "eval_mar_100_duck": 0.7948, "eval_mar_100_plant": 0.8828, "eval_mar_large": 0.8382, "eval_mar_medium": 0.8287, "eval_mar_small": 0.3423, "eval_runtime": 12.595, "eval_samples_per_second": 15.879, "eval_steps_per_second": 1.985, "step": 68000 }, { "epoch": 68.01, "grad_norm": 45.32460403442383, "learning_rate": 5.72937916323529e-06, "loss": 0.3811, "step": 68010 }, { "epoch": 68.04, "grad_norm": 47.77042770385742, "learning_rate": 5.726271032721864e-06, "loss": 0.4686, "step": 68040 }, { "epoch": 68.07, "grad_norm": 0.0006070107920095325, "learning_rate": 5.723162615488137e-06, "loss": 0.3117, "step": 68070 }, { "epoch": 68.1, "grad_norm": 54.879634857177734, "learning_rate": 5.720053912761261e-06, "loss": 0.3715, "step": 68100 }, { "epoch": 68.13, "grad_norm": 57.85285949707031, "learning_rate": 5.716944925768505e-06, "loss": 0.3397, "step": 68130 }, { "epoch": 68.16, "grad_norm": 41.92170715332031, "learning_rate": 5.7138356557372444e-06, "loss": 0.3272, "step": 68160 }, { "epoch": 68.19, "grad_norm": 56.07709884643555, "learning_rate": 5.710726103894974e-06, "loss": 0.3121, "step": 68190 }, { "epoch": 68.22, "grad_norm": 52.071388244628906, "learning_rate": 5.707616271469293e-06, "loss": 0.2956, "step": 68220 }, { "epoch": 68.25, "grad_norm": 44.796112060546875, "learning_rate": 5.704506159687914e-06, "loss": 0.3782, "step": 68250 }, { "epoch": 68.28, "grad_norm": 57.30556106567383, "learning_rate": 5.70139576977866e-06, "loss": 0.3483, "step": 68280 }, { "epoch": 68.31, "grad_norm": 40.7261962890625, "learning_rate": 5.6982851029694645e-06, "loss": 0.3469, "step": 68310 }, { "epoch": 68.34, "grad_norm": 54.71175003051758, "learning_rate": 5.695174160488369e-06, "loss": 0.4215, "step": 68340 }, { "epoch": 68.37, "grad_norm": 79.7336196899414, "learning_rate": 5.692062943563525e-06, "loss": 0.3228, "step": 68370 }, { "epoch": 68.4, "grad_norm": 26.427806854248047, "learning_rate": 5.68895145342319e-06, "loss": 0.4407, "step": 68400 }, { "epoch": 68.43, "grad_norm": 120.46192932128906, "learning_rate": 5.685839691295734e-06, "loss": 0.4245, "step": 68430 }, { "epoch": 68.46, "grad_norm": 48.99250411987305, "learning_rate": 5.682727658409628e-06, "loss": 0.4054, "step": 68460 }, { "epoch": 68.49, "grad_norm": 73.3100814819336, "learning_rate": 5.679615355993455e-06, "loss": 0.4075, "step": 68490 }, { "epoch": 68.52, "grad_norm": 43.2242431640625, "learning_rate": 5.6765027852759015e-06, "loss": 0.3822, "step": 68520 }, { "epoch": 68.55, "grad_norm": 52.65727615356445, "learning_rate": 5.673389947485763e-06, "loss": 0.3209, "step": 68550 }, { "epoch": 68.58, "grad_norm": 56.2143669128418, "learning_rate": 5.670276843851939e-06, "loss": 0.2954, "step": 68580 }, { "epoch": 68.61, "grad_norm": 121.09834289550781, "learning_rate": 5.66716347560343e-06, "loss": 0.3024, "step": 68610 }, { "epoch": 68.64, "grad_norm": 94.31746673583984, "learning_rate": 5.664049843969348e-06, "loss": 0.34, "step": 68640 }, { "epoch": 68.67, "grad_norm": 73.45506286621094, "learning_rate": 5.660935950178904e-06, "loss": 0.3091, "step": 68670 }, { "epoch": 68.7, "grad_norm": 99.91253662109375, "learning_rate": 5.657821795461413e-06, "loss": 0.3845, "step": 68700 }, { "epoch": 68.73, "grad_norm": 29.421653747558594, "learning_rate": 5.654707381046296e-06, "loss": 0.3639, "step": 68730 }, { "epoch": 68.76, "grad_norm": 55.20606994628906, "learning_rate": 5.651592708163074e-06, "loss": 0.3613, "step": 68760 }, { "epoch": 68.79, "grad_norm": 38.0544319152832, "learning_rate": 5.648477778041369e-06, "loss": 0.5823, "step": 68790 }, { "epoch": 68.82, "grad_norm": 41.477378845214844, "learning_rate": 5.645362591910908e-06, "loss": 0.3317, "step": 68820 }, { "epoch": 68.85, "grad_norm": 69.14557647705078, "learning_rate": 5.642247151001515e-06, "loss": 0.3684, "step": 68850 }, { "epoch": 68.88, "grad_norm": 52.48639678955078, "learning_rate": 5.639131456543119e-06, "loss": 0.3629, "step": 68880 }, { "epoch": 68.91, "grad_norm": 62.84488296508789, "learning_rate": 5.636015509765747e-06, "loss": 0.4796, "step": 68910 }, { "epoch": 68.94, "grad_norm": 0.037693195044994354, "learning_rate": 5.6328993118995215e-06, "loss": 0.3382, "step": 68940 }, { "epoch": 68.97, "grad_norm": 40.709842681884766, "learning_rate": 5.629782864174672e-06, "loss": 0.3577, "step": 68970 }, { "epoch": 69.0, "grad_norm": 22.956600189208984, "learning_rate": 5.626666167821522e-06, "loss": 0.4403, "step": 69000 }, { "epoch": 69.0, "eval_loss": 0.26752638816833496, "eval_map": 0.8138, "eval_map_50": 0.9639, "eval_map_75": 0.9218, "eval_map_chicken": 0.8099, "eval_map_duck": 0.7637, "eval_map_large": 0.8154, "eval_map_medium": 0.8104, "eval_map_plant": 0.8677, "eval_map_small": 0.2836, "eval_mar_1": 0.3296, "eval_mar_10": 0.8477, "eval_mar_100": 0.8539, "eval_mar_100_chicken": 0.8592, "eval_mar_100_duck": 0.801, "eval_mar_100_plant": 0.9016, "eval_mar_large": 0.856, "eval_mar_medium": 0.8523, "eval_mar_small": 0.4037, "eval_runtime": 12.7529, "eval_samples_per_second": 15.683, "eval_steps_per_second": 1.96, "step": 69000 }, { "epoch": 69.03, "grad_norm": 92.33494567871094, "learning_rate": 5.623549224070494e-06, "loss": 0.3644, "step": 69030 }, { "epoch": 69.06, "grad_norm": 57.187320709228516, "learning_rate": 5.620432034152107e-06, "loss": 0.5907, "step": 69060 }, { "epoch": 69.09, "grad_norm": 96.00457000732422, "learning_rate": 5.617314599296977e-06, "loss": 0.3478, "step": 69090 }, { "epoch": 69.12, "grad_norm": 91.53950500488281, "learning_rate": 5.614196920735822e-06, "loss": 0.3611, "step": 69120 }, { "epoch": 69.15, "grad_norm": 38.11248779296875, "learning_rate": 5.611078999699448e-06, "loss": 0.4038, "step": 69150 }, { "epoch": 69.18, "grad_norm": 35.21555709838867, "learning_rate": 5.607960837418763e-06, "loss": 0.3148, "step": 69180 }, { "epoch": 69.21, "grad_norm": 46.0137825012207, "learning_rate": 5.604842435124769e-06, "loss": 0.3339, "step": 69210 }, { "epoch": 69.24, "grad_norm": 5.773234079242684e-05, "learning_rate": 5.601723794048558e-06, "loss": 0.3055, "step": 69240 }, { "epoch": 69.27, "grad_norm": 41.68159484863281, "learning_rate": 5.598604915421324e-06, "loss": 0.4445, "step": 69270 }, { "epoch": 69.3, "grad_norm": 51.43182373046875, "learning_rate": 5.59548580047435e-06, "loss": 0.4392, "step": 69300 }, { "epoch": 69.33, "grad_norm": 7.141924288589507e-05, "learning_rate": 5.592366450439012e-06, "loss": 0.3419, "step": 69330 }, { "epoch": 69.36, "grad_norm": 46.471370697021484, "learning_rate": 5.58924686654678e-06, "loss": 0.3784, "step": 69360 }, { "epoch": 69.39, "grad_norm": 47.60498046875, "learning_rate": 5.586127050029218e-06, "loss": 0.3363, "step": 69390 }, { "epoch": 69.42, "grad_norm": 28.752321243286133, "learning_rate": 5.5830070021179785e-06, "loss": 0.3755, "step": 69420 }, { "epoch": 69.45, "grad_norm": 90.1480712890625, "learning_rate": 5.5798867240448075e-06, "loss": 0.3435, "step": 69450 }, { "epoch": 69.48, "grad_norm": 44.21235275268555, "learning_rate": 5.576766217041541e-06, "loss": 0.3347, "step": 69480 }, { "epoch": 69.51, "grad_norm": 0.00018045748583972454, "learning_rate": 5.573645482340107e-06, "loss": 0.3765, "step": 69510 }, { "epoch": 69.54, "grad_norm": 44.780357360839844, "learning_rate": 5.570524521172523e-06, "loss": 0.3763, "step": 69540 }, { "epoch": 69.57, "grad_norm": 41.22140884399414, "learning_rate": 5.567403334770891e-06, "loss": 0.3662, "step": 69570 }, { "epoch": 69.6, "grad_norm": 88.37141418457031, "learning_rate": 5.5642819243674085e-06, "loss": 0.3863, "step": 69600 }, { "epoch": 69.63, "grad_norm": 106.2341079711914, "learning_rate": 5.56116029119436e-06, "loss": 0.322, "step": 69630 }, { "epoch": 69.66, "grad_norm": 55.32559585571289, "learning_rate": 5.558038436484116e-06, "loss": 0.5255, "step": 69660 }, { "epoch": 69.69, "grad_norm": 34.43977737426758, "learning_rate": 5.554916361469133e-06, "loss": 0.3735, "step": 69690 }, { "epoch": 69.72, "grad_norm": 0.00011480472312541679, "learning_rate": 5.551794067381959e-06, "loss": 0.3575, "step": 69720 }, { "epoch": 69.75, "grad_norm": 50.7276496887207, "learning_rate": 5.548671555455226e-06, "loss": 0.3473, "step": 69750 }, { "epoch": 69.78, "grad_norm": 46.63409423828125, "learning_rate": 5.545548826921653e-06, "loss": 0.4183, "step": 69780 }, { "epoch": 69.81, "grad_norm": 57.60160446166992, "learning_rate": 5.5424258830140434e-06, "loss": 0.3672, "step": 69810 }, { "epoch": 69.84, "grad_norm": 34.1253547668457, "learning_rate": 5.5393027249652844e-06, "loss": 0.3397, "step": 69840 }, { "epoch": 69.87, "grad_norm": 84.11296844482422, "learning_rate": 5.536179354008351e-06, "loss": 0.3478, "step": 69870 }, { "epoch": 69.9, "grad_norm": 39.57813262939453, "learning_rate": 5.5330557713763e-06, "loss": 0.3385, "step": 69900 }, { "epoch": 69.93, "grad_norm": 41.71063995361328, "learning_rate": 5.529931978302272e-06, "loss": 0.3416, "step": 69930 }, { "epoch": 69.96, "grad_norm": 53.061790466308594, "learning_rate": 5.526807976019492e-06, "loss": 0.3314, "step": 69960 }, { "epoch": 69.99, "grad_norm": 31.697439193725586, "learning_rate": 5.523683765761266e-06, "loss": 0.3707, "step": 69990 }, { "epoch": 70.0, "eval_loss": 0.2790444791316986, "eval_map": 0.8017, "eval_map_50": 0.9582, "eval_map_75": 0.923, "eval_map_chicken": 0.8024, "eval_map_duck": 0.7398, "eval_map_large": 0.8042, "eval_map_medium": 0.7982, "eval_map_plant": 0.8629, "eval_map_small": 0.2908, "eval_mar_1": 0.3243, "eval_mar_10": 0.8376, "eval_mar_100": 0.8421, "eval_mar_100_chicken": 0.8515, "eval_mar_100_duck": 0.7809, "eval_mar_100_plant": 0.8939, "eval_mar_large": 0.8458, "eval_mar_medium": 0.8389, "eval_mar_small": 0.398, "eval_runtime": 13.3516, "eval_samples_per_second": 14.979, "eval_steps_per_second": 1.872, "step": 70000 }, { "epoch": 70.02, "grad_norm": 64.09541320800781, "learning_rate": 5.520559348760984e-06, "loss": 0.3038, "step": 70020 }, { "epoch": 70.05, "grad_norm": 29.771255493164062, "learning_rate": 5.517434726252113e-06, "loss": 0.3184, "step": 70050 }, { "epoch": 70.08, "grad_norm": 62.01903533935547, "learning_rate": 5.514309899468209e-06, "loss": 0.3745, "step": 70080 }, { "epoch": 70.11, "grad_norm": 80.8722915649414, "learning_rate": 5.5111848696429005e-06, "loss": 0.3501, "step": 70110 }, { "epoch": 70.14, "grad_norm": 33.8654670715332, "learning_rate": 5.5080596380099e-06, "loss": 0.3372, "step": 70140 }, { "epoch": 70.17, "grad_norm": 51.104618072509766, "learning_rate": 5.504934205803002e-06, "loss": 0.3181, "step": 70170 }, { "epoch": 70.2, "grad_norm": 167.69021606445312, "learning_rate": 5.5018085742560745e-06, "loss": 0.3563, "step": 70200 }, { "epoch": 70.23, "grad_norm": 57.787498474121094, "learning_rate": 5.498682744603071e-06, "loss": 0.4323, "step": 70230 }, { "epoch": 70.26, "grad_norm": 37.657772064208984, "learning_rate": 5.495556718078017e-06, "loss": 0.2698, "step": 70260 }, { "epoch": 70.29, "grad_norm": 56.136741638183594, "learning_rate": 5.492430495915018e-06, "loss": 0.444, "step": 70290 }, { "epoch": 70.32, "grad_norm": 122.20323944091797, "learning_rate": 5.489304079348259e-06, "loss": 0.3838, "step": 70320 }, { "epoch": 70.35, "grad_norm": 78.22285461425781, "learning_rate": 5.486177469611999e-06, "loss": 0.3603, "step": 70350 }, { "epoch": 70.38, "grad_norm": 31.24813461303711, "learning_rate": 5.483050667940571e-06, "loss": 0.3721, "step": 70380 }, { "epoch": 70.41, "grad_norm": 31.084609985351562, "learning_rate": 5.4799236755683916e-06, "loss": 0.3713, "step": 70410 }, { "epoch": 70.44, "grad_norm": 83.49063873291016, "learning_rate": 5.476796493729943e-06, "loss": 0.3351, "step": 70440 }, { "epoch": 70.47, "grad_norm": 108.77242279052734, "learning_rate": 5.473669123659793e-06, "loss": 0.3812, "step": 70470 }, { "epoch": 70.5, "grad_norm": 31.273191452026367, "learning_rate": 5.470541566592573e-06, "loss": 0.3376, "step": 70500 }, { "epoch": 70.53, "grad_norm": 44.0676155090332, "learning_rate": 5.467413823762994e-06, "loss": 0.3808, "step": 70530 }, { "epoch": 70.56, "grad_norm": 59.492095947265625, "learning_rate": 5.46428589640584e-06, "loss": 0.3879, "step": 70560 }, { "epoch": 70.59, "grad_norm": 71.71273803710938, "learning_rate": 5.4611577857559676e-06, "loss": 0.3732, "step": 70590 }, { "epoch": 70.62, "grad_norm": 82.30093383789062, "learning_rate": 5.458029493048303e-06, "loss": 0.3454, "step": 70620 }, { "epoch": 70.65, "grad_norm": 69.27047729492188, "learning_rate": 5.454901019517851e-06, "loss": 0.377, "step": 70650 }, { "epoch": 70.68, "grad_norm": 73.74759674072266, "learning_rate": 5.451772366399678e-06, "loss": 0.3732, "step": 70680 }, { "epoch": 70.71, "grad_norm": 36.474246978759766, "learning_rate": 5.448643534928931e-06, "loss": 0.2819, "step": 70710 }, { "epoch": 70.74, "grad_norm": 62.831058502197266, "learning_rate": 5.445514526340822e-06, "loss": 0.4339, "step": 70740 }, { "epoch": 70.77, "grad_norm": 35.08445739746094, "learning_rate": 5.442385341870633e-06, "loss": 0.3934, "step": 70770 }, { "epoch": 70.8, "grad_norm": 87.93460845947266, "learning_rate": 5.439255982753717e-06, "loss": 0.3537, "step": 70800 }, { "epoch": 70.83, "grad_norm": 43.69154357910156, "learning_rate": 5.436126450225495e-06, "loss": 0.3763, "step": 70830 }, { "epoch": 70.86, "grad_norm": 69.78821563720703, "learning_rate": 5.432996745521458e-06, "loss": 0.3705, "step": 70860 }, { "epoch": 70.89, "grad_norm": 47.48979568481445, "learning_rate": 5.429866869877163e-06, "loss": 0.2934, "step": 70890 }, { "epoch": 70.92, "grad_norm": 67.55951690673828, "learning_rate": 5.426736824528236e-06, "loss": 0.3479, "step": 70920 }, { "epoch": 70.95, "grad_norm": 42.52537155151367, "learning_rate": 5.423606610710368e-06, "loss": 0.4489, "step": 70950 }, { "epoch": 70.98, "grad_norm": 40.568565368652344, "learning_rate": 5.420476229659319e-06, "loss": 0.3141, "step": 70980 }, { "epoch": 71.0, "eval_loss": 0.2896789312362671, "eval_map": 0.791, "eval_map_50": 0.9656, "eval_map_75": 0.9244, "eval_map_chicken": 0.7686, "eval_map_duck": 0.7479, "eval_map_large": 0.81, "eval_map_medium": 0.7846, "eval_map_plant": 0.8564, "eval_map_small": 0.2887, "eval_mar_1": 0.3221, "eval_mar_10": 0.8283, "eval_mar_100": 0.8329, "eval_mar_100_chicken": 0.8171, "eval_mar_100_duck": 0.7892, "eval_mar_100_plant": 0.8924, "eval_mar_large": 0.8479, "eval_mar_medium": 0.8297, "eval_mar_small": 0.3876, "eval_runtime": 12.7086, "eval_samples_per_second": 15.737, "eval_steps_per_second": 1.967, "step": 71000 }, { "epoch": 71.01, "grad_norm": 152.8446502685547, "learning_rate": 5.417345682610914e-06, "loss": 0.3195, "step": 71010 }, { "epoch": 71.04, "grad_norm": 264.1171875, "learning_rate": 5.414214970801041e-06, "loss": 0.3315, "step": 71040 }, { "epoch": 71.07, "grad_norm": 43.089656829833984, "learning_rate": 5.411084095465661e-06, "loss": 0.3085, "step": 71070 }, { "epoch": 71.1, "grad_norm": 74.4775619506836, "learning_rate": 5.4079530578407895e-06, "loss": 0.3434, "step": 71100 }, { "epoch": 71.13, "grad_norm": 55.117305755615234, "learning_rate": 5.404821859162509e-06, "loss": 0.301, "step": 71130 }, { "epoch": 71.16, "grad_norm": 42.02001953125, "learning_rate": 5.401690500666972e-06, "loss": 0.3933, "step": 71160 }, { "epoch": 71.19, "grad_norm": 174.26234436035156, "learning_rate": 5.398558983590385e-06, "loss": 0.3807, "step": 71190 }, { "epoch": 71.22, "grad_norm": 71.86639404296875, "learning_rate": 5.3954273091690245e-06, "loss": 0.3754, "step": 71220 }, { "epoch": 71.25, "grad_norm": 40.899742126464844, "learning_rate": 5.392295478639226e-06, "loss": 0.4152, "step": 71250 }, { "epoch": 71.28, "grad_norm": 90.44598388671875, "learning_rate": 5.389163493237382e-06, "loss": 0.3662, "step": 71280 }, { "epoch": 71.31, "grad_norm": 73.8045654296875, "learning_rate": 5.386031354199956e-06, "loss": 0.3433, "step": 71310 }, { "epoch": 71.34, "grad_norm": 28.936880111694336, "learning_rate": 5.3828990627634655e-06, "loss": 0.284, "step": 71340 }, { "epoch": 71.37, "grad_norm": 68.62113189697266, "learning_rate": 5.379766620164488e-06, "loss": 0.3144, "step": 71370 }, { "epoch": 71.4, "grad_norm": 38.200645446777344, "learning_rate": 5.376634027639664e-06, "loss": 0.3682, "step": 71400 }, { "epoch": 71.43, "grad_norm": 47.5620231628418, "learning_rate": 5.373501286425691e-06, "loss": 0.3, "step": 71430 }, { "epoch": 71.46, "grad_norm": 28.642446517944336, "learning_rate": 5.370368397759324e-06, "loss": 0.3528, "step": 71460 }, { "epoch": 71.49, "grad_norm": 55.42530822753906, "learning_rate": 5.367235362877379e-06, "loss": 0.3612, "step": 71490 }, { "epoch": 71.52, "grad_norm": 55.24734115600586, "learning_rate": 5.36410218301673e-06, "loss": 0.3402, "step": 71520 }, { "epoch": 71.55, "grad_norm": 44.40134048461914, "learning_rate": 5.360968859414305e-06, "loss": 0.3927, "step": 71550 }, { "epoch": 71.58, "grad_norm": 34.57560348510742, "learning_rate": 5.357835393307089e-06, "loss": 0.3442, "step": 71580 }, { "epoch": 71.61, "grad_norm": 49.06620407104492, "learning_rate": 5.354701785932129e-06, "loss": 0.3375, "step": 71610 }, { "epoch": 71.64, "grad_norm": 39.409034729003906, "learning_rate": 5.35156803852652e-06, "loss": 0.3407, "step": 71640 }, { "epoch": 71.67, "grad_norm": 47.109378814697266, "learning_rate": 5.348434152327418e-06, "loss": 0.4018, "step": 71670 }, { "epoch": 71.7, "grad_norm": 85.60104370117188, "learning_rate": 5.345300128572031e-06, "loss": 0.39, "step": 71700 }, { "epoch": 71.73, "grad_norm": 67.191650390625, "learning_rate": 5.3421659684976205e-06, "loss": 0.3743, "step": 71730 }, { "epoch": 71.76, "grad_norm": 89.3816909790039, "learning_rate": 5.339031673341505e-06, "loss": 0.3318, "step": 71760 }, { "epoch": 71.79, "grad_norm": 57.99645233154297, "learning_rate": 5.335897244341054e-06, "loss": 0.3275, "step": 71790 }, { "epoch": 71.82, "grad_norm": 61.56242752075195, "learning_rate": 5.3327626827336906e-06, "loss": 0.3804, "step": 71820 }, { "epoch": 71.85, "grad_norm": 36.96889114379883, "learning_rate": 5.32962798975689e-06, "loss": 0.3385, "step": 71850 }, { "epoch": 71.88, "grad_norm": 31.221084594726562, "learning_rate": 5.326493166648179e-06, "loss": 0.3096, "step": 71880 }, { "epoch": 71.91, "grad_norm": 80.34669494628906, "learning_rate": 5.3233582146451375e-06, "loss": 0.3685, "step": 71910 }, { "epoch": 71.94, "grad_norm": 72.47837829589844, "learning_rate": 5.320223134985393e-06, "loss": 0.3546, "step": 71940 }, { "epoch": 71.97, "grad_norm": 0.00015040226571727544, "learning_rate": 5.3170879289066265e-06, "loss": 0.3805, "step": 71970 }, { "epoch": 72.0, "grad_norm": 63.801326751708984, "learning_rate": 5.3139525976465675e-06, "loss": 0.3931, "step": 72000 }, { "epoch": 72.0, "eval_loss": 0.2895579934120178, "eval_map": 0.7999, "eval_map_50": 0.9558, "eval_map_75": 0.9205, "eval_map_chicken": 0.7957, "eval_map_duck": 0.7506, "eval_map_large": 0.8152, "eval_map_medium": 0.7915, "eval_map_plant": 0.8534, "eval_map_small": 0.2829, "eval_mar_1": 0.3241, "eval_mar_10": 0.8333, "eval_mar_100": 0.8371, "eval_mar_100_chicken": 0.8416, "eval_mar_100_duck": 0.784, "eval_mar_100_plant": 0.8857, "eval_mar_large": 0.8531, "eval_mar_medium": 0.8314, "eval_mar_small": 0.3537, "eval_runtime": 12.5091, "eval_samples_per_second": 15.988, "eval_steps_per_second": 1.999, "step": 72000 }, { "epoch": 72.03, "grad_norm": 27.810344696044922, "learning_rate": 5.310817142442995e-06, "loss": 0.3973, "step": 72030 }, { "epoch": 72.06, "grad_norm": 67.50765991210938, "learning_rate": 5.307681564533736e-06, "loss": 0.3664, "step": 72060 }, { "epoch": 72.09, "grad_norm": 69.35010528564453, "learning_rate": 5.304545865156669e-06, "loss": 0.3599, "step": 72090 }, { "epoch": 72.12, "grad_norm": 100.0123062133789, "learning_rate": 5.301410045549719e-06, "loss": 0.3718, "step": 72120 }, { "epoch": 72.15, "grad_norm": 53.47022247314453, "learning_rate": 5.298274106950855e-06, "loss": 0.3781, "step": 72150 }, { "epoch": 72.18, "grad_norm": 44.83963394165039, "learning_rate": 5.295138050598097e-06, "loss": 0.3334, "step": 72180 }, { "epoch": 72.21, "grad_norm": 79.38215637207031, "learning_rate": 5.29200187772951e-06, "loss": 0.3533, "step": 72210 }, { "epoch": 72.24, "grad_norm": 80.02742004394531, "learning_rate": 5.2888655895832075e-06, "loss": 0.3456, "step": 72240 }, { "epoch": 72.27, "grad_norm": 54.46184158325195, "learning_rate": 5.285729187397344e-06, "loss": 0.4275, "step": 72270 }, { "epoch": 72.3, "grad_norm": 65.68153381347656, "learning_rate": 5.282592672410124e-06, "loss": 0.3015, "step": 72300 }, { "epoch": 72.33, "grad_norm": 27.298728942871094, "learning_rate": 5.27945604585979e-06, "loss": 0.3267, "step": 72330 }, { "epoch": 72.36, "grad_norm": 42.07027053833008, "learning_rate": 5.276319308984637e-06, "loss": 0.3504, "step": 72360 }, { "epoch": 72.39, "grad_norm": 108.28820037841797, "learning_rate": 5.2731824630229955e-06, "loss": 0.3578, "step": 72390 }, { "epoch": 72.42, "grad_norm": 42.1929817199707, "learning_rate": 5.270045509213244e-06, "loss": 0.3428, "step": 72420 }, { "epoch": 72.45, "grad_norm": 22.351905822753906, "learning_rate": 5.2669084487938025e-06, "loss": 0.3426, "step": 72450 }, { "epoch": 72.48, "grad_norm": 85.86829376220703, "learning_rate": 5.263771283003133e-06, "loss": 0.3663, "step": 72480 }, { "epoch": 72.51, "grad_norm": 37.760894775390625, "learning_rate": 5.260634013079738e-06, "loss": 0.3303, "step": 72510 }, { "epoch": 72.54, "grad_norm": 24.47450828552246, "learning_rate": 5.2574966402621615e-06, "loss": 0.3886, "step": 72540 }, { "epoch": 72.57, "grad_norm": 56.139835357666016, "learning_rate": 5.25435916578899e-06, "loss": 0.3657, "step": 72570 }, { "epoch": 72.6, "grad_norm": 116.27320861816406, "learning_rate": 5.251221590898848e-06, "loss": 0.2946, "step": 72600 }, { "epoch": 72.63, "grad_norm": 50.227054595947266, "learning_rate": 5.2480839168304e-06, "loss": 0.3076, "step": 72630 }, { "epoch": 72.66, "grad_norm": 26.720067977905273, "learning_rate": 5.244946144822351e-06, "loss": 0.4743, "step": 72660 }, { "epoch": 72.69, "grad_norm": 40.32868194580078, "learning_rate": 5.241808276113445e-06, "loss": 0.3434, "step": 72690 }, { "epoch": 72.72, "grad_norm": 145.30947875976562, "learning_rate": 5.238670311942459e-06, "loss": 0.3375, "step": 72720 }, { "epoch": 72.75, "grad_norm": 0.00016111430886667222, "learning_rate": 5.235532253548213e-06, "loss": 0.3079, "step": 72750 }, { "epoch": 72.78, "grad_norm": 53.08110046386719, "learning_rate": 5.232394102169566e-06, "loss": 0.3614, "step": 72780 }, { "epoch": 72.81, "grad_norm": 77.44256591796875, "learning_rate": 5.229255859045405e-06, "loss": 0.4356, "step": 72810 }, { "epoch": 72.84, "grad_norm": 40.184104919433594, "learning_rate": 5.226117525414663e-06, "loss": 0.3704, "step": 72840 }, { "epoch": 72.87, "grad_norm": 29.905786514282227, "learning_rate": 5.222979102516304e-06, "loss": 0.4326, "step": 72870 }, { "epoch": 72.9, "grad_norm": 68.46060180664062, "learning_rate": 5.219840591589325e-06, "loss": 0.327, "step": 72900 }, { "epoch": 72.93, "grad_norm": 59.05925750732422, "learning_rate": 5.216701993872763e-06, "loss": 0.3277, "step": 72930 }, { "epoch": 72.96, "grad_norm": 28.046104431152344, "learning_rate": 5.213563310605686e-06, "loss": 0.372, "step": 72960 }, { "epoch": 72.99, "grad_norm": 56.09440994262695, "learning_rate": 5.210424543027195e-06, "loss": 0.3108, "step": 72990 }, { "epoch": 73.0, "eval_loss": 0.2754177451133728, "eval_map": 0.8067, "eval_map_50": 0.9674, "eval_map_75": 0.9317, "eval_map_chicken": 0.7807, "eval_map_duck": 0.7709, "eval_map_large": 0.802, "eval_map_medium": 0.8031, "eval_map_plant": 0.8684, "eval_map_small": 0.309, "eval_mar_1": 0.3267, "eval_mar_10": 0.8413, "eval_mar_100": 0.8465, "eval_mar_100_chicken": 0.8288, "eval_mar_100_duck": 0.8124, "eval_mar_100_plant": 0.8983, "eval_mar_large": 0.8441, "eval_mar_medium": 0.8443, "eval_mar_small": 0.4119, "eval_runtime": 13.2614, "eval_samples_per_second": 15.081, "eval_steps_per_second": 1.885, "step": 73000 }, { "epoch": 73.02, "grad_norm": 56.69845962524414, "learning_rate": 5.207285692376427e-06, "loss": 0.3384, "step": 73020 }, { "epoch": 73.05, "grad_norm": 67.52530670166016, "learning_rate": 5.204146759892551e-06, "loss": 0.2758, "step": 73050 }, { "epoch": 73.08, "grad_norm": 47.884361267089844, "learning_rate": 5.201007746814767e-06, "loss": 0.3268, "step": 73080 }, { "epoch": 73.11, "grad_norm": 36.75482940673828, "learning_rate": 5.197868654382307e-06, "loss": 0.4091, "step": 73110 }, { "epoch": 73.14, "grad_norm": 60.41285705566406, "learning_rate": 5.194729483834438e-06, "loss": 0.3125, "step": 73140 }, { "epoch": 73.17, "grad_norm": 25.865018844604492, "learning_rate": 5.191590236410451e-06, "loss": 0.2942, "step": 73170 }, { "epoch": 73.2, "grad_norm": 0.0005106661119498312, "learning_rate": 5.188450913349674e-06, "loss": 0.3198, "step": 73200 }, { "epoch": 73.23, "grad_norm": 81.65201568603516, "learning_rate": 5.1853115158914595e-06, "loss": 0.3949, "step": 73230 }, { "epoch": 73.26, "grad_norm": 277.5521545410156, "learning_rate": 5.1821720452751945e-06, "loss": 0.3401, "step": 73260 }, { "epoch": 73.29, "grad_norm": 37.81293869018555, "learning_rate": 5.179032502740291e-06, "loss": 0.3983, "step": 73290 }, { "epoch": 73.32, "grad_norm": 27.489107131958008, "learning_rate": 5.175892889526189e-06, "loss": 0.4043, "step": 73320 }, { "epoch": 73.35, "grad_norm": 53.394649505615234, "learning_rate": 5.172753206872363e-06, "loss": 0.4057, "step": 73350 }, { "epoch": 73.38, "grad_norm": 105.8538818359375, "learning_rate": 5.1696134560183045e-06, "loss": 0.3575, "step": 73380 }, { "epoch": 73.41, "grad_norm": 46.337913513183594, "learning_rate": 5.166473638203539e-06, "loss": 0.361, "step": 73410 }, { "epoch": 73.44, "grad_norm": 43.64438247680664, "learning_rate": 5.16333375466762e-06, "loss": 0.4102, "step": 73440 }, { "epoch": 73.47, "grad_norm": 62.29459762573242, "learning_rate": 5.16019380665012e-06, "loss": 0.3643, "step": 73470 }, { "epoch": 73.5, "grad_norm": 66.70515441894531, "learning_rate": 5.157053795390642e-06, "loss": 0.343, "step": 73500 }, { "epoch": 73.53, "grad_norm": 63.99614715576172, "learning_rate": 5.153913722128813e-06, "loss": 0.4231, "step": 73530 }, { "epoch": 73.56, "grad_norm": 56.739009857177734, "learning_rate": 5.150773588104284e-06, "loss": 0.3856, "step": 73560 }, { "epoch": 73.59, "grad_norm": 76.93829345703125, "learning_rate": 5.147633394556731e-06, "loss": 0.4386, "step": 73590 }, { "epoch": 73.62, "grad_norm": 7.817478763172403e-05, "learning_rate": 5.144493142725851e-06, "loss": 0.3486, "step": 73620 }, { "epoch": 73.65, "grad_norm": 63.201168060302734, "learning_rate": 5.141352833851367e-06, "loss": 0.3324, "step": 73650 }, { "epoch": 73.68, "grad_norm": 57.90425491333008, "learning_rate": 5.138212469173022e-06, "loss": 0.3737, "step": 73680 }, { "epoch": 73.71, "grad_norm": 38.61542510986328, "learning_rate": 5.135072049930584e-06, "loss": 0.3024, "step": 73710 }, { "epoch": 73.74, "grad_norm": 43.024593353271484, "learning_rate": 5.13193157736384e-06, "loss": 0.3908, "step": 73740 }, { "epoch": 73.77, "grad_norm": 50.603065490722656, "learning_rate": 5.128791052712597e-06, "loss": 0.3225, "step": 73770 }, { "epoch": 73.8, "grad_norm": 113.19688415527344, "learning_rate": 5.1256504772166885e-06, "loss": 0.4171, "step": 73800 }, { "epoch": 73.83, "grad_norm": 73.72309875488281, "learning_rate": 5.122509852115959e-06, "loss": 0.395, "step": 73830 }, { "epoch": 73.86, "grad_norm": 76.61473083496094, "learning_rate": 5.119369178650282e-06, "loss": 0.4381, "step": 73860 }, { "epoch": 73.89, "grad_norm": 58.34016036987305, "learning_rate": 5.116228458059544e-06, "loss": 0.2847, "step": 73890 }, { "epoch": 73.92, "grad_norm": 67.36154174804688, "learning_rate": 5.1130876915836495e-06, "loss": 0.3588, "step": 73920 }, { "epoch": 73.95, "grad_norm": 1.792973853298463e-05, "learning_rate": 5.109946880462526e-06, "loss": 0.3316, "step": 73950 }, { "epoch": 73.98, "grad_norm": 49.80463790893555, "learning_rate": 5.1068060259361155e-06, "loss": 0.3259, "step": 73980 }, { "epoch": 74.0, "eval_loss": 0.2741202414035797, "eval_map": 0.8073, "eval_map_50": 0.9645, "eval_map_75": 0.9314, "eval_map_chicken": 0.8031, "eval_map_duck": 0.7619, "eval_map_large": 0.8422, "eval_map_medium": 0.7985, "eval_map_plant": 0.857, "eval_map_small": 0.2921, "eval_mar_1": 0.327, "eval_mar_10": 0.8429, "eval_mar_100": 0.8477, "eval_mar_100_chicken": 0.8497, "eval_mar_100_duck": 0.8031, "eval_mar_100_plant": 0.8902, "eval_mar_large": 0.8717, "eval_mar_medium": 0.8404, "eval_mar_small": 0.4056, "eval_runtime": 12.4942, "eval_samples_per_second": 16.007, "eval_steps_per_second": 2.001, "step": 74000 }, { "epoch": 74.01, "grad_norm": 69.7426528930664, "learning_rate": 5.1036651292443774e-06, "loss": 0.2974, "step": 74010 }, { "epoch": 74.04, "grad_norm": 56.65920639038086, "learning_rate": 5.100524191627289e-06, "loss": 0.3357, "step": 74040 }, { "epoch": 74.07, "grad_norm": 81.04058074951172, "learning_rate": 5.0973832143248405e-06, "loss": 0.3524, "step": 74070 }, { "epoch": 74.1, "grad_norm": 69.03898620605469, "learning_rate": 5.0942421985770415e-06, "loss": 0.3457, "step": 74100 }, { "epoch": 74.13, "grad_norm": 93.84957885742188, "learning_rate": 5.091101145623916e-06, "loss": 0.2919, "step": 74130 }, { "epoch": 74.16, "grad_norm": 73.12716674804688, "learning_rate": 5.087960056705499e-06, "loss": 0.3447, "step": 74160 }, { "epoch": 74.19, "grad_norm": 39.820064544677734, "learning_rate": 5.084818933061846e-06, "loss": 0.3187, "step": 74190 }, { "epoch": 74.22, "grad_norm": 260.0841979980469, "learning_rate": 5.0816777759330215e-06, "loss": 0.3785, "step": 74220 }, { "epoch": 74.25, "grad_norm": 33.61313247680664, "learning_rate": 5.078536586559104e-06, "loss": 0.3682, "step": 74250 }, { "epoch": 74.28, "grad_norm": 43.0290412902832, "learning_rate": 5.075395366180186e-06, "loss": 0.3695, "step": 74280 }, { "epoch": 74.31, "grad_norm": 73.47203826904297, "learning_rate": 5.072254116036372e-06, "loss": 0.3838, "step": 74310 }, { "epoch": 74.34, "grad_norm": 208.06161499023438, "learning_rate": 5.069112837367777e-06, "loss": 0.4251, "step": 74340 }, { "epoch": 74.37, "grad_norm": 49.6225700378418, "learning_rate": 5.065971531414528e-06, "loss": 0.4086, "step": 74370 }, { "epoch": 74.4, "grad_norm": 133.38600158691406, "learning_rate": 5.062830199416764e-06, "loss": 0.3682, "step": 74400 }, { "epoch": 74.43, "grad_norm": 88.1202392578125, "learning_rate": 5.0596888426146325e-06, "loss": 0.3222, "step": 74430 }, { "epoch": 74.46, "grad_norm": 39.13841247558594, "learning_rate": 5.05654746224829e-06, "loss": 0.29, "step": 74460 }, { "epoch": 74.49, "grad_norm": 151.9517059326172, "learning_rate": 5.053406059557906e-06, "loss": 0.4219, "step": 74490 }, { "epoch": 74.52, "grad_norm": 110.44225311279297, "learning_rate": 5.050264635783654e-06, "loss": 0.3411, "step": 74520 }, { "epoch": 74.55, "grad_norm": 147.96983337402344, "learning_rate": 5.047123192165722e-06, "loss": 0.3825, "step": 74550 }, { "epoch": 74.58, "grad_norm": 37.00914001464844, "learning_rate": 5.043981729944298e-06, "loss": 0.3879, "step": 74580 }, { "epoch": 74.61, "grad_norm": 171.95347595214844, "learning_rate": 5.040840250359584e-06, "loss": 0.4454, "step": 74610 }, { "epoch": 74.64, "grad_norm": 181.46224975585938, "learning_rate": 5.037698754651786e-06, "loss": 0.3304, "step": 74640 }, { "epoch": 74.67, "grad_norm": 51.033164978027344, "learning_rate": 5.034557244061117e-06, "loss": 0.3387, "step": 74670 }, { "epoch": 74.7, "grad_norm": 50.1196403503418, "learning_rate": 5.031415719827796e-06, "loss": 0.3201, "step": 74700 }, { "epoch": 74.73, "grad_norm": 61.087486267089844, "learning_rate": 5.028274183192046e-06, "loss": 0.3098, "step": 74730 }, { "epoch": 74.76, "grad_norm": 78.54173278808594, "learning_rate": 5.025132635394095e-06, "loss": 0.3411, "step": 74760 }, { "epoch": 74.79, "grad_norm": 25.6117000579834, "learning_rate": 5.021991077674179e-06, "loss": 0.3667, "step": 74790 }, { "epoch": 74.82, "grad_norm": 121.82492065429688, "learning_rate": 5.018849511272532e-06, "loss": 0.3731, "step": 74820 }, { "epoch": 74.85, "grad_norm": 109.08409881591797, "learning_rate": 5.015707937429398e-06, "loss": 0.4268, "step": 74850 }, { "epoch": 74.88, "grad_norm": 123.82302856445312, "learning_rate": 5.0125663573850204e-06, "loss": 0.345, "step": 74880 }, { "epoch": 74.91, "grad_norm": 78.83377838134766, "learning_rate": 5.0094247723796405e-06, "loss": 0.3758, "step": 74910 }, { "epoch": 74.94, "grad_norm": 56.73345184326172, "learning_rate": 5.006283183653513e-06, "loss": 0.3211, "step": 74940 }, { "epoch": 74.97, "grad_norm": 175.68844604492188, "learning_rate": 5.0031415924468816e-06, "loss": 0.3845, "step": 74970 }, { "epoch": 75.0, "grad_norm": 62.250694274902344, "learning_rate": 5e-06, "loss": 0.3673, "step": 75000 }, { "epoch": 75.0, "eval_loss": 0.2774085998535156, "eval_map": 0.8075, "eval_map_50": 0.9697, "eval_map_75": 0.9246, "eval_map_chicken": 0.7883, "eval_map_duck": 0.7663, "eval_map_large": 0.822, "eval_map_medium": 0.8, "eval_map_plant": 0.8681, "eval_map_small": 0.2445, "eval_mar_1": 0.3252, "eval_mar_10": 0.8427, "eval_mar_100": 0.8478, "eval_mar_100_chicken": 0.8348, "eval_mar_100_duck": 0.8098, "eval_mar_100_plant": 0.8988, "eval_mar_large": 0.8602, "eval_mar_medium": 0.8408, "eval_mar_small": 0.4028, "eval_runtime": 12.682, "eval_samples_per_second": 15.77, "eval_steps_per_second": 1.971, "step": 75000 }, { "epoch": 75.03, "grad_norm": 96.09950256347656, "learning_rate": 4.996858407553119e-06, "loss": 0.4151, "step": 75030 }, { "epoch": 75.06, "grad_norm": 60.19258117675781, "learning_rate": 4.99371681634649e-06, "loss": 0.3816, "step": 75060 }, { "epoch": 75.09, "grad_norm": 65.95404052734375, "learning_rate": 4.990575227620359e-06, "loss": 0.3561, "step": 75090 }, { "epoch": 75.12, "grad_norm": 16.284788131713867, "learning_rate": 4.987433642614981e-06, "loss": 0.3475, "step": 75120 }, { "epoch": 75.15, "grad_norm": 101.0333251953125, "learning_rate": 4.984292062570603e-06, "loss": 0.3657, "step": 75150 }, { "epoch": 75.18, "grad_norm": 106.7493667602539, "learning_rate": 4.981150488727469e-06, "loss": 0.4163, "step": 75180 }, { "epoch": 75.21, "grad_norm": 62.466365814208984, "learning_rate": 4.9780089223258235e-06, "loss": 0.3467, "step": 75210 }, { "epoch": 75.24, "grad_norm": 119.90625762939453, "learning_rate": 4.974867364605906e-06, "loss": 0.388, "step": 75240 }, { "epoch": 75.27, "grad_norm": 53.914730072021484, "learning_rate": 4.971725816807956e-06, "loss": 0.3436, "step": 75270 }, { "epoch": 75.3, "grad_norm": 48.6357421875, "learning_rate": 4.968584280172206e-06, "loss": 0.3435, "step": 75300 }, { "epoch": 75.33, "grad_norm": 67.97750854492188, "learning_rate": 4.9654427559388845e-06, "loss": 0.3318, "step": 75330 }, { "epoch": 75.36, "grad_norm": 20.656139373779297, "learning_rate": 4.962301245348215e-06, "loss": 0.3684, "step": 75360 }, { "epoch": 75.39, "grad_norm": 37.04246139526367, "learning_rate": 4.9591597496404165e-06, "loss": 0.4572, "step": 75390 }, { "epoch": 75.42, "grad_norm": 39.73920822143555, "learning_rate": 4.956018270055703e-06, "loss": 0.3956, "step": 75420 }, { "epoch": 75.45, "grad_norm": 152.38929748535156, "learning_rate": 4.952876807834281e-06, "loss": 0.36, "step": 75450 }, { "epoch": 75.48, "grad_norm": 105.62199401855469, "learning_rate": 4.949735364216348e-06, "loss": 0.371, "step": 75480 }, { "epoch": 75.51, "grad_norm": 23.39995765686035, "learning_rate": 4.946593940442097e-06, "loss": 0.3813, "step": 75510 }, { "epoch": 75.54, "grad_norm": 102.3947982788086, "learning_rate": 4.9434525377517115e-06, "loss": 0.4903, "step": 75540 }, { "epoch": 75.57, "grad_norm": 22.25848388671875, "learning_rate": 4.940311157385369e-06, "loss": 0.3805, "step": 75570 }, { "epoch": 75.6, "grad_norm": 0.0001872582215582952, "learning_rate": 4.937169800583237e-06, "loss": 0.284, "step": 75600 }, { "epoch": 75.63, "grad_norm": 34.863197326660156, "learning_rate": 4.934028468585473e-06, "loss": 0.349, "step": 75630 }, { "epoch": 75.66, "grad_norm": 93.2557144165039, "learning_rate": 4.930887162632225e-06, "loss": 0.4011, "step": 75660 }, { "epoch": 75.69, "grad_norm": 30.740509033203125, "learning_rate": 4.927745883963629e-06, "loss": 0.3169, "step": 75690 }, { "epoch": 75.72, "grad_norm": 94.27088928222656, "learning_rate": 4.924604633819815e-06, "loss": 0.3957, "step": 75720 }, { "epoch": 75.75, "grad_norm": 48.359195709228516, "learning_rate": 4.921463413440898e-06, "loss": 0.3467, "step": 75750 }, { "epoch": 75.78, "grad_norm": 69.16544342041016, "learning_rate": 4.91832222406698e-06, "loss": 0.3988, "step": 75780 }, { "epoch": 75.81, "grad_norm": 46.77901077270508, "learning_rate": 4.915181066938156e-06, "loss": 0.2973, "step": 75810 }, { "epoch": 75.84, "grad_norm": 35.41349411010742, "learning_rate": 4.912039943294502e-06, "loss": 0.3225, "step": 75840 }, { "epoch": 75.87, "grad_norm": 67.3143081665039, "learning_rate": 4.908898854376086e-06, "loss": 0.4036, "step": 75870 }, { "epoch": 75.9, "grad_norm": 39.68790054321289, "learning_rate": 4.90575780142296e-06, "loss": 0.3369, "step": 75900 }, { "epoch": 75.93, "grad_norm": 45.38121032714844, "learning_rate": 4.902616785675162e-06, "loss": 0.2764, "step": 75930 }, { "epoch": 75.96, "grad_norm": 59.727970123291016, "learning_rate": 4.899475808372714e-06, "loss": 0.3688, "step": 75960 }, { "epoch": 75.99, "grad_norm": 42.325172424316406, "learning_rate": 4.896334870755623e-06, "loss": 0.3785, "step": 75990 }, { "epoch": 76.0, "eval_loss": 0.286672443151474, "eval_map": 0.8005, "eval_map_50": 0.962, "eval_map_75": 0.9259, "eval_map_chicken": 0.789, "eval_map_duck": 0.7507, "eval_map_large": 0.8152, "eval_map_medium": 0.7912, "eval_map_plant": 0.8619, "eval_map_small": 0.2625, "eval_mar_1": 0.3243, "eval_mar_10": 0.8353, "eval_mar_100": 0.8393, "eval_mar_100_chicken": 0.8382, "eval_mar_100_duck": 0.7887, "eval_mar_100_plant": 0.8911, "eval_mar_large": 0.8547, "eval_mar_medium": 0.8332, "eval_mar_small": 0.3776, "eval_runtime": 13.0403, "eval_samples_per_second": 15.337, "eval_steps_per_second": 1.917, "step": 76000 }, { "epoch": 76.02, "grad_norm": 36.72150802612305, "learning_rate": 4.893193974063885e-06, "loss": 0.3384, "step": 76020 }, { "epoch": 76.05, "grad_norm": 69.30453491210938, "learning_rate": 4.890053119537475e-06, "loss": 0.3335, "step": 76050 }, { "epoch": 76.08, "grad_norm": 47.583763122558594, "learning_rate": 4.886912308416353e-06, "loss": 0.3777, "step": 76080 }, { "epoch": 76.11, "grad_norm": 63.60360336303711, "learning_rate": 4.8837715419404596e-06, "loss": 0.369, "step": 76110 }, { "epoch": 76.14, "grad_norm": 58.973724365234375, "learning_rate": 4.880630821349718e-06, "loss": 0.358, "step": 76140 }, { "epoch": 76.17, "grad_norm": 56.310760498046875, "learning_rate": 4.877490147884042e-06, "loss": 0.4052, "step": 76170 }, { "epoch": 76.2, "grad_norm": 51.5562744140625, "learning_rate": 4.874349522783313e-06, "loss": 0.3757, "step": 76200 }, { "epoch": 76.23, "grad_norm": 95.08251190185547, "learning_rate": 4.871208947287404e-06, "loss": 0.3849, "step": 76230 }, { "epoch": 76.26, "grad_norm": 37.81222152709961, "learning_rate": 4.8680684226361624e-06, "loss": 0.3451, "step": 76260 }, { "epoch": 76.29, "grad_norm": 43.137271881103516, "learning_rate": 4.864927950069417e-06, "loss": 0.3296, "step": 76290 }, { "epoch": 76.32, "grad_norm": 99.55328369140625, "learning_rate": 4.861787530826979e-06, "loss": 0.3995, "step": 76320 }, { "epoch": 76.35, "grad_norm": 38.03496170043945, "learning_rate": 4.8586471661486345e-06, "loss": 0.3465, "step": 76350 }, { "epoch": 76.38, "grad_norm": 59.183528900146484, "learning_rate": 4.85550685727415e-06, "loss": 0.3352, "step": 76380 }, { "epoch": 76.41, "grad_norm": 60.860984802246094, "learning_rate": 4.852366605443271e-06, "loss": 0.3531, "step": 76410 }, { "epoch": 76.44, "grad_norm": 61.353477478027344, "learning_rate": 4.8492264118957165e-06, "loss": 0.3592, "step": 76440 }, { "epoch": 76.47, "grad_norm": 24.048362731933594, "learning_rate": 4.846086277871188e-06, "loss": 0.3213, "step": 76470 }, { "epoch": 76.5, "grad_norm": 103.81499481201172, "learning_rate": 4.842946204609359e-06, "loss": 0.3596, "step": 76500 }, { "epoch": 76.53, "grad_norm": 77.99012756347656, "learning_rate": 4.839806193349882e-06, "loss": 0.3519, "step": 76530 }, { "epoch": 76.56, "grad_norm": 44.73519515991211, "learning_rate": 4.8366662453323826e-06, "loss": 0.3502, "step": 76560 }, { "epoch": 76.59, "grad_norm": 45.329368591308594, "learning_rate": 4.833526361796461e-06, "loss": 0.3998, "step": 76590 }, { "epoch": 76.62, "grad_norm": 103.1388931274414, "learning_rate": 4.830386543981696e-06, "loss": 0.4817, "step": 76620 }, { "epoch": 76.65, "grad_norm": 69.26411437988281, "learning_rate": 4.827246793127639e-06, "loss": 0.3279, "step": 76650 }, { "epoch": 76.68, "grad_norm": 66.59178924560547, "learning_rate": 4.8241071104738115e-06, "loss": 0.344, "step": 76680 }, { "epoch": 76.71, "grad_norm": 51.41876220703125, "learning_rate": 4.8209674972597116e-06, "loss": 0.4629, "step": 76710 }, { "epoch": 76.74, "grad_norm": 80.95065307617188, "learning_rate": 4.8178279547248055e-06, "loss": 0.3847, "step": 76740 }, { "epoch": 76.77, "grad_norm": 45.63530349731445, "learning_rate": 4.8146884841085405e-06, "loss": 0.3705, "step": 76770 }, { "epoch": 76.8, "grad_norm": 64.1922378540039, "learning_rate": 4.811549086650327e-06, "loss": 0.5386, "step": 76800 }, { "epoch": 76.83, "grad_norm": 47.3631591796875, "learning_rate": 4.8084097635895505e-06, "loss": 0.3069, "step": 76830 }, { "epoch": 76.86, "grad_norm": 80.86715698242188, "learning_rate": 4.805270516165564e-06, "loss": 0.4204, "step": 76860 }, { "epoch": 76.89, "grad_norm": 169.99560546875, "learning_rate": 4.802131345617694e-06, "loss": 0.3696, "step": 76890 }, { "epoch": 76.92, "grad_norm": 4.7745103074703366e-05, "learning_rate": 4.798992253185233e-06, "loss": 0.399, "step": 76920 }, { "epoch": 76.95, "grad_norm": 68.79605102539062, "learning_rate": 4.79585324010745e-06, "loss": 0.3323, "step": 76950 }, { "epoch": 76.98, "grad_norm": 27.288585662841797, "learning_rate": 4.792714307623574e-06, "loss": 0.3441, "step": 76980 }, { "epoch": 77.0, "eval_loss": 0.2761538624763489, "eval_map": 0.8081, "eval_map_50": 0.9584, "eval_map_75": 0.9263, "eval_map_chicken": 0.7948, "eval_map_duck": 0.7573, "eval_map_large": 0.8231, "eval_map_medium": 0.802, "eval_map_plant": 0.8721, "eval_map_small": 0.3086, "eval_mar_1": 0.3269, "eval_mar_10": 0.8434, "eval_mar_100": 0.8469, "eval_mar_100_chicken": 0.8435, "eval_mar_100_duck": 0.7964, "eval_mar_100_plant": 0.9009, "eval_mar_large": 0.8575, "eval_mar_medium": 0.8449, "eval_mar_small": 0.3839, "eval_runtime": 12.5899, "eval_samples_per_second": 15.886, "eval_steps_per_second": 1.986, "step": 77000 }, { "epoch": 77.01, "grad_norm": 56.2859992980957, "learning_rate": 4.789575456972807e-06, "loss": 0.3218, "step": 77010 }, { "epoch": 77.04, "grad_norm": 26.68988037109375, "learning_rate": 4.786436689394317e-06, "loss": 0.3065, "step": 77040 }, { "epoch": 77.07, "grad_norm": 52.37255096435547, "learning_rate": 4.783298006127238e-06, "loss": 0.3823, "step": 77070 }, { "epoch": 77.1, "grad_norm": 101.42407989501953, "learning_rate": 4.780159408410677e-06, "loss": 0.3017, "step": 77100 }, { "epoch": 77.13, "grad_norm": 73.09725189208984, "learning_rate": 4.777020897483697e-06, "loss": 0.3216, "step": 77130 }, { "epoch": 77.16, "grad_norm": 45.0621337890625, "learning_rate": 4.773882474585338e-06, "loss": 0.3998, "step": 77160 }, { "epoch": 77.19, "grad_norm": 44.780799865722656, "learning_rate": 4.770744140954596e-06, "loss": 0.3039, "step": 77190 }, { "epoch": 77.22, "grad_norm": 125.50296020507812, "learning_rate": 4.767605897830436e-06, "loss": 0.3239, "step": 77220 }, { "epoch": 77.25, "grad_norm": 57.67860412597656, "learning_rate": 4.7644677464517874e-06, "loss": 0.5612, "step": 77250 }, { "epoch": 77.28, "grad_norm": 47.23957443237305, "learning_rate": 4.761329688057543e-06, "loss": 0.3223, "step": 77280 }, { "epoch": 77.31, "grad_norm": 43.83549499511719, "learning_rate": 4.7581917238865565e-06, "loss": 0.3358, "step": 77310 }, { "epoch": 77.34, "grad_norm": 5.528132896870375e-05, "learning_rate": 4.7550538551776495e-06, "loss": 0.2885, "step": 77340 }, { "epoch": 77.37, "grad_norm": 36.684425354003906, "learning_rate": 4.7519160831696e-06, "loss": 0.3639, "step": 77370 }, { "epoch": 77.4, "grad_norm": 36.439453125, "learning_rate": 4.748778409101153e-06, "loss": 0.3286, "step": 77400 }, { "epoch": 77.43, "grad_norm": 0.00010466995445312932, "learning_rate": 4.745640834211012e-06, "loss": 0.3636, "step": 77430 }, { "epoch": 77.46, "grad_norm": 45.679927825927734, "learning_rate": 4.742503359737841e-06, "loss": 0.3527, "step": 77460 }, { "epoch": 77.49, "grad_norm": 56.845550537109375, "learning_rate": 4.739365986920265e-06, "loss": 0.4575, "step": 77490 }, { "epoch": 77.52, "grad_norm": 54.34233093261719, "learning_rate": 4.736228716996868e-06, "loss": 0.3215, "step": 77520 }, { "epoch": 77.55, "grad_norm": 232.83746337890625, "learning_rate": 4.733091551206198e-06, "loss": 0.3501, "step": 77550 }, { "epoch": 77.58, "grad_norm": 45.703609466552734, "learning_rate": 4.7299544907867576e-06, "loss": 0.326, "step": 77580 }, { "epoch": 77.61, "grad_norm": 89.6006088256836, "learning_rate": 4.726817536977006e-06, "loss": 0.334, "step": 77610 }, { "epoch": 77.64, "grad_norm": 88.63821411132812, "learning_rate": 4.723680691015366e-06, "loss": 0.5104, "step": 77640 }, { "epoch": 77.67, "grad_norm": 71.7905502319336, "learning_rate": 4.72054395414021e-06, "loss": 0.4446, "step": 77670 }, { "epoch": 77.7, "grad_norm": 53.75556945800781, "learning_rate": 4.717407327589878e-06, "loss": 0.3117, "step": 77700 }, { "epoch": 77.73, "grad_norm": 78.02815246582031, "learning_rate": 4.714270812602657e-06, "loss": 0.4198, "step": 77730 }, { "epoch": 77.76, "grad_norm": 161.20706176757812, "learning_rate": 4.711134410416794e-06, "loss": 0.3559, "step": 77760 }, { "epoch": 77.79, "grad_norm": 67.65469360351562, "learning_rate": 4.707998122270493e-06, "loss": 0.3805, "step": 77790 }, { "epoch": 77.82, "grad_norm": 43.17831039428711, "learning_rate": 4.704861949401904e-06, "loss": 0.3327, "step": 77820 }, { "epoch": 77.85, "grad_norm": 68.78608703613281, "learning_rate": 4.7017258930491474e-06, "loss": 0.339, "step": 77850 }, { "epoch": 77.88, "grad_norm": 58.352054595947266, "learning_rate": 4.6985899544502835e-06, "loss": 0.3207, "step": 77880 }, { "epoch": 77.91, "grad_norm": 42.44630813598633, "learning_rate": 4.695454134843332e-06, "loss": 0.4118, "step": 77910 }, { "epoch": 77.94, "grad_norm": 104.02799224853516, "learning_rate": 4.692318435466265e-06, "loss": 0.3496, "step": 77940 }, { "epoch": 77.97, "grad_norm": 54.272804260253906, "learning_rate": 4.689182857557006e-06, "loss": 0.3509, "step": 77970 }, { "epoch": 78.0, "grad_norm": 89.3212661743164, "learning_rate": 4.686047402353433e-06, "loss": 0.344, "step": 78000 }, { "epoch": 78.0, "eval_loss": 0.2862946689128876, "eval_map": 0.8003, "eval_map_50": 0.9617, "eval_map_75": 0.9286, "eval_map_chicken": 0.785, "eval_map_duck": 0.751, "eval_map_large": 0.8183, "eval_map_medium": 0.793, "eval_map_plant": 0.8648, "eval_map_small": 0.2783, "eval_mar_1": 0.3227, "eval_mar_10": 0.8338, "eval_mar_100": 0.8376, "eval_mar_100_chicken": 0.8336, "eval_mar_100_duck": 0.7866, "eval_mar_100_plant": 0.8926, "eval_mar_large": 0.8544, "eval_mar_medium": 0.8327, "eval_mar_small": 0.3869, "eval_runtime": 12.5561, "eval_samples_per_second": 15.929, "eval_steps_per_second": 1.991, "step": 78000 }, { "epoch": 78.03, "grad_norm": 41.207908630371094, "learning_rate": 4.682912071093374e-06, "loss": 0.3365, "step": 78030 }, { "epoch": 78.06, "grad_norm": 44.792633056640625, "learning_rate": 4.679776865014609e-06, "loss": 0.3202, "step": 78060 }, { "epoch": 78.09, "grad_norm": 66.00247192382812, "learning_rate": 4.676641785354865e-06, "loss": 0.3341, "step": 78090 }, { "epoch": 78.12, "grad_norm": 35.228843688964844, "learning_rate": 4.673506833351821e-06, "loss": 0.3818, "step": 78120 }, { "epoch": 78.15, "grad_norm": 0.0001028688348014839, "learning_rate": 4.670372010243111e-06, "loss": 0.3335, "step": 78150 }, { "epoch": 78.18, "grad_norm": 62.809871673583984, "learning_rate": 4.667237317266311e-06, "loss": 0.3612, "step": 78180 }, { "epoch": 78.21, "grad_norm": 65.34758758544922, "learning_rate": 4.664102755658948e-06, "loss": 0.3327, "step": 78210 }, { "epoch": 78.24, "grad_norm": 45.041709899902344, "learning_rate": 4.660968326658497e-06, "loss": 0.427, "step": 78240 }, { "epoch": 78.27, "grad_norm": 199.43502807617188, "learning_rate": 4.65783403150238e-06, "loss": 0.3688, "step": 78270 }, { "epoch": 78.3, "grad_norm": 221.42352294921875, "learning_rate": 4.654699871427972e-06, "loss": 0.3036, "step": 78300 }, { "epoch": 78.33, "grad_norm": 87.0208511352539, "learning_rate": 4.651565847672584e-06, "loss": 0.3372, "step": 78330 }, { "epoch": 78.36, "grad_norm": 28.67781639099121, "learning_rate": 4.648431961473482e-06, "loss": 0.3293, "step": 78360 }, { "epoch": 78.39, "grad_norm": 77.96605682373047, "learning_rate": 4.6452982140678735e-06, "loss": 0.2971, "step": 78390 }, { "epoch": 78.42, "grad_norm": 67.1175765991211, "learning_rate": 4.642164606692912e-06, "loss": 0.354, "step": 78420 }, { "epoch": 78.45, "grad_norm": 7.320206350414082e-05, "learning_rate": 4.639031140585697e-06, "loss": 0.281, "step": 78450 }, { "epoch": 78.48, "grad_norm": 73.01411437988281, "learning_rate": 4.635897816983272e-06, "loss": 0.3224, "step": 78480 }, { "epoch": 78.51, "grad_norm": 152.59507751464844, "learning_rate": 4.632764637122622e-06, "loss": 0.3246, "step": 78510 }, { "epoch": 78.54, "grad_norm": 82.95006561279297, "learning_rate": 4.629631602240678e-06, "loss": 0.3625, "step": 78540 }, { "epoch": 78.57, "grad_norm": 73.2284927368164, "learning_rate": 4.626498713574311e-06, "loss": 0.3468, "step": 78570 }, { "epoch": 78.6, "grad_norm": 75.01376342773438, "learning_rate": 4.6233659723603374e-06, "loss": 0.4001, "step": 78600 }, { "epoch": 78.63, "grad_norm": 66.83414459228516, "learning_rate": 4.620233379835513e-06, "loss": 0.4147, "step": 78630 }, { "epoch": 78.66, "grad_norm": 47.5323600769043, "learning_rate": 4.617100937236535e-06, "loss": 0.3971, "step": 78660 }, { "epoch": 78.69, "grad_norm": 45.6131477355957, "learning_rate": 4.6139686458000445e-06, "loss": 0.305, "step": 78690 }, { "epoch": 78.72, "grad_norm": 79.29166412353516, "learning_rate": 4.610836506762618e-06, "loss": 0.3049, "step": 78720 }, { "epoch": 78.75, "grad_norm": 77.18326568603516, "learning_rate": 4.6077045213607765e-06, "loss": 0.3979, "step": 78750 }, { "epoch": 78.78, "grad_norm": 89.47647094726562, "learning_rate": 4.604572690830976e-06, "loss": 0.3349, "step": 78780 }, { "epoch": 78.81, "grad_norm": 162.22079467773438, "learning_rate": 4.601441016409616e-06, "loss": 0.2974, "step": 78810 }, { "epoch": 78.84, "grad_norm": 72.96746826171875, "learning_rate": 4.59830949933303e-06, "loss": 0.3756, "step": 78840 }, { "epoch": 78.87, "grad_norm": 49.24949264526367, "learning_rate": 4.5951781408374915e-06, "loss": 0.2905, "step": 78870 }, { "epoch": 78.9, "grad_norm": 81.47919464111328, "learning_rate": 4.592046942159213e-06, "loss": 0.3644, "step": 78900 }, { "epoch": 78.93, "grad_norm": 70.19947814941406, "learning_rate": 4.588915904534341e-06, "loss": 0.4723, "step": 78930 }, { "epoch": 78.96, "grad_norm": 41.834815979003906, "learning_rate": 4.5857850291989596e-06, "loss": 0.3737, "step": 78960 }, { "epoch": 78.99, "grad_norm": 52.08351516723633, "learning_rate": 4.582654317389088e-06, "loss": 0.339, "step": 78990 }, { "epoch": 79.0, "eval_loss": 0.2687438130378723, "eval_map": 0.8186, "eval_map_50": 0.9613, "eval_map_75": 0.9276, "eval_map_chicken": 0.8079, "eval_map_duck": 0.7733, "eval_map_large": 0.8268, "eval_map_medium": 0.8113, "eval_map_plant": 0.8747, "eval_map_small": 0.2672, "eval_mar_1": 0.3303, "eval_mar_10": 0.8499, "eval_mar_100": 0.8542, "eval_mar_100_chicken": 0.8519, "eval_mar_100_duck": 0.8098, "eval_mar_100_plant": 0.901, "eval_mar_large": 0.8578, "eval_mar_medium": 0.85, "eval_mar_small": 0.3919, "eval_runtime": 12.7992, "eval_samples_per_second": 15.626, "eval_steps_per_second": 1.953, "step": 79000 }, { "epoch": 79.02, "grad_norm": 65.77619934082031, "learning_rate": 4.579523770340681e-06, "loss": 0.3391, "step": 79020 }, { "epoch": 79.05, "grad_norm": 44.03147506713867, "learning_rate": 4.576393389289633e-06, "loss": 0.4114, "step": 79050 }, { "epoch": 79.08, "grad_norm": 33.46986389160156, "learning_rate": 4.573263175471766e-06, "loss": 0.3052, "step": 79080 }, { "epoch": 79.11, "grad_norm": 116.59758758544922, "learning_rate": 4.5701331301228395e-06, "loss": 0.3896, "step": 79110 }, { "epoch": 79.14, "grad_norm": 26.716373443603516, "learning_rate": 4.567003254478545e-06, "loss": 0.4215, "step": 79140 }, { "epoch": 79.17, "grad_norm": 56.31059265136719, "learning_rate": 4.563873549774506e-06, "loss": 0.3464, "step": 79170 }, { "epoch": 79.2, "grad_norm": 55.68206787109375, "learning_rate": 4.560744017246284e-06, "loss": 0.3936, "step": 79200 }, { "epoch": 79.23, "grad_norm": 39.73326110839844, "learning_rate": 4.557614658129369e-06, "loss": 0.3069, "step": 79230 }, { "epoch": 79.26, "grad_norm": 59.32463836669922, "learning_rate": 4.55448547365918e-06, "loss": 0.4183, "step": 79260 }, { "epoch": 79.29, "grad_norm": 54.604915618896484, "learning_rate": 4.55135646507107e-06, "loss": 0.369, "step": 79290 }, { "epoch": 79.32, "grad_norm": 30.115066528320312, "learning_rate": 4.548227633600322e-06, "loss": 0.3704, "step": 79320 }, { "epoch": 79.35, "grad_norm": 50.072021484375, "learning_rate": 4.545098980482151e-06, "loss": 0.3368, "step": 79350 }, { "epoch": 79.38, "grad_norm": 31.9080753326416, "learning_rate": 4.541970506951698e-06, "loss": 0.3165, "step": 79380 }, { "epoch": 79.41, "grad_norm": 133.45260620117188, "learning_rate": 4.538842214244035e-06, "loss": 0.3379, "step": 79410 }, { "epoch": 79.44, "grad_norm": 98.52044677734375, "learning_rate": 4.535714103594162e-06, "loss": 0.3274, "step": 79440 }, { "epoch": 79.47, "grad_norm": 0.0004271268844604492, "learning_rate": 4.532586176237007e-06, "loss": 0.5187, "step": 79470 }, { "epoch": 79.5, "grad_norm": 79.19873809814453, "learning_rate": 4.529458433407429e-06, "loss": 0.348, "step": 79500 }, { "epoch": 79.53, "grad_norm": 265.66766357421875, "learning_rate": 4.526330876340209e-06, "loss": 0.4504, "step": 79530 }, { "epoch": 79.56, "grad_norm": 31.870468139648438, "learning_rate": 4.523203506270058e-06, "loss": 0.3241, "step": 79560 }, { "epoch": 79.59, "grad_norm": 65.54637145996094, "learning_rate": 4.520076324431612e-06, "loss": 0.4266, "step": 79590 }, { "epoch": 79.62, "grad_norm": 40.75666809082031, "learning_rate": 4.516949332059429e-06, "loss": 0.3729, "step": 79620 }, { "epoch": 79.65, "grad_norm": 72.28760528564453, "learning_rate": 4.513822530388004e-06, "loss": 0.3513, "step": 79650 }, { "epoch": 79.68, "grad_norm": 9.84502476057969e-05, "learning_rate": 4.510695920651742e-06, "loss": 0.3271, "step": 79680 }, { "epoch": 79.71, "grad_norm": 115.8056869506836, "learning_rate": 4.507569504084983e-06, "loss": 0.386, "step": 79710 }, { "epoch": 79.74, "grad_norm": 98.94456481933594, "learning_rate": 4.504443281921985e-06, "loss": 0.4488, "step": 79740 }, { "epoch": 79.77, "grad_norm": 63.09760665893555, "learning_rate": 4.501317255396931e-06, "loss": 0.377, "step": 79770 }, { "epoch": 79.8, "grad_norm": 27.420446395874023, "learning_rate": 4.4981914257439254e-06, "loss": 0.352, "step": 79800 }, { "epoch": 79.83, "grad_norm": 38.316017150878906, "learning_rate": 4.495065794196999e-06, "loss": 0.3761, "step": 79830 }, { "epoch": 79.86, "grad_norm": 67.256591796875, "learning_rate": 4.491940361990101e-06, "loss": 0.3733, "step": 79860 }, { "epoch": 79.89, "grad_norm": 60.85757827758789, "learning_rate": 4.488815130357103e-06, "loss": 0.3661, "step": 79890 }, { "epoch": 79.92, "grad_norm": 131.9546356201172, "learning_rate": 4.485690100531793e-06, "loss": 0.4037, "step": 79920 }, { "epoch": 79.95, "grad_norm": 52.427223205566406, "learning_rate": 4.482565273747888e-06, "loss": 0.4483, "step": 79950 }, { "epoch": 79.98, "grad_norm": 78.66764068603516, "learning_rate": 4.4794406512390175e-06, "loss": 0.2642, "step": 79980 }, { "epoch": 80.0, "eval_loss": 0.2586832344532013, "eval_map": 0.8227, "eval_map_50": 0.9658, "eval_map_75": 0.9254, "eval_map_chicken": 0.8123, "eval_map_duck": 0.78, "eval_map_large": 0.8269, "eval_map_medium": 0.8182, "eval_map_plant": 0.8757, "eval_map_small": 0.2459, "eval_mar_1": 0.3321, "eval_mar_10": 0.8543, "eval_mar_100": 0.8583, "eval_mar_100_chicken": 0.8551, "eval_mar_100_duck": 0.8165, "eval_mar_100_plant": 0.9033, "eval_mar_large": 0.858, "eval_mar_medium": 0.8576, "eval_mar_small": 0.39, "eval_runtime": 12.774, "eval_samples_per_second": 15.657, "eval_steps_per_second": 1.957, "step": 80000 }, { "epoch": 80.01, "grad_norm": 63.85955810546875, "learning_rate": 4.476316234238735e-06, "loss": 0.3357, "step": 80010 }, { "epoch": 80.04, "grad_norm": 22.43913459777832, "learning_rate": 4.473192023980509e-06, "loss": 0.3146, "step": 80040 }, { "epoch": 80.07, "grad_norm": 70.78424835205078, "learning_rate": 4.470068021697728e-06, "loss": 0.2919, "step": 80070 }, { "epoch": 80.1, "grad_norm": 101.76507568359375, "learning_rate": 4.466944228623701e-06, "loss": 0.3306, "step": 80100 }, { "epoch": 80.13, "grad_norm": 51.99180603027344, "learning_rate": 4.463820645991651e-06, "loss": 0.2971, "step": 80130 }, { "epoch": 80.16, "grad_norm": 110.32954406738281, "learning_rate": 4.460697275034717e-06, "loss": 0.4023, "step": 80160 }, { "epoch": 80.19, "grad_norm": 103.43629455566406, "learning_rate": 4.457574116985958e-06, "loss": 0.3389, "step": 80190 }, { "epoch": 80.22, "grad_norm": 75.853271484375, "learning_rate": 4.454451173078347e-06, "loss": 0.3809, "step": 80220 }, { "epoch": 80.25, "grad_norm": 0.00023755144502501935, "learning_rate": 4.451328444544774e-06, "loss": 0.2541, "step": 80250 }, { "epoch": 80.28, "grad_norm": 55.720802307128906, "learning_rate": 4.448205932618042e-06, "loss": 0.3117, "step": 80280 }, { "epoch": 80.31, "grad_norm": 30.30018424987793, "learning_rate": 4.44508363853087e-06, "loss": 0.3481, "step": 80310 }, { "epoch": 80.34, "grad_norm": 4.0790320781525224e-05, "learning_rate": 4.4419615635158875e-06, "loss": 0.3422, "step": 80340 }, { "epoch": 80.37, "grad_norm": 115.61272430419922, "learning_rate": 4.43883970880564e-06, "loss": 0.3754, "step": 80370 }, { "epoch": 80.4, "grad_norm": 73.1986312866211, "learning_rate": 4.4357180756325915e-06, "loss": 0.3675, "step": 80400 }, { "epoch": 80.43, "grad_norm": 59.47132873535156, "learning_rate": 4.43259666522911e-06, "loss": 0.358, "step": 80430 }, { "epoch": 80.46, "grad_norm": 53.91272735595703, "learning_rate": 4.42947547882748e-06, "loss": 0.2943, "step": 80460 }, { "epoch": 80.49, "grad_norm": 43.175437927246094, "learning_rate": 4.426354517659894e-06, "loss": 0.4473, "step": 80490 }, { "epoch": 80.52, "grad_norm": 66.0586166381836, "learning_rate": 4.423233782958459e-06, "loss": 0.4308, "step": 80520 }, { "epoch": 80.55, "grad_norm": 77.06782531738281, "learning_rate": 4.420113275955193e-06, "loss": 0.4187, "step": 80550 }, { "epoch": 80.58, "grad_norm": 63.33230209350586, "learning_rate": 4.416992997882023e-06, "loss": 0.3624, "step": 80580 }, { "epoch": 80.61, "grad_norm": 42.74773025512695, "learning_rate": 4.413872949970785e-06, "loss": 0.327, "step": 80610 }, { "epoch": 80.64, "grad_norm": 97.04639434814453, "learning_rate": 4.410753133453222e-06, "loss": 0.4088, "step": 80640 }, { "epoch": 80.67, "grad_norm": 164.1802978515625, "learning_rate": 4.407633549560991e-06, "loss": 0.3173, "step": 80670 }, { "epoch": 80.7, "grad_norm": 156.2711181640625, "learning_rate": 4.404514199525651e-06, "loss": 0.3476, "step": 80700 }, { "epoch": 80.73, "grad_norm": 52.91155242919922, "learning_rate": 4.401395084578677e-06, "loss": 0.3384, "step": 80730 }, { "epoch": 80.76, "grad_norm": 104.0030517578125, "learning_rate": 4.398276205951443e-06, "loss": 0.2822, "step": 80760 }, { "epoch": 80.79, "grad_norm": 58.91257858276367, "learning_rate": 4.3951575648752346e-06, "loss": 0.3457, "step": 80790 }, { "epoch": 80.82, "grad_norm": 79.34716033935547, "learning_rate": 4.392039162581239e-06, "loss": 0.4457, "step": 80820 }, { "epoch": 80.85, "grad_norm": 61.21530532836914, "learning_rate": 4.388921000300553e-06, "loss": 0.3555, "step": 80850 }, { "epoch": 80.88, "grad_norm": 43.29530334472656, "learning_rate": 4.38580307926418e-06, "loss": 0.3628, "step": 80880 }, { "epoch": 80.91, "grad_norm": 185.1835479736328, "learning_rate": 4.382685400703024e-06, "loss": 0.4157, "step": 80910 }, { "epoch": 80.94, "grad_norm": 81.77842712402344, "learning_rate": 4.379567965847896e-06, "loss": 0.3101, "step": 80940 }, { "epoch": 80.97, "grad_norm": 74.31657409667969, "learning_rate": 4.37645077592951e-06, "loss": 0.3573, "step": 80970 }, { "epoch": 81.0, "grad_norm": 58.33173370361328, "learning_rate": 4.373333832178478e-06, "loss": 0.3122, "step": 81000 }, { "epoch": 81.0, "eval_loss": 0.2692364454269409, "eval_map": 0.8148, "eval_map_50": 0.9609, "eval_map_75": 0.9193, "eval_map_chicken": 0.8096, "eval_map_duck": 0.7596, "eval_map_large": 0.8143, "eval_map_medium": 0.8076, "eval_map_plant": 0.8751, "eval_map_small": 0.2916, "eval_mar_1": 0.3303, "eval_mar_10": 0.8479, "eval_mar_100": 0.8517, "eval_mar_100_chicken": 0.8557, "eval_mar_100_duck": 0.7974, "eval_mar_100_plant": 0.902, "eval_mar_large": 0.851, "eval_mar_medium": 0.8495, "eval_mar_small": 0.3864, "eval_runtime": 12.3822, "eval_samples_per_second": 16.152, "eval_steps_per_second": 2.019, "step": 81000 }, { "epoch": 81.03, "grad_norm": 126.294921875, "learning_rate": 4.370217135825329e-06, "loss": 0.3287, "step": 81030 }, { "epoch": 81.06, "grad_norm": 68.54559326171875, "learning_rate": 4.36710068810048e-06, "loss": 0.3301, "step": 81060 }, { "epoch": 81.09, "grad_norm": 97.65863037109375, "learning_rate": 4.363984490234257e-06, "loss": 0.2969, "step": 81090 }, { "epoch": 81.12, "grad_norm": 42.647953033447266, "learning_rate": 4.360868543456883e-06, "loss": 0.3167, "step": 81120 }, { "epoch": 81.15, "grad_norm": 68.58827209472656, "learning_rate": 4.357752848998486e-06, "loss": 0.2767, "step": 81150 }, { "epoch": 81.18, "grad_norm": 46.38288879394531, "learning_rate": 4.354637408089093e-06, "loss": 0.3167, "step": 81180 }, { "epoch": 81.21, "grad_norm": 0.0006489719962701201, "learning_rate": 4.351522221958633e-06, "loss": 0.3562, "step": 81210 }, { "epoch": 81.24, "grad_norm": 50.47576141357422, "learning_rate": 4.348407291836928e-06, "loss": 0.3788, "step": 81240 }, { "epoch": 81.27, "grad_norm": 84.53024291992188, "learning_rate": 4.3452926189537056e-06, "loss": 0.4037, "step": 81270 }, { "epoch": 81.3, "grad_norm": 40.026676177978516, "learning_rate": 4.342178204538588e-06, "loss": 0.3905, "step": 81300 }, { "epoch": 81.33, "grad_norm": 73.51455688476562, "learning_rate": 4.339064049821098e-06, "loss": 0.2883, "step": 81330 }, { "epoch": 81.36, "grad_norm": 55.855445861816406, "learning_rate": 4.335950156030653e-06, "loss": 0.3561, "step": 81360 }, { "epoch": 81.39, "grad_norm": 41.767547607421875, "learning_rate": 4.332836524396571e-06, "loss": 0.4711, "step": 81390 }, { "epoch": 81.42, "grad_norm": 42.107303619384766, "learning_rate": 4.329723156148064e-06, "loss": 0.4199, "step": 81420 }, { "epoch": 81.45, "grad_norm": 67.1325454711914, "learning_rate": 4.326610052514238e-06, "loss": 0.3289, "step": 81450 }, { "epoch": 81.48, "grad_norm": 121.80181121826172, "learning_rate": 4.323497214724099e-06, "loss": 0.3953, "step": 81480 }, { "epoch": 81.51, "grad_norm": 96.99967956542969, "learning_rate": 4.320384644006546e-06, "loss": 0.3347, "step": 81510 }, { "epoch": 81.54, "grad_norm": 41.085933685302734, "learning_rate": 4.317272341590373e-06, "loss": 0.4183, "step": 81540 }, { "epoch": 81.57, "grad_norm": 35.67473220825195, "learning_rate": 4.314160308704269e-06, "loss": 0.3876, "step": 81570 }, { "epoch": 81.6, "grad_norm": 61.16136932373047, "learning_rate": 4.31104854657681e-06, "loss": 0.3012, "step": 81600 }, { "epoch": 81.63, "grad_norm": 128.5059051513672, "learning_rate": 4.307937056436476e-06, "loss": 0.3388, "step": 81630 }, { "epoch": 81.66, "grad_norm": 133.8731689453125, "learning_rate": 4.3048258395116326e-06, "loss": 0.3692, "step": 81660 }, { "epoch": 81.69, "grad_norm": 0.00010825125355040655, "learning_rate": 4.301714897030537e-06, "loss": 0.2702, "step": 81690 }, { "epoch": 81.72, "grad_norm": 50.41769027709961, "learning_rate": 4.298604230221341e-06, "loss": 0.3527, "step": 81720 }, { "epoch": 81.75, "grad_norm": 79.51022338867188, "learning_rate": 4.295493840312087e-06, "loss": 0.3529, "step": 81750 }, { "epoch": 81.78, "grad_norm": 35.30931854248047, "learning_rate": 4.2923837285307085e-06, "loss": 0.345, "step": 81780 }, { "epoch": 81.81, "grad_norm": 220.96678161621094, "learning_rate": 4.289273896105027e-06, "loss": 0.3729, "step": 81810 }, { "epoch": 81.84, "grad_norm": 63.90647888183594, "learning_rate": 4.286164344262756e-06, "loss": 0.4443, "step": 81840 }, { "epoch": 81.87, "grad_norm": 47.41940689086914, "learning_rate": 4.283055074231498e-06, "loss": 0.3509, "step": 81870 }, { "epoch": 81.9, "grad_norm": 34.70399856567383, "learning_rate": 4.279946087238739e-06, "loss": 0.3092, "step": 81900 }, { "epoch": 81.93, "grad_norm": 70.23284149169922, "learning_rate": 4.276837384511864e-06, "loss": 0.3799, "step": 81930 }, { "epoch": 81.96, "grad_norm": 40.635921478271484, "learning_rate": 4.273728967278137e-06, "loss": 0.3229, "step": 81960 }, { "epoch": 81.99, "grad_norm": 32.52659225463867, "learning_rate": 4.2706208367647115e-06, "loss": 0.3475, "step": 81990 }, { "epoch": 82.0, "eval_loss": 0.2804587185382843, "eval_map": 0.797, "eval_map_50": 0.9577, "eval_map_75": 0.9257, "eval_map_chicken": 0.7938, "eval_map_duck": 0.7365, "eval_map_large": 0.816, "eval_map_medium": 0.789, "eval_map_plant": 0.8607, "eval_map_small": 0.2179, "eval_mar_1": 0.3231, "eval_mar_10": 0.8329, "eval_mar_100": 0.8366, "eval_mar_100_chicken": 0.8421, "eval_mar_100_duck": 0.7789, "eval_mar_100_plant": 0.8888, "eval_mar_large": 0.8536, "eval_mar_medium": 0.831, "eval_mar_small": 0.3324, "eval_runtime": 13.222, "eval_samples_per_second": 15.126, "eval_steps_per_second": 1.891, "step": 82000 }, { "epoch": 82.02, "grad_norm": 43.95146942138672, "learning_rate": 4.267512994198629e-06, "loss": 0.2703, "step": 82020 }, { "epoch": 82.05, "grad_norm": 36.164974212646484, "learning_rate": 4.264405440806813e-06, "loss": 0.3427, "step": 82050 }, { "epoch": 82.08, "grad_norm": 40.10499572753906, "learning_rate": 4.261298177816082e-06, "loss": 0.3489, "step": 82080 }, { "epoch": 82.11, "grad_norm": 69.1693344116211, "learning_rate": 4.258191206453132e-06, "loss": 0.343, "step": 82110 }, { "epoch": 82.14, "grad_norm": 127.9786376953125, "learning_rate": 4.2550845279445455e-06, "loss": 0.3467, "step": 82140 }, { "epoch": 82.17, "grad_norm": 48.19189453125, "learning_rate": 4.251978143516789e-06, "loss": 0.3154, "step": 82170 }, { "epoch": 82.2, "grad_norm": 34.93656921386719, "learning_rate": 4.248872054396215e-06, "loss": 0.4107, "step": 82200 }, { "epoch": 82.23, "grad_norm": 40.05592346191406, "learning_rate": 4.245766261809059e-06, "loss": 0.4614, "step": 82230 }, { "epoch": 82.26, "grad_norm": 112.90399169921875, "learning_rate": 4.242660766981439e-06, "loss": 0.3491, "step": 82260 }, { "epoch": 82.29, "grad_norm": 25.270891189575195, "learning_rate": 4.239555571139353e-06, "loss": 0.3104, "step": 82290 }, { "epoch": 82.32, "grad_norm": 46.698097229003906, "learning_rate": 4.2364506755086856e-06, "loss": 0.2918, "step": 82320 }, { "epoch": 82.35, "grad_norm": 41.55739212036133, "learning_rate": 4.233346081315197e-06, "loss": 0.3246, "step": 82350 }, { "epoch": 82.38, "grad_norm": 62.291481018066406, "learning_rate": 4.230241789784535e-06, "loss": 0.3612, "step": 82380 }, { "epoch": 82.41, "grad_norm": 55.15680694580078, "learning_rate": 4.227137802142225e-06, "loss": 0.3091, "step": 82410 }, { "epoch": 82.44, "grad_norm": 37.55520248413086, "learning_rate": 4.224034119613671e-06, "loss": 0.2671, "step": 82440 }, { "epoch": 82.47, "grad_norm": 129.82908630371094, "learning_rate": 4.220930743424157e-06, "loss": 0.357, "step": 82470 }, { "epoch": 82.5, "grad_norm": 78.8675537109375, "learning_rate": 4.217827674798845e-06, "loss": 0.3776, "step": 82500 }, { "epoch": 82.53, "grad_norm": 142.05535888671875, "learning_rate": 4.2147249149627826e-06, "loss": 0.4349, "step": 82530 }, { "epoch": 82.56, "grad_norm": 52.57366180419922, "learning_rate": 4.211622465140887e-06, "loss": 0.3843, "step": 82560 }, { "epoch": 82.59, "grad_norm": 37.13884353637695, "learning_rate": 4.208520326557957e-06, "loss": 0.3646, "step": 82590 }, { "epoch": 82.62, "grad_norm": 86.63589477539062, "learning_rate": 4.2054185004386675e-06, "loss": 0.3212, "step": 82620 }, { "epoch": 82.65, "grad_norm": 48.998992919921875, "learning_rate": 4.202316988007568e-06, "loss": 0.3318, "step": 82650 }, { "epoch": 82.68, "grad_norm": 35.714149475097656, "learning_rate": 4.199215790489091e-06, "loss": 0.2511, "step": 82680 }, { "epoch": 82.71, "grad_norm": 55.05647659301758, "learning_rate": 4.196114909107538e-06, "loss": 0.3567, "step": 82710 }, { "epoch": 82.74, "grad_norm": 32.65850830078125, "learning_rate": 4.193014345087088e-06, "loss": 0.3211, "step": 82740 }, { "epoch": 82.77, "grad_norm": 57.23600769042969, "learning_rate": 4.1899140996517934e-06, "loss": 0.3144, "step": 82770 }, { "epoch": 82.8, "grad_norm": 123.17206573486328, "learning_rate": 4.186814174025582e-06, "loss": 0.3126, "step": 82800 }, { "epoch": 82.83, "grad_norm": 80.32073974609375, "learning_rate": 4.183714569432259e-06, "loss": 0.3556, "step": 82830 }, { "epoch": 82.86, "grad_norm": 69.4549789428711, "learning_rate": 4.180615287095494e-06, "loss": 0.2601, "step": 82860 }, { "epoch": 82.89, "grad_norm": 4.959073703503236e-05, "learning_rate": 4.177516328238838e-06, "loss": 0.302, "step": 82890 }, { "epoch": 82.92, "grad_norm": 67.3753662109375, "learning_rate": 4.174417694085711e-06, "loss": 0.3682, "step": 82920 }, { "epoch": 82.95, "grad_norm": 75.9146957397461, "learning_rate": 4.171319385859402e-06, "loss": 0.3106, "step": 82950 }, { "epoch": 82.98, "grad_norm": 47.08323287963867, "learning_rate": 4.168221404783076e-06, "loss": 0.4223, "step": 82980 }, { "epoch": 83.0, "eval_loss": 0.2651881277561188, "eval_map": 0.8082, "eval_map_50": 0.9656, "eval_map_75": 0.9314, "eval_map_chicken": 0.797, "eval_map_duck": 0.7561, "eval_map_large": 0.8239, "eval_map_medium": 0.8025, "eval_map_plant": 0.8714, "eval_map_small": 0.2965, "eval_mar_1": 0.3252, "eval_mar_10": 0.8455, "eval_mar_100": 0.8485, "eval_mar_100_chicken": 0.8447, "eval_mar_100_duck": 0.8005, "eval_mar_100_plant": 0.9003, "eval_mar_large": 0.8643, "eval_mar_medium": 0.8477, "eval_mar_small": 0.3701, "eval_runtime": 12.8289, "eval_samples_per_second": 15.59, "eval_steps_per_second": 1.949, "step": 83000 }, { "epoch": 83.01, "grad_norm": 29.939552307128906, "learning_rate": 4.165123752079768e-06, "loss": 0.3257, "step": 83010 }, { "epoch": 83.04, "grad_norm": 67.82865905761719, "learning_rate": 4.16202642897238e-06, "loss": 0.298, "step": 83040 }, { "epoch": 83.07, "grad_norm": 74.05908203125, "learning_rate": 4.158929436683687e-06, "loss": 0.3322, "step": 83070 }, { "epoch": 83.1, "grad_norm": 27.848155975341797, "learning_rate": 4.155832776436331e-06, "loss": 0.4286, "step": 83100 }, { "epoch": 83.13, "grad_norm": 205.10157775878906, "learning_rate": 4.152736449452827e-06, "loss": 0.3564, "step": 83130 }, { "epoch": 83.16, "grad_norm": 87.05416870117188, "learning_rate": 4.149640456955555e-06, "loss": 0.3552, "step": 83160 }, { "epoch": 83.19, "grad_norm": 515.0001220703125, "learning_rate": 4.146544800166764e-06, "loss": 0.3796, "step": 83190 }, { "epoch": 83.22, "grad_norm": 29.725467681884766, "learning_rate": 4.143449480308569e-06, "loss": 0.3208, "step": 83220 }, { "epoch": 83.25, "grad_norm": 58.52930450439453, "learning_rate": 4.140354498602952e-06, "loss": 0.3857, "step": 83250 }, { "epoch": 83.28, "grad_norm": 105.94011688232422, "learning_rate": 4.137259856271767e-06, "loss": 0.4048, "step": 83280 }, { "epoch": 83.31, "grad_norm": 91.10416412353516, "learning_rate": 4.134165554536728e-06, "loss": 0.322, "step": 83310 }, { "epoch": 83.34, "grad_norm": 67.30046081542969, "learning_rate": 4.131071594619416e-06, "loss": 0.3048, "step": 83340 }, { "epoch": 83.37, "grad_norm": 83.18740844726562, "learning_rate": 4.127977977741277e-06, "loss": 0.3481, "step": 83370 }, { "epoch": 83.4, "grad_norm": 75.9678726196289, "learning_rate": 4.124884705123619e-06, "loss": 0.3297, "step": 83400 }, { "epoch": 83.43, "grad_norm": 43.1819953918457, "learning_rate": 4.121791777987624e-06, "loss": 0.3341, "step": 83430 }, { "epoch": 83.46, "grad_norm": 61.54958724975586, "learning_rate": 4.118699197554327e-06, "loss": 0.3382, "step": 83460 }, { "epoch": 83.49, "grad_norm": 48.813873291015625, "learning_rate": 4.115606965044628e-06, "loss": 0.3135, "step": 83490 }, { "epoch": 83.52, "grad_norm": 63.883052825927734, "learning_rate": 4.112515081679295e-06, "loss": 0.2988, "step": 83520 }, { "epoch": 83.55, "grad_norm": 49.02377700805664, "learning_rate": 4.109423548678949e-06, "loss": 0.2873, "step": 83550 }, { "epoch": 83.58, "grad_norm": 86.0242691040039, "learning_rate": 4.106332367264085e-06, "loss": 0.3628, "step": 83580 }, { "epoch": 83.61, "grad_norm": 69.9734878540039, "learning_rate": 4.103241538655049e-06, "loss": 0.3489, "step": 83610 }, { "epoch": 83.64, "grad_norm": 76.19281768798828, "learning_rate": 4.1001510640720525e-06, "loss": 0.3615, "step": 83640 }, { "epoch": 83.67, "grad_norm": 279.1346435546875, "learning_rate": 4.0970609447351635e-06, "loss": 0.4275, "step": 83670 }, { "epoch": 83.7, "grad_norm": 60.1577033996582, "learning_rate": 4.093971181864313e-06, "loss": 0.3526, "step": 83700 }, { "epoch": 83.73, "grad_norm": 30.274580001831055, "learning_rate": 4.090881776679293e-06, "loss": 0.3522, "step": 83730 }, { "epoch": 83.76, "grad_norm": 62.87873077392578, "learning_rate": 4.087792730399749e-06, "loss": 0.2758, "step": 83760 }, { "epoch": 83.79, "grad_norm": 70.08503723144531, "learning_rate": 4.08470404424519e-06, "loss": 0.3605, "step": 83790 }, { "epoch": 83.82, "grad_norm": 69.52849578857422, "learning_rate": 4.081615719434978e-06, "loss": 0.3352, "step": 83820 }, { "epoch": 83.85, "grad_norm": 87.45691680908203, "learning_rate": 4.078527757188333e-06, "loss": 0.3296, "step": 83850 }, { "epoch": 83.88, "grad_norm": 71.50956726074219, "learning_rate": 4.075440158724339e-06, "loss": 0.4129, "step": 83880 }, { "epoch": 83.91, "grad_norm": 33.176918029785156, "learning_rate": 4.0723529252619276e-06, "loss": 0.2661, "step": 83910 }, { "epoch": 83.94, "grad_norm": 64.66427612304688, "learning_rate": 4.0692660580198905e-06, "loss": 0.3307, "step": 83940 }, { "epoch": 83.97, "grad_norm": 38.23569107055664, "learning_rate": 4.066179558216874e-06, "loss": 0.3588, "step": 83970 }, { "epoch": 84.0, "grad_norm": 64.9631118774414, "learning_rate": 4.063093427071376e-06, "loss": 0.3136, "step": 84000 }, { "epoch": 84.0, "eval_loss": 0.26553860306739807, "eval_map": 0.8157, "eval_map_50": 0.9692, "eval_map_75": 0.9236, "eval_map_chicken": 0.8092, "eval_map_duck": 0.7676, "eval_map_large": 0.8269, "eval_map_medium": 0.8136, "eval_map_plant": 0.8703, "eval_map_small": 0.3004, "eval_mar_1": 0.3272, "eval_mar_10": 0.85, "eval_mar_100": 0.8539, "eval_mar_100_chicken": 0.8573, "eval_mar_100_duck": 0.8031, "eval_mar_100_plant": 0.9012, "eval_mar_large": 0.8615, "eval_mar_medium": 0.8524, "eval_mar_small": 0.4066, "eval_runtime": 12.9008, "eval_samples_per_second": 15.503, "eval_steps_per_second": 1.938, "step": 84000 }, { "epoch": 84.03, "grad_norm": 20.19791030883789, "learning_rate": 4.0600076658017585e-06, "loss": 0.3384, "step": 84030 }, { "epoch": 84.06, "grad_norm": 170.56761169433594, "learning_rate": 4.056922275626227e-06, "loss": 0.3152, "step": 84060 }, { "epoch": 84.09, "grad_norm": 67.12255096435547, "learning_rate": 4.053837257762846e-06, "loss": 0.4929, "step": 84090 }, { "epoch": 84.12, "grad_norm": 32.32781982421875, "learning_rate": 4.0507526134295314e-06, "loss": 0.3457, "step": 84120 }, { "epoch": 84.15, "grad_norm": 61.14581298828125, "learning_rate": 4.047668343844051e-06, "loss": 0.3791, "step": 84150 }, { "epoch": 84.18, "grad_norm": 114.13729858398438, "learning_rate": 4.044584450224026e-06, "loss": 0.2964, "step": 84180 }, { "epoch": 84.21, "grad_norm": 89.34334564208984, "learning_rate": 4.04150093378693e-06, "loss": 0.3439, "step": 84210 }, { "epoch": 84.24, "grad_norm": 82.59429168701172, "learning_rate": 4.038417795750086e-06, "loss": 0.3035, "step": 84240 }, { "epoch": 84.27, "grad_norm": 83.0162582397461, "learning_rate": 4.035335037330668e-06, "loss": 0.2918, "step": 84270 }, { "epoch": 84.3, "grad_norm": 36.42682647705078, "learning_rate": 4.032252659745699e-06, "loss": 0.3563, "step": 84300 }, { "epoch": 84.33, "grad_norm": 67.92389678955078, "learning_rate": 4.0291706642120545e-06, "loss": 0.2922, "step": 84330 }, { "epoch": 84.36, "grad_norm": 65.03959655761719, "learning_rate": 4.0260890519464565e-06, "loss": 0.3676, "step": 84360 }, { "epoch": 84.39, "grad_norm": 51.363224029541016, "learning_rate": 4.023007824165476e-06, "loss": 0.3078, "step": 84390 }, { "epoch": 84.42, "grad_norm": 34.96634292602539, "learning_rate": 4.019926982085536e-06, "loss": 0.452, "step": 84420 }, { "epoch": 84.45, "grad_norm": 79.77047729492188, "learning_rate": 4.016846526922901e-06, "loss": 0.3825, "step": 84450 }, { "epoch": 84.48, "grad_norm": 49.688419342041016, "learning_rate": 4.013766459893686e-06, "loss": 0.4007, "step": 84480 }, { "epoch": 84.51, "grad_norm": 142.57315063476562, "learning_rate": 4.010686782213855e-06, "loss": 0.3645, "step": 84510 }, { "epoch": 84.54, "grad_norm": 57.285888671875, "learning_rate": 4.007607495099215e-06, "loss": 0.4522, "step": 84540 }, { "epoch": 84.57, "grad_norm": 30.3106689453125, "learning_rate": 4.004528599765419e-06, "loss": 0.3362, "step": 84570 }, { "epoch": 84.6, "grad_norm": 95.64994049072266, "learning_rate": 4.001450097427965e-06, "loss": 0.337, "step": 84600 }, { "epoch": 84.63, "grad_norm": 48.63984298706055, "learning_rate": 3.9983719893022e-06, "loss": 0.4045, "step": 84630 }, { "epoch": 84.66, "grad_norm": 60.29104232788086, "learning_rate": 3.995294276603312e-06, "loss": 0.3819, "step": 84660 }, { "epoch": 84.69, "grad_norm": 54.07719421386719, "learning_rate": 3.9922169605463305e-06, "loss": 0.3462, "step": 84690 }, { "epoch": 84.72, "grad_norm": 205.91461181640625, "learning_rate": 3.989140042346134e-06, "loss": 0.3415, "step": 84720 }, { "epoch": 84.75, "grad_norm": 48.897457122802734, "learning_rate": 3.986063523217439e-06, "loss": 0.3623, "step": 84750 }, { "epoch": 84.78, "grad_norm": 73.14813232421875, "learning_rate": 3.9829874043748064e-06, "loss": 0.3552, "step": 84780 }, { "epoch": 84.81, "grad_norm": 76.24201202392578, "learning_rate": 3.979911687032642e-06, "loss": 0.2554, "step": 84810 }, { "epoch": 84.84, "grad_norm": 46.18905258178711, "learning_rate": 3.9768363724051875e-06, "loss": 0.3797, "step": 84840 }, { "epoch": 84.87, "grad_norm": 25.6661319732666, "learning_rate": 3.97376146170653e-06, "loss": 0.3261, "step": 84870 }, { "epoch": 84.9, "grad_norm": 65.86599731445312, "learning_rate": 3.970686956150595e-06, "loss": 0.3218, "step": 84900 }, { "epoch": 84.93, "grad_norm": 55.88581848144531, "learning_rate": 3.967612856951146e-06, "loss": 0.3119, "step": 84930 }, { "epoch": 84.96, "grad_norm": 60.460121154785156, "learning_rate": 3.964539165321795e-06, "loss": 0.3533, "step": 84960 }, { "epoch": 84.99, "grad_norm": 58.134727478027344, "learning_rate": 3.9614658824759815e-06, "loss": 0.2849, "step": 84990 }, { "epoch": 85.0, "eval_loss": 0.26585152745246887, "eval_map": 0.8126, "eval_map_50": 0.9654, "eval_map_75": 0.9284, "eval_map_chicken": 0.7995, "eval_map_duck": 0.7634, "eval_map_large": 0.8435, "eval_map_medium": 0.8039, "eval_map_plant": 0.8748, "eval_map_small": 0.2584, "eval_mar_1": 0.3268, "eval_mar_10": 0.8465, "eval_mar_100": 0.85, "eval_mar_100_chicken": 0.8459, "eval_mar_100_duck": 0.7995, "eval_mar_100_plant": 0.9045, "eval_mar_large": 0.8762, "eval_mar_medium": 0.8451, "eval_mar_small": 0.3664, "eval_runtime": 13.1202, "eval_samples_per_second": 15.244, "eval_steps_per_second": 1.905, "step": 85000 }, { "epoch": 85.02, "grad_norm": 60.16609191894531, "learning_rate": 3.95839300962699e-06, "loss": 0.3734, "step": 85020 }, { "epoch": 85.05, "grad_norm": 73.15736389160156, "learning_rate": 3.955320547987943e-06, "loss": 0.2968, "step": 85050 }, { "epoch": 85.08, "grad_norm": 75.577392578125, "learning_rate": 3.952248498771797e-06, "loss": 0.3165, "step": 85080 }, { "epoch": 85.11, "grad_norm": 80.26549530029297, "learning_rate": 3.949176863191353e-06, "loss": 0.2937, "step": 85110 }, { "epoch": 85.14, "grad_norm": 36.90617370605469, "learning_rate": 3.946105642459241e-06, "loss": 0.4145, "step": 85140 }, { "epoch": 85.17, "grad_norm": 53.00357437133789, "learning_rate": 3.9430348377879315e-06, "loss": 0.3111, "step": 85170 }, { "epoch": 85.2, "grad_norm": 50.074554443359375, "learning_rate": 3.939964450389728e-06, "loss": 0.3524, "step": 85200 }, { "epoch": 85.23, "grad_norm": 111.48501586914062, "learning_rate": 3.9368944814767704e-06, "loss": 0.3934, "step": 85230 }, { "epoch": 85.26, "grad_norm": 75.2789306640625, "learning_rate": 3.9338249322610375e-06, "loss": 0.3257, "step": 85260 }, { "epoch": 85.29, "grad_norm": 63.08626174926758, "learning_rate": 3.9307558039543355e-06, "loss": 0.3148, "step": 85290 }, { "epoch": 85.32, "grad_norm": 47.67877960205078, "learning_rate": 3.927687097768309e-06, "loss": 0.3993, "step": 85320 }, { "epoch": 85.35, "grad_norm": 60.695404052734375, "learning_rate": 3.924618814914435e-06, "loss": 0.3349, "step": 85350 }, { "epoch": 85.38, "grad_norm": 76.16695404052734, "learning_rate": 3.921550956604019e-06, "loss": 0.3315, "step": 85380 }, { "epoch": 85.41, "grad_norm": 0.00012966377835255116, "learning_rate": 3.918483524048208e-06, "loss": 0.3288, "step": 85410 }, { "epoch": 85.44, "grad_norm": 58.6002311706543, "learning_rate": 3.915416518457974e-06, "loss": 0.3345, "step": 85440 }, { "epoch": 85.47, "grad_norm": 0.00010767208004835993, "learning_rate": 3.912349941044122e-06, "loss": 0.3153, "step": 85470 }, { "epoch": 85.5, "grad_norm": 74.4843978881836, "learning_rate": 3.909283793017289e-06, "loss": 0.3126, "step": 85500 }, { "epoch": 85.53, "grad_norm": 95.2237777709961, "learning_rate": 3.906218075587938e-06, "loss": 0.3689, "step": 85530 }, { "epoch": 85.56, "grad_norm": 66.24314880371094, "learning_rate": 3.9031527899663705e-06, "loss": 0.352, "step": 85560 }, { "epoch": 85.59, "grad_norm": 35.54780960083008, "learning_rate": 3.900087937362711e-06, "loss": 0.3243, "step": 85590 }, { "epoch": 85.62, "grad_norm": 38.614864349365234, "learning_rate": 3.897023518986915e-06, "loss": 0.3498, "step": 85620 }, { "epoch": 85.65, "grad_norm": 0.00013191276229918003, "learning_rate": 3.8939595360487655e-06, "loss": 0.3304, "step": 85650 }, { "epoch": 85.68, "grad_norm": 9.524432971375063e-05, "learning_rate": 3.890895989757874e-06, "loss": 0.2491, "step": 85680 }, { "epoch": 85.71, "grad_norm": 26.40590476989746, "learning_rate": 3.887832881323681e-06, "loss": 0.393, "step": 85710 }, { "epoch": 85.74, "grad_norm": 52.44107437133789, "learning_rate": 3.884770211955454e-06, "loss": 0.3528, "step": 85740 }, { "epoch": 85.77, "grad_norm": 43.583675384521484, "learning_rate": 3.8817079828622855e-06, "loss": 0.3657, "step": 85770 }, { "epoch": 85.8, "grad_norm": 48.950401306152344, "learning_rate": 3.8786461952530955e-06, "loss": 0.3474, "step": 85800 }, { "epoch": 85.83, "grad_norm": 68.08635711669922, "learning_rate": 3.875584850336627e-06, "loss": 0.3348, "step": 85830 }, { "epoch": 85.86, "grad_norm": 36.340003967285156, "learning_rate": 3.872523949321454e-06, "loss": 0.2759, "step": 85860 }, { "epoch": 85.89, "grad_norm": 97.86219787597656, "learning_rate": 3.869463493415969e-06, "loss": 0.2911, "step": 85890 }, { "epoch": 85.92, "grad_norm": 89.8642578125, "learning_rate": 3.866403483828392e-06, "loss": 0.3529, "step": 85920 }, { "epoch": 85.95, "grad_norm": 81.55255126953125, "learning_rate": 3.863343921766769e-06, "loss": 0.3172, "step": 85950 }, { "epoch": 85.98, "grad_norm": 96.72074127197266, "learning_rate": 3.860284808438962e-06, "loss": 0.3634, "step": 85980 }, { "epoch": 86.0, "eval_loss": 0.2642298638820648, "eval_map": 0.8194, "eval_map_50": 0.9604, "eval_map_75": 0.9231, "eval_map_chicken": 0.8109, "eval_map_duck": 0.7685, "eval_map_large": 0.8379, "eval_map_medium": 0.8143, "eval_map_plant": 0.8789, "eval_map_small": 0.2498, "eval_mar_1": 0.3308, "eval_mar_10": 0.8509, "eval_mar_100": 0.8538, "eval_mar_100_chicken": 0.8541, "eval_mar_100_duck": 0.8021, "eval_mar_100_plant": 0.9051, "eval_mar_large": 0.8711, "eval_mar_medium": 0.8519, "eval_mar_small": 0.2906, "eval_runtime": 12.7656, "eval_samples_per_second": 15.667, "eval_steps_per_second": 1.958, "step": 86000 }, { "epoch": 86.01, "grad_norm": 37.73887252807617, "learning_rate": 3.857226145052665e-06, "loss": 0.377, "step": 86010 }, { "epoch": 86.04, "grad_norm": 96.7146224975586, "learning_rate": 3.854167932815387e-06, "loss": 0.3278, "step": 86040 }, { "epoch": 86.07, "grad_norm": 51.56999588012695, "learning_rate": 3.851110172934463e-06, "loss": 0.3336, "step": 86070 }, { "epoch": 86.1, "grad_norm": 132.88180541992188, "learning_rate": 3.8480528666170495e-06, "loss": 0.2838, "step": 86100 }, { "epoch": 86.13, "grad_norm": 56.52324295043945, "learning_rate": 3.8449960150701175e-06, "loss": 0.3308, "step": 86130 }, { "epoch": 86.16, "grad_norm": 97.8554916381836, "learning_rate": 3.841939619500468e-06, "loss": 0.2807, "step": 86160 }, { "epoch": 86.19, "grad_norm": 60.00990295410156, "learning_rate": 3.838883681114718e-06, "loss": 0.2808, "step": 86190 }, { "epoch": 86.22, "grad_norm": 168.734130859375, "learning_rate": 3.835828201119302e-06, "loss": 0.4305, "step": 86220 }, { "epoch": 86.25, "grad_norm": 136.18235778808594, "learning_rate": 3.832773180720475e-06, "loss": 0.3312, "step": 86250 }, { "epoch": 86.28, "grad_norm": 93.48727416992188, "learning_rate": 3.8297186211243085e-06, "loss": 0.3465, "step": 86280 }, { "epoch": 86.31, "grad_norm": 8.836453343974426e-05, "learning_rate": 3.8266645235367e-06, "loss": 0.3126, "step": 86310 }, { "epoch": 86.34, "grad_norm": 32.67022705078125, "learning_rate": 3.823610889163354e-06, "loss": 0.376, "step": 86340 }, { "epoch": 86.37, "grad_norm": 31.43949317932129, "learning_rate": 3.820557719209799e-06, "loss": 0.2897, "step": 86370 }, { "epoch": 86.4, "grad_norm": 71.29381561279297, "learning_rate": 3.817505014881378e-06, "loss": 0.2769, "step": 86400 }, { "epoch": 86.43, "grad_norm": 72.68944549560547, "learning_rate": 3.814452777383248e-06, "loss": 0.3056, "step": 86430 }, { "epoch": 86.46, "grad_norm": 87.3520736694336, "learning_rate": 3.8114010079203877e-06, "loss": 0.2856, "step": 86460 }, { "epoch": 86.49, "grad_norm": 115.8770751953125, "learning_rate": 3.8083497076975863e-06, "loss": 0.3738, "step": 86490 }, { "epoch": 86.52, "grad_norm": 0.0001854299771366641, "learning_rate": 3.8052988779194478e-06, "loss": 0.3078, "step": 86520 }, { "epoch": 86.55, "grad_norm": 60.985191345214844, "learning_rate": 3.8022485197903924e-06, "loss": 0.3331, "step": 86550 }, { "epoch": 86.58, "grad_norm": 39.975135803222656, "learning_rate": 3.7991986345146503e-06, "loss": 0.3402, "step": 86580 }, { "epoch": 86.61, "grad_norm": 21.72452735900879, "learning_rate": 3.796149223296272e-06, "loss": 0.3744, "step": 86610 }, { "epoch": 86.64, "grad_norm": 91.82344055175781, "learning_rate": 3.7931002873391156e-06, "loss": 0.3605, "step": 86640 }, { "epoch": 86.67, "grad_norm": 115.16731262207031, "learning_rate": 3.790051827846851e-06, "loss": 0.3041, "step": 86670 }, { "epoch": 86.7, "grad_norm": 29.067026138305664, "learning_rate": 3.787003846022964e-06, "loss": 0.3237, "step": 86700 }, { "epoch": 86.73, "grad_norm": 143.17469787597656, "learning_rate": 3.783956343070746e-06, "loss": 0.3841, "step": 86730 }, { "epoch": 86.76, "grad_norm": 64.04499053955078, "learning_rate": 3.7809093201933078e-06, "loss": 0.281, "step": 86760 }, { "epoch": 86.79, "grad_norm": 53.69562911987305, "learning_rate": 3.7778627785935627e-06, "loss": 0.3384, "step": 86790 }, { "epoch": 86.82, "grad_norm": 64.10151672363281, "learning_rate": 3.774816719474238e-06, "loss": 0.2846, "step": 86820 }, { "epoch": 86.85, "grad_norm": 60.3651008605957, "learning_rate": 3.7717711440378695e-06, "loss": 0.2902, "step": 86850 }, { "epoch": 86.88, "grad_norm": 54.901451110839844, "learning_rate": 3.7687260534868e-06, "loss": 0.3788, "step": 86880 }, { "epoch": 86.91, "grad_norm": 89.72407531738281, "learning_rate": 3.7656814490231864e-06, "loss": 0.3692, "step": 86910 }, { "epoch": 86.94, "grad_norm": 42.89752197265625, "learning_rate": 3.762637331848989e-06, "loss": 0.3348, "step": 86940 }, { "epoch": 86.97, "grad_norm": 102.75811767578125, "learning_rate": 3.7595937031659775e-06, "loss": 0.3862, "step": 86970 }, { "epoch": 87.0, "grad_norm": 53.49305725097656, "learning_rate": 3.756550564175727e-06, "loss": 0.4086, "step": 87000 }, { "epoch": 87.0, "eval_loss": 0.26547977328300476, "eval_map": 0.8124, "eval_map_50": 0.9649, "eval_map_75": 0.924, "eval_map_chicken": 0.7972, "eval_map_duck": 0.7569, "eval_map_large": 0.8237, "eval_map_medium": 0.8076, "eval_map_plant": 0.8832, "eval_map_small": 0.3069, "eval_mar_1": 0.3279, "eval_mar_10": 0.8446, "eval_mar_100": 0.8484, "eval_mar_100_chicken": 0.8421, "eval_mar_100_duck": 0.7938, "eval_mar_100_plant": 0.9095, "eval_mar_large": 0.8606, "eval_mar_medium": 0.845, "eval_mar_small": 0.3824, "eval_runtime": 12.7707, "eval_samples_per_second": 15.661, "eval_steps_per_second": 1.958, "step": 87000 }, { "epoch": 87.03, "grad_norm": 39.80609130859375, "learning_rate": 3.7535079160796207e-06, "loss": 0.4166, "step": 87030 }, { "epoch": 87.06, "grad_norm": 32.69429016113281, "learning_rate": 3.7504657600788484e-06, "loss": 0.2969, "step": 87060 }, { "epoch": 87.09, "grad_norm": 114.84684753417969, "learning_rate": 3.7474240973744063e-06, "loss": 0.3284, "step": 87090 }, { "epoch": 87.12, "grad_norm": 40.37647247314453, "learning_rate": 3.744382929167094e-06, "loss": 0.3487, "step": 87120 }, { "epoch": 87.15, "grad_norm": 59.3717041015625, "learning_rate": 3.7413422566575153e-06, "loss": 0.3087, "step": 87150 }, { "epoch": 87.18, "grad_norm": 48.4447135925293, "learning_rate": 3.73830208104608e-06, "loss": 0.3016, "step": 87180 }, { "epoch": 87.21, "grad_norm": 31.026857376098633, "learning_rate": 3.735262403533002e-06, "loss": 0.4084, "step": 87210 }, { "epoch": 87.24, "grad_norm": 68.91844940185547, "learning_rate": 3.7322232253182984e-06, "loss": 0.3885, "step": 87240 }, { "epoch": 87.27, "grad_norm": 65.4188232421875, "learning_rate": 3.729184547601786e-06, "loss": 0.3927, "step": 87270 }, { "epoch": 87.3, "grad_norm": 0.00015406549209728837, "learning_rate": 3.7261463715830902e-06, "loss": 0.3533, "step": 87300 }, { "epoch": 87.33, "grad_norm": 120.36640167236328, "learning_rate": 3.7231086984616312e-06, "loss": 0.4106, "step": 87330 }, { "epoch": 87.36, "grad_norm": 31.269691467285156, "learning_rate": 3.7200715294366376e-06, "loss": 0.4366, "step": 87360 }, { "epoch": 87.39, "grad_norm": 54.8677864074707, "learning_rate": 3.717034865707133e-06, "loss": 0.273, "step": 87390 }, { "epoch": 87.42, "grad_norm": 100.52903747558594, "learning_rate": 3.7139987084719463e-06, "loss": 0.3616, "step": 87420 }, { "epoch": 87.45, "grad_norm": 59.32597351074219, "learning_rate": 3.7109630589297014e-06, "loss": 0.3159, "step": 87450 }, { "epoch": 87.48, "grad_norm": 45.1649284362793, "learning_rate": 3.7079279182788263e-06, "loss": 0.322, "step": 87480 }, { "epoch": 87.51, "grad_norm": 66.52740478515625, "learning_rate": 3.704893287717548e-06, "loss": 0.2672, "step": 87510 }, { "epoch": 87.54, "grad_norm": 67.82477569580078, "learning_rate": 3.70185916844389e-06, "loss": 0.3564, "step": 87540 }, { "epoch": 87.57, "grad_norm": 109.45364379882812, "learning_rate": 3.6988255616556725e-06, "loss": 0.4437, "step": 87570 }, { "epoch": 87.6, "grad_norm": 63.78138732910156, "learning_rate": 3.695792468550517e-06, "loss": 0.2685, "step": 87600 }, { "epoch": 87.63, "grad_norm": 51.303985595703125, "learning_rate": 3.6927598903258375e-06, "loss": 0.3425, "step": 87630 }, { "epoch": 87.66, "grad_norm": 58.48736572265625, "learning_rate": 3.689727828178854e-06, "loss": 0.3662, "step": 87660 }, { "epoch": 87.69, "grad_norm": 55.86106491088867, "learning_rate": 3.686696283306572e-06, "loss": 0.2612, "step": 87690 }, { "epoch": 87.72, "grad_norm": 57.764408111572266, "learning_rate": 3.6836652569057994e-06, "loss": 0.4982, "step": 87720 }, { "epoch": 87.75, "grad_norm": 84.55885314941406, "learning_rate": 3.680634750173137e-06, "loss": 0.3033, "step": 87750 }, { "epoch": 87.78, "grad_norm": 51.23897933959961, "learning_rate": 3.6776047643049777e-06, "loss": 0.3505, "step": 87780 }, { "epoch": 87.81, "grad_norm": 64.15745544433594, "learning_rate": 3.674575300497517e-06, "loss": 0.5061, "step": 87810 }, { "epoch": 87.84, "grad_norm": 41.513938903808594, "learning_rate": 3.6715463599467372e-06, "loss": 0.6107, "step": 87840 }, { "epoch": 87.87, "grad_norm": 45.170021057128906, "learning_rate": 3.668517943848416e-06, "loss": 0.333, "step": 87870 }, { "epoch": 87.9, "grad_norm": 93.24050903320312, "learning_rate": 3.6654900533981234e-06, "loss": 0.333, "step": 87900 }, { "epoch": 87.93, "grad_norm": 43.659339904785156, "learning_rate": 3.6624626897912213e-06, "loss": 0.3786, "step": 87930 }, { "epoch": 87.96, "grad_norm": 35.089202880859375, "learning_rate": 3.659435854222869e-06, "loss": 0.3665, "step": 87960 }, { "epoch": 87.99, "grad_norm": 58.08119201660156, "learning_rate": 3.6564095478880114e-06, "loss": 0.3238, "step": 87990 }, { "epoch": 88.0, "eval_loss": 0.25432494282722473, "eval_map": 0.822, "eval_map_50": 0.9675, "eval_map_75": 0.9315, "eval_map_chicken": 0.8092, "eval_map_duck": 0.7755, "eval_map_large": 0.8393, "eval_map_medium": 0.8157, "eval_map_plant": 0.8814, "eval_map_small": 0.2809, "eval_mar_1": 0.3303, "eval_mar_10": 0.8551, "eval_mar_100": 0.8586, "eval_mar_100_chicken": 0.8541, "eval_mar_100_duck": 0.8144, "eval_mar_100_plant": 0.9073, "eval_mar_large": 0.8723, "eval_mar_medium": 0.8548, "eval_mar_small": 0.3553, "eval_runtime": 13.37, "eval_samples_per_second": 14.959, "eval_steps_per_second": 1.87, "step": 88000 }, { "epoch": 88.02, "grad_norm": 51.55034255981445, "learning_rate": 3.653383771981385e-06, "loss": 0.3944, "step": 88020 }, { "epoch": 88.05, "grad_norm": 53.67812728881836, "learning_rate": 3.6503585276975196e-06, "loss": 0.2862, "step": 88050 }, { "epoch": 88.08, "grad_norm": 39.243526458740234, "learning_rate": 3.6473338162307314e-06, "loss": 0.3254, "step": 88080 }, { "epoch": 88.11, "grad_norm": 32.0158576965332, "learning_rate": 3.644309638775132e-06, "loss": 0.3912, "step": 88110 }, { "epoch": 88.14, "grad_norm": 29.435768127441406, "learning_rate": 3.6412859965246173e-06, "loss": 0.3573, "step": 88140 }, { "epoch": 88.17, "grad_norm": 40.1513786315918, "learning_rate": 3.6382628906728735e-06, "loss": 0.2964, "step": 88170 }, { "epoch": 88.2, "grad_norm": 60.409568786621094, "learning_rate": 3.635240322413375e-06, "loss": 0.3221, "step": 88200 }, { "epoch": 88.23, "grad_norm": 74.93896484375, "learning_rate": 3.6322182929393833e-06, "loss": 0.3499, "step": 88230 }, { "epoch": 88.26, "grad_norm": 78.86674499511719, "learning_rate": 3.6291968034439463e-06, "loss": 0.339, "step": 88260 }, { "epoch": 88.29, "grad_norm": 6.8061170168221e-05, "learning_rate": 3.6261758551199033e-06, "loss": 0.2991, "step": 88290 }, { "epoch": 88.32, "grad_norm": 70.96842956542969, "learning_rate": 3.6231554491598766e-06, "loss": 0.2787, "step": 88320 }, { "epoch": 88.35, "grad_norm": 42.552452087402344, "learning_rate": 3.6201355867562725e-06, "loss": 0.307, "step": 88350 }, { "epoch": 88.38, "grad_norm": 40.119956970214844, "learning_rate": 3.617116269101286e-06, "loss": 0.3465, "step": 88380 }, { "epoch": 88.41, "grad_norm": 23.125516891479492, "learning_rate": 3.614097497386894e-06, "loss": 0.3411, "step": 88410 }, { "epoch": 88.44, "grad_norm": 62.419044494628906, "learning_rate": 3.6110792728048636e-06, "loss": 0.3215, "step": 88440 }, { "epoch": 88.47, "grad_norm": 108.49114990234375, "learning_rate": 3.60806159654674e-06, "loss": 0.347, "step": 88470 }, { "epoch": 88.5, "grad_norm": 83.80387878417969, "learning_rate": 3.6050444698038547e-06, "loss": 0.3362, "step": 88500 }, { "epoch": 88.53, "grad_norm": 74.07994842529297, "learning_rate": 3.6020278937673202e-06, "loss": 0.3782, "step": 88530 }, { "epoch": 88.56, "grad_norm": 58.45664978027344, "learning_rate": 3.599011869628033e-06, "loss": 0.2838, "step": 88560 }, { "epoch": 88.59, "grad_norm": 23.82346534729004, "learning_rate": 3.595996398576672e-06, "loss": 0.3734, "step": 88590 }, { "epoch": 88.62, "grad_norm": 105.38636779785156, "learning_rate": 3.592981481803699e-06, "loss": 0.2881, "step": 88620 }, { "epoch": 88.65, "grad_norm": 35.59571838378906, "learning_rate": 3.5899671204993535e-06, "loss": 0.4293, "step": 88650 }, { "epoch": 88.68, "grad_norm": 36.647891998291016, "learning_rate": 3.5869533158536583e-06, "loss": 0.4212, "step": 88680 }, { "epoch": 88.71, "grad_norm": 27.596078872680664, "learning_rate": 3.583940069056415e-06, "loss": 0.2567, "step": 88710 }, { "epoch": 88.74, "grad_norm": 0.00022858002921566367, "learning_rate": 3.5809273812972078e-06, "loss": 0.3554, "step": 88740 }, { "epoch": 88.77, "grad_norm": 63.96984100341797, "learning_rate": 3.577915253765396e-06, "loss": 0.3208, "step": 88770 }, { "epoch": 88.8, "grad_norm": 79.72564697265625, "learning_rate": 3.5749036876501196e-06, "loss": 0.3069, "step": 88800 }, { "epoch": 88.83, "grad_norm": 46.0412483215332, "learning_rate": 3.5718926841402993e-06, "loss": 0.2983, "step": 88830 }, { "epoch": 88.86, "grad_norm": 26.93777084350586, "learning_rate": 3.5688822444246297e-06, "loss": 0.3477, "step": 88860 }, { "epoch": 88.89, "grad_norm": 51.916900634765625, "learning_rate": 3.5658723696915864e-06, "loss": 0.3159, "step": 88890 }, { "epoch": 88.92, "grad_norm": 44.57571792602539, "learning_rate": 3.562863061129419e-06, "loss": 0.3451, "step": 88920 }, { "epoch": 88.95, "grad_norm": 43.55433654785156, "learning_rate": 3.559854319926156e-06, "loss": 0.3332, "step": 88950 }, { "epoch": 88.98, "grad_norm": 76.697021484375, "learning_rate": 3.556846147269598e-06, "loss": 0.465, "step": 88980 }, { "epoch": 89.0, "eval_loss": 0.26904526352882385, "eval_map": 0.818, "eval_map_50": 0.9661, "eval_map_75": 0.9275, "eval_map_chicken": 0.7979, "eval_map_duck": 0.7725, "eval_map_large": 0.8314, "eval_map_medium": 0.8111, "eval_map_plant": 0.8835, "eval_map_small": 0.3431, "eval_mar_1": 0.3274, "eval_mar_10": 0.8482, "eval_mar_100": 0.8524, "eval_mar_100_chicken": 0.8396, "eval_mar_100_duck": 0.8046, "eval_mar_100_plant": 0.913, "eval_mar_large": 0.8649, "eval_mar_medium": 0.8484, "eval_mar_small": 0.4379, "eval_runtime": 12.7628, "eval_samples_per_second": 15.67, "eval_steps_per_second": 1.959, "step": 89000 }, { "epoch": 89.01, "grad_norm": 46.47408676147461, "learning_rate": 3.553838544347326e-06, "loss": 0.33, "step": 89010 }, { "epoch": 89.04, "grad_norm": 611.4906005859375, "learning_rate": 3.550831512346695e-06, "loss": 0.327, "step": 89040 }, { "epoch": 89.07, "grad_norm": 40.71622848510742, "learning_rate": 3.547825052454833e-06, "loss": 0.3237, "step": 89070 }, { "epoch": 89.1, "grad_norm": 39.560523986816406, "learning_rate": 3.5448191658586423e-06, "loss": 0.3457, "step": 89100 }, { "epoch": 89.13, "grad_norm": 71.82794189453125, "learning_rate": 3.541813853744799e-06, "loss": 0.3524, "step": 89130 }, { "epoch": 89.16, "grad_norm": 61.15816879272461, "learning_rate": 3.538809117299751e-06, "loss": 0.3093, "step": 89160 }, { "epoch": 89.19, "grad_norm": 39.94947052001953, "learning_rate": 3.535804957709724e-06, "loss": 0.3283, "step": 89190 }, { "epoch": 89.22, "grad_norm": 2.1249281417112797e-05, "learning_rate": 3.532801376160713e-06, "loss": 0.307, "step": 89220 }, { "epoch": 89.25, "grad_norm": 64.9295425415039, "learning_rate": 3.5297983738384813e-06, "loss": 0.3303, "step": 89250 }, { "epoch": 89.28, "grad_norm": 50.43611145019531, "learning_rate": 3.526795951928569e-06, "loss": 0.3654, "step": 89280 }, { "epoch": 89.31, "grad_norm": 59.45797348022461, "learning_rate": 3.5237941116162812e-06, "loss": 0.3082, "step": 89310 }, { "epoch": 89.34, "grad_norm": 58.05674362182617, "learning_rate": 3.520792854086702e-06, "loss": 0.3891, "step": 89340 }, { "epoch": 89.37, "grad_norm": 57.51795196533203, "learning_rate": 3.5177921805246772e-06, "loss": 0.2621, "step": 89370 }, { "epoch": 89.4, "grad_norm": 30.999542236328125, "learning_rate": 3.5147920921148267e-06, "loss": 0.3627, "step": 89400 }, { "epoch": 89.43, "grad_norm": 45.953887939453125, "learning_rate": 3.511792590041537e-06, "loss": 0.3322, "step": 89430 }, { "epoch": 89.46, "grad_norm": 420.60186767578125, "learning_rate": 3.5087936754889614e-06, "loss": 0.317, "step": 89460 }, { "epoch": 89.49, "grad_norm": 38.25226593017578, "learning_rate": 3.505795349641029e-06, "loss": 0.33, "step": 89490 }, { "epoch": 89.52, "grad_norm": 58.498321533203125, "learning_rate": 3.502797613681429e-06, "loss": 0.3605, "step": 89520 }, { "epoch": 89.55, "grad_norm": 103.08924865722656, "learning_rate": 3.49980046879362e-06, "loss": 0.3132, "step": 89550 }, { "epoch": 89.58, "grad_norm": 50.665252685546875, "learning_rate": 3.496803916160827e-06, "loss": 0.358, "step": 89580 }, { "epoch": 89.61, "grad_norm": 35.272216796875, "learning_rate": 3.4938079569660398e-06, "loss": 0.3568, "step": 89610 }, { "epoch": 89.64, "grad_norm": 59.047454833984375, "learning_rate": 3.4908125923920204e-06, "loss": 0.284, "step": 89640 }, { "epoch": 89.67, "grad_norm": 179.12362670898438, "learning_rate": 3.4878178236212883e-06, "loss": 0.3554, "step": 89670 }, { "epoch": 89.7, "grad_norm": 35.158905029296875, "learning_rate": 3.484823651836131e-06, "loss": 0.3178, "step": 89700 }, { "epoch": 89.73, "grad_norm": 58.436668395996094, "learning_rate": 3.4818300782186e-06, "loss": 0.3839, "step": 89730 }, { "epoch": 89.76, "grad_norm": 47.66510772705078, "learning_rate": 3.478837103950509e-06, "loss": 0.3277, "step": 89760 }, { "epoch": 89.79, "grad_norm": 78.29817199707031, "learning_rate": 3.4758447302134414e-06, "loss": 0.2835, "step": 89790 }, { "epoch": 89.82, "grad_norm": 121.37081146240234, "learning_rate": 3.472852958188736e-06, "loss": 0.3749, "step": 89820 }, { "epoch": 89.85, "grad_norm": 158.51480102539062, "learning_rate": 3.4698617890574972e-06, "loss": 0.3812, "step": 89850 }, { "epoch": 89.88, "grad_norm": 111.1998519897461, "learning_rate": 3.4668712240005912e-06, "loss": 0.4212, "step": 89880 }, { "epoch": 89.91, "grad_norm": 43.89581298828125, "learning_rate": 3.463881264198645e-06, "loss": 0.3064, "step": 89910 }, { "epoch": 89.94, "grad_norm": 77.7400894165039, "learning_rate": 3.4608919108320488e-06, "loss": 0.2893, "step": 89940 }, { "epoch": 89.97, "grad_norm": 24.88041114807129, "learning_rate": 3.457903165080952e-06, "loss": 0.3734, "step": 89970 }, { "epoch": 90.0, "grad_norm": 41.57048034667969, "learning_rate": 3.4549150281252635e-06, "loss": 0.37, "step": 90000 }, { "epoch": 90.0, "eval_loss": 0.26016053557395935, "eval_map": 0.8221, "eval_map_50": 0.9593, "eval_map_75": 0.9235, "eval_map_chicken": 0.8138, "eval_map_duck": 0.7712, "eval_map_large": 0.8417, "eval_map_medium": 0.8197, "eval_map_plant": 0.8813, "eval_map_small": 0.2761, "eval_mar_1": 0.3301, "eval_mar_10": 0.8531, "eval_mar_100": 0.8564, "eval_mar_100_chicken": 0.8584, "eval_mar_100_duck": 0.801, "eval_mar_100_plant": 0.9098, "eval_mar_large": 0.8744, "eval_mar_medium": 0.8569, "eval_mar_small": 0.3389, "eval_runtime": 12.6911, "eval_samples_per_second": 15.759, "eval_steps_per_second": 1.97, "step": 90000 }, { "epoch": 90.03, "grad_norm": 52.27177047729492, "learning_rate": 3.451927501144653e-06, "loss": 0.4317, "step": 90030 }, { "epoch": 90.06, "grad_norm": 34.04719161987305, "learning_rate": 3.4489405853185465e-06, "loss": 0.3575, "step": 90060 }, { "epoch": 90.09, "grad_norm": 50.2287712097168, "learning_rate": 3.445954281826134e-06, "loss": 0.2949, "step": 90090 }, { "epoch": 90.12, "grad_norm": 29.608713150024414, "learning_rate": 3.442968591846359e-06, "loss": 0.3328, "step": 90120 }, { "epoch": 90.15, "grad_norm": 42.25761032104492, "learning_rate": 3.4399835165579266e-06, "loss": 0.323, "step": 90150 }, { "epoch": 90.18, "grad_norm": 87.7109146118164, "learning_rate": 3.436999057139295e-06, "loss": 0.3235, "step": 90180 }, { "epoch": 90.21, "grad_norm": 45.1929931640625, "learning_rate": 3.4340152147686824e-06, "loss": 0.3688, "step": 90210 }, { "epoch": 90.24, "grad_norm": 65.5608139038086, "learning_rate": 3.431031990624063e-06, "loss": 0.3102, "step": 90240 }, { "epoch": 90.27, "grad_norm": 26.552268981933594, "learning_rate": 3.4280493858831665e-06, "loss": 0.3385, "step": 90270 }, { "epoch": 90.3, "grad_norm": 56.29317092895508, "learning_rate": 3.4250674017234774e-06, "loss": 0.356, "step": 90300 }, { "epoch": 90.33, "grad_norm": 56.497127532958984, "learning_rate": 3.4220860393222347e-06, "loss": 0.3461, "step": 90330 }, { "epoch": 90.36, "grad_norm": 51.76903533935547, "learning_rate": 3.4191052998564344e-06, "loss": 0.3499, "step": 90360 }, { "epoch": 90.39, "grad_norm": 50.63872146606445, "learning_rate": 3.4161251845028264e-06, "loss": 0.3409, "step": 90390 }, { "epoch": 90.42, "grad_norm": 91.92029571533203, "learning_rate": 3.4131456944379126e-06, "loss": 0.3199, "step": 90420 }, { "epoch": 90.45, "grad_norm": 36.55586242675781, "learning_rate": 3.4101668308379465e-06, "loss": 0.3261, "step": 90450 }, { "epoch": 90.48, "grad_norm": 67.41263580322266, "learning_rate": 3.407188594878938e-06, "loss": 0.2894, "step": 90480 }, { "epoch": 90.51, "grad_norm": 74.78584289550781, "learning_rate": 3.4042109877366447e-06, "loss": 0.3116, "step": 90510 }, { "epoch": 90.54, "grad_norm": 23.95796012878418, "learning_rate": 3.401234010586583e-06, "loss": 0.3469, "step": 90540 }, { "epoch": 90.57, "grad_norm": 82.01903533935547, "learning_rate": 3.398257664604015e-06, "loss": 0.3158, "step": 90570 }, { "epoch": 90.6, "grad_norm": 35.768714904785156, "learning_rate": 3.3952819509639534e-06, "loss": 0.3388, "step": 90600 }, { "epoch": 90.63, "grad_norm": 53.178489685058594, "learning_rate": 3.3923068708411645e-06, "loss": 0.3081, "step": 90630 }, { "epoch": 90.66, "grad_norm": 70.48599243164062, "learning_rate": 3.38933242541016e-06, "loss": 0.3565, "step": 90660 }, { "epoch": 90.69, "grad_norm": 63.03761672973633, "learning_rate": 3.3863586158452074e-06, "loss": 0.378, "step": 90690 }, { "epoch": 90.72, "grad_norm": 53.49186706542969, "learning_rate": 3.3833854433203185e-06, "loss": 0.3065, "step": 90720 }, { "epoch": 90.75, "grad_norm": 44.166812896728516, "learning_rate": 3.3804129090092542e-06, "loss": 0.315, "step": 90750 }, { "epoch": 90.78, "grad_norm": 8.893087215255946e-05, "learning_rate": 3.377441014085524e-06, "loss": 0.3304, "step": 90780 }, { "epoch": 90.81, "grad_norm": 73.69300842285156, "learning_rate": 3.374469759722383e-06, "loss": 0.4061, "step": 90810 }, { "epoch": 90.84, "grad_norm": 42.10377502441406, "learning_rate": 3.3714991470928393e-06, "loss": 0.3538, "step": 90840 }, { "epoch": 90.87, "grad_norm": 45.363765716552734, "learning_rate": 3.3685291773696425e-06, "loss": 0.3352, "step": 90870 }, { "epoch": 90.9, "grad_norm": 61.1765251159668, "learning_rate": 3.3655598517252886e-06, "loss": 0.3248, "step": 90900 }, { "epoch": 90.93, "grad_norm": 31.50785255432129, "learning_rate": 3.3625911713320205e-06, "loss": 0.4373, "step": 90930 }, { "epoch": 90.96, "grad_norm": 30.814041137695312, "learning_rate": 3.359623137361825e-06, "loss": 0.4208, "step": 90960 }, { "epoch": 90.99, "grad_norm": 49.471004486083984, "learning_rate": 3.3566557509864374e-06, "loss": 0.3063, "step": 90990 }, { "epoch": 91.0, "eval_loss": 0.26169589161872864, "eval_map": 0.8144, "eval_map_50": 0.9619, "eval_map_75": 0.9289, "eval_map_chicken": 0.8069, "eval_map_duck": 0.7583, "eval_map_large": 0.8147, "eval_map_medium": 0.8134, "eval_map_plant": 0.8779, "eval_map_small": 0.3302, "eval_mar_1": 0.3274, "eval_mar_10": 0.8485, "eval_mar_100": 0.8515, "eval_mar_100_chicken": 0.8575, "eval_mar_100_duck": 0.7902, "eval_mar_100_plant": 0.9068, "eval_mar_large": 0.8532, "eval_mar_medium": 0.8521, "eval_mar_small": 0.4172, "eval_runtime": 13.317, "eval_samples_per_second": 15.018, "eval_steps_per_second": 1.877, "step": 91000 }, { "epoch": 91.02, "grad_norm": 52.58911895751953, "learning_rate": 3.3536890133773346e-06, "loss": 0.3383, "step": 91020 }, { "epoch": 91.05, "grad_norm": 20.826831817626953, "learning_rate": 3.350722925705736e-06, "loss": 0.3566, "step": 91050 }, { "epoch": 91.08, "grad_norm": 19.2310733795166, "learning_rate": 3.347757489142608e-06, "loss": 0.3123, "step": 91080 }, { "epoch": 91.11, "grad_norm": 54.414798736572266, "learning_rate": 3.3447927048586538e-06, "loss": 0.368, "step": 91110 }, { "epoch": 91.14, "grad_norm": 123.11380004882812, "learning_rate": 3.3418285740243285e-06, "loss": 0.2497, "step": 91140 }, { "epoch": 91.17, "grad_norm": 64.96946716308594, "learning_rate": 3.3388650978098218e-06, "loss": 0.3119, "step": 91170 }, { "epoch": 91.2, "grad_norm": 44.66423797607422, "learning_rate": 3.3359022773850673e-06, "loss": 0.3627, "step": 91200 }, { "epoch": 91.23, "grad_norm": 99.88350677490234, "learning_rate": 3.3329401139197393e-06, "loss": 0.2681, "step": 91230 }, { "epoch": 91.26, "grad_norm": 134.87847900390625, "learning_rate": 3.329978608583252e-06, "loss": 0.3374, "step": 91260 }, { "epoch": 91.29, "grad_norm": 115.56591033935547, "learning_rate": 3.3270177625447632e-06, "loss": 0.2886, "step": 91290 }, { "epoch": 91.32, "grad_norm": 3.729057789314538e-05, "learning_rate": 3.3240575769731662e-06, "loss": 0.3311, "step": 91320 }, { "epoch": 91.35, "grad_norm": 45.575443267822266, "learning_rate": 3.3210980530370974e-06, "loss": 0.3438, "step": 91350 }, { "epoch": 91.38, "grad_norm": 62.94804763793945, "learning_rate": 3.3181391919049277e-06, "loss": 0.3834, "step": 91380 }, { "epoch": 91.41, "grad_norm": 61.31102752685547, "learning_rate": 3.315180994744769e-06, "loss": 0.3979, "step": 91410 }, { "epoch": 91.44, "grad_norm": 35.14569854736328, "learning_rate": 3.312223462724472e-06, "loss": 0.3055, "step": 91440 }, { "epoch": 91.47, "grad_norm": 57.5945930480957, "learning_rate": 3.309266597011621e-06, "loss": 0.3426, "step": 91470 }, { "epoch": 91.5, "grad_norm": 74.72227478027344, "learning_rate": 3.3063103987735433e-06, "loss": 0.3089, "step": 91500 }, { "epoch": 91.53, "grad_norm": 65.88390350341797, "learning_rate": 3.3033548691772976e-06, "loss": 0.344, "step": 91530 }, { "epoch": 91.56, "grad_norm": 77.1495132446289, "learning_rate": 3.300400009389678e-06, "loss": 0.2721, "step": 91560 }, { "epoch": 91.59, "grad_norm": 255.01686096191406, "learning_rate": 3.2974458205772197e-06, "loss": 0.3798, "step": 91590 }, { "epoch": 91.62, "grad_norm": 61.13650131225586, "learning_rate": 3.294492303906188e-06, "loss": 0.3267, "step": 91620 }, { "epoch": 91.65, "grad_norm": 98.7277603149414, "learning_rate": 3.2915394605425836e-06, "loss": 0.3103, "step": 91650 }, { "epoch": 91.68, "grad_norm": 76.50540161132812, "learning_rate": 3.2885872916521445e-06, "loss": 0.3055, "step": 91680 }, { "epoch": 91.71, "grad_norm": 77.27986907958984, "learning_rate": 3.2856357984003382e-06, "loss": 0.4186, "step": 91710 }, { "epoch": 91.74, "grad_norm": 72.93263244628906, "learning_rate": 3.282684981952369e-06, "loss": 0.3064, "step": 91740 }, { "epoch": 91.77, "grad_norm": 53.27059555053711, "learning_rate": 3.2797348434731725e-06, "loss": 0.4196, "step": 91770 }, { "epoch": 91.8, "grad_norm": 54.93994140625, "learning_rate": 3.2767853841274154e-06, "loss": 0.2854, "step": 91800 }, { "epoch": 91.83, "grad_norm": 119.52019500732422, "learning_rate": 3.273836605079499e-06, "loss": 0.2906, "step": 91830 }, { "epoch": 91.86, "grad_norm": 75.00179290771484, "learning_rate": 3.2708885074935515e-06, "loss": 0.3099, "step": 91860 }, { "epoch": 91.89, "grad_norm": 72.6570053100586, "learning_rate": 3.2679410925334394e-06, "loss": 0.3196, "step": 91890 }, { "epoch": 91.92, "grad_norm": 63.013675689697266, "learning_rate": 3.264994361362753e-06, "loss": 0.2664, "step": 91920 }, { "epoch": 91.95, "grad_norm": 45.469181060791016, "learning_rate": 3.262048315144816e-06, "loss": 0.316, "step": 91950 }, { "epoch": 91.98, "grad_norm": 40.8947868347168, "learning_rate": 3.25910295504268e-06, "loss": 0.2721, "step": 91980 }, { "epoch": 92.0, "eval_loss": 0.2699207365512848, "eval_map": 0.8123, "eval_map_50": 0.961, "eval_map_75": 0.9226, "eval_map_chicken": 0.8041, "eval_map_duck": 0.7592, "eval_map_large": 0.8225, "eval_map_medium": 0.8072, "eval_map_plant": 0.8736, "eval_map_small": 0.2827, "eval_mar_1": 0.3289, "eval_mar_10": 0.8461, "eval_mar_100": 0.8492, "eval_mar_100_chicken": 0.8467, "eval_mar_100_duck": 0.8, "eval_mar_100_plant": 0.9009, "eval_mar_large": 0.86, "eval_mar_medium": 0.848, "eval_mar_small": 0.3659, "eval_runtime": 12.7181, "eval_samples_per_second": 15.726, "eval_steps_per_second": 1.966, "step": 92000 }, { "epoch": 92.01, "grad_norm": 61.37200927734375, "learning_rate": 3.2561582822191273e-06, "loss": 0.3481, "step": 92010 }, { "epoch": 92.04, "grad_norm": 58.87767028808594, "learning_rate": 3.2532142978366654e-06, "loss": 0.3013, "step": 92040 }, { "epoch": 92.07, "grad_norm": 57.568016052246094, "learning_rate": 3.250271003057537e-06, "loss": 0.4476, "step": 92070 }, { "epoch": 92.1, "grad_norm": 24.546072006225586, "learning_rate": 3.247328399043706e-06, "loss": 0.3295, "step": 92100 }, { "epoch": 92.13, "grad_norm": 55.678497314453125, "learning_rate": 3.2443864869568666e-06, "loss": 0.2976, "step": 92130 }, { "epoch": 92.16, "grad_norm": 82.04127502441406, "learning_rate": 3.241445267958438e-06, "loss": 0.3107, "step": 92160 }, { "epoch": 92.19, "grad_norm": 55.92733383178711, "learning_rate": 3.2385047432095656e-06, "loss": 0.3849, "step": 92190 }, { "epoch": 92.22, "grad_norm": 115.82174682617188, "learning_rate": 3.235564913871126e-06, "loss": 0.3227, "step": 92220 }, { "epoch": 92.25, "grad_norm": 40.58470153808594, "learning_rate": 3.2326257811037154e-06, "loss": 0.3944, "step": 92250 }, { "epoch": 92.28, "grad_norm": 37.417667388916016, "learning_rate": 3.2296873460676557e-06, "loss": 0.3285, "step": 92280 }, { "epoch": 92.31, "grad_norm": 83.07000732421875, "learning_rate": 3.226749609922997e-06, "loss": 0.3063, "step": 92310 }, { "epoch": 92.34, "grad_norm": 898.8633422851562, "learning_rate": 3.2238125738295063e-06, "loss": 0.4254, "step": 92340 }, { "epoch": 92.37, "grad_norm": 72.9728775024414, "learning_rate": 3.220876238946684e-06, "loss": 0.2707, "step": 92370 }, { "epoch": 92.4, "grad_norm": 96.72753143310547, "learning_rate": 3.217940606433747e-06, "loss": 0.3307, "step": 92400 }, { "epoch": 92.43, "grad_norm": 63.87384033203125, "learning_rate": 3.2150056774496363e-06, "loss": 0.3418, "step": 92430 }, { "epoch": 92.46, "grad_norm": 64.26819610595703, "learning_rate": 3.212071453153015e-06, "loss": 0.382, "step": 92460 }, { "epoch": 92.49, "grad_norm": 43.40830993652344, "learning_rate": 3.209137934702267e-06, "loss": 0.3428, "step": 92490 }, { "epoch": 92.52, "grad_norm": 77.27102661132812, "learning_rate": 3.2062051232555024e-06, "loss": 0.4198, "step": 92520 }, { "epoch": 92.55, "grad_norm": 42.003028869628906, "learning_rate": 3.2032730199705477e-06, "loss": 0.4347, "step": 92550 }, { "epoch": 92.58, "grad_norm": 98.9157485961914, "learning_rate": 3.2003416260049493e-06, "loss": 0.2813, "step": 92580 }, { "epoch": 92.61, "grad_norm": 46.460384368896484, "learning_rate": 3.1974109425159754e-06, "loss": 0.3337, "step": 92610 }, { "epoch": 92.64, "grad_norm": 62.594608306884766, "learning_rate": 3.1944809706606123e-06, "loss": 0.3992, "step": 92640 }, { "epoch": 92.67, "grad_norm": 33.41252899169922, "learning_rate": 3.1915517115955704e-06, "loss": 0.3938, "step": 92670 }, { "epoch": 92.7, "grad_norm": 0.0006249400321394205, "learning_rate": 3.188623166477272e-06, "loss": 0.2687, "step": 92700 }, { "epoch": 92.73, "grad_norm": 76.48248291015625, "learning_rate": 3.185695336461861e-06, "loss": 0.2832, "step": 92730 }, { "epoch": 92.76, "grad_norm": 6.797236710553989e-05, "learning_rate": 3.182768222705198e-06, "loss": 0.3388, "step": 92760 }, { "epoch": 92.79, "grad_norm": 157.74392700195312, "learning_rate": 3.1798418263628595e-06, "loss": 0.2933, "step": 92790 }, { "epoch": 92.82, "grad_norm": 88.5262451171875, "learning_rate": 3.1769161485901445e-06, "loss": 0.3502, "step": 92820 }, { "epoch": 92.85, "grad_norm": 41.884063720703125, "learning_rate": 3.1739911905420617e-06, "loss": 0.29, "step": 92850 }, { "epoch": 92.88, "grad_norm": 91.06442260742188, "learning_rate": 3.171066953373338e-06, "loss": 0.2766, "step": 92880 }, { "epoch": 92.91, "grad_norm": 88.70669555664062, "learning_rate": 3.168143438238417e-06, "loss": 0.2588, "step": 92910 }, { "epoch": 92.94, "grad_norm": 37.735050201416016, "learning_rate": 3.1652206462914542e-06, "loss": 0.2605, "step": 92940 }, { "epoch": 92.97, "grad_norm": 45.69199752807617, "learning_rate": 3.1622985786863236e-06, "loss": 0.3247, "step": 92970 }, { "epoch": 93.0, "grad_norm": 79.71293640136719, "learning_rate": 3.1593772365766107e-06, "loss": 0.2704, "step": 93000 }, { "epoch": 93.0, "eval_loss": 0.25306951999664307, "eval_map": 0.8251, "eval_map_50": 0.9587, "eval_map_75": 0.9243, "eval_map_chicken": 0.8164, "eval_map_duck": 0.7768, "eval_map_large": 0.8461, "eval_map_medium": 0.8192, "eval_map_plant": 0.882, "eval_map_small": 0.3119, "eval_mar_1": 0.3351, "eval_mar_10": 0.8579, "eval_mar_100": 0.8611, "eval_mar_100_chicken": 0.864, "eval_mar_100_duck": 0.8113, "eval_mar_100_plant": 0.908, "eval_mar_large": 0.8786, "eval_mar_medium": 0.8575, "eval_mar_small": 0.3758, "eval_runtime": 12.641, "eval_samples_per_second": 15.821, "eval_steps_per_second": 1.978, "step": 93000 }, { "epoch": 93.03, "grad_norm": 16.544862747192383, "learning_rate": 3.156456621115615e-06, "loss": 0.4152, "step": 93030 }, { "epoch": 93.06, "grad_norm": 0.0003417735861148685, "learning_rate": 3.1535367334563493e-06, "loss": 0.329, "step": 93060 }, { "epoch": 93.09, "grad_norm": 51.385284423828125, "learning_rate": 3.1506175747515384e-06, "loss": 0.3272, "step": 93090 }, { "epoch": 93.12, "grad_norm": 56.81093215942383, "learning_rate": 3.147699146153621e-06, "loss": 0.4187, "step": 93120 }, { "epoch": 93.15, "grad_norm": 53.25189971923828, "learning_rate": 3.144781448814746e-06, "loss": 0.3608, "step": 93150 }, { "epoch": 93.18, "grad_norm": 76.44778442382812, "learning_rate": 3.141864483886774e-06, "loss": 0.3416, "step": 93180 }, { "epoch": 93.21, "grad_norm": 79.82946014404297, "learning_rate": 3.1389482525212753e-06, "loss": 0.3872, "step": 93210 }, { "epoch": 93.24, "grad_norm": 127.37114715576172, "learning_rate": 3.1360327558695336e-06, "loss": 0.3833, "step": 93240 }, { "epoch": 93.27, "grad_norm": 52.35741424560547, "learning_rate": 3.1331179950825415e-06, "loss": 0.3992, "step": 93270 }, { "epoch": 93.3, "grad_norm": 111.60072326660156, "learning_rate": 3.130203971310999e-06, "loss": 0.3202, "step": 93300 }, { "epoch": 93.33, "grad_norm": 58.31892395019531, "learning_rate": 3.1272906857053164e-06, "loss": 0.4026, "step": 93330 }, { "epoch": 93.36, "grad_norm": 27.048818588256836, "learning_rate": 3.1243781394156138e-06, "loss": 0.4084, "step": 93360 }, { "epoch": 93.39, "grad_norm": 74.36424255371094, "learning_rate": 3.121466333591715e-06, "loss": 0.3722, "step": 93390 }, { "epoch": 93.42, "grad_norm": 65.34441375732422, "learning_rate": 3.1185552693831595e-06, "loss": 0.3215, "step": 93420 }, { "epoch": 93.45, "grad_norm": 1089.331787109375, "learning_rate": 3.1156449479391876e-06, "loss": 0.3657, "step": 93450 }, { "epoch": 93.48, "grad_norm": 63.42023468017578, "learning_rate": 3.1127353704087477e-06, "loss": 0.3505, "step": 93480 }, { "epoch": 93.51, "grad_norm": 195.43772888183594, "learning_rate": 3.1098265379404953e-06, "loss": 0.3704, "step": 93510 }, { "epoch": 93.54, "grad_norm": 28.342546463012695, "learning_rate": 3.106918451682789e-06, "loss": 0.38, "step": 93540 }, { "epoch": 93.57, "grad_norm": 411.14837646484375, "learning_rate": 3.1040111127836994e-06, "loss": 0.3031, "step": 93570 }, { "epoch": 93.6, "grad_norm": 24.98753547668457, "learning_rate": 3.1011045223909954e-06, "loss": 0.3878, "step": 93600 }, { "epoch": 93.63, "grad_norm": 49.6890869140625, "learning_rate": 3.0981986816521536e-06, "loss": 0.3282, "step": 93630 }, { "epoch": 93.66, "grad_norm": 107.6459732055664, "learning_rate": 3.0952935917143533e-06, "loss": 0.2803, "step": 93660 }, { "epoch": 93.69, "grad_norm": 53.84072494506836, "learning_rate": 3.092389253724476e-06, "loss": 0.3601, "step": 93690 }, { "epoch": 93.72, "grad_norm": 68.09239196777344, "learning_rate": 3.089485668829113e-06, "loss": 0.3314, "step": 93720 }, { "epoch": 93.75, "grad_norm": 65.32845306396484, "learning_rate": 3.0865828381745515e-06, "loss": 0.2704, "step": 93750 }, { "epoch": 93.78, "grad_norm": 37.913272857666016, "learning_rate": 3.0836807629067828e-06, "loss": 0.2727, "step": 93780 }, { "epoch": 93.81, "grad_norm": 100.89753723144531, "learning_rate": 3.0807794441715e-06, "loss": 0.2616, "step": 93810 }, { "epoch": 93.84, "grad_norm": 6.159136682981625e-05, "learning_rate": 3.077878883114096e-06, "loss": 0.3216, "step": 93840 }, { "epoch": 93.87, "grad_norm": 42.20660400390625, "learning_rate": 3.074979080879671e-06, "loss": 0.3235, "step": 93870 }, { "epoch": 93.9, "grad_norm": 37.488800048828125, "learning_rate": 3.0720800386130176e-06, "loss": 0.3487, "step": 93900 }, { "epoch": 93.93, "grad_norm": 69.13880920410156, "learning_rate": 3.069181757458633e-06, "loss": 0.2578, "step": 93930 }, { "epoch": 93.96, "grad_norm": 289.8492736816406, "learning_rate": 3.066284238560713e-06, "loss": 0.3331, "step": 93960 }, { "epoch": 93.99, "grad_norm": 126.32936096191406, "learning_rate": 3.063387483063148e-06, "loss": 0.3274, "step": 93990 }, { "epoch": 94.0, "eval_loss": 0.2599104344844818, "eval_map": 0.8229, "eval_map_50": 0.968, "eval_map_75": 0.9292, "eval_map_chicken": 0.8147, "eval_map_duck": 0.7744, "eval_map_large": 0.8409, "eval_map_medium": 0.8176, "eval_map_plant": 0.8797, "eval_map_small": 0.3198, "eval_mar_1": 0.3306, "eval_mar_10": 0.8551, "eval_mar_100": 0.8583, "eval_mar_100_chicken": 0.8602, "eval_mar_100_duck": 0.8082, "eval_mar_100_plant": 0.9066, "eval_mar_large": 0.8716, "eval_mar_medium": 0.8564, "eval_mar_small": 0.4022, "eval_runtime": 13.0936, "eval_samples_per_second": 15.275, "eval_steps_per_second": 1.909, "step": 94000 }, { "epoch": 94.02, "grad_norm": 94.75921630859375, "learning_rate": 3.0604914921095373e-06, "loss": 0.2777, "step": 94020 }, { "epoch": 94.05, "grad_norm": 67.7879867553711, "learning_rate": 3.0575962668431704e-06, "loss": 0.3008, "step": 94050 }, { "epoch": 94.08, "grad_norm": 0.00014549352636095136, "learning_rate": 3.0547018084070344e-06, "loss": 0.3421, "step": 94080 }, { "epoch": 94.11, "grad_norm": 62.49932861328125, "learning_rate": 3.0518081179438173e-06, "loss": 0.316, "step": 94110 }, { "epoch": 94.14, "grad_norm": 70.30824279785156, "learning_rate": 3.0489151965958998e-06, "loss": 0.3504, "step": 94140 }, { "epoch": 94.17, "grad_norm": 47.12671661376953, "learning_rate": 3.046023045505366e-06, "loss": 0.3037, "step": 94170 }, { "epoch": 94.2, "grad_norm": 39.03264236450195, "learning_rate": 3.043131665813988e-06, "loss": 0.3498, "step": 94200 }, { "epoch": 94.23, "grad_norm": 82.71261596679688, "learning_rate": 3.040241058663238e-06, "loss": 0.2708, "step": 94230 }, { "epoch": 94.26, "grad_norm": 56.22880554199219, "learning_rate": 3.0373512251942817e-06, "loss": 0.4234, "step": 94260 }, { "epoch": 94.29, "grad_norm": 92.04325866699219, "learning_rate": 3.034462166547978e-06, "loss": 0.2823, "step": 94290 }, { "epoch": 94.32, "grad_norm": 66.09696197509766, "learning_rate": 3.031573883864882e-06, "loss": 0.3334, "step": 94320 }, { "epoch": 94.35, "grad_norm": 41.22340393066406, "learning_rate": 3.028686378285245e-06, "loss": 0.3521, "step": 94350 }, { "epoch": 94.38, "grad_norm": 48.8659782409668, "learning_rate": 3.025799650949006e-06, "loss": 0.3026, "step": 94380 }, { "epoch": 94.41, "grad_norm": 38.40583038330078, "learning_rate": 3.0229137029957993e-06, "loss": 0.3035, "step": 94410 }, { "epoch": 94.44, "grad_norm": 81.63008880615234, "learning_rate": 3.0200285355649504e-06, "loss": 0.3157, "step": 94440 }, { "epoch": 94.47, "grad_norm": 63.38157653808594, "learning_rate": 3.0171441497954807e-06, "loss": 0.298, "step": 94470 }, { "epoch": 94.5, "grad_norm": 3.025247497134842e-05, "learning_rate": 3.0142605468260976e-06, "loss": 0.3207, "step": 94500 }, { "epoch": 94.53, "grad_norm": 93.58333587646484, "learning_rate": 3.011377727795202e-06, "loss": 0.2754, "step": 94530 }, { "epoch": 94.56, "grad_norm": 54.05022048950195, "learning_rate": 3.0084956938408873e-06, "loss": 0.3954, "step": 94560 }, { "epoch": 94.59, "grad_norm": 122.21818542480469, "learning_rate": 3.0056144461009313e-06, "loss": 0.3387, "step": 94590 }, { "epoch": 94.62, "grad_norm": 51.41754150390625, "learning_rate": 3.002733985712808e-06, "loss": 0.3429, "step": 94620 }, { "epoch": 94.65, "grad_norm": 83.7321548461914, "learning_rate": 2.9998543138136773e-06, "loss": 0.3391, "step": 94650 }, { "epoch": 94.68, "grad_norm": 69.21733093261719, "learning_rate": 2.9969754315403865e-06, "loss": 0.3298, "step": 94680 }, { "epoch": 94.71, "grad_norm": 49.32567596435547, "learning_rate": 2.994097340029474e-06, "loss": 0.3575, "step": 94710 }, { "epoch": 94.74, "grad_norm": 72.66204071044922, "learning_rate": 2.991220040417162e-06, "loss": 0.2803, "step": 94740 }, { "epoch": 94.77, "grad_norm": 173.75938415527344, "learning_rate": 2.9883435338393674e-06, "loss": 0.3244, "step": 94770 }, { "epoch": 94.8, "grad_norm": 81.99342346191406, "learning_rate": 2.9854678214316875e-06, "loss": 0.3305, "step": 94800 }, { "epoch": 94.83, "grad_norm": 49.36806106567383, "learning_rate": 2.982592904329407e-06, "loss": 0.3494, "step": 94830 }, { "epoch": 94.86, "grad_norm": 96.76710510253906, "learning_rate": 2.979718783667499e-06, "loss": 0.3967, "step": 94860 }, { "epoch": 94.89, "grad_norm": 55.12672424316406, "learning_rate": 2.9768454605806176e-06, "loss": 0.2927, "step": 94890 }, { "epoch": 94.92, "grad_norm": 139.1917266845703, "learning_rate": 2.97397293620311e-06, "loss": 0.3827, "step": 94920 }, { "epoch": 94.95, "grad_norm": 29.917402267456055, "learning_rate": 2.9711012116690007e-06, "loss": 0.289, "step": 94950 }, { "epoch": 94.98, "grad_norm": 46.63770294189453, "learning_rate": 2.968230288112002e-06, "loss": 0.3198, "step": 94980 }, { "epoch": 95.0, "eval_loss": 0.2560811936855316, "eval_map": 0.8264, "eval_map_50": 0.9661, "eval_map_75": 0.9288, "eval_map_chicken": 0.8153, "eval_map_duck": 0.7869, "eval_map_large": 0.8412, "eval_map_medium": 0.8232, "eval_map_plant": 0.8768, "eval_map_small": 0.303, "eval_mar_1": 0.3327, "eval_mar_10": 0.859, "eval_mar_100": 0.8625, "eval_mar_100_chicken": 0.8628, "eval_mar_100_duck": 0.8191, "eval_mar_100_plant": 0.9057, "eval_mar_large": 0.8739, "eval_mar_medium": 0.8619, "eval_mar_small": 0.3882, "eval_runtime": 12.7034, "eval_samples_per_second": 15.744, "eval_steps_per_second": 1.968, "step": 95000 }, { "epoch": 95.01, "grad_norm": 45.346439361572266, "learning_rate": 2.965360166665508e-06, "loss": 0.3301, "step": 95010 }, { "epoch": 95.04, "grad_norm": 50.0007438659668, "learning_rate": 2.962490848462596e-06, "loss": 0.4534, "step": 95040 }, { "epoch": 95.07, "grad_norm": 43.47236251831055, "learning_rate": 2.959622334636031e-06, "loss": 0.2897, "step": 95070 }, { "epoch": 95.1, "grad_norm": 38.356231689453125, "learning_rate": 2.9567546263182554e-06, "loss": 0.3391, "step": 95100 }, { "epoch": 95.13, "grad_norm": 49.38764572143555, "learning_rate": 2.9538877246413943e-06, "loss": 0.3202, "step": 95130 }, { "epoch": 95.16, "grad_norm": 35.81418991088867, "learning_rate": 2.951021630737255e-06, "loss": 0.3689, "step": 95160 }, { "epoch": 95.19, "grad_norm": 67.5494613647461, "learning_rate": 2.9481563457373247e-06, "loss": 0.3969, "step": 95190 }, { "epoch": 95.22, "grad_norm": 97.90336608886719, "learning_rate": 2.945291870772776e-06, "loss": 0.2766, "step": 95220 }, { "epoch": 95.25, "grad_norm": 49.82147979736328, "learning_rate": 2.9424282069744564e-06, "loss": 0.3972, "step": 95250 }, { "epoch": 95.28, "grad_norm": 0.09147245436906815, "learning_rate": 2.9395653554728955e-06, "loss": 0.3158, "step": 95280 }, { "epoch": 95.31, "grad_norm": 44.0391731262207, "learning_rate": 2.9367033173983006e-06, "loss": 0.3292, "step": 95310 }, { "epoch": 95.34, "grad_norm": 107.96846771240234, "learning_rate": 2.933842093880558e-06, "loss": 0.3448, "step": 95340 }, { "epoch": 95.37, "grad_norm": 122.39090728759766, "learning_rate": 2.930981686049237e-06, "loss": 0.3559, "step": 95370 }, { "epoch": 95.4, "grad_norm": 102.42393493652344, "learning_rate": 2.92812209503358e-06, "loss": 0.2796, "step": 95400 }, { "epoch": 95.43, "grad_norm": 53.30547332763672, "learning_rate": 2.9252633219625073e-06, "loss": 0.2969, "step": 95430 }, { "epoch": 95.46, "grad_norm": 67.60681915283203, "learning_rate": 2.922405367964617e-06, "loss": 0.3117, "step": 95460 }, { "epoch": 95.49, "grad_norm": 78.22626495361328, "learning_rate": 2.919548234168183e-06, "loss": 0.3304, "step": 95490 }, { "epoch": 95.52, "grad_norm": 44.90492248535156, "learning_rate": 2.91669192170116e-06, "loss": 0.3349, "step": 95520 }, { "epoch": 95.55, "grad_norm": 54.41095733642578, "learning_rate": 2.9138364316911747e-06, "loss": 0.4368, "step": 95550 }, { "epoch": 95.58, "grad_norm": 76.27407836914062, "learning_rate": 2.9109817652655253e-06, "loss": 0.3964, "step": 95580 }, { "epoch": 95.61, "grad_norm": 129.2054901123047, "learning_rate": 2.908127923551194e-06, "loss": 0.3165, "step": 95610 }, { "epoch": 95.64, "grad_norm": 50.66813659667969, "learning_rate": 2.9052749076748266e-06, "loss": 0.3057, "step": 95640 }, { "epoch": 95.67, "grad_norm": 124.43778228759766, "learning_rate": 2.902422718762752e-06, "loss": 0.4234, "step": 95670 }, { "epoch": 95.7, "grad_norm": 48.916481018066406, "learning_rate": 2.899571357940969e-06, "loss": 0.3301, "step": 95700 }, { "epoch": 95.73, "grad_norm": 65.13150024414062, "learning_rate": 2.896720826335151e-06, "loss": 0.2859, "step": 95730 }, { "epoch": 95.76, "grad_norm": 41.94382095336914, "learning_rate": 2.8938711250706397e-06, "loss": 0.2765, "step": 95760 }, { "epoch": 95.79, "grad_norm": 46.28852844238281, "learning_rate": 2.8910222552724552e-06, "loss": 0.3321, "step": 95790 }, { "epoch": 95.82, "grad_norm": 40.582393646240234, "learning_rate": 2.8881742180652813e-06, "loss": 0.2965, "step": 95820 }, { "epoch": 95.85, "grad_norm": 79.8348388671875, "learning_rate": 2.8853270145734846e-06, "loss": 0.3348, "step": 95850 }, { "epoch": 95.88, "grad_norm": 38.64009475708008, "learning_rate": 2.8824806459210907e-06, "loss": 0.2745, "step": 95880 }, { "epoch": 95.91, "grad_norm": 75.61990356445312, "learning_rate": 2.8796351132318047e-06, "loss": 0.3027, "step": 95910 }, { "epoch": 95.94, "grad_norm": 29.466428756713867, "learning_rate": 2.876790417628994e-06, "loss": 0.3292, "step": 95940 }, { "epoch": 95.97, "grad_norm": 62.237430572509766, "learning_rate": 2.8739465602357014e-06, "loss": 0.2929, "step": 95970 }, { "epoch": 96.0, "grad_norm": 45.7805290222168, "learning_rate": 2.871103542174637e-06, "loss": 0.3286, "step": 96000 }, { "epoch": 96.0, "eval_loss": 0.2624129056930542, "eval_map": 0.8178, "eval_map_50": 0.9671, "eval_map_75": 0.9258, "eval_map_chicken": 0.817, "eval_map_duck": 0.7705, "eval_map_large": 0.8334, "eval_map_medium": 0.8123, "eval_map_plant": 0.8658, "eval_map_small": 0.3039, "eval_mar_1": 0.3303, "eval_mar_10": 0.8535, "eval_mar_100": 0.8567, "eval_mar_100_chicken": 0.8624, "eval_mar_100_duck": 0.8082, "eval_mar_100_plant": 0.8996, "eval_mar_large": 0.87, "eval_mar_medium": 0.8552, "eval_mar_small": 0.3876, "eval_runtime": 12.7277, "eval_samples_per_second": 15.714, "eval_steps_per_second": 1.964, "step": 96000 }, { "epoch": 96.03, "grad_norm": 128.1754913330078, "learning_rate": 2.86826136456818e-06, "loss": 0.3577, "step": 96030 }, { "epoch": 96.06, "grad_norm": 109.0239486694336, "learning_rate": 2.865420028538375e-06, "loss": 0.3622, "step": 96060 }, { "epoch": 96.09, "grad_norm": 85.77645874023438, "learning_rate": 2.8625795352069385e-06, "loss": 0.4461, "step": 96090 }, { "epoch": 96.12, "grad_norm": 78.34368896484375, "learning_rate": 2.8597398856952473e-06, "loss": 0.3482, "step": 96120 }, { "epoch": 96.15, "grad_norm": 54.10435104370117, "learning_rate": 2.856901081124359e-06, "loss": 0.3319, "step": 96150 }, { "epoch": 96.18, "grad_norm": 62.302024841308594, "learning_rate": 2.8540631226149813e-06, "loss": 0.3905, "step": 96180 }, { "epoch": 96.21, "grad_norm": 53.964996337890625, "learning_rate": 2.8512260112874994e-06, "loss": 0.2538, "step": 96210 }, { "epoch": 96.24, "grad_norm": 31.712064743041992, "learning_rate": 2.8483897482619566e-06, "loss": 0.2928, "step": 96240 }, { "epoch": 96.27, "grad_norm": 49.39045715332031, "learning_rate": 2.845554334658066e-06, "loss": 0.3622, "step": 96270 }, { "epoch": 96.3, "grad_norm": 4.637540769181214e-05, "learning_rate": 2.8427197715952047e-06, "loss": 0.2819, "step": 96300 }, { "epoch": 96.33, "grad_norm": 65.57496643066406, "learning_rate": 2.8398860601924145e-06, "loss": 0.2278, "step": 96330 }, { "epoch": 96.36, "grad_norm": 56.46002197265625, "learning_rate": 2.837053201568396e-06, "loss": 0.2873, "step": 96360 }, { "epoch": 96.39, "grad_norm": 51.32892990112305, "learning_rate": 2.834221196841521e-06, "loss": 0.3509, "step": 96390 }, { "epoch": 96.42, "grad_norm": 31.760278701782227, "learning_rate": 2.831390047129815e-06, "loss": 0.3349, "step": 96420 }, { "epoch": 96.45, "grad_norm": 123.59374237060547, "learning_rate": 2.8285597535509775e-06, "loss": 0.3812, "step": 96450 }, { "epoch": 96.48, "grad_norm": 62.53190612792969, "learning_rate": 2.825730317222358e-06, "loss": 0.3742, "step": 96480 }, { "epoch": 96.51, "grad_norm": 52.945831298828125, "learning_rate": 2.8229017392609782e-06, "loss": 0.3029, "step": 96510 }, { "epoch": 96.54, "grad_norm": 57.25663757324219, "learning_rate": 2.820074020783511e-06, "loss": 0.3157, "step": 96540 }, { "epoch": 96.57, "grad_norm": 93.4737319946289, "learning_rate": 2.817247162906297e-06, "loss": 0.3198, "step": 96570 }, { "epoch": 96.6, "grad_norm": 38.074241638183594, "learning_rate": 2.814421166745337e-06, "loss": 0.4224, "step": 96600 }, { "epoch": 96.63, "grad_norm": 60.76255416870117, "learning_rate": 2.811596033416285e-06, "loss": 0.3727, "step": 96630 }, { "epoch": 96.66, "grad_norm": 73.3246078491211, "learning_rate": 2.808771764034462e-06, "loss": 0.2725, "step": 96660 }, { "epoch": 96.69, "grad_norm": 84.76152801513672, "learning_rate": 2.8059483597148457e-06, "loss": 0.3186, "step": 96690 }, { "epoch": 96.72, "grad_norm": 33.37944030761719, "learning_rate": 2.803125821572068e-06, "loss": 0.2871, "step": 96720 }, { "epoch": 96.75, "grad_norm": 39.63099670410156, "learning_rate": 2.800304150720424e-06, "loss": 0.3482, "step": 96750 }, { "epoch": 96.78, "grad_norm": 28.743057250976562, "learning_rate": 2.7974833482738674e-06, "loss": 0.389, "step": 96780 }, { "epoch": 96.81, "grad_norm": 64.76252746582031, "learning_rate": 2.7946634153460016e-06, "loss": 0.2837, "step": 96810 }, { "epoch": 96.84, "grad_norm": 63.94390869140625, "learning_rate": 2.791844353050094e-06, "loss": 0.3926, "step": 96840 }, { "epoch": 96.87, "grad_norm": 36.9520263671875, "learning_rate": 2.7890261624990643e-06, "loss": 0.2898, "step": 96870 }, { "epoch": 96.9, "grad_norm": 96.72816467285156, "learning_rate": 2.7862088448054936e-06, "loss": 0.2874, "step": 96900 }, { "epoch": 96.93, "grad_norm": 103.20718383789062, "learning_rate": 2.783392401081609e-06, "loss": 0.3841, "step": 96930 }, { "epoch": 96.96, "grad_norm": 113.7406234741211, "learning_rate": 2.7805768324393017e-06, "loss": 0.33, "step": 96960 }, { "epoch": 96.99, "grad_norm": 39.94602584838867, "learning_rate": 2.77776213999011e-06, "loss": 0.35, "step": 96990 }, { "epoch": 97.0, "eval_loss": 0.24359101057052612, "eval_map": 0.8304, "eval_map_50": 0.9659, "eval_map_75": 0.9291, "eval_map_chicken": 0.8168, "eval_map_duck": 0.7886, "eval_map_large": 0.8375, "eval_map_medium": 0.8274, "eval_map_plant": 0.8857, "eval_map_small": 0.3281, "eval_mar_1": 0.3336, "eval_mar_10": 0.866, "eval_mar_100": 0.8697, "eval_mar_100_chicken": 0.8644, "eval_mar_100_duck": 0.8299, "eval_mar_100_plant": 0.9147, "eval_mar_large": 0.8707, "eval_mar_medium": 0.8701, "eval_mar_small": 0.4291, "eval_runtime": 13.1728, "eval_samples_per_second": 15.183, "eval_steps_per_second": 1.898, "step": 97000 }, { "epoch": 97.02, "grad_norm": 50.85993576049805, "learning_rate": 2.7749483248452324e-06, "loss": 0.3065, "step": 97020 }, { "epoch": 97.05, "grad_norm": 52.248836517333984, "learning_rate": 2.772135388115519e-06, "loss": 0.327, "step": 97050 }, { "epoch": 97.08, "grad_norm": 53.525875091552734, "learning_rate": 2.769323330911472e-06, "loss": 0.3065, "step": 97080 }, { "epoch": 97.11, "grad_norm": 23.823579788208008, "learning_rate": 2.766512154343246e-06, "loss": 0.2949, "step": 97110 }, { "epoch": 97.14, "grad_norm": 33.226966857910156, "learning_rate": 2.763701859520652e-06, "loss": 0.288, "step": 97140 }, { "epoch": 97.17, "grad_norm": 86.92098236083984, "learning_rate": 2.760892447553143e-06, "loss": 0.3291, "step": 97170 }, { "epoch": 97.2, "grad_norm": 83.12812042236328, "learning_rate": 2.7580839195498397e-06, "loss": 0.3141, "step": 97200 }, { "epoch": 97.23, "grad_norm": 92.2624740600586, "learning_rate": 2.7552762766194975e-06, "loss": 0.3157, "step": 97230 }, { "epoch": 97.26, "grad_norm": 50.790191650390625, "learning_rate": 2.752469519870534e-06, "loss": 0.3409, "step": 97260 }, { "epoch": 97.29, "grad_norm": 75.48125457763672, "learning_rate": 2.7496636504110077e-06, "loss": 0.3807, "step": 97290 }, { "epoch": 97.32, "grad_norm": 41.778255462646484, "learning_rate": 2.746858669348634e-06, "loss": 0.3387, "step": 97320 }, { "epoch": 97.35, "grad_norm": 63.31541442871094, "learning_rate": 2.7440545777907747e-06, "loss": 0.3761, "step": 97350 }, { "epoch": 97.38, "grad_norm": 32.030731201171875, "learning_rate": 2.741251376844443e-06, "loss": 0.2966, "step": 97380 }, { "epoch": 97.41, "grad_norm": 101.18850708007812, "learning_rate": 2.7384490676162932e-06, "loss": 0.3148, "step": 97410 }, { "epoch": 97.44, "grad_norm": 66.32122039794922, "learning_rate": 2.7356476512126386e-06, "loss": 0.3452, "step": 97440 }, { "epoch": 97.47, "grad_norm": 72.26156616210938, "learning_rate": 2.7328471287394265e-06, "loss": 0.3617, "step": 97470 }, { "epoch": 97.5, "grad_norm": 33.9433708190918, "learning_rate": 2.7300475013022666e-06, "loss": 0.2853, "step": 97500 }, { "epoch": 97.53, "grad_norm": 0.0005007755244150758, "learning_rate": 2.7272487700064027e-06, "loss": 0.3028, "step": 97530 }, { "epoch": 97.56, "grad_norm": 84.46577453613281, "learning_rate": 2.724450935956733e-06, "loss": 0.339, "step": 97560 }, { "epoch": 97.59, "grad_norm": 74.83903503417969, "learning_rate": 2.7216540002577933e-06, "loss": 0.3913, "step": 97590 }, { "epoch": 97.62, "grad_norm": 33.90381622314453, "learning_rate": 2.718857964013773e-06, "loss": 0.3353, "step": 97620 }, { "epoch": 97.65, "grad_norm": 35.307613372802734, "learning_rate": 2.716062828328502e-06, "loss": 0.4098, "step": 97650 }, { "epoch": 97.68, "grad_norm": 68.27301788330078, "learning_rate": 2.713268594305458e-06, "loss": 0.3117, "step": 97680 }, { "epoch": 97.71, "grad_norm": 43.770877838134766, "learning_rate": 2.710475263047756e-06, "loss": 0.3919, "step": 97710 }, { "epoch": 97.74, "grad_norm": 42.806400299072266, "learning_rate": 2.707682835658163e-06, "loss": 0.2731, "step": 97740 }, { "epoch": 97.77, "grad_norm": 55.939788818359375, "learning_rate": 2.7048913132390787e-06, "loss": 0.3612, "step": 97770 }, { "epoch": 97.8, "grad_norm": 65.48059844970703, "learning_rate": 2.7021006968925613e-06, "loss": 0.2845, "step": 97800 }, { "epoch": 97.83, "grad_norm": 100.27794647216797, "learning_rate": 2.6993109877202942e-06, "loss": 0.3663, "step": 97830 }, { "epoch": 97.86, "grad_norm": 0.00013208930613473058, "learning_rate": 2.6965221868236156e-06, "loss": 0.3032, "step": 97860 }, { "epoch": 97.89, "grad_norm": 5.762829096056521e-05, "learning_rate": 2.6937342953034963e-06, "loss": 0.3252, "step": 97890 }, { "epoch": 97.92, "grad_norm": 55.94629669189453, "learning_rate": 2.6909473142605522e-06, "loss": 0.3066, "step": 97920 }, { "epoch": 97.95, "grad_norm": 64.99919128417969, "learning_rate": 2.6881612447950425e-06, "loss": 0.3342, "step": 97950 }, { "epoch": 97.98, "grad_norm": 74.64478302001953, "learning_rate": 2.6853760880068587e-06, "loss": 0.3377, "step": 97980 }, { "epoch": 98.0, "eval_loss": 0.26262038946151733, "eval_map": 0.819, "eval_map_50": 0.9551, "eval_map_75": 0.921, "eval_map_chicken": 0.8114, "eval_map_duck": 0.7658, "eval_map_large": 0.819, "eval_map_medium": 0.8177, "eval_map_plant": 0.8798, "eval_map_small": 0.294, "eval_mar_1": 0.3325, "eval_mar_10": 0.8525, "eval_mar_100": 0.8559, "eval_mar_100_chicken": 0.8588, "eval_mar_100_duck": 0.8005, "eval_mar_100_plant": 0.9083, "eval_mar_large": 0.8551, "eval_mar_medium": 0.8563, "eval_mar_small": 0.3603, "eval_runtime": 12.5825, "eval_samples_per_second": 15.895, "eval_steps_per_second": 1.987, "step": 98000 }, { "epoch": 98.01, "grad_norm": 65.87945556640625, "learning_rate": 2.6825918449955386e-06, "loss": 0.3547, "step": 98010 }, { "epoch": 98.04, "grad_norm": 68.42288970947266, "learning_rate": 2.6798085168602595e-06, "loss": 0.2583, "step": 98040 }, { "epoch": 98.07, "grad_norm": 55.21492004394531, "learning_rate": 2.6770261046998315e-06, "loss": 0.3658, "step": 98070 }, { "epoch": 98.1, "grad_norm": 49.67182159423828, "learning_rate": 2.6742446096127086e-06, "loss": 0.3059, "step": 98100 }, { "epoch": 98.13, "grad_norm": 38.464027404785156, "learning_rate": 2.671464032696982e-06, "loss": 0.3206, "step": 98130 }, { "epoch": 98.16, "grad_norm": 35.08512878417969, "learning_rate": 2.668684375050378e-06, "loss": 0.4094, "step": 98160 }, { "epoch": 98.19, "grad_norm": 52.293609619140625, "learning_rate": 2.6659056377702606e-06, "loss": 0.4436, "step": 98190 }, { "epoch": 98.22, "grad_norm": 36.339420318603516, "learning_rate": 2.663127821953633e-06, "loss": 0.2873, "step": 98220 }, { "epoch": 98.25, "grad_norm": 62.543975830078125, "learning_rate": 2.6603509286971342e-06, "loss": 0.3461, "step": 98250 }, { "epoch": 98.28, "grad_norm": 27.11348533630371, "learning_rate": 2.6575749590970336e-06, "loss": 0.3193, "step": 98280 }, { "epoch": 98.31, "grad_norm": 34.61616897583008, "learning_rate": 2.654799914249245e-06, "loss": 0.283, "step": 98310 }, { "epoch": 98.34, "grad_norm": 41.453433990478516, "learning_rate": 2.6520257952493066e-06, "loss": 0.3736, "step": 98340 }, { "epoch": 98.37, "grad_norm": 35.38328170776367, "learning_rate": 2.6492526031924005e-06, "loss": 0.3223, "step": 98370 }, { "epoch": 98.4, "grad_norm": 47.43199920654297, "learning_rate": 2.646480339173337e-06, "loss": 0.3485, "step": 98400 }, { "epoch": 98.43, "grad_norm": 67.0950927734375, "learning_rate": 2.6437090042865655e-06, "loss": 0.3542, "step": 98430 }, { "epoch": 98.46, "grad_norm": 0.0004951642476953566, "learning_rate": 2.6409385996261606e-06, "loss": 0.4301, "step": 98460 }, { "epoch": 98.49, "grad_norm": 65.99639892578125, "learning_rate": 2.6381691262858385e-06, "loss": 0.3016, "step": 98490 }, { "epoch": 98.52, "grad_norm": 58.26409149169922, "learning_rate": 2.635400585358937e-06, "loss": 0.3173, "step": 98520 }, { "epoch": 98.55, "grad_norm": 64.86679077148438, "learning_rate": 2.6326329779384397e-06, "loss": 0.2387, "step": 98550 }, { "epoch": 98.58, "grad_norm": 44.577388763427734, "learning_rate": 2.62986630511695e-06, "loss": 0.3322, "step": 98580 }, { "epoch": 98.61, "grad_norm": 7.959528738865629e-05, "learning_rate": 2.627100567986709e-06, "loss": 0.2594, "step": 98610 }, { "epoch": 98.64, "grad_norm": 70.58832550048828, "learning_rate": 2.624335767639582e-06, "loss": 0.2606, "step": 98640 }, { "epoch": 98.67, "grad_norm": 39.48030471801758, "learning_rate": 2.6215719051670705e-06, "loss": 0.3476, "step": 98670 }, { "epoch": 98.7, "grad_norm": 47.59482955932617, "learning_rate": 2.618808981660304e-06, "loss": 0.2819, "step": 98700 }, { "epoch": 98.73, "grad_norm": 34.72926712036133, "learning_rate": 2.6160469982100426e-06, "loss": 0.2436, "step": 98730 }, { "epoch": 98.76, "grad_norm": 55.71095275878906, "learning_rate": 2.6132859559066704e-06, "loss": 0.3752, "step": 98760 }, { "epoch": 98.79, "grad_norm": 86.14274597167969, "learning_rate": 2.610525855840206e-06, "loss": 0.286, "step": 98790 }, { "epoch": 98.82, "grad_norm": 0.0002997389528900385, "learning_rate": 2.607766699100288e-06, "loss": 0.3165, "step": 98820 }, { "epoch": 98.85, "grad_norm": 172.19790649414062, "learning_rate": 2.6050084867761953e-06, "loss": 0.3315, "step": 98850 }, { "epoch": 98.88, "grad_norm": 51.42770767211914, "learning_rate": 2.6022512199568205e-06, "loss": 0.3008, "step": 98880 }, { "epoch": 98.91, "grad_norm": 38.146026611328125, "learning_rate": 2.5994948997306935e-06, "loss": 0.3183, "step": 98910 }, { "epoch": 98.94, "grad_norm": 415.650390625, "learning_rate": 2.5967395271859614e-06, "loss": 0.2789, "step": 98940 }, { "epoch": 98.97, "grad_norm": 27.245256423950195, "learning_rate": 2.5939851034104035e-06, "loss": 0.3954, "step": 98970 }, { "epoch": 99.0, "grad_norm": 124.74019622802734, "learning_rate": 2.5912316294914232e-06, "loss": 0.3617, "step": 99000 }, { "epoch": 99.0, "eval_loss": 0.26728200912475586, "eval_map": 0.8161, "eval_map_50": 0.9609, "eval_map_75": 0.9234, "eval_map_chicken": 0.8023, "eval_map_duck": 0.7725, "eval_map_large": 0.8305, "eval_map_medium": 0.8086, "eval_map_plant": 0.8736, "eval_map_small": 0.2854, "eval_mar_1": 0.3311, "eval_mar_10": 0.8506, "eval_mar_100": 0.8545, "eval_mar_100_chicken": 0.8485, "eval_mar_100_duck": 0.8134, "eval_mar_100_plant": 0.9016, "eval_mar_large": 0.87, "eval_mar_medium": 0.8498, "eval_mar_small": 0.3632, "eval_runtime": 12.5844, "eval_samples_per_second": 15.893, "eval_steps_per_second": 1.987, "step": 99000 }, { "epoch": 99.03, "grad_norm": 76.72203063964844, "learning_rate": 2.58847910651605e-06, "loss": 0.3671, "step": 99030 }, { "epoch": 99.06, "grad_norm": 95.86083984375, "learning_rate": 2.5857275355709317e-06, "loss": 0.4423, "step": 99060 }, { "epoch": 99.09, "grad_norm": 61.35476303100586, "learning_rate": 2.5829769177423504e-06, "loss": 0.3049, "step": 99090 }, { "epoch": 99.12, "grad_norm": 50.42326736450195, "learning_rate": 2.580227254116199e-06, "loss": 0.2927, "step": 99120 }, { "epoch": 99.15, "grad_norm": 47.52080535888672, "learning_rate": 2.5774785457780107e-06, "loss": 0.395, "step": 99150 }, { "epoch": 99.18, "grad_norm": 111.17036437988281, "learning_rate": 2.5747307938129245e-06, "loss": 0.4257, "step": 99180 }, { "epoch": 99.21, "grad_norm": 52.92369842529297, "learning_rate": 2.5719839993057143e-06, "loss": 0.4292, "step": 99210 }, { "epoch": 99.24, "grad_norm": 0.00012048759526805952, "learning_rate": 2.5692381633407672e-06, "loss": 0.2571, "step": 99240 }, { "epoch": 99.27, "grad_norm": 83.98875427246094, "learning_rate": 2.5664932870020966e-06, "loss": 0.2613, "step": 99270 }, { "epoch": 99.3, "grad_norm": 53.04889678955078, "learning_rate": 2.5637493713733376e-06, "loss": 0.337, "step": 99300 }, { "epoch": 99.33, "grad_norm": 71.96670532226562, "learning_rate": 2.5610064175377456e-06, "loss": 0.3027, "step": 99330 }, { "epoch": 99.36, "grad_norm": 56.4972038269043, "learning_rate": 2.558264426578192e-06, "loss": 0.3497, "step": 99360 }, { "epoch": 99.39, "grad_norm": 92.76538848876953, "learning_rate": 2.5555233995771757e-06, "loss": 0.3721, "step": 99390 }, { "epoch": 99.42, "grad_norm": 48.93081283569336, "learning_rate": 2.5527833376168055e-06, "loss": 0.324, "step": 99420 }, { "epoch": 99.45, "grad_norm": 77.30941772460938, "learning_rate": 2.5500442417788176e-06, "loss": 0.3502, "step": 99450 }, { "epoch": 99.48, "grad_norm": 33.05656433105469, "learning_rate": 2.547306113144564e-06, "loss": 0.3326, "step": 99480 }, { "epoch": 99.51, "grad_norm": 58.29331588745117, "learning_rate": 2.5445689527950135e-06, "loss": 0.2651, "step": 99510 }, { "epoch": 99.54, "grad_norm": 48.23080062866211, "learning_rate": 2.541832761810753e-06, "loss": 0.3556, "step": 99540 }, { "epoch": 99.57, "grad_norm": 44.910823822021484, "learning_rate": 2.5390975412719897e-06, "loss": 0.3359, "step": 99570 }, { "epoch": 99.6, "grad_norm": 26.720394134521484, "learning_rate": 2.536363292258543e-06, "loss": 0.2934, "step": 99600 }, { "epoch": 99.63, "grad_norm": 55.45170974731445, "learning_rate": 2.5336300158498518e-06, "loss": 0.2581, "step": 99630 }, { "epoch": 99.66, "grad_norm": 100.24824523925781, "learning_rate": 2.5308977131249724e-06, "loss": 0.3375, "step": 99660 }, { "epoch": 99.69, "grad_norm": 39.15064239501953, "learning_rate": 2.5281663851625703e-06, "loss": 0.3296, "step": 99690 }, { "epoch": 99.72, "grad_norm": 67.25764465332031, "learning_rate": 2.5254360330409343e-06, "loss": 0.2481, "step": 99720 }, { "epoch": 99.75, "grad_norm": 72.34410858154297, "learning_rate": 2.5227066578379624e-06, "loss": 0.26, "step": 99750 }, { "epoch": 99.78, "grad_norm": 56.84872817993164, "learning_rate": 2.5199782606311708e-06, "loss": 0.4139, "step": 99780 }, { "epoch": 99.81, "grad_norm": 9.12968025659211e-05, "learning_rate": 2.5172508424976837e-06, "loss": 0.5398, "step": 99810 }, { "epoch": 99.84, "grad_norm": 59.67491912841797, "learning_rate": 2.514524404514248e-06, "loss": 0.3011, "step": 99840 }, { "epoch": 99.87, "grad_norm": 33.41859436035156, "learning_rate": 2.5117989477572126e-06, "loss": 0.282, "step": 99870 }, { "epoch": 99.9, "grad_norm": 21.118412017822266, "learning_rate": 2.509074473302546e-06, "loss": 0.4413, "step": 99900 }, { "epoch": 99.93, "grad_norm": 56.47477722167969, "learning_rate": 2.50635098222583e-06, "loss": 0.5079, "step": 99930 }, { "epoch": 99.96, "grad_norm": 55.89588928222656, "learning_rate": 2.503628475602256e-06, "loss": 0.3917, "step": 99960 }, { "epoch": 99.99, "grad_norm": 49.626739501953125, "learning_rate": 2.500906954506623e-06, "loss": 0.364, "step": 99990 }, { "epoch": 100.0, "eval_loss": 0.2604798972606659, "eval_map": 0.824, "eval_map_50": 0.9626, "eval_map_75": 0.9338, "eval_map_chicken": 0.8203, "eval_map_duck": 0.7746, "eval_map_large": 0.8453, "eval_map_medium": 0.8195, "eval_map_plant": 0.8771, "eval_map_small": 0.323, "eval_mar_1": 0.3306, "eval_mar_10": 0.8547, "eval_mar_100": 0.8589, "eval_mar_100_chicken": 0.861, "eval_mar_100_duck": 0.8093, "eval_mar_100_plant": 0.9064, "eval_mar_large": 0.8753, "eval_mar_medium": 0.8573, "eval_mar_small": 0.4057, "eval_runtime": 13.1776, "eval_samples_per_second": 15.177, "eval_steps_per_second": 1.897, "step": 100000 }, { "epoch": 100.02, "grad_norm": 85.0333480834961, "learning_rate": 2.4981864200133483e-06, "loss": 0.2908, "step": 100020 }, { "epoch": 100.05, "grad_norm": 48.80693054199219, "learning_rate": 2.49546687319645e-06, "loss": 0.3578, "step": 100050 }, { "epoch": 100.08, "grad_norm": 43.178653717041016, "learning_rate": 2.49274831512957e-06, "loss": 0.3218, "step": 100080 }, { "epoch": 100.11, "grad_norm": 60.674842834472656, "learning_rate": 2.490030746885946e-06, "loss": 0.3521, "step": 100110 }, { "epoch": 100.14, "grad_norm": 71.40019989013672, "learning_rate": 2.4873141695384346e-06, "loss": 0.4224, "step": 100140 }, { "epoch": 100.17, "grad_norm": 35.24929428100586, "learning_rate": 2.484598584159494e-06, "loss": 0.3132, "step": 100170 }, { "epoch": 100.2, "grad_norm": 54.195438385009766, "learning_rate": 2.4818839918211963e-06, "loss": 0.3404, "step": 100200 }, { "epoch": 100.23, "grad_norm": 39.87151336669922, "learning_rate": 2.4791703935952193e-06, "loss": 0.2918, "step": 100230 }, { "epoch": 100.26, "grad_norm": 63.39114761352539, "learning_rate": 2.4764577905528503e-06, "loss": 0.3684, "step": 100260 }, { "epoch": 100.29, "grad_norm": 113.64999389648438, "learning_rate": 2.4737461837649784e-06, "loss": 0.3595, "step": 100290 }, { "epoch": 100.32, "grad_norm": 46.2447624206543, "learning_rate": 2.4710355743021077e-06, "loss": 0.3233, "step": 100320 }, { "epoch": 100.35, "grad_norm": 91.08055877685547, "learning_rate": 2.4683259632343363e-06, "loss": 0.2787, "step": 100350 }, { "epoch": 100.38, "grad_norm": 36.746707916259766, "learning_rate": 2.4656173516313852e-06, "loss": 0.385, "step": 100380 }, { "epoch": 100.41, "grad_norm": 56.71092224121094, "learning_rate": 2.4629097405625645e-06, "loss": 0.2738, "step": 100410 }, { "epoch": 100.44, "grad_norm": 45.57659912109375, "learning_rate": 2.4602031310968013e-06, "loss": 0.2905, "step": 100440 }, { "epoch": 100.47, "grad_norm": 46.59294891357422, "learning_rate": 2.457497524302616e-06, "loss": 0.2952, "step": 100470 }, { "epoch": 100.5, "grad_norm": 64.8698501586914, "learning_rate": 2.4547929212481436e-06, "loss": 0.3854, "step": 100500 }, { "epoch": 100.53, "grad_norm": 50.016109466552734, "learning_rate": 2.4520893230011174e-06, "loss": 0.3251, "step": 100530 }, { "epoch": 100.56, "grad_norm": 44.69855880737305, "learning_rate": 2.4493867306288772e-06, "loss": 0.3825, "step": 100560 }, { "epoch": 100.59, "grad_norm": 74.99469757080078, "learning_rate": 2.44668514519836e-06, "loss": 0.3465, "step": 100590 }, { "epoch": 100.62, "grad_norm": 57.389556884765625, "learning_rate": 2.4439845677761124e-06, "loss": 0.2624, "step": 100620 }, { "epoch": 100.65, "grad_norm": 219.28419494628906, "learning_rate": 2.4412849994282744e-06, "loss": 0.2516, "step": 100650 }, { "epoch": 100.68, "grad_norm": 95.02751922607422, "learning_rate": 2.4385864412206e-06, "loss": 0.3689, "step": 100680 }, { "epoch": 100.71, "grad_norm": 174.82374572753906, "learning_rate": 2.4358888942184324e-06, "loss": 0.333, "step": 100710 }, { "epoch": 100.74, "grad_norm": 408.4792785644531, "learning_rate": 2.433192359486723e-06, "loss": 0.3251, "step": 100740 }, { "epoch": 100.77, "grad_norm": 45.323577880859375, "learning_rate": 2.430496838090017e-06, "loss": 0.3761, "step": 100770 }, { "epoch": 100.8, "grad_norm": 54.754798889160156, "learning_rate": 2.4278023310924676e-06, "loss": 0.301, "step": 100800 }, { "epoch": 100.83, "grad_norm": 58.31977844238281, "learning_rate": 2.4251088395578214e-06, "loss": 0.3238, "step": 100830 }, { "epoch": 100.86, "grad_norm": 64.82942962646484, "learning_rate": 2.422416364549429e-06, "loss": 0.2805, "step": 100860 }, { "epoch": 100.89, "grad_norm": 71.43714141845703, "learning_rate": 2.419724907130233e-06, "loss": 0.3151, "step": 100890 }, { "epoch": 100.92, "grad_norm": 43.47178649902344, "learning_rate": 2.417034468362782e-06, "loss": 0.3549, "step": 100920 }, { "epoch": 100.95, "grad_norm": 39.203853607177734, "learning_rate": 2.4143450493092146e-06, "loss": 0.3238, "step": 100950 }, { "epoch": 100.98, "grad_norm": 90.06908416748047, "learning_rate": 2.4116566510312734e-06, "loss": 0.3617, "step": 100980 }, { "epoch": 101.0, "eval_loss": 0.25038665533065796, "eval_map": 0.8272, "eval_map_50": 0.9633, "eval_map_75": 0.9247, "eval_map_chicken": 0.8188, "eval_map_duck": 0.7768, "eval_map_large": 0.8427, "eval_map_medium": 0.8225, "eval_map_plant": 0.886, "eval_map_small": 0.2991, "eval_mar_1": 0.3337, "eval_mar_10": 0.859, "eval_mar_100": 0.8635, "eval_mar_100_chicken": 0.862, "eval_mar_100_duck": 0.8139, "eval_mar_100_plant": 0.9146, "eval_mar_large": 0.8814, "eval_mar_medium": 0.8625, "eval_mar_small": 0.3882, "eval_runtime": 12.9142, "eval_samples_per_second": 15.487, "eval_steps_per_second": 1.936, "step": 101000 }, { "epoch": 101.01, "grad_norm": 61.029300689697266, "learning_rate": 2.408969274590296e-06, "loss": 0.2618, "step": 101010 }, { "epoch": 101.04, "grad_norm": 69.63468933105469, "learning_rate": 2.406282921047213e-06, "loss": 0.3434, "step": 101040 }, { "epoch": 101.07, "grad_norm": 32.13951873779297, "learning_rate": 2.403597591462557e-06, "loss": 0.2697, "step": 101070 }, { "epoch": 101.1, "grad_norm": 40.814666748046875, "learning_rate": 2.4009132868964525e-06, "loss": 0.3236, "step": 101100 }, { "epoch": 101.13, "grad_norm": 44.477325439453125, "learning_rate": 2.3982300084086224e-06, "loss": 0.3305, "step": 101130 }, { "epoch": 101.16, "grad_norm": 74.7467269897461, "learning_rate": 2.395547757058379e-06, "loss": 0.3437, "step": 101160 }, { "epoch": 101.19, "grad_norm": 117.67740631103516, "learning_rate": 2.3928665339046363e-06, "loss": 0.3072, "step": 101190 }, { "epoch": 101.22, "grad_norm": 68.71038818359375, "learning_rate": 2.3901863400058954e-06, "loss": 0.303, "step": 101220 }, { "epoch": 101.25, "grad_norm": 40.906612396240234, "learning_rate": 2.387507176420256e-06, "loss": 0.3234, "step": 101250 }, { "epoch": 101.28, "grad_norm": 112.5341567993164, "learning_rate": 2.3848290442054096e-06, "loss": 0.3283, "step": 101280 }, { "epoch": 101.31, "grad_norm": 97.07479095458984, "learning_rate": 2.382151944418642e-06, "loss": 0.351, "step": 101310 }, { "epoch": 101.34, "grad_norm": 37.94271469116211, "learning_rate": 2.379475878116826e-06, "loss": 0.3132, "step": 101340 }, { "epoch": 101.37, "grad_norm": 1115.914306640625, "learning_rate": 2.376800846356434e-06, "loss": 0.2897, "step": 101370 }, { "epoch": 101.4, "grad_norm": 63.31821060180664, "learning_rate": 2.3741268501935212e-06, "loss": 0.3253, "step": 101400 }, { "epoch": 101.43, "grad_norm": 77.8603286743164, "learning_rate": 2.3714538906837452e-06, "loss": 0.3477, "step": 101430 }, { "epoch": 101.46, "grad_norm": 46.80801010131836, "learning_rate": 2.368781968882343e-06, "loss": 0.3682, "step": 101460 }, { "epoch": 101.49, "grad_norm": 74.0793685913086, "learning_rate": 2.3661110858441517e-06, "loss": 0.4072, "step": 101490 }, { "epoch": 101.52, "grad_norm": 93.3003921508789, "learning_rate": 2.3634412426235886e-06, "loss": 0.3219, "step": 101520 }, { "epoch": 101.55, "grad_norm": 52.63278579711914, "learning_rate": 2.3607724402746685e-06, "loss": 0.2792, "step": 101550 }, { "epoch": 101.58, "grad_norm": 68.7774429321289, "learning_rate": 2.358104679850991e-06, "loss": 0.3707, "step": 101580 }, { "epoch": 101.61, "grad_norm": 157.89881896972656, "learning_rate": 2.3554379624057485e-06, "loss": 0.2892, "step": 101610 }, { "epoch": 101.64, "grad_norm": 105.74935150146484, "learning_rate": 2.3527722889917147e-06, "loss": 0.3149, "step": 101640 }, { "epoch": 101.67, "grad_norm": 27.93893051147461, "learning_rate": 2.3501076606612587e-06, "loss": 0.3225, "step": 101670 }, { "epoch": 101.7, "grad_norm": 242.32102966308594, "learning_rate": 2.3474440784663287e-06, "loss": 0.2682, "step": 101700 }, { "epoch": 101.73, "grad_norm": 46.66429901123047, "learning_rate": 2.344781543458472e-06, "loss": 0.3345, "step": 101730 }, { "epoch": 101.76, "grad_norm": 33.549198150634766, "learning_rate": 2.3421200566888096e-06, "loss": 0.3117, "step": 101760 }, { "epoch": 101.79, "grad_norm": 126.8124008178711, "learning_rate": 2.3394596192080575e-06, "loss": 0.279, "step": 101790 }, { "epoch": 101.82, "grad_norm": 94.76863098144531, "learning_rate": 2.3368002320665118e-06, "loss": 0.2954, "step": 101820 }, { "epoch": 101.85, "grad_norm": 64.32222747802734, "learning_rate": 2.334141896314057e-06, "loss": 0.3767, "step": 101850 }, { "epoch": 101.88, "grad_norm": 41.3961181640625, "learning_rate": 2.3314846130001622e-06, "loss": 0.3157, "step": 101880 }, { "epoch": 101.91, "grad_norm": 36.182281494140625, "learning_rate": 2.3288283831738834e-06, "loss": 0.2918, "step": 101910 }, { "epoch": 101.94, "grad_norm": 101.90792846679688, "learning_rate": 2.326173207883854e-06, "loss": 0.2557, "step": 101940 }, { "epoch": 101.97, "grad_norm": 84.4319839477539, "learning_rate": 2.323519088178299e-06, "loss": 0.4543, "step": 101970 }, { "epoch": 102.0, "grad_norm": 122.67556762695312, "learning_rate": 2.320866025105016e-06, "loss": 0.2855, "step": 102000 }, { "epoch": 102.0, "eval_loss": 0.2508144676685333, "eval_map": 0.8213, "eval_map_50": 0.9647, "eval_map_75": 0.9329, "eval_map_chicken": 0.8165, "eval_map_duck": 0.7631, "eval_map_large": 0.8351, "eval_map_medium": 0.8192, "eval_map_plant": 0.8843, "eval_map_small": 0.3075, "eval_mar_1": 0.3294, "eval_mar_10": 0.8561, "eval_mar_100": 0.86, "eval_mar_100_chicken": 0.865, "eval_mar_100_duck": 0.8031, "eval_mar_100_plant": 0.9118, "eval_mar_large": 0.8727, "eval_mar_medium": 0.8619, "eval_mar_small": 0.392, "eval_runtime": 12.5729, "eval_samples_per_second": 15.907, "eval_steps_per_second": 1.988, "step": 102000 }, { "epoch": 102.03, "grad_norm": 60.3825798034668, "learning_rate": 2.318214019711401e-06, "loss": 0.3122, "step": 102030 }, { "epoch": 102.06, "grad_norm": 3.619626659201458e-05, "learning_rate": 2.3155630730444185e-06, "loss": 0.3544, "step": 102060 }, { "epoch": 102.09, "grad_norm": 58.06379318237305, "learning_rate": 2.3129131861506225e-06, "loss": 0.358, "step": 102090 }, { "epoch": 102.12, "grad_norm": 44.004215240478516, "learning_rate": 2.3102643600761445e-06, "loss": 0.3085, "step": 102120 }, { "epoch": 102.15, "grad_norm": 28.617773056030273, "learning_rate": 2.3076165958666992e-06, "loss": 0.3558, "step": 102150 }, { "epoch": 102.18, "grad_norm": 61.483184814453125, "learning_rate": 2.3049698945675826e-06, "loss": 0.2916, "step": 102180 }, { "epoch": 102.21, "grad_norm": 46.97686004638672, "learning_rate": 2.3023242572236728e-06, "loss": 0.3124, "step": 102210 }, { "epoch": 102.24, "grad_norm": 78.1906967163086, "learning_rate": 2.299679684879421e-06, "loss": 0.3298, "step": 102240 }, { "epoch": 102.27, "grad_norm": 50.59100341796875, "learning_rate": 2.2970361785788673e-06, "loss": 0.2998, "step": 102270 }, { "epoch": 102.3, "grad_norm": 70.06971740722656, "learning_rate": 2.294393739365621e-06, "loss": 0.3162, "step": 102300 }, { "epoch": 102.33, "grad_norm": 126.2216567993164, "learning_rate": 2.291752368282879e-06, "loss": 0.3468, "step": 102330 }, { "epoch": 102.36, "grad_norm": 60.906272888183594, "learning_rate": 2.289112066373411e-06, "loss": 0.3134, "step": 102360 }, { "epoch": 102.39, "grad_norm": 94.25902557373047, "learning_rate": 2.2864728346795686e-06, "loss": 0.334, "step": 102390 }, { "epoch": 102.42, "grad_norm": 25.635311126708984, "learning_rate": 2.2838346742432753e-06, "loss": 0.3264, "step": 102420 }, { "epoch": 102.45, "grad_norm": 65.8117904663086, "learning_rate": 2.281197586106037e-06, "loss": 0.3061, "step": 102450 }, { "epoch": 102.48, "grad_norm": 61.342586517333984, "learning_rate": 2.2785615713089363e-06, "loss": 0.3189, "step": 102480 }, { "epoch": 102.51, "grad_norm": 91.99687957763672, "learning_rate": 2.2759266308926257e-06, "loss": 0.3387, "step": 102510 }, { "epoch": 102.54, "grad_norm": 50.61820602416992, "learning_rate": 2.2732927658973427e-06, "loss": 0.2439, "step": 102540 }, { "epoch": 102.57, "grad_norm": 66.88514709472656, "learning_rate": 2.2706599773628906e-06, "loss": 0.3848, "step": 102570 }, { "epoch": 102.6, "grad_norm": 112.15740966796875, "learning_rate": 2.268028266328655e-06, "loss": 0.3972, "step": 102600 }, { "epoch": 102.63, "grad_norm": 37.6876220703125, "learning_rate": 2.2653976338335936e-06, "loss": 0.2637, "step": 102630 }, { "epoch": 102.66, "grad_norm": 73.63807678222656, "learning_rate": 2.262768080916241e-06, "loss": 0.3063, "step": 102660 }, { "epoch": 102.69, "grad_norm": 57.28899383544922, "learning_rate": 2.260139608614699e-06, "loss": 0.3008, "step": 102690 }, { "epoch": 102.72, "grad_norm": 42.878684997558594, "learning_rate": 2.25751221796665e-06, "loss": 0.3828, "step": 102720 }, { "epoch": 102.75, "grad_norm": 90.36102294921875, "learning_rate": 2.254885910009341e-06, "loss": 0.3758, "step": 102750 }, { "epoch": 102.78, "grad_norm": 28.182830810546875, "learning_rate": 2.2522606857796036e-06, "loss": 0.3613, "step": 102780 }, { "epoch": 102.81, "grad_norm": 99.58158111572266, "learning_rate": 2.249636546313831e-06, "loss": 0.2665, "step": 102810 }, { "epoch": 102.84, "grad_norm": 49.25252914428711, "learning_rate": 2.247013492647994e-06, "loss": 0.2757, "step": 102840 }, { "epoch": 102.87, "grad_norm": 43.810909271240234, "learning_rate": 2.2443915258176283e-06, "loss": 0.3454, "step": 102870 }, { "epoch": 102.9, "grad_norm": 61.715576171875, "learning_rate": 2.2417706468578495e-06, "loss": 0.277, "step": 102900 }, { "epoch": 102.93, "grad_norm": 79.8405532836914, "learning_rate": 2.239150856803336e-06, "loss": 0.3071, "step": 102930 }, { "epoch": 102.96, "grad_norm": 44.50522994995117, "learning_rate": 2.2365321566883437e-06, "loss": 0.2985, "step": 102960 }, { "epoch": 102.99, "grad_norm": 68.21998596191406, "learning_rate": 2.2339145475466885e-06, "loss": 0.3384, "step": 102990 }, { "epoch": 103.0, "eval_loss": 0.2511653006076813, "eval_map": 0.828, "eval_map_50": 0.9619, "eval_map_75": 0.934, "eval_map_chicken": 0.8158, "eval_map_duck": 0.7848, "eval_map_large": 0.8439, "eval_map_medium": 0.8276, "eval_map_plant": 0.8833, "eval_map_small": 0.3208, "eval_mar_1": 0.3343, "eval_mar_10": 0.8606, "eval_mar_100": 0.8646, "eval_mar_100_chicken": 0.8602, "eval_mar_100_duck": 0.8216, "eval_mar_100_plant": 0.9119, "eval_mar_large": 0.8757, "eval_mar_medium": 0.8657, "eval_mar_small": 0.4164, "eval_runtime": 12.9173, "eval_samples_per_second": 15.483, "eval_steps_per_second": 1.935, "step": 103000 }, { "epoch": 103.02, "grad_norm": 62.947235107421875, "learning_rate": 2.2312980304117656e-06, "loss": 0.3145, "step": 103020 }, { "epoch": 103.05, "grad_norm": 36.934574127197266, "learning_rate": 2.228682606316529e-06, "loss": 0.2659, "step": 103050 }, { "epoch": 103.08, "grad_norm": 156.98410034179688, "learning_rate": 2.2260682762935137e-06, "loss": 0.2886, "step": 103080 }, { "epoch": 103.11, "grad_norm": 24.20302391052246, "learning_rate": 2.2234550413748106e-06, "loss": 0.4062, "step": 103110 }, { "epoch": 103.14, "grad_norm": 134.32875061035156, "learning_rate": 2.220842902592087e-06, "loss": 0.2614, "step": 103140 }, { "epoch": 103.17, "grad_norm": 43.908565521240234, "learning_rate": 2.2182318609765703e-06, "loss": 0.3216, "step": 103170 }, { "epoch": 103.2, "grad_norm": 312.3174133300781, "learning_rate": 2.2156219175590623e-06, "loss": 0.3144, "step": 103200 }, { "epoch": 103.23, "grad_norm": 43.799591064453125, "learning_rate": 2.2130130733699206e-06, "loss": 0.3458, "step": 103230 }, { "epoch": 103.26, "grad_norm": 78.38511657714844, "learning_rate": 2.2104053294390847e-06, "loss": 0.3516, "step": 103260 }, { "epoch": 103.29, "grad_norm": 73.01293182373047, "learning_rate": 2.2077986867960436e-06, "loss": 0.3513, "step": 103290 }, { "epoch": 103.32, "grad_norm": 77.19669342041016, "learning_rate": 2.2051931464698636e-06, "loss": 0.336, "step": 103320 }, { "epoch": 103.35, "grad_norm": 912.9269409179688, "learning_rate": 2.2025887094891657e-06, "loss": 0.3546, "step": 103350 }, { "epoch": 103.38, "grad_norm": 27.250869750976562, "learning_rate": 2.1999853768821433e-06, "loss": 0.2604, "step": 103380 }, { "epoch": 103.41, "grad_norm": 59.26245880126953, "learning_rate": 2.1973831496765503e-06, "loss": 0.3117, "step": 103410 }, { "epoch": 103.44, "grad_norm": 45.40540313720703, "learning_rate": 2.1947820288997067e-06, "loss": 0.3625, "step": 103440 }, { "epoch": 103.47, "grad_norm": 6.854474486317486e-05, "learning_rate": 2.19218201557849e-06, "loss": 0.3335, "step": 103470 }, { "epoch": 103.5, "grad_norm": 66.07411193847656, "learning_rate": 2.1895831107393485e-06, "loss": 0.2868, "step": 103500 }, { "epoch": 103.53, "grad_norm": 4.3078551243525e-05, "learning_rate": 2.1869853154082828e-06, "loss": 0.375, "step": 103530 }, { "epoch": 103.56, "grad_norm": 92.22895812988281, "learning_rate": 2.1843886306108686e-06, "loss": 0.2935, "step": 103560 }, { "epoch": 103.59, "grad_norm": 66.51820373535156, "learning_rate": 2.18179305737223e-06, "loss": 0.323, "step": 103590 }, { "epoch": 103.62, "grad_norm": 45.70412063598633, "learning_rate": 2.179198596717063e-06, "loss": 0.3348, "step": 103620 }, { "epoch": 103.65, "grad_norm": 72.18115997314453, "learning_rate": 2.1766052496696155e-06, "loss": 0.2994, "step": 103650 }, { "epoch": 103.68, "grad_norm": 76.4967269897461, "learning_rate": 2.174013017253701e-06, "loss": 0.291, "step": 103680 }, { "epoch": 103.71, "grad_norm": 53.50641632080078, "learning_rate": 2.1714219004926923e-06, "loss": 0.3134, "step": 103710 }, { "epoch": 103.74, "grad_norm": 64.19667053222656, "learning_rate": 2.168831900409523e-06, "loss": 0.3557, "step": 103740 }, { "epoch": 103.77, "grad_norm": 123.80131530761719, "learning_rate": 2.1662430180266808e-06, "loss": 0.411, "step": 103770 }, { "epoch": 103.8, "grad_norm": 177.2345733642578, "learning_rate": 2.1636552543662187e-06, "loss": 0.3523, "step": 103800 }, { "epoch": 103.83, "grad_norm": 45.02297592163086, "learning_rate": 2.1610686104497413e-06, "loss": 0.3573, "step": 103830 }, { "epoch": 103.86, "grad_norm": 73.64077758789062, "learning_rate": 2.158483087298417e-06, "loss": 0.3679, "step": 103860 }, { "epoch": 103.89, "grad_norm": 112.79368591308594, "learning_rate": 2.15589868593297e-06, "loss": 0.3121, "step": 103890 }, { "epoch": 103.92, "grad_norm": 6.195154855959117e-05, "learning_rate": 2.153315407373679e-06, "loss": 0.2903, "step": 103920 }, { "epoch": 103.95, "grad_norm": 67.447265625, "learning_rate": 2.1507332526403814e-06, "loss": 0.3023, "step": 103950 }, { "epoch": 103.98, "grad_norm": 82.7964096069336, "learning_rate": 2.1481522227524725e-06, "loss": 0.3331, "step": 103980 }, { "epoch": 104.0, "eval_loss": 0.2545425593852997, "eval_map": 0.8195, "eval_map_50": 0.9621, "eval_map_75": 0.9333, "eval_map_chicken": 0.813, "eval_map_duck": 0.7652, "eval_map_large": 0.8263, "eval_map_medium": 0.8183, "eval_map_plant": 0.8803, "eval_map_small": 0.3061, "eval_mar_1": 0.331, "eval_mar_10": 0.8535, "eval_mar_100": 0.8564, "eval_mar_100_chicken": 0.8569, "eval_mar_100_duck": 0.8052, "eval_mar_100_plant": 0.9071, "eval_mar_large": 0.8602, "eval_mar_medium": 0.8573, "eval_mar_small": 0.4039, "eval_runtime": 12.7038, "eval_samples_per_second": 15.743, "eval_steps_per_second": 1.968, "step": 104000 }, { "epoch": 104.01, "grad_norm": 56.906036376953125, "learning_rate": 2.1455723187289028e-06, "loss": 0.3692, "step": 104010 }, { "epoch": 104.04, "grad_norm": 72.5718765258789, "learning_rate": 2.1429935415881753e-06, "loss": 0.353, "step": 104040 }, { "epoch": 104.07, "grad_norm": 60.61161422729492, "learning_rate": 2.1404158923483524e-06, "loss": 0.3477, "step": 104070 }, { "epoch": 104.1, "grad_norm": 33.45027160644531, "learning_rate": 2.137839372027047e-06, "loss": 0.2641, "step": 104100 }, { "epoch": 104.13, "grad_norm": 39.38404846191406, "learning_rate": 2.135263981641429e-06, "loss": 0.295, "step": 104130 }, { "epoch": 104.16, "grad_norm": 91.07547760009766, "learning_rate": 2.132689722208223e-06, "loss": 0.3602, "step": 104160 }, { "epoch": 104.19, "grad_norm": 4.448373874765821e-05, "learning_rate": 2.1301165947437064e-06, "loss": 0.2545, "step": 104190 }, { "epoch": 104.22, "grad_norm": 74.86077880859375, "learning_rate": 2.1275446002637063e-06, "loss": 0.3214, "step": 104220 }, { "epoch": 104.25, "grad_norm": 103.2509536743164, "learning_rate": 2.124973739783609e-06, "loss": 0.4231, "step": 104250 }, { "epoch": 104.28, "grad_norm": 69.51040649414062, "learning_rate": 2.1224040143183444e-06, "loss": 0.4107, "step": 104280 }, { "epoch": 104.31, "grad_norm": 31.188066482543945, "learning_rate": 2.1198354248824057e-06, "loss": 0.3611, "step": 104310 }, { "epoch": 104.34, "grad_norm": 78.05699920654297, "learning_rate": 2.1172679724898264e-06, "loss": 0.4143, "step": 104340 }, { "epoch": 104.37, "grad_norm": 73.48011779785156, "learning_rate": 2.1147016581542e-06, "loss": 0.352, "step": 104370 }, { "epoch": 104.4, "grad_norm": 118.53067016601562, "learning_rate": 2.112136482888663e-06, "loss": 0.3032, "step": 104400 }, { "epoch": 104.43, "grad_norm": 44.73308563232422, "learning_rate": 2.1095724477059077e-06, "loss": 0.3028, "step": 104430 }, { "epoch": 104.46, "grad_norm": 40.50200271606445, "learning_rate": 2.107009553618174e-06, "loss": 0.3042, "step": 104460 }, { "epoch": 104.49, "grad_norm": 71.992431640625, "learning_rate": 2.1044478016372544e-06, "loss": 0.3323, "step": 104490 }, { "epoch": 104.52, "grad_norm": 103.34069061279297, "learning_rate": 2.1018871927744844e-06, "loss": 0.3679, "step": 104520 }, { "epoch": 104.55, "grad_norm": 99.35243225097656, "learning_rate": 2.0993277280407547e-06, "loss": 0.4147, "step": 104550 }, { "epoch": 104.58, "grad_norm": 55.58158493041992, "learning_rate": 2.0967694084464973e-06, "loss": 0.302, "step": 104580 }, { "epoch": 104.61, "grad_norm": 35.0411376953125, "learning_rate": 2.0942122350017023e-06, "loss": 0.2767, "step": 104610 }, { "epoch": 104.64, "grad_norm": 36.08042526245117, "learning_rate": 2.0916562087158964e-06, "loss": 0.2559, "step": 104640 }, { "epoch": 104.67, "grad_norm": 137.35403442382812, "learning_rate": 2.0891013305981615e-06, "loss": 0.3401, "step": 104670 }, { "epoch": 104.7, "grad_norm": 30.13197135925293, "learning_rate": 2.0865476016571206e-06, "loss": 0.3265, "step": 104700 }, { "epoch": 104.73, "grad_norm": 120.36119842529297, "learning_rate": 2.083995022900946e-06, "loss": 0.364, "step": 104730 }, { "epoch": 104.76, "grad_norm": 34.020904541015625, "learning_rate": 2.0814435953373554e-06, "loss": 0.3071, "step": 104760 }, { "epoch": 104.79, "grad_norm": 6.172459688968956e-05, "learning_rate": 2.0788933199736145e-06, "loss": 0.2864, "step": 104790 }, { "epoch": 104.82, "grad_norm": 57.665565490722656, "learning_rate": 2.076344197816527e-06, "loss": 0.3432, "step": 104820 }, { "epoch": 104.85, "grad_norm": 35.08525848388672, "learning_rate": 2.0737962298724513e-06, "loss": 0.3342, "step": 104850 }, { "epoch": 104.88, "grad_norm": 29.563392639160156, "learning_rate": 2.0712494171472776e-06, "loss": 0.3357, "step": 104880 }, { "epoch": 104.91, "grad_norm": 108.29151153564453, "learning_rate": 2.0687037606464554e-06, "loss": 0.2898, "step": 104910 }, { "epoch": 104.94, "grad_norm": 19.089824676513672, "learning_rate": 2.066159261374964e-06, "loss": 0.3056, "step": 104940 }, { "epoch": 104.97, "grad_norm": 51.8682746887207, "learning_rate": 2.0636159203373344e-06, "loss": 0.2571, "step": 104970 }, { "epoch": 105.0, "grad_norm": 50.808685302734375, "learning_rate": 2.061073738537635e-06, "loss": 0.3158, "step": 105000 }, { "epoch": 105.0, "eval_loss": 0.2530730068683624, "eval_map": 0.8304, "eval_map_50": 0.9634, "eval_map_75": 0.9254, "eval_map_chicken": 0.8222, "eval_map_duck": 0.7788, "eval_map_large": 0.8411, "eval_map_medium": 0.8263, "eval_map_plant": 0.89, "eval_map_small": 0.2998, "eval_mar_1": 0.3356, "eval_mar_10": 0.8632, "eval_mar_100": 0.8671, "eval_mar_100_chicken": 0.867, "eval_mar_100_duck": 0.8196, "eval_mar_100_plant": 0.9148, "eval_mar_large": 0.8794, "eval_mar_medium": 0.865, "eval_mar_small": 0.3968, "eval_runtime": 12.6701, "eval_samples_per_second": 15.785, "eval_steps_per_second": 1.973, "step": 105000 }, { "epoch": 105.03, "grad_norm": 26.10096549987793, "learning_rate": 2.0585327169794796e-06, "loss": 0.3782, "step": 105030 }, { "epoch": 105.06, "grad_norm": 96.70814514160156, "learning_rate": 2.0559928566660235e-06, "loss": 0.2806, "step": 105060 }, { "epoch": 105.09, "grad_norm": 36.11616516113281, "learning_rate": 2.0534541585999656e-06, "loss": 0.2919, "step": 105090 }, { "epoch": 105.12, "grad_norm": 75.42376708984375, "learning_rate": 2.0509166237835398e-06, "loss": 0.3285, "step": 105120 }, { "epoch": 105.15, "grad_norm": 93.80187225341797, "learning_rate": 2.0483802532185286e-06, "loss": 0.3454, "step": 105150 }, { "epoch": 105.18, "grad_norm": 71.72364044189453, "learning_rate": 2.0458450479062465e-06, "loss": 0.391, "step": 105180 }, { "epoch": 105.21, "grad_norm": 67.4125747680664, "learning_rate": 2.043311008847555e-06, "loss": 0.2992, "step": 105210 }, { "epoch": 105.24, "grad_norm": 8.454218914266676e-05, "learning_rate": 2.040778137042852e-06, "loss": 0.2799, "step": 105240 }, { "epoch": 105.27, "grad_norm": 71.04151916503906, "learning_rate": 2.0382464334920774e-06, "loss": 0.5559, "step": 105270 }, { "epoch": 105.3, "grad_norm": 51.83363723754883, "learning_rate": 2.035715899194704e-06, "loss": 0.3032, "step": 105300 }, { "epoch": 105.33, "grad_norm": 48.12313461303711, "learning_rate": 2.033186535149748e-06, "loss": 0.3774, "step": 105330 }, { "epoch": 105.36, "grad_norm": 44.62849044799805, "learning_rate": 2.030658342355765e-06, "loss": 0.354, "step": 105360 }, { "epoch": 105.39, "grad_norm": 64.9791030883789, "learning_rate": 2.028131321810841e-06, "loss": 0.3605, "step": 105390 }, { "epoch": 105.42, "grad_norm": 39.539546966552734, "learning_rate": 2.025605474512608e-06, "loss": 0.3553, "step": 105420 }, { "epoch": 105.45, "grad_norm": 75.35477447509766, "learning_rate": 2.0230808014582265e-06, "loss": 0.3944, "step": 105450 }, { "epoch": 105.48, "grad_norm": 65.32703399658203, "learning_rate": 2.0205573036443994e-06, "loss": 0.3212, "step": 105480 }, { "epoch": 105.51, "grad_norm": 181.7245330810547, "learning_rate": 2.018034982067363e-06, "loss": 0.3469, "step": 105510 }, { "epoch": 105.54, "grad_norm": 93.93302154541016, "learning_rate": 2.0155138377228924e-06, "loss": 0.2711, "step": 105540 }, { "epoch": 105.57, "grad_norm": 49.45730972290039, "learning_rate": 2.0129938716062917e-06, "loss": 0.338, "step": 105570 }, { "epoch": 105.6, "grad_norm": 54.22216796875, "learning_rate": 2.0104750847124075e-06, "loss": 0.4103, "step": 105600 }, { "epoch": 105.63, "grad_norm": 44.252559661865234, "learning_rate": 2.0079574780356116e-06, "loss": 0.3136, "step": 105630 }, { "epoch": 105.66, "grad_norm": 79.4963150024414, "learning_rate": 2.0054410525698217e-06, "loss": 0.4003, "step": 105660 }, { "epoch": 105.69, "grad_norm": 182.1365966796875, "learning_rate": 2.0029258093084774e-06, "loss": 0.3815, "step": 105690 }, { "epoch": 105.72, "grad_norm": 45.57594299316406, "learning_rate": 2.0004117492445614e-06, "loss": 0.3637, "step": 105720 }, { "epoch": 105.75, "grad_norm": 47.79154586791992, "learning_rate": 1.9978988733705807e-06, "loss": 0.3502, "step": 105750 }, { "epoch": 105.78, "grad_norm": 63.35527038574219, "learning_rate": 1.9953871826785804e-06, "loss": 0.3138, "step": 105780 }, { "epoch": 105.81, "grad_norm": 63.23777389526367, "learning_rate": 1.9928766781601366e-06, "loss": 0.3653, "step": 105810 }, { "epoch": 105.84, "grad_norm": 62.715694427490234, "learning_rate": 1.990367360806359e-06, "loss": 0.4091, "step": 105840 }, { "epoch": 105.87, "grad_norm": 1241.5020751953125, "learning_rate": 1.9878592316078813e-06, "loss": 0.3384, "step": 105870 }, { "epoch": 105.9, "grad_norm": 65.9755859375, "learning_rate": 1.9853522915548777e-06, "loss": 0.3038, "step": 105900 }, { "epoch": 105.93, "grad_norm": 40.31246566772461, "learning_rate": 1.9828465416370434e-06, "loss": 0.2752, "step": 105930 }, { "epoch": 105.96, "grad_norm": 37.16298294067383, "learning_rate": 1.980341982843616e-06, "loss": 0.3361, "step": 105960 }, { "epoch": 105.99, "grad_norm": 131.2887725830078, "learning_rate": 1.977838616163349e-06, "loss": 0.301, "step": 105990 }, { "epoch": 106.0, "eval_loss": 0.25964710116386414, "eval_map": 0.8221, "eval_map_50": 0.961, "eval_map_75": 0.9276, "eval_map_chicken": 0.8082, "eval_map_duck": 0.7767, "eval_map_large": 0.8335, "eval_map_medium": 0.8191, "eval_map_plant": 0.8815, "eval_map_small": 0.2968, "eval_mar_1": 0.3325, "eval_mar_10": 0.8551, "eval_mar_100": 0.8587, "eval_mar_100_chicken": 0.8543, "eval_mar_100_duck": 0.8149, "eval_mar_100_plant": 0.9067, "eval_mar_large": 0.8703, "eval_mar_medium": 0.8586, "eval_mar_small": 0.3629, "eval_runtime": 12.6577, "eval_samples_per_second": 15.801, "eval_steps_per_second": 1.975, "step": 106000 }, { "epoch": 106.02, "grad_norm": 97.86959075927734, "learning_rate": 1.975336442584537e-06, "loss": 0.2732, "step": 106020 }, { "epoch": 106.05, "grad_norm": 59.18266296386719, "learning_rate": 1.9728354630949935e-06, "loss": 0.3655, "step": 106050 }, { "epoch": 106.08, "grad_norm": 46.19010543823242, "learning_rate": 1.9703356786820687e-06, "loss": 0.3915, "step": 106080 }, { "epoch": 106.11, "grad_norm": 119.53529357910156, "learning_rate": 1.967837090332637e-06, "loss": 0.2545, "step": 106110 }, { "epoch": 106.14, "grad_norm": 65.8073959350586, "learning_rate": 1.9653396990331043e-06, "loss": 0.3228, "step": 106140 }, { "epoch": 106.17, "grad_norm": 40.57026672363281, "learning_rate": 1.9628435057693963e-06, "loss": 0.3128, "step": 106170 }, { "epoch": 106.2, "grad_norm": 6.559910252690315e-05, "learning_rate": 1.9603485115269743e-06, "loss": 0.3014, "step": 106200 }, { "epoch": 106.23, "grad_norm": 43.270286560058594, "learning_rate": 1.9578547172908185e-06, "loss": 0.2841, "step": 106230 }, { "epoch": 106.26, "grad_norm": 168.57339477539062, "learning_rate": 1.955362124045445e-06, "loss": 0.2592, "step": 106260 }, { "epoch": 106.29, "grad_norm": 105.64681243896484, "learning_rate": 1.9528707327748853e-06, "loss": 0.2811, "step": 106290 }, { "epoch": 106.32, "grad_norm": 95.56590270996094, "learning_rate": 1.9503805444627054e-06, "loss": 0.2831, "step": 106320 }, { "epoch": 106.35, "grad_norm": 34.25845718383789, "learning_rate": 1.9478915600919877e-06, "loss": 0.3337, "step": 106350 }, { "epoch": 106.38, "grad_norm": 29.421674728393555, "learning_rate": 1.945403780645346e-06, "loss": 0.2431, "step": 106380 }, { "epoch": 106.41, "grad_norm": 44.345455169677734, "learning_rate": 1.942917207104917e-06, "loss": 0.3748, "step": 106410 }, { "epoch": 106.44, "grad_norm": 53.64118576049805, "learning_rate": 1.9404318404523605e-06, "loss": 0.3771, "step": 106440 }, { "epoch": 106.47, "grad_norm": 23.742046356201172, "learning_rate": 1.937947681668858e-06, "loss": 0.3309, "step": 106470 }, { "epoch": 106.5, "grad_norm": 31.696788787841797, "learning_rate": 1.9354647317351187e-06, "loss": 0.3088, "step": 106500 }, { "epoch": 106.53, "grad_norm": 117.04497528076172, "learning_rate": 1.9329829916313684e-06, "loss": 0.2771, "step": 106530 }, { "epoch": 106.56, "grad_norm": 56.607704162597656, "learning_rate": 1.930502462337362e-06, "loss": 0.297, "step": 106560 }, { "epoch": 106.59, "grad_norm": 39.30537033081055, "learning_rate": 1.928023144832371e-06, "loss": 0.3055, "step": 106590 }, { "epoch": 106.62, "grad_norm": 106.04175567626953, "learning_rate": 1.9255450400951937e-06, "loss": 0.2973, "step": 106620 }, { "epoch": 106.65, "grad_norm": 69.02726745605469, "learning_rate": 1.9230681491041425e-06, "loss": 0.2414, "step": 106650 }, { "epoch": 106.68, "grad_norm": 57.96987533569336, "learning_rate": 1.920592472837057e-06, "loss": 0.3787, "step": 106680 }, { "epoch": 106.71, "grad_norm": 51.3906364440918, "learning_rate": 1.918118012271297e-06, "loss": 0.3492, "step": 106710 }, { "epoch": 106.74, "grad_norm": 56.02267837524414, "learning_rate": 1.9156447683837365e-06, "loss": 0.3529, "step": 106740 }, { "epoch": 106.77, "grad_norm": 56.653987884521484, "learning_rate": 1.913172742150774e-06, "loss": 0.2748, "step": 106770 }, { "epoch": 106.8, "grad_norm": 128.84120178222656, "learning_rate": 1.910701934548329e-06, "loss": 0.4086, "step": 106800 }, { "epoch": 106.83, "grad_norm": 94.21256256103516, "learning_rate": 1.908232346551834e-06, "loss": 0.3239, "step": 106830 }, { "epoch": 106.86, "grad_norm": 37.86124801635742, "learning_rate": 1.9057639791362437e-06, "loss": 0.2975, "step": 106860 }, { "epoch": 106.89, "grad_norm": 145.65882873535156, "learning_rate": 1.9032968332760331e-06, "loss": 0.3039, "step": 106890 }, { "epoch": 106.92, "grad_norm": 57.877140045166016, "learning_rate": 1.900830909945189e-06, "loss": 0.2829, "step": 106920 }, { "epoch": 106.95, "grad_norm": 50.45780563354492, "learning_rate": 1.8983662101172217e-06, "loss": 0.3274, "step": 106950 }, { "epoch": 106.98, "grad_norm": 44.34297561645508, "learning_rate": 1.8959027347651527e-06, "loss": 0.3579, "step": 106980 }, { "epoch": 107.0, "eval_loss": 0.2433866560459137, "eval_map": 0.8339, "eval_map_50": 0.9647, "eval_map_75": 0.9304, "eval_map_chicken": 0.8228, "eval_map_duck": 0.7883, "eval_map_large": 0.8464, "eval_map_medium": 0.832, "eval_map_plant": 0.8905, "eval_map_small": 0.3098, "eval_mar_1": 0.3344, "eval_mar_10": 0.8641, "eval_mar_100": 0.8685, "eval_mar_100_chicken": 0.8678, "eval_mar_100_duck": 0.8227, "eval_mar_100_plant": 0.915, "eval_mar_large": 0.877, "eval_mar_medium": 0.8705, "eval_mar_small": 0.3989, "eval_runtime": 12.6813, "eval_samples_per_second": 15.771, "eval_steps_per_second": 1.971, "step": 107000 }, { "epoch": 107.01, "grad_norm": 42.31320571899414, "learning_rate": 1.8934404848615245e-06, "loss": 0.3824, "step": 107010 }, { "epoch": 107.04, "grad_norm": 33.91014862060547, "learning_rate": 1.8909794613783943e-06, "loss": 0.3951, "step": 107040 }, { "epoch": 107.07, "grad_norm": 405.43304443359375, "learning_rate": 1.8885196652873372e-06, "loss": 0.2906, "step": 107070 }, { "epoch": 107.1, "grad_norm": 62.29679870605469, "learning_rate": 1.8860610975594384e-06, "loss": 0.2507, "step": 107100 }, { "epoch": 107.13, "grad_norm": 42.76564025878906, "learning_rate": 1.8836037591653044e-06, "loss": 0.3353, "step": 107130 }, { "epoch": 107.16, "grad_norm": 120.91712951660156, "learning_rate": 1.8811476510750486e-06, "loss": 0.2769, "step": 107160 }, { "epoch": 107.19, "grad_norm": 120.77377319335938, "learning_rate": 1.8786927742583111e-06, "loss": 0.3511, "step": 107190 }, { "epoch": 107.22, "grad_norm": 66.76496124267578, "learning_rate": 1.876239129684232e-06, "loss": 0.3459, "step": 107220 }, { "epoch": 107.25, "grad_norm": 56.006126403808594, "learning_rate": 1.873786718321476e-06, "loss": 0.3252, "step": 107250 }, { "epoch": 107.28, "grad_norm": 53.25779724121094, "learning_rate": 1.8713355411382117e-06, "loss": 0.3138, "step": 107280 }, { "epoch": 107.31, "grad_norm": 109.7735824584961, "learning_rate": 1.8688855991021272e-06, "loss": 0.3223, "step": 107310 }, { "epoch": 107.34, "grad_norm": 90.90961456298828, "learning_rate": 1.8664368931804211e-06, "loss": 0.3313, "step": 107340 }, { "epoch": 107.37, "grad_norm": 64.66145324707031, "learning_rate": 1.8639894243398055e-06, "loss": 0.2391, "step": 107370 }, { "epoch": 107.4, "grad_norm": 45.553070068359375, "learning_rate": 1.8615431935464984e-06, "loss": 0.2899, "step": 107400 }, { "epoch": 107.43, "grad_norm": 131.3832550048828, "learning_rate": 1.8590982017662362e-06, "loss": 0.3778, "step": 107430 }, { "epoch": 107.46, "grad_norm": 132.2882843017578, "learning_rate": 1.8566544499642587e-06, "loss": 0.3506, "step": 107460 }, { "epoch": 107.49, "grad_norm": 128.610595703125, "learning_rate": 1.854211939105327e-06, "loss": 0.2857, "step": 107490 }, { "epoch": 107.52, "grad_norm": 55.774070739746094, "learning_rate": 1.8517706701536998e-06, "loss": 0.3453, "step": 107520 }, { "epoch": 107.55, "grad_norm": 108.1532974243164, "learning_rate": 1.8493306440731557e-06, "loss": 0.3417, "step": 107550 }, { "epoch": 107.58, "grad_norm": 42.774131774902344, "learning_rate": 1.8468918618269749e-06, "loss": 0.2983, "step": 107580 }, { "epoch": 107.61, "grad_norm": 91.56114196777344, "learning_rate": 1.8444543243779512e-06, "loss": 0.3016, "step": 107610 }, { "epoch": 107.64, "grad_norm": 106.45355987548828, "learning_rate": 1.8420180326883857e-06, "loss": 0.2982, "step": 107640 }, { "epoch": 107.67, "grad_norm": 37.1694450378418, "learning_rate": 1.8395829877200904e-06, "loss": 0.4032, "step": 107670 }, { "epoch": 107.7, "grad_norm": 95.86736297607422, "learning_rate": 1.837149190434378e-06, "loss": 0.3667, "step": 107700 }, { "epoch": 107.73, "grad_norm": 63.11392593383789, "learning_rate": 1.8347166417920776e-06, "loss": 0.2517, "step": 107730 }, { "epoch": 107.76, "grad_norm": 139.97535705566406, "learning_rate": 1.8322853427535148e-06, "loss": 0.3557, "step": 107760 }, { "epoch": 107.79, "grad_norm": 87.70382690429688, "learning_rate": 1.8298552942785352e-06, "loss": 0.3013, "step": 107790 }, { "epoch": 107.82, "grad_norm": 71.18769073486328, "learning_rate": 1.8274264973264782e-06, "loss": 0.2942, "step": 107820 }, { "epoch": 107.85, "grad_norm": 105.11351013183594, "learning_rate": 1.824998952856198e-06, "loss": 0.2926, "step": 107850 }, { "epoch": 107.88, "grad_norm": 58.507572174072266, "learning_rate": 1.822572661826047e-06, "loss": 0.3244, "step": 107880 }, { "epoch": 107.91, "grad_norm": 50.747554779052734, "learning_rate": 1.8201476251938888e-06, "loss": 0.3169, "step": 107910 }, { "epoch": 107.94, "grad_norm": 80.7756576538086, "learning_rate": 1.8177238439170885e-06, "loss": 0.2872, "step": 107940 }, { "epoch": 107.97, "grad_norm": 99.96867370605469, "learning_rate": 1.8153013189525192e-06, "loss": 0.269, "step": 107970 }, { "epoch": 108.0, "grad_norm": 86.83390808105469, "learning_rate": 1.8128800512565514e-06, "loss": 0.3682, "step": 108000 }, { "epoch": 108.0, "eval_loss": 0.24401062726974487, "eval_map": 0.8324, "eval_map_50": 0.9624, "eval_map_75": 0.9341, "eval_map_chicken": 0.8208, "eval_map_duck": 0.7831, "eval_map_large": 0.8428, "eval_map_medium": 0.8311, "eval_map_plant": 0.8932, "eval_map_small": 0.3185, "eval_mar_1": 0.3348, "eval_mar_10": 0.8644, "eval_mar_100": 0.8681, "eval_mar_100_chicken": 0.8648, "eval_mar_100_duck": 0.8227, "eval_mar_100_plant": 0.9167, "eval_mar_large": 0.8751, "eval_mar_medium": 0.8697, "eval_mar_small": 0.404, "eval_runtime": 13.0636, "eval_samples_per_second": 15.31, "eval_steps_per_second": 1.914, "step": 108000 }, { "epoch": 108.03, "grad_norm": 32.29725646972656, "learning_rate": 1.810460041785067e-06, "loss": 0.34, "step": 108030 }, { "epoch": 108.06, "grad_norm": 58.399410247802734, "learning_rate": 1.8080412914934436e-06, "loss": 0.347, "step": 108060 }, { "epoch": 108.09, "grad_norm": 89.20189666748047, "learning_rate": 1.8056238013365679e-06, "loss": 0.2849, "step": 108090 }, { "epoch": 108.12, "grad_norm": 56.73651885986328, "learning_rate": 1.803207572268826e-06, "loss": 0.3524, "step": 108120 }, { "epoch": 108.15, "grad_norm": 37.779911041259766, "learning_rate": 1.800792605244109e-06, "loss": 0.3608, "step": 108150 }, { "epoch": 108.18, "grad_norm": 41.75631332397461, "learning_rate": 1.7983789012158037e-06, "loss": 0.3795, "step": 108180 }, { "epoch": 108.21, "grad_norm": 55.996253967285156, "learning_rate": 1.7959664611368043e-06, "loss": 0.2726, "step": 108210 }, { "epoch": 108.24, "grad_norm": 54.929996490478516, "learning_rate": 1.7935552859595058e-06, "loss": 0.3085, "step": 108240 }, { "epoch": 108.27, "grad_norm": 60.716922760009766, "learning_rate": 1.7911453766357983e-06, "loss": 0.305, "step": 108270 }, { "epoch": 108.3, "grad_norm": 34.98912048339844, "learning_rate": 1.7887367341170781e-06, "loss": 0.2892, "step": 108300 }, { "epoch": 108.33, "grad_norm": 3.5572964407037944e-05, "learning_rate": 1.7863293593542402e-06, "loss": 0.254, "step": 108330 }, { "epoch": 108.36, "grad_norm": 157.20236206054688, "learning_rate": 1.7839232532976746e-06, "loss": 0.3757, "step": 108360 }, { "epoch": 108.39, "grad_norm": 43.239540100097656, "learning_rate": 1.781518416897276e-06, "loss": 0.3189, "step": 108390 }, { "epoch": 108.42, "grad_norm": 70.89527893066406, "learning_rate": 1.779114851102437e-06, "loss": 0.2754, "step": 108420 }, { "epoch": 108.45, "grad_norm": 51.442359924316406, "learning_rate": 1.7767125568620442e-06, "loss": 0.3187, "step": 108450 }, { "epoch": 108.48, "grad_norm": 53.7587890625, "learning_rate": 1.7743115351244883e-06, "loss": 0.3929, "step": 108480 }, { "epoch": 108.51, "grad_norm": 40.34379196166992, "learning_rate": 1.77191178683765e-06, "loss": 0.3149, "step": 108510 }, { "epoch": 108.54, "grad_norm": 57.369510650634766, "learning_rate": 1.769513312948919e-06, "loss": 0.3142, "step": 108540 }, { "epoch": 108.57, "grad_norm": 89.70407104492188, "learning_rate": 1.767116114405169e-06, "loss": 0.3763, "step": 108570 }, { "epoch": 108.6, "grad_norm": 64.8481216430664, "learning_rate": 1.7647201921527802e-06, "loss": 0.3013, "step": 108600 }, { "epoch": 108.63, "grad_norm": 27.814350128173828, "learning_rate": 1.7623255471376199e-06, "loss": 0.271, "step": 108630 }, { "epoch": 108.66, "grad_norm": 69.15274810791016, "learning_rate": 1.7599321803050595e-06, "loss": 0.2784, "step": 108660 }, { "epoch": 108.69, "grad_norm": 51.49317169189453, "learning_rate": 1.7575400925999613e-06, "loss": 0.3287, "step": 108690 }, { "epoch": 108.72, "grad_norm": 90.85767364501953, "learning_rate": 1.7551492849666857e-06, "loss": 0.2815, "step": 108720 }, { "epoch": 108.75, "grad_norm": 32.34882736206055, "learning_rate": 1.7527597583490825e-06, "loss": 0.2931, "step": 108750 }, { "epoch": 108.78, "grad_norm": 33.53123092651367, "learning_rate": 1.7503715136905014e-06, "loss": 0.279, "step": 108780 }, { "epoch": 108.81, "grad_norm": 64.9784164428711, "learning_rate": 1.7479845519337795e-06, "loss": 0.3057, "step": 108810 }, { "epoch": 108.84, "grad_norm": 27.3458251953125, "learning_rate": 1.7455988740212576e-06, "loss": 0.3574, "step": 108840 }, { "epoch": 108.87, "grad_norm": 309.8020935058594, "learning_rate": 1.7432144808947587e-06, "loss": 0.3436, "step": 108870 }, { "epoch": 108.9, "grad_norm": 78.69480895996094, "learning_rate": 1.7408313734956074e-06, "loss": 0.277, "step": 108900 }, { "epoch": 108.93, "grad_norm": 90.77513122558594, "learning_rate": 1.7384495527646127e-06, "loss": 0.2566, "step": 108930 }, { "epoch": 108.96, "grad_norm": 5.7214998378185555e-05, "learning_rate": 1.7360690196420816e-06, "loss": 0.2879, "step": 108960 }, { "epoch": 108.99, "grad_norm": 68.66263580322266, "learning_rate": 1.7336897750678106e-06, "loss": 0.3234, "step": 108990 }, { "epoch": 109.0, "eval_loss": 0.25315338373184204, "eval_map": 0.8224, "eval_map_50": 0.9605, "eval_map_75": 0.9338, "eval_map_chicken": 0.8134, "eval_map_duck": 0.7699, "eval_map_large": 0.8253, "eval_map_medium": 0.8203, "eval_map_plant": 0.884, "eval_map_small": 0.3491, "eval_mar_1": 0.3316, "eval_mar_10": 0.8569, "eval_mar_100": 0.8609, "eval_mar_100_chicken": 0.8622, "eval_mar_100_duck": 0.8108, "eval_mar_100_plant": 0.9098, "eval_mar_large": 0.8643, "eval_mar_medium": 0.8603, "eval_mar_small": 0.4301, "eval_runtime": 12.568, "eval_samples_per_second": 15.913, "eval_steps_per_second": 1.989, "step": 109000 }, { "epoch": 109.02, "grad_norm": 34.28153991699219, "learning_rate": 1.7313118199810897e-06, "loss": 0.3332, "step": 109020 }, { "epoch": 109.05, "grad_norm": 46.55137634277344, "learning_rate": 1.7289351553206952e-06, "loss": 0.3039, "step": 109050 }, { "epoch": 109.08, "grad_norm": 46.402557373046875, "learning_rate": 1.7265597820248987e-06, "loss": 0.341, "step": 109080 }, { "epoch": 109.11, "grad_norm": 42.930049896240234, "learning_rate": 1.7241857010314555e-06, "loss": 0.3541, "step": 109110 }, { "epoch": 109.14, "grad_norm": 76.71400451660156, "learning_rate": 1.7218129132776224e-06, "loss": 0.3471, "step": 109140 }, { "epoch": 109.17, "grad_norm": 71.37537384033203, "learning_rate": 1.719441419700133e-06, "loss": 0.2756, "step": 109170 }, { "epoch": 109.2, "grad_norm": 71.17495727539062, "learning_rate": 1.7170712212352187e-06, "loss": 0.3127, "step": 109200 }, { "epoch": 109.23, "grad_norm": 27.208005905151367, "learning_rate": 1.714702318818593e-06, "loss": 0.32, "step": 109230 }, { "epoch": 109.26, "grad_norm": 29.75830078125, "learning_rate": 1.7123347133854628e-06, "loss": 0.3132, "step": 109260 }, { "epoch": 109.29, "grad_norm": 75.63294982910156, "learning_rate": 1.7099684058705212e-06, "loss": 0.3299, "step": 109290 }, { "epoch": 109.32, "grad_norm": 82.87824249267578, "learning_rate": 1.7076033972079503e-06, "loss": 0.3024, "step": 109320 }, { "epoch": 109.35, "grad_norm": 44.17387008666992, "learning_rate": 1.7052396883314154e-06, "loss": 0.315, "step": 109350 }, { "epoch": 109.38, "grad_norm": 45.85847473144531, "learning_rate": 1.702877280174074e-06, "loss": 0.3052, "step": 109380 }, { "epoch": 109.41, "grad_norm": 37.659217834472656, "learning_rate": 1.700516173668565e-06, "loss": 0.5055, "step": 109410 }, { "epoch": 109.44, "grad_norm": 67.85224914550781, "learning_rate": 1.698156369747016e-06, "loss": 0.3147, "step": 109440 }, { "epoch": 109.47, "grad_norm": 39.72770690917969, "learning_rate": 1.6957978693410414e-06, "loss": 0.3163, "step": 109470 }, { "epoch": 109.5, "grad_norm": 48.38703155517578, "learning_rate": 1.6934406733817417e-06, "loss": 0.3395, "step": 109500 }, { "epoch": 109.53, "grad_norm": 50.6041374206543, "learning_rate": 1.6910847827996961e-06, "loss": 0.2886, "step": 109530 }, { "epoch": 109.56, "grad_norm": 35.03040313720703, "learning_rate": 1.6887301985249754e-06, "loss": 0.2544, "step": 109560 }, { "epoch": 109.59, "grad_norm": 36.23527908325195, "learning_rate": 1.6863769214871334e-06, "loss": 0.3576, "step": 109590 }, { "epoch": 109.62, "grad_norm": 82.93131256103516, "learning_rate": 1.6840249526152036e-06, "loss": 0.332, "step": 109620 }, { "epoch": 109.65, "grad_norm": 100.89730072021484, "learning_rate": 1.6816742928377072e-06, "loss": 0.3221, "step": 109650 }, { "epoch": 109.68, "grad_norm": 64.72860717773438, "learning_rate": 1.6793249430826502e-06, "loss": 0.2594, "step": 109680 }, { "epoch": 109.71, "grad_norm": 43.99069595336914, "learning_rate": 1.6769769042775141e-06, "loss": 0.2811, "step": 109710 }, { "epoch": 109.74, "grad_norm": 74.57102966308594, "learning_rate": 1.6746301773492701e-06, "loss": 0.2845, "step": 109740 }, { "epoch": 109.77, "grad_norm": 65.6674575805664, "learning_rate": 1.6722847632243699e-06, "loss": 0.2772, "step": 109770 }, { "epoch": 109.8, "grad_norm": 59.639739990234375, "learning_rate": 1.6699406628287423e-06, "loss": 0.3326, "step": 109800 }, { "epoch": 109.83, "grad_norm": 57.53010559082031, "learning_rate": 1.6675978770878055e-06, "loss": 0.2938, "step": 109830 }, { "epoch": 109.86, "grad_norm": 55.99067687988281, "learning_rate": 1.6652564069264476e-06, "loss": 0.315, "step": 109860 }, { "epoch": 109.89, "grad_norm": 30.19736671447754, "learning_rate": 1.6629162532690517e-06, "loss": 0.3294, "step": 109890 }, { "epoch": 109.92, "grad_norm": 111.66912078857422, "learning_rate": 1.6605774170394683e-06, "loss": 0.5145, "step": 109920 }, { "epoch": 109.95, "grad_norm": 24.310741424560547, "learning_rate": 1.658239899161036e-06, "loss": 0.262, "step": 109950 }, { "epoch": 109.98, "grad_norm": 64.34194946289062, "learning_rate": 1.6559037005565665e-06, "loss": 0.3412, "step": 109980 }, { "epoch": 110.0, "eval_loss": 0.24004921317100525, "eval_map": 0.8375, "eval_map_50": 0.966, "eval_map_75": 0.9346, "eval_map_chicken": 0.829, "eval_map_duck": 0.7912, "eval_map_large": 0.8387, "eval_map_medium": 0.8398, "eval_map_plant": 0.8923, "eval_map_small": 0.3187, "eval_mar_1": 0.3368, "eval_mar_10": 0.8692, "eval_mar_100": 0.8733, "eval_mar_100_chicken": 0.8722, "eval_mar_100_duck": 0.8309, "eval_mar_100_plant": 0.9167, "eval_mar_large": 0.8725, "eval_mar_medium": 0.878, "eval_mar_small": 0.4148, "eval_runtime": 12.62, "eval_samples_per_second": 15.848, "eval_steps_per_second": 1.981, "step": 110000 }, { "epoch": 110.01, "grad_norm": 36.40325927734375, "learning_rate": 1.6535688221483558e-06, "loss": 0.3147, "step": 110010 }, { "epoch": 110.04, "grad_norm": 60.60414123535156, "learning_rate": 1.651235264858177e-06, "loss": 0.3513, "step": 110040 }, { "epoch": 110.07, "grad_norm": 62.319000244140625, "learning_rate": 1.6489030296072833e-06, "loss": 0.2867, "step": 110070 }, { "epoch": 110.1, "grad_norm": 51.38201141357422, "learning_rate": 1.6465721173164e-06, "loss": 0.3468, "step": 110100 }, { "epoch": 110.13, "grad_norm": 67.61336517333984, "learning_rate": 1.644242528905739e-06, "loss": 0.343, "step": 110130 }, { "epoch": 110.16, "grad_norm": 43.806617736816406, "learning_rate": 1.6419142652949793e-06, "loss": 0.2926, "step": 110160 }, { "epoch": 110.19, "grad_norm": 120.5438232421875, "learning_rate": 1.6395873274032887e-06, "loss": 0.3238, "step": 110190 }, { "epoch": 110.22, "grad_norm": 98.48339080810547, "learning_rate": 1.6372617161493014e-06, "loss": 0.3293, "step": 110220 }, { "epoch": 110.25, "grad_norm": 30.474123001098633, "learning_rate": 1.6349374324511347e-06, "loss": 0.3108, "step": 110250 }, { "epoch": 110.28, "grad_norm": 259.0688171386719, "learning_rate": 1.6326144772263752e-06, "loss": 0.3717, "step": 110280 }, { "epoch": 110.31, "grad_norm": 102.05205535888672, "learning_rate": 1.6302928513920912e-06, "loss": 0.2929, "step": 110310 }, { "epoch": 110.34, "grad_norm": 40.07794952392578, "learning_rate": 1.627972555864824e-06, "loss": 0.3012, "step": 110340 }, { "epoch": 110.37, "grad_norm": 46.2303466796875, "learning_rate": 1.6256535915605904e-06, "loss": 0.3161, "step": 110370 }, { "epoch": 110.4, "grad_norm": 53.52840805053711, "learning_rate": 1.6233359593948777e-06, "loss": 0.3523, "step": 110400 }, { "epoch": 110.43, "grad_norm": 43.13352966308594, "learning_rate": 1.6210196602826544e-06, "loss": 0.3057, "step": 110430 }, { "epoch": 110.46, "grad_norm": 40.59793472290039, "learning_rate": 1.618704695138353e-06, "loss": 0.3638, "step": 110460 }, { "epoch": 110.49, "grad_norm": 44.563777923583984, "learning_rate": 1.6163910648758919e-06, "loss": 0.3996, "step": 110490 }, { "epoch": 110.52, "grad_norm": 33.67497634887695, "learning_rate": 1.6140787704086502e-06, "loss": 0.309, "step": 110520 }, { "epoch": 110.55, "grad_norm": 85.30017852783203, "learning_rate": 1.6117678126494895e-06, "loss": 0.3548, "step": 110550 }, { "epoch": 110.58, "grad_norm": 72.46063232421875, "learning_rate": 1.6094581925107356e-06, "loss": 0.3384, "step": 110580 }, { "epoch": 110.61, "grad_norm": 95.61538696289062, "learning_rate": 1.607149910904191e-06, "loss": 0.2903, "step": 110610 }, { "epoch": 110.64, "grad_norm": 40.76462936401367, "learning_rate": 1.6048429687411294e-06, "loss": 0.2687, "step": 110640 }, { "epoch": 110.67, "grad_norm": 51.01396560668945, "learning_rate": 1.6025373669322963e-06, "loss": 0.4053, "step": 110670 }, { "epoch": 110.7, "grad_norm": 193.05052185058594, "learning_rate": 1.600233106387904e-06, "loss": 0.354, "step": 110700 }, { "epoch": 110.73, "grad_norm": 91.88577270507812, "learning_rate": 1.5979301880176407e-06, "loss": 0.2925, "step": 110730 }, { "epoch": 110.76, "grad_norm": 71.4330062866211, "learning_rate": 1.5956286127306591e-06, "loss": 0.3246, "step": 110760 }, { "epoch": 110.79, "grad_norm": 35.81876754760742, "learning_rate": 1.5933283814355871e-06, "loss": 0.3443, "step": 110790 }, { "epoch": 110.82, "grad_norm": 89.28417205810547, "learning_rate": 1.591029495040518e-06, "loss": 0.3094, "step": 110820 }, { "epoch": 110.85, "grad_norm": 71.15819549560547, "learning_rate": 1.5887319544530182e-06, "loss": 0.2735, "step": 110850 }, { "epoch": 110.88, "grad_norm": 47.88679504394531, "learning_rate": 1.586435760580118e-06, "loss": 0.2807, "step": 110880 }, { "epoch": 110.91, "grad_norm": 7.304209430003539e-05, "learning_rate": 1.58414091432832e-06, "loss": 0.2748, "step": 110910 }, { "epoch": 110.94, "grad_norm": 74.33480834960938, "learning_rate": 1.5818474166035907e-06, "loss": 0.3502, "step": 110940 }, { "epoch": 110.97, "grad_norm": 44.882591247558594, "learning_rate": 1.5795552683113679e-06, "loss": 0.2738, "step": 110970 }, { "epoch": 111.0, "grad_norm": 81.43742370605469, "learning_rate": 1.5772644703565564e-06, "loss": 0.4866, "step": 111000 }, { "epoch": 111.0, "eval_loss": 0.2558477818965912, "eval_map": 0.8248, "eval_map_50": 0.9649, "eval_map_75": 0.9282, "eval_map_chicken": 0.8161, "eval_map_duck": 0.7728, "eval_map_large": 0.8365, "eval_map_medium": 0.8222, "eval_map_plant": 0.8855, "eval_map_small": 0.3024, "eval_mar_1": 0.3305, "eval_mar_10": 0.8537, "eval_mar_100": 0.8579, "eval_mar_100_chicken": 0.8596, "eval_mar_100_duck": 0.8052, "eval_mar_100_plant": 0.9089, "eval_mar_large": 0.8692, "eval_mar_medium": 0.858, "eval_mar_small": 0.3818, "eval_runtime": 12.9357, "eval_samples_per_second": 15.461, "eval_steps_per_second": 1.933, "step": 111000 }, { "epoch": 111.03, "grad_norm": 60.771053314208984, "learning_rate": 1.5749750236435279e-06, "loss": 0.2951, "step": 111030 }, { "epoch": 111.06, "grad_norm": 47.95699691772461, "learning_rate": 1.572686929076116e-06, "loss": 0.3532, "step": 111060 }, { "epoch": 111.09, "grad_norm": 55.603973388671875, "learning_rate": 1.570400187557627e-06, "loss": 0.3171, "step": 111090 }, { "epoch": 111.12, "grad_norm": 61.563411712646484, "learning_rate": 1.5681147999908308e-06, "loss": 0.2389, "step": 111120 }, { "epoch": 111.15, "grad_norm": 39.736122131347656, "learning_rate": 1.5658307672779594e-06, "loss": 0.3812, "step": 111150 }, { "epoch": 111.18, "grad_norm": 50.103450775146484, "learning_rate": 1.5635480903207139e-06, "loss": 0.2959, "step": 111180 }, { "epoch": 111.21, "grad_norm": 34.244468688964844, "learning_rate": 1.5612667700202616e-06, "loss": 0.3078, "step": 111210 }, { "epoch": 111.24, "grad_norm": 103.32115173339844, "learning_rate": 1.5589868072772279e-06, "loss": 0.3083, "step": 111240 }, { "epoch": 111.27, "grad_norm": 58.7523193359375, "learning_rate": 1.5567082029917074e-06, "loss": 0.2726, "step": 111270 }, { "epoch": 111.3, "grad_norm": 202.6095733642578, "learning_rate": 1.554430958063259e-06, "loss": 0.285, "step": 111300 }, { "epoch": 111.33, "grad_norm": 81.46269989013672, "learning_rate": 1.552155073390899e-06, "loss": 0.3709, "step": 111330 }, { "epoch": 111.36, "grad_norm": 168.8040313720703, "learning_rate": 1.5498805498731146e-06, "loss": 0.2495, "step": 111360 }, { "epoch": 111.39, "grad_norm": 58.93038558959961, "learning_rate": 1.5476073884078463e-06, "loss": 0.2735, "step": 111390 }, { "epoch": 111.42, "grad_norm": 32.826995849609375, "learning_rate": 1.5453355898925094e-06, "loss": 0.2845, "step": 111420 }, { "epoch": 111.45, "grad_norm": 80.41813659667969, "learning_rate": 1.5430651552239684e-06, "loss": 0.3241, "step": 111450 }, { "epoch": 111.48, "grad_norm": 52.441078186035156, "learning_rate": 1.5407960852985582e-06, "loss": 0.2561, "step": 111480 }, { "epoch": 111.51, "grad_norm": 86.9748306274414, "learning_rate": 1.5385283810120688e-06, "loss": 0.2846, "step": 111510 }, { "epoch": 111.54, "grad_norm": 164.47743225097656, "learning_rate": 1.5362620432597559e-06, "loss": 0.3208, "step": 111540 }, { "epoch": 111.57, "grad_norm": 54.85179901123047, "learning_rate": 1.533997072936333e-06, "loss": 0.2958, "step": 111570 }, { "epoch": 111.6, "grad_norm": 36.446346282958984, "learning_rate": 1.531733470935976e-06, "loss": 0.3234, "step": 111600 }, { "epoch": 111.63, "grad_norm": 55.29444122314453, "learning_rate": 1.5294712381523168e-06, "loss": 0.3248, "step": 111630 }, { "epoch": 111.66, "grad_norm": 119.53633117675781, "learning_rate": 1.5272103754784517e-06, "loss": 0.4102, "step": 111660 }, { "epoch": 111.69, "grad_norm": 45.40870666503906, "learning_rate": 1.5249508838069287e-06, "loss": 0.3339, "step": 111690 }, { "epoch": 111.72, "grad_norm": 77.22078704833984, "learning_rate": 1.5226927640297663e-06, "loss": 0.4043, "step": 111720 }, { "epoch": 111.75, "grad_norm": 51.50632095336914, "learning_rate": 1.5204360170384286e-06, "loss": 0.2681, "step": 111750 }, { "epoch": 111.78, "grad_norm": 45.92066955566406, "learning_rate": 1.5181806437238472e-06, "loss": 0.2878, "step": 111780 }, { "epoch": 111.81, "grad_norm": 34.85203552246094, "learning_rate": 1.5159266449764048e-06, "loss": 0.2494, "step": 111810 }, { "epoch": 111.84, "grad_norm": 590.0502319335938, "learning_rate": 1.5136740216859464e-06, "loss": 0.316, "step": 111840 }, { "epoch": 111.87, "grad_norm": 26.550508499145508, "learning_rate": 1.511422774741771e-06, "loss": 0.2557, "step": 111870 }, { "epoch": 111.9, "grad_norm": 37.13541793823242, "learning_rate": 1.5091729050326376e-06, "loss": 0.3249, "step": 111900 }, { "epoch": 111.93, "grad_norm": 43.40486145019531, "learning_rate": 1.5069244134467553e-06, "loss": 0.3057, "step": 111930 }, { "epoch": 111.96, "grad_norm": 99.5147476196289, "learning_rate": 1.5046773008717968e-06, "loss": 0.266, "step": 111960 }, { "epoch": 111.99, "grad_norm": 55.345924377441406, "learning_rate": 1.5024315681948815e-06, "loss": 0.2781, "step": 111990 }, { "epoch": 112.0, "eval_loss": 0.24606649577617645, "eval_map": 0.8324, "eval_map_50": 0.9616, "eval_map_75": 0.9294, "eval_map_chicken": 0.8247, "eval_map_duck": 0.7794, "eval_map_large": 0.8408, "eval_map_medium": 0.8333, "eval_map_plant": 0.893, "eval_map_small": 0.3114, "eval_mar_1": 0.3349, "eval_mar_10": 0.8623, "eval_mar_100": 0.8659, "eval_mar_100_chicken": 0.8686, "eval_mar_100_duck": 0.8124, "eval_mar_100_plant": 0.9166, "eval_mar_large": 0.8706, "eval_mar_medium": 0.8679, "eval_mar_small": 0.3745, "eval_runtime": 12.7227, "eval_samples_per_second": 15.72, "eval_steps_per_second": 1.965, "step": 112000 }, { "epoch": 112.02, "grad_norm": 35.96322250366211, "learning_rate": 1.5001872163025954e-06, "loss": 0.2868, "step": 112020 }, { "epoch": 112.05, "grad_norm": 83.16947937011719, "learning_rate": 1.4979442460809684e-06, "loss": 0.2984, "step": 112050 }, { "epoch": 112.08, "grad_norm": 131.19912719726562, "learning_rate": 1.4957026584154926e-06, "loss": 0.3227, "step": 112080 }, { "epoch": 112.11, "grad_norm": 181.14096069335938, "learning_rate": 1.4934624541911086e-06, "loss": 0.2373, "step": 112110 }, { "epoch": 112.14, "grad_norm": 53.037715911865234, "learning_rate": 1.4912236342922143e-06, "loss": 0.3163, "step": 112140 }, { "epoch": 112.17, "grad_norm": 41.767093658447266, "learning_rate": 1.4889861996026617e-06, "loss": 0.2919, "step": 112170 }, { "epoch": 112.2, "grad_norm": 98.0342788696289, "learning_rate": 1.4867501510057548e-06, "loss": 0.3041, "step": 112200 }, { "epoch": 112.23, "grad_norm": 86.63310241699219, "learning_rate": 1.4845154893842473e-06, "loss": 0.2952, "step": 112230 }, { "epoch": 112.26, "grad_norm": 40.27363204956055, "learning_rate": 1.482282215620352e-06, "loss": 0.2918, "step": 112260 }, { "epoch": 112.29, "grad_norm": 67.10263061523438, "learning_rate": 1.4800503305957264e-06, "loss": 0.2713, "step": 112290 }, { "epoch": 112.32, "grad_norm": 393.9277648925781, "learning_rate": 1.4778198351914853e-06, "loss": 0.3121, "step": 112320 }, { "epoch": 112.35, "grad_norm": 39.3964729309082, "learning_rate": 1.4755907302881927e-06, "loss": 0.385, "step": 112350 }, { "epoch": 112.38, "grad_norm": 53.74249267578125, "learning_rate": 1.4733630167658652e-06, "loss": 0.3554, "step": 112380 }, { "epoch": 112.41, "grad_norm": 41.2756233215332, "learning_rate": 1.4711366955039664e-06, "loss": 0.3396, "step": 112410 }, { "epoch": 112.44, "grad_norm": 38.36640930175781, "learning_rate": 1.4689117673814135e-06, "loss": 0.2669, "step": 112440 }, { "epoch": 112.47, "grad_norm": 87.13140869140625, "learning_rate": 1.4666882332765747e-06, "loss": 0.3377, "step": 112470 }, { "epoch": 112.5, "grad_norm": 62.073848724365234, "learning_rate": 1.4644660940672628e-06, "loss": 0.2967, "step": 112500 }, { "epoch": 112.53, "grad_norm": 81.44989776611328, "learning_rate": 1.462245350630745e-06, "loss": 0.248, "step": 112530 }, { "epoch": 112.56, "grad_norm": 52.43824768066406, "learning_rate": 1.4600260038437376e-06, "loss": 0.3754, "step": 112560 }, { "epoch": 112.59, "grad_norm": 68.97999572753906, "learning_rate": 1.4578080545823991e-06, "loss": 0.3139, "step": 112590 }, { "epoch": 112.62, "grad_norm": 65.02044677734375, "learning_rate": 1.4555915037223438e-06, "loss": 0.2757, "step": 112620 }, { "epoch": 112.65, "grad_norm": 1160.73388671875, "learning_rate": 1.4533763521386319e-06, "loss": 0.2668, "step": 112650 }, { "epoch": 112.68, "grad_norm": 68.99584197998047, "learning_rate": 1.4511626007057667e-06, "loss": 0.3178, "step": 112680 }, { "epoch": 112.71, "grad_norm": 40.52623748779297, "learning_rate": 1.4489502502977037e-06, "loss": 0.279, "step": 112710 }, { "epoch": 112.74, "grad_norm": 0.0001034984816215001, "learning_rate": 1.4467393017878444e-06, "loss": 0.3348, "step": 112740 }, { "epoch": 112.77, "grad_norm": 38.7282600402832, "learning_rate": 1.4445297560490373e-06, "loss": 0.2844, "step": 112770 }, { "epoch": 112.8, "grad_norm": 62.98735046386719, "learning_rate": 1.4423216139535735e-06, "loss": 0.35, "step": 112800 }, { "epoch": 112.83, "grad_norm": 66.61173248291016, "learning_rate": 1.4401148763731953e-06, "loss": 0.2788, "step": 112830 }, { "epoch": 112.86, "grad_norm": 37.82440185546875, "learning_rate": 1.4379095441790847e-06, "loss": 0.3208, "step": 112860 }, { "epoch": 112.89, "grad_norm": 109.86479949951172, "learning_rate": 1.4357056182418727e-06, "loss": 0.2746, "step": 112890 }, { "epoch": 112.92, "grad_norm": 50.276817321777344, "learning_rate": 1.4335030994316357e-06, "loss": 0.3386, "step": 112920 }, { "epoch": 112.95, "grad_norm": 57.35239791870117, "learning_rate": 1.4313019886178942e-06, "loss": 0.3076, "step": 112950 }, { "epoch": 112.98, "grad_norm": 111.79991912841797, "learning_rate": 1.4291022866696086e-06, "loss": 0.3233, "step": 112980 }, { "epoch": 113.0, "eval_loss": 0.2467425912618637, "eval_map": 0.8333, "eval_map_50": 0.9634, "eval_map_75": 0.9328, "eval_map_chicken": 0.8201, "eval_map_duck": 0.7881, "eval_map_large": 0.8462, "eval_map_medium": 0.8308, "eval_map_plant": 0.8917, "eval_map_small": 0.3187, "eval_mar_1": 0.3342, "eval_mar_10": 0.865, "eval_mar_100": 0.8689, "eval_mar_100_chicken": 0.867, "eval_mar_100_duck": 0.8222, "eval_mar_100_plant": 0.9176, "eval_mar_large": 0.8788, "eval_mar_medium": 0.8696, "eval_mar_small": 0.4206, "eval_runtime": 12.7787, "eval_samples_per_second": 15.651, "eval_steps_per_second": 1.956, "step": 113000 }, { "epoch": 113.01, "grad_norm": 30.920276641845703, "learning_rate": 1.42690399445519e-06, "loss": 0.2925, "step": 113010 }, { "epoch": 113.04, "grad_norm": 49.06583786010742, "learning_rate": 1.4247071128424838e-06, "loss": 0.2764, "step": 113040 }, { "epoch": 113.07, "grad_norm": 109.119384765625, "learning_rate": 1.4225116426987916e-06, "loss": 0.3576, "step": 113070 }, { "epoch": 113.1, "grad_norm": 71.10040283203125, "learning_rate": 1.420317584890844e-06, "loss": 0.3122, "step": 113100 }, { "epoch": 113.13, "grad_norm": 54.722835540771484, "learning_rate": 1.4181249402848246e-06, "loss": 0.2687, "step": 113130 }, { "epoch": 113.16, "grad_norm": 64.2225112915039, "learning_rate": 1.4159337097463515e-06, "loss": 0.299, "step": 113160 }, { "epoch": 113.19, "grad_norm": 65.18496704101562, "learning_rate": 1.413743894140489e-06, "loss": 0.271, "step": 113190 }, { "epoch": 113.22, "grad_norm": 367.8233642578125, "learning_rate": 1.4115554943317416e-06, "loss": 0.3889, "step": 113220 }, { "epoch": 113.25, "grad_norm": 72.89109802246094, "learning_rate": 1.4093685111840567e-06, "loss": 0.3695, "step": 113250 }, { "epoch": 113.28, "grad_norm": 53.926483154296875, "learning_rate": 1.407182945560817e-06, "loss": 0.3647, "step": 113280 }, { "epoch": 113.31, "grad_norm": 52.478050231933594, "learning_rate": 1.404998798324853e-06, "loss": 0.2729, "step": 113310 }, { "epoch": 113.34, "grad_norm": 96.05762481689453, "learning_rate": 1.402816070338427e-06, "loss": 0.2856, "step": 113340 }, { "epoch": 113.37, "grad_norm": 89.84333801269531, "learning_rate": 1.4006347624632505e-06, "loss": 0.3246, "step": 113370 }, { "epoch": 113.4, "grad_norm": 86.69061279296875, "learning_rate": 1.3984548755604655e-06, "loss": 0.311, "step": 113400 }, { "epoch": 113.43, "grad_norm": 63.412357330322266, "learning_rate": 1.3962764104906596e-06, "loss": 0.2874, "step": 113430 }, { "epoch": 113.46, "grad_norm": 32.9383430480957, "learning_rate": 1.3940993681138533e-06, "loss": 0.3062, "step": 113460 }, { "epoch": 113.49, "grad_norm": 44.63263702392578, "learning_rate": 1.39192374928951e-06, "loss": 0.334, "step": 113490 }, { "epoch": 113.52, "grad_norm": 64.07804107666016, "learning_rate": 1.38974955487653e-06, "loss": 0.3063, "step": 113520 }, { "epoch": 113.55, "grad_norm": 40.1292610168457, "learning_rate": 1.3875767857332512e-06, "loss": 0.3753, "step": 113550 }, { "epoch": 113.58, "grad_norm": 44.38451385498047, "learning_rate": 1.3854054427174468e-06, "loss": 0.335, "step": 113580 }, { "epoch": 113.61, "grad_norm": 69.53759765625, "learning_rate": 1.3832355266863307e-06, "loss": 0.2558, "step": 113610 }, { "epoch": 113.64, "grad_norm": 78.4531478881836, "learning_rate": 1.3810670384965469e-06, "loss": 0.2971, "step": 113640 }, { "epoch": 113.67, "grad_norm": 42.34751892089844, "learning_rate": 1.3788999790041867e-06, "loss": 0.5052, "step": 113670 }, { "epoch": 113.7, "grad_norm": 39.1889533996582, "learning_rate": 1.3767343490647668e-06, "loss": 0.3076, "step": 113700 }, { "epoch": 113.73, "grad_norm": 514.1843872070312, "learning_rate": 1.3745701495332447e-06, "loss": 0.3016, "step": 113730 }, { "epoch": 113.76, "grad_norm": 46.99753189086914, "learning_rate": 1.372407381264011e-06, "loss": 0.2562, "step": 113760 }, { "epoch": 113.79, "grad_norm": 48.368412017822266, "learning_rate": 1.3702460451108934e-06, "loss": 0.3005, "step": 113790 }, { "epoch": 113.82, "grad_norm": 50.54788589477539, "learning_rate": 1.368086141927154e-06, "loss": 0.2568, "step": 113820 }, { "epoch": 113.85, "grad_norm": 29.061279296875, "learning_rate": 1.3659276725654863e-06, "loss": 0.2994, "step": 113850 }, { "epoch": 113.88, "grad_norm": 36.15702819824219, "learning_rate": 1.3637706378780209e-06, "loss": 0.3201, "step": 113880 }, { "epoch": 113.91, "grad_norm": 322.1993713378906, "learning_rate": 1.361615038716322e-06, "loss": 0.3048, "step": 113910 }, { "epoch": 113.94, "grad_norm": 145.29837036132812, "learning_rate": 1.3594608759313832e-06, "loss": 0.284, "step": 113940 }, { "epoch": 113.97, "grad_norm": 90.51874542236328, "learning_rate": 1.3573081503736362e-06, "loss": 0.3144, "step": 113970 }, { "epoch": 114.0, "grad_norm": 149.41188049316406, "learning_rate": 1.3551568628929434e-06, "loss": 0.2915, "step": 114000 }, { "epoch": 114.0, "eval_loss": 0.23929466307163239, "eval_map": 0.8366, "eval_map_50": 0.9605, "eval_map_75": 0.9276, "eval_map_chicken": 0.8273, "eval_map_duck": 0.79, "eval_map_large": 0.8412, "eval_map_medium": 0.8387, "eval_map_plant": 0.8925, "eval_map_small": 0.3392, "eval_mar_1": 0.336, "eval_mar_10": 0.8677, "eval_mar_100": 0.8707, "eval_mar_100_chicken": 0.8734, "eval_mar_100_duck": 0.8216, "eval_mar_100_plant": 0.917, "eval_mar_large": 0.8747, "eval_mar_medium": 0.8751, "eval_mar_small": 0.3934, "eval_runtime": 13.3384, "eval_samples_per_second": 14.994, "eval_steps_per_second": 1.874, "step": 114000 }, { "epoch": 114.03, "grad_norm": 124.3449935913086, "learning_rate": 1.3530070143385966e-06, "loss": 0.2961, "step": 114030 }, { "epoch": 114.06, "grad_norm": 35.864681243896484, "learning_rate": 1.350858605559323e-06, "loss": 0.3684, "step": 114060 }, { "epoch": 114.09, "grad_norm": 60.432472229003906, "learning_rate": 1.3487116374032811e-06, "loss": 0.2642, "step": 114090 }, { "epoch": 114.12, "grad_norm": 68.29483795166016, "learning_rate": 1.346566110718061e-06, "loss": 0.2861, "step": 114120 }, { "epoch": 114.15, "grad_norm": 43.01601028442383, "learning_rate": 1.3444220263506797e-06, "loss": 0.2491, "step": 114150 }, { "epoch": 114.18, "grad_norm": 71.6007308959961, "learning_rate": 1.3422793851475907e-06, "loss": 0.335, "step": 114180 }, { "epoch": 114.21, "grad_norm": 34.29472351074219, "learning_rate": 1.3401381879546716e-06, "loss": 0.2985, "step": 114210 }, { "epoch": 114.24, "grad_norm": 43.90507125854492, "learning_rate": 1.337998435617235e-06, "loss": 0.3349, "step": 114240 }, { "epoch": 114.27, "grad_norm": 50.22809600830078, "learning_rate": 1.3358601289800211e-06, "loss": 0.2644, "step": 114270 }, { "epoch": 114.3, "grad_norm": 70.90999603271484, "learning_rate": 1.333723268887201e-06, "loss": 0.3364, "step": 114300 }, { "epoch": 114.33, "grad_norm": 37.243858337402344, "learning_rate": 1.3315878561823697e-06, "loss": 0.3111, "step": 114330 }, { "epoch": 114.36, "grad_norm": 34.69538116455078, "learning_rate": 1.3294538917085586e-06, "loss": 0.3178, "step": 114360 }, { "epoch": 114.39, "grad_norm": 72.23746490478516, "learning_rate": 1.3273213763082188e-06, "loss": 0.3273, "step": 114390 }, { "epoch": 114.42, "grad_norm": 49.03205871582031, "learning_rate": 1.3251903108232362e-06, "loss": 0.3016, "step": 114420 }, { "epoch": 114.45, "grad_norm": 39.62008285522461, "learning_rate": 1.3230606960949204e-06, "loss": 0.3107, "step": 114450 }, { "epoch": 114.48, "grad_norm": 62.427490234375, "learning_rate": 1.3209325329640126e-06, "loss": 0.2825, "step": 114480 }, { "epoch": 114.51, "grad_norm": 55.44679260253906, "learning_rate": 1.3188058222706735e-06, "loss": 0.3087, "step": 114510 }, { "epoch": 114.54, "grad_norm": 97.2188491821289, "learning_rate": 1.316680564854499e-06, "loss": 0.3142, "step": 114540 }, { "epoch": 114.57, "grad_norm": 47.15496826171875, "learning_rate": 1.3145567615545013e-06, "loss": 0.3254, "step": 114570 }, { "epoch": 114.6, "grad_norm": 107.84281921386719, "learning_rate": 1.312434413209131e-06, "loss": 0.2698, "step": 114600 }, { "epoch": 114.63, "grad_norm": 65.63544464111328, "learning_rate": 1.3103135206562535e-06, "loss": 0.3101, "step": 114630 }, { "epoch": 114.66, "grad_norm": 67.24834442138672, "learning_rate": 1.3081940847331658e-06, "loss": 0.3419, "step": 114660 }, { "epoch": 114.69, "grad_norm": 5.6102009693859145e-05, "learning_rate": 1.3060761062765853e-06, "loss": 0.2229, "step": 114690 }, { "epoch": 114.72, "grad_norm": 76.00422668457031, "learning_rate": 1.3039595861226579e-06, "loss": 0.29, "step": 114720 }, { "epoch": 114.75, "grad_norm": 73.79566955566406, "learning_rate": 1.301844525106951e-06, "loss": 0.3473, "step": 114750 }, { "epoch": 114.78, "grad_norm": 43.69677734375, "learning_rate": 1.2997309240644607e-06, "loss": 0.3287, "step": 114780 }, { "epoch": 114.81, "grad_norm": 73.95443725585938, "learning_rate": 1.2976187838295984e-06, "loss": 0.3031, "step": 114810 }, { "epoch": 114.84, "grad_norm": 32.22563171386719, "learning_rate": 1.2955081052362072e-06, "loss": 0.337, "step": 114840 }, { "epoch": 114.87, "grad_norm": 33.562255859375, "learning_rate": 1.2933988891175458e-06, "loss": 0.2736, "step": 114870 }, { "epoch": 114.9, "grad_norm": 55.682254791259766, "learning_rate": 1.2912911363063048e-06, "loss": 0.3011, "step": 114900 }, { "epoch": 114.93, "grad_norm": 56.983455657958984, "learning_rate": 1.2891848476345864e-06, "loss": 0.323, "step": 114930 }, { "epoch": 114.96, "grad_norm": 125.78273010253906, "learning_rate": 1.2870800239339237e-06, "loss": 0.3636, "step": 114960 }, { "epoch": 114.99, "grad_norm": 45.95997619628906, "learning_rate": 1.2849766660352652e-06, "loss": 0.3298, "step": 114990 }, { "epoch": 115.0, "eval_loss": 0.24736888706684113, "eval_map": 0.8323, "eval_map_50": 0.9637, "eval_map_75": 0.9268, "eval_map_chicken": 0.8222, "eval_map_duck": 0.7779, "eval_map_large": 0.8287, "eval_map_medium": 0.8334, "eval_map_plant": 0.8966, "eval_map_small": 0.3252, "eval_mar_1": 0.3324, "eval_mar_10": 0.8629, "eval_mar_100": 0.8667, "eval_mar_100_chicken": 0.8672, "eval_mar_100_duck": 0.8129, "eval_mar_100_plant": 0.9199, "eval_mar_large": 0.8656, "eval_mar_medium": 0.8709, "eval_mar_small": 0.4067, "eval_runtime": 12.6448, "eval_samples_per_second": 15.817, "eval_steps_per_second": 1.977, "step": 115000 }, { "epoch": 115.02, "grad_norm": 104.96399688720703, "learning_rate": 1.2828747747689846e-06, "loss": 0.3291, "step": 115020 }, { "epoch": 115.05, "grad_norm": 38.06474304199219, "learning_rate": 1.2807743509648745e-06, "loss": 0.3928, "step": 115050 }, { "epoch": 115.08, "grad_norm": 36.52778244018555, "learning_rate": 1.2786753954521508e-06, "loss": 0.3688, "step": 115080 }, { "epoch": 115.11, "grad_norm": 235.37173461914062, "learning_rate": 1.2765779090594454e-06, "loss": 0.3083, "step": 115110 }, { "epoch": 115.14, "grad_norm": 56.57828140258789, "learning_rate": 1.2744818926148157e-06, "loss": 0.2581, "step": 115140 }, { "epoch": 115.17, "grad_norm": 44.17103958129883, "learning_rate": 1.2723873469457304e-06, "loss": 0.3582, "step": 115170 }, { "epoch": 115.2, "grad_norm": 77.74555969238281, "learning_rate": 1.2702942728790897e-06, "loss": 0.3503, "step": 115200 }, { "epoch": 115.23, "grad_norm": 70.21920013427734, "learning_rate": 1.2682026712412016e-06, "loss": 0.3137, "step": 115230 }, { "epoch": 115.26, "grad_norm": 70.18965911865234, "learning_rate": 1.2661125428577998e-06, "loss": 0.3609, "step": 115260 }, { "epoch": 115.29, "grad_norm": 104.90813446044922, "learning_rate": 1.2640238885540313e-06, "loss": 0.3816, "step": 115290 }, { "epoch": 115.32, "grad_norm": 151.09361267089844, "learning_rate": 1.2619367091544654e-06, "loss": 0.2998, "step": 115320 }, { "epoch": 115.35, "grad_norm": 111.99593353271484, "learning_rate": 1.2598510054830888e-06, "loss": 0.322, "step": 115350 }, { "epoch": 115.38, "grad_norm": 39.08504104614258, "learning_rate": 1.2577667783633007e-06, "loss": 0.3151, "step": 115380 }, { "epoch": 115.41, "grad_norm": 51.8145751953125, "learning_rate": 1.2556840286179234e-06, "loss": 0.4006, "step": 115410 }, { "epoch": 115.44, "grad_norm": 42.65520477294922, "learning_rate": 1.2536027570691938e-06, "loss": 0.2749, "step": 115440 }, { "epoch": 115.47, "grad_norm": 5.748760304413736e-05, "learning_rate": 1.2515229645387639e-06, "loss": 0.3786, "step": 115470 }, { "epoch": 115.5, "grad_norm": 48.603851318359375, "learning_rate": 1.2494446518477022e-06, "loss": 0.3125, "step": 115500 }, { "epoch": 115.53, "grad_norm": 302.6282958984375, "learning_rate": 1.2473678198164967e-06, "loss": 0.2999, "step": 115530 }, { "epoch": 115.56, "grad_norm": 49.433990478515625, "learning_rate": 1.2452924692650443e-06, "loss": 0.3022, "step": 115560 }, { "epoch": 115.59, "grad_norm": 33.62896728515625, "learning_rate": 1.2432186010126613e-06, "loss": 0.2858, "step": 115590 }, { "epoch": 115.62, "grad_norm": 44.40532302856445, "learning_rate": 1.2411462158780791e-06, "loss": 0.2857, "step": 115620 }, { "epoch": 115.65, "grad_norm": 33.839805603027344, "learning_rate": 1.2390753146794438e-06, "loss": 0.3682, "step": 115650 }, { "epoch": 115.68, "grad_norm": 43.377647399902344, "learning_rate": 1.2370058982343109e-06, "loss": 0.3085, "step": 115680 }, { "epoch": 115.71, "grad_norm": 45.93667221069336, "learning_rate": 1.2349379673596568e-06, "loss": 0.2774, "step": 115710 }, { "epoch": 115.74, "grad_norm": 43.49879455566406, "learning_rate": 1.232871522871864e-06, "loss": 0.3314, "step": 115740 }, { "epoch": 115.77, "grad_norm": 110.21099853515625, "learning_rate": 1.2308065655867346e-06, "loss": 0.3003, "step": 115770 }, { "epoch": 115.8, "grad_norm": 36.44283676147461, "learning_rate": 1.2287430963194807e-06, "loss": 0.2877, "step": 115800 }, { "epoch": 115.83, "grad_norm": 64.81216430664062, "learning_rate": 1.2266811158847285e-06, "loss": 0.2805, "step": 115830 }, { "epoch": 115.86, "grad_norm": 91.93071746826172, "learning_rate": 1.2246206250965127e-06, "loss": 0.3157, "step": 115860 }, { "epoch": 115.89, "grad_norm": 36.35700607299805, "learning_rate": 1.2225616247682848e-06, "loss": 0.3428, "step": 115890 }, { "epoch": 115.92, "grad_norm": 73.2782211303711, "learning_rate": 1.2205041157129017e-06, "loss": 0.3106, "step": 115920 }, { "epoch": 115.95, "grad_norm": 45.069480895996094, "learning_rate": 1.218448098742641e-06, "loss": 0.3076, "step": 115950 }, { "epoch": 115.98, "grad_norm": 68.000244140625, "learning_rate": 1.2163935746691807e-06, "loss": 0.3928, "step": 115980 }, { "epoch": 116.0, "eval_loss": 0.24245685338974, "eval_map": 0.8386, "eval_map_50": 0.9669, "eval_map_75": 0.9343, "eval_map_chicken": 0.8272, "eval_map_duck": 0.7927, "eval_map_large": 0.8459, "eval_map_medium": 0.8384, "eval_map_plant": 0.896, "eval_map_small": 0.3302, "eval_mar_1": 0.336, "eval_mar_10": 0.8695, "eval_mar_100": 0.8739, "eval_mar_100_chicken": 0.872, "eval_mar_100_duck": 0.8284, "eval_mar_100_plant": 0.9213, "eval_mar_large": 0.879, "eval_mar_medium": 0.8764, "eval_mar_small": 0.4186, "eval_runtime": 12.5862, "eval_samples_per_second": 15.89, "eval_steps_per_second": 1.986, "step": 116000 }, { "epoch": 116.01, "grad_norm": 65.84066009521484, "learning_rate": 1.2143405443036182e-06, "loss": 0.2957, "step": 116010 }, { "epoch": 116.04, "grad_norm": 31.770423889160156, "learning_rate": 1.2122890084564542e-06, "loss": 0.2519, "step": 116040 }, { "epoch": 116.07, "grad_norm": 66.86061096191406, "learning_rate": 1.2102389679376037e-06, "loss": 0.366, "step": 116070 }, { "epoch": 116.1, "grad_norm": 58.58472442626953, "learning_rate": 1.2081904235563908e-06, "loss": 0.2942, "step": 116100 }, { "epoch": 116.13, "grad_norm": 47.948490142822266, "learning_rate": 1.206143376121549e-06, "loss": 0.3286, "step": 116130 }, { "epoch": 116.16, "grad_norm": 41.05491256713867, "learning_rate": 1.204097826441218e-06, "loss": 0.3368, "step": 116160 }, { "epoch": 116.19, "grad_norm": 41.1511344909668, "learning_rate": 1.2020537753229506e-06, "loss": 0.3012, "step": 116190 }, { "epoch": 116.22, "grad_norm": 82.31441497802734, "learning_rate": 1.200011223573702e-06, "loss": 0.3028, "step": 116220 }, { "epoch": 116.25, "grad_norm": 47.29643630981445, "learning_rate": 1.1979701719998454e-06, "loss": 0.3297, "step": 116250 }, { "epoch": 116.28, "grad_norm": 46.353294372558594, "learning_rate": 1.1959306214071508e-06, "loss": 0.3846, "step": 116280 }, { "epoch": 116.31, "grad_norm": 51.256736755371094, "learning_rate": 1.193892572600804e-06, "loss": 0.3169, "step": 116310 }, { "epoch": 116.34, "grad_norm": 28.250940322875977, "learning_rate": 1.1918560263853902e-06, "loss": 0.3795, "step": 116340 }, { "epoch": 116.37, "grad_norm": 40.305206298828125, "learning_rate": 1.1898209835649083e-06, "loss": 0.3364, "step": 116370 }, { "epoch": 116.4, "grad_norm": 43.21030044555664, "learning_rate": 1.18778744494276e-06, "loss": 0.3171, "step": 116400 }, { "epoch": 116.43, "grad_norm": 54.95117950439453, "learning_rate": 1.1857554113217568e-06, "loss": 0.2974, "step": 116430 }, { "epoch": 116.46, "grad_norm": 33.02461242675781, "learning_rate": 1.1837248835041093e-06, "loss": 0.3039, "step": 116460 }, { "epoch": 116.49, "grad_norm": 302.1905517578125, "learning_rate": 1.181695862291441e-06, "loss": 0.2664, "step": 116490 }, { "epoch": 116.52, "grad_norm": 48.875770568847656, "learning_rate": 1.1796683484847731e-06, "loss": 0.2485, "step": 116520 }, { "epoch": 116.55, "grad_norm": 61.11624526977539, "learning_rate": 1.1776423428845423e-06, "loss": 0.312, "step": 116550 }, { "epoch": 116.58, "grad_norm": 45.63688278198242, "learning_rate": 1.1756178462905782e-06, "loss": 0.3392, "step": 116580 }, { "epoch": 116.61, "grad_norm": 37.25294876098633, "learning_rate": 1.1735948595021234e-06, "loss": 0.2682, "step": 116610 }, { "epoch": 116.64, "grad_norm": 94.44135284423828, "learning_rate": 1.1715733833178178e-06, "loss": 0.2981, "step": 116640 }, { "epoch": 116.67, "grad_norm": 51.25745391845703, "learning_rate": 1.16955341853571e-06, "loss": 0.3095, "step": 116670 }, { "epoch": 116.7, "grad_norm": 68.81034088134766, "learning_rate": 1.1675349659532514e-06, "loss": 0.3542, "step": 116700 }, { "epoch": 116.73, "grad_norm": 46.76045608520508, "learning_rate": 1.1655180263672928e-06, "loss": 0.2845, "step": 116730 }, { "epoch": 116.76, "grad_norm": 71.26046752929688, "learning_rate": 1.1635026005740902e-06, "loss": 0.3088, "step": 116760 }, { "epoch": 116.79, "grad_norm": 80.9001235961914, "learning_rate": 1.1614886893693044e-06, "loss": 0.3249, "step": 116790 }, { "epoch": 116.82, "grad_norm": 93.72239685058594, "learning_rate": 1.159476293547992e-06, "loss": 0.2698, "step": 116820 }, { "epoch": 116.85, "grad_norm": 342.6943054199219, "learning_rate": 1.1574654139046171e-06, "loss": 0.3149, "step": 116850 }, { "epoch": 116.88, "grad_norm": 39.89041519165039, "learning_rate": 1.1554560512330437e-06, "loss": 0.343, "step": 116880 }, { "epoch": 116.91, "grad_norm": 68.13346862792969, "learning_rate": 1.1534482063265346e-06, "loss": 0.276, "step": 116910 }, { "epoch": 116.94, "grad_norm": 63.337554931640625, "learning_rate": 1.1514418799777554e-06, "loss": 0.305, "step": 116940 }, { "epoch": 116.97, "grad_norm": 37.85152816772461, "learning_rate": 1.149437072978773e-06, "loss": 0.3091, "step": 116970 }, { "epoch": 117.0, "grad_norm": 58.31180953979492, "learning_rate": 1.1474337861210543e-06, "loss": 0.3156, "step": 117000 }, { "epoch": 117.0, "eval_loss": 0.2513735890388489, "eval_map": 0.8285, "eval_map_50": 0.9597, "eval_map_75": 0.9311, "eval_map_chicken": 0.8205, "eval_map_duck": 0.7744, "eval_map_large": 0.827, "eval_map_medium": 0.8317, "eval_map_plant": 0.8907, "eval_map_small": 0.3177, "eval_mar_1": 0.3333, "eval_mar_10": 0.8605, "eval_mar_100": 0.8645, "eval_mar_100_chicken": 0.8646, "eval_mar_100_duck": 0.8119, "eval_mar_100_plant": 0.917, "eval_mar_large": 0.8634, "eval_mar_medium": 0.8681, "eval_mar_small": 0.392, "eval_runtime": 12.9412, "eval_samples_per_second": 15.455, "eval_steps_per_second": 1.932, "step": 117000 }, { "epoch": 117.03, "grad_norm": 88.5228271484375, "learning_rate": 1.1454320201954628e-06, "loss": 0.3123, "step": 117030 }, { "epoch": 117.06, "grad_norm": 99.66394805908203, "learning_rate": 1.1434317759922664e-06, "loss": 0.2677, "step": 117060 }, { "epoch": 117.09, "grad_norm": 26.195398330688477, "learning_rate": 1.141433054301127e-06, "loss": 0.2934, "step": 117090 }, { "epoch": 117.12, "grad_norm": 53.116329193115234, "learning_rate": 1.1394358559111101e-06, "loss": 0.285, "step": 117120 }, { "epoch": 117.15, "grad_norm": 62.40752029418945, "learning_rate": 1.1374401816106778e-06, "loss": 0.3201, "step": 117150 }, { "epoch": 117.18, "grad_norm": 176.06365966796875, "learning_rate": 1.135446032187692e-06, "loss": 0.2631, "step": 117180 }, { "epoch": 117.21, "grad_norm": 36.219512939453125, "learning_rate": 1.1334534084294084e-06, "loss": 0.2779, "step": 117210 }, { "epoch": 117.24, "grad_norm": 0.0018512321403250098, "learning_rate": 1.1314623111224865e-06, "loss": 0.3335, "step": 117240 }, { "epoch": 117.27, "grad_norm": 314.8780517578125, "learning_rate": 1.1294727410529754e-06, "loss": 0.5026, "step": 117270 }, { "epoch": 117.3, "grad_norm": 44.36787033081055, "learning_rate": 1.1274846990063314e-06, "loss": 0.3207, "step": 117300 }, { "epoch": 117.33, "grad_norm": 58.95147705078125, "learning_rate": 1.125498185767398e-06, "loss": 0.2939, "step": 117330 }, { "epoch": 117.36, "grad_norm": 34.95223617553711, "learning_rate": 1.1235132021204226e-06, "loss": 0.3162, "step": 117360 }, { "epoch": 117.39, "grad_norm": 58.721309661865234, "learning_rate": 1.1215297488490412e-06, "loss": 0.3504, "step": 117390 }, { "epoch": 117.42, "grad_norm": 53.60651779174805, "learning_rate": 1.1195478267362924e-06, "loss": 0.3201, "step": 117420 }, { "epoch": 117.45, "grad_norm": 30.908355712890625, "learning_rate": 1.1175674365646067e-06, "loss": 0.2437, "step": 117450 }, { "epoch": 117.48, "grad_norm": 53.9553337097168, "learning_rate": 1.1155885791158128e-06, "loss": 0.2973, "step": 117480 }, { "epoch": 117.51, "grad_norm": 89.2034683227539, "learning_rate": 1.1136112551711293e-06, "loss": 0.2948, "step": 117510 }, { "epoch": 117.54, "grad_norm": 47.13205337524414, "learning_rate": 1.111635465511175e-06, "loss": 0.3301, "step": 117540 }, { "epoch": 117.57, "grad_norm": 92.46783447265625, "learning_rate": 1.109661210915956e-06, "loss": 0.3829, "step": 117570 }, { "epoch": 117.6, "grad_norm": 121.87944793701172, "learning_rate": 1.1076884921648834e-06, "loss": 0.3195, "step": 117600 }, { "epoch": 117.63, "grad_norm": 60.715206146240234, "learning_rate": 1.1057173100367495e-06, "loss": 0.2806, "step": 117630 }, { "epoch": 117.66, "grad_norm": 81.54305267333984, "learning_rate": 1.1037476653097501e-06, "loss": 0.2571, "step": 117660 }, { "epoch": 117.69, "grad_norm": 33.13121032714844, "learning_rate": 1.101779558761466e-06, "loss": 0.2792, "step": 117690 }, { "epoch": 117.72, "grad_norm": 53.67279052734375, "learning_rate": 1.0998129911688766e-06, "loss": 0.2649, "step": 117720 }, { "epoch": 117.75, "grad_norm": 139.10862731933594, "learning_rate": 1.097847963308351e-06, "loss": 0.3488, "step": 117750 }, { "epoch": 117.78, "grad_norm": 61.68399429321289, "learning_rate": 1.0958844759556525e-06, "loss": 0.37, "step": 117780 }, { "epoch": 117.81, "grad_norm": 4.33638779213652e-05, "learning_rate": 1.0939225298859324e-06, "loss": 0.2904, "step": 117810 }, { "epoch": 117.84, "grad_norm": 33.36587905883789, "learning_rate": 1.0919621258737384e-06, "loss": 0.3684, "step": 117840 }, { "epoch": 117.87, "grad_norm": 45.22385025024414, "learning_rate": 1.090003264693003e-06, "loss": 0.3218, "step": 117870 }, { "epoch": 117.9, "grad_norm": 30.70008659362793, "learning_rate": 1.0880459471170597e-06, "loss": 0.2821, "step": 117900 }, { "epoch": 117.93, "grad_norm": 61.99275207519531, "learning_rate": 1.0860901739186209e-06, "loss": 0.3618, "step": 117930 }, { "epoch": 117.96, "grad_norm": 56.93696212768555, "learning_rate": 1.0841359458697986e-06, "loss": 0.2827, "step": 117960 }, { "epoch": 117.99, "grad_norm": 63.12413787841797, "learning_rate": 1.0821832637420887e-06, "loss": 0.3184, "step": 117990 }, { "epoch": 118.0, "eval_loss": 0.24968189001083374, "eval_map": 0.8294, "eval_map_50": 0.9636, "eval_map_75": 0.9301, "eval_map_chicken": 0.8231, "eval_map_duck": 0.7759, "eval_map_large": 0.8352, "eval_map_medium": 0.8321, "eval_map_plant": 0.8892, "eval_map_small": 0.3286, "eval_mar_1": 0.3334, "eval_mar_10": 0.8631, "eval_mar_100": 0.8669, "eval_mar_100_chicken": 0.8672, "eval_mar_100_duck": 0.817, "eval_mar_100_plant": 0.9166, "eval_mar_large": 0.8738, "eval_mar_medium": 0.8691, "eval_mar_small": 0.3995, "eval_runtime": 12.6987, "eval_samples_per_second": 15.75, "eval_steps_per_second": 1.969, "step": 118000 }, { "epoch": 118.02, "grad_norm": 43.703067779541016, "learning_rate": 1.0802321283063794e-06, "loss": 0.3308, "step": 118020 }, { "epoch": 118.05, "grad_norm": 772.25146484375, "learning_rate": 1.0782825403329488e-06, "loss": 0.4412, "step": 118050 }, { "epoch": 118.08, "grad_norm": 113.36270141601562, "learning_rate": 1.0763345005914649e-06, "loss": 0.3575, "step": 118080 }, { "epoch": 118.11, "grad_norm": 98.24491882324219, "learning_rate": 1.0743880098509802e-06, "loss": 0.2888, "step": 118110 }, { "epoch": 118.14, "grad_norm": 48.9073600769043, "learning_rate": 1.0724430688799402e-06, "loss": 0.2739, "step": 118140 }, { "epoch": 118.17, "grad_norm": 30.04938316345215, "learning_rate": 1.0704996784461753e-06, "loss": 0.322, "step": 118170 }, { "epoch": 118.2, "grad_norm": 36.7311897277832, "learning_rate": 1.0685578393169054e-06, "loss": 0.3117, "step": 118200 }, { "epoch": 118.23, "grad_norm": 26.93063735961914, "learning_rate": 1.0666175522587402e-06, "loss": 0.2596, "step": 118230 }, { "epoch": 118.26, "grad_norm": 51.703678131103516, "learning_rate": 1.0646788180376716e-06, "loss": 0.2896, "step": 118260 }, { "epoch": 118.29, "grad_norm": 20.9184627532959, "learning_rate": 1.0627416374190818e-06, "loss": 0.3321, "step": 118290 }, { "epoch": 118.32, "grad_norm": 28.498327255249023, "learning_rate": 1.0608060111677409e-06, "loss": 0.3443, "step": 118320 }, { "epoch": 118.35, "grad_norm": 61.52924728393555, "learning_rate": 1.0588719400478004e-06, "loss": 0.3166, "step": 118350 }, { "epoch": 118.38, "grad_norm": 53.04189682006836, "learning_rate": 1.0569394248228026e-06, "loss": 0.3309, "step": 118380 }, { "epoch": 118.41, "grad_norm": 42.258113861083984, "learning_rate": 1.0550084662556753e-06, "loss": 0.2852, "step": 118410 }, { "epoch": 118.44, "grad_norm": 48.20222854614258, "learning_rate": 1.053079065108728e-06, "loss": 0.3201, "step": 118440 }, { "epoch": 118.47, "grad_norm": 75.4854965209961, "learning_rate": 1.0511512221436581e-06, "loss": 0.3758, "step": 118470 }, { "epoch": 118.5, "grad_norm": 33.50846862792969, "learning_rate": 1.049224938121548e-06, "loss": 0.3173, "step": 118500 }, { "epoch": 118.53, "grad_norm": 45.05970764160156, "learning_rate": 1.0473002138028654e-06, "loss": 0.3661, "step": 118530 }, { "epoch": 118.56, "grad_norm": 44.69578170776367, "learning_rate": 1.0453770499474585e-06, "loss": 0.4152, "step": 118560 }, { "epoch": 118.59, "grad_norm": 39.160099029541016, "learning_rate": 1.0434554473145646e-06, "loss": 0.317, "step": 118590 }, { "epoch": 118.62, "grad_norm": 86.883544921875, "learning_rate": 1.0415354066627993e-06, "loss": 0.3816, "step": 118620 }, { "epoch": 118.65, "grad_norm": 41.29216003417969, "learning_rate": 1.0396169287501652e-06, "loss": 0.2935, "step": 118650 }, { "epoch": 118.68, "grad_norm": 797.63671875, "learning_rate": 1.037700014334047e-06, "loss": 0.2985, "step": 118680 }, { "epoch": 118.71, "grad_norm": 62.420230865478516, "learning_rate": 1.0357846641712143e-06, "loss": 0.2677, "step": 118710 }, { "epoch": 118.74, "grad_norm": 34.74366760253906, "learning_rate": 1.0338708790178136e-06, "loss": 0.3341, "step": 118740 }, { "epoch": 118.77, "grad_norm": 197.43185424804688, "learning_rate": 1.03195865962938e-06, "loss": 0.2676, "step": 118770 }, { "epoch": 118.8, "grad_norm": 55.87971496582031, "learning_rate": 1.0300480067608232e-06, "loss": 0.2998, "step": 118800 }, { "epoch": 118.83, "grad_norm": 53.527164459228516, "learning_rate": 1.0281389211664439e-06, "loss": 0.3818, "step": 118830 }, { "epoch": 118.86, "grad_norm": 35.31074523925781, "learning_rate": 1.026231403599915e-06, "loss": 0.3269, "step": 118860 }, { "epoch": 118.89, "grad_norm": 140.10302734375, "learning_rate": 1.0243254548142973e-06, "loss": 0.3048, "step": 118890 }, { "epoch": 118.92, "grad_norm": 41.628047943115234, "learning_rate": 1.0224210755620257e-06, "loss": 0.3587, "step": 118920 }, { "epoch": 118.95, "grad_norm": 136.59413146972656, "learning_rate": 1.020518266594921e-06, "loss": 0.293, "step": 118950 }, { "epoch": 118.98, "grad_norm": 34.68840789794922, "learning_rate": 1.0186170286641816e-06, "loss": 0.2561, "step": 118980 }, { "epoch": 119.0, "eval_loss": 0.24401848018169403, "eval_map": 0.8339, "eval_map_50": 0.963, "eval_map_75": 0.929, "eval_map_chicken": 0.8266, "eval_map_duck": 0.785, "eval_map_large": 0.8349, "eval_map_medium": 0.8324, "eval_map_plant": 0.8901, "eval_map_small": 0.3126, "eval_mar_1": 0.3359, "eval_mar_10": 0.8661, "eval_mar_100": 0.8701, "eval_mar_100_chicken": 0.8702, "eval_mar_100_duck": 0.8232, "eval_mar_100_plant": 0.917, "eval_mar_large": 0.8724, "eval_mar_medium": 0.8711, "eval_mar_small": 0.3844, "eval_runtime": 12.916, "eval_samples_per_second": 15.485, "eval_steps_per_second": 1.936, "step": 119000 }, { "epoch": 119.01, "grad_norm": 79.59882354736328, "learning_rate": 1.016717362520388e-06, "loss": 0.287, "step": 119010 }, { "epoch": 119.04, "grad_norm": 4.899946361547336e-05, "learning_rate": 1.014819268913495e-06, "loss": 0.2657, "step": 119040 }, { "epoch": 119.07, "grad_norm": 85.00459289550781, "learning_rate": 1.0129227485928432e-06, "loss": 0.316, "step": 119070 }, { "epoch": 119.1, "grad_norm": 88.0716323852539, "learning_rate": 1.0110278023071445e-06, "loss": 0.3115, "step": 119100 }, { "epoch": 119.13, "grad_norm": 28.01144790649414, "learning_rate": 1.0091344308044987e-06, "loss": 0.3072, "step": 119130 }, { "epoch": 119.16, "grad_norm": 83.44989013671875, "learning_rate": 1.0072426348323754e-06, "loss": 0.3146, "step": 119160 }, { "epoch": 119.19, "grad_norm": 23.89536476135254, "learning_rate": 1.0053524151376283e-06, "loss": 0.2639, "step": 119190 }, { "epoch": 119.22, "grad_norm": 59.83538818359375, "learning_rate": 1.0034637724664832e-06, "loss": 0.3178, "step": 119220 }, { "epoch": 119.25, "grad_norm": 83.98983764648438, "learning_rate": 1.0015767075645472e-06, "loss": 0.4583, "step": 119250 }, { "epoch": 119.28, "grad_norm": 20.30184555053711, "learning_rate": 9.99691221176805e-07, "loss": 0.2935, "step": 119280 }, { "epoch": 119.31, "grad_norm": 48.67921447753906, "learning_rate": 9.978073140476169e-07, "loss": 0.291, "step": 119310 }, { "epoch": 119.34, "grad_norm": 60.34995651245117, "learning_rate": 9.959249869207177e-07, "loss": 0.3298, "step": 119340 }, { "epoch": 119.37, "grad_norm": 59.32844161987305, "learning_rate": 9.940442405392226e-07, "loss": 0.2902, "step": 119370 }, { "epoch": 119.4, "grad_norm": 127.55361938476562, "learning_rate": 9.921650756456164e-07, "loss": 0.2851, "step": 119400 }, { "epoch": 119.43, "grad_norm": 38.41410446166992, "learning_rate": 9.902874929817696e-07, "loss": 0.2932, "step": 119430 }, { "epoch": 119.46, "grad_norm": 46.175048828125, "learning_rate": 9.884114932889172e-07, "loss": 0.3097, "step": 119460 }, { "epoch": 119.49, "grad_norm": 82.69879150390625, "learning_rate": 9.86537077307677e-07, "loss": 0.385, "step": 119490 }, { "epoch": 119.52, "grad_norm": 77.70191955566406, "learning_rate": 9.84664245778037e-07, "loss": 0.3804, "step": 119520 }, { "epoch": 119.55, "grad_norm": 46.23442077636719, "learning_rate": 9.82792999439362e-07, "loss": 0.3341, "step": 119550 }, { "epoch": 119.58, "grad_norm": 32.586116790771484, "learning_rate": 9.809233390303901e-07, "loss": 0.3493, "step": 119580 }, { "epoch": 119.61, "grad_norm": 242.12603759765625, "learning_rate": 9.79055265289236e-07, "loss": 0.3059, "step": 119610 }, { "epoch": 119.64, "grad_norm": 38.090396881103516, "learning_rate": 9.771887789533818e-07, "loss": 0.3175, "step": 119640 }, { "epoch": 119.67, "grad_norm": 39.20272445678711, "learning_rate": 9.753238807596903e-07, "loss": 0.29, "step": 119670 }, { "epoch": 119.7, "grad_norm": 58.96403121948242, "learning_rate": 9.734605714443906e-07, "loss": 0.321, "step": 119700 }, { "epoch": 119.73, "grad_norm": 66.43486785888672, "learning_rate": 9.715988517430896e-07, "loss": 0.248, "step": 119730 }, { "epoch": 119.76, "grad_norm": 0.00010922289220616221, "learning_rate": 9.69738722390765e-07, "loss": 0.2966, "step": 119760 }, { "epoch": 119.79, "grad_norm": 63.542842864990234, "learning_rate": 9.67880184121765e-07, "loss": 0.3399, "step": 119790 }, { "epoch": 119.82, "grad_norm": 63.45317459106445, "learning_rate": 9.66023237669812e-07, "loss": 0.3216, "step": 119820 }, { "epoch": 119.85, "grad_norm": 44.85850143432617, "learning_rate": 9.641678837679985e-07, "loss": 0.2499, "step": 119850 }, { "epoch": 119.88, "grad_norm": 31.49958038330078, "learning_rate": 9.623141231487904e-07, "loss": 0.2871, "step": 119880 }, { "epoch": 119.91, "grad_norm": 26.224300384521484, "learning_rate": 9.60461956544021e-07, "loss": 0.3991, "step": 119910 }, { "epoch": 119.94, "grad_norm": 28.243581771850586, "learning_rate": 9.586113846848982e-07, "loss": 0.2259, "step": 119940 }, { "epoch": 119.97, "grad_norm": 55.22255325317383, "learning_rate": 9.567624083019966e-07, "loss": 0.3521, "step": 119970 }, { "epoch": 120.0, "grad_norm": 64.13876342773438, "learning_rate": 9.549150281252633e-07, "loss": 0.2776, "step": 120000 }, { "epoch": 120.0, "eval_loss": 0.24415385723114014, "eval_map": 0.8358, "eval_map_50": 0.9622, "eval_map_75": 0.9283, "eval_map_chicken": 0.8278, "eval_map_duck": 0.7874, "eval_map_large": 0.835, "eval_map_medium": 0.8371, "eval_map_plant": 0.8921, "eval_map_small": 0.3166, "eval_mar_1": 0.3366, "eval_mar_10": 0.8677, "eval_mar_100": 0.8717, "eval_mar_100_chicken": 0.8724, "eval_mar_100_duck": 0.8242, "eval_mar_100_plant": 0.9183, "eval_mar_large": 0.8706, "eval_mar_medium": 0.8759, "eval_mar_small": 0.3809, "eval_runtime": 12.981, "eval_samples_per_second": 15.407, "eval_steps_per_second": 1.926, "step": 120000 }, { "epoch": 120.03, "grad_norm": 42.1312370300293, "learning_rate": 9.530692448840151e-07, "loss": 0.2645, "step": 120030 }, { "epoch": 120.06, "grad_norm": 94.06922912597656, "learning_rate": 9.512250593069394e-07, "loss": 0.3051, "step": 120060 }, { "epoch": 120.09, "grad_norm": 33.870872497558594, "learning_rate": 9.49382472122089e-07, "loss": 0.2851, "step": 120090 }, { "epoch": 120.12, "grad_norm": 35.32245635986328, "learning_rate": 9.475414840568903e-07, "loss": 0.2386, "step": 120120 }, { "epoch": 120.15, "grad_norm": 47.44566345214844, "learning_rate": 9.457020958381324e-07, "loss": 0.2825, "step": 120150 }, { "epoch": 120.18, "grad_norm": 19.2061710357666, "learning_rate": 9.438643081919818e-07, "loss": 0.3237, "step": 120180 }, { "epoch": 120.21, "grad_norm": 32.5858039855957, "learning_rate": 9.420281218439648e-07, "loss": 0.2668, "step": 120210 }, { "epoch": 120.24, "grad_norm": 41.57807540893555, "learning_rate": 9.401935375189802e-07, "loss": 0.3473, "step": 120240 }, { "epoch": 120.27, "grad_norm": 80.34391784667969, "learning_rate": 9.383605559412911e-07, "loss": 0.3395, "step": 120270 }, { "epoch": 120.3, "grad_norm": 75.64508819580078, "learning_rate": 9.365291778345303e-07, "loss": 0.367, "step": 120300 }, { "epoch": 120.33, "grad_norm": 89.43112182617188, "learning_rate": 9.346994039216972e-07, "loss": 0.3338, "step": 120330 }, { "epoch": 120.36, "grad_norm": 83.30877685546875, "learning_rate": 9.32871234925159e-07, "loss": 0.3123, "step": 120360 }, { "epoch": 120.39, "grad_norm": 38.849365234375, "learning_rate": 9.310446715666449e-07, "loss": 0.288, "step": 120390 }, { "epoch": 120.42, "grad_norm": 37.362911224365234, "learning_rate": 9.29219714567256e-07, "loss": 0.3621, "step": 120420 }, { "epoch": 120.45, "grad_norm": 45.99931716918945, "learning_rate": 9.273963646474527e-07, "loss": 0.2807, "step": 120450 }, { "epoch": 120.48, "grad_norm": 37.47846221923828, "learning_rate": 9.255746225270689e-07, "loss": 0.2962, "step": 120480 }, { "epoch": 120.51, "grad_norm": 49.93107986450195, "learning_rate": 9.237544889252969e-07, "loss": 0.2756, "step": 120510 }, { "epoch": 120.54, "grad_norm": 80.15019226074219, "learning_rate": 9.21935964560699e-07, "loss": 0.535, "step": 120540 }, { "epoch": 120.57, "grad_norm": 41.61152267456055, "learning_rate": 9.201190501511964e-07, "loss": 0.3592, "step": 120570 }, { "epoch": 120.6, "grad_norm": 45.39149475097656, "learning_rate": 9.183037464140804e-07, "loss": 0.2998, "step": 120600 }, { "epoch": 120.63, "grad_norm": 46.62757110595703, "learning_rate": 9.164900540660032e-07, "loss": 0.3632, "step": 120630 }, { "epoch": 120.66, "grad_norm": 76.2103271484375, "learning_rate": 9.146779738229838e-07, "loss": 0.2786, "step": 120660 }, { "epoch": 120.69, "grad_norm": 53.28227233886719, "learning_rate": 9.128675064004006e-07, "loss": 0.2721, "step": 120690 }, { "epoch": 120.72, "grad_norm": 50.21246337890625, "learning_rate": 9.110586525129988e-07, "loss": 0.2878, "step": 120720 }, { "epoch": 120.75, "grad_norm": 113.7528076171875, "learning_rate": 9.09251412874882e-07, "loss": 0.3505, "step": 120750 }, { "epoch": 120.78, "grad_norm": 46.23336410522461, "learning_rate": 9.074457881995252e-07, "loss": 0.3381, "step": 120780 }, { "epoch": 120.81, "grad_norm": 32.41233444213867, "learning_rate": 9.056417791997568e-07, "loss": 0.3456, "step": 120810 }, { "epoch": 120.84, "grad_norm": 0.00013031107664573938, "learning_rate": 9.038393865877725e-07, "loss": 0.3772, "step": 120840 }, { "epoch": 120.87, "grad_norm": 37.69630432128906, "learning_rate": 9.020386110751267e-07, "loss": 0.284, "step": 120870 }, { "epoch": 120.9, "grad_norm": 65.08125305175781, "learning_rate": 9.002394533727382e-07, "loss": 0.3774, "step": 120900 }, { "epoch": 120.93, "grad_norm": 93.68107604980469, "learning_rate": 8.984419141908857e-07, "loss": 0.3423, "step": 120930 }, { "epoch": 120.96, "grad_norm": 161.4302215576172, "learning_rate": 8.966459942392108e-07, "loss": 0.3577, "step": 120960 }, { "epoch": 120.99, "grad_norm": 90.91012573242188, "learning_rate": 8.94851694226711e-07, "loss": 0.2591, "step": 120990 }, { "epoch": 121.0, "eval_loss": 0.24734477698802948, "eval_map": 0.8302, "eval_map_50": 0.9627, "eval_map_75": 0.93, "eval_map_chicken": 0.819, "eval_map_duck": 0.7774, "eval_map_large": 0.8306, "eval_map_medium": 0.8283, "eval_map_plant": 0.8942, "eval_map_small": 0.3066, "eval_mar_1": 0.3334, "eval_mar_10": 0.8615, "eval_mar_100": 0.8653, "eval_mar_100_chicken": 0.863, "eval_mar_100_duck": 0.8144, "eval_mar_100_plant": 0.9185, "eval_mar_large": 0.8695, "eval_mar_medium": 0.8663, "eval_mar_small": 0.3879, "eval_runtime": 12.6602, "eval_samples_per_second": 15.797, "eval_steps_per_second": 1.975, "step": 121000 }, { "epoch": 121.02, "grad_norm": 43.917686462402344, "learning_rate": 8.930590148617513e-07, "loss": 0.3135, "step": 121020 }, { "epoch": 121.05, "grad_norm": 47.407798767089844, "learning_rate": 8.912679568520494e-07, "loss": 0.2686, "step": 121050 }, { "epoch": 121.08, "grad_norm": 38.72737503051758, "learning_rate": 8.894785209046886e-07, "loss": 0.2596, "step": 121080 }, { "epoch": 121.11, "grad_norm": 57.526275634765625, "learning_rate": 8.876907077261093e-07, "loss": 0.2928, "step": 121110 }, { "epoch": 121.14, "grad_norm": 35.45844268798828, "learning_rate": 8.859045180221137e-07, "loss": 0.357, "step": 121140 }, { "epoch": 121.17, "grad_norm": 274.0605163574219, "learning_rate": 8.841199524978583e-07, "loss": 0.2615, "step": 121170 }, { "epoch": 121.2, "grad_norm": 74.42118072509766, "learning_rate": 8.823370118578628e-07, "loss": 0.3201, "step": 121200 }, { "epoch": 121.23, "grad_norm": 45.51139450073242, "learning_rate": 8.805556968060047e-07, "loss": 0.281, "step": 121230 }, { "epoch": 121.26, "grad_norm": 62.175201416015625, "learning_rate": 8.787760080455171e-07, "loss": 0.3795, "step": 121260 }, { "epoch": 121.29, "grad_norm": 75.42108154296875, "learning_rate": 8.769979462789957e-07, "loss": 0.2986, "step": 121290 }, { "epoch": 121.32, "grad_norm": 43.88665771484375, "learning_rate": 8.752215122083874e-07, "loss": 0.269, "step": 121320 }, { "epoch": 121.35, "grad_norm": 44.19907760620117, "learning_rate": 8.734467065350022e-07, "loss": 0.2894, "step": 121350 }, { "epoch": 121.38, "grad_norm": 204.92965698242188, "learning_rate": 8.716735299595059e-07, "loss": 0.2856, "step": 121380 }, { "epoch": 121.41, "grad_norm": 112.35313415527344, "learning_rate": 8.699019831819206e-07, "loss": 0.2488, "step": 121410 }, { "epoch": 121.44, "grad_norm": 70.0196762084961, "learning_rate": 8.68132066901623e-07, "loss": 0.3113, "step": 121440 }, { "epoch": 121.47, "grad_norm": 130.3570556640625, "learning_rate": 8.663637818173504e-07, "loss": 0.2488, "step": 121470 }, { "epoch": 121.5, "grad_norm": 44.26051330566406, "learning_rate": 8.645971286271903e-07, "loss": 0.439, "step": 121500 }, { "epoch": 121.53, "grad_norm": 29.16769027709961, "learning_rate": 8.628321080285945e-07, "loss": 0.3191, "step": 121530 }, { "epoch": 121.56, "grad_norm": 76.77120971679688, "learning_rate": 8.610687207183604e-07, "loss": 0.2753, "step": 121560 }, { "epoch": 121.59, "grad_norm": 34.87427520751953, "learning_rate": 8.59306967392649e-07, "loss": 0.4959, "step": 121590 }, { "epoch": 121.62, "grad_norm": 86.26604461669922, "learning_rate": 8.575468487469696e-07, "loss": 0.3427, "step": 121620 }, { "epoch": 121.65, "grad_norm": 52.9260139465332, "learning_rate": 8.557883654761906e-07, "loss": 0.2978, "step": 121650 }, { "epoch": 121.68, "grad_norm": 46.798622131347656, "learning_rate": 8.540315182745329e-07, "loss": 0.3954, "step": 121680 }, { "epoch": 121.71, "grad_norm": 55.85752868652344, "learning_rate": 8.522763078355739e-07, "loss": 0.3338, "step": 121710 }, { "epoch": 121.74, "grad_norm": 62.068572998046875, "learning_rate": 8.505227348522404e-07, "loss": 0.3277, "step": 121740 }, { "epoch": 121.77, "grad_norm": 64.33328247070312, "learning_rate": 8.487708000168166e-07, "loss": 0.2501, "step": 121770 }, { "epoch": 121.8, "grad_norm": 70.65706634521484, "learning_rate": 8.470205040209362e-07, "loss": 0.3291, "step": 121800 }, { "epoch": 121.83, "grad_norm": 33.4031867980957, "learning_rate": 8.452718475555927e-07, "loss": 0.3321, "step": 121830 }, { "epoch": 121.86, "grad_norm": 46.883602142333984, "learning_rate": 8.435248313111244e-07, "loss": 0.2723, "step": 121860 }, { "epoch": 121.89, "grad_norm": 37.142555236816406, "learning_rate": 8.41779455977228e-07, "loss": 0.3394, "step": 121890 }, { "epoch": 121.92, "grad_norm": 34.430599212646484, "learning_rate": 8.400357222429473e-07, "loss": 0.2995, "step": 121920 }, { "epoch": 121.95, "grad_norm": 65.86578369140625, "learning_rate": 8.382936307966838e-07, "loss": 0.4001, "step": 121950 }, { "epoch": 121.98, "grad_norm": 40.32062530517578, "learning_rate": 8.365531823261841e-07, "loss": 0.3241, "step": 121980 }, { "epoch": 122.0, "eval_loss": 0.2518215477466583, "eval_map": 0.831, "eval_map_50": 0.9615, "eval_map_75": 0.9294, "eval_map_chicken": 0.82, "eval_map_duck": 0.7825, "eval_map_large": 0.8359, "eval_map_medium": 0.8319, "eval_map_plant": 0.8905, "eval_map_small": 0.3046, "eval_mar_1": 0.3329, "eval_mar_10": 0.8637, "eval_mar_100": 0.8668, "eval_mar_100_chicken": 0.8662, "eval_mar_100_duck": 0.818, "eval_mar_100_plant": 0.9163, "eval_mar_large": 0.8747, "eval_mar_medium": 0.8693, "eval_mar_small": 0.3782, "eval_runtime": 12.66, "eval_samples_per_second": 15.798, "eval_steps_per_second": 1.975, "step": 122000 }, { "epoch": 122.01, "grad_norm": 26.775012969970703, "learning_rate": 8.348143775185536e-07, "loss": 0.3348, "step": 122010 }, { "epoch": 122.04, "grad_norm": 81.80377960205078, "learning_rate": 8.330772170602424e-07, "loss": 0.3074, "step": 122040 }, { "epoch": 122.07, "grad_norm": 101.98365020751953, "learning_rate": 8.313417016370557e-07, "loss": 0.3586, "step": 122070 }, { "epoch": 122.1, "grad_norm": 80.21438598632812, "learning_rate": 8.296078319341444e-07, "loss": 0.2769, "step": 122100 }, { "epoch": 122.13, "grad_norm": 106.71128845214844, "learning_rate": 8.278756086360157e-07, "loss": 0.278, "step": 122130 }, { "epoch": 122.16, "grad_norm": 95.8992691040039, "learning_rate": 8.261450324265225e-07, "loss": 0.3181, "step": 122160 }, { "epoch": 122.19, "grad_norm": 61.61162567138672, "learning_rate": 8.244161039888709e-07, "loss": 0.2333, "step": 122190 }, { "epoch": 122.22, "grad_norm": 44.09703063964844, "learning_rate": 8.226888240056114e-07, "loss": 0.2849, "step": 122220 }, { "epoch": 122.25, "grad_norm": 187.04525756835938, "learning_rate": 8.209631931586499e-07, "loss": 0.271, "step": 122250 }, { "epoch": 122.28, "grad_norm": 69.08055114746094, "learning_rate": 8.192392121292336e-07, "loss": 0.3909, "step": 122280 }, { "epoch": 122.31, "grad_norm": 25.231712341308594, "learning_rate": 8.175168815979689e-07, "loss": 0.2648, "step": 122310 }, { "epoch": 122.34, "grad_norm": 38.346405029296875, "learning_rate": 8.157962022448001e-07, "loss": 0.2941, "step": 122340 }, { "epoch": 122.37, "grad_norm": 67.84364318847656, "learning_rate": 8.140771747490273e-07, "loss": 0.2372, "step": 122370 }, { "epoch": 122.4, "grad_norm": 63.16106033325195, "learning_rate": 8.123597997892918e-07, "loss": 0.3158, "step": 122400 }, { "epoch": 122.43, "grad_norm": 44.99759292602539, "learning_rate": 8.106440780435881e-07, "loss": 0.2799, "step": 122430 }, { "epoch": 122.46, "grad_norm": 30.418960571289062, "learning_rate": 8.089300101892561e-07, "loss": 0.3517, "step": 122460 }, { "epoch": 122.49, "grad_norm": 93.41539001464844, "learning_rate": 8.072175969029832e-07, "loss": 0.2718, "step": 122490 }, { "epoch": 122.52, "grad_norm": 34.000553131103516, "learning_rate": 8.055068388608011e-07, "loss": 0.2803, "step": 122520 }, { "epoch": 122.55, "grad_norm": 55.75439453125, "learning_rate": 8.037977367380922e-07, "loss": 0.2392, "step": 122550 }, { "epoch": 122.58, "grad_norm": 66.77178955078125, "learning_rate": 8.020902912095807e-07, "loss": 0.3133, "step": 122580 }, { "epoch": 122.61, "grad_norm": 58.356624603271484, "learning_rate": 8.003845029493407e-07, "loss": 0.3344, "step": 122610 }, { "epoch": 122.64, "grad_norm": 39.11393737792969, "learning_rate": 7.986803726307901e-07, "loss": 0.3285, "step": 122640 }, { "epoch": 122.67, "grad_norm": 51.441226959228516, "learning_rate": 7.969779009266915e-07, "loss": 0.2687, "step": 122670 }, { "epoch": 122.7, "grad_norm": 43.96261215209961, "learning_rate": 7.952770885091548e-07, "loss": 0.3003, "step": 122700 }, { "epoch": 122.73, "grad_norm": 139.54248046875, "learning_rate": 7.935779360496337e-07, "loss": 0.2612, "step": 122730 }, { "epoch": 122.76, "grad_norm": 61.93021774291992, "learning_rate": 7.918804442189271e-07, "loss": 0.4367, "step": 122760 }, { "epoch": 122.79, "grad_norm": 73.30164337158203, "learning_rate": 7.901846136871766e-07, "loss": 0.2763, "step": 122790 }, { "epoch": 122.82, "grad_norm": 28.21575355529785, "learning_rate": 7.884904451238712e-07, "loss": 0.3018, "step": 122820 }, { "epoch": 122.85, "grad_norm": 3.031083724636119e-05, "learning_rate": 7.867979391978398e-07, "loss": 0.3096, "step": 122850 }, { "epoch": 122.88, "grad_norm": 69.66227722167969, "learning_rate": 7.851070965772572e-07, "loss": 0.2917, "step": 122880 }, { "epoch": 122.91, "grad_norm": 117.62500762939453, "learning_rate": 7.834179179296419e-07, "loss": 0.3295, "step": 122910 }, { "epoch": 122.94, "grad_norm": 78.29336547851562, "learning_rate": 7.81730403921856e-07, "loss": 0.3869, "step": 122940 }, { "epoch": 122.97, "grad_norm": 34.98335647583008, "learning_rate": 7.800445552201014e-07, "loss": 0.2978, "step": 122970 }, { "epoch": 123.0, "grad_norm": 66.36168670654297, "learning_rate": 7.783603724899258e-07, "loss": 0.3599, "step": 123000 }, { "epoch": 123.0, "eval_loss": 0.23615802824497223, "eval_map": 0.84, "eval_map_50": 0.9602, "eval_map_75": 0.9317, "eval_map_chicken": 0.8325, "eval_map_duck": 0.7897, "eval_map_large": 0.8521, "eval_map_medium": 0.8382, "eval_map_plant": 0.8978, "eval_map_small": 0.3253, "eval_mar_1": 0.3385, "eval_mar_10": 0.8713, "eval_mar_100": 0.8747, "eval_mar_100_chicken": 0.8775, "eval_mar_100_duck": 0.8253, "eval_mar_100_plant": 0.9214, "eval_mar_large": 0.8869, "eval_mar_medium": 0.8767, "eval_mar_small": 0.3742, "eval_runtime": 13.0714, "eval_samples_per_second": 15.301, "eval_steps_per_second": 1.913, "step": 123000 }, { "epoch": 123.03, "grad_norm": 37.66360855102539, "learning_rate": 7.766778563962152e-07, "loss": 0.3303, "step": 123030 }, { "epoch": 123.06, "grad_norm": 53.345645904541016, "learning_rate": 7.749970076032048e-07, "loss": 0.2474, "step": 123060 }, { "epoch": 123.09, "grad_norm": 41.67295455932617, "learning_rate": 7.733178267744634e-07, "loss": 0.3337, "step": 123090 }, { "epoch": 123.12, "grad_norm": 184.02207946777344, "learning_rate": 7.716403145729073e-07, "loss": 0.3489, "step": 123120 }, { "epoch": 123.15, "grad_norm": 76.17407989501953, "learning_rate": 7.699644716607896e-07, "loss": 0.267, "step": 123150 }, { "epoch": 123.18, "grad_norm": 30.68532371520996, "learning_rate": 7.682902986997076e-07, "loss": 0.265, "step": 123180 }, { "epoch": 123.21, "grad_norm": 38.67988967895508, "learning_rate": 7.666177963505989e-07, "loss": 0.316, "step": 123210 }, { "epoch": 123.24, "grad_norm": 51.03422164916992, "learning_rate": 7.649469652737407e-07, "loss": 0.3024, "step": 123240 }, { "epoch": 123.27, "grad_norm": 46.50461196899414, "learning_rate": 7.632778061287494e-07, "loss": 0.2915, "step": 123270 }, { "epoch": 123.3, "grad_norm": 94.87723541259766, "learning_rate": 7.61610319574585e-07, "loss": 0.3022, "step": 123300 }, { "epoch": 123.33, "grad_norm": 56.503021240234375, "learning_rate": 7.599445062695404e-07, "loss": 0.2911, "step": 123330 }, { "epoch": 123.36, "grad_norm": 43.81483459472656, "learning_rate": 7.582803668712579e-07, "loss": 0.2901, "step": 123360 }, { "epoch": 123.39, "grad_norm": 64.88619232177734, "learning_rate": 7.566179020367098e-07, "loss": 0.2843, "step": 123390 }, { "epoch": 123.42, "grad_norm": 62.56235122680664, "learning_rate": 7.549571124222127e-07, "loss": 0.314, "step": 123420 }, { "epoch": 123.45, "grad_norm": 368.8034362792969, "learning_rate": 7.532979986834177e-07, "loss": 0.2845, "step": 123450 }, { "epoch": 123.48, "grad_norm": 62.246337890625, "learning_rate": 7.51640561475318e-07, "loss": 0.2606, "step": 123480 }, { "epoch": 123.51, "grad_norm": 36.4081916809082, "learning_rate": 7.499848014522443e-07, "loss": 0.2976, "step": 123510 }, { "epoch": 123.54, "grad_norm": 40.07600402832031, "learning_rate": 7.48330719267864e-07, "loss": 0.299, "step": 123540 }, { "epoch": 123.57, "grad_norm": 39.588951110839844, "learning_rate": 7.466783155751816e-07, "loss": 0.3613, "step": 123570 }, { "epoch": 123.6, "grad_norm": 49.725975036621094, "learning_rate": 7.450275910265415e-07, "loss": 0.2703, "step": 123600 }, { "epoch": 123.63, "grad_norm": 6.096890865592286e-05, "learning_rate": 7.433785462736209e-07, "loss": 0.2786, "step": 123630 }, { "epoch": 123.66, "grad_norm": 69.359130859375, "learning_rate": 7.4173118196744e-07, "loss": 0.2697, "step": 123660 }, { "epoch": 123.69, "grad_norm": 29.63197135925293, "learning_rate": 7.4008549875835e-07, "loss": 0.2976, "step": 123690 }, { "epoch": 123.72, "grad_norm": 85.306640625, "learning_rate": 7.384414972960419e-07, "loss": 0.2892, "step": 123720 }, { "epoch": 123.75, "grad_norm": 52.62541580200195, "learning_rate": 7.367991782295392e-07, "loss": 0.2942, "step": 123750 }, { "epoch": 123.78, "grad_norm": 96.29012298583984, "learning_rate": 7.351585422072049e-07, "loss": 0.3642, "step": 123780 }, { "epoch": 123.81, "grad_norm": 65.81732940673828, "learning_rate": 7.335195898767367e-07, "loss": 0.3137, "step": 123810 }, { "epoch": 123.84, "grad_norm": 64.91571044921875, "learning_rate": 7.318823218851668e-07, "loss": 0.3077, "step": 123840 }, { "epoch": 123.87, "grad_norm": 3.174061203026213e-05, "learning_rate": 7.302467388788614e-07, "loss": 0.3104, "step": 123870 }, { "epoch": 123.9, "grad_norm": 40.80809783935547, "learning_rate": 7.286128415035249e-07, "loss": 0.2829, "step": 123900 }, { "epoch": 123.93, "grad_norm": 34.45760726928711, "learning_rate": 7.269806304041915e-07, "loss": 0.3077, "step": 123930 }, { "epoch": 123.96, "grad_norm": 31.825748443603516, "learning_rate": 7.253501062252338e-07, "loss": 0.4126, "step": 123960 }, { "epoch": 123.99, "grad_norm": 283.079345703125, "learning_rate": 7.237212696103568e-07, "loss": 0.2938, "step": 123990 }, { "epoch": 124.0, "eval_loss": 0.24026469886302948, "eval_map": 0.84, "eval_map_50": 0.9615, "eval_map_75": 0.935, "eval_map_chicken": 0.8314, "eval_map_duck": 0.7929, "eval_map_large": 0.8457, "eval_map_medium": 0.8403, "eval_map_plant": 0.8958, "eval_map_small": 0.3254, "eval_mar_1": 0.337, "eval_mar_10": 0.8711, "eval_mar_100": 0.8742, "eval_mar_100_chicken": 0.8753, "eval_mar_100_duck": 0.8263, "eval_mar_100_plant": 0.921, "eval_mar_large": 0.8804, "eval_mar_medium": 0.8781, "eval_mar_small": 0.3856, "eval_runtime": 12.897, "eval_samples_per_second": 15.507, "eval_steps_per_second": 1.938, "step": 124000 }, { "epoch": 124.02, "grad_norm": 78.27993774414062, "learning_rate": 7.220941212026005e-07, "loss": 0.2595, "step": 124020 }, { "epoch": 124.05, "grad_norm": 32.914878845214844, "learning_rate": 7.204686616443352e-07, "loss": 0.3383, "step": 124050 }, { "epoch": 124.08, "grad_norm": 89.54714965820312, "learning_rate": 7.188448915772673e-07, "loss": 0.2896, "step": 124080 }, { "epoch": 124.11, "grad_norm": 54.36312484741211, "learning_rate": 7.172228116424374e-07, "loss": 0.3349, "step": 124110 }, { "epoch": 124.14, "grad_norm": 0.0002560427528806031, "learning_rate": 7.156024224802139e-07, "loss": 0.2819, "step": 124140 }, { "epoch": 124.17, "grad_norm": 31.058639526367188, "learning_rate": 7.139837247303027e-07, "loss": 0.3045, "step": 124170 }, { "epoch": 124.2, "grad_norm": 113.28618621826172, "learning_rate": 7.123667190317396e-07, "loss": 0.2774, "step": 124200 }, { "epoch": 124.23, "grad_norm": 27.305908203125, "learning_rate": 7.107514060228921e-07, "loss": 0.291, "step": 124230 }, { "epoch": 124.26, "grad_norm": 41.28659439086914, "learning_rate": 7.091377863414611e-07, "loss": 0.2813, "step": 124260 }, { "epoch": 124.29, "grad_norm": 33.01529312133789, "learning_rate": 7.07525860624479e-07, "loss": 0.3154, "step": 124290 }, { "epoch": 124.32, "grad_norm": 121.67820739746094, "learning_rate": 7.059156295083064e-07, "loss": 0.32, "step": 124320 }, { "epoch": 124.35, "grad_norm": 3.980648398282938e-05, "learning_rate": 7.043070936286395e-07, "loss": 0.497, "step": 124350 }, { "epoch": 124.38, "grad_norm": 58.941593170166016, "learning_rate": 7.027002536204986e-07, "loss": 0.3235, "step": 124380 }, { "epoch": 124.41, "grad_norm": 42.37139129638672, "learning_rate": 7.010951101182439e-07, "loss": 0.3331, "step": 124410 }, { "epoch": 124.44, "grad_norm": 58.06804656982422, "learning_rate": 6.994916637555571e-07, "loss": 0.2945, "step": 124440 }, { "epoch": 124.47, "grad_norm": 31.65445899963379, "learning_rate": 6.978899151654556e-07, "loss": 0.2946, "step": 124470 }, { "epoch": 124.5, "grad_norm": 59.976715087890625, "learning_rate": 6.962898649802824e-07, "loss": 0.2631, "step": 124500 }, { "epoch": 124.53, "grad_norm": 48.193023681640625, "learning_rate": 6.946915138317129e-07, "loss": 0.2668, "step": 124530 }, { "epoch": 124.56, "grad_norm": 86.22926330566406, "learning_rate": 6.930948623507505e-07, "loss": 0.2721, "step": 124560 }, { "epoch": 124.59, "grad_norm": 53.43819808959961, "learning_rate": 6.914999111677295e-07, "loss": 0.2878, "step": 124590 }, { "epoch": 124.62, "grad_norm": 52.74860763549805, "learning_rate": 6.89906660912309e-07, "loss": 0.2701, "step": 124620 }, { "epoch": 124.65, "grad_norm": 63.48703384399414, "learning_rate": 6.883151122134812e-07, "loss": 0.2611, "step": 124650 }, { "epoch": 124.68, "grad_norm": 29.78040313720703, "learning_rate": 6.86725265699561e-07, "loss": 0.3021, "step": 124680 }, { "epoch": 124.71, "grad_norm": 120.81752014160156, "learning_rate": 6.851371219981989e-07, "loss": 0.323, "step": 124710 }, { "epoch": 124.74, "grad_norm": 70.92333984375, "learning_rate": 6.835506817363657e-07, "loss": 0.2239, "step": 124740 }, { "epoch": 124.77, "grad_norm": 70.9956283569336, "learning_rate": 6.819659455403654e-07, "loss": 0.4855, "step": 124770 }, { "epoch": 124.8, "grad_norm": 78.31332397460938, "learning_rate": 6.803829140358237e-07, "loss": 0.2877, "step": 124800 }, { "epoch": 124.83, "grad_norm": 75.43861389160156, "learning_rate": 6.788015878476983e-07, "loss": 0.3071, "step": 124830 }, { "epoch": 124.86, "grad_norm": 59.99592971801758, "learning_rate": 6.772219676002717e-07, "loss": 0.432, "step": 124860 }, { "epoch": 124.89, "grad_norm": 24.188940048217773, "learning_rate": 6.756440539171533e-07, "loss": 0.2994, "step": 124890 }, { "epoch": 124.92, "grad_norm": 41.65231704711914, "learning_rate": 6.74067847421277e-07, "loss": 0.262, "step": 124920 }, { "epoch": 124.95, "grad_norm": 66.81924438476562, "learning_rate": 6.724933487349061e-07, "loss": 0.3295, "step": 124950 }, { "epoch": 124.98, "grad_norm": 42.67380905151367, "learning_rate": 6.709205584796241e-07, "loss": 0.2533, "step": 124980 }, { "epoch": 125.0, "eval_loss": 0.2422100007534027, "eval_map": 0.8363, "eval_map_50": 0.9619, "eval_map_75": 0.9322, "eval_map_chicken": 0.8245, "eval_map_duck": 0.7883, "eval_map_large": 0.8377, "eval_map_medium": 0.8376, "eval_map_plant": 0.8962, "eval_map_small": 0.3405, "eval_mar_1": 0.3355, "eval_mar_10": 0.8677, "eval_mar_100": 0.8709, "eval_mar_100_chicken": 0.8686, "eval_mar_100_duck": 0.8232, "eval_mar_100_plant": 0.9208, "eval_mar_large": 0.8757, "eval_mar_medium": 0.8744, "eval_mar_small": 0.4067, "eval_runtime": 12.6956, "eval_samples_per_second": 15.753, "eval_steps_per_second": 1.969, "step": 125000 }, { "epoch": 125.01, "grad_norm": 54.74504852294922, "learning_rate": 6.693494772763487e-07, "loss": 0.3601, "step": 125010 }, { "epoch": 125.04, "grad_norm": 151.3241729736328, "learning_rate": 6.677801057453143e-07, "loss": 0.2874, "step": 125040 }, { "epoch": 125.07, "grad_norm": 29.46302032470703, "learning_rate": 6.662124445060863e-07, "loss": 0.336, "step": 125070 }, { "epoch": 125.1, "grad_norm": 56.365474700927734, "learning_rate": 6.646464941775499e-07, "loss": 0.3829, "step": 125100 }, { "epoch": 125.13, "grad_norm": 71.3403549194336, "learning_rate": 6.630822553779193e-07, "loss": 0.341, "step": 125130 }, { "epoch": 125.16, "grad_norm": 114.94991302490234, "learning_rate": 6.615197287247299e-07, "loss": 0.3265, "step": 125160 }, { "epoch": 125.19, "grad_norm": 52.7833251953125, "learning_rate": 6.599589148348451e-07, "loss": 0.2805, "step": 125190 }, { "epoch": 125.22, "grad_norm": 55.476470947265625, "learning_rate": 6.583998143244463e-07, "loss": 0.2886, "step": 125220 }, { "epoch": 125.25, "grad_norm": 29.350927352905273, "learning_rate": 6.568424278090446e-07, "loss": 0.3258, "step": 125250 }, { "epoch": 125.28, "grad_norm": 38.303245544433594, "learning_rate": 6.552867559034687e-07, "loss": 0.3075, "step": 125280 }, { "epoch": 125.31, "grad_norm": 47.25801467895508, "learning_rate": 6.537327992218745e-07, "loss": 0.2995, "step": 125310 }, { "epoch": 125.34, "grad_norm": 134.0709686279297, "learning_rate": 6.521805583777396e-07, "loss": 0.3237, "step": 125340 }, { "epoch": 125.37, "grad_norm": 83.15758514404297, "learning_rate": 6.506300339838656e-07, "loss": 0.2994, "step": 125370 }, { "epoch": 125.4, "grad_norm": 33.70622634887695, "learning_rate": 6.490812266523716e-07, "loss": 0.3003, "step": 125400 }, { "epoch": 125.43, "grad_norm": 38.261714935302734, "learning_rate": 6.475341369947047e-07, "loss": 0.3292, "step": 125430 }, { "epoch": 125.46, "grad_norm": 38.17567443847656, "learning_rate": 6.459887656216318e-07, "loss": 0.319, "step": 125460 }, { "epoch": 125.49, "grad_norm": 32.35736846923828, "learning_rate": 6.444451131432383e-07, "loss": 0.3452, "step": 125490 }, { "epoch": 125.52, "grad_norm": 73.73548889160156, "learning_rate": 6.429031801689362e-07, "loss": 0.3695, "step": 125520 }, { "epoch": 125.55, "grad_norm": 63.63335037231445, "learning_rate": 6.413629673074562e-07, "loss": 0.2671, "step": 125550 }, { "epoch": 125.58, "grad_norm": 61.89775085449219, "learning_rate": 6.398244751668481e-07, "loss": 0.4638, "step": 125580 }, { "epoch": 125.61, "grad_norm": 50.46353530883789, "learning_rate": 6.382877043544855e-07, "loss": 0.3056, "step": 125610 }, { "epoch": 125.64, "grad_norm": 62.52997970581055, "learning_rate": 6.36752655477062e-07, "loss": 0.2809, "step": 125640 }, { "epoch": 125.67, "grad_norm": 88.85185241699219, "learning_rate": 6.352193291405884e-07, "loss": 0.3362, "step": 125670 }, { "epoch": 125.7, "grad_norm": 100.70507049560547, "learning_rate": 6.336877259504004e-07, "loss": 0.3567, "step": 125700 }, { "epoch": 125.73, "grad_norm": 25.62355613708496, "learning_rate": 6.321578465111478e-07, "loss": 0.3117, "step": 125730 }, { "epoch": 125.76, "grad_norm": 61.03522491455078, "learning_rate": 6.30629691426804e-07, "loss": 0.285, "step": 125760 }, { "epoch": 125.79, "grad_norm": 73.00994873046875, "learning_rate": 6.291032613006604e-07, "loss": 0.268, "step": 125790 }, { "epoch": 125.82, "grad_norm": 51.71257400512695, "learning_rate": 6.275785567353293e-07, "loss": 0.275, "step": 125820 }, { "epoch": 125.85, "grad_norm": 86.31132507324219, "learning_rate": 6.260555783327366e-07, "loss": 0.3317, "step": 125850 }, { "epoch": 125.88, "grad_norm": 44.231224060058594, "learning_rate": 6.245343266941328e-07, "loss": 0.2886, "step": 125880 }, { "epoch": 125.91, "grad_norm": 35.50563049316406, "learning_rate": 6.23014802420081e-07, "loss": 0.2416, "step": 125910 }, { "epoch": 125.94, "grad_norm": 62.51797103881836, "learning_rate": 6.214970061104686e-07, "loss": 0.3304, "step": 125940 }, { "epoch": 125.97, "grad_norm": 43.939842224121094, "learning_rate": 6.199809383644956e-07, "loss": 0.3594, "step": 125970 }, { "epoch": 126.0, "grad_norm": 31.046838760375977, "learning_rate": 6.184665997806832e-07, "loss": 0.3822, "step": 126000 }, { "epoch": 126.0, "eval_loss": 0.24273502826690674, "eval_map": 0.8376, "eval_map_50": 0.9645, "eval_map_75": 0.9307, "eval_map_chicken": 0.8244, "eval_map_duck": 0.7953, "eval_map_large": 0.8477, "eval_map_medium": 0.8354, "eval_map_plant": 0.8932, "eval_map_small": 0.3178, "eval_mar_1": 0.3366, "eval_mar_10": 0.8695, "eval_mar_100": 0.8726, "eval_mar_100_chicken": 0.8696, "eval_mar_100_duck": 0.8299, "eval_mar_100_plant": 0.9182, "eval_mar_large": 0.8793, "eval_mar_medium": 0.8739, "eval_mar_small": 0.4005, "eval_runtime": 13.0225, "eval_samples_per_second": 15.358, "eval_steps_per_second": 1.92, "step": 126000 }, { "epoch": 126.03, "grad_norm": 30.411670684814453, "learning_rate": 6.169539909568656e-07, "loss": 0.3202, "step": 126030 }, { "epoch": 126.06, "grad_norm": 3.498875958030112e-05, "learning_rate": 6.154431124901983e-07, "loss": 0.4622, "step": 126060 }, { "epoch": 126.09, "grad_norm": 141.02651977539062, "learning_rate": 6.139339649771525e-07, "loss": 0.3271, "step": 126090 }, { "epoch": 126.12, "grad_norm": 49.18608474731445, "learning_rate": 6.124265490135161e-07, "loss": 0.2643, "step": 126120 }, { "epoch": 126.15, "grad_norm": 85.54029083251953, "learning_rate": 6.109208651943921e-07, "loss": 0.3162, "step": 126150 }, { "epoch": 126.18, "grad_norm": 50.23918533325195, "learning_rate": 6.094169141142014e-07, "loss": 0.2979, "step": 126180 }, { "epoch": 126.21, "grad_norm": 53.57966995239258, "learning_rate": 6.079146963666777e-07, "loss": 0.3825, "step": 126210 }, { "epoch": 126.24, "grad_norm": 40.75045394897461, "learning_rate": 6.064142125448763e-07, "loss": 0.3581, "step": 126240 }, { "epoch": 126.27, "grad_norm": 75.1888427734375, "learning_rate": 6.049154632411625e-07, "loss": 0.3291, "step": 126270 }, { "epoch": 126.3, "grad_norm": 48.21953582763672, "learning_rate": 6.034184490472195e-07, "loss": 0.3119, "step": 126300 }, { "epoch": 126.33, "grad_norm": 48.28759002685547, "learning_rate": 6.019231705540435e-07, "loss": 0.2807, "step": 126330 }, { "epoch": 126.36, "grad_norm": 77.10408020019531, "learning_rate": 6.004296283519478e-07, "loss": 0.2492, "step": 126360 }, { "epoch": 126.39, "grad_norm": 47.968502044677734, "learning_rate": 5.989378230305592e-07, "loss": 0.301, "step": 126390 }, { "epoch": 126.42, "grad_norm": 77.88301086425781, "learning_rate": 5.974477551788194e-07, "loss": 0.2294, "step": 126420 }, { "epoch": 126.45, "grad_norm": 77.80766296386719, "learning_rate": 5.959594253849821e-07, "loss": 0.3, "step": 126450 }, { "epoch": 126.48, "grad_norm": 104.58247375488281, "learning_rate": 5.944728342366179e-07, "loss": 0.3196, "step": 126480 }, { "epoch": 126.51, "grad_norm": 111.57868194580078, "learning_rate": 5.92987982320607e-07, "loss": 0.3264, "step": 126510 }, { "epoch": 126.54, "grad_norm": 108.09015655517578, "learning_rate": 5.915048702231491e-07, "loss": 0.3277, "step": 126540 }, { "epoch": 126.57, "grad_norm": 67.20681762695312, "learning_rate": 5.900234985297498e-07, "loss": 0.2748, "step": 126570 }, { "epoch": 126.6, "grad_norm": 34.24258804321289, "learning_rate": 5.885438678252342e-07, "loss": 0.3123, "step": 126600 }, { "epoch": 126.63, "grad_norm": 32.15423583984375, "learning_rate": 5.870659786937344e-07, "loss": 0.3126, "step": 126630 }, { "epoch": 126.66, "grad_norm": 70.37622833251953, "learning_rate": 5.855898317186992e-07, "loss": 0.2733, "step": 126660 }, { "epoch": 126.69, "grad_norm": 60.04710006713867, "learning_rate": 5.841154274828869e-07, "loss": 0.2866, "step": 126690 }, { "epoch": 126.72, "grad_norm": 36.75681686401367, "learning_rate": 5.826427665683715e-07, "loss": 0.3225, "step": 126720 }, { "epoch": 126.75, "grad_norm": 263.40679931640625, "learning_rate": 5.811718495565327e-07, "loss": 0.4722, "step": 126750 }, { "epoch": 126.78, "grad_norm": 27.41478157043457, "learning_rate": 5.797026770280683e-07, "loss": 0.3257, "step": 126780 }, { "epoch": 126.81, "grad_norm": 43.023651123046875, "learning_rate": 5.782352495629822e-07, "loss": 0.2747, "step": 126810 }, { "epoch": 126.84, "grad_norm": 33.9915771484375, "learning_rate": 5.767695677405921e-07, "loss": 0.2788, "step": 126840 }, { "epoch": 126.87, "grad_norm": 168.3291778564453, "learning_rate": 5.753056321395267e-07, "loss": 0.3006, "step": 126870 }, { "epoch": 126.9, "grad_norm": 44.38801956176758, "learning_rate": 5.738434433377244e-07, "loss": 0.257, "step": 126900 }, { "epoch": 126.93, "grad_norm": 41.41050720214844, "learning_rate": 5.723830019124332e-07, "loss": 0.2661, "step": 126930 }, { "epoch": 126.96, "grad_norm": 112.23653411865234, "learning_rate": 5.709243084402128e-07, "loss": 0.3153, "step": 126960 }, { "epoch": 126.99, "grad_norm": 150.71221923828125, "learning_rate": 5.694673634969334e-07, "loss": 0.3135, "step": 126990 }, { "epoch": 127.0, "eval_loss": 0.24615387618541718, "eval_map": 0.8335, "eval_map_50": 0.9645, "eval_map_75": 0.9305, "eval_map_chicken": 0.8224, "eval_map_duck": 0.7884, "eval_map_large": 0.8431, "eval_map_medium": 0.8314, "eval_map_plant": 0.8898, "eval_map_small": 0.3185, "eval_mar_1": 0.3348, "eval_mar_10": 0.8659, "eval_mar_100": 0.8691, "eval_mar_100_chicken": 0.8674, "eval_mar_100_duck": 0.8242, "eval_mar_100_plant": 0.9156, "eval_mar_large": 0.8765, "eval_mar_medium": 0.8697, "eval_mar_small": 0.4007, "eval_runtime": 12.6275, "eval_samples_per_second": 15.838, "eval_steps_per_second": 1.98, "step": 127000 }, { "epoch": 127.02, "grad_norm": 46.768829345703125, "learning_rate": 5.680121676577721e-07, "loss": 0.3093, "step": 127020 }, { "epoch": 127.05, "grad_norm": 44.9498291015625, "learning_rate": 5.665587214972173e-07, "loss": 0.3454, "step": 127050 }, { "epoch": 127.08, "grad_norm": 166.06199645996094, "learning_rate": 5.651070255890689e-07, "loss": 0.2819, "step": 127080 }, { "epoch": 127.11, "grad_norm": 71.01069641113281, "learning_rate": 5.6365708050643e-07, "loss": 0.2834, "step": 127110 }, { "epoch": 127.14, "grad_norm": 60.89025115966797, "learning_rate": 5.622088868217179e-07, "loss": 0.405, "step": 127140 }, { "epoch": 127.17, "grad_norm": 34.34323501586914, "learning_rate": 5.607624451066568e-07, "loss": 0.3836, "step": 127170 }, { "epoch": 127.2, "grad_norm": 121.6922378540039, "learning_rate": 5.593177559322776e-07, "loss": 0.2905, "step": 127200 }, { "epoch": 127.23, "grad_norm": 64.93509674072266, "learning_rate": 5.578748198689226e-07, "loss": 0.4497, "step": 127230 }, { "epoch": 127.26, "grad_norm": 41.17024612426758, "learning_rate": 5.564336374862373e-07, "loss": 0.3383, "step": 127260 }, { "epoch": 127.29, "grad_norm": 83.46958923339844, "learning_rate": 5.549942093531812e-07, "loss": 0.2255, "step": 127290 }, { "epoch": 127.32, "grad_norm": 62.96837615966797, "learning_rate": 5.535565360380146e-07, "loss": 0.2642, "step": 127320 }, { "epoch": 127.35, "grad_norm": 51.5461540222168, "learning_rate": 5.521206181083111e-07, "loss": 0.2369, "step": 127350 }, { "epoch": 127.38, "grad_norm": 157.1609649658203, "learning_rate": 5.506864561309455e-07, "loss": 0.2726, "step": 127380 }, { "epoch": 127.41, "grad_norm": 181.45111083984375, "learning_rate": 5.492540506721033e-07, "loss": 0.2918, "step": 127410 }, { "epoch": 127.44, "grad_norm": 185.44935607910156, "learning_rate": 5.478234022972756e-07, "loss": 0.2812, "step": 127440 }, { "epoch": 127.47, "grad_norm": 48.20285415649414, "learning_rate": 5.46394511571261e-07, "loss": 0.3158, "step": 127470 }, { "epoch": 127.5, "grad_norm": 58.19514083862305, "learning_rate": 5.449673790581611e-07, "loss": 0.2823, "step": 127500 }, { "epoch": 127.53, "grad_norm": 52.27482604980469, "learning_rate": 5.435420053213863e-07, "loss": 0.345, "step": 127530 }, { "epoch": 127.56, "grad_norm": 73.65628051757812, "learning_rate": 5.421183909236494e-07, "loss": 0.313, "step": 127560 }, { "epoch": 127.59, "grad_norm": 25.400362014770508, "learning_rate": 5.406965364269745e-07, "loss": 0.2838, "step": 127590 }, { "epoch": 127.62, "grad_norm": 47.6740608215332, "learning_rate": 5.392764423926844e-07, "loss": 0.2923, "step": 127620 }, { "epoch": 127.65, "grad_norm": 111.61190032958984, "learning_rate": 5.378581093814112e-07, "loss": 0.2951, "step": 127650 }, { "epoch": 127.68, "grad_norm": 58.15241622924805, "learning_rate": 5.364415379530891e-07, "loss": 0.3037, "step": 127680 }, { "epoch": 127.71, "grad_norm": 38.63493728637695, "learning_rate": 5.350267286669585e-07, "loss": 0.2793, "step": 127710 }, { "epoch": 127.74, "grad_norm": 149.19857788085938, "learning_rate": 5.33613682081564e-07, "loss": 0.3354, "step": 127740 }, { "epoch": 127.77, "grad_norm": 67.32189178466797, "learning_rate": 5.322023987547547e-07, "loss": 0.2995, "step": 127770 }, { "epoch": 127.8, "grad_norm": 40.641746520996094, "learning_rate": 5.307928792436812e-07, "loss": 0.2987, "step": 127800 }, { "epoch": 127.83, "grad_norm": 62.0726318359375, "learning_rate": 5.293851241048015e-07, "loss": 0.3321, "step": 127830 }, { "epoch": 127.86, "grad_norm": 64.79600524902344, "learning_rate": 5.279791338938717e-07, "loss": 0.3556, "step": 127860 }, { "epoch": 127.89, "grad_norm": 37.74106979370117, "learning_rate": 5.265749091659589e-07, "loss": 0.3038, "step": 127890 }, { "epoch": 127.92, "grad_norm": 65.1868667602539, "learning_rate": 5.251724504754258e-07, "loss": 0.3043, "step": 127920 }, { "epoch": 127.95, "grad_norm": 65.49906158447266, "learning_rate": 5.237717583759421e-07, "loss": 0.3259, "step": 127950 }, { "epoch": 127.98, "grad_norm": 54.348411560058594, "learning_rate": 5.22372833420478e-07, "loss": 0.4718, "step": 127980 }, { "epoch": 128.0, "eval_loss": 0.24144622683525085, "eval_map": 0.8367, "eval_map_50": 0.9644, "eval_map_75": 0.9302, "eval_map_chicken": 0.8257, "eval_map_duck": 0.7906, "eval_map_large": 0.8441, "eval_map_medium": 0.8356, "eval_map_plant": 0.8938, "eval_map_small": 0.3142, "eval_mar_1": 0.3355, "eval_mar_10": 0.8685, "eval_mar_100": 0.8717, "eval_mar_100_chicken": 0.8702, "eval_mar_100_duck": 0.8258, "eval_mar_100_plant": 0.9191, "eval_mar_large": 0.8785, "eval_mar_medium": 0.8733, "eval_mar_small": 0.3959, "eval_runtime": 12.8001, "eval_samples_per_second": 15.625, "eval_steps_per_second": 1.953, "step": 128000 }, { "epoch": 128.01, "grad_norm": 45.438018798828125, "learning_rate": 5.209756761613072e-07, "loss": 0.3013, "step": 128010 }, { "epoch": 128.04, "grad_norm": 75.49828338623047, "learning_rate": 5.19580287150005e-07, "loss": 0.3907, "step": 128040 }, { "epoch": 128.07, "grad_norm": 74.12843322753906, "learning_rate": 5.181866669374507e-07, "loss": 0.3114, "step": 128070 }, { "epoch": 128.1, "grad_norm": 25.690052032470703, "learning_rate": 5.167948160738206e-07, "loss": 0.3193, "step": 128100 }, { "epoch": 128.13, "grad_norm": 4.024134250357747e-05, "learning_rate": 5.154047351085983e-07, "loss": 0.2855, "step": 128130 }, { "epoch": 128.16, "grad_norm": 37.38752746582031, "learning_rate": 5.140164245905633e-07, "loss": 0.239, "step": 128160 }, { "epoch": 128.19, "grad_norm": 34.369468688964844, "learning_rate": 5.126298850677991e-07, "loss": 0.2496, "step": 128190 }, { "epoch": 128.22, "grad_norm": 65.20069122314453, "learning_rate": 5.112451170876903e-07, "loss": 0.3673, "step": 128220 }, { "epoch": 128.25, "grad_norm": 0.00011308489047223702, "learning_rate": 5.098621211969224e-07, "loss": 0.2792, "step": 128250 }, { "epoch": 128.28, "grad_norm": 65.7624740600586, "learning_rate": 5.084808979414779e-07, "loss": 0.2568, "step": 128280 }, { "epoch": 128.31, "grad_norm": 65.5888900756836, "learning_rate": 5.071014478666425e-07, "loss": 0.2661, "step": 128310 }, { "epoch": 128.34, "grad_norm": 2.9726164939347655e-05, "learning_rate": 5.057237715170032e-07, "loss": 0.3115, "step": 128340 }, { "epoch": 128.37, "grad_norm": 41.98238754272461, "learning_rate": 5.043478694364423e-07, "loss": 0.3439, "step": 128370 }, { "epoch": 128.4, "grad_norm": 24.932043075561523, "learning_rate": 5.029737421681446e-07, "loss": 0.3262, "step": 128400 }, { "epoch": 128.43, "grad_norm": 43.97943878173828, "learning_rate": 5.016013902545957e-07, "loss": 0.3036, "step": 128430 }, { "epoch": 128.46, "grad_norm": 25.283605575561523, "learning_rate": 5.002308142375762e-07, "loss": 0.286, "step": 128460 }, { "epoch": 128.49, "grad_norm": 41.36478042602539, "learning_rate": 4.988620146581685e-07, "loss": 0.3752, "step": 128490 }, { "epoch": 128.52, "grad_norm": 66.25089263916016, "learning_rate": 4.97494992056754e-07, "loss": 0.2878, "step": 128520 }, { "epoch": 128.55, "grad_norm": 35.06134796142578, "learning_rate": 4.961297469730097e-07, "loss": 0.2411, "step": 128550 }, { "epoch": 128.58, "grad_norm": 38.56977462768555, "learning_rate": 4.947662799459152e-07, "loss": 0.2539, "step": 128580 }, { "epoch": 128.61, "grad_norm": 68.8032455444336, "learning_rate": 4.934045915137419e-07, "loss": 0.2853, "step": 128610 }, { "epoch": 128.64, "grad_norm": 39.095062255859375, "learning_rate": 4.920446822140673e-07, "loss": 0.2807, "step": 128640 }, { "epoch": 128.67, "grad_norm": 75.20214080810547, "learning_rate": 4.90686552583759e-07, "loss": 0.262, "step": 128670 }, { "epoch": 128.7, "grad_norm": 46.51348876953125, "learning_rate": 4.893302031589864e-07, "loss": 0.2495, "step": 128700 }, { "epoch": 128.73, "grad_norm": 25.428678512573242, "learning_rate": 4.87975634475214e-07, "loss": 0.4572, "step": 128730 }, { "epoch": 128.76, "grad_norm": 51.83258056640625, "learning_rate": 4.866228470672041e-07, "loss": 0.3089, "step": 128760 }, { "epoch": 128.79, "grad_norm": 52.3604850769043, "learning_rate": 4.852718414690166e-07, "loss": 0.2806, "step": 128790 }, { "epoch": 128.82, "grad_norm": 32.221580505371094, "learning_rate": 4.839226182140072e-07, "loss": 0.394, "step": 128820 }, { "epoch": 128.85, "grad_norm": 74.38313293457031, "learning_rate": 4.825751778348259e-07, "loss": 0.358, "step": 128850 }, { "epoch": 128.88, "grad_norm": 47.746925354003906, "learning_rate": 4.812295208634238e-07, "loss": 0.2733, "step": 128880 }, { "epoch": 128.91, "grad_norm": 114.18741607666016, "learning_rate": 4.798856478310409e-07, "loss": 0.3309, "step": 128910 }, { "epoch": 128.94, "grad_norm": 86.55680847167969, "learning_rate": 4.78543559268222e-07, "loss": 0.2319, "step": 128940 }, { "epoch": 128.97, "grad_norm": 2.6186717150267214e-05, "learning_rate": 4.772032557047984e-07, "loss": 0.2603, "step": 128970 }, { "epoch": 129.0, "grad_norm": 56.36320114135742, "learning_rate": 4.758647376699033e-07, "loss": 0.2618, "step": 129000 }, { "epoch": 129.0, "eval_loss": 0.24308738112449646, "eval_map": 0.8372, "eval_map_50": 0.9628, "eval_map_75": 0.9374, "eval_map_chicken": 0.8263, "eval_map_duck": 0.7917, "eval_map_large": 0.8481, "eval_map_medium": 0.8355, "eval_map_plant": 0.8936, "eval_map_small": 0.3165, "eval_mar_1": 0.336, "eval_mar_10": 0.8684, "eval_mar_100": 0.8716, "eval_mar_100_chicken": 0.8688, "eval_mar_100_duck": 0.8268, "eval_mar_100_plant": 0.9191, "eval_mar_large": 0.88, "eval_mar_medium": 0.8742, "eval_mar_small": 0.39, "eval_runtime": 13.4115, "eval_samples_per_second": 14.913, "eval_steps_per_second": 1.864, "step": 129000 }, { "epoch": 129.03, "grad_norm": 72.43539428710938, "learning_rate": 4.7452800569195987e-07, "loss": 0.34, "step": 129030 }, { "epoch": 129.06, "grad_norm": 52.626033782958984, "learning_rate": 4.731930602986906e-07, "loss": 0.321, "step": 129060 }, { "epoch": 129.09, "grad_norm": 148.2350311279297, "learning_rate": 4.718599020171105e-07, "loss": 0.3197, "step": 129090 }, { "epoch": 129.12, "grad_norm": 86.8702392578125, "learning_rate": 4.705285313735297e-07, "loss": 0.3602, "step": 129120 }, { "epoch": 129.15, "grad_norm": 113.36012268066406, "learning_rate": 4.691989488935511e-07, "loss": 0.2808, "step": 129150 }, { "epoch": 129.18, "grad_norm": 69.45126342773438, "learning_rate": 4.678711551020743e-07, "loss": 0.2461, "step": 129180 }, { "epoch": 129.21, "grad_norm": 82.2297592163086, "learning_rate": 4.665451505232882e-07, "loss": 0.3019, "step": 129210 }, { "epoch": 129.24, "grad_norm": 38.60065460205078, "learning_rate": 4.6522093568068307e-07, "loss": 0.2908, "step": 129240 }, { "epoch": 129.27, "grad_norm": 120.44233703613281, "learning_rate": 4.63898511097034e-07, "loss": 0.4175, "step": 129270 }, { "epoch": 129.3, "grad_norm": 53.63652038574219, "learning_rate": 4.625778772944156e-07, "loss": 0.3221, "step": 129300 }, { "epoch": 129.33, "grad_norm": 47.46669006347656, "learning_rate": 4.61259034794192e-07, "loss": 0.2699, "step": 129330 }, { "epoch": 129.36, "grad_norm": 143.14111328125, "learning_rate": 4.599419841170216e-07, "loss": 0.2308, "step": 129360 }, { "epoch": 129.39, "grad_norm": 87.29920959472656, "learning_rate": 4.5862672578285475e-07, "loss": 0.2662, "step": 129390 }, { "epoch": 129.42, "grad_norm": 38.587608337402344, "learning_rate": 4.5731326031093645e-07, "loss": 0.2953, "step": 129420 }, { "epoch": 129.45, "grad_norm": 105.91064453125, "learning_rate": 4.5600158821979933e-07, "loss": 0.3093, "step": 129450 }, { "epoch": 129.48, "grad_norm": 71.81896209716797, "learning_rate": 4.546917100272735e-07, "loss": 0.3704, "step": 129480 }, { "epoch": 129.51, "grad_norm": 84.76165771484375, "learning_rate": 4.533836262504759e-07, "loss": 0.2853, "step": 129510 }, { "epoch": 129.54, "grad_norm": 49.157142639160156, "learning_rate": 4.520773374058179e-07, "loss": 0.2698, "step": 129540 }, { "epoch": 129.57, "grad_norm": 48.50545883178711, "learning_rate": 4.5077284400900147e-07, "loss": 0.377, "step": 129570 }, { "epoch": 129.6, "grad_norm": 114.16522216796875, "learning_rate": 4.494701465750217e-07, "loss": 0.3035, "step": 129600 }, { "epoch": 129.63, "grad_norm": 65.99445343017578, "learning_rate": 4.4816924561816076e-07, "loss": 0.2571, "step": 129630 }, { "epoch": 129.66, "grad_norm": 111.5808334350586, "learning_rate": 4.4687014165199547e-07, "loss": 0.3287, "step": 129660 }, { "epoch": 129.69, "grad_norm": 78.54638671875, "learning_rate": 4.455728351893895e-07, "loss": 0.2731, "step": 129690 }, { "epoch": 129.72, "grad_norm": 53.07883071899414, "learning_rate": 4.4427732674250045e-07, "loss": 0.3556, "step": 129720 }, { "epoch": 129.75, "grad_norm": 87.5382308959961, "learning_rate": 4.4298361682277355e-07, "loss": 0.2982, "step": 129750 }, { "epoch": 129.78, "grad_norm": 46.82167434692383, "learning_rate": 4.416917059409465e-07, "loss": 0.3209, "step": 129780 }, { "epoch": 129.81, "grad_norm": 55.39387512207031, "learning_rate": 4.404015946070439e-07, "loss": 0.3322, "step": 129810 }, { "epoch": 129.84, "grad_norm": 2.3562819478684105e-05, "learning_rate": 4.391132833303807e-07, "loss": 0.2358, "step": 129840 }, { "epoch": 129.87, "grad_norm": 38.88970947265625, "learning_rate": 4.378267726195645e-07, "loss": 0.3984, "step": 129870 }, { "epoch": 129.9, "grad_norm": 187.3661346435547, "learning_rate": 4.3654206298248625e-07, "loss": 0.2708, "step": 129900 }, { "epoch": 129.93, "grad_norm": 77.7891845703125, "learning_rate": 4.352591549263302e-07, "loss": 0.3104, "step": 129930 }, { "epoch": 129.96, "grad_norm": 0.00016013698768801987, "learning_rate": 4.3397804895756957e-07, "loss": 0.281, "step": 129960 }, { "epoch": 129.99, "grad_norm": 47.47621154785156, "learning_rate": 4.32698745581962e-07, "loss": 0.3085, "step": 129990 }, { "epoch": 130.0, "eval_loss": 0.24088603258132935, "eval_map": 0.8389, "eval_map_50": 0.9647, "eval_map_75": 0.9338, "eval_map_chicken": 0.8298, "eval_map_duck": 0.7931, "eval_map_large": 0.8435, "eval_map_medium": 0.8395, "eval_map_plant": 0.8938, "eval_map_small": 0.3262, "eval_mar_1": 0.3365, "eval_mar_10": 0.8714, "eval_mar_100": 0.8748, "eval_mar_100_chicken": 0.8738, "eval_mar_100_duck": 0.8299, "eval_mar_100_plant": 0.9207, "eval_mar_large": 0.8782, "eval_mar_medium": 0.8795, "eval_mar_small": 0.4021, "eval_runtime": 13.0706, "eval_samples_per_second": 15.302, "eval_steps_per_second": 1.913, "step": 130000 }, { "epoch": 130.02, "grad_norm": 47.53729248046875, "learning_rate": 4.31421245304558e-07, "loss": 0.2705, "step": 130020 }, { "epoch": 130.05, "grad_norm": 69.62460327148438, "learning_rate": 4.301455486296946e-07, "loss": 0.3002, "step": 130050 }, { "epoch": 130.08, "grad_norm": 64.60865020751953, "learning_rate": 4.2887165606099513e-07, "loss": 0.3145, "step": 130080 }, { "epoch": 130.11, "grad_norm": 39.483036041259766, "learning_rate": 4.275995681013745e-07, "loss": 0.3141, "step": 130110 }, { "epoch": 130.14, "grad_norm": 31.121074676513672, "learning_rate": 4.263292852530293e-07, "loss": 0.3361, "step": 130140 }, { "epoch": 130.17, "grad_norm": 44.49150848388672, "learning_rate": 4.250608080174512e-07, "loss": 0.2798, "step": 130170 }, { "epoch": 130.2, "grad_norm": 45.69865798950195, "learning_rate": 4.237941368954124e-07, "loss": 0.2961, "step": 130200 }, { "epoch": 130.23, "grad_norm": 297.6024169921875, "learning_rate": 4.225292723869762e-07, "loss": 0.3922, "step": 130230 }, { "epoch": 130.26, "grad_norm": 187.6627655029297, "learning_rate": 4.212662149914887e-07, "loss": 0.3099, "step": 130260 }, { "epoch": 130.29, "grad_norm": 105.57393646240234, "learning_rate": 4.200049652075866e-07, "loss": 0.3199, "step": 130290 }, { "epoch": 130.32, "grad_norm": 140.3778839111328, "learning_rate": 4.1874552353319107e-07, "loss": 0.2458, "step": 130320 }, { "epoch": 130.35, "grad_norm": 71.5981216430664, "learning_rate": 4.1748789046551055e-07, "loss": 0.3115, "step": 130350 }, { "epoch": 130.38, "grad_norm": 22.47512435913086, "learning_rate": 4.162320665010372e-07, "loss": 0.2711, "step": 130380 }, { "epoch": 130.41, "grad_norm": 129.10415649414062, "learning_rate": 4.149780521355523e-07, "loss": 0.3628, "step": 130410 }, { "epoch": 130.44, "grad_norm": 49.88002395629883, "learning_rate": 4.137258478641176e-07, "loss": 0.2835, "step": 130440 }, { "epoch": 130.47, "grad_norm": 67.61511993408203, "learning_rate": 4.1247545418108715e-07, "loss": 0.3411, "step": 130470 }, { "epoch": 130.5, "grad_norm": 98.08781433105469, "learning_rate": 4.112268715800943e-07, "loss": 0.301, "step": 130500 }, { "epoch": 130.53, "grad_norm": 46.390193939208984, "learning_rate": 4.099801005540616e-07, "loss": 0.2877, "step": 130530 }, { "epoch": 130.56, "grad_norm": 122.86546325683594, "learning_rate": 4.087351415951918e-07, "loss": 0.313, "step": 130560 }, { "epoch": 130.59, "grad_norm": 58.362483978271484, "learning_rate": 4.0749199519497686e-07, "loss": 0.291, "step": 130590 }, { "epoch": 130.62, "grad_norm": 29.56102180480957, "learning_rate": 4.062506618441908e-07, "loss": 0.2641, "step": 130620 }, { "epoch": 130.65, "grad_norm": 40.742366790771484, "learning_rate": 4.0501114203289395e-07, "loss": 0.3561, "step": 130650 }, { "epoch": 130.68, "grad_norm": 56.48892593383789, "learning_rate": 4.0377343625042587e-07, "loss": 0.3385, "step": 130680 }, { "epoch": 130.71, "grad_norm": 127.87748718261719, "learning_rate": 4.025375449854163e-07, "loss": 0.2796, "step": 130710 }, { "epoch": 130.74, "grad_norm": 50.04518127441406, "learning_rate": 4.013034687257727e-07, "loss": 0.2926, "step": 130740 }, { "epoch": 130.77, "grad_norm": 137.6788330078125, "learning_rate": 4.000712079586916e-07, "loss": 0.2775, "step": 130770 }, { "epoch": 130.8, "grad_norm": 37.225006103515625, "learning_rate": 3.9884076317064813e-07, "loss": 0.2995, "step": 130800 }, { "epoch": 130.83, "grad_norm": 54.066131591796875, "learning_rate": 3.9761213484740435e-07, "loss": 0.278, "step": 130830 }, { "epoch": 130.86, "grad_norm": 54.09221649169922, "learning_rate": 3.963853234740006e-07, "loss": 0.4246, "step": 130860 }, { "epoch": 130.89, "grad_norm": 33.868167877197266, "learning_rate": 3.951603295347639e-07, "loss": 0.3073, "step": 130890 }, { "epoch": 130.92, "grad_norm": 40.55158996582031, "learning_rate": 3.9393715351330243e-07, "loss": 0.3718, "step": 130920 }, { "epoch": 130.95, "grad_norm": 39.677825927734375, "learning_rate": 3.9271579589250817e-07, "loss": 0.3453, "step": 130950 }, { "epoch": 130.98, "grad_norm": 71.55697631835938, "learning_rate": 3.9149625715455107e-07, "loss": 0.3682, "step": 130980 }, { "epoch": 131.0, "eval_loss": 0.2402298003435135, "eval_map": 0.8385, "eval_map_50": 0.9641, "eval_map_75": 0.9329, "eval_map_chicken": 0.8265, "eval_map_duck": 0.7923, "eval_map_large": 0.8421, "eval_map_medium": 0.8393, "eval_map_plant": 0.8966, "eval_map_small": 0.3243, "eval_mar_1": 0.3366, "eval_mar_10": 0.8699, "eval_mar_100": 0.8734, "eval_mar_100_chicken": 0.871, "eval_mar_100_duck": 0.8278, "eval_mar_100_plant": 0.9213, "eval_mar_large": 0.8787, "eval_mar_medium": 0.8773, "eval_mar_small": 0.3914, "eval_runtime": 12.9916, "eval_samples_per_second": 15.395, "eval_steps_per_second": 1.924, "step": 131000 }, { "epoch": 131.01, "grad_norm": 38.73418426513672, "learning_rate": 3.902785377808882e-07, "loss": 0.2881, "step": 131010 }, { "epoch": 131.04, "grad_norm": 47.15311813354492, "learning_rate": 3.890626382522539e-07, "loss": 0.2847, "step": 131040 }, { "epoch": 131.07, "grad_norm": 60.66126251220703, "learning_rate": 3.8784855904866637e-07, "loss": 0.3206, "step": 131070 }, { "epoch": 131.1, "grad_norm": 68.07306671142578, "learning_rate": 3.866363006494256e-07, "loss": 0.3064, "step": 131100 }, { "epoch": 131.13, "grad_norm": 81.74397277832031, "learning_rate": 3.8542586353311264e-07, "loss": 0.301, "step": 131130 }, { "epoch": 131.16, "grad_norm": 87.38013458251953, "learning_rate": 3.8421724817758745e-07, "loss": 0.2594, "step": 131160 }, { "epoch": 131.19, "grad_norm": 64.26652526855469, "learning_rate": 3.830104550599922e-07, "loss": 0.3123, "step": 131190 }, { "epoch": 131.22, "grad_norm": 55.31968307495117, "learning_rate": 3.818054846567515e-07, "loss": 0.3663, "step": 131220 }, { "epoch": 131.25, "grad_norm": 24.944095611572266, "learning_rate": 3.8060233744356634e-07, "loss": 0.3323, "step": 131250 }, { "epoch": 131.28, "grad_norm": 62.49150848388672, "learning_rate": 3.794010138954213e-07, "loss": 0.2889, "step": 131280 }, { "epoch": 131.31, "grad_norm": 92.13021850585938, "learning_rate": 3.782015144865808e-07, "loss": 0.3182, "step": 131310 }, { "epoch": 131.34, "grad_norm": 116.17314910888672, "learning_rate": 3.770038396905862e-07, "loss": 0.3192, "step": 131340 }, { "epoch": 131.37, "grad_norm": 63.436214447021484, "learning_rate": 3.758079899802619e-07, "loss": 0.2634, "step": 131370 }, { "epoch": 131.4, "grad_norm": 47.677650451660156, "learning_rate": 3.7461396582771035e-07, "loss": 0.2652, "step": 131400 }, { "epoch": 131.43, "grad_norm": 90.6214599609375, "learning_rate": 3.7342176770431284e-07, "loss": 0.3176, "step": 131430 }, { "epoch": 131.46, "grad_norm": 47.414894104003906, "learning_rate": 3.7223139608073e-07, "loss": 0.2473, "step": 131460 }, { "epoch": 131.49, "grad_norm": 37.193355560302734, "learning_rate": 3.710428514269027e-07, "loss": 0.3486, "step": 131490 }, { "epoch": 131.52, "grad_norm": 131.3379669189453, "learning_rate": 3.698561342120499e-07, "loss": 0.3077, "step": 131520 }, { "epoch": 131.55, "grad_norm": 73.29032135009766, "learning_rate": 3.6867124490466697e-07, "loss": 0.3132, "step": 131550 }, { "epoch": 131.58, "grad_norm": 253.30690002441406, "learning_rate": 3.674881839725314e-07, "loss": 0.3526, "step": 131580 }, { "epoch": 131.61, "grad_norm": 77.48766326904297, "learning_rate": 3.6630695188269505e-07, "loss": 0.3444, "step": 131610 }, { "epoch": 131.64, "grad_norm": 49.08409118652344, "learning_rate": 3.651275491014905e-07, "loss": 0.3102, "step": 131640 }, { "epoch": 131.67, "grad_norm": 41.39360427856445, "learning_rate": 3.6394997609452755e-07, "loss": 0.2959, "step": 131670 }, { "epoch": 131.7, "grad_norm": 189.17108154296875, "learning_rate": 3.627742333266937e-07, "loss": 0.3414, "step": 131700 }, { "epoch": 131.73, "grad_norm": 59.7651252746582, "learning_rate": 3.6160032126215274e-07, "loss": 0.3362, "step": 131730 }, { "epoch": 131.76, "grad_norm": 36.15788650512695, "learning_rate": 3.604282403643472e-07, "loss": 0.2959, "step": 131760 }, { "epoch": 131.79, "grad_norm": 55.95381546020508, "learning_rate": 3.5925799109599426e-07, "loss": 0.3011, "step": 131790 }, { "epoch": 131.82, "grad_norm": 3.4396427508909255e-05, "learning_rate": 3.5808957391909315e-07, "loss": 0.3618, "step": 131820 }, { "epoch": 131.85, "grad_norm": 36.6750602722168, "learning_rate": 3.569229892949133e-07, "loss": 0.2505, "step": 131850 }, { "epoch": 131.88, "grad_norm": 142.3834228515625, "learning_rate": 3.557582376840063e-07, "loss": 0.2953, "step": 131880 }, { "epoch": 131.91, "grad_norm": 3.971091427956708e-05, "learning_rate": 3.545953195461954e-07, "loss": 0.3077, "step": 131910 }, { "epoch": 131.94, "grad_norm": 43.25530242919922, "learning_rate": 3.534342353405834e-07, "loss": 0.3041, "step": 131940 }, { "epoch": 131.97, "grad_norm": 81.1157455444336, "learning_rate": 3.5227498552554805e-07, "loss": 0.3349, "step": 131970 }, { "epoch": 132.0, "grad_norm": 71.26033782958984, "learning_rate": 3.511175705587433e-07, "loss": 0.2999, "step": 132000 }, { "epoch": 132.0, "eval_loss": 0.24112722277641296, "eval_map": 0.8374, "eval_map_50": 0.9641, "eval_map_75": 0.9296, "eval_map_chicken": 0.8243, "eval_map_duck": 0.7922, "eval_map_large": 0.8448, "eval_map_medium": 0.8377, "eval_map_plant": 0.8957, "eval_map_small": 0.3115, "eval_mar_1": 0.3362, "eval_mar_10": 0.869, "eval_mar_100": 0.8722, "eval_mar_100_chicken": 0.8702, "eval_mar_100_duck": 0.8263, "eval_mar_100_plant": 0.9201, "eval_mar_large": 0.8773, "eval_mar_medium": 0.8761, "eval_mar_small": 0.3776, "eval_runtime": 12.9848, "eval_samples_per_second": 15.403, "eval_steps_per_second": 1.925, "step": 132000 }, { "epoch": 132.03, "grad_norm": 102.34610748291016, "learning_rate": 3.4996199089709695e-07, "loss": 0.2605, "step": 132030 }, { "epoch": 132.06, "grad_norm": 75.8858642578125, "learning_rate": 3.488082469968146e-07, "loss": 0.2788, "step": 132060 }, { "epoch": 132.09, "grad_norm": 110.1533203125, "learning_rate": 3.476563393133747e-07, "loss": 0.3417, "step": 132090 }, { "epoch": 132.12, "grad_norm": 3.904820914613083e-05, "learning_rate": 3.465062683015341e-07, "loss": 0.2913, "step": 132120 }, { "epoch": 132.15, "grad_norm": 76.44367980957031, "learning_rate": 3.4535803441532125e-07, "loss": 0.274, "step": 132150 }, { "epoch": 132.18, "grad_norm": 73.05992889404297, "learning_rate": 3.442116381080418e-07, "loss": 0.3166, "step": 132180 }, { "epoch": 132.21, "grad_norm": 42.857181549072266, "learning_rate": 3.430670798322733e-07, "loss": 0.2873, "step": 132210 }, { "epoch": 132.24, "grad_norm": 0.00011815642938017845, "learning_rate": 3.419243600398703e-07, "loss": 0.3181, "step": 132240 }, { "epoch": 132.27, "grad_norm": 49.21725082397461, "learning_rate": 3.407834791819603e-07, "loss": 0.3477, "step": 132270 }, { "epoch": 132.3, "grad_norm": 75.97196960449219, "learning_rate": 3.396444377089453e-07, "loss": 0.2409, "step": 132300 }, { "epoch": 132.33, "grad_norm": 95.13970184326172, "learning_rate": 3.3850723607049994e-07, "loss": 0.284, "step": 132330 }, { "epoch": 132.36, "grad_norm": 42.67320251464844, "learning_rate": 3.373718747155752e-07, "loss": 0.3891, "step": 132360 }, { "epoch": 132.39, "grad_norm": 122.92071533203125, "learning_rate": 3.3623835409239023e-07, "loss": 0.3954, "step": 132390 }, { "epoch": 132.42, "grad_norm": 48.22465515136719, "learning_rate": 3.351066746484455e-07, "loss": 0.261, "step": 132420 }, { "epoch": 132.45, "grad_norm": 82.07262420654297, "learning_rate": 3.3397683683050685e-07, "loss": 0.3348, "step": 132450 }, { "epoch": 132.48, "grad_norm": 89.3011474609375, "learning_rate": 3.328488410846187e-07, "loss": 0.2351, "step": 132480 }, { "epoch": 132.51, "grad_norm": 43.032745361328125, "learning_rate": 3.317226878560931e-07, "loss": 0.3182, "step": 132510 }, { "epoch": 132.54, "grad_norm": 50.274810791015625, "learning_rate": 3.3059837758951995e-07, "loss": 0.3633, "step": 132540 }, { "epoch": 132.57, "grad_norm": 85.09639739990234, "learning_rate": 3.294759107287582e-07, "loss": 0.3046, "step": 132570 }, { "epoch": 132.6, "grad_norm": 82.05386352539062, "learning_rate": 3.283552877169399e-07, "loss": 0.2617, "step": 132600 }, { "epoch": 132.63, "grad_norm": 56.337608337402344, "learning_rate": 3.272365089964691e-07, "loss": 0.2844, "step": 132630 }, { "epoch": 132.66, "grad_norm": 67.90081024169922, "learning_rate": 3.2611957500902345e-07, "loss": 0.3603, "step": 132660 }, { "epoch": 132.69, "grad_norm": 26.008312225341797, "learning_rate": 3.250044861955487e-07, "loss": 0.2229, "step": 132690 }, { "epoch": 132.72, "grad_norm": 33.81447982788086, "learning_rate": 3.2389124299626483e-07, "loss": 0.3146, "step": 132720 }, { "epoch": 132.75, "grad_norm": 45.818565368652344, "learning_rate": 3.227798458506637e-07, "loss": 0.2921, "step": 132750 }, { "epoch": 132.78, "grad_norm": 2.555555875005666e-05, "learning_rate": 3.216702951975059e-07, "loss": 0.3041, "step": 132780 }, { "epoch": 132.81, "grad_norm": 650.5856323242188, "learning_rate": 3.205625914748256e-07, "loss": 0.3248, "step": 132810 }, { "epoch": 132.84, "grad_norm": 54.398929595947266, "learning_rate": 3.194567351199257e-07, "loss": 0.2801, "step": 132840 }, { "epoch": 132.87, "grad_norm": 47.778228759765625, "learning_rate": 3.18352726569382e-07, "loss": 0.3551, "step": 132870 }, { "epoch": 132.9, "grad_norm": 48.30999755859375, "learning_rate": 3.172505662590386e-07, "loss": 0.3678, "step": 132900 }, { "epoch": 132.93, "grad_norm": 114.25973510742188, "learning_rate": 3.161502546240114e-07, "loss": 0.2583, "step": 132930 }, { "epoch": 132.96, "grad_norm": 31.366130828857422, "learning_rate": 3.150517920986851e-07, "loss": 0.3061, "step": 132960 }, { "epoch": 132.99, "grad_norm": 53.6392936706543, "learning_rate": 3.1395517911671613e-07, "loss": 0.3393, "step": 132990 }, { "epoch": 133.0, "eval_loss": 0.23990611732006073, "eval_map": 0.839, "eval_map_50": 0.9641, "eval_map_75": 0.9345, "eval_map_chicken": 0.8276, "eval_map_duck": 0.7974, "eval_map_large": 0.847, "eval_map_medium": 0.838, "eval_map_plant": 0.8919, "eval_map_small": 0.3157, "eval_mar_1": 0.3372, "eval_mar_10": 0.87, "eval_mar_100": 0.8732, "eval_mar_100_chicken": 0.8706, "eval_mar_100_duck": 0.8309, "eval_mar_100_plant": 0.918, "eval_mar_large": 0.8797, "eval_mar_medium": 0.8758, "eval_mar_small": 0.3913, "eval_runtime": 12.91, "eval_samples_per_second": 15.492, "eval_steps_per_second": 1.936, "step": 133000 }, { "epoch": 133.02, "grad_norm": 69.37965393066406, "learning_rate": 3.128604161110299e-07, "loss": 0.3037, "step": 133020 }, { "epoch": 133.05, "grad_norm": 39.11929702758789, "learning_rate": 3.1176750351382235e-07, "loss": 0.4716, "step": 133050 }, { "epoch": 133.08, "grad_norm": 38.760353088378906, "learning_rate": 3.106764417565561e-07, "loss": 0.3493, "step": 133080 }, { "epoch": 133.11, "grad_norm": 8.436179632553831e-05, "learning_rate": 3.095872312699666e-07, "loss": 0.3896, "step": 133110 }, { "epoch": 133.14, "grad_norm": 78.86498260498047, "learning_rate": 3.08499872484056e-07, "loss": 0.2766, "step": 133140 }, { "epoch": 133.17, "grad_norm": 35.481605529785156, "learning_rate": 3.07414365828097e-07, "loss": 0.3017, "step": 133170 }, { "epoch": 133.2, "grad_norm": 44.75107192993164, "learning_rate": 3.0633071173062966e-07, "loss": 0.3879, "step": 133200 }, { "epoch": 133.23, "grad_norm": 53.748897552490234, "learning_rate": 3.052489106194645e-07, "loss": 0.2851, "step": 133230 }, { "epoch": 133.26, "grad_norm": 53.68852233886719, "learning_rate": 3.0416896292167873e-07, "loss": 0.2891, "step": 133260 }, { "epoch": 133.29, "grad_norm": 37.728267669677734, "learning_rate": 3.030908690636192e-07, "loss": 0.2803, "step": 133290 }, { "epoch": 133.32, "grad_norm": 44.40903091430664, "learning_rate": 3.0201462947089865e-07, "loss": 0.2301, "step": 133320 }, { "epoch": 133.35, "grad_norm": 56.701873779296875, "learning_rate": 3.0094024456840176e-07, "loss": 0.3826, "step": 133350 }, { "epoch": 133.38, "grad_norm": 36.01105880737305, "learning_rate": 2.99867714780277e-07, "loss": 0.3037, "step": 133380 }, { "epoch": 133.41, "grad_norm": 3.489242590148933e-05, "learning_rate": 2.9879704052994395e-07, "loss": 0.2813, "step": 133410 }, { "epoch": 133.44, "grad_norm": 43.10932540893555, "learning_rate": 2.9772822224008515e-07, "loss": 0.3238, "step": 133440 }, { "epoch": 133.47, "grad_norm": 32.43568801879883, "learning_rate": 2.9666126033265517e-07, "loss": 0.3311, "step": 133470 }, { "epoch": 133.5, "grad_norm": 80.9537353515625, "learning_rate": 2.9559615522887275e-07, "loss": 0.3141, "step": 133500 }, { "epoch": 133.53, "grad_norm": 57.36110305786133, "learning_rate": 2.9453290734922537e-07, "loss": 0.381, "step": 133530 }, { "epoch": 133.56, "grad_norm": 28.574535369873047, "learning_rate": 2.9347151711346556e-07, "loss": 0.3717, "step": 133560 }, { "epoch": 133.59, "grad_norm": 63.255035400390625, "learning_rate": 2.9241198494061427e-07, "loss": 0.3173, "step": 133590 }, { "epoch": 133.62, "grad_norm": 60.05976104736328, "learning_rate": 2.913543112489564e-07, "loss": 0.3414, "step": 133620 }, { "epoch": 133.65, "grad_norm": 57.024742126464844, "learning_rate": 2.9029849645604735e-07, "loss": 0.2356, "step": 133650 }, { "epoch": 133.68, "grad_norm": 142.78050231933594, "learning_rate": 2.892445409787037e-07, "loss": 0.295, "step": 133680 }, { "epoch": 133.71, "grad_norm": 66.85313415527344, "learning_rate": 2.8819244523301206e-07, "loss": 0.2642, "step": 133710 }, { "epoch": 133.74, "grad_norm": 96.01446533203125, "learning_rate": 2.8714220963432125e-07, "loss": 0.3711, "step": 133740 }, { "epoch": 133.77, "grad_norm": 36.59801483154297, "learning_rate": 2.8609383459724915e-07, "loss": 0.2824, "step": 133770 }, { "epoch": 133.8, "grad_norm": 77.94417572021484, "learning_rate": 2.850473205356774e-07, "loss": 0.282, "step": 133800 }, { "epoch": 133.83, "grad_norm": 71.38693237304688, "learning_rate": 2.8400266786275387e-07, "loss": 0.3117, "step": 133830 }, { "epoch": 133.86, "grad_norm": 38.100215911865234, "learning_rate": 2.829598769908892e-07, "loss": 0.3251, "step": 133860 }, { "epoch": 133.89, "grad_norm": 34.341590881347656, "learning_rate": 2.819189483317625e-07, "loss": 0.3062, "step": 133890 }, { "epoch": 133.92, "grad_norm": 36.638065338134766, "learning_rate": 2.8087988229631325e-07, "loss": 0.3245, "step": 133920 }, { "epoch": 133.95, "grad_norm": 47.399959564208984, "learning_rate": 2.798426792947517e-07, "loss": 0.4888, "step": 133950 }, { "epoch": 133.98, "grad_norm": 48.093441009521484, "learning_rate": 2.788073397365465e-07, "loss": 0.3064, "step": 133980 }, { "epoch": 134.0, "eval_loss": 0.2376776933670044, "eval_map": 0.8417, "eval_map_50": 0.964, "eval_map_75": 0.9382, "eval_map_chicken": 0.8329, "eval_map_duck": 0.7981, "eval_map_large": 0.8561, "eval_map_medium": 0.8393, "eval_map_plant": 0.8943, "eval_map_small": 0.3154, "eval_mar_1": 0.3387, "eval_mar_10": 0.8727, "eval_mar_100": 0.8754, "eval_mar_100_chicken": 0.8769, "eval_mar_100_duck": 0.8304, "eval_mar_100_plant": 0.9191, "eval_mar_large": 0.8869, "eval_mar_medium": 0.8772, "eval_mar_small": 0.3835, "eval_runtime": 12.7904, "eval_samples_per_second": 15.637, "eval_steps_per_second": 1.955, "step": 134000 }, { "epoch": 134.01, "grad_norm": 112.90911102294922, "learning_rate": 2.77773864030435e-07, "loss": 0.3274, "step": 134010 }, { "epoch": 134.04, "grad_norm": 36.7672119140625, "learning_rate": 2.76742252584416e-07, "loss": 0.308, "step": 134040 }, { "epoch": 134.07, "grad_norm": 79.12322235107422, "learning_rate": 2.757125058057536e-07, "loss": 0.2506, "step": 134070 }, { "epoch": 134.1, "grad_norm": 7.293569069588557e-05, "learning_rate": 2.746846241009765e-07, "loss": 0.2553, "step": 134100 }, { "epoch": 134.13, "grad_norm": 42.98638153076172, "learning_rate": 2.7365860787587405e-07, "loss": 0.259, "step": 134130 }, { "epoch": 134.16, "grad_norm": 85.73538208007812, "learning_rate": 2.7263445753550275e-07, "loss": 0.2363, "step": 134160 }, { "epoch": 134.19, "grad_norm": 70.6085433959961, "learning_rate": 2.716121734841814e-07, "loss": 0.3002, "step": 134190 }, { "epoch": 134.22, "grad_norm": 34.53990936279297, "learning_rate": 2.7059175612548947e-07, "loss": 0.2666, "step": 134220 }, { "epoch": 134.25, "grad_norm": 68.78221130371094, "learning_rate": 2.6957320586227354e-07, "loss": 0.2736, "step": 134250 }, { "epoch": 134.28, "grad_norm": 126.67739868164062, "learning_rate": 2.685565230966408e-07, "loss": 0.3677, "step": 134280 }, { "epoch": 134.31, "grad_norm": 60.462303161621094, "learning_rate": 2.6754170822996026e-07, "loss": 0.2831, "step": 134310 }, { "epoch": 134.34, "grad_norm": 80.9090347290039, "learning_rate": 2.665287616628659e-07, "loss": 0.3336, "step": 134340 }, { "epoch": 134.37, "grad_norm": 38.44847869873047, "learning_rate": 2.655176837952528e-07, "loss": 0.2694, "step": 134370 }, { "epoch": 134.4, "grad_norm": 34.712867736816406, "learning_rate": 2.6450847502627883e-07, "loss": 0.2834, "step": 134400 }, { "epoch": 134.43, "grad_norm": 72.30377197265625, "learning_rate": 2.6350113575436266e-07, "loss": 0.3044, "step": 134430 }, { "epoch": 134.46, "grad_norm": 43.366329193115234, "learning_rate": 2.6249566637718714e-07, "loss": 0.3433, "step": 134460 }, { "epoch": 134.49, "grad_norm": 214.68313598632812, "learning_rate": 2.614920672916943e-07, "loss": 0.3633, "step": 134490 }, { "epoch": 134.52, "grad_norm": 102.60550689697266, "learning_rate": 2.604903388940899e-07, "loss": 0.2496, "step": 134520 }, { "epoch": 134.55, "grad_norm": 117.38189697265625, "learning_rate": 2.594904815798399e-07, "loss": 0.2893, "step": 134550 }, { "epoch": 134.58, "grad_norm": 52.96833801269531, "learning_rate": 2.584924957436735e-07, "loss": 0.2762, "step": 134580 }, { "epoch": 134.61, "grad_norm": 32.06796646118164, "learning_rate": 2.5749638177957834e-07, "loss": 0.2991, "step": 134610 }, { "epoch": 134.64, "grad_norm": 25.51287841796875, "learning_rate": 2.5650214008080544e-07, "loss": 0.3057, "step": 134640 }, { "epoch": 134.67, "grad_norm": 69.3886489868164, "learning_rate": 2.555097710398635e-07, "loss": 0.2714, "step": 134670 }, { "epoch": 134.7, "grad_norm": 31.11724090576172, "learning_rate": 2.5451927504852757e-07, "loss": 0.298, "step": 134700 }, { "epoch": 134.73, "grad_norm": 50.63080596923828, "learning_rate": 2.5353065249782647e-07, "loss": 0.2976, "step": 134730 }, { "epoch": 134.76, "grad_norm": 40.258792877197266, "learning_rate": 2.525439037780558e-07, "loss": 0.3479, "step": 134760 }, { "epoch": 134.79, "grad_norm": 79.93506622314453, "learning_rate": 2.515590292787656e-07, "loss": 0.295, "step": 134790 }, { "epoch": 134.82, "grad_norm": 25.239242553710938, "learning_rate": 2.505760293887699e-07, "loss": 0.3349, "step": 134820 }, { "epoch": 134.85, "grad_norm": 144.5735321044922, "learning_rate": 2.49594904496141e-07, "loss": 0.3429, "step": 134850 }, { "epoch": 134.88, "grad_norm": 88.77118682861328, "learning_rate": 2.486156549882135e-07, "loss": 0.3669, "step": 134880 }, { "epoch": 134.91, "grad_norm": 63.17409133911133, "learning_rate": 2.4763828125157654e-07, "loss": 0.2863, "step": 134910 }, { "epoch": 134.94, "grad_norm": 76.33804321289062, "learning_rate": 2.4666278367208417e-07, "loss": 0.3808, "step": 134940 }, { "epoch": 134.97, "grad_norm": 28.844863891601562, "learning_rate": 2.456891626348451e-07, "loss": 0.3181, "step": 134970 }, { "epoch": 135.0, "grad_norm": 73.51531982421875, "learning_rate": 2.447174185242324e-07, "loss": 0.2612, "step": 135000 }, { "epoch": 135.0, "eval_loss": 0.2374763935804367, "eval_map": 0.8423, "eval_map_50": 0.9644, "eval_map_75": 0.9339, "eval_map_chicken": 0.8308, "eval_map_duck": 0.7974, "eval_map_large": 0.849, "eval_map_medium": 0.842, "eval_map_plant": 0.8986, "eval_map_small": 0.3126, "eval_mar_1": 0.3391, "eval_mar_10": 0.8727, "eval_mar_100": 0.8761, "eval_mar_100_chicken": 0.8748, "eval_mar_100_duck": 0.8314, "eval_mar_100_plant": 0.922, "eval_mar_large": 0.8812, "eval_mar_medium": 0.8801, "eval_mar_small": 0.3803, "eval_runtime": 13.4176, "eval_samples_per_second": 14.906, "eval_steps_per_second": 1.863, "step": 135000 }, { "epoch": 135.03, "grad_norm": 123.0885238647461, "learning_rate": 2.43747551723873e-07, "loss": 0.2689, "step": 135030 }, { "epoch": 135.06, "grad_norm": 106.5955810546875, "learning_rate": 2.4277956261665624e-07, "loss": 0.3568, "step": 135060 }, { "epoch": 135.09, "grad_norm": 45.890052795410156, "learning_rate": 2.418134515847276e-07, "loss": 0.3034, "step": 135090 }, { "epoch": 135.12, "grad_norm": 59.46106719970703, "learning_rate": 2.40849219009493e-07, "loss": 0.2978, "step": 135120 }, { "epoch": 135.15, "grad_norm": 160.30067443847656, "learning_rate": 2.3988686527161686e-07, "loss": 0.4037, "step": 135150 }, { "epoch": 135.18, "grad_norm": 47.3480110168457, "learning_rate": 2.389263907510209e-07, "loss": 0.2951, "step": 135180 }, { "epoch": 135.21, "grad_norm": 65.01859283447266, "learning_rate": 2.3796779582688444e-07, "loss": 0.2953, "step": 135210 }, { "epoch": 135.24, "grad_norm": 162.04942321777344, "learning_rate": 2.3701108087764657e-07, "loss": 0.3527, "step": 135240 }, { "epoch": 135.27, "grad_norm": 36.01948547363281, "learning_rate": 2.3605624628100178e-07, "loss": 0.2777, "step": 135270 }, { "epoch": 135.3, "grad_norm": 34.9932861328125, "learning_rate": 2.351032924139063e-07, "loss": 0.2984, "step": 135300 }, { "epoch": 135.33, "grad_norm": 546.1378173828125, "learning_rate": 2.3415221965256807e-07, "loss": 0.2734, "step": 135330 }, { "epoch": 135.36, "grad_norm": 48.31810760498047, "learning_rate": 2.3320302837245846e-07, "loss": 0.3204, "step": 135360 }, { "epoch": 135.39, "grad_norm": 57.29185485839844, "learning_rate": 2.3225571894830047e-07, "loss": 0.352, "step": 135390 }, { "epoch": 135.42, "grad_norm": 95.8145980834961, "learning_rate": 2.3131029175407883e-07, "loss": 0.2391, "step": 135420 }, { "epoch": 135.45, "grad_norm": 56.86247634887695, "learning_rate": 2.3036674716303277e-07, "loss": 0.3626, "step": 135450 }, { "epoch": 135.48, "grad_norm": 74.483154296875, "learning_rate": 2.2942508554765764e-07, "loss": 0.3074, "step": 135480 }, { "epoch": 135.51, "grad_norm": 73.7916259765625, "learning_rate": 2.2848530727970775e-07, "loss": 0.3211, "step": 135510 }, { "epoch": 135.54, "grad_norm": 27.43739128112793, "learning_rate": 2.27547412730193e-07, "loss": 0.2786, "step": 135540 }, { "epoch": 135.57, "grad_norm": 28.07783317565918, "learning_rate": 2.2661140226937773e-07, "loss": 0.2877, "step": 135570 }, { "epoch": 135.6, "grad_norm": 67.46879577636719, "learning_rate": 2.2567727626678527e-07, "loss": 0.3054, "step": 135600 }, { "epoch": 135.63, "grad_norm": 34.095306396484375, "learning_rate": 2.2474503509119394e-07, "loss": 0.2887, "step": 135630 }, { "epoch": 135.66, "grad_norm": 54.77301025390625, "learning_rate": 2.2381467911063658e-07, "loss": 0.3264, "step": 135660 }, { "epoch": 135.69, "grad_norm": 57.4935302734375, "learning_rate": 2.2288620869240384e-07, "loss": 0.305, "step": 135690 }, { "epoch": 135.72, "grad_norm": 91.25618743896484, "learning_rate": 2.2195962420304083e-07, "loss": 0.2808, "step": 135720 }, { "epoch": 135.75, "grad_norm": 48.622493743896484, "learning_rate": 2.210349260083494e-07, "loss": 0.3321, "step": 135750 }, { "epoch": 135.78, "grad_norm": 51.08588409423828, "learning_rate": 2.2011211447338477e-07, "loss": 0.3489, "step": 135780 }, { "epoch": 135.81, "grad_norm": 54.52326202392578, "learning_rate": 2.191911899624588e-07, "loss": 0.3167, "step": 135810 }, { "epoch": 135.84, "grad_norm": 33.04013442993164, "learning_rate": 2.1827215283913683e-07, "loss": 0.2851, "step": 135840 }, { "epoch": 135.87, "grad_norm": 67.35446166992188, "learning_rate": 2.1735500346624083e-07, "loss": 0.304, "step": 135870 }, { "epoch": 135.9, "grad_norm": 49.69088363647461, "learning_rate": 2.1643974220584729e-07, "loss": 0.3329, "step": 135900 }, { "epoch": 135.93, "grad_norm": 67.20279693603516, "learning_rate": 2.1552636941928717e-07, "loss": 0.3378, "step": 135930 }, { "epoch": 135.96, "grad_norm": 67.39599609375, "learning_rate": 2.1461488546714425e-07, "loss": 0.2682, "step": 135960 }, { "epoch": 135.99, "grad_norm": 119.83210754394531, "learning_rate": 2.137052907092596e-07, "loss": 0.2906, "step": 135990 }, { "epoch": 136.0, "eval_loss": 0.23846592009067535, "eval_map": 0.8392, "eval_map_50": 0.9641, "eval_map_75": 0.9341, "eval_map_chicken": 0.8299, "eval_map_duck": 0.7909, "eval_map_large": 0.8443, "eval_map_medium": 0.8415, "eval_map_plant": 0.8969, "eval_map_small": 0.3276, "eval_mar_1": 0.3364, "eval_mar_10": 0.8705, "eval_mar_100": 0.874, "eval_mar_100_chicken": 0.8742, "eval_mar_100_duck": 0.8258, "eval_mar_100_plant": 0.9218, "eval_mar_large": 0.8775, "eval_mar_medium": 0.8796, "eval_mar_small": 0.3938, "eval_runtime": 12.6135, "eval_samples_per_second": 15.856, "eval_steps_per_second": 1.982, "step": 136000 }, { "epoch": 136.02, "grad_norm": 31.28265953063965, "learning_rate": 2.127975855047243e-07, "loss": 0.3239, "step": 136020 }, { "epoch": 136.05, "grad_norm": 63.03318786621094, "learning_rate": 2.1189177021188888e-07, "loss": 0.3189, "step": 136050 }, { "epoch": 136.08, "grad_norm": 6.829766789451241e-05, "learning_rate": 2.1098784518835292e-07, "loss": 0.3005, "step": 136080 }, { "epoch": 136.11, "grad_norm": 78.09432220458984, "learning_rate": 2.1008581079097312e-07, "loss": 0.3421, "step": 136110 }, { "epoch": 136.14, "grad_norm": 34.74638366699219, "learning_rate": 2.0918566737585688e-07, "loss": 0.3838, "step": 136140 }, { "epoch": 136.17, "grad_norm": 72.68283081054688, "learning_rate": 2.0828741529836771e-07, "loss": 0.3085, "step": 136170 }, { "epoch": 136.2, "grad_norm": 87.65298461914062, "learning_rate": 2.0739105491312028e-07, "loss": 0.2839, "step": 136200 }, { "epoch": 136.23, "grad_norm": 188.88485717773438, "learning_rate": 2.0649658657398487e-07, "loss": 0.3181, "step": 136230 }, { "epoch": 136.26, "grad_norm": 49.13550567626953, "learning_rate": 2.056040106340823e-07, "loss": 0.2549, "step": 136260 }, { "epoch": 136.29, "grad_norm": 26.52992057800293, "learning_rate": 2.0471332744578853e-07, "loss": 0.2842, "step": 136290 }, { "epoch": 136.32, "grad_norm": 46.20585250854492, "learning_rate": 2.0382453736072838e-07, "loss": 0.2924, "step": 136320 }, { "epoch": 136.35, "grad_norm": 66.662109375, "learning_rate": 2.0293764072978618e-07, "loss": 0.2849, "step": 136350 }, { "epoch": 136.38, "grad_norm": 94.07794189453125, "learning_rate": 2.0205263790309125e-07, "loss": 0.2408, "step": 136380 }, { "epoch": 136.41, "grad_norm": 75.1082992553711, "learning_rate": 2.0116952923003142e-07, "loss": 0.3371, "step": 136410 }, { "epoch": 136.44, "grad_norm": 51.17622756958008, "learning_rate": 2.0028831505924162e-07, "loss": 0.2866, "step": 136440 }, { "epoch": 136.47, "grad_norm": 34.178009033203125, "learning_rate": 1.9940899573861195e-07, "loss": 0.25, "step": 136470 }, { "epoch": 136.5, "grad_norm": 56.812564849853516, "learning_rate": 1.9853157161528468e-07, "loss": 0.2711, "step": 136500 }, { "epoch": 136.53, "grad_norm": 83.26190948486328, "learning_rate": 1.9765604303565223e-07, "loss": 0.2463, "step": 136530 }, { "epoch": 136.56, "grad_norm": 25.360902786254883, "learning_rate": 1.967824103453597e-07, "loss": 0.2424, "step": 136560 }, { "epoch": 136.59, "grad_norm": 62.677852630615234, "learning_rate": 1.959106738893035e-07, "loss": 0.2764, "step": 136590 }, { "epoch": 136.62, "grad_norm": 57.45070266723633, "learning_rate": 1.9504083401162999e-07, "loss": 0.2936, "step": 136620 }, { "epoch": 136.65, "grad_norm": 50.50303268432617, "learning_rate": 1.9417289105574054e-07, "loss": 0.3131, "step": 136650 }, { "epoch": 136.68, "grad_norm": 35.340476989746094, "learning_rate": 1.9330684536428335e-07, "loss": 0.3018, "step": 136680 }, { "epoch": 136.71, "grad_norm": 37.114158630371094, "learning_rate": 1.9244269727916097e-07, "loss": 0.3123, "step": 136710 }, { "epoch": 136.74, "grad_norm": 0.00014646242198068649, "learning_rate": 1.9158044714152447e-07, "loss": 0.3179, "step": 136740 }, { "epoch": 136.77, "grad_norm": 50.72560501098633, "learning_rate": 1.907200952917765e-07, "loss": 0.3113, "step": 136770 }, { "epoch": 136.8, "grad_norm": 95.05646514892578, "learning_rate": 1.8986164206957037e-07, "loss": 0.2783, "step": 136800 }, { "epoch": 136.83, "grad_norm": 127.39823913574219, "learning_rate": 1.8900508781381056e-07, "loss": 0.2434, "step": 136830 }, { "epoch": 136.86, "grad_norm": 63.642906188964844, "learning_rate": 1.8815043286265044e-07, "loss": 0.3685, "step": 136860 }, { "epoch": 136.89, "grad_norm": 41.83503341674805, "learning_rate": 1.8729767755349514e-07, "loss": 0.2715, "step": 136890 }, { "epoch": 136.92, "grad_norm": 32.18874740600586, "learning_rate": 1.8644682222299703e-07, "loss": 0.3355, "step": 136920 }, { "epoch": 136.95, "grad_norm": 47.56570053100586, "learning_rate": 1.8559786720706185e-07, "loss": 0.2538, "step": 136950 }, { "epoch": 136.98, "grad_norm": 34.584434509277344, "learning_rate": 1.8475081284084428e-07, "loss": 0.2954, "step": 136980 }, { "epoch": 137.0, "eval_loss": 0.2362699657678604, "eval_map": 0.8422, "eval_map_50": 0.9641, "eval_map_75": 0.9382, "eval_map_chicken": 0.8326, "eval_map_duck": 0.7944, "eval_map_large": 0.8493, "eval_map_medium": 0.8414, "eval_map_plant": 0.8996, "eval_map_small": 0.3286, "eval_mar_1": 0.3381, "eval_mar_10": 0.8725, "eval_mar_100": 0.8764, "eval_mar_100_chicken": 0.8757, "eval_mar_100_duck": 0.8304, "eval_mar_100_plant": 0.9231, "eval_mar_large": 0.8806, "eval_mar_medium": 0.8793, "eval_mar_small": 0.4059, "eval_runtime": 12.7737, "eval_samples_per_second": 15.657, "eval_steps_per_second": 1.957, "step": 137000 }, { "epoch": 137.01, "grad_norm": 68.13064575195312, "learning_rate": 1.8390565945874572e-07, "loss": 0.3266, "step": 137010 }, { "epoch": 137.04, "grad_norm": 55.27497100830078, "learning_rate": 1.8306240739442094e-07, "loss": 0.2567, "step": 137040 }, { "epoch": 137.07, "grad_norm": 64.2692642211914, "learning_rate": 1.8222105698077253e-07, "loss": 0.4386, "step": 137070 }, { "epoch": 137.1, "grad_norm": 80.51256561279297, "learning_rate": 1.8138160854995145e-07, "loss": 0.258, "step": 137100 }, { "epoch": 137.13, "grad_norm": 28.94463539123535, "learning_rate": 1.805440624333593e-07, "loss": 0.2605, "step": 137130 }, { "epoch": 137.16, "grad_norm": 50.57966613769531, "learning_rate": 1.7970841896164658e-07, "loss": 0.2873, "step": 137160 }, { "epoch": 137.19, "grad_norm": 168.8895721435547, "learning_rate": 1.788746784647105e-07, "loss": 0.2886, "step": 137190 }, { "epoch": 137.22, "grad_norm": 48.154117584228516, "learning_rate": 1.7804284127169946e-07, "loss": 0.2826, "step": 137220 }, { "epoch": 137.25, "grad_norm": 65.65048217773438, "learning_rate": 1.7721290771100964e-07, "loss": 0.3499, "step": 137250 }, { "epoch": 137.28, "grad_norm": 39.36507034301758, "learning_rate": 1.7638487811028616e-07, "loss": 0.3592, "step": 137280 }, { "epoch": 137.31, "grad_norm": 3.9918140828376636e-05, "learning_rate": 1.7555875279642087e-07, "loss": 0.2876, "step": 137310 }, { "epoch": 137.34, "grad_norm": 54.13775634765625, "learning_rate": 1.7473453209555625e-07, "loss": 0.2783, "step": 137340 }, { "epoch": 137.37, "grad_norm": 80.78361511230469, "learning_rate": 1.7391221633308032e-07, "loss": 0.3372, "step": 137370 }, { "epoch": 137.4, "grad_norm": 56.548744201660156, "learning_rate": 1.7309180583363062e-07, "loss": 0.3109, "step": 137400 }, { "epoch": 137.43, "grad_norm": 80.68622589111328, "learning_rate": 1.7227330092109306e-07, "loss": 0.2908, "step": 137430 }, { "epoch": 137.46, "grad_norm": 25.322471618652344, "learning_rate": 1.7145670191859977e-07, "loss": 0.3728, "step": 137460 }, { "epoch": 137.49, "grad_norm": 96.24398040771484, "learning_rate": 1.7064200914853112e-07, "loss": 0.2647, "step": 137490 }, { "epoch": 137.52, "grad_norm": 83.2372817993164, "learning_rate": 1.6982922293251548e-07, "loss": 0.2709, "step": 137520 }, { "epoch": 137.55, "grad_norm": 39.50093078613281, "learning_rate": 1.690183435914261e-07, "loss": 0.3095, "step": 137550 }, { "epoch": 137.58, "grad_norm": 70.94627380371094, "learning_rate": 1.6820937144538807e-07, "loss": 0.2679, "step": 137580 }, { "epoch": 137.61, "grad_norm": 65.37276458740234, "learning_rate": 1.6740230681376867e-07, "loss": 0.3017, "step": 137610 }, { "epoch": 137.64, "grad_norm": 104.22909545898438, "learning_rate": 1.6659715001518583e-07, "loss": 0.3464, "step": 137640 }, { "epoch": 137.67, "grad_norm": 52.15321350097656, "learning_rate": 1.6579390136750086e-07, "loss": 0.263, "step": 137670 }, { "epoch": 137.7, "grad_norm": 5.983114169794135e-05, "learning_rate": 1.6499256118782503e-07, "loss": 0.2545, "step": 137700 }, { "epoch": 137.73, "grad_norm": 47.72504425048828, "learning_rate": 1.6419312979251368e-07, "loss": 0.3004, "step": 137730 }, { "epoch": 137.76, "grad_norm": 116.74971771240234, "learning_rate": 1.6339560749717154e-07, "loss": 0.2953, "step": 137760 }, { "epoch": 137.79, "grad_norm": 108.49244689941406, "learning_rate": 1.6259999461664567e-07, "loss": 0.3553, "step": 137790 }, { "epoch": 137.82, "grad_norm": 81.96439361572266, "learning_rate": 1.6180629146503256e-07, "loss": 0.2575, "step": 137820 }, { "epoch": 137.85, "grad_norm": 33.68819808959961, "learning_rate": 1.6101449835567273e-07, "loss": 0.2336, "step": 137850 }, { "epoch": 137.88, "grad_norm": 125.3701171875, "learning_rate": 1.6022461560115498e-07, "loss": 0.2917, "step": 137880 }, { "epoch": 137.91, "grad_norm": 69.24799346923828, "learning_rate": 1.59436643513311e-07, "loss": 0.2775, "step": 137910 }, { "epoch": 137.94, "grad_norm": 81.26785278320312, "learning_rate": 1.586505824032214e-07, "loss": 0.332, "step": 137940 }, { "epoch": 137.97, "grad_norm": 58.87031173706055, "learning_rate": 1.5786643258120905e-07, "loss": 0.3515, "step": 137970 }, { "epoch": 138.0, "grad_norm": 45.44074249267578, "learning_rate": 1.5708419435684463e-07, "loss": 0.304, "step": 138000 }, { "epoch": 138.0, "eval_loss": 0.24126343429088593, "eval_map": 0.8374, "eval_map_50": 0.9641, "eval_map_75": 0.934, "eval_map_chicken": 0.8264, "eval_map_duck": 0.7891, "eval_map_large": 0.8419, "eval_map_medium": 0.8376, "eval_map_plant": 0.8966, "eval_map_small": 0.3224, "eval_mar_1": 0.3362, "eval_mar_10": 0.8684, "eval_mar_100": 0.872, "eval_mar_100_chicken": 0.8712, "eval_mar_100_duck": 0.8253, "eval_mar_100_plant": 0.9195, "eval_mar_large": 0.8758, "eval_mar_medium": 0.8749, "eval_mar_small": 0.3962, "eval_runtime": 13.2447, "eval_samples_per_second": 15.1, "eval_steps_per_second": 1.888, "step": 138000 }, { "epoch": 138.03, "grad_norm": 91.21382904052734, "learning_rate": 1.563038680389428e-07, "loss": 0.3465, "step": 138030 }, { "epoch": 138.06, "grad_norm": 59.0599479675293, "learning_rate": 1.555254539355655e-07, "loss": 0.2661, "step": 138060 }, { "epoch": 138.09, "grad_norm": 48.07148361206055, "learning_rate": 1.5474895235401688e-07, "loss": 0.3511, "step": 138090 }, { "epoch": 138.12, "grad_norm": 50.89740753173828, "learning_rate": 1.5397436360084784e-07, "loss": 0.3122, "step": 138120 }, { "epoch": 138.15, "grad_norm": 78.71317291259766, "learning_rate": 1.532016879818532e-07, "loss": 0.2792, "step": 138150 }, { "epoch": 138.18, "grad_norm": 57.20781707763672, "learning_rate": 1.5243092580207507e-07, "loss": 0.2789, "step": 138180 }, { "epoch": 138.21, "grad_norm": 121.16020202636719, "learning_rate": 1.5166207736579564e-07, "loss": 0.2839, "step": 138210 }, { "epoch": 138.24, "grad_norm": 40.64909744262695, "learning_rate": 1.5089514297654594e-07, "loss": 0.3112, "step": 138240 }, { "epoch": 138.27, "grad_norm": 65.25508117675781, "learning_rate": 1.5013012293709828e-07, "loss": 0.3288, "step": 138270 }, { "epoch": 138.3, "grad_norm": 39.81812286376953, "learning_rate": 1.4936701754947104e-07, "loss": 0.3029, "step": 138300 }, { "epoch": 138.33, "grad_norm": 33.33059310913086, "learning_rate": 1.4860582711492544e-07, "loss": 0.2521, "step": 138330 }, { "epoch": 138.36, "grad_norm": 74.51206970214844, "learning_rate": 1.4784655193396947e-07, "loss": 0.2791, "step": 138360 }, { "epoch": 138.39, "grad_norm": 51.74424362182617, "learning_rate": 1.4708919230635054e-07, "loss": 0.4274, "step": 138390 }, { "epoch": 138.42, "grad_norm": 87.70036315917969, "learning_rate": 1.463337485310634e-07, "loss": 0.2608, "step": 138420 }, { "epoch": 138.45, "grad_norm": 52.942955017089844, "learning_rate": 1.4558022090634504e-07, "loss": 0.2849, "step": 138450 }, { "epoch": 138.48, "grad_norm": 98.33261108398438, "learning_rate": 1.448286097296764e-07, "loss": 0.3167, "step": 138480 }, { "epoch": 138.51, "grad_norm": 37.00206756591797, "learning_rate": 1.4407891529778172e-07, "loss": 0.3867, "step": 138510 }, { "epoch": 138.54, "grad_norm": 77.3591079711914, "learning_rate": 1.4333113790662822e-07, "loss": 0.306, "step": 138540 }, { "epoch": 138.57, "grad_norm": 157.75489807128906, "learning_rate": 1.425852778514264e-07, "loss": 0.2781, "step": 138570 }, { "epoch": 138.6, "grad_norm": 38.123634338378906, "learning_rate": 1.4184133542663014e-07, "loss": 0.2693, "step": 138600 }, { "epoch": 138.63, "grad_norm": 38.28602600097656, "learning_rate": 1.4109931092593732e-07, "loss": 0.2756, "step": 138630 }, { "epoch": 138.66, "grad_norm": 76.58318328857422, "learning_rate": 1.4035920464228525e-07, "loss": 0.2917, "step": 138660 }, { "epoch": 138.69, "grad_norm": 50.124961853027344, "learning_rate": 1.39621016867858e-07, "loss": 0.2438, "step": 138690 }, { "epoch": 138.72, "grad_norm": 54.18952560424805, "learning_rate": 1.388847478940797e-07, "loss": 0.2826, "step": 138720 }, { "epoch": 138.75, "grad_norm": 29.58094024658203, "learning_rate": 1.3815039801161723e-07, "loss": 0.3124, "step": 138750 }, { "epoch": 138.78, "grad_norm": 25.52207374572754, "learning_rate": 1.3741796751038095e-07, "loss": 0.3331, "step": 138780 }, { "epoch": 138.81, "grad_norm": 47.852413177490234, "learning_rate": 1.366874566795229e-07, "loss": 0.2891, "step": 138810 }, { "epoch": 138.84, "grad_norm": 71.01753234863281, "learning_rate": 1.3595886580743677e-07, "loss": 0.3068, "step": 138840 }, { "epoch": 138.87, "grad_norm": 55.47769546508789, "learning_rate": 1.3523219518175924e-07, "loss": 0.2797, "step": 138870 }, { "epoch": 138.9, "grad_norm": 65.89171600341797, "learning_rate": 1.3450744508936687e-07, "loss": 0.4404, "step": 138900 }, { "epoch": 138.93, "grad_norm": 57.51774215698242, "learning_rate": 1.337846158163819e-07, "loss": 0.2987, "step": 138930 }, { "epoch": 138.96, "grad_norm": 34.96419906616211, "learning_rate": 1.330637076481639e-07, "loss": 0.3036, "step": 138960 }, { "epoch": 138.99, "grad_norm": 66.54273223876953, "learning_rate": 1.3234472086931738e-07, "loss": 0.2716, "step": 138990 }, { "epoch": 139.0, "eval_loss": 0.2420237958431244, "eval_map": 0.8401, "eval_map_50": 0.964, "eval_map_75": 0.9343, "eval_map_chicken": 0.8264, "eval_map_duck": 0.7967, "eval_map_large": 0.8493, "eval_map_medium": 0.8406, "eval_map_plant": 0.8971, "eval_map_small": 0.3285, "eval_mar_1": 0.3376, "eval_mar_10": 0.8709, "eval_mar_100": 0.8745, "eval_mar_100_chicken": 0.8716, "eval_mar_100_duck": 0.8309, "eval_mar_100_plant": 0.9208, "eval_mar_large": 0.88, "eval_mar_medium": 0.8779, "eval_mar_small": 0.3978, "eval_runtime": 12.8363, "eval_samples_per_second": 15.581, "eval_steps_per_second": 1.948, "step": 139000 }, { "epoch": 139.02, "grad_norm": 31.081195831298828, "learning_rate": 1.3162765576368587e-07, "loss": 0.2772, "step": 139020 }, { "epoch": 139.05, "grad_norm": 98.94862365722656, "learning_rate": 1.3091251261435568e-07, "loss": 0.3426, "step": 139050 }, { "epoch": 139.08, "grad_norm": 37.268611907958984, "learning_rate": 1.3019929170365376e-07, "loss": 0.2749, "step": 139080 }, { "epoch": 139.11, "grad_norm": 97.38902282714844, "learning_rate": 1.2948799331314933e-07, "loss": 0.2576, "step": 139110 }, { "epoch": 139.14, "grad_norm": 36.92425537109375, "learning_rate": 1.287786177236511e-07, "loss": 0.2962, "step": 139140 }, { "epoch": 139.17, "grad_norm": 55.09945297241211, "learning_rate": 1.2807116521520947e-07, "loss": 0.3605, "step": 139170 }, { "epoch": 139.2, "grad_norm": 198.84095764160156, "learning_rate": 1.2736563606711384e-07, "loss": 0.2592, "step": 139200 }, { "epoch": 139.23, "grad_norm": 42.63358688354492, "learning_rate": 1.2666203055789915e-07, "loss": 0.3251, "step": 139230 }, { "epoch": 139.26, "grad_norm": 45.17707824707031, "learning_rate": 1.259603489653355e-07, "loss": 0.3057, "step": 139260 }, { "epoch": 139.29, "grad_norm": 61.087432861328125, "learning_rate": 1.252605915664362e-07, "loss": 0.2787, "step": 139290 }, { "epoch": 139.32, "grad_norm": 102.43244934082031, "learning_rate": 1.2456275863745426e-07, "loss": 0.2449, "step": 139320 }, { "epoch": 139.35, "grad_norm": 61.925628662109375, "learning_rate": 1.2386685045388313e-07, "loss": 0.2645, "step": 139350 }, { "epoch": 139.38, "grad_norm": 73.36103820800781, "learning_rate": 1.2317286729045586e-07, "loss": 0.2803, "step": 139380 }, { "epoch": 139.41, "grad_norm": 58.304744720458984, "learning_rate": 1.224808094211477e-07, "loss": 0.3068, "step": 139410 }, { "epoch": 139.44, "grad_norm": 287.02545166015625, "learning_rate": 1.2179067711917015e-07, "loss": 0.2751, "step": 139440 }, { "epoch": 139.47, "grad_norm": 37.6390495300293, "learning_rate": 1.21102470656978e-07, "loss": 0.3145, "step": 139470 }, { "epoch": 139.5, "grad_norm": 45.643707275390625, "learning_rate": 1.2041619030626283e-07, "loss": 0.2779, "step": 139500 }, { "epoch": 139.53, "grad_norm": 35.93082809448242, "learning_rate": 1.1973183633795849e-07, "loss": 0.3279, "step": 139530 }, { "epoch": 139.56, "grad_norm": 4.199631803203374e-05, "learning_rate": 1.1904940902223661e-07, "loss": 0.3325, "step": 139560 }, { "epoch": 139.59, "grad_norm": 53.78736114501953, "learning_rate": 1.1836890862850892e-07, "loss": 0.4185, "step": 139590 }, { "epoch": 139.62, "grad_norm": 58.38619613647461, "learning_rate": 1.1769033542542552e-07, "loss": 0.3033, "step": 139620 }, { "epoch": 139.65, "grad_norm": 58.04499435424805, "learning_rate": 1.1701368968087711e-07, "loss": 0.3376, "step": 139650 }, { "epoch": 139.68, "grad_norm": 98.6880874633789, "learning_rate": 1.1633897166199227e-07, "loss": 0.3358, "step": 139680 }, { "epoch": 139.71, "grad_norm": 77.15230560302734, "learning_rate": 1.1566618163513954e-07, "loss": 0.2561, "step": 139710 }, { "epoch": 139.74, "grad_norm": 74.55155944824219, "learning_rate": 1.1499531986592482e-07, "loss": 0.2779, "step": 139740 }, { "epoch": 139.77, "grad_norm": 161.90643310546875, "learning_rate": 1.1432638661919515e-07, "loss": 0.2906, "step": 139770 }, { "epoch": 139.8, "grad_norm": 41.2548942565918, "learning_rate": 1.136593821590326e-07, "loss": 0.364, "step": 139800 }, { "epoch": 139.83, "grad_norm": 31.869827270507812, "learning_rate": 1.129943067487621e-07, "loss": 0.246, "step": 139830 }, { "epoch": 139.86, "grad_norm": 77.08648681640625, "learning_rate": 1.1233116065094363e-07, "loss": 0.3231, "step": 139860 }, { "epoch": 139.89, "grad_norm": 36.88239288330078, "learning_rate": 1.1166994412737774e-07, "loss": 0.375, "step": 139890 }, { "epoch": 139.92, "grad_norm": 51.21638870239258, "learning_rate": 1.1101065743910122e-07, "loss": 0.3598, "step": 139920 }, { "epoch": 139.95, "grad_norm": 55.65005111694336, "learning_rate": 1.1035330084639084e-07, "loss": 0.268, "step": 139950 }, { "epoch": 139.98, "grad_norm": 58.46250534057617, "learning_rate": 1.0969787460876013e-07, "loss": 0.3027, "step": 139980 }, { "epoch": 140.0, "eval_loss": 0.24009007215499878, "eval_map": 0.8416, "eval_map_50": 0.9644, "eval_map_75": 0.9344, "eval_map_chicken": 0.8312, "eval_map_duck": 0.795, "eval_map_large": 0.8482, "eval_map_medium": 0.8409, "eval_map_plant": 0.8985, "eval_map_small": 0.3279, "eval_mar_1": 0.3375, "eval_mar_10": 0.8715, "eval_mar_100": 0.8755, "eval_mar_100_chicken": 0.874, "eval_mar_100_duck": 0.8304, "eval_mar_100_plant": 0.922, "eval_mar_large": 0.8787, "eval_mar_medium": 0.8797, "eval_mar_small": 0.3965, "eval_runtime": 12.6676, "eval_samples_per_second": 15.788, "eval_steps_per_second": 1.974, "step": 140000 }, { "epoch": 140.01, "grad_norm": 67.37495422363281, "learning_rate": 1.09044378984961e-07, "loss": 0.3308, "step": 140010 }, { "epoch": 140.04, "grad_norm": 39.61513900756836, "learning_rate": 1.0839281423298375e-07, "loss": 0.3441, "step": 140040 }, { "epoch": 140.07, "grad_norm": 65.87903594970703, "learning_rate": 1.0774318061005484e-07, "loss": 0.2941, "step": 140070 }, { "epoch": 140.1, "grad_norm": 23.792728424072266, "learning_rate": 1.0709547837263967e-07, "loss": 0.2638, "step": 140100 }, { "epoch": 140.13, "grad_norm": 43.72016525268555, "learning_rate": 1.0644970777644093e-07, "loss": 0.2507, "step": 140130 }, { "epoch": 140.16, "grad_norm": 34.209049224853516, "learning_rate": 1.0580586907639912e-07, "loss": 0.2717, "step": 140160 }, { "epoch": 140.19, "grad_norm": 46.558387756347656, "learning_rate": 1.0516396252669092e-07, "loss": 0.3319, "step": 140190 }, { "epoch": 140.22, "grad_norm": 59.507568359375, "learning_rate": 1.0452398838073141e-07, "loss": 0.2908, "step": 140220 }, { "epoch": 140.25, "grad_norm": 30.305130004882812, "learning_rate": 1.038859468911707e-07, "loss": 0.2668, "step": 140250 }, { "epoch": 140.28, "grad_norm": 50.259681701660156, "learning_rate": 1.032498383099001e-07, "loss": 0.3573, "step": 140280 }, { "epoch": 140.31, "grad_norm": 38.37261962890625, "learning_rate": 1.0261566288804315e-07, "loss": 0.2752, "step": 140310 }, { "epoch": 140.34, "grad_norm": 86.87332916259766, "learning_rate": 1.0198342087596292e-07, "loss": 0.2542, "step": 140340 }, { "epoch": 140.37, "grad_norm": 70.43539428710938, "learning_rate": 1.0135311252325863e-07, "loss": 0.3502, "step": 140370 }, { "epoch": 140.4, "grad_norm": 59.70405960083008, "learning_rate": 1.007247380787657e-07, "loss": 0.2706, "step": 140400 }, { "epoch": 140.43, "grad_norm": 45.32529067993164, "learning_rate": 1.0009829779055679e-07, "loss": 0.3461, "step": 140430 }, { "epoch": 140.46, "grad_norm": 137.25991821289062, "learning_rate": 9.947379190594076e-08, "loss": 0.2685, "step": 140460 }, { "epoch": 140.49, "grad_norm": 30.88907814025879, "learning_rate": 9.885122067146147e-08, "loss": 0.2905, "step": 140490 }, { "epoch": 140.52, "grad_norm": 44.08793258666992, "learning_rate": 9.823058433290178e-08, "loss": 0.354, "step": 140520 }, { "epoch": 140.55, "grad_norm": 54.0882682800293, "learning_rate": 9.761188313527792e-08, "loss": 0.3987, "step": 140550 }, { "epoch": 140.58, "grad_norm": 56.39140701293945, "learning_rate": 9.699511732284395e-08, "loss": 0.3414, "step": 140580 }, { "epoch": 140.61, "grad_norm": 92.50699615478516, "learning_rate": 9.638028713908898e-08, "loss": 0.2498, "step": 140610 }, { "epoch": 140.64, "grad_norm": 30.601844787597656, "learning_rate": 9.576739282673886e-08, "loss": 0.3422, "step": 140640 }, { "epoch": 140.67, "grad_norm": 258.62567138671875, "learning_rate": 9.515643462775337e-08, "loss": 0.2562, "step": 140670 }, { "epoch": 140.7, "grad_norm": 34.915225982666016, "learning_rate": 9.454741278333013e-08, "loss": 0.3039, "step": 140700 }, { "epoch": 140.73, "grad_norm": 49.48041915893555, "learning_rate": 9.394032753390014e-08, "loss": 0.2699, "step": 140730 }, { "epoch": 140.76, "grad_norm": 46.959190368652344, "learning_rate": 9.333517911913281e-08, "loss": 0.3754, "step": 140760 }, { "epoch": 140.79, "grad_norm": 53.09184265136719, "learning_rate": 9.273196777792926e-08, "loss": 0.3154, "step": 140790 }, { "epoch": 140.82, "grad_norm": 74.48652648925781, "learning_rate": 9.213069374842953e-08, "loss": 0.3026, "step": 140820 }, { "epoch": 140.85, "grad_norm": 56.52968978881836, "learning_rate": 9.153135726800599e-08, "loss": 0.2439, "step": 140850 }, { "epoch": 140.88, "grad_norm": 40.503753662109375, "learning_rate": 9.093395857326714e-08, "loss": 0.2784, "step": 140880 }, { "epoch": 140.91, "grad_norm": 40.962669372558594, "learning_rate": 9.03384979000571e-08, "loss": 0.2684, "step": 140910 }, { "epoch": 140.94, "grad_norm": 5.757583130616695e-05, "learning_rate": 8.974497548345396e-08, "loss": 0.2698, "step": 140940 }, { "epoch": 140.97, "grad_norm": 52.89075469970703, "learning_rate": 8.915339155777136e-08, "loss": 0.2618, "step": 140970 }, { "epoch": 141.0, "grad_norm": 85.5951919555664, "learning_rate": 8.856374635655696e-08, "loss": 0.2667, "step": 141000 }, { "epoch": 141.0, "eval_loss": 0.23998680710792542, "eval_map": 0.8399, "eval_map_50": 0.9641, "eval_map_75": 0.9341, "eval_map_chicken": 0.8283, "eval_map_duck": 0.7945, "eval_map_large": 0.8463, "eval_map_medium": 0.8405, "eval_map_plant": 0.8968, "eval_map_small": 0.3252, "eval_mar_1": 0.3374, "eval_mar_10": 0.8704, "eval_mar_100": 0.8743, "eval_mar_100_chicken": 0.8728, "eval_mar_100_duck": 0.8294, "eval_mar_100_plant": 0.9208, "eval_mar_large": 0.8801, "eval_mar_medium": 0.8781, "eval_mar_small": 0.3943, "eval_runtime": 12.5066, "eval_samples_per_second": 15.992, "eval_steps_per_second": 1.999, "step": 141000 }, { "epoch": 141.03, "grad_norm": 48.95118713378906, "learning_rate": 8.797604011259287e-08, "loss": 0.2594, "step": 141030 }, { "epoch": 141.06, "grad_norm": 114.50353240966797, "learning_rate": 8.739027305789682e-08, "loss": 0.2967, "step": 141060 }, { "epoch": 141.09, "grad_norm": 47.038002014160156, "learning_rate": 8.680644542372052e-08, "loss": 0.2466, "step": 141090 }, { "epoch": 141.12, "grad_norm": 67.99077606201172, "learning_rate": 8.622455744054958e-08, "loss": 0.3461, "step": 141120 }, { "epoch": 141.15, "grad_norm": 43.070858001708984, "learning_rate": 8.564460933810414e-08, "loss": 0.2927, "step": 141150 }, { "epoch": 141.18, "grad_norm": 95.92662048339844, "learning_rate": 8.506660134533828e-08, "loss": 0.2726, "step": 141180 }, { "epoch": 141.21, "grad_norm": 57.91007995605469, "learning_rate": 8.449053369044058e-08, "loss": 0.2748, "step": 141210 }, { "epoch": 141.24, "grad_norm": 56.54653549194336, "learning_rate": 8.391640660083411e-08, "loss": 0.2932, "step": 141240 }, { "epoch": 141.27, "grad_norm": 46.14498519897461, "learning_rate": 8.334422030317424e-08, "loss": 0.2738, "step": 141270 }, { "epoch": 141.3, "grad_norm": 34.719970703125, "learning_rate": 8.277397502335194e-08, "loss": 0.262, "step": 141300 }, { "epoch": 141.33, "grad_norm": 68.12794494628906, "learning_rate": 8.22056709864899e-08, "loss": 0.3084, "step": 141330 }, { "epoch": 141.36, "grad_norm": 71.2370376586914, "learning_rate": 8.163930841694589e-08, "loss": 0.3092, "step": 141360 }, { "epoch": 141.39, "grad_norm": 56.48031234741211, "learning_rate": 8.107488753831161e-08, "loss": 0.2726, "step": 141390 }, { "epoch": 141.42, "grad_norm": 43.603981018066406, "learning_rate": 8.051240857341102e-08, "loss": 0.2756, "step": 141420 }, { "epoch": 141.45, "grad_norm": 1.7962971469387412e-05, "learning_rate": 7.995187174430152e-08, "loss": 0.2281, "step": 141450 }, { "epoch": 141.48, "grad_norm": 38.507606506347656, "learning_rate": 7.939327727227441e-08, "loss": 0.351, "step": 141480 }, { "epoch": 141.51, "grad_norm": 46.96717071533203, "learning_rate": 7.883662537785442e-08, "loss": 0.2579, "step": 141510 }, { "epoch": 141.54, "grad_norm": 0.00010439289326313883, "learning_rate": 7.828191628079851e-08, "loss": 0.3787, "step": 141540 }, { "epoch": 141.57, "grad_norm": 59.97026824951172, "learning_rate": 7.772915020009708e-08, "loss": 0.3482, "step": 141570 }, { "epoch": 141.6, "grad_norm": 55.872989654541016, "learning_rate": 7.717832735397335e-08, "loss": 0.2785, "step": 141600 }, { "epoch": 141.63, "grad_norm": 93.1576156616211, "learning_rate": 7.662944795988337e-08, "loss": 0.3685, "step": 141630 }, { "epoch": 141.66, "grad_norm": 19.671382904052734, "learning_rate": 7.608251223451601e-08, "loss": 0.421, "step": 141660 }, { "epoch": 141.69, "grad_norm": 44.653167724609375, "learning_rate": 7.553752039379359e-08, "loss": 0.3231, "step": 141690 }, { "epoch": 141.72, "grad_norm": 40.182701110839844, "learning_rate": 7.499447265286952e-08, "loss": 0.2875, "step": 141720 }, { "epoch": 141.75, "grad_norm": 44.59953689575195, "learning_rate": 7.445336922613067e-08, "loss": 0.3123, "step": 141750 }, { "epoch": 141.78, "grad_norm": 61.42672348022461, "learning_rate": 7.39142103271956e-08, "loss": 0.3198, "step": 141780 }, { "epoch": 141.81, "grad_norm": 65.91622924804688, "learning_rate": 7.33769961689168e-08, "loss": 0.2859, "step": 141810 }, { "epoch": 141.84, "grad_norm": 38.00876235961914, "learning_rate": 7.284172696337688e-08, "loss": 0.2915, "step": 141840 }, { "epoch": 141.87, "grad_norm": 33.1888313293457, "learning_rate": 7.230840292189179e-08, "loss": 0.2849, "step": 141870 }, { "epoch": 141.9, "grad_norm": 55.5238151550293, "learning_rate": 7.177702425500977e-08, "loss": 0.3016, "step": 141900 }, { "epoch": 141.93, "grad_norm": 36.1206169128418, "learning_rate": 7.124759117251078e-08, "loss": 0.3262, "step": 141930 }, { "epoch": 141.96, "grad_norm": 60.590179443359375, "learning_rate": 7.072010388340656e-08, "loss": 0.3632, "step": 141960 }, { "epoch": 141.99, "grad_norm": 9.311408939538524e-05, "learning_rate": 7.019456259594049e-08, "loss": 0.2245, "step": 141990 }, { "epoch": 142.0, "eval_loss": 0.2408023625612259, "eval_map": 0.8404, "eval_map_50": 0.9641, "eval_map_75": 0.9337, "eval_map_chicken": 0.8276, "eval_map_duck": 0.797, "eval_map_large": 0.8463, "eval_map_medium": 0.8406, "eval_map_plant": 0.8967, "eval_map_small": 0.3176, "eval_mar_1": 0.3371, "eval_mar_10": 0.8702, "eval_mar_100": 0.874, "eval_mar_100_chicken": 0.8716, "eval_mar_100_duck": 0.8304, "eval_mar_100_plant": 0.9201, "eval_mar_large": 0.878, "eval_mar_medium": 0.8783, "eval_mar_small": 0.3929, "eval_runtime": 13.3732, "eval_samples_per_second": 14.955, "eval_steps_per_second": 1.869, "step": 142000 }, { "epoch": 142.02, "grad_norm": 22.433849334716797, "learning_rate": 6.967096751758773e-08, "loss": 0.3396, "step": 142020 }, { "epoch": 142.05, "grad_norm": 155.00833129882812, "learning_rate": 6.914931885505626e-08, "loss": 0.375, "step": 142050 }, { "epoch": 142.08, "grad_norm": 178.66238403320312, "learning_rate": 6.862961681428304e-08, "loss": 0.2966, "step": 142080 }, { "epoch": 142.11, "grad_norm": 121.82696533203125, "learning_rate": 6.811186160044004e-08, "loss": 0.361, "step": 142110 }, { "epoch": 142.14, "grad_norm": 101.87291717529297, "learning_rate": 6.759605341792819e-08, "loss": 0.3233, "step": 142140 }, { "epoch": 142.17, "grad_norm": 26.818071365356445, "learning_rate": 6.708219247038017e-08, "loss": 0.3177, "step": 142170 }, { "epoch": 142.2, "grad_norm": 69.10475158691406, "learning_rate": 6.657027896065982e-08, "loss": 0.4748, "step": 142200 }, { "epoch": 142.23, "grad_norm": 68.20149993896484, "learning_rate": 6.606031309086269e-08, "loss": 0.2715, "step": 142230 }, { "epoch": 142.26, "grad_norm": 56.48291778564453, "learning_rate": 6.555229506231608e-08, "loss": 0.3472, "step": 142260 }, { "epoch": 142.29, "grad_norm": 37.22603225708008, "learning_rate": 6.504622507557679e-08, "loss": 0.4289, "step": 142290 }, { "epoch": 142.32, "grad_norm": 52.79399490356445, "learning_rate": 6.454210333043275e-08, "loss": 0.3039, "step": 142320 }, { "epoch": 142.35, "grad_norm": 46.64039993286133, "learning_rate": 6.403993002590425e-08, "loss": 0.253, "step": 142350 }, { "epoch": 142.38, "grad_norm": 49.42123794555664, "learning_rate": 6.353970536024045e-08, "loss": 0.3166, "step": 142380 }, { "epoch": 142.41, "grad_norm": 21.660137176513672, "learning_rate": 6.304142953092285e-08, "loss": 0.3348, "step": 142410 }, { "epoch": 142.44, "grad_norm": 39.82235336303711, "learning_rate": 6.254510273466186e-08, "loss": 0.2907, "step": 142440 }, { "epoch": 142.47, "grad_norm": 124.7683334350586, "learning_rate": 6.205072516740129e-08, "loss": 0.3199, "step": 142470 }, { "epoch": 142.5, "grad_norm": 170.65469360351562, "learning_rate": 6.15582970243117e-08, "loss": 0.3111, "step": 142500 }, { "epoch": 142.53, "grad_norm": 53.45037841796875, "learning_rate": 6.106781849979648e-08, "loss": 0.2517, "step": 142530 }, { "epoch": 142.56, "grad_norm": 64.79625701904297, "learning_rate": 6.057928978748906e-08, "loss": 0.2742, "step": 142560 }, { "epoch": 142.59, "grad_norm": 58.203224182128906, "learning_rate": 6.009271108025294e-08, "loss": 0.283, "step": 142590 }, { "epoch": 142.62, "grad_norm": 139.23681640625, "learning_rate": 5.960808257018113e-08, "loss": 0.2678, "step": 142620 }, { "epoch": 142.65, "grad_norm": 38.096248626708984, "learning_rate": 5.9125404448597825e-08, "loss": 0.2757, "step": 142650 }, { "epoch": 142.68, "grad_norm": 39.13498306274414, "learning_rate": 5.864467690605613e-08, "loss": 0.2689, "step": 142680 }, { "epoch": 142.71, "grad_norm": 26.894298553466797, "learning_rate": 5.8165900132340356e-08, "loss": 0.2712, "step": 142710 }, { "epoch": 142.74, "grad_norm": 175.10646057128906, "learning_rate": 5.76890743164632e-08, "loss": 0.2924, "step": 142740 }, { "epoch": 142.77, "grad_norm": 56.27349090576172, "learning_rate": 5.7214199646669076e-08, "loss": 0.3226, "step": 142770 }, { "epoch": 142.8, "grad_norm": 39.49784469604492, "learning_rate": 5.674127631043025e-08, "loss": 0.3803, "step": 142800 }, { "epoch": 142.83, "grad_norm": 56.85136795043945, "learning_rate": 5.6270304494449035e-08, "loss": 0.3562, "step": 142830 }, { "epoch": 142.86, "grad_norm": 83.35325622558594, "learning_rate": 5.580128438465837e-08, "loss": 0.276, "step": 142860 }, { "epoch": 142.89, "grad_norm": 61.11635971069336, "learning_rate": 5.533421616621903e-08, "loss": 0.3355, "step": 142890 }, { "epoch": 142.92, "grad_norm": 139.5830078125, "learning_rate": 5.4869100023523526e-08, "loss": 0.2944, "step": 142920 }, { "epoch": 142.95, "grad_norm": 52.97641372680664, "learning_rate": 5.440593614019107e-08, "loss": 0.3449, "step": 142950 }, { "epoch": 142.98, "grad_norm": 43.80342102050781, "learning_rate": 5.394472469907208e-08, "loss": 0.3448, "step": 142980 }, { "epoch": 143.0, "eval_loss": 0.2393781840801239, "eval_map": 0.8421, "eval_map_50": 0.9641, "eval_map_75": 0.9344, "eval_map_chicken": 0.8322, "eval_map_duck": 0.7968, "eval_map_large": 0.851, "eval_map_medium": 0.8427, "eval_map_plant": 0.8974, "eval_map_small": 0.3271, "eval_mar_1": 0.3385, "eval_mar_10": 0.8725, "eval_mar_100": 0.8763, "eval_mar_100_chicken": 0.8753, "eval_mar_100_duck": 0.8325, "eval_mar_100_plant": 0.9213, "eval_mar_large": 0.8815, "eval_mar_medium": 0.8804, "eval_mar_small": 0.3978, "eval_runtime": 12.8486, "eval_samples_per_second": 15.566, "eval_steps_per_second": 1.946, "step": 143000 }, { "epoch": 143.01, "grad_norm": 100.4185562133789, "learning_rate": 5.348546588224535e-08, "loss": 0.2466, "step": 143010 }, { "epoch": 143.04, "grad_norm": 33.929100036621094, "learning_rate": 5.302815987101917e-08, "loss": 0.3202, "step": 143040 }, { "epoch": 143.07, "grad_norm": 43.5810432434082, "learning_rate": 5.2572806845930244e-08, "loss": 0.2467, "step": 143070 }, { "epoch": 143.1, "grad_norm": 47.08922576904297, "learning_rate": 5.2119406986745336e-08, "loss": 0.3851, "step": 143100 }, { "epoch": 143.13, "grad_norm": 46.243961334228516, "learning_rate": 5.1667960472459034e-08, "loss": 0.3798, "step": 143130 }, { "epoch": 143.16, "grad_norm": 39.1627197265625, "learning_rate": 5.121846748129544e-08, "loss": 0.3011, "step": 143160 }, { "epoch": 143.19, "grad_norm": 6.975133146625012e-05, "learning_rate": 5.077092819070761e-08, "loss": 0.2489, "step": 143190 }, { "epoch": 143.22, "grad_norm": 58.999969482421875, "learning_rate": 5.032534277737644e-08, "loss": 0.2989, "step": 143220 }, { "epoch": 143.25, "grad_norm": 43.22013854980469, "learning_rate": 4.988171141721232e-08, "loss": 0.2743, "step": 143250 }, { "epoch": 143.28, "grad_norm": 79.94573974609375, "learning_rate": 4.944003428535349e-08, "loss": 0.2756, "step": 143280 }, { "epoch": 143.31, "grad_norm": 59.805904388427734, "learning_rate": 4.900031155616769e-08, "loss": 0.3333, "step": 143310 }, { "epoch": 143.34, "grad_norm": 42.778018951416016, "learning_rate": 4.856254340325051e-08, "loss": 0.3207, "step": 143340 }, { "epoch": 143.37, "grad_norm": 202.1372833251953, "learning_rate": 4.81267299994248e-08, "loss": 0.2779, "step": 143370 }, { "epoch": 143.4, "grad_norm": 46.233829498291016, "learning_rate": 4.769287151674407e-08, "loss": 0.2659, "step": 143400 }, { "epoch": 143.43, "grad_norm": 52.271053314208984, "learning_rate": 4.726096812648795e-08, "loss": 0.3338, "step": 143430 }, { "epoch": 143.46, "grad_norm": 61.19332504272461, "learning_rate": 4.683101999916562e-08, "loss": 0.2867, "step": 143460 }, { "epoch": 143.49, "grad_norm": 41.802677154541016, "learning_rate": 4.6403027304513513e-08, "loss": 0.2946, "step": 143490 }, { "epoch": 143.52, "grad_norm": 31.53972816467285, "learning_rate": 4.597699021149649e-08, "loss": 0.3314, "step": 143520 }, { "epoch": 143.55, "grad_norm": 33.493900299072266, "learning_rate": 4.5552908888306654e-08, "loss": 0.3229, "step": 143550 }, { "epoch": 143.58, "grad_norm": 38.010982513427734, "learning_rate": 4.5130783502365106e-08, "loss": 0.3194, "step": 143580 }, { "epoch": 143.61, "grad_norm": 80.27337646484375, "learning_rate": 4.4710614220320746e-08, "loss": 0.2764, "step": 143610 }, { "epoch": 143.64, "grad_norm": 53.67249298095703, "learning_rate": 4.429240120804923e-08, "loss": 0.2817, "step": 143640 }, { "epoch": 143.67, "grad_norm": 109.22937774658203, "learning_rate": 4.387614463065404e-08, "loss": 0.3688, "step": 143670 }, { "epoch": 143.7, "grad_norm": 112.37272644042969, "learning_rate": 4.346184465246761e-08, "loss": 0.3005, "step": 143700 }, { "epoch": 143.73, "grad_norm": 65.33618927001953, "learning_rate": 4.304950143704745e-08, "loss": 0.3486, "step": 143730 }, { "epoch": 143.76, "grad_norm": 78.86051940917969, "learning_rate": 4.263911514718222e-08, "loss": 0.3125, "step": 143760 }, { "epoch": 143.79, "grad_norm": 139.55819702148438, "learning_rate": 4.2230685944884554e-08, "loss": 0.3264, "step": 143790 }, { "epoch": 143.82, "grad_norm": 65.05883026123047, "learning_rate": 4.1824213991396024e-08, "loss": 0.2538, "step": 143820 }, { "epoch": 143.85, "grad_norm": 54.82771682739258, "learning_rate": 4.1419699447186045e-08, "loss": 0.4264, "step": 143850 }, { "epoch": 143.88, "grad_norm": 5.0194004870718345e-05, "learning_rate": 4.10171424719491e-08, "loss": 0.2926, "step": 143880 }, { "epoch": 143.91, "grad_norm": 34.01335144042969, "learning_rate": 4.061654322460973e-08, "loss": 0.2721, "step": 143910 }, { "epoch": 143.94, "grad_norm": 74.68109893798828, "learning_rate": 4.0217901863317534e-08, "loss": 0.2855, "step": 143940 }, { "epoch": 143.97, "grad_norm": 59.768760681152344, "learning_rate": 3.9821218545449956e-08, "loss": 0.2749, "step": 143970 }, { "epoch": 144.0, "grad_norm": 87.50907897949219, "learning_rate": 3.9426493427611177e-08, "loss": 0.3681, "step": 144000 }, { "epoch": 144.0, "eval_loss": 0.2400701642036438, "eval_map": 0.8413, "eval_map_50": 0.9642, "eval_map_75": 0.9341, "eval_map_chicken": 0.8309, "eval_map_duck": 0.7958, "eval_map_large": 0.8491, "eval_map_medium": 0.8413, "eval_map_plant": 0.8974, "eval_map_small": 0.3267, "eval_mar_1": 0.3377, "eval_mar_10": 0.8713, "eval_mar_100": 0.8749, "eval_mar_100_chicken": 0.8738, "eval_mar_100_duck": 0.8299, "eval_mar_100_plant": 0.921, "eval_mar_large": 0.8802, "eval_mar_medium": 0.879, "eval_mar_small": 0.396, "eval_runtime": 13.545, "eval_samples_per_second": 14.766, "eval_steps_per_second": 1.846, "step": 144000 }, { "epoch": 144.03, "grad_norm": 40.8502082824707, "learning_rate": 3.9033726665632096e-08, "loss": 0.3148, "step": 144030 }, { "epoch": 144.06, "grad_norm": 91.63549041748047, "learning_rate": 3.864291841457146e-08, "loss": 0.3048, "step": 144060 }, { "epoch": 144.09, "grad_norm": 45.578216552734375, "learning_rate": 3.825406882871363e-08, "loss": 0.3251, "step": 144090 }, { "epoch": 144.12, "grad_norm": 80.2746810913086, "learning_rate": 3.786717806157136e-08, "loss": 0.3168, "step": 144120 }, { "epoch": 144.15, "grad_norm": 35.266990661621094, "learning_rate": 3.748224626588137e-08, "loss": 0.2221, "step": 144150 }, { "epoch": 144.18, "grad_norm": 60.2780647277832, "learning_rate": 3.7099273593609316e-08, "loss": 0.2992, "step": 144180 }, { "epoch": 144.21, "grad_norm": 52.68596649169922, "learning_rate": 3.6718260195947594e-08, "loss": 0.2666, "step": 144210 }, { "epoch": 144.24, "grad_norm": 48.9094352722168, "learning_rate": 3.633920622331311e-08, "loss": 0.3037, "step": 144240 }, { "epoch": 144.27, "grad_norm": 45.596031188964844, "learning_rate": 3.5962111825350585e-08, "loss": 0.2763, "step": 144270 }, { "epoch": 144.3, "grad_norm": 46.69120407104492, "learning_rate": 3.558697715093207e-08, "loss": 0.2787, "step": 144300 }, { "epoch": 144.33, "grad_norm": 48.58835983276367, "learning_rate": 3.521380234815297e-08, "loss": 0.3193, "step": 144330 }, { "epoch": 144.36, "grad_norm": 73.88804626464844, "learning_rate": 3.4842587564337674e-08, "loss": 0.2766, "step": 144360 }, { "epoch": 144.39, "grad_norm": 58.77862548828125, "learning_rate": 3.4473332946036164e-08, "loss": 0.3523, "step": 144390 }, { "epoch": 144.42, "grad_norm": 52.23748016357422, "learning_rate": 3.410603863902406e-08, "loss": 0.3865, "step": 144420 }, { "epoch": 144.45, "grad_norm": 850.9139404296875, "learning_rate": 3.374070478830316e-08, "loss": 0.3134, "step": 144450 }, { "epoch": 144.48, "grad_norm": 37.850624084472656, "learning_rate": 3.337733153810141e-08, "loss": 0.2591, "step": 144480 }, { "epoch": 144.51, "grad_norm": 71.7245101928711, "learning_rate": 3.301591903187351e-08, "loss": 0.3478, "step": 144510 }, { "epoch": 144.54, "grad_norm": 31.965068817138672, "learning_rate": 3.2656467412298665e-08, "loss": 0.2478, "step": 144540 }, { "epoch": 144.57, "grad_norm": 105.37738800048828, "learning_rate": 3.2298976821282804e-08, "loss": 0.3375, "step": 144570 }, { "epoch": 144.6, "grad_norm": 31.52541160583496, "learning_rate": 3.194344739995803e-08, "loss": 0.2834, "step": 144600 }, { "epoch": 144.63, "grad_norm": 46.88361358642578, "learning_rate": 3.158987928868151e-08, "loss": 0.3163, "step": 144630 }, { "epoch": 144.66, "grad_norm": 71.95752716064453, "learning_rate": 3.1238272627035494e-08, "loss": 0.3428, "step": 144660 }, { "epoch": 144.69, "grad_norm": 46.4920539855957, "learning_rate": 3.088862755383004e-08, "loss": 0.3118, "step": 144690 }, { "epoch": 144.72, "grad_norm": 37.72059631347656, "learning_rate": 3.054094420709863e-08, "loss": 0.3177, "step": 144720 }, { "epoch": 144.75, "grad_norm": 31.390506744384766, "learning_rate": 3.019522272410202e-08, "loss": 0.2941, "step": 144750 }, { "epoch": 144.78, "grad_norm": 128.3743438720703, "learning_rate": 2.985146324132438e-08, "loss": 0.3301, "step": 144780 }, { "epoch": 144.81, "grad_norm": 86.62149810791016, "learning_rate": 2.950966589447657e-08, "loss": 0.2494, "step": 144810 }, { "epoch": 144.84, "grad_norm": 60.859710693359375, "learning_rate": 2.9169830818496226e-08, "loss": 0.288, "step": 144840 }, { "epoch": 144.87, "grad_norm": 103.16439056396484, "learning_rate": 2.8831958147543805e-08, "loss": 0.313, "step": 144870 }, { "epoch": 144.9, "grad_norm": 477.981689453125, "learning_rate": 2.8496048015005385e-08, "loss": 0.3733, "step": 144900 }, { "epoch": 144.93, "grad_norm": 89.3824462890625, "learning_rate": 2.8162100553494887e-08, "loss": 0.3099, "step": 144930 }, { "epoch": 144.96, "grad_norm": 95.60401153564453, "learning_rate": 2.783011589484741e-08, "loss": 0.3444, "step": 144960 }, { "epoch": 144.99, "grad_norm": 3.576817107386887e-05, "learning_rate": 2.7500094170126447e-08, "loss": 0.2593, "step": 144990 }, { "epoch": 145.0, "eval_loss": 0.23933224380016327, "eval_map": 0.8414, "eval_map_50": 0.9641, "eval_map_75": 0.9344, "eval_map_chicken": 0.8313, "eval_map_duck": 0.7955, "eval_map_large": 0.8522, "eval_map_medium": 0.8411, "eval_map_plant": 0.8973, "eval_map_small": 0.3271, "eval_mar_1": 0.3378, "eval_mar_10": 0.8713, "eval_mar_100": 0.875, "eval_mar_100_chicken": 0.8742, "eval_mar_100_duck": 0.8294, "eval_mar_100_plant": 0.9213, "eval_mar_large": 0.8821, "eval_mar_medium": 0.879, "eval_mar_small": 0.3947, "eval_runtime": 12.7573, "eval_samples_per_second": 15.677, "eval_steps_per_second": 1.96, "step": 145000 }, { "epoch": 145.02, "grad_norm": 29.352874755859375, "learning_rate": 2.7172035509619442e-08, "loss": 0.2993, "step": 145020 }, { "epoch": 145.05, "grad_norm": 26.896503448486328, "learning_rate": 2.684594004283836e-08, "loss": 0.3853, "step": 145050 }, { "epoch": 145.08, "grad_norm": 0.0003059447917621583, "learning_rate": 2.6521807898520214e-08, "loss": 0.3302, "step": 145080 }, { "epoch": 145.11, "grad_norm": 101.3816909790039, "learning_rate": 2.6199639204628202e-08, "loss": 0.2852, "step": 145110 }, { "epoch": 145.14, "grad_norm": 86.7889633178711, "learning_rate": 2.5879434088348364e-08, "loss": 0.2624, "step": 145140 }, { "epoch": 145.17, "grad_norm": 100.32111358642578, "learning_rate": 2.556119267609347e-08, "loss": 0.3053, "step": 145170 }, { "epoch": 145.2, "grad_norm": 54.667118072509766, "learning_rate": 2.5244915093499134e-08, "loss": 0.2713, "step": 145200 }, { "epoch": 145.23, "grad_norm": 36.59981918334961, "learning_rate": 2.493060146542825e-08, "loss": 0.3028, "step": 145230 }, { "epoch": 145.26, "grad_norm": 28.94950294494629, "learning_rate": 2.46182519159649e-08, "loss": 0.3106, "step": 145260 }, { "epoch": 145.29, "grad_norm": 35.86062240600586, "learning_rate": 2.43078665684221e-08, "loss": 0.325, "step": 145290 }, { "epoch": 145.32, "grad_norm": 31.792987823486328, "learning_rate": 2.3999445545332955e-08, "loss": 0.2756, "step": 145320 }, { "epoch": 145.35, "grad_norm": 53.422367095947266, "learning_rate": 2.3692988968458398e-08, "loss": 0.3122, "step": 145350 }, { "epoch": 145.38, "grad_norm": 26.477041244506836, "learning_rate": 2.3388496958782203e-08, "loss": 0.2739, "step": 145380 }, { "epoch": 145.41, "grad_norm": 37.903465270996094, "learning_rate": 2.3085969636513217e-08, "loss": 0.2659, "step": 145410 }, { "epoch": 145.44, "grad_norm": 38.776676177978516, "learning_rate": 2.2785407121084236e-08, "loss": 0.3372, "step": 145440 }, { "epoch": 145.47, "grad_norm": 32.46723937988281, "learning_rate": 2.2486809531152563e-08, "loss": 0.2959, "step": 145470 }, { "epoch": 145.5, "grad_norm": 37.740718841552734, "learning_rate": 2.219017698460002e-08, "loss": 0.2828, "step": 145500 }, { "epoch": 145.53, "grad_norm": 43.8394889831543, "learning_rate": 2.1895509598532372e-08, "loss": 0.2822, "step": 145530 }, { "epoch": 145.56, "grad_norm": 35.27846908569336, "learning_rate": 2.1602807489279344e-08, "loss": 0.2926, "step": 145560 }, { "epoch": 145.59, "grad_norm": 47.12772750854492, "learning_rate": 2.1312070772395165e-08, "loss": 0.2896, "step": 145590 }, { "epoch": 145.62, "grad_norm": 5.450280514196493e-05, "learning_rate": 2.1023299562658584e-08, "loss": 0.255, "step": 145620 }, { "epoch": 145.65, "grad_norm": 70.20610046386719, "learning_rate": 2.0736493974071736e-08, "loss": 0.3111, "step": 145650 }, { "epoch": 145.68, "grad_norm": 25.958900451660156, "learning_rate": 2.0451654119860164e-08, "loss": 0.2551, "step": 145680 }, { "epoch": 145.71, "grad_norm": 36.5417366027832, "learning_rate": 2.016878011247503e-08, "loss": 0.3437, "step": 145710 }, { "epoch": 145.74, "grad_norm": 41.20005416870117, "learning_rate": 1.988787206359033e-08, "loss": 0.2652, "step": 145740 }, { "epoch": 145.77, "grad_norm": 331.1378479003906, "learning_rate": 1.9608930084104027e-08, "loss": 0.2832, "step": 145770 }, { "epoch": 145.8, "grad_norm": 50.504859924316406, "learning_rate": 1.9331954284137476e-08, "loss": 0.3033, "step": 145800 }, { "epoch": 145.83, "grad_norm": 61.66500473022461, "learning_rate": 1.9056944773037656e-08, "loss": 0.2964, "step": 145830 }, { "epoch": 145.86, "grad_norm": 47.21424102783203, "learning_rate": 1.8783901659372162e-08, "loss": 0.2866, "step": 145860 }, { "epoch": 145.89, "grad_norm": 27.87393569946289, "learning_rate": 1.8512825050935323e-08, "loss": 0.2597, "step": 145890 }, { "epoch": 145.92, "grad_norm": 40.78389358520508, "learning_rate": 1.8243715054744315e-08, "loss": 0.4559, "step": 145920 }, { "epoch": 145.95, "grad_norm": 69.98662567138672, "learning_rate": 1.7976571777038044e-08, "loss": 0.2694, "step": 145950 }, { "epoch": 145.98, "grad_norm": 58.9811897277832, "learning_rate": 1.7711395323281588e-08, "loss": 0.4266, "step": 145980 }, { "epoch": 146.0, "eval_loss": 0.2398325800895691, "eval_map": 0.8403, "eval_map_50": 0.9641, "eval_map_75": 0.934, "eval_map_chicken": 0.8302, "eval_map_duck": 0.7933, "eval_map_large": 0.8493, "eval_map_medium": 0.8405, "eval_map_plant": 0.8974, "eval_map_small": 0.3289, "eval_mar_1": 0.3368, "eval_mar_10": 0.8705, "eval_mar_100": 0.8745, "eval_mar_100_chicken": 0.8728, "eval_mar_100_duck": 0.8289, "eval_mar_100_plant": 0.9218, "eval_mar_large": 0.8799, "eval_mar_medium": 0.8786, "eval_mar_small": 0.4008, "eval_runtime": 12.7481, "eval_samples_per_second": 15.689, "eval_steps_per_second": 1.961, "step": 146000 }, { "epoch": 146.01, "grad_norm": 81.97119140625, "learning_rate": 1.7448185798161765e-08, "loss": 0.2942, "step": 146010 }, { "epoch": 146.04, "grad_norm": 58.759220123291016, "learning_rate": 1.71869433055899e-08, "loss": 0.3594, "step": 146040 }, { "epoch": 146.07, "grad_norm": 72.47972106933594, "learning_rate": 1.6927667948700155e-08, "loss": 0.3132, "step": 146070 }, { "epoch": 146.1, "grad_norm": 104.22921752929688, "learning_rate": 1.6670359829850657e-08, "loss": 0.3061, "step": 146100 }, { "epoch": 146.13, "grad_norm": 202.91465759277344, "learning_rate": 1.6415019050622373e-08, "loss": 0.2848, "step": 146130 }, { "epoch": 146.16, "grad_norm": 40.29695510864258, "learning_rate": 1.6161645711819664e-08, "loss": 0.331, "step": 146160 }, { "epoch": 146.19, "grad_norm": 39.34471893310547, "learning_rate": 1.5910239913470292e-08, "loss": 0.2468, "step": 146190 }, { "epoch": 146.22, "grad_norm": 46.67807388305664, "learning_rate": 1.5660801754825983e-08, "loss": 0.3019, "step": 146220 }, { "epoch": 146.25, "grad_norm": 178.4644775390625, "learning_rate": 1.541333133436018e-08, "loss": 0.4656, "step": 146250 }, { "epoch": 146.28, "grad_norm": 39.73921203613281, "learning_rate": 1.5167828749770853e-08, "loss": 0.2642, "step": 146280 }, { "epoch": 146.31, "grad_norm": 33.14162063598633, "learning_rate": 1.4924294097977687e-08, "loss": 0.2813, "step": 146310 }, { "epoch": 146.34, "grad_norm": 32.47589111328125, "learning_rate": 1.4682727475124891e-08, "loss": 0.2671, "step": 146340 }, { "epoch": 146.37, "grad_norm": 86.15673065185547, "learning_rate": 1.4443128976579513e-08, "loss": 0.2797, "step": 146370 }, { "epoch": 146.4, "grad_norm": 88.22364044189453, "learning_rate": 1.4205498696930332e-08, "loss": 0.3081, "step": 146400 }, { "epoch": 146.43, "grad_norm": 2.4407499950029887e-05, "learning_rate": 1.3969836729990637e-08, "loss": 0.277, "step": 146430 }, { "epoch": 146.46, "grad_norm": 66.0626449584961, "learning_rate": 1.3736143168796012e-08, "loss": 0.3188, "step": 146460 }, { "epoch": 146.49, "grad_norm": 41.01301574707031, "learning_rate": 1.3504418105604877e-08, "loss": 0.3178, "step": 146490 }, { "epoch": 146.52, "grad_norm": 76.22486877441406, "learning_rate": 1.3274661631899055e-08, "loss": 0.3259, "step": 146520 }, { "epoch": 146.55, "grad_norm": 68.09413146972656, "learning_rate": 1.3046873838381546e-08, "loss": 0.2799, "step": 146550 }, { "epoch": 146.58, "grad_norm": 56.76133346557617, "learning_rate": 1.2821054814980971e-08, "loss": 0.3215, "step": 146580 }, { "epoch": 146.61, "grad_norm": 174.92391967773438, "learning_rate": 1.2597204650845463e-08, "loss": 0.3041, "step": 146610 }, { "epoch": 146.64, "grad_norm": 66.56468963623047, "learning_rate": 1.2375323434348773e-08, "loss": 0.338, "step": 146640 }, { "epoch": 146.67, "grad_norm": 63.390689849853516, "learning_rate": 1.2155411253085835e-08, "loss": 0.3014, "step": 146670 }, { "epoch": 146.7, "grad_norm": 60.35597610473633, "learning_rate": 1.1937468193873869e-08, "loss": 0.2775, "step": 146700 }, { "epoch": 146.73, "grad_norm": 79.57076263427734, "learning_rate": 1.1721494342754048e-08, "loss": 0.3024, "step": 146730 }, { "epoch": 146.76, "grad_norm": 59.98455047607422, "learning_rate": 1.1507489784989278e-08, "loss": 0.2532, "step": 146760 }, { "epoch": 146.79, "grad_norm": 9.17486977414228e-05, "learning_rate": 1.129545460506476e-08, "loss": 0.3333, "step": 146790 }, { "epoch": 146.82, "grad_norm": 103.91783142089844, "learning_rate": 1.1085388886689085e-08, "loss": 0.3236, "step": 146820 }, { "epoch": 146.85, "grad_norm": 41.20187759399414, "learning_rate": 1.0877292712792586e-08, "loss": 0.2667, "step": 146850 }, { "epoch": 146.88, "grad_norm": 36.82316207885742, "learning_rate": 1.067116616552899e-08, "loss": 0.344, "step": 146880 }, { "epoch": 146.91, "grad_norm": 60.35491180419922, "learning_rate": 1.0467009326272648e-08, "loss": 0.3455, "step": 146910 }, { "epoch": 146.94, "grad_norm": 59.31444549560547, "learning_rate": 1.026482227562242e-08, "loss": 0.3441, "step": 146940 }, { "epoch": 146.97, "grad_norm": 45.72970199584961, "learning_rate": 1.0064605093397794e-08, "loss": 0.3488, "step": 146970 }, { "epoch": 147.0, "grad_norm": 27.167434692382812, "learning_rate": 9.866357858642206e-09, "loss": 0.2411, "step": 147000 }, { "epoch": 147.0, "eval_loss": 0.23921215534210205, "eval_map": 0.8412, "eval_map_50": 0.9641, "eval_map_75": 0.934, "eval_map_chicken": 0.8315, "eval_map_duck": 0.7949, "eval_map_large": 0.8507, "eval_map_medium": 0.8403, "eval_map_plant": 0.8973, "eval_map_small": 0.3283, "eval_mar_1": 0.3374, "eval_mar_10": 0.8709, "eval_mar_100": 0.8747, "eval_mar_100_chicken": 0.8736, "eval_mar_100_duck": 0.8289, "eval_mar_100_plant": 0.9217, "eval_mar_large": 0.8811, "eval_mar_medium": 0.8787, "eval_mar_small": 0.3978, "eval_runtime": 12.7983, "eval_samples_per_second": 15.627, "eval_steps_per_second": 1.953, "step": 147000 }, { "epoch": 147.03, "grad_norm": 68.04301452636719, "learning_rate": 9.670080649619717e-09, "loss": 0.2671, "step": 147030 }, { "epoch": 147.06, "grad_norm": 49.406471252441406, "learning_rate": 9.475773543818345e-09, "loss": 0.3022, "step": 147060 }, { "epoch": 147.09, "grad_norm": 133.29685974121094, "learning_rate": 9.283436617946173e-09, "loss": 0.3135, "step": 147090 }, { "epoch": 147.12, "grad_norm": 52.03335189819336, "learning_rate": 9.09306994793635e-09, "loss": 0.2778, "step": 147120 }, { "epoch": 147.15, "grad_norm": 77.40728759765625, "learning_rate": 8.904673608940983e-09, "loss": 0.2809, "step": 147150 }, { "epoch": 147.18, "grad_norm": 24.626285552978516, "learning_rate": 8.718247675337243e-09, "loss": 0.2886, "step": 147180 }, { "epoch": 147.21, "grad_norm": 53.9749755859375, "learning_rate": 8.53379222072237e-09, "loss": 0.281, "step": 147210 }, { "epoch": 147.24, "grad_norm": 142.32505798339844, "learning_rate": 8.351307317917002e-09, "loss": 0.3224, "step": 147240 }, { "epoch": 147.27, "grad_norm": 90.94452667236328, "learning_rate": 8.17079303896351e-09, "loss": 0.2979, "step": 147270 }, { "epoch": 147.3, "grad_norm": 49.234397888183594, "learning_rate": 7.992249455124889e-09, "loss": 0.3423, "step": 147300 }, { "epoch": 147.33, "grad_norm": 82.35476684570312, "learning_rate": 7.815676636888093e-09, "loss": 0.2933, "step": 147330 }, { "epoch": 147.36, "grad_norm": 30.57929039001465, "learning_rate": 7.641074653961244e-09, "loss": 0.2503, "step": 147360 }, { "epoch": 147.39, "grad_norm": 73.26297760009766, "learning_rate": 7.468443575274764e-09, "loss": 0.3509, "step": 147390 }, { "epoch": 147.42, "grad_norm": 103.71812438964844, "learning_rate": 7.297783468980246e-09, "loss": 0.297, "step": 147420 }, { "epoch": 147.45, "grad_norm": 0.0001304660690948367, "learning_rate": 7.129094402451575e-09, "loss": 0.2427, "step": 147450 }, { "epoch": 147.48, "grad_norm": 55.81462097167969, "learning_rate": 6.962376442284368e-09, "loss": 0.3383, "step": 147480 }, { "epoch": 147.51, "grad_norm": 93.39418029785156, "learning_rate": 6.797629654296533e-09, "loss": 0.286, "step": 147510 }, { "epoch": 147.54, "grad_norm": 53.413330078125, "learning_rate": 6.63485410352771e-09, "loss": 0.2571, "step": 147540 }, { "epoch": 147.57, "grad_norm": 32.398216247558594, "learning_rate": 6.4740498542387174e-09, "loss": 0.3188, "step": 147570 }, { "epoch": 147.6, "grad_norm": 97.60844421386719, "learning_rate": 6.315216969912663e-09, "loss": 0.3163, "step": 147600 }, { "epoch": 147.63, "grad_norm": 67.30962371826172, "learning_rate": 6.1583555132543886e-09, "loss": 0.2585, "step": 147630 }, { "epoch": 147.66, "grad_norm": 67.46554565429688, "learning_rate": 6.003465546189358e-09, "loss": 0.3162, "step": 147660 }, { "epoch": 147.69, "grad_norm": 80.94842529296875, "learning_rate": 5.850547129867546e-09, "loss": 0.2829, "step": 147690 }, { "epoch": 147.72, "grad_norm": 104.3691177368164, "learning_rate": 5.699600324657328e-09, "loss": 0.3874, "step": 147720 }, { "epoch": 147.75, "grad_norm": 126.78202819824219, "learning_rate": 5.5506251901504825e-09, "loss": 0.2896, "step": 147750 }, { "epoch": 147.78, "grad_norm": 47.32594680786133, "learning_rate": 5.403621785159407e-09, "loss": 0.2368, "step": 147780 }, { "epoch": 147.81, "grad_norm": 105.59767150878906, "learning_rate": 5.258590167719901e-09, "loss": 0.3576, "step": 147810 }, { "epoch": 147.84, "grad_norm": 39.52690505981445, "learning_rate": 5.115530395087276e-09, "loss": 0.2919, "step": 147840 }, { "epoch": 147.87, "grad_norm": 142.16648864746094, "learning_rate": 4.9744425237396865e-09, "loss": 0.4269, "step": 147870 }, { "epoch": 147.9, "grad_norm": 27.467119216918945, "learning_rate": 4.835326609376468e-09, "loss": 0.2951, "step": 147900 }, { "epoch": 147.93, "grad_norm": 120.70387268066406, "learning_rate": 4.6981827069181305e-09, "loss": 0.2924, "step": 147930 }, { "epoch": 147.96, "grad_norm": 39.85201644897461, "learning_rate": 4.5630108705063684e-09, "loss": 0.2397, "step": 147960 }, { "epoch": 147.99, "grad_norm": 0.00013461867638397962, "learning_rate": 4.429811153505714e-09, "loss": 0.2675, "step": 147990 }, { "epoch": 148.0, "eval_loss": 0.23934325575828552, "eval_map": 0.8412, "eval_map_50": 0.9641, "eval_map_75": 0.934, "eval_map_chicken": 0.8315, "eval_map_duck": 0.795, "eval_map_large": 0.8507, "eval_map_medium": 0.8407, "eval_map_plant": 0.8973, "eval_map_small": 0.3268, "eval_mar_1": 0.3374, "eval_mar_10": 0.871, "eval_mar_100": 0.8748, "eval_mar_100_chicken": 0.8738, "eval_mar_100_duck": 0.8289, "eval_mar_100_plant": 0.9215, "eval_mar_large": 0.881, "eval_mar_medium": 0.879, "eval_mar_small": 0.3947, "eval_runtime": 12.9965, "eval_samples_per_second": 15.389, "eval_steps_per_second": 1.924, "step": 148000 }, { "epoch": 148.02, "grad_norm": 23.768529891967773, "learning_rate": 4.298583608501328e-09, "loss": 0.2354, "step": 148020 }, { "epoch": 148.05, "grad_norm": 74.93134307861328, "learning_rate": 4.169328287299545e-09, "loss": 0.3732, "step": 148050 }, { "epoch": 148.08, "grad_norm": 37.484535217285156, "learning_rate": 4.042045240927883e-09, "loss": 0.4355, "step": 148080 }, { "epoch": 148.11, "grad_norm": 49.3876838684082, "learning_rate": 3.9167345196361454e-09, "loss": 0.2641, "step": 148110 }, { "epoch": 148.14, "grad_norm": 85.84772491455078, "learning_rate": 3.793396172895314e-09, "loss": 0.2854, "step": 148140 }, { "epoch": 148.17, "grad_norm": 42.09956359863281, "learning_rate": 3.672030249396441e-09, "loss": 0.3354, "step": 148170 }, { "epoch": 148.2, "grad_norm": 62.16353988647461, "learning_rate": 3.5526367970539765e-09, "loss": 0.3456, "step": 148200 }, { "epoch": 148.23, "grad_norm": 63.247642517089844, "learning_rate": 3.4352158630018837e-09, "loss": 0.242, "step": 148230 }, { "epoch": 148.26, "grad_norm": 28.344297409057617, "learning_rate": 3.31976749359586e-09, "loss": 0.321, "step": 148260 }, { "epoch": 148.29, "grad_norm": 29.875398635864258, "learning_rate": 3.206291734413891e-09, "loss": 0.3615, "step": 148290 }, { "epoch": 148.32, "grad_norm": 136.5042724609375, "learning_rate": 3.094788630254031e-09, "loss": 0.3272, "step": 148320 }, { "epoch": 148.35, "grad_norm": 45.96992492675781, "learning_rate": 2.9852582251355124e-09, "loss": 0.316, "step": 148350 }, { "epoch": 148.38, "grad_norm": 73.72418975830078, "learning_rate": 2.8777005622998567e-09, "loss": 0.274, "step": 148380 }, { "epoch": 148.41, "grad_norm": 56.12461471557617, "learning_rate": 2.772115684209209e-09, "loss": 0.3025, "step": 148410 }, { "epoch": 148.44, "grad_norm": 64.38630676269531, "learning_rate": 2.6685036325457826e-09, "loss": 0.2964, "step": 148440 }, { "epoch": 148.47, "grad_norm": 78.73645782470703, "learning_rate": 2.5668644482151892e-09, "loss": 0.2818, "step": 148470 }, { "epoch": 148.5, "grad_norm": 51.524024963378906, "learning_rate": 2.4671981713420003e-09, "loss": 0.2586, "step": 148500 }, { "epoch": 148.53, "grad_norm": 61.21031188964844, "learning_rate": 2.369504841273629e-09, "loss": 0.2708, "step": 148530 }, { "epoch": 148.56, "grad_norm": 36.56247329711914, "learning_rate": 2.2737844965775578e-09, "loss": 0.3043, "step": 148560 }, { "epoch": 148.59, "grad_norm": 50.6976432800293, "learning_rate": 2.1800371750430037e-09, "loss": 0.2613, "step": 148590 }, { "epoch": 148.62, "grad_norm": 104.1881103515625, "learning_rate": 2.088262913679251e-09, "loss": 0.2819, "step": 148620 }, { "epoch": 148.65, "grad_norm": 77.15486907958984, "learning_rate": 1.9984617487173174e-09, "loss": 0.4705, "step": 148650 }, { "epoch": 148.68, "grad_norm": 257.2453308105469, "learning_rate": 1.9106337156099553e-09, "loss": 0.3147, "step": 148680 }, { "epoch": 148.71, "grad_norm": 94.51361083984375, "learning_rate": 1.8247788490299846e-09, "loss": 0.2612, "step": 148710 }, { "epoch": 148.74, "grad_norm": 54.03693771362305, "learning_rate": 1.740897182871404e-09, "loss": 0.2791, "step": 148740 }, { "epoch": 148.77, "grad_norm": 49.87909698486328, "learning_rate": 1.6589887502493907e-09, "loss": 0.3407, "step": 148770 }, { "epoch": 148.8, "grad_norm": 28.649518966674805, "learning_rate": 1.5790535835003006e-09, "loss": 0.3216, "step": 148800 }, { "epoch": 148.83, "grad_norm": 4.305846232455224e-05, "learning_rate": 1.5010917141811132e-09, "loss": 0.2869, "step": 148830 }, { "epoch": 148.86, "grad_norm": 52.96532440185547, "learning_rate": 1.425103173069986e-09, "loss": 0.3236, "step": 148860 }, { "epoch": 148.89, "grad_norm": 67.40062713623047, "learning_rate": 1.3510879901657003e-09, "loss": 0.3153, "step": 148890 }, { "epoch": 148.92, "grad_norm": 231.43629455566406, "learning_rate": 1.2790461946887712e-09, "loss": 0.3195, "step": 148920 }, { "epoch": 148.95, "grad_norm": 79.40615844726562, "learning_rate": 1.2089778150797816e-09, "loss": 0.3172, "step": 148950 }, { "epoch": 148.98, "grad_norm": 70.4350357055664, "learning_rate": 1.1408828790010484e-09, "loss": 0.2945, "step": 148980 }, { "epoch": 149.0, "eval_loss": 0.23936256766319275, "eval_map": 0.8413, "eval_map_50": 0.9641, "eval_map_75": 0.9341, "eval_map_chicken": 0.8309, "eval_map_duck": 0.7956, "eval_map_large": 0.8507, "eval_map_medium": 0.8408, "eval_map_plant": 0.8973, "eval_map_small": 0.3268, "eval_mar_1": 0.3376, "eval_mar_10": 0.8711, "eval_mar_100": 0.8749, "eval_mar_100_chicken": 0.8738, "eval_mar_100_duck": 0.8294, "eval_mar_100_plant": 0.9215, "eval_mar_large": 0.881, "eval_mar_medium": 0.8792, "eval_mar_small": 0.3947, "eval_runtime": 13.2099, "eval_samples_per_second": 15.14, "eval_steps_per_second": 1.893, "step": 149000 }, { "epoch": 149.01, "grad_norm": 54.10673141479492, "learning_rate": 1.074761413334957e-09, "loss": 0.3406, "step": 149010 }, { "epoch": 149.04, "grad_norm": 151.46478271484375, "learning_rate": 1.0106134441850712e-09, "loss": 0.2945, "step": 149040 }, { "epoch": 149.07, "grad_norm": 34.716007232666016, "learning_rate": 9.484389968766882e-10, "loss": 0.3195, "step": 149070 }, { "epoch": 149.1, "grad_norm": 45.507484436035156, "learning_rate": 8.88238095955174e-10, "loss": 0.2918, "step": 149100 }, { "epoch": 149.13, "grad_norm": 36.53764724731445, "learning_rate": 8.300107651859623e-10, "loss": 0.3201, "step": 149130 }, { "epoch": 149.16, "grad_norm": 83.43916320800781, "learning_rate": 7.737570275573314e-10, "loss": 0.2829, "step": 149160 }, { "epoch": 149.19, "grad_norm": 32.47480773925781, "learning_rate": 7.194769052765171e-10, "loss": 0.354, "step": 149190 }, { "epoch": 149.22, "grad_norm": 33.71446990966797, "learning_rate": 6.671704197735995e-10, "loss": 0.2461, "step": 149220 }, { "epoch": 149.25, "grad_norm": 60.157772064208984, "learning_rate": 6.168375916970615e-10, "loss": 0.2918, "step": 149250 }, { "epoch": 149.28, "grad_norm": 36.07717514038086, "learning_rate": 5.684784409182298e-10, "loss": 0.4074, "step": 149280 }, { "epoch": 149.31, "grad_norm": 87.16706085205078, "learning_rate": 5.220929865284996e-10, "loss": 0.3109, "step": 149310 }, { "epoch": 149.34, "grad_norm": 51.461036682128906, "learning_rate": 4.776812468398895e-10, "loss": 0.3124, "step": 149340 }, { "epoch": 149.37, "grad_norm": 53.12019729614258, "learning_rate": 4.3524323938559655e-10, "loss": 0.26, "step": 149370 }, { "epoch": 149.4, "grad_norm": 31.98765754699707, "learning_rate": 3.9477898091944135e-10, "loss": 0.2911, "step": 149400 }, { "epoch": 149.43, "grad_norm": 45.87418746948242, "learning_rate": 3.562884874158679e-10, "loss": 0.3493, "step": 149430 }, { "epoch": 149.46, "grad_norm": 50.62438201904297, "learning_rate": 3.1977177407105376e-10, "loss": 0.3425, "step": 149460 }, { "epoch": 149.49, "grad_norm": 118.38064575195312, "learning_rate": 2.8522885530013475e-10, "loss": 0.4231, "step": 149490 }, { "epoch": 149.52, "grad_norm": 50.225257873535156, "learning_rate": 2.5265974474109054e-10, "loss": 0.2677, "step": 149520 }, { "epoch": 149.55, "grad_norm": 253.9156951904297, "learning_rate": 2.2206445525085886e-10, "loss": 0.3737, "step": 149550 }, { "epoch": 149.58, "grad_norm": 46.01321029663086, "learning_rate": 1.9344299890866614e-10, "loss": 0.4259, "step": 149580 }, { "epoch": 149.61, "grad_norm": 127.60037994384766, "learning_rate": 1.6679538701325215e-10, "loss": 0.3044, "step": 149610 }, { "epoch": 149.64, "grad_norm": 74.46773529052734, "learning_rate": 1.4212163008509028e-10, "loss": 0.3339, "step": 149640 }, { "epoch": 149.67, "grad_norm": 92.41695404052734, "learning_rate": 1.1942173786527732e-10, "loss": 0.285, "step": 149670 }, { "epoch": 149.7, "grad_norm": 43.35792541503906, "learning_rate": 9.869571931442334e-11, "loss": 0.3565, "step": 149700 }, { "epoch": 149.73, "grad_norm": 63.197906494140625, "learning_rate": 7.994358261542712e-11, "loss": 0.2996, "step": 149730 }, { "epoch": 149.76, "grad_norm": 38.37726974487305, "learning_rate": 6.316533517125578e-11, "loss": 0.2403, "step": 149760 }, { "epoch": 149.79, "grad_norm": 30.22950553894043, "learning_rate": 4.83609836054999e-11, "loss": 0.2904, "step": 149790 }, { "epoch": 149.82, "grad_norm": 50.76008987426758, "learning_rate": 3.55305337634837e-11, "loss": 0.3492, "step": 149820 }, { "epoch": 149.85, "grad_norm": 64.36622619628906, "learning_rate": 2.467399070893439e-11, "loss": 0.3102, "step": 149850 }, { "epoch": 149.88, "grad_norm": 34.911922454833984, "learning_rate": 1.57913587295333e-11, "loss": 0.2852, "step": 149880 }, { "epoch": 149.91, "grad_norm": 77.51419067382812, "learning_rate": 8.882641330809627e-12, "loss": 0.3452, "step": 149910 }, { "epoch": 149.94, "grad_norm": 37.67948532104492, "learning_rate": 3.947841241136452e-12, "loss": 0.2644, "step": 149940 }, { "epoch": 149.97, "grad_norm": 105.01229095458984, "learning_rate": 9.869604078449612e-13, "loss": 0.3033, "step": 149970 }, { "epoch": 150.0, "grad_norm": 62.513954162597656, "learning_rate": 0.0, "loss": 0.2848, "step": 150000 }, { "epoch": 150.0, "eval_loss": 0.23936401307582855, "eval_map": 0.8413, "eval_map_50": 0.9641, "eval_map_75": 0.9341, "eval_map_chicken": 0.8309, "eval_map_duck": 0.7956, "eval_map_large": 0.8507, "eval_map_medium": 0.8408, "eval_map_plant": 0.8973, "eval_map_small": 0.3268, "eval_mar_1": 0.3376, "eval_mar_10": 0.8711, "eval_mar_100": 0.8749, "eval_mar_100_chicken": 0.8738, "eval_mar_100_duck": 0.8294, "eval_mar_100_plant": 0.9215, "eval_mar_large": 0.881, "eval_mar_medium": 0.8792, "eval_mar_small": 0.3947, "eval_runtime": 14.2916, "eval_samples_per_second": 13.994, "eval_steps_per_second": 1.749, "step": 150000 } ], "logging_steps": 30, "max_steps": 150000, "num_input_tokens_seen": 0, "num_train_epochs": 150, "save_steps": 10, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.160346693632e+19, "train_batch_size": 2, "trial_name": null, "trial_params": null }