{ "best_metric": 0.894109396914446, "best_model_checkpoint": "swin-tiny-patch4-window7-224-classification/checkpoint-1000", "epoch": 9.975062344139651, "eval_steps": 500, "global_step": 1000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.1, "grad_norm": 5.313328742980957, "learning_rate": 1e-05, "loss": 2.0631, "step": 10 }, { "epoch": 0.2, "grad_norm": 5.775210857391357, "learning_rate": 2e-05, "loss": 1.7746, "step": 20 }, { "epoch": 0.3, "grad_norm": 4.816771984100342, "learning_rate": 3e-05, "loss": 1.5563, "step": 30 }, { "epoch": 0.4, "grad_norm": 6.007688999176025, "learning_rate": 4e-05, "loss": 1.2361, "step": 40 }, { "epoch": 0.5, "grad_norm": 7.807107925415039, "learning_rate": 5e-05, "loss": 1.0685, "step": 50 }, { "epoch": 0.6, "grad_norm": 6.920202732086182, "learning_rate": 6e-05, "loss": 0.9716, "step": 60 }, { "epoch": 0.7, "grad_norm": 5.678104877471924, "learning_rate": 7e-05, "loss": 0.9274, "step": 70 }, { "epoch": 0.8, "grad_norm": 7.1128010749816895, "learning_rate": 8e-05, "loss": 0.8962, "step": 80 }, { "epoch": 0.9, "grad_norm": 8.700326919555664, "learning_rate": 9e-05, "loss": 0.8924, "step": 90 }, { "epoch": 1.0, "grad_norm": 7.496997833251953, "learning_rate": 0.0001, "loss": 0.8514, "step": 100 }, { "epoch": 1.0, "eval_accuracy": 0.705820476858345, "eval_loss": 0.7862311005592346, "eval_runtime": 38.0587, "eval_samples_per_second": 74.937, "eval_steps_per_second": 1.182, "step": 100 }, { "epoch": 1.1, "grad_norm": 5.842683792114258, "learning_rate": 9.888888888888889e-05, "loss": 0.7961, "step": 110 }, { "epoch": 1.2, "grad_norm": 7.082647323608398, "learning_rate": 9.777777777777778e-05, "loss": 0.8179, "step": 120 }, { "epoch": 1.3, "grad_norm": 6.531412601470947, "learning_rate": 9.666666666666667e-05, "loss": 0.772, "step": 130 }, { "epoch": 1.4, "grad_norm": 4.981847763061523, "learning_rate": 9.555555555555557e-05, "loss": 0.7298, "step": 140 }, { "epoch": 1.5, "grad_norm": 6.9995198249816895, "learning_rate": 9.444444444444444e-05, "loss": 0.6882, "step": 150 }, { "epoch": 1.6, "grad_norm": 4.9766693115234375, "learning_rate": 9.333333333333334e-05, "loss": 0.7048, "step": 160 }, { "epoch": 1.7, "grad_norm": 5.899988174438477, "learning_rate": 9.222222222222223e-05, "loss": 0.7254, "step": 170 }, { "epoch": 1.8, "grad_norm": 4.040374755859375, "learning_rate": 9.111111111111112e-05, "loss": 0.6543, "step": 180 }, { "epoch": 1.9, "grad_norm": 6.156548023223877, "learning_rate": 9e-05, "loss": 0.6329, "step": 190 }, { "epoch": 2.0, "grad_norm": 6.268504619598389, "learning_rate": 8.888888888888889e-05, "loss": 0.631, "step": 200 }, { "epoch": 2.0, "eval_accuracy": 0.7885694249649369, "eval_loss": 0.5491923093795776, "eval_runtime": 38.032, "eval_samples_per_second": 74.989, "eval_steps_per_second": 1.183, "step": 200 }, { "epoch": 2.09, "grad_norm": 6.425785541534424, "learning_rate": 8.777777777777778e-05, "loss": 0.612, "step": 210 }, { "epoch": 2.19, "grad_norm": 5.754735469818115, "learning_rate": 8.666666666666667e-05, "loss": 0.6117, "step": 220 }, { "epoch": 2.29, "grad_norm": 5.51466703414917, "learning_rate": 8.555555555555556e-05, "loss": 0.6041, "step": 230 }, { "epoch": 2.39, "grad_norm": 5.170792102813721, "learning_rate": 8.444444444444444e-05, "loss": 0.6057, "step": 240 }, { "epoch": 2.49, "grad_norm": 5.319341659545898, "learning_rate": 8.333333333333334e-05, "loss": 0.556, "step": 250 }, { "epoch": 2.59, "grad_norm": 7.033453941345215, "learning_rate": 8.222222222222222e-05, "loss": 0.5918, "step": 260 }, { "epoch": 2.69, "grad_norm": 6.37441349029541, "learning_rate": 8.111111111111112e-05, "loss": 0.5603, "step": 270 }, { "epoch": 2.79, "grad_norm": 4.926444053649902, "learning_rate": 8e-05, "loss": 0.5827, "step": 280 }, { "epoch": 2.89, "grad_norm": 4.273934841156006, "learning_rate": 7.88888888888889e-05, "loss": 0.5406, "step": 290 }, { "epoch": 2.99, "grad_norm": 5.528937816619873, "learning_rate": 7.777777777777778e-05, "loss": 0.5516, "step": 300 }, { "epoch": 2.99, "eval_accuracy": 0.8078541374474053, "eval_loss": 0.5194684863090515, "eval_runtime": 38.7637, "eval_samples_per_second": 73.574, "eval_steps_per_second": 1.161, "step": 300 }, { "epoch": 3.09, "grad_norm": 9.39556884765625, "learning_rate": 7.666666666666667e-05, "loss": 0.5039, "step": 310 }, { "epoch": 3.19, "grad_norm": 5.658567428588867, "learning_rate": 7.555555555555556e-05, "loss": 0.497, "step": 320 }, { "epoch": 3.29, "grad_norm": 5.974184513092041, "learning_rate": 7.444444444444444e-05, "loss": 0.5041, "step": 330 }, { "epoch": 3.39, "grad_norm": 5.876557350158691, "learning_rate": 7.333333333333333e-05, "loss": 0.4978, "step": 340 }, { "epoch": 3.49, "grad_norm": 7.005821704864502, "learning_rate": 7.222222222222222e-05, "loss": 0.5145, "step": 350 }, { "epoch": 3.59, "grad_norm": 4.967439651489258, "learning_rate": 7.111111111111112e-05, "loss": 0.4803, "step": 360 }, { "epoch": 3.69, "grad_norm": 4.976831436157227, "learning_rate": 7e-05, "loss": 0.4816, "step": 370 }, { "epoch": 3.79, "grad_norm": 6.872544288635254, "learning_rate": 6.88888888888889e-05, "loss": 0.4811, "step": 380 }, { "epoch": 3.89, "grad_norm": 5.97269868850708, "learning_rate": 6.777777777777778e-05, "loss": 0.4978, "step": 390 }, { "epoch": 3.99, "grad_norm": 5.982359886169434, "learning_rate": 6.666666666666667e-05, "loss": 0.4597, "step": 400 }, { "epoch": 4.0, "eval_accuracy": 0.8229312762973352, "eval_loss": 0.48398053646087646, "eval_runtime": 38.5586, "eval_samples_per_second": 73.965, "eval_steps_per_second": 1.167, "step": 401 }, { "epoch": 4.09, "grad_norm": 4.751348972320557, "learning_rate": 6.555555555555556e-05, "loss": 0.4443, "step": 410 }, { "epoch": 4.19, "grad_norm": 4.549768447875977, "learning_rate": 6.444444444444446e-05, "loss": 0.423, "step": 420 }, { "epoch": 4.29, "grad_norm": 6.269441604614258, "learning_rate": 6.333333333333333e-05, "loss": 0.449, "step": 430 }, { "epoch": 4.39, "grad_norm": 4.581325531005859, "learning_rate": 6.222222222222222e-05, "loss": 0.4099, "step": 440 }, { "epoch": 4.49, "grad_norm": 5.31665563583374, "learning_rate": 6.111111111111112e-05, "loss": 0.4075, "step": 450 }, { "epoch": 4.59, "grad_norm": 5.4510416984558105, "learning_rate": 6e-05, "loss": 0.4195, "step": 460 }, { "epoch": 4.69, "grad_norm": 5.312283515930176, "learning_rate": 5.8888888888888896e-05, "loss": 0.3829, "step": 470 }, { "epoch": 4.79, "grad_norm": 5.865201950073242, "learning_rate": 5.7777777777777776e-05, "loss": 0.4209, "step": 480 }, { "epoch": 4.89, "grad_norm": 5.123416423797607, "learning_rate": 5.666666666666667e-05, "loss": 0.4094, "step": 490 }, { "epoch": 4.99, "grad_norm": 4.805253982543945, "learning_rate": 5.555555555555556e-05, "loss": 0.4094, "step": 500 }, { "epoch": 5.0, "eval_accuracy": 0.8597475455820477, "eval_loss": 0.3945947587490082, "eval_runtime": 38.9464, "eval_samples_per_second": 73.229, "eval_steps_per_second": 1.155, "step": 501 }, { "epoch": 5.09, "grad_norm": 4.796308517456055, "learning_rate": 5.4444444444444446e-05, "loss": 0.3555, "step": 510 }, { "epoch": 5.19, "grad_norm": 4.460272789001465, "learning_rate": 5.333333333333333e-05, "loss": 0.3719, "step": 520 }, { "epoch": 5.29, "grad_norm": 5.035310745239258, "learning_rate": 5.222222222222223e-05, "loss": 0.3785, "step": 530 }, { "epoch": 5.39, "grad_norm": 5.145148754119873, "learning_rate": 5.111111111111111e-05, "loss": 0.3881, "step": 540 }, { "epoch": 5.49, "grad_norm": 6.646022319793701, "learning_rate": 5e-05, "loss": 0.3776, "step": 550 }, { "epoch": 5.59, "grad_norm": 6.039920330047607, "learning_rate": 4.888888888888889e-05, "loss": 0.3891, "step": 560 }, { "epoch": 5.69, "grad_norm": 4.1056623458862305, "learning_rate": 4.7777777777777784e-05, "loss": 0.3369, "step": 570 }, { "epoch": 5.79, "grad_norm": 5.633683204650879, "learning_rate": 4.666666666666667e-05, "loss": 0.3606, "step": 580 }, { "epoch": 5.89, "grad_norm": 5.873307704925537, "learning_rate": 4.555555555555556e-05, "loss": 0.3386, "step": 590 }, { "epoch": 5.99, "grad_norm": 5.3684983253479, "learning_rate": 4.4444444444444447e-05, "loss": 0.3289, "step": 600 }, { "epoch": 6.0, "eval_accuracy": 0.8572931276297335, "eval_loss": 0.3786664605140686, "eval_runtime": 40.3623, "eval_samples_per_second": 70.66, "eval_steps_per_second": 1.115, "step": 601 }, { "epoch": 6.08, "grad_norm": 4.093196392059326, "learning_rate": 4.3333333333333334e-05, "loss": 0.3251, "step": 610 }, { "epoch": 6.18, "grad_norm": 4.739712715148926, "learning_rate": 4.222222222222222e-05, "loss": 0.3029, "step": 620 }, { "epoch": 6.28, "grad_norm": 3.9042539596557617, "learning_rate": 4.111111111111111e-05, "loss": 0.3269, "step": 630 }, { "epoch": 6.38, "grad_norm": 5.695065498352051, "learning_rate": 4e-05, "loss": 0.3318, "step": 640 }, { "epoch": 6.48, "grad_norm": 4.814879894256592, "learning_rate": 3.888888888888889e-05, "loss": 0.3087, "step": 650 }, { "epoch": 6.58, "grad_norm": 4.633285999298096, "learning_rate": 3.777777777777778e-05, "loss": 0.3416, "step": 660 }, { "epoch": 6.68, "grad_norm": 4.862430095672607, "learning_rate": 3.6666666666666666e-05, "loss": 0.3063, "step": 670 }, { "epoch": 6.78, "grad_norm": 4.930345058441162, "learning_rate": 3.555555555555556e-05, "loss": 0.3225, "step": 680 }, { "epoch": 6.88, "grad_norm": 4.587200164794922, "learning_rate": 3.444444444444445e-05, "loss": 0.3139, "step": 690 }, { "epoch": 6.98, "grad_norm": 5.570033550262451, "learning_rate": 3.3333333333333335e-05, "loss": 0.3145, "step": 700 }, { "epoch": 6.99, "eval_accuracy": 0.8772791023842917, "eval_loss": 0.3536257743835449, "eval_runtime": 42.7073, "eval_samples_per_second": 66.78, "eval_steps_per_second": 1.054, "step": 701 }, { "epoch": 7.08, "grad_norm": 5.562831401824951, "learning_rate": 3.222222222222223e-05, "loss": 0.2998, "step": 710 }, { "epoch": 7.18, "grad_norm": 5.306057453155518, "learning_rate": 3.111111111111111e-05, "loss": 0.2807, "step": 720 }, { "epoch": 7.28, "grad_norm": 5.671096324920654, "learning_rate": 3e-05, "loss": 0.2794, "step": 730 }, { "epoch": 7.38, "grad_norm": 5.412108421325684, "learning_rate": 2.8888888888888888e-05, "loss": 0.2814, "step": 740 }, { "epoch": 7.48, "grad_norm": 4.249139308929443, "learning_rate": 2.777777777777778e-05, "loss": 0.2531, "step": 750 }, { "epoch": 7.58, "grad_norm": 5.714820861816406, "learning_rate": 2.6666666666666667e-05, "loss": 0.2712, "step": 760 }, { "epoch": 7.68, "grad_norm": 5.9426469802856445, "learning_rate": 2.5555555555555554e-05, "loss": 0.2589, "step": 770 }, { "epoch": 7.78, "grad_norm": 4.787610054016113, "learning_rate": 2.4444444444444445e-05, "loss": 0.2745, "step": 780 }, { "epoch": 7.88, "grad_norm": 3.9021761417388916, "learning_rate": 2.3333333333333336e-05, "loss": 0.2728, "step": 790 }, { "epoch": 7.98, "grad_norm": 5.153603553771973, "learning_rate": 2.2222222222222223e-05, "loss": 0.2911, "step": 800 }, { "epoch": 8.0, "eval_accuracy": 0.885343618513324, "eval_loss": 0.32486388087272644, "eval_runtime": 42.9219, "eval_samples_per_second": 66.446, "eval_steps_per_second": 1.048, "step": 802 }, { "epoch": 8.08, "grad_norm": 5.644151210784912, "learning_rate": 2.111111111111111e-05, "loss": 0.2559, "step": 810 }, { "epoch": 8.18, "grad_norm": 4.023806095123291, "learning_rate": 2e-05, "loss": 0.2415, "step": 820 }, { "epoch": 8.28, "grad_norm": 6.214153289794922, "learning_rate": 1.888888888888889e-05, "loss": 0.2447, "step": 830 }, { "epoch": 8.38, "grad_norm": 3.903085231781006, "learning_rate": 1.777777777777778e-05, "loss": 0.2602, "step": 840 }, { "epoch": 8.48, "grad_norm": 5.1828131675720215, "learning_rate": 1.6666666666666667e-05, "loss": 0.2529, "step": 850 }, { "epoch": 8.58, "grad_norm": 4.840915203094482, "learning_rate": 1.5555555555555555e-05, "loss": 0.2302, "step": 860 }, { "epoch": 8.68, "grad_norm": 5.424919128417969, "learning_rate": 1.4444444444444444e-05, "loss": 0.246, "step": 870 }, { "epoch": 8.78, "grad_norm": 6.8232316970825195, "learning_rate": 1.3333333333333333e-05, "loss": 0.2435, "step": 880 }, { "epoch": 8.88, "grad_norm": 4.856207847595215, "learning_rate": 1.2222222222222222e-05, "loss": 0.2376, "step": 890 }, { "epoch": 8.98, "grad_norm": 5.239475250244141, "learning_rate": 1.1111111111111112e-05, "loss": 0.2487, "step": 900 }, { "epoch": 9.0, "eval_accuracy": 0.8937587657784011, "eval_loss": 0.30939263105392456, "eval_runtime": 39.5102, "eval_samples_per_second": 72.184, "eval_steps_per_second": 1.139, "step": 902 }, { "epoch": 9.08, "grad_norm": 5.607595443725586, "learning_rate": 1e-05, "loss": 0.2253, "step": 910 }, { "epoch": 9.18, "grad_norm": 3.9586544036865234, "learning_rate": 8.88888888888889e-06, "loss": 0.2478, "step": 920 }, { "epoch": 9.28, "grad_norm": 3.8814537525177, "learning_rate": 7.777777777777777e-06, "loss": 0.2173, "step": 930 }, { "epoch": 9.38, "grad_norm": 7.456895351409912, "learning_rate": 6.666666666666667e-06, "loss": 0.2315, "step": 940 }, { "epoch": 9.48, "grad_norm": 4.543576717376709, "learning_rate": 5.555555555555556e-06, "loss": 0.2281, "step": 950 }, { "epoch": 9.58, "grad_norm": 4.659777641296387, "learning_rate": 4.444444444444445e-06, "loss": 0.2324, "step": 960 }, { "epoch": 9.68, "grad_norm": 4.048426151275635, "learning_rate": 3.3333333333333333e-06, "loss": 0.2132, "step": 970 }, { "epoch": 9.78, "grad_norm": 3.969958543777466, "learning_rate": 2.2222222222222225e-06, "loss": 0.2076, "step": 980 }, { "epoch": 9.88, "grad_norm": 5.621995449066162, "learning_rate": 1.1111111111111112e-06, "loss": 0.2262, "step": 990 }, { "epoch": 9.98, "grad_norm": 4.668450355529785, "learning_rate": 0.0, "loss": 0.2291, "step": 1000 }, { "epoch": 9.98, "eval_accuracy": 0.894109396914446, "eval_loss": 0.303189218044281, "eval_runtime": 39.0907, "eval_samples_per_second": 72.958, "eval_steps_per_second": 1.151, "step": 1000 }, { "epoch": 9.98, "step": 1000, "total_flos": 6.364199987970048e+18, "train_loss": 0.48602974390983583, "train_runtime": 5351.8475, "train_samples_per_second": 47.954, "train_steps_per_second": 0.187 } ], "logging_steps": 10, "max_steps": 1000, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "total_flos": 6.364199987970048e+18, "train_batch_size": 64, "trial_name": null, "trial_params": null }