|
{ |
|
"best_metric": 0.9825666666666667, |
|
"best_model_checkpoint": "ConvNeXT_AI_image_detector/checkpoint-9843", |
|
"epoch": 9.993142857142857, |
|
"global_step": 10930, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.574565416285453e-07, |
|
"loss": 0.7008, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.149130832570906e-07, |
|
"loss": 0.7003, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.372369624885636e-06, |
|
"loss": 0.6998, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.8298261665141813e-06, |
|
"loss": 0.7049, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 2.287282708142727e-06, |
|
"loss": 0.7058, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 2.744739249771272e-06, |
|
"loss": 0.6932, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 3.2021957913998174e-06, |
|
"loss": 0.6891, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 3.6596523330283626e-06, |
|
"loss": 0.6989, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.1171088746569085e-06, |
|
"loss": 0.6856, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.574565416285454e-06, |
|
"loss": 0.6816, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 5.032021957913998e-06, |
|
"loss": 0.6759, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 5.489478499542544e-06, |
|
"loss": 0.6651, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 5.946935041171089e-06, |
|
"loss": 0.6603, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 6.404391582799635e-06, |
|
"loss": 0.6629, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 6.86184812442818e-06, |
|
"loss": 0.6433, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 7.319304666056725e-06, |
|
"loss": 0.643, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 7.77676120768527e-06, |
|
"loss": 0.6202, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.234217749313817e-06, |
|
"loss": 0.6105, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.691674290942362e-06, |
|
"loss": 0.6101, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.149130832570907e-06, |
|
"loss": 0.5807, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.606587374199452e-06, |
|
"loss": 0.5529, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.0064043915827996e-05, |
|
"loss": 0.5322, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.0521500457456541e-05, |
|
"loss": 0.5229, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.0978956999085088e-05, |
|
"loss": 0.4976, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.1436413540713633e-05, |
|
"loss": 0.4647, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.1893870082342178e-05, |
|
"loss": 0.4669, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.2351326623970723e-05, |
|
"loss": 0.4221, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.280878316559927e-05, |
|
"loss": 0.4471, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.3266239707227815e-05, |
|
"loss": 0.4232, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.372369624885636e-05, |
|
"loss": 0.4298, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.4181152790484905e-05, |
|
"loss": 0.3925, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.463860933211345e-05, |
|
"loss": 0.4, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.5096065873741997e-05, |
|
"loss": 0.3907, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.555352241537054e-05, |
|
"loss": 0.3892, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.6010978956999087e-05, |
|
"loss": 0.3593, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.6468435498627634e-05, |
|
"loss": 0.3609, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.6925892040256177e-05, |
|
"loss": 0.3451, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.7383348581884724e-05, |
|
"loss": 0.3043, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.7840805123513268e-05, |
|
"loss": 0.376, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.8298261665141814e-05, |
|
"loss": 0.3395, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.8755718206770358e-05, |
|
"loss": 0.3348, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.9213174748398905e-05, |
|
"loss": 0.3165, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.9670631290027448e-05, |
|
"loss": 0.3079, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 2.012808783165599e-05, |
|
"loss": 0.3513, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 2.0585544373284538e-05, |
|
"loss": 0.3309, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 2.1043000914913082e-05, |
|
"loss": 0.3129, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 2.150045745654163e-05, |
|
"loss": 0.2876, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 2.1957913998170175e-05, |
|
"loss": 0.2744, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 2.241537053979872e-05, |
|
"loss": 0.2849, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 2.2872827081427266e-05, |
|
"loss": 0.2864, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 2.333028362305581e-05, |
|
"loss": 0.2885, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 2.3787740164684356e-05, |
|
"loss": 0.2808, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 2.42451967063129e-05, |
|
"loss": 0.307, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 2.4702653247941446e-05, |
|
"loss": 0.3107, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 2.5160109789569993e-05, |
|
"loss": 0.3062, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 2.561756633119854e-05, |
|
"loss": 0.3214, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 2.6075022872827083e-05, |
|
"loss": 0.2785, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 2.653247941445563e-05, |
|
"loss": 0.2914, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 2.6989935956084173e-05, |
|
"loss": 0.2793, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 2.744739249771272e-05, |
|
"loss": 0.2943, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 2.7904849039341263e-05, |
|
"loss": 0.2301, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 2.836230558096981e-05, |
|
"loss": 0.2968, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 2.8819762122598354e-05, |
|
"loss": 0.292, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 2.92772186642269e-05, |
|
"loss": 0.2695, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 2.9734675205855444e-05, |
|
"loss": 0.2298, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.0192131747483994e-05, |
|
"loss": 0.2619, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.064958828911254e-05, |
|
"loss": 0.2452, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.110704483074108e-05, |
|
"loss": 0.2977, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.1564501372369624e-05, |
|
"loss": 0.243, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.2021957913998174e-05, |
|
"loss": 0.2695, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.247941445562672e-05, |
|
"loss": 0.252, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.293687099725527e-05, |
|
"loss": 0.2631, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.3394327538883805e-05, |
|
"loss": 0.2919, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.3851784080512355e-05, |
|
"loss": 0.2688, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.43092406221409e-05, |
|
"loss": 0.2855, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.476669716376945e-05, |
|
"loss": 0.2782, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.522415370539799e-05, |
|
"loss": 0.2534, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.5681610247026535e-05, |
|
"loss": 0.2005, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.613906678865508e-05, |
|
"loss": 0.2167, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.659652333028363e-05, |
|
"loss": 0.289, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.705397987191217e-05, |
|
"loss": 0.2775, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.7511436413540716e-05, |
|
"loss": 0.2553, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.796889295516926e-05, |
|
"loss": 0.2529, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.842634949679781e-05, |
|
"loss": 0.2649, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.888380603842635e-05, |
|
"loss": 0.2394, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.9341262580054896e-05, |
|
"loss": 0.2067, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.979871912168344e-05, |
|
"loss": 0.2748, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 4.025617566331198e-05, |
|
"loss": 0.2251, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 4.071363220494053e-05, |
|
"loss": 0.2436, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 4.1171088746569077e-05, |
|
"loss": 0.2296, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 4.162854528819763e-05, |
|
"loss": 0.2791, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 4.2086001829826163e-05, |
|
"loss": 0.2627, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 4.2543458371454714e-05, |
|
"loss": 0.2042, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.300091491308326e-05, |
|
"loss": 0.2193, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.345837145471181e-05, |
|
"loss": 0.2399, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 4.391582799634035e-05, |
|
"loss": 0.2554, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 4.4373284537968894e-05, |
|
"loss": 0.2245, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 4.483074107959744e-05, |
|
"loss": 0.2211, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.528819762122599e-05, |
|
"loss": 0.2306, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.574565416285453e-05, |
|
"loss": 0.2271, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 4.6203110704483074e-05, |
|
"loss": 0.3005, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 4.666056724611162e-05, |
|
"loss": 0.248, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 4.711802378774017e-05, |
|
"loss": 0.2582, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 4.757548032936871e-05, |
|
"loss": 0.2792, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 4.8032936870997255e-05, |
|
"loss": 0.2444, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.84903934126258e-05, |
|
"loss": 0.2219, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 4.894784995425435e-05, |
|
"loss": 0.2214, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 4.940530649588289e-05, |
|
"loss": 0.2059, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 4.986276303751144e-05, |
|
"loss": 0.2164, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.9444333333333333, |
|
"eval_loss": 0.14764423668384552, |
|
"eval_runtime": 103.8431, |
|
"eval_samples_per_second": 288.897, |
|
"eval_steps_per_second": 18.056, |
|
"step": 1093 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 4.996442004676223e-05, |
|
"loss": 0.2307, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 4.9913591542136836e-05, |
|
"loss": 0.1882, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 4.986276303751144e-05, |
|
"loss": 0.2297, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 4.981193453288605e-05, |
|
"loss": 0.2255, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 4.976110602826065e-05, |
|
"loss": 0.2034, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 4.9710277523635254e-05, |
|
"loss": 0.2139, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 4.965944901900986e-05, |
|
"loss": 0.1942, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 4.960862051438447e-05, |
|
"loss": 0.2598, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 4.955779200975907e-05, |
|
"loss": 0.2086, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 4.950696350513368e-05, |
|
"loss": 0.2201, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 4.9456135000508286e-05, |
|
"loss": 0.1858, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 4.940530649588289e-05, |
|
"loss": 0.2062, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 4.93544779912575e-05, |
|
"loss": 0.213, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 4.9303649486632104e-05, |
|
"loss": 0.2231, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 4.925282098200672e-05, |
|
"loss": 0.2187, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 4.920199247738132e-05, |
|
"loss": 0.1906, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 4.915116397275592e-05, |
|
"loss": 0.2123, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 4.910033546813053e-05, |
|
"loss": 0.2157, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 4.9049506963505136e-05, |
|
"loss": 0.2075, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 4.899867845887974e-05, |
|
"loss": 0.2215, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 4.894784995425435e-05, |
|
"loss": 0.2208, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 4.8897021449628955e-05, |
|
"loss": 0.25, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 4.884619294500356e-05, |
|
"loss": 0.1994, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 4.879536444037817e-05, |
|
"loss": 0.2184, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 4.8744535935752774e-05, |
|
"loss": 0.2235, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 4.869370743112738e-05, |
|
"loss": 0.1856, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 4.8642878926501986e-05, |
|
"loss": 0.1842, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 4.859205042187659e-05, |
|
"loss": 0.2039, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 4.854122191725119e-05, |
|
"loss": 0.196, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 4.84903934126258e-05, |
|
"loss": 0.23, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 4.8439564908000405e-05, |
|
"loss": 0.2282, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 4.838873640337501e-05, |
|
"loss": 0.2078, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 4.8337907898749624e-05, |
|
"loss": 0.2308, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 4.828707939412423e-05, |
|
"loss": 0.226, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 4.8236250889498837e-05, |
|
"loss": 0.1782, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 4.818542238487344e-05, |
|
"loss": 0.1885, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 4.813459388024805e-05, |
|
"loss": 0.1993, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 4.8083765375622655e-05, |
|
"loss": 0.1759, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 4.8032936870997255e-05, |
|
"loss": 0.1748, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 4.798210836637186e-05, |
|
"loss": 0.1764, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 4.793127986174647e-05, |
|
"loss": 0.2233, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 4.7880451357121074e-05, |
|
"loss": 0.1792, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 4.782962285249568e-05, |
|
"loss": 0.1693, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 4.7778794347870286e-05, |
|
"loss": 0.162, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 4.772796584324489e-05, |
|
"loss": 0.1932, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 4.76771373386195e-05, |
|
"loss": 0.2128, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 4.7626308833994105e-05, |
|
"loss": 0.22, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 4.757548032936871e-05, |
|
"loss": 0.2337, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 4.752465182474332e-05, |
|
"loss": 0.185, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 4.7473823320117924e-05, |
|
"loss": 0.1849, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 4.742299481549253e-05, |
|
"loss": 0.1653, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 4.737216631086714e-05, |
|
"loss": 0.1747, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 4.732133780624174e-05, |
|
"loss": 0.1859, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 4.727050930161635e-05, |
|
"loss": 0.2241, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 4.7219680796990955e-05, |
|
"loss": 0.2332, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 4.716885229236556e-05, |
|
"loss": 0.2009, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 4.711802378774017e-05, |
|
"loss": 0.1926, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 4.7067195283114774e-05, |
|
"loss": 0.2122, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 4.701636677848938e-05, |
|
"loss": 0.2322, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.696553827386399e-05, |
|
"loss": 0.1813, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.691470976923859e-05, |
|
"loss": 0.1635, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 4.68638812646132e-05, |
|
"loss": 0.202, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.68130527599878e-05, |
|
"loss": 0.2074, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 4.6762224255362405e-05, |
|
"loss": 0.1835, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 4.671139575073701e-05, |
|
"loss": 0.1786, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 4.666056724611162e-05, |
|
"loss": 0.1985, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 4.6609738741486224e-05, |
|
"loss": 0.1782, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 4.655891023686084e-05, |
|
"loss": 0.1976, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 4.6508081732235444e-05, |
|
"loss": 0.1964, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 4.645725322761005e-05, |
|
"loss": 0.1991, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 4.6406424722984656e-05, |
|
"loss": 0.1826, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 4.635559621835926e-05, |
|
"loss": 0.2241, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 4.630476771373387e-05, |
|
"loss": 0.2132, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 4.625393920910847e-05, |
|
"loss": 0.184, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.6203110704483074e-05, |
|
"loss": 0.169, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 4.615228219985768e-05, |
|
"loss": 0.194, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.610145369523229e-05, |
|
"loss": 0.1915, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.605062519060689e-05, |
|
"loss": 0.2084, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 4.59997966859815e-05, |
|
"loss": 0.2079, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 4.5948968181356106e-05, |
|
"loss": 0.1748, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 4.589813967673071e-05, |
|
"loss": 0.1503, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 4.584731117210532e-05, |
|
"loss": 0.1674, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 4.5796482667479925e-05, |
|
"loss": 0.1609, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 4.574565416285453e-05, |
|
"loss": 0.1921, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 4.569482565822914e-05, |
|
"loss": 0.206, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 4.5643997153603744e-05, |
|
"loss": 0.1457, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 4.559316864897835e-05, |
|
"loss": 0.1794, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 4.5542340144352956e-05, |
|
"loss": 0.2206, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 4.549151163972756e-05, |
|
"loss": 0.1735, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 4.544068313510217e-05, |
|
"loss": 0.1473, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 4.5389854630476775e-05, |
|
"loss": 0.1444, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 4.533902612585138e-05, |
|
"loss": 0.1372, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 4.528819762122599e-05, |
|
"loss": 0.1942, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 4.5237369116600594e-05, |
|
"loss": 0.1598, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 4.51865406119752e-05, |
|
"loss": 0.1636, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 4.5135712107349806e-05, |
|
"loss": 0.1917, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 4.5084883602724406e-05, |
|
"loss": 0.1642, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 4.503405509809901e-05, |
|
"loss": 0.1566, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 4.498322659347362e-05, |
|
"loss": 0.1994, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 4.4932398088848225e-05, |
|
"loss": 0.1809, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 4.488156958422283e-05, |
|
"loss": 0.1595, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 4.483074107959744e-05, |
|
"loss": 0.1663, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 4.4779912574972044e-05, |
|
"loss": 0.1879, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 4.472908407034666e-05, |
|
"loss": 0.1641, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 4.467825556572126e-05, |
|
"loss": 0.181, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 4.462742706109587e-05, |
|
"loss": 0.1705, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 4.4576598556470476e-05, |
|
"loss": 0.1792, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 4.4525770051845075e-05, |
|
"loss": 0.1623, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 4.447494154721968e-05, |
|
"loss": 0.1647, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.9672666666666667, |
|
"eval_loss": 0.0872458964586258, |
|
"eval_runtime": 111.6471, |
|
"eval_samples_per_second": 268.704, |
|
"eval_steps_per_second": 16.794, |
|
"step": 2187 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 4.442411304259429e-05, |
|
"loss": 0.1572, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 4.4373284537968894e-05, |
|
"loss": 0.1677, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 4.43224560333435e-05, |
|
"loss": 0.1581, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 4.4271627528718107e-05, |
|
"loss": 0.2149, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 4.422079902409271e-05, |
|
"loss": 0.186, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 4.416997051946732e-05, |
|
"loss": 0.1687, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 4.4119142014841925e-05, |
|
"loss": 0.1698, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 4.406831351021653e-05, |
|
"loss": 0.1663, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 4.401748500559114e-05, |
|
"loss": 0.1742, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 4.3966656500965744e-05, |
|
"loss": 0.1841, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 4.391582799634035e-05, |
|
"loss": 0.1742, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 4.386499949171495e-05, |
|
"loss": 0.1699, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 4.381417098708956e-05, |
|
"loss": 0.1797, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 4.376334248246417e-05, |
|
"loss": 0.1928, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 4.3712513977838776e-05, |
|
"loss": 0.1441, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 4.366168547321338e-05, |
|
"loss": 0.2079, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 4.361085696858799e-05, |
|
"loss": 0.1892, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 4.3560028463962595e-05, |
|
"loss": 0.164, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 4.35091999593372e-05, |
|
"loss": 0.1716, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 4.345837145471181e-05, |
|
"loss": 0.1589, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 4.3407542950086413e-05, |
|
"loss": 0.1585, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 4.335671444546101e-05, |
|
"loss": 0.1476, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 4.330588594083562e-05, |
|
"loss": 0.1224, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 4.3255057436210226e-05, |
|
"loss": 0.1368, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 4.320422893158483e-05, |
|
"loss": 0.1522, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 4.315340042695944e-05, |
|
"loss": 0.14, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 4.3102571922334044e-05, |
|
"loss": 0.1476, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 4.305174341770865e-05, |
|
"loss": 0.1646, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 4.300091491308326e-05, |
|
"loss": 0.1685, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 4.295008640845786e-05, |
|
"loss": 0.1966, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 4.2899257903832476e-05, |
|
"loss": 0.1786, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 4.284842939920708e-05, |
|
"loss": 0.1237, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 4.279760089458168e-05, |
|
"loss": 0.209, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 4.274677238995629e-05, |
|
"loss": 0.1838, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 4.2695943885330895e-05, |
|
"loss": 0.155, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 4.26451153807055e-05, |
|
"loss": 0.1306, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 4.259428687608011e-05, |
|
"loss": 0.147, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 4.2543458371454714e-05, |
|
"loss": 0.1762, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 4.249262986682932e-05, |
|
"loss": 0.1485, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.2441801362203926e-05, |
|
"loss": 0.2248, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 4.239097285757853e-05, |
|
"loss": 0.1692, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 4.234014435295314e-05, |
|
"loss": 0.2101, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 4.2289315848327745e-05, |
|
"loss": 0.1871, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 4.223848734370235e-05, |
|
"loss": 0.2135, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 4.218765883907696e-05, |
|
"loss": 0.1416, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 4.213683033445156e-05, |
|
"loss": 0.1609, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 4.2086001829826163e-05, |
|
"loss": 0.147, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 4.2035173325200776e-05, |
|
"loss": 0.1684, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 4.198434482057538e-05, |
|
"loss": 0.1683, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 4.193351631594999e-05, |
|
"loss": 0.1958, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 4.1882687811324595e-05, |
|
"loss": 0.163, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 4.18318593066992e-05, |
|
"loss": 0.1654, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 4.178103080207381e-05, |
|
"loss": 0.1906, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 4.1730202297448414e-05, |
|
"loss": 0.1611, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 4.167937379282302e-05, |
|
"loss": 0.199, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 4.162854528819763e-05, |
|
"loss": 0.1649, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 4.1577716783572226e-05, |
|
"loss": 0.1777, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 4.152688827894683e-05, |
|
"loss": 0.2043, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 4.147605977432144e-05, |
|
"loss": 0.1599, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 4.1425231269696045e-05, |
|
"loss": 0.129, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 4.137440276507065e-05, |
|
"loss": 0.1764, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 4.132357426044526e-05, |
|
"loss": 0.1519, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 4.1272745755819864e-05, |
|
"loss": 0.1629, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 4.122191725119447e-05, |
|
"loss": 0.1469, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 4.1171088746569077e-05, |
|
"loss": 0.1732, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 4.112026024194369e-05, |
|
"loss": 0.124, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 4.106943173731829e-05, |
|
"loss": 0.1505, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 4.1018603232692895e-05, |
|
"loss": 0.1848, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 4.09677747280675e-05, |
|
"loss": 0.1194, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 4.091694622344211e-05, |
|
"loss": 0.1401, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 4.0866117718816714e-05, |
|
"loss": 0.1875, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 4.081528921419132e-05, |
|
"loss": 0.1982, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 4.076446070956593e-05, |
|
"loss": 0.1719, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 4.071363220494053e-05, |
|
"loss": 0.1837, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 4.066280370031514e-05, |
|
"loss": 0.2217, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 4.0611975195689746e-05, |
|
"loss": 0.1561, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 4.056114669106435e-05, |
|
"loss": 0.1587, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 4.051031818643896e-05, |
|
"loss": 0.167, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.0459489681813565e-05, |
|
"loss": 0.1372, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.0408661177188164e-05, |
|
"loss": 0.1302, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.035783267256277e-05, |
|
"loss": 0.1464, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 4.030700416793738e-05, |
|
"loss": 0.1584, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 4.025617566331198e-05, |
|
"loss": 0.1505, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 4.0205347158686596e-05, |
|
"loss": 0.1454, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 4.01545186540612e-05, |
|
"loss": 0.145, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 4.010369014943581e-05, |
|
"loss": 0.1717, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 4.0052861644810415e-05, |
|
"loss": 0.1565, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 4.000203314018502e-05, |
|
"loss": 0.1575, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 3.995120463555963e-05, |
|
"loss": 0.1593, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 3.9900376130934234e-05, |
|
"loss": 0.1687, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 3.984954762630883e-05, |
|
"loss": 0.1476, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 3.979871912168344e-05, |
|
"loss": 0.1763, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 3.9747890617058046e-05, |
|
"loss": 0.1234, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 3.969706211243265e-05, |
|
"loss": 0.1396, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 3.964623360780726e-05, |
|
"loss": 0.203, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 3.9595405103181865e-05, |
|
"loss": 0.1515, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 3.954457659855647e-05, |
|
"loss": 0.147, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 3.949374809393108e-05, |
|
"loss": 0.156, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 3.9442919589305684e-05, |
|
"loss": 0.1416, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 3.939209108468029e-05, |
|
"loss": 0.1578, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 3.9341262580054896e-05, |
|
"loss": 0.2017, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 3.92904340754295e-05, |
|
"loss": 0.1584, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 3.923960557080411e-05, |
|
"loss": 0.1858, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 3.9188777066178715e-05, |
|
"loss": 0.1603, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 3.913794856155332e-05, |
|
"loss": 0.1672, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 3.908712005692793e-05, |
|
"loss": 0.1342, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 3.9036291552302534e-05, |
|
"loss": 0.1689, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 3.898546304767714e-05, |
|
"loss": 0.1509, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 3.8934634543051746e-05, |
|
"loss": 0.1455, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 3.888380603842635e-05, |
|
"loss": 0.1784, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.9708666666666667, |
|
"eval_loss": 0.07605975866317749, |
|
"eval_runtime": 102.048, |
|
"eval_samples_per_second": 293.979, |
|
"eval_steps_per_second": 18.374, |
|
"step": 3281 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 3.883297753380096e-05, |
|
"loss": 0.1603, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 3.8782149029175565e-05, |
|
"loss": 0.1591, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 3.873132052455017e-05, |
|
"loss": 0.1236, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 3.868049201992477e-05, |
|
"loss": 0.1304, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 3.862966351529938e-05, |
|
"loss": 0.1566, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 3.8578835010673984e-05, |
|
"loss": 0.1745, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 3.852800650604859e-05, |
|
"loss": 0.1399, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 3.8477178001423196e-05, |
|
"loss": 0.1467, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 3.842634949679781e-05, |
|
"loss": 0.1402, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 3.8375520992172416e-05, |
|
"loss": 0.1356, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 3.832469248754702e-05, |
|
"loss": 0.1732, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 3.827386398292163e-05, |
|
"loss": 0.1344, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 3.8223035478296234e-05, |
|
"loss": 0.1193, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 3.817220697367084e-05, |
|
"loss": 0.1673, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 3.812137846904544e-05, |
|
"loss": 0.12, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 3.8070549964420047e-05, |
|
"loss": 0.1433, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 3.801972145979465e-05, |
|
"loss": 0.17, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 3.796889295516926e-05, |
|
"loss": 0.1584, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 3.7918064450543865e-05, |
|
"loss": 0.1456, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 3.786723594591847e-05, |
|
"loss": 0.1574, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 3.781640744129308e-05, |
|
"loss": 0.1577, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 3.7765578936667684e-05, |
|
"loss": 0.1771, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 3.771475043204229e-05, |
|
"loss": 0.127, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 3.76639219274169e-05, |
|
"loss": 0.1679, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 3.76130934227915e-05, |
|
"loss": 0.1194, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 3.756226491816611e-05, |
|
"loss": 0.1338, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 3.7511436413540716e-05, |
|
"loss": 0.1326, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 3.746060790891532e-05, |
|
"loss": 0.1309, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 3.740977940428993e-05, |
|
"loss": 0.1464, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 3.7358950899664535e-05, |
|
"loss": 0.1019, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 3.730812239503914e-05, |
|
"loss": 0.1395, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 3.725729389041375e-05, |
|
"loss": 0.1495, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 3.7206465385788353e-05, |
|
"loss": 0.1398, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 3.715563688116296e-05, |
|
"loss": 0.1291, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 3.7104808376537566e-05, |
|
"loss": 0.1487, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 3.705397987191217e-05, |
|
"loss": 0.1274, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 3.700315136728678e-05, |
|
"loss": 0.1288, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 3.6952322862661385e-05, |
|
"loss": 0.1178, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 3.6901494358035984e-05, |
|
"loss": 0.1471, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 3.685066585341059e-05, |
|
"loss": 0.1444, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 3.67998373487852e-05, |
|
"loss": 0.1251, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 3.67490088441598e-05, |
|
"loss": 0.1185, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 3.669818033953441e-05, |
|
"loss": 0.135, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 3.6647351834909016e-05, |
|
"loss": 0.1694, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 3.659652333028363e-05, |
|
"loss": 0.1581, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 3.6545694825658235e-05, |
|
"loss": 0.1113, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 3.649486632103284e-05, |
|
"loss": 0.1336, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 3.644403781640745e-05, |
|
"loss": 0.1459, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 3.639320931178205e-05, |
|
"loss": 0.1836, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 3.6342380807156654e-05, |
|
"loss": 0.1551, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 3.629155230253126e-05, |
|
"loss": 0.1296, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 3.6240723797905866e-05, |
|
"loss": 0.1584, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 3.618989529328047e-05, |
|
"loss": 0.1561, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 3.613906678865508e-05, |
|
"loss": 0.1359, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 3.6088238284029685e-05, |
|
"loss": 0.1685, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 3.603740977940429e-05, |
|
"loss": 0.1339, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 3.59865812747789e-05, |
|
"loss": 0.1385, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 3.5935752770153504e-05, |
|
"loss": 0.1122, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 3.588492426552811e-05, |
|
"loss": 0.1291, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 3.5834095760902716e-05, |
|
"loss": 0.1137, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 3.578326725627732e-05, |
|
"loss": 0.1466, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 3.573243875165192e-05, |
|
"loss": 0.1007, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 3.5681610247026535e-05, |
|
"loss": 0.1285, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"learning_rate": 3.563078174240114e-05, |
|
"loss": 0.0934, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 3.557995323777575e-05, |
|
"loss": 0.1168, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 3.5529124733150354e-05, |
|
"loss": 0.1241, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 3.547829622852496e-05, |
|
"loss": 0.1488, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 3.542746772389957e-05, |
|
"loss": 0.1595, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 3.537663921927417e-05, |
|
"loss": 0.1184, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 3.532581071464878e-05, |
|
"loss": 0.1222, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 3.5274982210023386e-05, |
|
"loss": 0.1684, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 3.522415370539799e-05, |
|
"loss": 0.1684, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 3.517332520077259e-05, |
|
"loss": 0.1223, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 3.51224966961472e-05, |
|
"loss": 0.1191, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 3.5071668191521804e-05, |
|
"loss": 0.1456, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 3.502083968689641e-05, |
|
"loss": 0.1518, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 3.4970011182271017e-05, |
|
"loss": 0.12, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 3.491918267764562e-05, |
|
"loss": 0.1362, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"learning_rate": 3.486835417302023e-05, |
|
"loss": 0.1388, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 3.4817525668394835e-05, |
|
"loss": 0.1208, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"learning_rate": 3.476669716376945e-05, |
|
"loss": 0.1378, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 3.4715868659144055e-05, |
|
"loss": 0.1435, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"learning_rate": 3.4665040154518654e-05, |
|
"loss": 0.1161, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 3.461421164989326e-05, |
|
"loss": 0.1244, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 3.456338314526787e-05, |
|
"loss": 0.1258, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 3.451255464064247e-05, |
|
"loss": 0.1365, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 3.446172613601708e-05, |
|
"loss": 0.123, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 3.4410897631391686e-05, |
|
"loss": 0.163, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 3.436006912676629e-05, |
|
"loss": 0.0955, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 3.43092406221409e-05, |
|
"loss": 0.1748, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 3.4258412117515505e-05, |
|
"loss": 0.1224, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 3.420758361289011e-05, |
|
"loss": 0.1374, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 3.415675510826472e-05, |
|
"loss": 0.1356, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 3.4105926603639323e-05, |
|
"loss": 0.1348, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"learning_rate": 3.405509809901393e-05, |
|
"loss": 0.0868, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 3.400426959438853e-05, |
|
"loss": 0.1597, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 3.3953441089763136e-05, |
|
"loss": 0.1374, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 3.390261258513775e-05, |
|
"loss": 0.1277, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 3.3851784080512355e-05, |
|
"loss": 0.1403, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 3.380095557588696e-05, |
|
"loss": 0.1342, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"learning_rate": 3.375012707126157e-05, |
|
"loss": 0.1304, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 3.3699298566636174e-05, |
|
"loss": 0.0999, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 3.364847006201078e-05, |
|
"loss": 0.1232, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 3.3597641557385386e-05, |
|
"loss": 0.1745, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 3.354681305275999e-05, |
|
"loss": 0.1354, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 3.34959845481346e-05, |
|
"loss": 0.1221, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"learning_rate": 3.34451560435092e-05, |
|
"loss": 0.1222, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 3.3394327538883805e-05, |
|
"loss": 0.1411, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 3.334349903425841e-05, |
|
"loss": 0.1247, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.9666, |
|
"eval_loss": 0.08694037795066833, |
|
"eval_runtime": 102.1698, |
|
"eval_samples_per_second": 293.629, |
|
"eval_steps_per_second": 18.352, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 3.329267052963302e-05, |
|
"loss": 0.1312, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 3.3241842025007624e-05, |
|
"loss": 0.1268, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 3.319101352038223e-05, |
|
"loss": 0.1054, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 3.3140185015756836e-05, |
|
"loss": 0.1293, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 3.308935651113144e-05, |
|
"loss": 0.1052, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"learning_rate": 3.303852800650605e-05, |
|
"loss": 0.1292, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"learning_rate": 3.298769950188066e-05, |
|
"loss": 0.1292, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 3.293687099725527e-05, |
|
"loss": 0.1014, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 4.08, |
|
"learning_rate": 3.288604249262987e-05, |
|
"loss": 0.1177, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"learning_rate": 3.2835213988004474e-05, |
|
"loss": 0.0986, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 3.278438548337908e-05, |
|
"loss": 0.1245, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"learning_rate": 3.2733556978753686e-05, |
|
"loss": 0.1129, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"learning_rate": 3.268272847412829e-05, |
|
"loss": 0.1276, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 4.12, |
|
"learning_rate": 3.26318999695029e-05, |
|
"loss": 0.1214, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"learning_rate": 3.2581071464877505e-05, |
|
"loss": 0.1215, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 3.253024296025211e-05, |
|
"loss": 0.1237, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"learning_rate": 3.247941445562672e-05, |
|
"loss": 0.1325, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"learning_rate": 3.2428585951001324e-05, |
|
"loss": 0.1056, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 3.237775744637593e-05, |
|
"loss": 0.1106, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 4.18, |
|
"learning_rate": 3.232692894175054e-05, |
|
"loss": 0.1024, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 3.227610043712514e-05, |
|
"loss": 0.1329, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"learning_rate": 3.222527193249974e-05, |
|
"loss": 0.1181, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 3.217444342787435e-05, |
|
"loss": 0.1267, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 3.2123614923248955e-05, |
|
"loss": 0.1586, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 3.207278641862357e-05, |
|
"loss": 0.1158, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 4.23, |
|
"learning_rate": 3.2021957913998174e-05, |
|
"loss": 0.1376, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"learning_rate": 3.197112940937278e-05, |
|
"loss": 0.1081, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 3.192030090474739e-05, |
|
"loss": 0.1024, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 4.26, |
|
"learning_rate": 3.186947240012199e-05, |
|
"loss": 0.1139, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"learning_rate": 3.18186438954966e-05, |
|
"loss": 0.1408, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 3.1767815390871206e-05, |
|
"loss": 0.1218, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 4.29, |
|
"learning_rate": 3.1716986886245805e-05, |
|
"loss": 0.1342, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 3.166615838162041e-05, |
|
"loss": 0.1161, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 3.161532987699502e-05, |
|
"loss": 0.1305, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 4.32, |
|
"learning_rate": 3.1564501372369624e-05, |
|
"loss": 0.1199, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 4.32, |
|
"learning_rate": 3.151367286774423e-05, |
|
"loss": 0.1249, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 3.146284436311884e-05, |
|
"loss": 0.138, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 3.141201585849344e-05, |
|
"loss": 0.1424, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 4.35, |
|
"learning_rate": 3.136118735386805e-05, |
|
"loss": 0.1289, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 3.1310358849242656e-05, |
|
"loss": 0.096, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 3.125953034461726e-05, |
|
"loss": 0.1028, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 4.38, |
|
"learning_rate": 3.120870183999187e-05, |
|
"loss": 0.1208, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"learning_rate": 3.1157873335366475e-05, |
|
"loss": 0.1472, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 3.110704483074108e-05, |
|
"loss": 0.1191, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 3.105621632611569e-05, |
|
"loss": 0.1565, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"learning_rate": 3.1005387821490293e-05, |
|
"loss": 0.157, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 4.43, |
|
"learning_rate": 3.09545593168649e-05, |
|
"loss": 0.1375, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 4.43, |
|
"learning_rate": 3.0903730812239506e-05, |
|
"loss": 0.118, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 3.085290230761411e-05, |
|
"loss": 0.1108, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 4.45, |
|
"learning_rate": 3.080207380298872e-05, |
|
"loss": 0.1257, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 3.0751245298363325e-05, |
|
"loss": 0.1223, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 3.070041679373793e-05, |
|
"loss": 0.1116, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 4.48, |
|
"learning_rate": 3.064958828911254e-05, |
|
"loss": 0.1285, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"learning_rate": 3.0598759784487144e-05, |
|
"loss": 0.1289, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"learning_rate": 3.054793127986175e-05, |
|
"loss": 0.1461, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"learning_rate": 3.0497102775236353e-05, |
|
"loss": 0.1155, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 3.044627427061096e-05, |
|
"loss": 0.1066, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 3.0395445765985565e-05, |
|
"loss": 0.1093, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 3.034461726136017e-05, |
|
"loss": 0.1343, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 3.0293788756734775e-05, |
|
"loss": 0.0961, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"learning_rate": 3.0242960252109388e-05, |
|
"loss": 0.127, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 4.56, |
|
"learning_rate": 3.0192131747483994e-05, |
|
"loss": 0.0846, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 3.0141303242858597e-05, |
|
"loss": 0.1538, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"learning_rate": 3.0090474738233203e-05, |
|
"loss": 0.1338, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"learning_rate": 3.003964623360781e-05, |
|
"loss": 0.123, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 2.9988817728982416e-05, |
|
"loss": 0.1192, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 2.9937989224357022e-05, |
|
"loss": 0.1462, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"learning_rate": 2.988716071973163e-05, |
|
"loss": 0.0944, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 4.63, |
|
"learning_rate": 2.9836332215106235e-05, |
|
"loss": 0.1048, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 2.9785503710480838e-05, |
|
"loss": 0.1334, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 2.9734675205855444e-05, |
|
"loss": 0.1048, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"learning_rate": 2.968384670123005e-05, |
|
"loss": 0.1562, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 2.9633018196604656e-05, |
|
"loss": 0.076, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"learning_rate": 2.9582189691979263e-05, |
|
"loss": 0.1313, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 2.953136118735387e-05, |
|
"loss": 0.1707, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 4.69, |
|
"learning_rate": 2.9480532682728472e-05, |
|
"loss": 0.0979, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"learning_rate": 2.9429704178103078e-05, |
|
"loss": 0.1088, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 4.71, |
|
"learning_rate": 2.937887567347769e-05, |
|
"loss": 0.1176, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 2.9328047168852298e-05, |
|
"loss": 0.1372, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 2.92772186642269e-05, |
|
"loss": 0.1035, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"learning_rate": 2.9226390159601507e-05, |
|
"loss": 0.1355, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 2.9175561654976113e-05, |
|
"loss": 0.1396, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 2.912473315035072e-05, |
|
"loss": 0.125, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 4.76, |
|
"learning_rate": 2.9073904645725326e-05, |
|
"loss": 0.1185, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"learning_rate": 2.9023076141099932e-05, |
|
"loss": 0.1066, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 4.78, |
|
"learning_rate": 2.8972247636474538e-05, |
|
"loss": 0.0853, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 2.892141913184914e-05, |
|
"loss": 0.1374, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"learning_rate": 2.8870590627223747e-05, |
|
"loss": 0.1178, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 2.8819762122598354e-05, |
|
"loss": 0.0968, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"learning_rate": 2.876893361797296e-05, |
|
"loss": 0.1314, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 4.83, |
|
"learning_rate": 2.8718105113347566e-05, |
|
"loss": 0.116, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"learning_rate": 2.8667276608722172e-05, |
|
"loss": 0.1184, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 4.85, |
|
"learning_rate": 2.8616448104096775e-05, |
|
"loss": 0.1644, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 4.85, |
|
"learning_rate": 2.856561959947138e-05, |
|
"loss": 0.146, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 2.8514791094845988e-05, |
|
"loss": 0.1018, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 2.84639625902206e-05, |
|
"loss": 0.1751, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 4.88, |
|
"learning_rate": 2.8413134085595204e-05, |
|
"loss": 0.1067, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 2.836230558096981e-05, |
|
"loss": 0.12, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"learning_rate": 2.8311477076344416e-05, |
|
"loss": 0.1146, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"learning_rate": 2.8260648571719023e-05, |
|
"loss": 0.1168, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 4.92, |
|
"learning_rate": 2.820982006709363e-05, |
|
"loss": 0.0926, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 2.8158991562468235e-05, |
|
"loss": 0.1016, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 4.94, |
|
"learning_rate": 2.810816305784284e-05, |
|
"loss": 0.1052, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 2.8057334553217445e-05, |
|
"loss": 0.1434, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 2.800650604859205e-05, |
|
"loss": 0.1074, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 2.7955677543966657e-05, |
|
"loss": 0.1112, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 4.97, |
|
"learning_rate": 2.7904849039341263e-05, |
|
"loss": 0.1219, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"learning_rate": 2.785402053471587e-05, |
|
"loss": 0.1132, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"learning_rate": 2.7803192030090476e-05, |
|
"loss": 0.0994, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_accuracy": 0.9695666666666667, |
|
"eval_loss": 0.08079633861780167, |
|
"eval_runtime": 101.8886, |
|
"eval_samples_per_second": 294.439, |
|
"eval_steps_per_second": 18.402, |
|
"step": 5468 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 2.775236352546508e-05, |
|
"loss": 0.1391, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 2.7701535020839685e-05, |
|
"loss": 0.1461, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 2.765070651621429e-05, |
|
"loss": 0.1145, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 2.7599878011588898e-05, |
|
"loss": 0.1075, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 2.7549049506963507e-05, |
|
"loss": 0.1242, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"learning_rate": 2.7498221002338114e-05, |
|
"loss": 0.1185, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 5.06, |
|
"learning_rate": 2.744739249771272e-05, |
|
"loss": 0.1264, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"learning_rate": 2.7396563993087326e-05, |
|
"loss": 0.1225, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"learning_rate": 2.7345735488461933e-05, |
|
"loss": 0.1093, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 5.08, |
|
"learning_rate": 2.729490698383654e-05, |
|
"loss": 0.1401, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 5.09, |
|
"learning_rate": 2.7244078479211145e-05, |
|
"loss": 0.14, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 5.1, |
|
"learning_rate": 2.7193249974585748e-05, |
|
"loss": 0.1029, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 5.11, |
|
"learning_rate": 2.7142421469960354e-05, |
|
"loss": 0.1439, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 5.12, |
|
"learning_rate": 2.709159296533496e-05, |
|
"loss": 0.1313, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 5.13, |
|
"learning_rate": 2.7040764460709567e-05, |
|
"loss": 0.1124, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 2.6989935956084173e-05, |
|
"loss": 0.1053, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 5.15, |
|
"learning_rate": 2.693910745145878e-05, |
|
"loss": 0.1356, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 5.16, |
|
"learning_rate": 2.6888278946833382e-05, |
|
"loss": 0.0995, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 5.17, |
|
"learning_rate": 2.683745044220799e-05, |
|
"loss": 0.1129, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 5.17, |
|
"learning_rate": 2.6786621937582595e-05, |
|
"loss": 0.1015, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 5.18, |
|
"learning_rate": 2.67357934329572e-05, |
|
"loss": 0.1189, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 5.19, |
|
"learning_rate": 2.6684964928331808e-05, |
|
"loss": 0.1295, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 5.2, |
|
"learning_rate": 2.6634136423706417e-05, |
|
"loss": 0.0881, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"learning_rate": 2.6583307919081023e-05, |
|
"loss": 0.1108, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 5.22, |
|
"learning_rate": 2.653247941445563e-05, |
|
"loss": 0.0811, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 5.23, |
|
"learning_rate": 2.6481650909830236e-05, |
|
"loss": 0.1191, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 5.24, |
|
"learning_rate": 2.6430822405204842e-05, |
|
"loss": 0.1216, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 5.25, |
|
"learning_rate": 2.637999390057945e-05, |
|
"loss": 0.1006, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 5.26, |
|
"learning_rate": 2.632916539595405e-05, |
|
"loss": 0.0943, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 5.27, |
|
"learning_rate": 2.6278336891328658e-05, |
|
"loss": 0.123, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 5.28, |
|
"learning_rate": 2.6227508386703264e-05, |
|
"loss": 0.0993, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 5.28, |
|
"learning_rate": 2.617667988207787e-05, |
|
"loss": 0.1109, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 5.29, |
|
"learning_rate": 2.6125851377452477e-05, |
|
"loss": 0.0953, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 5.3, |
|
"learning_rate": 2.6075022872827083e-05, |
|
"loss": 0.1008, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 5.31, |
|
"learning_rate": 2.6024194368201686e-05, |
|
"loss": 0.1115, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 5.32, |
|
"learning_rate": 2.5973365863576292e-05, |
|
"loss": 0.0955, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 5.33, |
|
"learning_rate": 2.59225373589509e-05, |
|
"loss": 0.1287, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 5.34, |
|
"learning_rate": 2.5871708854325505e-05, |
|
"loss": 0.0831, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 5.35, |
|
"learning_rate": 2.582088034970011e-05, |
|
"loss": 0.1183, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 5.36, |
|
"learning_rate": 2.5770051845074717e-05, |
|
"loss": 0.1151, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 5.37, |
|
"learning_rate": 2.5719223340449327e-05, |
|
"loss": 0.1049, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 5.38, |
|
"learning_rate": 2.5668394835823933e-05, |
|
"loss": 0.0925, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 5.39, |
|
"learning_rate": 2.561756633119854e-05, |
|
"loss": 0.1157, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 5.39, |
|
"learning_rate": 2.5566737826573146e-05, |
|
"loss": 0.1019, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 5.4, |
|
"learning_rate": 2.5515909321947752e-05, |
|
"loss": 0.0825, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 5.41, |
|
"learning_rate": 2.5465080817322355e-05, |
|
"loss": 0.0889, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 5.42, |
|
"learning_rate": 2.541425231269696e-05, |
|
"loss": 0.136, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 5.43, |
|
"learning_rate": 2.5363423808071568e-05, |
|
"loss": 0.1255, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 5.44, |
|
"learning_rate": 2.5312595303446174e-05, |
|
"loss": 0.1396, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 5.45, |
|
"learning_rate": 2.526176679882078e-05, |
|
"loss": 0.1023, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 5.46, |
|
"learning_rate": 2.5210938294195386e-05, |
|
"loss": 0.0913, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 5.47, |
|
"learning_rate": 2.5160109789569993e-05, |
|
"loss": 0.0917, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 5.48, |
|
"learning_rate": 2.5109281284944596e-05, |
|
"loss": 0.0956, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 5.49, |
|
"learning_rate": 2.5058452780319202e-05, |
|
"loss": 0.1073, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 5.49, |
|
"learning_rate": 2.5007624275693808e-05, |
|
"loss": 0.105, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"learning_rate": 2.4956795771068418e-05, |
|
"loss": 0.1031, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 5.51, |
|
"learning_rate": 2.4905967266443024e-05, |
|
"loss": 0.0799, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 5.52, |
|
"learning_rate": 2.4855138761817627e-05, |
|
"loss": 0.115, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"learning_rate": 2.4804310257192233e-05, |
|
"loss": 0.0797, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 5.54, |
|
"learning_rate": 2.475348175256684e-05, |
|
"loss": 0.1361, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 5.55, |
|
"learning_rate": 2.4702653247941446e-05, |
|
"loss": 0.0957, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 5.56, |
|
"learning_rate": 2.4651824743316052e-05, |
|
"loss": 0.1263, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 5.57, |
|
"learning_rate": 2.460099623869066e-05, |
|
"loss": 0.1299, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 5.58, |
|
"learning_rate": 2.4550167734065265e-05, |
|
"loss": 0.11, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 5.59, |
|
"learning_rate": 2.449933922943987e-05, |
|
"loss": 0.1418, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 5.6, |
|
"learning_rate": 2.4448510724814477e-05, |
|
"loss": 0.1418, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 5.6, |
|
"learning_rate": 2.4397682220189084e-05, |
|
"loss": 0.0887, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 5.61, |
|
"learning_rate": 2.434685371556369e-05, |
|
"loss": 0.103, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 5.62, |
|
"learning_rate": 2.4296025210938296e-05, |
|
"loss": 0.0839, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 5.63, |
|
"learning_rate": 2.42451967063129e-05, |
|
"loss": 0.1393, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 5.64, |
|
"learning_rate": 2.4194368201687505e-05, |
|
"loss": 0.1092, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 5.65, |
|
"learning_rate": 2.4143539697062115e-05, |
|
"loss": 0.1245, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 5.66, |
|
"learning_rate": 2.409271119243672e-05, |
|
"loss": 0.1068, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 5.67, |
|
"learning_rate": 2.4041882687811328e-05, |
|
"loss": 0.1221, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 5.68, |
|
"learning_rate": 2.399105418318593e-05, |
|
"loss": 0.0857, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 2.3940225678560537e-05, |
|
"loss": 0.1215, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 5.7, |
|
"learning_rate": 2.3889397173935143e-05, |
|
"loss": 0.1101, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 5.71, |
|
"learning_rate": 2.383856866930975e-05, |
|
"loss": 0.1308, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 5.71, |
|
"learning_rate": 2.3787740164684356e-05, |
|
"loss": 0.0737, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 5.72, |
|
"learning_rate": 2.3736911660058962e-05, |
|
"loss": 0.1157, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 5.73, |
|
"learning_rate": 2.368608315543357e-05, |
|
"loss": 0.1328, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 5.74, |
|
"learning_rate": 2.3635254650808175e-05, |
|
"loss": 0.1183, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 5.75, |
|
"learning_rate": 2.358442614618278e-05, |
|
"loss": 0.104, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 5.76, |
|
"learning_rate": 2.3533597641557387e-05, |
|
"loss": 0.1231, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 5.77, |
|
"learning_rate": 2.3482769136931993e-05, |
|
"loss": 0.1211, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 5.78, |
|
"learning_rate": 2.34319406323066e-05, |
|
"loss": 0.1034, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 5.79, |
|
"learning_rate": 2.3381112127681203e-05, |
|
"loss": 0.1187, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"learning_rate": 2.333028362305581e-05, |
|
"loss": 0.124, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 5.81, |
|
"learning_rate": 2.327945511843042e-05, |
|
"loss": 0.0865, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 5.81, |
|
"learning_rate": 2.3228626613805025e-05, |
|
"loss": 0.1136, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 2.317779810917963e-05, |
|
"loss": 0.1108, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 5.83, |
|
"learning_rate": 2.3126969604554234e-05, |
|
"loss": 0.1211, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 5.84, |
|
"learning_rate": 2.307614109992884e-05, |
|
"loss": 0.0907, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 5.85, |
|
"learning_rate": 2.3025312595303447e-05, |
|
"loss": 0.0942, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 5.86, |
|
"learning_rate": 2.2974484090678053e-05, |
|
"loss": 0.138, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 5.87, |
|
"learning_rate": 2.292365558605266e-05, |
|
"loss": 0.1244, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 5.88, |
|
"learning_rate": 2.2872827081427266e-05, |
|
"loss": 0.1086, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 5.89, |
|
"learning_rate": 2.2821998576801872e-05, |
|
"loss": 0.0759, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 5.9, |
|
"learning_rate": 2.2771170072176478e-05, |
|
"loss": 0.082, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 5.91, |
|
"learning_rate": 2.2720341567551084e-05, |
|
"loss": 0.1167, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 5.92, |
|
"learning_rate": 2.266951306292569e-05, |
|
"loss": 0.0743, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 5.92, |
|
"learning_rate": 2.2618684558300297e-05, |
|
"loss": 0.1057, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 5.93, |
|
"learning_rate": 2.2567856053674903e-05, |
|
"loss": 0.1445, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 5.94, |
|
"learning_rate": 2.2517027549049506e-05, |
|
"loss": 0.13, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 5.95, |
|
"learning_rate": 2.2466199044424112e-05, |
|
"loss": 0.1032, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 5.96, |
|
"learning_rate": 2.241537053979872e-05, |
|
"loss": 0.1166, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 5.97, |
|
"learning_rate": 2.236454203517333e-05, |
|
"loss": 0.1184, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 5.98, |
|
"learning_rate": 2.2313713530547935e-05, |
|
"loss": 0.1078, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 5.99, |
|
"learning_rate": 2.2262885025922538e-05, |
|
"loss": 0.123, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 2.2212056521297144e-05, |
|
"loss": 0.073, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_accuracy": 0.9820333333333333, |
|
"eval_loss": 0.051426343619823456, |
|
"eval_runtime": 102.5724, |
|
"eval_samples_per_second": 292.476, |
|
"eval_steps_per_second": 18.28, |
|
"step": 6562 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 2.216122801667175e-05, |
|
"loss": 0.1078, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 2.2110399512046356e-05, |
|
"loss": 0.1048, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 2.2059571007420963e-05, |
|
"loss": 0.1445, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 2.200874250279557e-05, |
|
"loss": 0.0772, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 2.1957913998170175e-05, |
|
"loss": 0.1061, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 6.05, |
|
"learning_rate": 2.190708549354478e-05, |
|
"loss": 0.1112, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 6.06, |
|
"learning_rate": 2.1856256988919388e-05, |
|
"loss": 0.1076, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"learning_rate": 2.1805428484293994e-05, |
|
"loss": 0.0861, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 6.08, |
|
"learning_rate": 2.17545999796686e-05, |
|
"loss": 0.1148, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"learning_rate": 2.1703771475043207e-05, |
|
"loss": 0.0875, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 6.1, |
|
"learning_rate": 2.165294297041781e-05, |
|
"loss": 0.0723, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 6.11, |
|
"learning_rate": 2.1602114465792416e-05, |
|
"loss": 0.0933, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 6.12, |
|
"learning_rate": 2.1551285961167022e-05, |
|
"loss": 0.1057, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 6.13, |
|
"learning_rate": 2.150045745654163e-05, |
|
"loss": 0.0954, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 6.13, |
|
"learning_rate": 2.1449628951916238e-05, |
|
"loss": 0.1079, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 6.14, |
|
"learning_rate": 2.139880044729084e-05, |
|
"loss": 0.1028, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 6.15, |
|
"learning_rate": 2.1347971942665447e-05, |
|
"loss": 0.1368, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 6.16, |
|
"learning_rate": 2.1297143438040054e-05, |
|
"loss": 0.0769, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 6.17, |
|
"learning_rate": 2.124631493341466e-05, |
|
"loss": 0.1254, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 6.18, |
|
"learning_rate": 2.1195486428789266e-05, |
|
"loss": 0.1206, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 6.19, |
|
"learning_rate": 2.1144657924163873e-05, |
|
"loss": 0.1022, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 6.2, |
|
"learning_rate": 2.109382941953848e-05, |
|
"loss": 0.113, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 6.21, |
|
"learning_rate": 2.1043000914913082e-05, |
|
"loss": 0.1041, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 6.22, |
|
"learning_rate": 2.099217241028769e-05, |
|
"loss": 0.1096, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 6.23, |
|
"learning_rate": 2.0941343905662298e-05, |
|
"loss": 0.0917, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 6.24, |
|
"learning_rate": 2.0890515401036904e-05, |
|
"loss": 0.102, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 6.24, |
|
"learning_rate": 2.083968689641151e-05, |
|
"loss": 0.1193, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 2.0788858391786113e-05, |
|
"loss": 0.1137, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 6.26, |
|
"learning_rate": 2.073802988716072e-05, |
|
"loss": 0.0728, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 6.27, |
|
"learning_rate": 2.0687201382535326e-05, |
|
"loss": 0.1141, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 6.28, |
|
"learning_rate": 2.0636372877909932e-05, |
|
"loss": 0.0876, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 6.29, |
|
"learning_rate": 2.0585544373284538e-05, |
|
"loss": 0.0897, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 6.3, |
|
"learning_rate": 2.0534715868659145e-05, |
|
"loss": 0.0996, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 6.31, |
|
"learning_rate": 2.048388736403375e-05, |
|
"loss": 0.0964, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 6.32, |
|
"learning_rate": 2.0433058859408357e-05, |
|
"loss": 0.1266, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 6.33, |
|
"learning_rate": 2.0382230354782963e-05, |
|
"loss": 0.1257, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 6.34, |
|
"learning_rate": 2.033140185015757e-05, |
|
"loss": 0.0987, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 6.35, |
|
"learning_rate": 2.0280573345532176e-05, |
|
"loss": 0.1201, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 6.35, |
|
"learning_rate": 2.0229744840906782e-05, |
|
"loss": 0.1349, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 6.36, |
|
"learning_rate": 2.0178916336281385e-05, |
|
"loss": 0.0879, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 6.37, |
|
"learning_rate": 2.012808783165599e-05, |
|
"loss": 0.0881, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"learning_rate": 2.00772593270306e-05, |
|
"loss": 0.1035, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 6.39, |
|
"learning_rate": 2.0026430822405207e-05, |
|
"loss": 0.0846, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 6.4, |
|
"learning_rate": 1.9975602317779814e-05, |
|
"loss": 0.0884, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 6.41, |
|
"learning_rate": 1.9924773813154417e-05, |
|
"loss": 0.0931, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 6.42, |
|
"learning_rate": 1.9873945308529023e-05, |
|
"loss": 0.1073, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 6.43, |
|
"learning_rate": 1.982311680390363e-05, |
|
"loss": 0.0938, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 6.44, |
|
"learning_rate": 1.9772288299278235e-05, |
|
"loss": 0.1037, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 6.45, |
|
"learning_rate": 1.9721459794652842e-05, |
|
"loss": 0.1279, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 6.45, |
|
"learning_rate": 1.9670631290027448e-05, |
|
"loss": 0.1044, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 6.46, |
|
"learning_rate": 1.9619802785402054e-05, |
|
"loss": 0.1065, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 6.47, |
|
"learning_rate": 1.956897428077666e-05, |
|
"loss": 0.0957, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 6.48, |
|
"learning_rate": 1.9518145776151267e-05, |
|
"loss": 0.1101, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 6.49, |
|
"learning_rate": 1.9467317271525873e-05, |
|
"loss": 0.1277, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"learning_rate": 1.941648876690048e-05, |
|
"loss": 0.124, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 6.51, |
|
"learning_rate": 1.9365660262275086e-05, |
|
"loss": 0.1077, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 6.52, |
|
"learning_rate": 1.931483175764969e-05, |
|
"loss": 0.1306, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 1.9264003253024295e-05, |
|
"loss": 0.108, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 6.54, |
|
"learning_rate": 1.9213174748398905e-05, |
|
"loss": 0.0956, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 6.55, |
|
"learning_rate": 1.916234624377351e-05, |
|
"loss": 0.0931, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 6.56, |
|
"learning_rate": 1.9111517739148117e-05, |
|
"loss": 0.1262, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 6.56, |
|
"learning_rate": 1.906068923452272e-05, |
|
"loss": 0.0925, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 6.57, |
|
"learning_rate": 1.9009860729897326e-05, |
|
"loss": 0.09, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 6.58, |
|
"learning_rate": 1.8959032225271933e-05, |
|
"loss": 0.1074, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 6.59, |
|
"learning_rate": 1.890820372064654e-05, |
|
"loss": 0.1037, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 1.8857375216021145e-05, |
|
"loss": 0.1026, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 6.61, |
|
"learning_rate": 1.880654671139575e-05, |
|
"loss": 0.1333, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 6.62, |
|
"learning_rate": 1.8755718206770358e-05, |
|
"loss": 0.0885, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 6.63, |
|
"learning_rate": 1.8704889702144964e-05, |
|
"loss": 0.1095, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 6.64, |
|
"learning_rate": 1.865406119751957e-05, |
|
"loss": 0.1133, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"learning_rate": 1.8603232692894177e-05, |
|
"loss": 0.1173, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 6.66, |
|
"learning_rate": 1.8552404188268783e-05, |
|
"loss": 0.0833, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 6.67, |
|
"learning_rate": 1.850157568364339e-05, |
|
"loss": 0.0924, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 6.67, |
|
"learning_rate": 1.8450747179017992e-05, |
|
"loss": 0.0854, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 6.68, |
|
"learning_rate": 1.83999186743926e-05, |
|
"loss": 0.1039, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 6.69, |
|
"learning_rate": 1.8349090169767205e-05, |
|
"loss": 0.111, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 6.7, |
|
"learning_rate": 1.8298261665141814e-05, |
|
"loss": 0.0932, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 6.71, |
|
"learning_rate": 1.824743316051642e-05, |
|
"loss": 0.1303, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 6.72, |
|
"learning_rate": 1.8196604655891024e-05, |
|
"loss": 0.0795, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 6.73, |
|
"learning_rate": 1.814577615126563e-05, |
|
"loss": 0.1012, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 6.74, |
|
"learning_rate": 1.8094947646640236e-05, |
|
"loss": 0.0965, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 6.75, |
|
"learning_rate": 1.8044119142014842e-05, |
|
"loss": 0.1429, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 6.76, |
|
"learning_rate": 1.799329063738945e-05, |
|
"loss": 0.0846, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"learning_rate": 1.7942462132764055e-05, |
|
"loss": 0.1217, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"learning_rate": 1.789163362813866e-05, |
|
"loss": 0.0834, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 6.78, |
|
"learning_rate": 1.7840805123513268e-05, |
|
"loss": 0.1285, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 6.79, |
|
"learning_rate": 1.7789976618887874e-05, |
|
"loss": 0.1155, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 6.8, |
|
"learning_rate": 1.773914811426248e-05, |
|
"loss": 0.0882, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 6.81, |
|
"learning_rate": 1.7688319609637087e-05, |
|
"loss": 0.0978, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 6.82, |
|
"learning_rate": 1.7637491105011693e-05, |
|
"loss": 0.1181, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 6.83, |
|
"learning_rate": 1.7586662600386296e-05, |
|
"loss": 0.0937, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 6.84, |
|
"learning_rate": 1.7535834095760902e-05, |
|
"loss": 0.1175, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 6.85, |
|
"learning_rate": 1.7485005591135508e-05, |
|
"loss": 0.0708, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 6.86, |
|
"learning_rate": 1.7434177086510115e-05, |
|
"loss": 0.0917, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 6.87, |
|
"learning_rate": 1.7383348581884724e-05, |
|
"loss": 0.1017, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 6.88, |
|
"learning_rate": 1.7332520077259327e-05, |
|
"loss": 0.112, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 6.88, |
|
"learning_rate": 1.7281691572633933e-05, |
|
"loss": 0.0953, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 6.89, |
|
"learning_rate": 1.723086306800854e-05, |
|
"loss": 0.098, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 6.9, |
|
"learning_rate": 1.7180034563383146e-05, |
|
"loss": 0.095, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 6.91, |
|
"learning_rate": 1.7129206058757752e-05, |
|
"loss": 0.0859, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 6.92, |
|
"learning_rate": 1.707837755413236e-05, |
|
"loss": 0.0875, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 6.93, |
|
"learning_rate": 1.7027549049506965e-05, |
|
"loss": 0.0916, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 6.94, |
|
"learning_rate": 1.6976720544881568e-05, |
|
"loss": 0.1097, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 6.95, |
|
"learning_rate": 1.6925892040256177e-05, |
|
"loss": 0.0594, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 6.96, |
|
"learning_rate": 1.6875063535630784e-05, |
|
"loss": 0.1163, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 6.97, |
|
"learning_rate": 1.682423503100539e-05, |
|
"loss": 0.1046, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 6.98, |
|
"learning_rate": 1.6773406526379996e-05, |
|
"loss": 0.1123, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 6.99, |
|
"learning_rate": 1.67225780217546e-05, |
|
"loss": 0.0895, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 6.99, |
|
"learning_rate": 1.6671749517129205e-05, |
|
"loss": 0.0924, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_accuracy": 0.9809333333333333, |
|
"eval_loss": 0.05264745652675629, |
|
"eval_runtime": 102.6038, |
|
"eval_samples_per_second": 292.387, |
|
"eval_steps_per_second": 18.274, |
|
"step": 7656 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 1.6620921012503812e-05, |
|
"loss": 0.0788, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 1.6570092507878418e-05, |
|
"loss": 0.0714, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 1.6519264003253024e-05, |
|
"loss": 0.1153, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 1.6468435498627634e-05, |
|
"loss": 0.086, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 1.6417606994002237e-05, |
|
"loss": 0.1047, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 7.05, |
|
"learning_rate": 1.6366778489376843e-05, |
|
"loss": 0.1012, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 7.06, |
|
"learning_rate": 1.631594998475145e-05, |
|
"loss": 0.1116, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 7.07, |
|
"learning_rate": 1.6265121480126056e-05, |
|
"loss": 0.1029, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 7.08, |
|
"learning_rate": 1.6214292975500662e-05, |
|
"loss": 0.0967, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 7.09, |
|
"learning_rate": 1.616346447087527e-05, |
|
"loss": 0.1088, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 7.09, |
|
"learning_rate": 1.611263596624987e-05, |
|
"loss": 0.1231, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 7.1, |
|
"learning_rate": 1.6061807461624478e-05, |
|
"loss": 0.0778, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 7.11, |
|
"learning_rate": 1.6010978956999087e-05, |
|
"loss": 0.0903, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 7.12, |
|
"learning_rate": 1.5960150452373694e-05, |
|
"loss": 0.0847, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 7.13, |
|
"learning_rate": 1.59093219477483e-05, |
|
"loss": 0.1004, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 7.14, |
|
"learning_rate": 1.5858493443122903e-05, |
|
"loss": 0.0938, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 7.15, |
|
"learning_rate": 1.580766493849751e-05, |
|
"loss": 0.0768, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 7.16, |
|
"learning_rate": 1.5756836433872115e-05, |
|
"loss": 0.0817, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 7.17, |
|
"learning_rate": 1.570600792924672e-05, |
|
"loss": 0.0988, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 7.18, |
|
"learning_rate": 1.5655179424621328e-05, |
|
"loss": 0.0967, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 7.19, |
|
"learning_rate": 1.5604350919995934e-05, |
|
"loss": 0.0977, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 7.2, |
|
"learning_rate": 1.555352241537054e-05, |
|
"loss": 0.092, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 7.2, |
|
"learning_rate": 1.5502693910745147e-05, |
|
"loss": 0.0934, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 7.21, |
|
"learning_rate": 1.5451865406119753e-05, |
|
"loss": 0.0964, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 7.22, |
|
"learning_rate": 1.540103690149436e-05, |
|
"loss": 0.1101, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 7.23, |
|
"learning_rate": 1.5350208396868966e-05, |
|
"loss": 0.1129, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 7.24, |
|
"learning_rate": 1.5299379892243572e-05, |
|
"loss": 0.0887, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 7.25, |
|
"learning_rate": 1.5248551387618176e-05, |
|
"loss": 0.0897, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 7.26, |
|
"learning_rate": 1.5197722882992783e-05, |
|
"loss": 0.0791, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 7.27, |
|
"learning_rate": 1.5146894378367387e-05, |
|
"loss": 0.0942, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 7.28, |
|
"learning_rate": 1.5096065873741997e-05, |
|
"loss": 0.0782, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 7.29, |
|
"learning_rate": 1.5045237369116602e-05, |
|
"loss": 0.1055, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 7.3, |
|
"learning_rate": 1.4994408864491208e-05, |
|
"loss": 0.0957, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 7.31, |
|
"learning_rate": 1.4943580359865814e-05, |
|
"loss": 0.0984, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 7.31, |
|
"learning_rate": 1.4892751855240419e-05, |
|
"loss": 0.0959, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 7.32, |
|
"learning_rate": 1.4841923350615025e-05, |
|
"loss": 0.1075, |
|
"step": 8010 |
|
}, |
|
{ |
|
"epoch": 7.33, |
|
"learning_rate": 1.4791094845989631e-05, |
|
"loss": 0.0908, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 7.34, |
|
"learning_rate": 1.4740266341364236e-05, |
|
"loss": 0.0928, |
|
"step": 8030 |
|
}, |
|
{ |
|
"epoch": 7.35, |
|
"learning_rate": 1.4689437836738846e-05, |
|
"loss": 0.099, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 7.36, |
|
"learning_rate": 1.463860933211345e-05, |
|
"loss": 0.0954, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 7.37, |
|
"learning_rate": 1.4587780827488056e-05, |
|
"loss": 0.0728, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 7.38, |
|
"learning_rate": 1.4536952322862663e-05, |
|
"loss": 0.1019, |
|
"step": 8070 |
|
}, |
|
{ |
|
"epoch": 7.39, |
|
"learning_rate": 1.4486123818237269e-05, |
|
"loss": 0.1178, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 7.4, |
|
"learning_rate": 1.4435295313611874e-05, |
|
"loss": 0.087, |
|
"step": 8090 |
|
}, |
|
{ |
|
"epoch": 7.41, |
|
"learning_rate": 1.438446680898648e-05, |
|
"loss": 0.09, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 7.41, |
|
"learning_rate": 1.4333638304361086e-05, |
|
"loss": 0.0804, |
|
"step": 8110 |
|
}, |
|
{ |
|
"epoch": 7.42, |
|
"learning_rate": 1.428280979973569e-05, |
|
"loss": 0.0766, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 7.43, |
|
"learning_rate": 1.42319812951103e-05, |
|
"loss": 0.0966, |
|
"step": 8130 |
|
}, |
|
{ |
|
"epoch": 7.44, |
|
"learning_rate": 1.4181152790484905e-05, |
|
"loss": 0.0942, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 7.45, |
|
"learning_rate": 1.4130324285859511e-05, |
|
"loss": 0.0837, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 7.46, |
|
"learning_rate": 1.4079495781234118e-05, |
|
"loss": 0.0912, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 7.47, |
|
"learning_rate": 1.4028667276608722e-05, |
|
"loss": 0.1134, |
|
"step": 8170 |
|
}, |
|
{ |
|
"epoch": 7.48, |
|
"learning_rate": 1.3977838771983329e-05, |
|
"loss": 0.1082, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 7.49, |
|
"learning_rate": 1.3927010267357935e-05, |
|
"loss": 0.1082, |
|
"step": 8190 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 1.387618176273254e-05, |
|
"loss": 0.0921, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 7.51, |
|
"learning_rate": 1.3825353258107146e-05, |
|
"loss": 0.1005, |
|
"step": 8210 |
|
}, |
|
{ |
|
"epoch": 7.52, |
|
"learning_rate": 1.3774524753481754e-05, |
|
"loss": 0.1213, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 7.52, |
|
"learning_rate": 1.372369624885636e-05, |
|
"loss": 0.0583, |
|
"step": 8230 |
|
}, |
|
{ |
|
"epoch": 7.53, |
|
"learning_rate": 1.3672867744230966e-05, |
|
"loss": 0.0846, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 7.54, |
|
"learning_rate": 1.3622039239605573e-05, |
|
"loss": 0.0881, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 7.55, |
|
"learning_rate": 1.3571210734980177e-05, |
|
"loss": 0.1069, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 7.56, |
|
"learning_rate": 1.3520382230354783e-05, |
|
"loss": 0.088, |
|
"step": 8270 |
|
}, |
|
{ |
|
"epoch": 7.57, |
|
"learning_rate": 1.346955372572939e-05, |
|
"loss": 0.0779, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 7.58, |
|
"learning_rate": 1.3418725221103994e-05, |
|
"loss": 0.1205, |
|
"step": 8290 |
|
}, |
|
{ |
|
"epoch": 7.59, |
|
"learning_rate": 1.33678967164786e-05, |
|
"loss": 0.1117, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 7.6, |
|
"learning_rate": 1.3317068211853209e-05, |
|
"loss": 0.1121, |
|
"step": 8310 |
|
}, |
|
{ |
|
"epoch": 7.61, |
|
"learning_rate": 1.3266239707227815e-05, |
|
"loss": 0.0909, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 7.62, |
|
"learning_rate": 1.3215411202602421e-05, |
|
"loss": 0.1003, |
|
"step": 8330 |
|
}, |
|
{ |
|
"epoch": 7.63, |
|
"learning_rate": 1.3164582697977026e-05, |
|
"loss": 0.0652, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 7.63, |
|
"learning_rate": 1.3113754193351632e-05, |
|
"loss": 0.0963, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 7.64, |
|
"learning_rate": 1.3062925688726238e-05, |
|
"loss": 0.1049, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 7.65, |
|
"learning_rate": 1.3012097184100843e-05, |
|
"loss": 0.0977, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 7.66, |
|
"learning_rate": 1.296126867947545e-05, |
|
"loss": 0.1228, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 7.67, |
|
"learning_rate": 1.2910440174850056e-05, |
|
"loss": 0.13, |
|
"step": 8390 |
|
}, |
|
{ |
|
"epoch": 7.68, |
|
"learning_rate": 1.2859611670224663e-05, |
|
"loss": 0.1099, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 7.69, |
|
"learning_rate": 1.280878316559927e-05, |
|
"loss": 0.1012, |
|
"step": 8410 |
|
}, |
|
{ |
|
"epoch": 7.7, |
|
"learning_rate": 1.2757954660973876e-05, |
|
"loss": 0.0781, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 7.71, |
|
"learning_rate": 1.270712615634848e-05, |
|
"loss": 0.0855, |
|
"step": 8430 |
|
}, |
|
{ |
|
"epoch": 7.72, |
|
"learning_rate": 1.2656297651723087e-05, |
|
"loss": 0.0958, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 7.73, |
|
"learning_rate": 1.2605469147097693e-05, |
|
"loss": 0.1045, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 7.73, |
|
"learning_rate": 1.2554640642472298e-05, |
|
"loss": 0.1095, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 7.74, |
|
"learning_rate": 1.2503812137846904e-05, |
|
"loss": 0.0886, |
|
"step": 8470 |
|
}, |
|
{ |
|
"epoch": 7.75, |
|
"learning_rate": 1.2452983633221512e-05, |
|
"loss": 0.0981, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 7.76, |
|
"learning_rate": 1.2402155128596117e-05, |
|
"loss": 0.0769, |
|
"step": 8490 |
|
}, |
|
{ |
|
"epoch": 7.77, |
|
"learning_rate": 1.2351326623970723e-05, |
|
"loss": 0.0889, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 7.78, |
|
"learning_rate": 1.230049811934533e-05, |
|
"loss": 0.0807, |
|
"step": 8510 |
|
}, |
|
{ |
|
"epoch": 7.79, |
|
"learning_rate": 1.2249669614719936e-05, |
|
"loss": 0.089, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 7.8, |
|
"learning_rate": 1.2198841110094542e-05, |
|
"loss": 0.1046, |
|
"step": 8530 |
|
}, |
|
{ |
|
"epoch": 7.81, |
|
"learning_rate": 1.2148012605469148e-05, |
|
"loss": 0.1205, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 7.82, |
|
"learning_rate": 1.2097184100843753e-05, |
|
"loss": 0.1335, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 7.83, |
|
"learning_rate": 1.204635559621836e-05, |
|
"loss": 0.0942, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 7.84, |
|
"learning_rate": 1.1995527091592965e-05, |
|
"loss": 0.1021, |
|
"step": 8570 |
|
}, |
|
{ |
|
"epoch": 7.84, |
|
"learning_rate": 1.1944698586967572e-05, |
|
"loss": 0.082, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 7.85, |
|
"learning_rate": 1.1893870082342178e-05, |
|
"loss": 0.1006, |
|
"step": 8590 |
|
}, |
|
{ |
|
"epoch": 7.86, |
|
"learning_rate": 1.1843041577716784e-05, |
|
"loss": 0.0961, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 7.87, |
|
"learning_rate": 1.179221307309139e-05, |
|
"loss": 0.1122, |
|
"step": 8610 |
|
}, |
|
{ |
|
"epoch": 7.88, |
|
"learning_rate": 1.1741384568465997e-05, |
|
"loss": 0.0573, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 7.89, |
|
"learning_rate": 1.1690556063840601e-05, |
|
"loss": 0.0848, |
|
"step": 8630 |
|
}, |
|
{ |
|
"epoch": 7.9, |
|
"learning_rate": 1.163972755921521e-05, |
|
"loss": 0.0722, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 7.91, |
|
"learning_rate": 1.1588899054589816e-05, |
|
"loss": 0.0939, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 7.92, |
|
"learning_rate": 1.153807054996442e-05, |
|
"loss": 0.0616, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 7.93, |
|
"learning_rate": 1.1487242045339026e-05, |
|
"loss": 0.1085, |
|
"step": 8670 |
|
}, |
|
{ |
|
"epoch": 7.94, |
|
"learning_rate": 1.1436413540713633e-05, |
|
"loss": 0.1256, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 7.95, |
|
"learning_rate": 1.1385585036088239e-05, |
|
"loss": 0.0937, |
|
"step": 8690 |
|
}, |
|
{ |
|
"epoch": 7.95, |
|
"learning_rate": 1.1334756531462845e-05, |
|
"loss": 0.1, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 7.96, |
|
"learning_rate": 1.1283928026837452e-05, |
|
"loss": 0.1087, |
|
"step": 8710 |
|
}, |
|
{ |
|
"epoch": 7.97, |
|
"learning_rate": 1.1233099522212056e-05, |
|
"loss": 0.0906, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 7.98, |
|
"learning_rate": 1.1182271017586664e-05, |
|
"loss": 0.1057, |
|
"step": 8730 |
|
}, |
|
{ |
|
"epoch": 7.99, |
|
"learning_rate": 1.1131442512961269e-05, |
|
"loss": 0.0938, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 1.1080614008335875e-05, |
|
"loss": 0.0779, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_accuracy": 0.9818666666666667, |
|
"eval_loss": 0.051762744784355164, |
|
"eval_runtime": 101.2563, |
|
"eval_samples_per_second": 296.278, |
|
"eval_steps_per_second": 18.517, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 1.1029785503710481e-05, |
|
"loss": 0.0775, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 1.0978956999085088e-05, |
|
"loss": 0.0886, |
|
"step": 8770 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 1.0928128494459694e-05, |
|
"loss": 0.0933, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 1.08772999898343e-05, |
|
"loss": 0.1209, |
|
"step": 8790 |
|
}, |
|
{ |
|
"epoch": 8.05, |
|
"learning_rate": 1.0826471485208905e-05, |
|
"loss": 0.0808, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 8.05, |
|
"learning_rate": 1.0775642980583511e-05, |
|
"loss": 0.0894, |
|
"step": 8810 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"learning_rate": 1.0724814475958119e-05, |
|
"loss": 0.079, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 8.07, |
|
"learning_rate": 1.0673985971332724e-05, |
|
"loss": 0.0719, |
|
"step": 8830 |
|
}, |
|
{ |
|
"epoch": 8.08, |
|
"learning_rate": 1.062315746670733e-05, |
|
"loss": 0.103, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 8.09, |
|
"learning_rate": 1.0572328962081936e-05, |
|
"loss": 0.0913, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 8.1, |
|
"learning_rate": 1.0521500457456541e-05, |
|
"loss": 0.072, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 8.11, |
|
"learning_rate": 1.0470671952831149e-05, |
|
"loss": 0.0739, |
|
"step": 8870 |
|
}, |
|
{ |
|
"epoch": 8.12, |
|
"learning_rate": 1.0419843448205755e-05, |
|
"loss": 0.0947, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 8.13, |
|
"learning_rate": 1.036901494358036e-05, |
|
"loss": 0.055, |
|
"step": 8890 |
|
}, |
|
{ |
|
"epoch": 8.14, |
|
"learning_rate": 1.0318186438954966e-05, |
|
"loss": 0.0903, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 8.15, |
|
"learning_rate": 1.0267357934329572e-05, |
|
"loss": 0.1091, |
|
"step": 8910 |
|
}, |
|
{ |
|
"epoch": 8.16, |
|
"learning_rate": 1.0216529429704179e-05, |
|
"loss": 0.0875, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 8.16, |
|
"learning_rate": 1.0165700925078785e-05, |
|
"loss": 0.0824, |
|
"step": 8930 |
|
}, |
|
{ |
|
"epoch": 8.17, |
|
"learning_rate": 1.0114872420453391e-05, |
|
"loss": 0.0918, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 8.18, |
|
"learning_rate": 1.0064043915827996e-05, |
|
"loss": 0.0996, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 8.19, |
|
"learning_rate": 1.0013215411202604e-05, |
|
"loss": 0.074, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 8.2, |
|
"learning_rate": 9.962386906577208e-06, |
|
"loss": 0.0574, |
|
"step": 8970 |
|
}, |
|
{ |
|
"epoch": 8.21, |
|
"learning_rate": 9.911558401951815e-06, |
|
"loss": 0.0887, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 8.22, |
|
"learning_rate": 9.860729897326421e-06, |
|
"loss": 0.1018, |
|
"step": 8990 |
|
}, |
|
{ |
|
"epoch": 8.23, |
|
"learning_rate": 9.809901392701027e-06, |
|
"loss": 0.0784, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 8.24, |
|
"learning_rate": 9.759072888075633e-06, |
|
"loss": 0.0852, |
|
"step": 9010 |
|
}, |
|
{ |
|
"epoch": 8.25, |
|
"learning_rate": 9.70824438345024e-06, |
|
"loss": 0.1074, |
|
"step": 9020 |
|
}, |
|
{ |
|
"epoch": 8.26, |
|
"learning_rate": 9.657415878824844e-06, |
|
"loss": 0.0827, |
|
"step": 9030 |
|
}, |
|
{ |
|
"epoch": 8.27, |
|
"learning_rate": 9.606587374199452e-06, |
|
"loss": 0.0841, |
|
"step": 9040 |
|
}, |
|
{ |
|
"epoch": 8.27, |
|
"learning_rate": 9.555758869574059e-06, |
|
"loss": 0.0828, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 8.28, |
|
"learning_rate": 9.504930364948663e-06, |
|
"loss": 0.0742, |
|
"step": 9060 |
|
}, |
|
{ |
|
"epoch": 8.29, |
|
"learning_rate": 9.45410186032327e-06, |
|
"loss": 0.1027, |
|
"step": 9070 |
|
}, |
|
{ |
|
"epoch": 8.3, |
|
"learning_rate": 9.403273355697876e-06, |
|
"loss": 0.0791, |
|
"step": 9080 |
|
}, |
|
{ |
|
"epoch": 8.31, |
|
"learning_rate": 9.352444851072482e-06, |
|
"loss": 0.1058, |
|
"step": 9090 |
|
}, |
|
{ |
|
"epoch": 8.32, |
|
"learning_rate": 9.301616346447088e-06, |
|
"loss": 0.079, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 8.33, |
|
"learning_rate": 9.250787841821695e-06, |
|
"loss": 0.1119, |
|
"step": 9110 |
|
}, |
|
{ |
|
"epoch": 8.34, |
|
"learning_rate": 9.1999593371963e-06, |
|
"loss": 0.0941, |
|
"step": 9120 |
|
}, |
|
{ |
|
"epoch": 8.35, |
|
"learning_rate": 9.149130832570907e-06, |
|
"loss": 0.0657, |
|
"step": 9130 |
|
}, |
|
{ |
|
"epoch": 8.36, |
|
"learning_rate": 9.098302327945512e-06, |
|
"loss": 0.0726, |
|
"step": 9140 |
|
}, |
|
{ |
|
"epoch": 8.37, |
|
"learning_rate": 9.047473823320118e-06, |
|
"loss": 0.0987, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 8.37, |
|
"learning_rate": 8.996645318694724e-06, |
|
"loss": 0.1138, |
|
"step": 9160 |
|
}, |
|
{ |
|
"epoch": 8.38, |
|
"learning_rate": 8.94581681406933e-06, |
|
"loss": 0.1244, |
|
"step": 9170 |
|
}, |
|
{ |
|
"epoch": 8.39, |
|
"learning_rate": 8.894988309443937e-06, |
|
"loss": 0.0844, |
|
"step": 9180 |
|
}, |
|
{ |
|
"epoch": 8.4, |
|
"learning_rate": 8.844159804818543e-06, |
|
"loss": 0.0853, |
|
"step": 9190 |
|
}, |
|
{ |
|
"epoch": 8.41, |
|
"learning_rate": 8.793331300193148e-06, |
|
"loss": 0.0705, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 8.42, |
|
"learning_rate": 8.742502795567754e-06, |
|
"loss": 0.1141, |
|
"step": 9210 |
|
}, |
|
{ |
|
"epoch": 8.43, |
|
"learning_rate": 8.691674290942362e-06, |
|
"loss": 0.1058, |
|
"step": 9220 |
|
}, |
|
{ |
|
"epoch": 8.44, |
|
"learning_rate": 8.640845786316967e-06, |
|
"loss": 0.0774, |
|
"step": 9230 |
|
}, |
|
{ |
|
"epoch": 8.45, |
|
"learning_rate": 8.590017281691573e-06, |
|
"loss": 0.0793, |
|
"step": 9240 |
|
}, |
|
{ |
|
"epoch": 8.46, |
|
"learning_rate": 8.53918877706618e-06, |
|
"loss": 0.0908, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 8.47, |
|
"learning_rate": 8.488360272440784e-06, |
|
"loss": 0.1035, |
|
"step": 9260 |
|
}, |
|
{ |
|
"epoch": 8.48, |
|
"learning_rate": 8.437531767815392e-06, |
|
"loss": 0.0774, |
|
"step": 9270 |
|
}, |
|
{ |
|
"epoch": 8.48, |
|
"learning_rate": 8.386703263189998e-06, |
|
"loss": 0.0838, |
|
"step": 9280 |
|
}, |
|
{ |
|
"epoch": 8.49, |
|
"learning_rate": 8.335874758564603e-06, |
|
"loss": 0.0879, |
|
"step": 9290 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"learning_rate": 8.285046253939209e-06, |
|
"loss": 0.0521, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 8.51, |
|
"learning_rate": 8.234217749313817e-06, |
|
"loss": 0.0696, |
|
"step": 9310 |
|
}, |
|
{ |
|
"epoch": 8.52, |
|
"learning_rate": 8.183389244688422e-06, |
|
"loss": 0.0575, |
|
"step": 9320 |
|
}, |
|
{ |
|
"epoch": 8.53, |
|
"learning_rate": 8.132560740063028e-06, |
|
"loss": 0.0922, |
|
"step": 9330 |
|
}, |
|
{ |
|
"epoch": 8.54, |
|
"learning_rate": 8.081732235437634e-06, |
|
"loss": 0.0964, |
|
"step": 9340 |
|
}, |
|
{ |
|
"epoch": 8.55, |
|
"learning_rate": 8.030903730812239e-06, |
|
"loss": 0.0887, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 8.56, |
|
"learning_rate": 7.980075226186847e-06, |
|
"loss": 0.1038, |
|
"step": 9360 |
|
}, |
|
{ |
|
"epoch": 8.57, |
|
"learning_rate": 7.929246721561451e-06, |
|
"loss": 0.1118, |
|
"step": 9370 |
|
}, |
|
{ |
|
"epoch": 8.58, |
|
"learning_rate": 7.878418216936058e-06, |
|
"loss": 0.1013, |
|
"step": 9380 |
|
}, |
|
{ |
|
"epoch": 8.59, |
|
"learning_rate": 7.827589712310664e-06, |
|
"loss": 0.1175, |
|
"step": 9390 |
|
}, |
|
{ |
|
"epoch": 8.59, |
|
"learning_rate": 7.77676120768527e-06, |
|
"loss": 0.0783, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 8.6, |
|
"learning_rate": 7.725932703059876e-06, |
|
"loss": 0.0808, |
|
"step": 9410 |
|
}, |
|
{ |
|
"epoch": 8.61, |
|
"learning_rate": 7.675104198434483e-06, |
|
"loss": 0.105, |
|
"step": 9420 |
|
}, |
|
{ |
|
"epoch": 8.62, |
|
"learning_rate": 7.624275693809088e-06, |
|
"loss": 0.1038, |
|
"step": 9430 |
|
}, |
|
{ |
|
"epoch": 8.63, |
|
"learning_rate": 7.573447189183694e-06, |
|
"loss": 0.0696, |
|
"step": 9440 |
|
}, |
|
{ |
|
"epoch": 8.64, |
|
"learning_rate": 7.522618684558301e-06, |
|
"loss": 0.1156, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 8.65, |
|
"learning_rate": 7.471790179932907e-06, |
|
"loss": 0.0767, |
|
"step": 9460 |
|
}, |
|
{ |
|
"epoch": 8.66, |
|
"learning_rate": 7.4209616753075125e-06, |
|
"loss": 0.0672, |
|
"step": 9470 |
|
}, |
|
{ |
|
"epoch": 8.67, |
|
"learning_rate": 7.370133170682118e-06, |
|
"loss": 0.0908, |
|
"step": 9480 |
|
}, |
|
{ |
|
"epoch": 8.68, |
|
"learning_rate": 7.319304666056725e-06, |
|
"loss": 0.0985, |
|
"step": 9490 |
|
}, |
|
{ |
|
"epoch": 8.69, |
|
"learning_rate": 7.268476161431331e-06, |
|
"loss": 0.0973, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 8.69, |
|
"learning_rate": 7.217647656805937e-06, |
|
"loss": 0.0895, |
|
"step": 9510 |
|
}, |
|
{ |
|
"epoch": 8.7, |
|
"learning_rate": 7.166819152180543e-06, |
|
"loss": 0.0847, |
|
"step": 9520 |
|
}, |
|
{ |
|
"epoch": 8.71, |
|
"learning_rate": 7.11599064755515e-06, |
|
"loss": 0.117, |
|
"step": 9530 |
|
}, |
|
{ |
|
"epoch": 8.72, |
|
"learning_rate": 7.065162142929756e-06, |
|
"loss": 0.0863, |
|
"step": 9540 |
|
}, |
|
{ |
|
"epoch": 8.73, |
|
"learning_rate": 7.014333638304361e-06, |
|
"loss": 0.0981, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 8.74, |
|
"learning_rate": 6.963505133678967e-06, |
|
"loss": 0.0899, |
|
"step": 9560 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"learning_rate": 6.912676629053573e-06, |
|
"loss": 0.0861, |
|
"step": 9570 |
|
}, |
|
{ |
|
"epoch": 8.76, |
|
"learning_rate": 6.86184812442818e-06, |
|
"loss": 0.1042, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 8.77, |
|
"learning_rate": 6.811019619802786e-06, |
|
"loss": 0.0927, |
|
"step": 9590 |
|
}, |
|
{ |
|
"epoch": 8.78, |
|
"learning_rate": 6.760191115177392e-06, |
|
"loss": 0.075, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 8.79, |
|
"learning_rate": 6.709362610551997e-06, |
|
"loss": 0.0615, |
|
"step": 9610 |
|
}, |
|
{ |
|
"epoch": 8.8, |
|
"learning_rate": 6.658534105926604e-06, |
|
"loss": 0.0934, |
|
"step": 9620 |
|
}, |
|
{ |
|
"epoch": 8.8, |
|
"learning_rate": 6.607705601301211e-06, |
|
"loss": 0.0871, |
|
"step": 9630 |
|
}, |
|
{ |
|
"epoch": 8.81, |
|
"learning_rate": 6.556877096675816e-06, |
|
"loss": 0.0855, |
|
"step": 9640 |
|
}, |
|
{ |
|
"epoch": 8.82, |
|
"learning_rate": 6.5060485920504215e-06, |
|
"loss": 0.1094, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 8.83, |
|
"learning_rate": 6.455220087425028e-06, |
|
"loss": 0.1232, |
|
"step": 9660 |
|
}, |
|
{ |
|
"epoch": 8.84, |
|
"learning_rate": 6.404391582799635e-06, |
|
"loss": 0.0762, |
|
"step": 9670 |
|
}, |
|
{ |
|
"epoch": 8.85, |
|
"learning_rate": 6.35356307817424e-06, |
|
"loss": 0.0699, |
|
"step": 9680 |
|
}, |
|
{ |
|
"epoch": 8.86, |
|
"learning_rate": 6.302734573548847e-06, |
|
"loss": 0.0934, |
|
"step": 9690 |
|
}, |
|
{ |
|
"epoch": 8.87, |
|
"learning_rate": 6.251906068923452e-06, |
|
"loss": 0.0564, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 8.88, |
|
"learning_rate": 6.201077564298058e-06, |
|
"loss": 0.0496, |
|
"step": 9710 |
|
}, |
|
{ |
|
"epoch": 8.89, |
|
"learning_rate": 6.150249059672665e-06, |
|
"loss": 0.0697, |
|
"step": 9720 |
|
}, |
|
{ |
|
"epoch": 8.9, |
|
"learning_rate": 6.099420555047271e-06, |
|
"loss": 0.093, |
|
"step": 9730 |
|
}, |
|
{ |
|
"epoch": 8.91, |
|
"learning_rate": 6.048592050421876e-06, |
|
"loss": 0.0947, |
|
"step": 9740 |
|
}, |
|
{ |
|
"epoch": 8.91, |
|
"learning_rate": 5.997763545796483e-06, |
|
"loss": 0.0931, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 8.92, |
|
"learning_rate": 5.946935041171089e-06, |
|
"loss": 0.0675, |
|
"step": 9760 |
|
}, |
|
{ |
|
"epoch": 8.93, |
|
"learning_rate": 5.896106536545695e-06, |
|
"loss": 0.0762, |
|
"step": 9770 |
|
}, |
|
{ |
|
"epoch": 8.94, |
|
"learning_rate": 5.845278031920301e-06, |
|
"loss": 0.0832, |
|
"step": 9780 |
|
}, |
|
{ |
|
"epoch": 8.95, |
|
"learning_rate": 5.794449527294908e-06, |
|
"loss": 0.0925, |
|
"step": 9790 |
|
}, |
|
{ |
|
"epoch": 8.96, |
|
"learning_rate": 5.743621022669513e-06, |
|
"loss": 0.0909, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 8.97, |
|
"learning_rate": 5.6927925180441195e-06, |
|
"loss": 0.0886, |
|
"step": 9810 |
|
}, |
|
{ |
|
"epoch": 8.98, |
|
"learning_rate": 5.641964013418726e-06, |
|
"loss": 0.0588, |
|
"step": 9820 |
|
}, |
|
{ |
|
"epoch": 8.99, |
|
"learning_rate": 5.591135508793332e-06, |
|
"loss": 0.0688, |
|
"step": 9830 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 5.5403070041679375e-06, |
|
"loss": 0.0903, |
|
"step": 9840 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_accuracy": 0.9825666666666667, |
|
"eval_loss": 0.05175173282623291, |
|
"eval_runtime": 101.3937, |
|
"eval_samples_per_second": 295.876, |
|
"eval_steps_per_second": 18.492, |
|
"step": 9843 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 5.489478499542544e-06, |
|
"loss": 0.0986, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 5.43864999491715e-06, |
|
"loss": 0.0729, |
|
"step": 9860 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.3878214902917556e-06, |
|
"loss": 0.0657, |
|
"step": 9870 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.336992985666362e-06, |
|
"loss": 0.0791, |
|
"step": 9880 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 5.286164481040968e-06, |
|
"loss": 0.1165, |
|
"step": 9890 |
|
}, |
|
{ |
|
"epoch": 9.05, |
|
"learning_rate": 5.235335976415574e-06, |
|
"loss": 0.0687, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"learning_rate": 5.18450747179018e-06, |
|
"loss": 0.0669, |
|
"step": 9910 |
|
}, |
|
{ |
|
"epoch": 9.07, |
|
"learning_rate": 5.133678967164786e-06, |
|
"loss": 0.0769, |
|
"step": 9920 |
|
}, |
|
{ |
|
"epoch": 9.08, |
|
"learning_rate": 5.0828504625393924e-06, |
|
"loss": 0.0733, |
|
"step": 9930 |
|
}, |
|
{ |
|
"epoch": 9.09, |
|
"learning_rate": 5.032021957913998e-06, |
|
"loss": 0.0912, |
|
"step": 9940 |
|
}, |
|
{ |
|
"epoch": 9.1, |
|
"learning_rate": 4.981193453288604e-06, |
|
"loss": 0.1091, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 9.11, |
|
"learning_rate": 4.9303649486632104e-06, |
|
"loss": 0.0967, |
|
"step": 9960 |
|
}, |
|
{ |
|
"epoch": 9.12, |
|
"learning_rate": 4.879536444037817e-06, |
|
"loss": 0.0636, |
|
"step": 9970 |
|
}, |
|
{ |
|
"epoch": 9.12, |
|
"learning_rate": 4.828707939412422e-06, |
|
"loss": 0.0849, |
|
"step": 9980 |
|
}, |
|
{ |
|
"epoch": 9.13, |
|
"learning_rate": 4.777879434787029e-06, |
|
"loss": 0.0975, |
|
"step": 9990 |
|
}, |
|
{ |
|
"epoch": 9.14, |
|
"learning_rate": 4.727050930161635e-06, |
|
"loss": 0.0965, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 9.15, |
|
"learning_rate": 4.676222425536241e-06, |
|
"loss": 0.0893, |
|
"step": 10010 |
|
}, |
|
{ |
|
"epoch": 9.16, |
|
"learning_rate": 4.625393920910847e-06, |
|
"loss": 0.0738, |
|
"step": 10020 |
|
}, |
|
{ |
|
"epoch": 9.17, |
|
"learning_rate": 4.574565416285454e-06, |
|
"loss": 0.0803, |
|
"step": 10030 |
|
}, |
|
{ |
|
"epoch": 9.18, |
|
"learning_rate": 4.523736911660059e-06, |
|
"loss": 0.0972, |
|
"step": 10040 |
|
}, |
|
{ |
|
"epoch": 9.19, |
|
"learning_rate": 4.472908407034665e-06, |
|
"loss": 0.0733, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 9.2, |
|
"learning_rate": 4.422079902409272e-06, |
|
"loss": 0.1134, |
|
"step": 10060 |
|
}, |
|
{ |
|
"epoch": 9.21, |
|
"learning_rate": 4.371251397783877e-06, |
|
"loss": 0.0716, |
|
"step": 10070 |
|
}, |
|
{ |
|
"epoch": 9.22, |
|
"learning_rate": 4.320422893158483e-06, |
|
"loss": 0.0927, |
|
"step": 10080 |
|
}, |
|
{ |
|
"epoch": 9.23, |
|
"learning_rate": 4.26959438853309e-06, |
|
"loss": 0.0786, |
|
"step": 10090 |
|
}, |
|
{ |
|
"epoch": 9.23, |
|
"learning_rate": 4.218765883907696e-06, |
|
"loss": 0.0624, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 9.24, |
|
"learning_rate": 4.167937379282301e-06, |
|
"loss": 0.0919, |
|
"step": 10110 |
|
}, |
|
{ |
|
"epoch": 9.25, |
|
"learning_rate": 4.1171088746569085e-06, |
|
"loss": 0.0825, |
|
"step": 10120 |
|
}, |
|
{ |
|
"epoch": 9.26, |
|
"learning_rate": 4.066280370031514e-06, |
|
"loss": 0.0912, |
|
"step": 10130 |
|
}, |
|
{ |
|
"epoch": 9.27, |
|
"learning_rate": 4.015451865406119e-06, |
|
"loss": 0.048, |
|
"step": 10140 |
|
}, |
|
{ |
|
"epoch": 9.28, |
|
"learning_rate": 3.964623360780726e-06, |
|
"loss": 0.0815, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 9.29, |
|
"learning_rate": 3.913794856155332e-06, |
|
"loss": 0.08, |
|
"step": 10160 |
|
}, |
|
{ |
|
"epoch": 9.3, |
|
"learning_rate": 3.862966351529938e-06, |
|
"loss": 0.0712, |
|
"step": 10170 |
|
}, |
|
{ |
|
"epoch": 9.31, |
|
"learning_rate": 3.812137846904544e-06, |
|
"loss": 0.0527, |
|
"step": 10180 |
|
}, |
|
{ |
|
"epoch": 9.32, |
|
"learning_rate": 3.7613093422791504e-06, |
|
"loss": 0.0653, |
|
"step": 10190 |
|
}, |
|
{ |
|
"epoch": 9.33, |
|
"learning_rate": 3.7104808376537563e-06, |
|
"loss": 0.043, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 9.33, |
|
"learning_rate": 3.6596523330283626e-06, |
|
"loss": 0.0821, |
|
"step": 10210 |
|
}, |
|
{ |
|
"epoch": 9.34, |
|
"learning_rate": 3.6088238284029684e-06, |
|
"loss": 0.0863, |
|
"step": 10220 |
|
}, |
|
{ |
|
"epoch": 9.35, |
|
"learning_rate": 3.557995323777575e-06, |
|
"loss": 0.0752, |
|
"step": 10230 |
|
}, |
|
{ |
|
"epoch": 9.36, |
|
"learning_rate": 3.5071668191521806e-06, |
|
"loss": 0.117, |
|
"step": 10240 |
|
}, |
|
{ |
|
"epoch": 9.37, |
|
"learning_rate": 3.4563383145267864e-06, |
|
"loss": 0.0767, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 9.38, |
|
"learning_rate": 3.405509809901393e-06, |
|
"loss": 0.0836, |
|
"step": 10260 |
|
}, |
|
{ |
|
"epoch": 9.39, |
|
"learning_rate": 3.3546813052759986e-06, |
|
"loss": 0.0983, |
|
"step": 10270 |
|
}, |
|
{ |
|
"epoch": 9.4, |
|
"learning_rate": 3.3038528006506053e-06, |
|
"loss": 0.0974, |
|
"step": 10280 |
|
}, |
|
{ |
|
"epoch": 9.41, |
|
"learning_rate": 3.2530242960252107e-06, |
|
"loss": 0.1011, |
|
"step": 10290 |
|
}, |
|
{ |
|
"epoch": 9.42, |
|
"learning_rate": 3.2021957913998174e-06, |
|
"loss": 0.0679, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 9.43, |
|
"learning_rate": 3.1513672867744233e-06, |
|
"loss": 0.083, |
|
"step": 10310 |
|
}, |
|
{ |
|
"epoch": 9.44, |
|
"learning_rate": 3.100538782149029e-06, |
|
"loss": 0.1026, |
|
"step": 10320 |
|
}, |
|
{ |
|
"epoch": 9.44, |
|
"learning_rate": 3.0497102775236355e-06, |
|
"loss": 0.0761, |
|
"step": 10330 |
|
}, |
|
{ |
|
"epoch": 9.45, |
|
"learning_rate": 2.9988817728982413e-06, |
|
"loss": 0.0795, |
|
"step": 10340 |
|
}, |
|
{ |
|
"epoch": 9.46, |
|
"learning_rate": 2.9480532682728476e-06, |
|
"loss": 0.1148, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 9.47, |
|
"learning_rate": 2.897224763647454e-06, |
|
"loss": 0.0704, |
|
"step": 10360 |
|
}, |
|
{ |
|
"epoch": 9.48, |
|
"learning_rate": 2.8463962590220598e-06, |
|
"loss": 0.0867, |
|
"step": 10370 |
|
}, |
|
{ |
|
"epoch": 9.49, |
|
"learning_rate": 2.795567754396666e-06, |
|
"loss": 0.0631, |
|
"step": 10380 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"learning_rate": 2.744739249771272e-06, |
|
"loss": 0.0751, |
|
"step": 10390 |
|
}, |
|
{ |
|
"epoch": 9.51, |
|
"learning_rate": 2.6939107451458778e-06, |
|
"loss": 0.1008, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 9.52, |
|
"learning_rate": 2.643082240520484e-06, |
|
"loss": 0.0671, |
|
"step": 10410 |
|
}, |
|
{ |
|
"epoch": 9.53, |
|
"learning_rate": 2.59225373589509e-06, |
|
"loss": 0.0774, |
|
"step": 10420 |
|
}, |
|
{ |
|
"epoch": 9.54, |
|
"learning_rate": 2.5414252312696962e-06, |
|
"loss": 0.0611, |
|
"step": 10430 |
|
}, |
|
{ |
|
"epoch": 9.55, |
|
"learning_rate": 2.490596726644302e-06, |
|
"loss": 0.0799, |
|
"step": 10440 |
|
}, |
|
{ |
|
"epoch": 9.55, |
|
"learning_rate": 2.4397682220189084e-06, |
|
"loss": 0.1153, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 9.56, |
|
"learning_rate": 2.3889397173935147e-06, |
|
"loss": 0.0569, |
|
"step": 10460 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"learning_rate": 2.3381112127681205e-06, |
|
"loss": 0.08, |
|
"step": 10470 |
|
}, |
|
{ |
|
"epoch": 9.58, |
|
"learning_rate": 2.287282708142727e-06, |
|
"loss": 0.0857, |
|
"step": 10480 |
|
}, |
|
{ |
|
"epoch": 9.59, |
|
"learning_rate": 2.2364542035173327e-06, |
|
"loss": 0.09, |
|
"step": 10490 |
|
}, |
|
{ |
|
"epoch": 9.6, |
|
"learning_rate": 2.1856256988919385e-06, |
|
"loss": 0.0815, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 9.61, |
|
"learning_rate": 2.134797194266545e-06, |
|
"loss": 0.0805, |
|
"step": 10510 |
|
}, |
|
{ |
|
"epoch": 9.62, |
|
"learning_rate": 2.0839686896411507e-06, |
|
"loss": 0.0895, |
|
"step": 10520 |
|
}, |
|
{ |
|
"epoch": 9.63, |
|
"learning_rate": 2.033140185015757e-06, |
|
"loss": 0.0943, |
|
"step": 10530 |
|
}, |
|
{ |
|
"epoch": 9.64, |
|
"learning_rate": 1.982311680390363e-06, |
|
"loss": 0.0836, |
|
"step": 10540 |
|
}, |
|
{ |
|
"epoch": 9.65, |
|
"learning_rate": 1.931483175764969e-06, |
|
"loss": 0.091, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 9.65, |
|
"learning_rate": 1.8806546711395752e-06, |
|
"loss": 0.0594, |
|
"step": 10560 |
|
}, |
|
{ |
|
"epoch": 9.66, |
|
"learning_rate": 1.8298261665141813e-06, |
|
"loss": 0.0641, |
|
"step": 10570 |
|
}, |
|
{ |
|
"epoch": 9.67, |
|
"learning_rate": 1.7789976618887876e-06, |
|
"loss": 0.0828, |
|
"step": 10580 |
|
}, |
|
{ |
|
"epoch": 9.68, |
|
"learning_rate": 1.7281691572633932e-06, |
|
"loss": 0.0984, |
|
"step": 10590 |
|
}, |
|
{ |
|
"epoch": 9.69, |
|
"learning_rate": 1.6773406526379993e-06, |
|
"loss": 0.1019, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 9.7, |
|
"learning_rate": 1.6265121480126054e-06, |
|
"loss": 0.0797, |
|
"step": 10610 |
|
}, |
|
{ |
|
"epoch": 9.71, |
|
"learning_rate": 1.5756836433872117e-06, |
|
"loss": 0.0828, |
|
"step": 10620 |
|
}, |
|
{ |
|
"epoch": 9.72, |
|
"learning_rate": 1.5248551387618177e-06, |
|
"loss": 0.0776, |
|
"step": 10630 |
|
}, |
|
{ |
|
"epoch": 9.73, |
|
"learning_rate": 1.4740266341364238e-06, |
|
"loss": 0.0883, |
|
"step": 10640 |
|
}, |
|
{ |
|
"epoch": 9.74, |
|
"learning_rate": 1.4231981295110299e-06, |
|
"loss": 0.0725, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 9.75, |
|
"learning_rate": 1.372369624885636e-06, |
|
"loss": 0.0799, |
|
"step": 10660 |
|
}, |
|
{ |
|
"epoch": 9.76, |
|
"learning_rate": 1.321541120260242e-06, |
|
"loss": 0.079, |
|
"step": 10670 |
|
}, |
|
{ |
|
"epoch": 9.76, |
|
"learning_rate": 1.2707126156348481e-06, |
|
"loss": 0.0643, |
|
"step": 10680 |
|
}, |
|
{ |
|
"epoch": 9.77, |
|
"learning_rate": 1.2198841110094542e-06, |
|
"loss": 0.0922, |
|
"step": 10690 |
|
}, |
|
{ |
|
"epoch": 9.78, |
|
"learning_rate": 1.1690556063840603e-06, |
|
"loss": 0.0822, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 9.79, |
|
"learning_rate": 1.1182271017586663e-06, |
|
"loss": 0.1008, |
|
"step": 10710 |
|
}, |
|
{ |
|
"epoch": 9.8, |
|
"learning_rate": 1.0673985971332724e-06, |
|
"loss": 0.0812, |
|
"step": 10720 |
|
}, |
|
{ |
|
"epoch": 9.81, |
|
"learning_rate": 1.0165700925078785e-06, |
|
"loss": 0.0655, |
|
"step": 10730 |
|
}, |
|
{ |
|
"epoch": 9.82, |
|
"learning_rate": 9.657415878824846e-07, |
|
"loss": 0.0824, |
|
"step": 10740 |
|
}, |
|
{ |
|
"epoch": 9.83, |
|
"learning_rate": 9.149130832570906e-07, |
|
"loss": 0.0865, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 9.84, |
|
"learning_rate": 8.640845786316966e-07, |
|
"loss": 0.0938, |
|
"step": 10760 |
|
}, |
|
{ |
|
"epoch": 9.85, |
|
"learning_rate": 8.132560740063027e-07, |
|
"loss": 0.1019, |
|
"step": 10770 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"learning_rate": 7.624275693809089e-07, |
|
"loss": 0.076, |
|
"step": 10780 |
|
}, |
|
{ |
|
"epoch": 9.87, |
|
"learning_rate": 7.115990647555149e-07, |
|
"loss": 0.0716, |
|
"step": 10790 |
|
}, |
|
{ |
|
"epoch": 9.87, |
|
"learning_rate": 6.60770560130121e-07, |
|
"loss": 0.057, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 9.88, |
|
"learning_rate": 6.099420555047271e-07, |
|
"loss": 0.0706, |
|
"step": 10810 |
|
}, |
|
{ |
|
"epoch": 9.89, |
|
"learning_rate": 5.591135508793332e-07, |
|
"loss": 0.0895, |
|
"step": 10820 |
|
}, |
|
{ |
|
"epoch": 9.9, |
|
"learning_rate": 5.082850462539392e-07, |
|
"loss": 0.0739, |
|
"step": 10830 |
|
}, |
|
{ |
|
"epoch": 9.91, |
|
"learning_rate": 4.574565416285453e-07, |
|
"loss": 0.114, |
|
"step": 10840 |
|
}, |
|
{ |
|
"epoch": 9.92, |
|
"learning_rate": 4.0662803700315134e-07, |
|
"loss": 0.0875, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 9.93, |
|
"learning_rate": 3.5579953237775747e-07, |
|
"loss": 0.1068, |
|
"step": 10860 |
|
}, |
|
{ |
|
"epoch": 9.94, |
|
"learning_rate": 3.0497102775236355e-07, |
|
"loss": 0.0756, |
|
"step": 10870 |
|
}, |
|
{ |
|
"epoch": 9.95, |
|
"learning_rate": 2.541425231269696e-07, |
|
"loss": 0.0867, |
|
"step": 10880 |
|
}, |
|
{ |
|
"epoch": 9.96, |
|
"learning_rate": 2.0331401850157567e-07, |
|
"loss": 0.0846, |
|
"step": 10890 |
|
}, |
|
{ |
|
"epoch": 9.97, |
|
"learning_rate": 1.5248551387618177e-07, |
|
"loss": 0.0767, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 9.97, |
|
"learning_rate": 1.0165700925078784e-07, |
|
"loss": 0.087, |
|
"step": 10910 |
|
}, |
|
{ |
|
"epoch": 9.98, |
|
"learning_rate": 5.082850462539392e-08, |
|
"loss": 0.0692, |
|
"step": 10920 |
|
}, |
|
{ |
|
"epoch": 9.99, |
|
"learning_rate": 0.0, |
|
"loss": 0.0721, |
|
"step": 10930 |
|
}, |
|
{ |
|
"epoch": 9.99, |
|
"eval_accuracy": 0.9817, |
|
"eval_loss": 0.05116609111428261, |
|
"eval_runtime": 101.3542, |
|
"eval_samples_per_second": 295.992, |
|
"eval_steps_per_second": 18.499, |
|
"step": 10930 |
|
}, |
|
{ |
|
"epoch": 9.99, |
|
"step": 10930, |
|
"total_flos": 1.7577285640071414e+19, |
|
"train_loss": 0.14659581507222977, |
|
"train_runtime": 6024.3375, |
|
"train_samples_per_second": 116.195, |
|
"train_steps_per_second": 1.814 |
|
} |
|
], |
|
"max_steps": 10930, |
|
"num_train_epochs": 10, |
|
"total_flos": 1.7577285640071414e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|