diff --git "a/idefics2/checkpoint-9000/trainer_state.json" "b/idefics2/checkpoint-9000/trainer_state.json" new file mode 100644--- /dev/null +++ "b/idefics2/checkpoint-9000/trainer_state.json" @@ -0,0 +1,6405 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 3.1212068666551067, + "eval_steps": 1000, + "global_step": 9000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0034680076296167853, + "grad_norm": 0.7528897523880005, + "learning_rate": 0.00019982, + "loss": 4.9086, + "step": 10 + }, + { + "epoch": 0.006936015259233571, + "grad_norm": 0.17007838189601898, + "learning_rate": 0.00019962000000000002, + "loss": 0.1029, + "step": 20 + }, + { + "epoch": 0.010404022888850355, + "grad_norm": 0.12162350863218307, + "learning_rate": 0.00019942, + "loss": 0.0687, + "step": 30 + }, + { + "epoch": 0.013872030518467141, + "grad_norm": 0.08915918320417404, + "learning_rate": 0.00019922, + "loss": 0.0636, + "step": 40 + }, + { + "epoch": 0.017340038148083926, + "grad_norm": 0.11356709897518158, + "learning_rate": 0.00019902, + "loss": 0.0623, + "step": 50 + }, + { + "epoch": 0.02080804577770071, + "grad_norm": 0.09400122612714767, + "learning_rate": 0.00019882, + "loss": 0.0561, + "step": 60 + }, + { + "epoch": 0.024276053407317495, + "grad_norm": 0.1102517619729042, + "learning_rate": 0.00019862000000000002, + "loss": 0.0603, + "step": 70 + }, + { + "epoch": 0.027744061036934282, + "grad_norm": 0.09301582723855972, + "learning_rate": 0.00019842000000000001, + "loss": 0.0559, + "step": 80 + }, + { + "epoch": 0.031212068666551067, + "grad_norm": 0.08400452882051468, + "learning_rate": 0.00019822, + "loss": 0.0645, + "step": 90 + }, + { + "epoch": 0.03468007629616785, + "grad_norm": 0.10444998741149902, + "learning_rate": 0.00019802, + "loss": 0.0644, + "step": 100 + }, + { + "epoch": 0.038148083925784636, + "grad_norm": 0.06524047255516052, + "learning_rate": 0.00019782, + "loss": 0.0599, + "step": 110 + }, + { + "epoch": 0.04161609155540142, + "grad_norm": 0.0640910267829895, + "learning_rate": 0.00019762, + "loss": 0.0542, + "step": 120 + }, + { + "epoch": 0.045084099185018205, + "grad_norm": 0.07332012802362442, + "learning_rate": 0.00019742000000000002, + "loss": 0.0567, + "step": 130 + }, + { + "epoch": 0.04855210681463499, + "grad_norm": 4.02905797958374, + "learning_rate": 0.00019722, + "loss": 0.0592, + "step": 140 + }, + { + "epoch": 0.05202011444425178, + "grad_norm": 0.09805350750684738, + "learning_rate": 0.00019702, + "loss": 0.0657, + "step": 150 + }, + { + "epoch": 0.055488122073868565, + "grad_norm": 0.056836508214473724, + "learning_rate": 0.00019682, + "loss": 0.0523, + "step": 160 + }, + { + "epoch": 0.05895612970348535, + "grad_norm": 0.08670804649591446, + "learning_rate": 0.00019662, + "loss": 0.0585, + "step": 170 + }, + { + "epoch": 0.062424137333102134, + "grad_norm": 0.08405158668756485, + "learning_rate": 0.00019642, + "loss": 0.0537, + "step": 180 + }, + { + "epoch": 0.06589214496271892, + "grad_norm": 0.0825357437133789, + "learning_rate": 0.00019622000000000002, + "loss": 0.055, + "step": 190 + }, + { + "epoch": 0.0693601525923357, + "grad_norm": 0.06040720269083977, + "learning_rate": 0.00019602, + "loss": 0.0554, + "step": 200 + }, + { + "epoch": 0.07282816022195249, + "grad_norm": 0.08198798447847366, + "learning_rate": 0.00019582, + "loss": 0.0599, + "step": 210 + }, + { + "epoch": 0.07629616785156927, + "grad_norm": 0.05527138710021973, + "learning_rate": 0.00019562, + "loss": 0.0533, + "step": 220 + }, + { + "epoch": 0.07976417548118606, + "grad_norm": 0.07315631955862045, + "learning_rate": 0.00019542, + "loss": 0.0528, + "step": 230 + }, + { + "epoch": 0.08323218311080284, + "grad_norm": 0.08420062810182571, + "learning_rate": 0.00019522, + "loss": 0.0518, + "step": 240 + }, + { + "epoch": 0.08670019074041962, + "grad_norm": 0.0529639795422554, + "learning_rate": 0.00019502, + "loss": 0.0568, + "step": 250 + }, + { + "epoch": 0.09016819837003641, + "grad_norm": 0.09339221566915512, + "learning_rate": 0.00019482, + "loss": 0.0557, + "step": 260 + }, + { + "epoch": 0.0936362059996532, + "grad_norm": 0.06453025341033936, + "learning_rate": 0.00019462, + "loss": 0.062, + "step": 270 + }, + { + "epoch": 0.09710421362926998, + "grad_norm": 0.07090363651514053, + "learning_rate": 0.00019442, + "loss": 0.0609, + "step": 280 + }, + { + "epoch": 0.10057222125888678, + "grad_norm": 0.1122497022151947, + "learning_rate": 0.00019422, + "loss": 0.0541, + "step": 290 + }, + { + "epoch": 0.10404022888850356, + "grad_norm": 0.09782398492097855, + "learning_rate": 0.00019402, + "loss": 0.0553, + "step": 300 + }, + { + "epoch": 0.10750823651812035, + "grad_norm": 0.06216060370206833, + "learning_rate": 0.00019382, + "loss": 0.0483, + "step": 310 + }, + { + "epoch": 0.11097624414773713, + "grad_norm": 0.08817891031503677, + "learning_rate": 0.00019362, + "loss": 0.058, + "step": 320 + }, + { + "epoch": 0.11444425177735391, + "grad_norm": 0.07251620292663574, + "learning_rate": 0.00019342, + "loss": 0.0477, + "step": 330 + }, + { + "epoch": 0.1179122594069707, + "grad_norm": 0.059537626802921295, + "learning_rate": 0.00019322, + "loss": 0.0559, + "step": 340 + }, + { + "epoch": 0.12138026703658748, + "grad_norm": 0.06329932063817978, + "learning_rate": 0.00019302, + "loss": 0.0557, + "step": 350 + }, + { + "epoch": 0.12484827466620427, + "grad_norm": 0.06601905822753906, + "learning_rate": 0.00019282000000000001, + "loss": 0.0575, + "step": 360 + }, + { + "epoch": 0.12831628229582104, + "grad_norm": 0.0639985054731369, + "learning_rate": 0.00019262, + "loss": 0.0519, + "step": 370 + }, + { + "epoch": 0.13178428992543784, + "grad_norm": 0.08187698572874069, + "learning_rate": 0.00019242, + "loss": 0.0545, + "step": 380 + }, + { + "epoch": 0.13525229755505463, + "grad_norm": 0.06790990382432938, + "learning_rate": 0.00019222, + "loss": 0.0474, + "step": 390 + }, + { + "epoch": 0.1387203051846714, + "grad_norm": 0.04906002804636955, + "learning_rate": 0.00019202, + "loss": 0.0579, + "step": 400 + }, + { + "epoch": 0.1421883128142882, + "grad_norm": 0.06125594303011894, + "learning_rate": 0.00019182, + "loss": 0.0421, + "step": 410 + }, + { + "epoch": 0.14565632044390497, + "grad_norm": 0.06252908706665039, + "learning_rate": 0.00019162, + "loss": 0.0577, + "step": 420 + }, + { + "epoch": 0.14912432807352177, + "grad_norm": 0.09376012533903122, + "learning_rate": 0.00019142, + "loss": 0.0546, + "step": 430 + }, + { + "epoch": 0.15259233570313854, + "grad_norm": 0.05025137588381767, + "learning_rate": 0.00019122, + "loss": 0.0505, + "step": 440 + }, + { + "epoch": 0.15606034333275534, + "grad_norm": 0.07034559547901154, + "learning_rate": 0.00019102, + "loss": 0.0495, + "step": 450 + }, + { + "epoch": 0.1595283509623721, + "grad_norm": 0.07475865632295609, + "learning_rate": 0.00019082, + "loss": 0.0533, + "step": 460 + }, + { + "epoch": 0.1629963585919889, + "grad_norm": 0.07459491491317749, + "learning_rate": 0.00019062, + "loss": 0.0551, + "step": 470 + }, + { + "epoch": 0.16646436622160568, + "grad_norm": 0.052578963339328766, + "learning_rate": 0.00019042, + "loss": 0.0564, + "step": 480 + }, + { + "epoch": 0.16993237385122248, + "grad_norm": 0.05785336345434189, + "learning_rate": 0.00019022, + "loss": 0.0549, + "step": 490 + }, + { + "epoch": 0.17340038148083925, + "grad_norm": 0.06621215492486954, + "learning_rate": 0.00019002, + "loss": 0.0528, + "step": 500 + }, + { + "epoch": 0.17686838911045605, + "grad_norm": 0.07456778734922409, + "learning_rate": 0.00018982000000000002, + "loss": 0.0496, + "step": 510 + }, + { + "epoch": 0.18033639674007282, + "grad_norm": 0.06371001899242401, + "learning_rate": 0.00018962000000000002, + "loss": 0.0509, + "step": 520 + }, + { + "epoch": 0.18380440436968962, + "grad_norm": 0.052791863679885864, + "learning_rate": 0.00018942, + "loss": 0.0504, + "step": 530 + }, + { + "epoch": 0.1872724119993064, + "grad_norm": 0.06491260975599289, + "learning_rate": 0.00018922, + "loss": 0.0527, + "step": 540 + }, + { + "epoch": 0.19074041962892319, + "grad_norm": 0.07835149765014648, + "learning_rate": 0.00018902000000000003, + "loss": 0.0564, + "step": 550 + }, + { + "epoch": 0.19420842725853996, + "grad_norm": 0.05977100506424904, + "learning_rate": 0.00018882000000000003, + "loss": 0.0477, + "step": 560 + }, + { + "epoch": 0.19767643488815675, + "grad_norm": 0.07109620422124863, + "learning_rate": 0.00018862000000000002, + "loss": 0.0537, + "step": 570 + }, + { + "epoch": 0.20114444251777355, + "grad_norm": 0.10347943007946014, + "learning_rate": 0.00018842000000000002, + "loss": 0.0542, + "step": 580 + }, + { + "epoch": 0.20461245014739032, + "grad_norm": 0.06503281742334366, + "learning_rate": 0.00018822, + "loss": 0.0484, + "step": 590 + }, + { + "epoch": 0.20808045777700712, + "grad_norm": 0.07903438806533813, + "learning_rate": 0.00018802, + "loss": 0.0597, + "step": 600 + }, + { + "epoch": 0.2115484654066239, + "grad_norm": 0.0713895708322525, + "learning_rate": 0.00018782000000000003, + "loss": 0.0498, + "step": 610 + }, + { + "epoch": 0.2150164730362407, + "grad_norm": 0.061313629150390625, + "learning_rate": 0.00018762000000000002, + "loss": 0.053, + "step": 620 + }, + { + "epoch": 0.21848448066585746, + "grad_norm": 0.07045572996139526, + "learning_rate": 0.00018742000000000002, + "loss": 0.0541, + "step": 630 + }, + { + "epoch": 0.22195248829547426, + "grad_norm": 0.07118247449398041, + "learning_rate": 0.00018722, + "loss": 0.0586, + "step": 640 + }, + { + "epoch": 0.22542049592509103, + "grad_norm": 0.05364071577787399, + "learning_rate": 0.00018702, + "loss": 0.0497, + "step": 650 + }, + { + "epoch": 0.22888850355470783, + "grad_norm": 0.07208040356636047, + "learning_rate": 0.00018682000000000003, + "loss": 0.0559, + "step": 660 + }, + { + "epoch": 0.2323565111843246, + "grad_norm": 0.07200731337070465, + "learning_rate": 0.00018662000000000003, + "loss": 0.048, + "step": 670 + }, + { + "epoch": 0.2358245188139414, + "grad_norm": 0.05730220302939415, + "learning_rate": 0.00018642000000000002, + "loss": 0.0495, + "step": 680 + }, + { + "epoch": 0.23929252644355817, + "grad_norm": 0.06378819793462753, + "learning_rate": 0.00018622000000000002, + "loss": 0.0552, + "step": 690 + }, + { + "epoch": 0.24276053407317497, + "grad_norm": 0.05866115912795067, + "learning_rate": 0.00018602, + "loss": 0.0541, + "step": 700 + }, + { + "epoch": 0.24622854170279174, + "grad_norm": 0.07053161412477493, + "learning_rate": 0.00018582, + "loss": 0.0533, + "step": 710 + }, + { + "epoch": 0.24969654933240854, + "grad_norm": 0.07450433820486069, + "learning_rate": 0.00018562000000000003, + "loss": 0.0527, + "step": 720 + }, + { + "epoch": 0.25316455696202533, + "grad_norm": 0.07172481715679169, + "learning_rate": 0.00018542000000000002, + "loss": 0.0583, + "step": 730 + }, + { + "epoch": 0.2566325645916421, + "grad_norm": 0.0536239892244339, + "learning_rate": 0.00018522000000000002, + "loss": 0.0486, + "step": 740 + }, + { + "epoch": 0.2601005722212589, + "grad_norm": 0.06321065127849579, + "learning_rate": 0.00018502000000000001, + "loss": 0.0412, + "step": 750 + }, + { + "epoch": 0.2635685798508757, + "grad_norm": 0.056946441531181335, + "learning_rate": 0.00018482, + "loss": 0.051, + "step": 760 + }, + { + "epoch": 0.26703658748049247, + "grad_norm": 0.05819573253393173, + "learning_rate": 0.00018462, + "loss": 0.0517, + "step": 770 + }, + { + "epoch": 0.27050459511010927, + "grad_norm": 0.05857665091753006, + "learning_rate": 0.00018442000000000003, + "loss": 0.0529, + "step": 780 + }, + { + "epoch": 0.273972602739726, + "grad_norm": 0.06014329940080643, + "learning_rate": 0.00018422000000000002, + "loss": 0.0544, + "step": 790 + }, + { + "epoch": 0.2774406103693428, + "grad_norm": 0.09187959134578705, + "learning_rate": 0.00018402000000000002, + "loss": 0.0439, + "step": 800 + }, + { + "epoch": 0.2809086179989596, + "grad_norm": 0.056131429970264435, + "learning_rate": 0.00018382, + "loss": 0.0389, + "step": 810 + }, + { + "epoch": 0.2843766256285764, + "grad_norm": 0.04884343966841698, + "learning_rate": 0.00018362, + "loss": 0.0475, + "step": 820 + }, + { + "epoch": 0.28784463325819315, + "grad_norm": 0.09032488614320755, + "learning_rate": 0.00018342, + "loss": 0.056, + "step": 830 + }, + { + "epoch": 0.29131264088780995, + "grad_norm": 0.09926522523164749, + "learning_rate": 0.00018322000000000002, + "loss": 0.0515, + "step": 840 + }, + { + "epoch": 0.29478064851742675, + "grad_norm": 0.04553750529885292, + "learning_rate": 0.00018302000000000002, + "loss": 0.0505, + "step": 850 + }, + { + "epoch": 0.29824865614704354, + "grad_norm": 0.058485984802246094, + "learning_rate": 0.00018282000000000001, + "loss": 0.0453, + "step": 860 + }, + { + "epoch": 0.3017166637766603, + "grad_norm": 0.052825070917606354, + "learning_rate": 0.00018262, + "loss": 0.053, + "step": 870 + }, + { + "epoch": 0.3051846714062771, + "grad_norm": 0.07626510411500931, + "learning_rate": 0.00018242, + "loss": 0.0474, + "step": 880 + }, + { + "epoch": 0.3086526790358939, + "grad_norm": 0.07244163751602173, + "learning_rate": 0.00018222, + "loss": 0.0456, + "step": 890 + }, + { + "epoch": 0.3121206866655107, + "grad_norm": 0.09651289880275726, + "learning_rate": 0.00018202000000000002, + "loss": 0.0499, + "step": 900 + }, + { + "epoch": 0.3155886942951274, + "grad_norm": 0.0737752839922905, + "learning_rate": 0.00018182000000000002, + "loss": 0.0517, + "step": 910 + }, + { + "epoch": 0.3190567019247442, + "grad_norm": 0.06631263345479965, + "learning_rate": 0.00018162, + "loss": 0.0499, + "step": 920 + }, + { + "epoch": 0.322524709554361, + "grad_norm": 0.10696552693843842, + "learning_rate": 0.00018142, + "loss": 0.0524, + "step": 930 + }, + { + "epoch": 0.3259927171839778, + "grad_norm": 0.06753025203943253, + "learning_rate": 0.00018122, + "loss": 0.0458, + "step": 940 + }, + { + "epoch": 0.32946072481359456, + "grad_norm": 0.05970798432826996, + "learning_rate": 0.00018102000000000003, + "loss": 0.0471, + "step": 950 + }, + { + "epoch": 0.33292873244321136, + "grad_norm": 0.05324438214302063, + "learning_rate": 0.00018082000000000002, + "loss": 0.0468, + "step": 960 + }, + { + "epoch": 0.33639674007282816, + "grad_norm": 0.0749637559056282, + "learning_rate": 0.00018062000000000002, + "loss": 0.0507, + "step": 970 + }, + { + "epoch": 0.33986474770244496, + "grad_norm": 0.07317759841680527, + "learning_rate": 0.00018042, + "loss": 0.0482, + "step": 980 + }, + { + "epoch": 0.34333275533206176, + "grad_norm": 0.11602938175201416, + "learning_rate": 0.00018022, + "loss": 0.0558, + "step": 990 + }, + { + "epoch": 0.3468007629616785, + "grad_norm": 0.07047244161367416, + "learning_rate": 0.00018002, + "loss": 0.0547, + "step": 1000 + }, + { + "epoch": 0.3468007629616785, + "eval_loss": 0.062305010855197906, + "eval_runtime": 714.4647, + "eval_samples_per_second": 12.852, + "eval_steps_per_second": 1.607, + "step": 1000 + }, + { + "epoch": 0.3502687705912953, + "grad_norm": 0.07715447247028351, + "learning_rate": 0.00017982000000000002, + "loss": 0.0445, + "step": 1010 + }, + { + "epoch": 0.3537367782209121, + "grad_norm": 0.08189492672681808, + "learning_rate": 0.00017962000000000002, + "loss": 0.0529, + "step": 1020 + }, + { + "epoch": 0.3572047858505289, + "grad_norm": 0.06011577695608139, + "learning_rate": 0.00017942, + "loss": 0.0469, + "step": 1030 + }, + { + "epoch": 0.36067279348014564, + "grad_norm": 0.06397314369678497, + "learning_rate": 0.00017922, + "loss": 0.0542, + "step": 1040 + }, + { + "epoch": 0.36414080110976244, + "grad_norm": 0.06121763586997986, + "learning_rate": 0.00017902, + "loss": 0.0474, + "step": 1050 + }, + { + "epoch": 0.36760880873937923, + "grad_norm": 0.0765228345990181, + "learning_rate": 0.00017882, + "loss": 0.0476, + "step": 1060 + }, + { + "epoch": 0.37107681636899603, + "grad_norm": 0.0813635066151619, + "learning_rate": 0.00017862000000000002, + "loss": 0.0488, + "step": 1070 + }, + { + "epoch": 0.3745448239986128, + "grad_norm": 0.06827688962221146, + "learning_rate": 0.00017842000000000002, + "loss": 0.0431, + "step": 1080 + }, + { + "epoch": 0.3780128316282296, + "grad_norm": 0.06176091730594635, + "learning_rate": 0.00017822, + "loss": 0.0513, + "step": 1090 + }, + { + "epoch": 0.38148083925784637, + "grad_norm": 0.07062922418117523, + "learning_rate": 0.00017802, + "loss": 0.0499, + "step": 1100 + }, + { + "epoch": 0.38494884688746317, + "grad_norm": 0.059431031346321106, + "learning_rate": 0.00017782, + "loss": 0.0556, + "step": 1110 + }, + { + "epoch": 0.3884168545170799, + "grad_norm": 0.06391894072294235, + "learning_rate": 0.00017762, + "loss": 0.0535, + "step": 1120 + }, + { + "epoch": 0.3918848621466967, + "grad_norm": 0.08487355709075928, + "learning_rate": 0.00017742000000000002, + "loss": 0.0509, + "step": 1130 + }, + { + "epoch": 0.3953528697763135, + "grad_norm": 0.06291911005973816, + "learning_rate": 0.00017722000000000001, + "loss": 0.0462, + "step": 1140 + }, + { + "epoch": 0.3988208774059303, + "grad_norm": 0.06936580687761307, + "learning_rate": 0.00017702, + "loss": 0.0465, + "step": 1150 + }, + { + "epoch": 0.4022888850355471, + "grad_norm": 0.06751543283462524, + "learning_rate": 0.00017682, + "loss": 0.0553, + "step": 1160 + }, + { + "epoch": 0.40575689266516385, + "grad_norm": 0.08026771247386932, + "learning_rate": 0.00017662, + "loss": 0.0503, + "step": 1170 + }, + { + "epoch": 0.40922490029478065, + "grad_norm": 0.05316636711359024, + "learning_rate": 0.00017642, + "loss": 0.0426, + "step": 1180 + }, + { + "epoch": 0.41269290792439745, + "grad_norm": 0.15491995215415955, + "learning_rate": 0.00017622000000000002, + "loss": 0.0481, + "step": 1190 + }, + { + "epoch": 0.41616091555401424, + "grad_norm": 0.059617578983306885, + "learning_rate": 0.00017602, + "loss": 0.0443, + "step": 1200 + }, + { + "epoch": 0.419628923183631, + "grad_norm": 0.08931437879800797, + "learning_rate": 0.00017582, + "loss": 0.0462, + "step": 1210 + }, + { + "epoch": 0.4230969308132478, + "grad_norm": 1.4069609642028809, + "learning_rate": 0.00017562, + "loss": 0.0577, + "step": 1220 + }, + { + "epoch": 0.4265649384428646, + "grad_norm": 7.056313514709473, + "learning_rate": 0.00017542, + "loss": 0.3501, + "step": 1230 + }, + { + "epoch": 0.4300329460724814, + "grad_norm": 0.8197808265686035, + "learning_rate": 0.00017522000000000002, + "loss": 0.1137, + "step": 1240 + }, + { + "epoch": 0.4335009537020981, + "grad_norm": 0.4104197025299072, + "learning_rate": 0.00017502000000000001, + "loss": 0.1134, + "step": 1250 + }, + { + "epoch": 0.4369689613317149, + "grad_norm": 1.3801881074905396, + "learning_rate": 0.00017482, + "loss": 0.0869, + "step": 1260 + }, + { + "epoch": 0.4404369689613317, + "grad_norm": 3.6437034606933594, + "learning_rate": 0.00017462, + "loss": 0.0544, + "step": 1270 + }, + { + "epoch": 0.4439049765909485, + "grad_norm": 1.1371792554855347, + "learning_rate": 0.00017442, + "loss": 0.0929, + "step": 1280 + }, + { + "epoch": 0.44737298422056526, + "grad_norm": 2.9385204315185547, + "learning_rate": 0.00017422, + "loss": 0.1166, + "step": 1290 + }, + { + "epoch": 0.45084099185018206, + "grad_norm": 0.8243001699447632, + "learning_rate": 0.00017402000000000002, + "loss": 0.0907, + "step": 1300 + }, + { + "epoch": 0.45430899947979886, + "grad_norm": 1.3375756740570068, + "learning_rate": 0.00017382, + "loss": 0.0641, + "step": 1310 + }, + { + "epoch": 0.45777700710941566, + "grad_norm": 0.7172356843948364, + "learning_rate": 0.00017362, + "loss": 0.0575, + "step": 1320 + }, + { + "epoch": 0.4612450147390324, + "grad_norm": 0.18255668878555298, + "learning_rate": 0.00017342, + "loss": 0.0628, + "step": 1330 + }, + { + "epoch": 0.4647130223686492, + "grad_norm": 0.11856569349765778, + "learning_rate": 0.00017322, + "loss": 0.1111, + "step": 1340 + }, + { + "epoch": 0.468181029998266, + "grad_norm": 0.16621063649654388, + "learning_rate": 0.00017302, + "loss": 0.0446, + "step": 1350 + }, + { + "epoch": 0.4716490376278828, + "grad_norm": 0.10158076882362366, + "learning_rate": 0.00017282000000000002, + "loss": 0.0532, + "step": 1360 + }, + { + "epoch": 0.4751170452574996, + "grad_norm": 0.08783379942178726, + "learning_rate": 0.00017262, + "loss": 0.0491, + "step": 1370 + }, + { + "epoch": 0.47858505288711634, + "grad_norm": 0.08198387920856476, + "learning_rate": 0.00017242, + "loss": 0.0485, + "step": 1380 + }, + { + "epoch": 0.48205306051673313, + "grad_norm": 0.09546195715665817, + "learning_rate": 0.00017222, + "loss": 0.0486, + "step": 1390 + }, + { + "epoch": 0.48552106814634993, + "grad_norm": 0.13258413970470428, + "learning_rate": 0.00017202, + "loss": 0.0544, + "step": 1400 + }, + { + "epoch": 0.48898907577596673, + "grad_norm": 0.09508573263883591, + "learning_rate": 0.00017182, + "loss": 0.0486, + "step": 1410 + }, + { + "epoch": 0.4924570834055835, + "grad_norm": 0.07258226722478867, + "learning_rate": 0.00017162000000000001, + "loss": 0.0448, + "step": 1420 + }, + { + "epoch": 0.49592509103520027, + "grad_norm": 0.05399150773882866, + "learning_rate": 0.00017142, + "loss": 0.0436, + "step": 1430 + }, + { + "epoch": 0.49939309866481707, + "grad_norm": 0.06922592222690582, + "learning_rate": 0.00017122, + "loss": 0.0451, + "step": 1440 + }, + { + "epoch": 0.5028611062944338, + "grad_norm": 0.05059856176376343, + "learning_rate": 0.00017102, + "loss": 0.0413, + "step": 1450 + }, + { + "epoch": 0.5063291139240507, + "grad_norm": 0.06339547783136368, + "learning_rate": 0.00017082, + "loss": 0.0495, + "step": 1460 + }, + { + "epoch": 0.5097971215536674, + "grad_norm": 0.0521874763071537, + "learning_rate": 0.00017062, + "loss": 0.0489, + "step": 1470 + }, + { + "epoch": 0.5132651291832842, + "grad_norm": 0.15334908664226532, + "learning_rate": 0.00017042, + "loss": 0.04, + "step": 1480 + }, + { + "epoch": 0.516733136812901, + "grad_norm": 0.07494404166936874, + "learning_rate": 0.00017022, + "loss": 0.0537, + "step": 1490 + }, + { + "epoch": 0.5202011444425177, + "grad_norm": 0.052238237112760544, + "learning_rate": 0.00017002, + "loss": 0.0492, + "step": 1500 + }, + { + "epoch": 0.5236691520721346, + "grad_norm": 0.06761351227760315, + "learning_rate": 0.00016982, + "loss": 0.051, + "step": 1510 + }, + { + "epoch": 0.5271371597017513, + "grad_norm": 0.07580805569887161, + "learning_rate": 0.00016962, + "loss": 0.0498, + "step": 1520 + }, + { + "epoch": 0.5306051673313681, + "grad_norm": 0.07397795468568802, + "learning_rate": 0.00016942000000000001, + "loss": 0.0471, + "step": 1530 + }, + { + "epoch": 0.5340731749609849, + "grad_norm": 0.04779529199004173, + "learning_rate": 0.00016922, + "loss": 0.0465, + "step": 1540 + }, + { + "epoch": 0.5375411825906017, + "grad_norm": 0.08508727699518204, + "learning_rate": 0.00016902, + "loss": 0.046, + "step": 1550 + }, + { + "epoch": 0.5410091902202185, + "grad_norm": 0.06656762957572937, + "learning_rate": 0.00016882, + "loss": 0.0433, + "step": 1560 + }, + { + "epoch": 0.5444771978498353, + "grad_norm": 0.05622195079922676, + "learning_rate": 0.00016862, + "loss": 0.0455, + "step": 1570 + }, + { + "epoch": 0.547945205479452, + "grad_norm": 0.06422954052686691, + "learning_rate": 0.00016842, + "loss": 0.0428, + "step": 1580 + }, + { + "epoch": 0.5514132131090689, + "grad_norm": 0.08891351521015167, + "learning_rate": 0.00016822, + "loss": 0.0495, + "step": 1590 + }, + { + "epoch": 0.5548812207386856, + "grad_norm": 0.08472294360399246, + "learning_rate": 0.00016802, + "loss": 0.0525, + "step": 1600 + }, + { + "epoch": 0.5583492283683024, + "grad_norm": 0.08518495410680771, + "learning_rate": 0.00016782, + "loss": 0.0518, + "step": 1610 + }, + { + "epoch": 0.5618172359979192, + "grad_norm": 0.1859533041715622, + "learning_rate": 0.00016762, + "loss": 0.0478, + "step": 1620 + }, + { + "epoch": 0.565285243627536, + "grad_norm": 0.05245356634259224, + "learning_rate": 0.00016742, + "loss": 0.0418, + "step": 1630 + }, + { + "epoch": 0.5687532512571528, + "grad_norm": 0.05982668697834015, + "learning_rate": 0.00016722, + "loss": 0.0458, + "step": 1640 + }, + { + "epoch": 0.5722212588867696, + "grad_norm": 0.0750059187412262, + "learning_rate": 0.00016702, + "loss": 0.0489, + "step": 1650 + }, + { + "epoch": 0.5756892665163863, + "grad_norm": 0.061683133244514465, + "learning_rate": 0.00016682, + "loss": 0.0461, + "step": 1660 + }, + { + "epoch": 0.5791572741460032, + "grad_norm": 0.06833604723215103, + "learning_rate": 0.00016662, + "loss": 0.0466, + "step": 1670 + }, + { + "epoch": 0.5826252817756199, + "grad_norm": 0.05507722124457359, + "learning_rate": 0.00016642, + "loss": 0.0416, + "step": 1680 + }, + { + "epoch": 0.5860932894052366, + "grad_norm": 0.09453442692756653, + "learning_rate": 0.00016622, + "loss": 0.0418, + "step": 1690 + }, + { + "epoch": 0.5895612970348535, + "grad_norm": 0.0574457123875618, + "learning_rate": 0.00016601999999999999, + "loss": 0.0495, + "step": 1700 + }, + { + "epoch": 0.5930293046644702, + "grad_norm": 0.0829281136393547, + "learning_rate": 0.00016582, + "loss": 0.0435, + "step": 1710 + }, + { + "epoch": 0.5964973122940871, + "grad_norm": 0.0569952018558979, + "learning_rate": 0.00016562, + "loss": 0.0476, + "step": 1720 + }, + { + "epoch": 0.5999653199237038, + "grad_norm": 0.09291055798530579, + "learning_rate": 0.00016542, + "loss": 0.0446, + "step": 1730 + }, + { + "epoch": 0.6034333275533206, + "grad_norm": 0.03767919912934303, + "learning_rate": 0.00016522, + "loss": 0.0426, + "step": 1740 + }, + { + "epoch": 0.6069013351829374, + "grad_norm": 0.07064680010080338, + "learning_rate": 0.00016502, + "loss": 0.0491, + "step": 1750 + }, + { + "epoch": 0.6103693428125542, + "grad_norm": 0.05687623471021652, + "learning_rate": 0.00016482, + "loss": 0.0454, + "step": 1760 + }, + { + "epoch": 0.613837350442171, + "grad_norm": 0.04933289438486099, + "learning_rate": 0.00016462, + "loss": 0.0385, + "step": 1770 + }, + { + "epoch": 0.6173053580717878, + "grad_norm": 0.0763295516371727, + "learning_rate": 0.00016442000000000003, + "loss": 0.0448, + "step": 1780 + }, + { + "epoch": 0.6207733657014045, + "grad_norm": 0.04926356300711632, + "learning_rate": 0.00016422000000000002, + "loss": 0.0484, + "step": 1790 + }, + { + "epoch": 0.6242413733310214, + "grad_norm": 0.07130167633295059, + "learning_rate": 0.00016402000000000002, + "loss": 0.0403, + "step": 1800 + }, + { + "epoch": 0.6277093809606381, + "grad_norm": 0.06025327742099762, + "learning_rate": 0.00016382000000000001, + "loss": 0.0483, + "step": 1810 + }, + { + "epoch": 0.6311773885902549, + "grad_norm": 0.06522911787033081, + "learning_rate": 0.00016362, + "loss": 0.0433, + "step": 1820 + }, + { + "epoch": 0.6346453962198717, + "grad_norm": 0.09565310180187225, + "learning_rate": 0.00016342, + "loss": 0.0539, + "step": 1830 + }, + { + "epoch": 0.6381134038494884, + "grad_norm": 0.08908990025520325, + "learning_rate": 0.00016322000000000003, + "loss": 0.0479, + "step": 1840 + }, + { + "epoch": 0.6415814114791053, + "grad_norm": 0.05405285581946373, + "learning_rate": 0.00016302000000000002, + "loss": 0.0416, + "step": 1850 + }, + { + "epoch": 0.645049419108722, + "grad_norm": 0.0656275674700737, + "learning_rate": 0.00016282000000000002, + "loss": 0.0455, + "step": 1860 + }, + { + "epoch": 0.6485174267383388, + "grad_norm": 0.30592814087867737, + "learning_rate": 0.00016262, + "loss": 0.0471, + "step": 1870 + }, + { + "epoch": 0.6519854343679556, + "grad_norm": 0.06433047354221344, + "learning_rate": 0.00016242, + "loss": 0.045, + "step": 1880 + }, + { + "epoch": 0.6554534419975724, + "grad_norm": 0.05807631090283394, + "learning_rate": 0.00016222000000000003, + "loss": 0.0502, + "step": 1890 + }, + { + "epoch": 0.6589214496271891, + "grad_norm": 0.09822454303503036, + "learning_rate": 0.00016202000000000002, + "loss": 0.0465, + "step": 1900 + }, + { + "epoch": 0.662389457256806, + "grad_norm": 0.06036192551255226, + "learning_rate": 0.00016182000000000002, + "loss": 0.0455, + "step": 1910 + }, + { + "epoch": 0.6658574648864227, + "grad_norm": 0.05637621134519577, + "learning_rate": 0.00016162000000000001, + "loss": 0.0459, + "step": 1920 + }, + { + "epoch": 0.6693254725160396, + "grad_norm": 0.062302861362695694, + "learning_rate": 0.00016142, + "loss": 0.0461, + "step": 1930 + }, + { + "epoch": 0.6727934801456563, + "grad_norm": 0.05843142420053482, + "learning_rate": 0.00016122, + "loss": 0.0455, + "step": 1940 + }, + { + "epoch": 0.6762614877752731, + "grad_norm": 0.052565112709999084, + "learning_rate": 0.00016102000000000003, + "loss": 0.0432, + "step": 1950 + }, + { + "epoch": 0.6797294954048899, + "grad_norm": 0.059093691408634186, + "learning_rate": 0.00016082000000000002, + "loss": 0.0433, + "step": 1960 + }, + { + "epoch": 0.6831975030345067, + "grad_norm": 0.07197156548500061, + "learning_rate": 0.00016062000000000002, + "loss": 0.0465, + "step": 1970 + }, + { + "epoch": 0.6866655106641235, + "grad_norm": 0.039968665689229965, + "learning_rate": 0.00016042, + "loss": 0.0381, + "step": 1980 + }, + { + "epoch": 0.6901335182937403, + "grad_norm": 0.05672089383006096, + "learning_rate": 0.00016022, + "loss": 0.0422, + "step": 1990 + }, + { + "epoch": 0.693601525923357, + "grad_norm": 0.10138531029224396, + "learning_rate": 0.00016002, + "loss": 0.0464, + "step": 2000 + }, + { + "epoch": 0.693601525923357, + "eval_loss": 0.06039171665906906, + "eval_runtime": 710.8114, + "eval_samples_per_second": 12.918, + "eval_steps_per_second": 1.615, + "step": 2000 + }, + { + "epoch": 0.6970695335529739, + "grad_norm": 0.0607437863945961, + "learning_rate": 0.00015982000000000002, + "loss": 0.0481, + "step": 2010 + }, + { + "epoch": 0.7005375411825906, + "grad_norm": 0.06681676208972931, + "learning_rate": 0.00015962000000000002, + "loss": 0.049, + "step": 2020 + }, + { + "epoch": 0.7040055488122073, + "grad_norm": 0.07225602120161057, + "learning_rate": 0.00015942000000000002, + "loss": 0.0465, + "step": 2030 + }, + { + "epoch": 0.7074735564418242, + "grad_norm": 0.056885506957769394, + "learning_rate": 0.00015922, + "loss": 0.0452, + "step": 2040 + }, + { + "epoch": 0.7109415640714409, + "grad_norm": 0.05572199076414108, + "learning_rate": 0.00015902, + "loss": 0.0423, + "step": 2050 + }, + { + "epoch": 0.7144095717010578, + "grad_norm": 0.0516788586974144, + "learning_rate": 0.00015882, + "loss": 0.0404, + "step": 2060 + }, + { + "epoch": 0.7178775793306745, + "grad_norm": 0.05214313417673111, + "learning_rate": 0.00015862000000000002, + "loss": 0.0428, + "step": 2070 + }, + { + "epoch": 0.7213455869602913, + "grad_norm": 0.06317329406738281, + "learning_rate": 0.00015842000000000002, + "loss": 0.0447, + "step": 2080 + }, + { + "epoch": 0.7248135945899081, + "grad_norm": 0.07574247568845749, + "learning_rate": 0.00015822, + "loss": 0.0395, + "step": 2090 + }, + { + "epoch": 0.7282816022195249, + "grad_norm": 0.06367363035678864, + "learning_rate": 0.00015802, + "loss": 0.0439, + "step": 2100 + }, + { + "epoch": 0.7317496098491416, + "grad_norm": 0.06190785393118858, + "learning_rate": 0.00015782, + "loss": 0.0404, + "step": 2110 + }, + { + "epoch": 0.7352176174787585, + "grad_norm": 0.07103675603866577, + "learning_rate": 0.00015762, + "loss": 0.0468, + "step": 2120 + }, + { + "epoch": 0.7386856251083752, + "grad_norm": 0.04957522451877594, + "learning_rate": 0.00015742000000000002, + "loss": 0.0453, + "step": 2130 + }, + { + "epoch": 0.7421536327379921, + "grad_norm": 0.07643826305866241, + "learning_rate": 0.00015722000000000002, + "loss": 0.0404, + "step": 2140 + }, + { + "epoch": 0.7456216403676088, + "grad_norm": 0.04660920053720474, + "learning_rate": 0.00015702, + "loss": 0.0373, + "step": 2150 + }, + { + "epoch": 0.7490896479972255, + "grad_norm": 0.06206243112683296, + "learning_rate": 0.00015682, + "loss": 0.0406, + "step": 2160 + }, + { + "epoch": 0.7525576556268424, + "grad_norm": 0.07452013343572617, + "learning_rate": 0.00015662, + "loss": 0.043, + "step": 2170 + }, + { + "epoch": 0.7560256632564591, + "grad_norm": 0.04379798844456673, + "learning_rate": 0.00015642000000000002, + "loss": 0.0494, + "step": 2180 + }, + { + "epoch": 0.759493670886076, + "grad_norm": 0.05709415674209595, + "learning_rate": 0.00015622000000000002, + "loss": 0.041, + "step": 2190 + }, + { + "epoch": 0.7629616785156927, + "grad_norm": 0.07468123733997345, + "learning_rate": 0.00015602000000000001, + "loss": 0.0403, + "step": 2200 + }, + { + "epoch": 0.7664296861453095, + "grad_norm": 0.08100239932537079, + "learning_rate": 0.00015582, + "loss": 0.0481, + "step": 2210 + }, + { + "epoch": 0.7698976937749263, + "grad_norm": 0.07269110530614853, + "learning_rate": 0.00015562, + "loss": 0.0394, + "step": 2220 + }, + { + "epoch": 0.7733657014045431, + "grad_norm": 0.060352034866809845, + "learning_rate": 0.00015542, + "loss": 0.04, + "step": 2230 + }, + { + "epoch": 0.7768337090341598, + "grad_norm": 0.05698138475418091, + "learning_rate": 0.00015522000000000002, + "loss": 0.0383, + "step": 2240 + }, + { + "epoch": 0.7803017166637767, + "grad_norm": 0.06030441075563431, + "learning_rate": 0.00015502000000000002, + "loss": 0.0456, + "step": 2250 + }, + { + "epoch": 0.7837697242933934, + "grad_norm": 0.08272086083889008, + "learning_rate": 0.00015482, + "loss": 0.0443, + "step": 2260 + }, + { + "epoch": 0.7872377319230103, + "grad_norm": 0.06746231019496918, + "learning_rate": 0.00015462, + "loss": 0.04, + "step": 2270 + }, + { + "epoch": 0.790705739552627, + "grad_norm": 0.08844996243715286, + "learning_rate": 0.00015442, + "loss": 0.0418, + "step": 2280 + }, + { + "epoch": 0.7941737471822438, + "grad_norm": 0.04906987398862839, + "learning_rate": 0.00015422, + "loss": 0.0452, + "step": 2290 + }, + { + "epoch": 0.7976417548118606, + "grad_norm": 0.053966376930475235, + "learning_rate": 0.00015402000000000002, + "loss": 0.0404, + "step": 2300 + }, + { + "epoch": 0.8011097624414774, + "grad_norm": 0.07681586593389511, + "learning_rate": 0.00015382000000000001, + "loss": 0.05, + "step": 2310 + }, + { + "epoch": 0.8045777700710942, + "grad_norm": 0.059431836009025574, + "learning_rate": 0.00015362, + "loss": 0.0496, + "step": 2320 + }, + { + "epoch": 0.808045777700711, + "grad_norm": 0.058449339121580124, + "learning_rate": 0.00015342, + "loss": 0.0419, + "step": 2330 + }, + { + "epoch": 0.8115137853303277, + "grad_norm": 0.05404651537537575, + "learning_rate": 0.00015322, + "loss": 0.0382, + "step": 2340 + }, + { + "epoch": 0.8149817929599446, + "grad_norm": 0.06095472350716591, + "learning_rate": 0.00015302, + "loss": 0.0487, + "step": 2350 + }, + { + "epoch": 0.8184498005895613, + "grad_norm": 0.07456117123365402, + "learning_rate": 0.00015282000000000002, + "loss": 0.0386, + "step": 2360 + }, + { + "epoch": 0.821917808219178, + "grad_norm": 0.06098613142967224, + "learning_rate": 0.00015262, + "loss": 0.0425, + "step": 2370 + }, + { + "epoch": 0.8253858158487949, + "grad_norm": 0.07008852809667587, + "learning_rate": 0.00015242, + "loss": 0.0457, + "step": 2380 + }, + { + "epoch": 0.8288538234784116, + "grad_norm": 0.06413611769676208, + "learning_rate": 0.00015222, + "loss": 0.0433, + "step": 2390 + }, + { + "epoch": 0.8323218311080285, + "grad_norm": 0.05168429762125015, + "learning_rate": 0.00015202, + "loss": 0.0441, + "step": 2400 + }, + { + "epoch": 0.8357898387376452, + "grad_norm": 0.0708162784576416, + "learning_rate": 0.00015182, + "loss": 0.0393, + "step": 2410 + }, + { + "epoch": 0.839257846367262, + "grad_norm": 0.05856487527489662, + "learning_rate": 0.00015162000000000002, + "loss": 0.0437, + "step": 2420 + }, + { + "epoch": 0.8427258539968788, + "grad_norm": 0.06069020926952362, + "learning_rate": 0.00015142, + "loss": 0.0484, + "step": 2430 + }, + { + "epoch": 0.8461938616264956, + "grad_norm": 0.0641227513551712, + "learning_rate": 0.00015122, + "loss": 0.046, + "step": 2440 + }, + { + "epoch": 0.8496618692561123, + "grad_norm": 0.06422239542007446, + "learning_rate": 0.00015102, + "loss": 0.043, + "step": 2450 + }, + { + "epoch": 0.8531298768857292, + "grad_norm": 0.07488572597503662, + "learning_rate": 0.00015082, + "loss": 0.0517, + "step": 2460 + }, + { + "epoch": 0.8565978845153459, + "grad_norm": 0.06784242391586304, + "learning_rate": 0.00015062000000000002, + "loss": 0.0437, + "step": 2470 + }, + { + "epoch": 0.8600658921449628, + "grad_norm": 0.07483550906181335, + "learning_rate": 0.00015042, + "loss": 0.0491, + "step": 2480 + }, + { + "epoch": 0.8635338997745795, + "grad_norm": 0.06362838298082352, + "learning_rate": 0.00015022, + "loss": 0.043, + "step": 2490 + }, + { + "epoch": 0.8670019074041962, + "grad_norm": 0.06175532937049866, + "learning_rate": 0.00015002, + "loss": 0.0425, + "step": 2500 + }, + { + "epoch": 0.8704699150338131, + "grad_norm": 0.08987358957529068, + "learning_rate": 0.00014982, + "loss": 0.0479, + "step": 2510 + }, + { + "epoch": 0.8739379226634298, + "grad_norm": 0.05209062620997429, + "learning_rate": 0.00014962, + "loss": 0.045, + "step": 2520 + }, + { + "epoch": 0.8774059302930467, + "grad_norm": 0.055044736713171005, + "learning_rate": 0.00014942000000000002, + "loss": 0.0469, + "step": 2530 + }, + { + "epoch": 0.8808739379226634, + "grad_norm": 0.0646812692284584, + "learning_rate": 0.00014922, + "loss": 0.0426, + "step": 2540 + }, + { + "epoch": 0.8843419455522802, + "grad_norm": 0.048389263451099396, + "learning_rate": 0.00014902, + "loss": 0.0417, + "step": 2550 + }, + { + "epoch": 0.887809953181897, + "grad_norm": 0.07316736876964569, + "learning_rate": 0.00014882, + "loss": 0.0412, + "step": 2560 + }, + { + "epoch": 0.8912779608115138, + "grad_norm": 0.06801818311214447, + "learning_rate": 0.00014862, + "loss": 0.0446, + "step": 2570 + }, + { + "epoch": 0.8947459684411305, + "grad_norm": 0.08190831542015076, + "learning_rate": 0.00014842, + "loss": 0.0423, + "step": 2580 + }, + { + "epoch": 0.8982139760707474, + "grad_norm": 0.05850045010447502, + "learning_rate": 0.00014822000000000001, + "loss": 0.0418, + "step": 2590 + }, + { + "epoch": 0.9016819837003641, + "grad_norm": 0.07633431255817413, + "learning_rate": 0.00014802, + "loss": 0.0439, + "step": 2600 + }, + { + "epoch": 0.905149991329981, + "grad_norm": 0.07250861078500748, + "learning_rate": 0.00014782, + "loss": 0.0456, + "step": 2610 + }, + { + "epoch": 0.9086179989595977, + "grad_norm": 0.05381698161363602, + "learning_rate": 0.00014762, + "loss": 0.0398, + "step": 2620 + }, + { + "epoch": 0.9120860065892145, + "grad_norm": 0.05847073718905449, + "learning_rate": 0.00014742, + "loss": 0.0426, + "step": 2630 + }, + { + "epoch": 0.9155540142188313, + "grad_norm": 0.04121188446879387, + "learning_rate": 0.00014722, + "loss": 0.0379, + "step": 2640 + }, + { + "epoch": 0.9190220218484481, + "grad_norm": 0.05500589683651924, + "learning_rate": 0.00014702, + "loss": 0.0429, + "step": 2650 + }, + { + "epoch": 0.9224900294780648, + "grad_norm": 0.07371719181537628, + "learning_rate": 0.00014682, + "loss": 0.0404, + "step": 2660 + }, + { + "epoch": 0.9259580371076817, + "grad_norm": 0.07182417809963226, + "learning_rate": 0.00014662, + "loss": 0.0425, + "step": 2670 + }, + { + "epoch": 0.9294260447372984, + "grad_norm": 0.07057616114616394, + "learning_rate": 0.00014642, + "loss": 0.0447, + "step": 2680 + }, + { + "epoch": 0.9328940523669152, + "grad_norm": 0.06059495732188225, + "learning_rate": 0.00014622, + "loss": 0.0467, + "step": 2690 + }, + { + "epoch": 0.936362059996532, + "grad_norm": 0.05471622198820114, + "learning_rate": 0.00014602, + "loss": 0.0429, + "step": 2700 + }, + { + "epoch": 0.9398300676261487, + "grad_norm": 0.2795173227787018, + "learning_rate": 0.00014582, + "loss": 0.0441, + "step": 2710 + }, + { + "epoch": 0.9432980752557656, + "grad_norm": 0.05505786091089249, + "learning_rate": 0.00014562, + "loss": 0.0481, + "step": 2720 + }, + { + "epoch": 0.9467660828853823, + "grad_norm": 0.06604549288749695, + "learning_rate": 0.00014542, + "loss": 0.0436, + "step": 2730 + }, + { + "epoch": 0.9502340905149992, + "grad_norm": 0.04876833036541939, + "learning_rate": 0.00014522, + "loss": 0.0427, + "step": 2740 + }, + { + "epoch": 0.9537020981446159, + "grad_norm": 0.06586755067110062, + "learning_rate": 0.00014502, + "loss": 0.0376, + "step": 2750 + }, + { + "epoch": 0.9571701057742327, + "grad_norm": 0.05991548299789429, + "learning_rate": 0.00014482, + "loss": 0.0425, + "step": 2760 + }, + { + "epoch": 0.9606381134038495, + "grad_norm": 0.05464167147874832, + "learning_rate": 0.00014462, + "loss": 0.0475, + "step": 2770 + }, + { + "epoch": 0.9641061210334663, + "grad_norm": 0.08428128808736801, + "learning_rate": 0.00014442, + "loss": 0.0407, + "step": 2780 + }, + { + "epoch": 0.967574128663083, + "grad_norm": 0.05811979994177818, + "learning_rate": 0.00014422, + "loss": 0.0417, + "step": 2790 + }, + { + "epoch": 0.9710421362926999, + "grad_norm": 0.06777170300483704, + "learning_rate": 0.00014402, + "loss": 0.0394, + "step": 2800 + }, + { + "epoch": 0.9745101439223166, + "grad_norm": 0.07404989749193192, + "learning_rate": 0.00014382, + "loss": 0.0451, + "step": 2810 + }, + { + "epoch": 0.9779781515519335, + "grad_norm": 0.08595024049282074, + "learning_rate": 0.00014362, + "loss": 0.0518, + "step": 2820 + }, + { + "epoch": 0.9814461591815502, + "grad_norm": 0.07936517149209976, + "learning_rate": 0.00014342, + "loss": 0.0368, + "step": 2830 + }, + { + "epoch": 0.984914166811167, + "grad_norm": 0.08811389654874802, + "learning_rate": 0.00014322, + "loss": 0.0444, + "step": 2840 + }, + { + "epoch": 0.9883821744407838, + "grad_norm": 0.06864507496356964, + "learning_rate": 0.00014302, + "loss": 0.0445, + "step": 2850 + }, + { + "epoch": 0.9918501820704005, + "grad_norm": 0.05267275124788284, + "learning_rate": 0.00014282, + "loss": 0.0485, + "step": 2860 + }, + { + "epoch": 0.9953181897000173, + "grad_norm": 0.051028452813625336, + "learning_rate": 0.00014261999999999999, + "loss": 0.0416, + "step": 2870 + }, + { + "epoch": 0.9987861973296341, + "grad_norm": 0.047300126403570175, + "learning_rate": 0.00014242, + "loss": 0.0443, + "step": 2880 + }, + { + "epoch": 1.0022542049592509, + "grad_norm": 0.06848949193954468, + "learning_rate": 0.00014222, + "loss": 0.0422, + "step": 2890 + }, + { + "epoch": 1.0057222125888676, + "grad_norm": 0.05861698463559151, + "learning_rate": 0.00014202, + "loss": 0.0298, + "step": 2900 + }, + { + "epoch": 1.0091902202184846, + "grad_norm": 0.0671234130859375, + "learning_rate": 0.00014182, + "loss": 0.0271, + "step": 2910 + }, + { + "epoch": 1.0126582278481013, + "grad_norm": 0.07007008045911789, + "learning_rate": 0.00014162, + "loss": 0.0279, + "step": 2920 + }, + { + "epoch": 1.016126235477718, + "grad_norm": 0.07409070432186127, + "learning_rate": 0.00014141999999999998, + "loss": 0.0332, + "step": 2930 + }, + { + "epoch": 1.0195942431073348, + "grad_norm": 0.0820993110537529, + "learning_rate": 0.00014122, + "loss": 0.032, + "step": 2940 + }, + { + "epoch": 1.0230622507369516, + "grad_norm": 0.0683741420507431, + "learning_rate": 0.00014102, + "loss": 0.0285, + "step": 2950 + }, + { + "epoch": 1.0265302583665683, + "grad_norm": 0.09559917449951172, + "learning_rate": 0.00014082, + "loss": 0.0328, + "step": 2960 + }, + { + "epoch": 1.0299982659961853, + "grad_norm": 0.07927672564983368, + "learning_rate": 0.00014062, + "loss": 0.0317, + "step": 2970 + }, + { + "epoch": 1.033466273625802, + "grad_norm": 0.06615123897790909, + "learning_rate": 0.00014042, + "loss": 0.0289, + "step": 2980 + }, + { + "epoch": 1.0369342812554188, + "grad_norm": 0.051615212112665176, + "learning_rate": 0.00014022, + "loss": 0.0274, + "step": 2990 + }, + { + "epoch": 1.0404022888850355, + "grad_norm": 0.060853827744722366, + "learning_rate": 0.00014002, + "loss": 0.033, + "step": 3000 + }, + { + "epoch": 1.0404022888850355, + "eval_loss": 0.060293715447187424, + "eval_runtime": 713.1843, + "eval_samples_per_second": 12.875, + "eval_steps_per_second": 1.61, + "step": 3000 + }, + { + "epoch": 1.0438702965146522, + "grad_norm": 0.05449477955698967, + "learning_rate": 0.00013982000000000003, + "loss": 0.0308, + "step": 3010 + }, + { + "epoch": 1.0473383041442692, + "grad_norm": 0.06479578465223312, + "learning_rate": 0.00013962000000000002, + "loss": 0.0343, + "step": 3020 + }, + { + "epoch": 1.050806311773886, + "grad_norm": 0.060166362673044205, + "learning_rate": 0.00013942000000000002, + "loss": 0.0366, + "step": 3030 + }, + { + "epoch": 1.0542743194035027, + "grad_norm": 0.07008329033851624, + "learning_rate": 0.00013922, + "loss": 0.0327, + "step": 3040 + }, + { + "epoch": 1.0577423270331194, + "grad_norm": 0.07188612222671509, + "learning_rate": 0.00013902, + "loss": 0.0317, + "step": 3050 + }, + { + "epoch": 1.0612103346627362, + "grad_norm": 0.06554035097360611, + "learning_rate": 0.00013882000000000003, + "loss": 0.031, + "step": 3060 + }, + { + "epoch": 1.0646783422923531, + "grad_norm": 0.0675990879535675, + "learning_rate": 0.00013862000000000002, + "loss": 0.0341, + "step": 3070 + }, + { + "epoch": 1.0681463499219699, + "grad_norm": 0.061157677322626114, + "learning_rate": 0.00013842000000000002, + "loss": 0.0301, + "step": 3080 + }, + { + "epoch": 1.0716143575515866, + "grad_norm": 0.07951588183641434, + "learning_rate": 0.00013822000000000001, + "loss": 0.0289, + "step": 3090 + }, + { + "epoch": 1.0750823651812034, + "grad_norm": 0.1063622459769249, + "learning_rate": 0.00013802, + "loss": 0.0325, + "step": 3100 + }, + { + "epoch": 1.0785503728108201, + "grad_norm": 0.07916730642318726, + "learning_rate": 0.00013782, + "loss": 0.0338, + "step": 3110 + }, + { + "epoch": 1.082018380440437, + "grad_norm": 0.0792151466012001, + "learning_rate": 0.00013762000000000003, + "loss": 0.0287, + "step": 3120 + }, + { + "epoch": 1.0854863880700538, + "grad_norm": 0.0631512925028801, + "learning_rate": 0.00013742000000000002, + "loss": 0.031, + "step": 3130 + }, + { + "epoch": 1.0889543956996706, + "grad_norm": 0.07032682001590729, + "learning_rate": 0.00013722000000000002, + "loss": 0.0284, + "step": 3140 + }, + { + "epoch": 1.0924224033292873, + "grad_norm": 0.07017088681459427, + "learning_rate": 0.00013702, + "loss": 0.0292, + "step": 3150 + }, + { + "epoch": 1.095890410958904, + "grad_norm": 0.06100435182452202, + "learning_rate": 0.00013682, + "loss": 0.0359, + "step": 3160 + }, + { + "epoch": 1.0993584185885208, + "grad_norm": 0.05606581270694733, + "learning_rate": 0.00013662, + "loss": 0.032, + "step": 3170 + }, + { + "epoch": 1.1028264262181378, + "grad_norm": 0.07687368988990784, + "learning_rate": 0.00013642000000000003, + "loss": 0.0353, + "step": 3180 + }, + { + "epoch": 1.1062944338477545, + "grad_norm": 0.05796977877616882, + "learning_rate": 0.00013622000000000002, + "loss": 0.0286, + "step": 3190 + }, + { + "epoch": 1.1097624414773712, + "grad_norm": 0.08586996048688889, + "learning_rate": 0.00013602000000000002, + "loss": 0.0285, + "step": 3200 + }, + { + "epoch": 1.113230449106988, + "grad_norm": 0.08199802041053772, + "learning_rate": 0.00013582, + "loss": 0.0346, + "step": 3210 + }, + { + "epoch": 1.1166984567366047, + "grad_norm": 0.06581319123506546, + "learning_rate": 0.00013562, + "loss": 0.029, + "step": 3220 + }, + { + "epoch": 1.1201664643662217, + "grad_norm": 0.06755177676677704, + "learning_rate": 0.00013542, + "loss": 0.0323, + "step": 3230 + }, + { + "epoch": 1.1236344719958384, + "grad_norm": 0.07359416782855988, + "learning_rate": 0.00013522000000000002, + "loss": 0.031, + "step": 3240 + }, + { + "epoch": 1.1271024796254552, + "grad_norm": 0.07140175998210907, + "learning_rate": 0.00013502000000000002, + "loss": 0.0341, + "step": 3250 + }, + { + "epoch": 1.130570487255072, + "grad_norm": 0.07570434361696243, + "learning_rate": 0.00013482000000000001, + "loss": 0.0307, + "step": 3260 + }, + { + "epoch": 1.1340384948846887, + "grad_norm": 0.0646577849984169, + "learning_rate": 0.00013462, + "loss": 0.0353, + "step": 3270 + }, + { + "epoch": 1.1375065025143056, + "grad_norm": 0.07016121596097946, + "learning_rate": 0.00013442, + "loss": 0.0367, + "step": 3280 + }, + { + "epoch": 1.1409745101439224, + "grad_norm": 0.06299825757741928, + "learning_rate": 0.00013422, + "loss": 0.0355, + "step": 3290 + }, + { + "epoch": 1.1444425177735391, + "grad_norm": 0.07232199609279633, + "learning_rate": 0.00013402000000000002, + "loss": 0.0302, + "step": 3300 + }, + { + "epoch": 1.1479105254031559, + "grad_norm": 0.06672387570142746, + "learning_rate": 0.00013382000000000002, + "loss": 0.0314, + "step": 3310 + }, + { + "epoch": 1.1513785330327726, + "grad_norm": 0.06597165018320084, + "learning_rate": 0.00013362, + "loss": 0.0316, + "step": 3320 + }, + { + "epoch": 1.1548465406623896, + "grad_norm": 0.07945774495601654, + "learning_rate": 0.00013342, + "loss": 0.0304, + "step": 3330 + }, + { + "epoch": 1.1583145482920063, + "grad_norm": 0.0883309543132782, + "learning_rate": 0.00013322, + "loss": 0.0313, + "step": 3340 + }, + { + "epoch": 1.161782555921623, + "grad_norm": 0.1610005646944046, + "learning_rate": 0.00013302000000000002, + "loss": 0.0379, + "step": 3350 + }, + { + "epoch": 1.1652505635512398, + "grad_norm": 0.08350630104541779, + "learning_rate": 0.00013282000000000002, + "loss": 0.0283, + "step": 3360 + }, + { + "epoch": 1.1687185711808565, + "grad_norm": 0.08908521384000778, + "learning_rate": 0.00013262000000000001, + "loss": 0.0271, + "step": 3370 + }, + { + "epoch": 1.1721865788104733, + "grad_norm": 0.06909502297639847, + "learning_rate": 0.00013242, + "loss": 0.0312, + "step": 3380 + }, + { + "epoch": 1.1756545864400902, + "grad_norm": 0.0837428942322731, + "learning_rate": 0.00013222, + "loss": 0.0351, + "step": 3390 + }, + { + "epoch": 1.179122594069707, + "grad_norm": 0.06636606156826019, + "learning_rate": 0.00013202, + "loss": 0.0284, + "step": 3400 + }, + { + "epoch": 1.1825906016993237, + "grad_norm": 0.0990837961435318, + "learning_rate": 0.00013182000000000002, + "loss": 0.0301, + "step": 3410 + }, + { + "epoch": 1.1860586093289405, + "grad_norm": 0.08313869684934616, + "learning_rate": 0.00013162000000000002, + "loss": 0.0335, + "step": 3420 + }, + { + "epoch": 1.1895266169585572, + "grad_norm": 0.07330479472875595, + "learning_rate": 0.00013142, + "loss": 0.0328, + "step": 3430 + }, + { + "epoch": 1.1929946245881742, + "grad_norm": 0.05378459393978119, + "learning_rate": 0.00013122, + "loss": 0.0306, + "step": 3440 + }, + { + "epoch": 1.196462632217791, + "grad_norm": 0.06030990183353424, + "learning_rate": 0.00013102, + "loss": 0.0303, + "step": 3450 + }, + { + "epoch": 1.1999306398474077, + "grad_norm": 0.07298003882169724, + "learning_rate": 0.00013082, + "loss": 0.0312, + "step": 3460 + }, + { + "epoch": 1.2033986474770244, + "grad_norm": 0.06307482719421387, + "learning_rate": 0.00013062000000000002, + "loss": 0.0324, + "step": 3470 + }, + { + "epoch": 1.2068666551066412, + "grad_norm": 0.06007950380444527, + "learning_rate": 0.00013042000000000002, + "loss": 0.0339, + "step": 3480 + }, + { + "epoch": 1.2103346627362581, + "grad_norm": 0.07138363271951675, + "learning_rate": 0.00013022, + "loss": 0.0371, + "step": 3490 + }, + { + "epoch": 1.2138026703658749, + "grad_norm": 0.06266158819198608, + "learning_rate": 0.00013002, + "loss": 0.0329, + "step": 3500 + }, + { + "epoch": 1.2172706779954916, + "grad_norm": 0.06397438049316406, + "learning_rate": 0.00012982, + "loss": 0.0296, + "step": 3510 + }, + { + "epoch": 1.2207386856251083, + "grad_norm": 0.061814188957214355, + "learning_rate": 0.00012962, + "loss": 0.0306, + "step": 3520 + }, + { + "epoch": 1.224206693254725, + "grad_norm": 0.060092389583587646, + "learning_rate": 0.00012942000000000002, + "loss": 0.0323, + "step": 3530 + }, + { + "epoch": 1.227674700884342, + "grad_norm": 0.10667088627815247, + "learning_rate": 0.00012922, + "loss": 0.0298, + "step": 3540 + }, + { + "epoch": 1.2311427085139588, + "grad_norm": 0.09048482030630112, + "learning_rate": 0.00012902, + "loss": 0.0319, + "step": 3550 + }, + { + "epoch": 1.2346107161435755, + "grad_norm": 0.09124518185853958, + "learning_rate": 0.00012882, + "loss": 0.0361, + "step": 3560 + }, + { + "epoch": 1.2380787237731923, + "grad_norm": 0.05594000220298767, + "learning_rate": 0.00012862, + "loss": 0.0341, + "step": 3570 + }, + { + "epoch": 1.241546731402809, + "grad_norm": 0.06354895979166031, + "learning_rate": 0.00012842, + "loss": 0.0281, + "step": 3580 + }, + { + "epoch": 1.2450147390324258, + "grad_norm": 0.059312548488378525, + "learning_rate": 0.00012822000000000002, + "loss": 0.0276, + "step": 3590 + }, + { + "epoch": 1.2484827466620427, + "grad_norm": 0.06291409581899643, + "learning_rate": 0.00012802, + "loss": 0.0315, + "step": 3600 + }, + { + "epoch": 1.2519507542916595, + "grad_norm": 0.05183565244078636, + "learning_rate": 0.00012782, + "loss": 0.0294, + "step": 3610 + }, + { + "epoch": 1.2554187619212762, + "grad_norm": 0.06372030079364777, + "learning_rate": 0.00012762, + "loss": 0.03, + "step": 3620 + }, + { + "epoch": 1.258886769550893, + "grad_norm": 0.1327325701713562, + "learning_rate": 0.00012742, + "loss": 0.0366, + "step": 3630 + }, + { + "epoch": 1.26235477718051, + "grad_norm": 0.07337796688079834, + "learning_rate": 0.00012722000000000002, + "loss": 0.0324, + "step": 3640 + }, + { + "epoch": 1.2658227848101267, + "grad_norm": 0.06626396626234055, + "learning_rate": 0.00012702000000000001, + "loss": 0.0322, + "step": 3650 + }, + { + "epoch": 1.2692907924397434, + "grad_norm": 0.07255198061466217, + "learning_rate": 0.00012682, + "loss": 0.036, + "step": 3660 + }, + { + "epoch": 1.2727588000693602, + "grad_norm": 0.0766686201095581, + "learning_rate": 0.00012662, + "loss": 0.0315, + "step": 3670 + }, + { + "epoch": 1.276226807698977, + "grad_norm": 0.06377042084932327, + "learning_rate": 0.00012642, + "loss": 0.0322, + "step": 3680 + }, + { + "epoch": 1.2796948153285936, + "grad_norm": 0.0723329707980156, + "learning_rate": 0.00012622, + "loss": 0.0356, + "step": 3690 + }, + { + "epoch": 1.2831628229582104, + "grad_norm": 0.0876326933503151, + "learning_rate": 0.00012602000000000002, + "loss": 0.0346, + "step": 3700 + }, + { + "epoch": 1.2866308305878273, + "grad_norm": 0.07476814091205597, + "learning_rate": 0.00012582, + "loss": 0.0316, + "step": 3710 + }, + { + "epoch": 1.290098838217444, + "grad_norm": 0.07503268122673035, + "learning_rate": 0.00012562, + "loss": 0.0318, + "step": 3720 + }, + { + "epoch": 1.2935668458470608, + "grad_norm": 0.10708837956190109, + "learning_rate": 0.00012542, + "loss": 0.0349, + "step": 3730 + }, + { + "epoch": 1.2970348534766776, + "grad_norm": 0.08280046284198761, + "learning_rate": 0.00012522, + "loss": 0.0324, + "step": 3740 + }, + { + "epoch": 1.3005028611062945, + "grad_norm": 0.06630868464708328, + "learning_rate": 0.00012502, + "loss": 0.0305, + "step": 3750 + }, + { + "epoch": 1.3039708687359113, + "grad_norm": 0.09755595773458481, + "learning_rate": 0.00012482000000000001, + "loss": 0.0338, + "step": 3760 + }, + { + "epoch": 1.307438876365528, + "grad_norm": 0.08033673465251923, + "learning_rate": 0.00012462, + "loss": 0.0312, + "step": 3770 + }, + { + "epoch": 1.3109068839951448, + "grad_norm": 0.0811261385679245, + "learning_rate": 0.00012442, + "loss": 0.0322, + "step": 3780 + }, + { + "epoch": 1.3143748916247615, + "grad_norm": 0.06264316290616989, + "learning_rate": 0.00012422, + "loss": 0.0292, + "step": 3790 + }, + { + "epoch": 1.3178428992543783, + "grad_norm": 0.07748369127511978, + "learning_rate": 0.00012402, + "loss": 0.0302, + "step": 3800 + }, + { + "epoch": 1.3213109068839952, + "grad_norm": 0.0690523013472557, + "learning_rate": 0.00012382, + "loss": 0.0336, + "step": 3810 + }, + { + "epoch": 1.324778914513612, + "grad_norm": 0.09423090517520905, + "learning_rate": 0.00012362, + "loss": 0.0295, + "step": 3820 + }, + { + "epoch": 1.3282469221432287, + "grad_norm": 0.08562049269676208, + "learning_rate": 0.00012342, + "loss": 0.0292, + "step": 3830 + }, + { + "epoch": 1.3317149297728454, + "grad_norm": 0.05059509724378586, + "learning_rate": 0.00012322, + "loss": 0.0266, + "step": 3840 + }, + { + "epoch": 1.3351829374024624, + "grad_norm": 0.06358881294727325, + "learning_rate": 0.00012302, + "loss": 0.0309, + "step": 3850 + }, + { + "epoch": 1.3386509450320792, + "grad_norm": 0.1298878937959671, + "learning_rate": 0.00012282, + "loss": 0.0356, + "step": 3860 + }, + { + "epoch": 1.342118952661696, + "grad_norm": 0.06405311822891235, + "learning_rate": 0.00012262, + "loss": 0.0317, + "step": 3870 + }, + { + "epoch": 1.3455869602913126, + "grad_norm": 0.07615106552839279, + "learning_rate": 0.00012242, + "loss": 0.0285, + "step": 3880 + }, + { + "epoch": 1.3490549679209294, + "grad_norm": 0.08331302553415298, + "learning_rate": 0.00012222, + "loss": 0.0359, + "step": 3890 + }, + { + "epoch": 1.3525229755505461, + "grad_norm": 0.06869524717330933, + "learning_rate": 0.00012202, + "loss": 0.0319, + "step": 3900 + }, + { + "epoch": 1.3559909831801629, + "grad_norm": 0.08540484309196472, + "learning_rate": 0.00012182, + "loss": 0.0328, + "step": 3910 + }, + { + "epoch": 1.3594589908097798, + "grad_norm": 0.07091011852025986, + "learning_rate": 0.00012162, + "loss": 0.0368, + "step": 3920 + }, + { + "epoch": 1.3629269984393966, + "grad_norm": 0.075434111058712, + "learning_rate": 0.00012142, + "loss": 0.0336, + "step": 3930 + }, + { + "epoch": 1.3663950060690133, + "grad_norm": 0.06716951727867126, + "learning_rate": 0.00012122, + "loss": 0.0335, + "step": 3940 + }, + { + "epoch": 1.36986301369863, + "grad_norm": 0.09082087874412537, + "learning_rate": 0.00012102, + "loss": 0.0345, + "step": 3950 + }, + { + "epoch": 1.373331021328247, + "grad_norm": 0.0903453379869461, + "learning_rate": 0.00012082, + "loss": 0.0415, + "step": 3960 + }, + { + "epoch": 1.3767990289578638, + "grad_norm": 0.052235305309295654, + "learning_rate": 0.00012062, + "loss": 0.0282, + "step": 3970 + }, + { + "epoch": 1.3802670365874805, + "grad_norm": 0.07253699749708176, + "learning_rate": 0.00012042, + "loss": 0.0316, + "step": 3980 + }, + { + "epoch": 1.3837350442170973, + "grad_norm": 0.0548410601913929, + "learning_rate": 0.00012022, + "loss": 0.0303, + "step": 3990 + }, + { + "epoch": 1.387203051846714, + "grad_norm": 0.08785740286111832, + "learning_rate": 0.00012001999999999999, + "loss": 0.0337, + "step": 4000 + }, + { + "epoch": 1.387203051846714, + "eval_loss": 0.060144323855638504, + "eval_runtime": 708.9302, + "eval_samples_per_second": 12.952, + "eval_steps_per_second": 1.619, + "step": 4000 + }, + { + "epoch": 1.3906710594763307, + "grad_norm": 0.0650157481431961, + "learning_rate": 0.00011982, + "loss": 0.0327, + "step": 4010 + }, + { + "epoch": 1.3941390671059477, + "grad_norm": 0.06679214537143707, + "learning_rate": 0.00011962, + "loss": 0.0352, + "step": 4020 + }, + { + "epoch": 1.3976070747355644, + "grad_norm": 0.08368890732526779, + "learning_rate": 0.00011942, + "loss": 0.0346, + "step": 4030 + }, + { + "epoch": 1.4010750823651812, + "grad_norm": 0.09027834981679916, + "learning_rate": 0.00011922, + "loss": 0.0304, + "step": 4040 + }, + { + "epoch": 1.404543089994798, + "grad_norm": 0.0602988600730896, + "learning_rate": 0.00011902, + "loss": 0.0278, + "step": 4050 + }, + { + "epoch": 1.408011097624415, + "grad_norm": 0.08348573744297028, + "learning_rate": 0.00011882, + "loss": 0.0316, + "step": 4060 + }, + { + "epoch": 1.4114791052540316, + "grad_norm": 0.065241239964962, + "learning_rate": 0.00011862, + "loss": 0.0308, + "step": 4070 + }, + { + "epoch": 1.4149471128836484, + "grad_norm": 0.058722469955682755, + "learning_rate": 0.00011842, + "loss": 0.0312, + "step": 4080 + }, + { + "epoch": 1.4184151205132651, + "grad_norm": 0.06701633334159851, + "learning_rate": 0.00011822, + "loss": 0.0303, + "step": 4090 + }, + { + "epoch": 1.4218831281428819, + "grad_norm": 0.06035483255982399, + "learning_rate": 0.00011802, + "loss": 0.0263, + "step": 4100 + }, + { + "epoch": 1.4253511357724986, + "grad_norm": 0.07209423929452896, + "learning_rate": 0.00011782, + "loss": 0.0299, + "step": 4110 + }, + { + "epoch": 1.4288191434021154, + "grad_norm": 0.08608460426330566, + "learning_rate": 0.00011762, + "loss": 0.0299, + "step": 4120 + }, + { + "epoch": 1.4322871510317323, + "grad_norm": 0.06970153748989105, + "learning_rate": 0.00011742, + "loss": 0.0345, + "step": 4130 + }, + { + "epoch": 1.435755158661349, + "grad_norm": 0.05995609238743782, + "learning_rate": 0.00011721999999999999, + "loss": 0.0284, + "step": 4140 + }, + { + "epoch": 1.4392231662909658, + "grad_norm": 0.03222940117120743, + "learning_rate": 0.00011702, + "loss": 0.0293, + "step": 4150 + }, + { + "epoch": 1.4426911739205825, + "grad_norm": 0.07076498866081238, + "learning_rate": 0.00011682, + "loss": 0.0362, + "step": 4160 + }, + { + "epoch": 1.4461591815501995, + "grad_norm": 0.07425186783075333, + "learning_rate": 0.00011661999999999999, + "loss": 0.0281, + "step": 4170 + }, + { + "epoch": 1.4496271891798163, + "grad_norm": 0.04875819757580757, + "learning_rate": 0.00011642, + "loss": 0.0361, + "step": 4180 + }, + { + "epoch": 1.453095196809433, + "grad_norm": 0.06577154994010925, + "learning_rate": 0.00011622, + "loss": 0.0291, + "step": 4190 + }, + { + "epoch": 1.4565632044390497, + "grad_norm": 0.08174604177474976, + "learning_rate": 0.00011601999999999999, + "loss": 0.0284, + "step": 4200 + }, + { + "epoch": 1.4600312120686665, + "grad_norm": 0.08212857693433762, + "learning_rate": 0.00011582, + "loss": 0.03, + "step": 4210 + }, + { + "epoch": 1.4634992196982832, + "grad_norm": 0.06090838089585304, + "learning_rate": 0.00011562, + "loss": 0.028, + "step": 4220 + }, + { + "epoch": 1.4669672273279002, + "grad_norm": 0.07029874622821808, + "learning_rate": 0.00011541999999999999, + "loss": 0.033, + "step": 4230 + }, + { + "epoch": 1.470435234957517, + "grad_norm": 0.10370688140392303, + "learning_rate": 0.00011522, + "loss": 0.0341, + "step": 4240 + }, + { + "epoch": 1.4739032425871337, + "grad_norm": 0.06497831642627716, + "learning_rate": 0.00011501999999999999, + "loss": 0.0355, + "step": 4250 + }, + { + "epoch": 1.4773712502167504, + "grad_norm": 0.047859255224466324, + "learning_rate": 0.00011482000000000002, + "loss": 0.0312, + "step": 4260 + }, + { + "epoch": 1.4808392578463674, + "grad_norm": 0.044814929366111755, + "learning_rate": 0.00011462000000000001, + "loss": 0.0325, + "step": 4270 + }, + { + "epoch": 1.4843072654759841, + "grad_norm": 0.08687663078308105, + "learning_rate": 0.00011442000000000002, + "loss": 0.0346, + "step": 4280 + }, + { + "epoch": 1.4877752731056009, + "grad_norm": 0.06687606126070023, + "learning_rate": 0.00011422000000000001, + "loss": 0.0318, + "step": 4290 + }, + { + "epoch": 1.4912432807352176, + "grad_norm": 0.0769667997956276, + "learning_rate": 0.00011402000000000001, + "loss": 0.0314, + "step": 4300 + }, + { + "epoch": 1.4947112883648344, + "grad_norm": 0.07912110537290573, + "learning_rate": 0.00011382000000000002, + "loss": 0.0269, + "step": 4310 + }, + { + "epoch": 1.498179295994451, + "grad_norm": 0.06801219284534454, + "learning_rate": 0.00011362000000000001, + "loss": 0.032, + "step": 4320 + }, + { + "epoch": 1.5016473036240678, + "grad_norm": 0.07353610545396805, + "learning_rate": 0.00011342000000000001, + "loss": 0.0328, + "step": 4330 + }, + { + "epoch": 1.5051153112536848, + "grad_norm": 0.05926644429564476, + "learning_rate": 0.00011322000000000002, + "loss": 0.0268, + "step": 4340 + }, + { + "epoch": 1.5085833188833015, + "grad_norm": 0.07942460477352142, + "learning_rate": 0.00011302000000000001, + "loss": 0.0299, + "step": 4350 + }, + { + "epoch": 1.5120513265129183, + "grad_norm": 0.09032566100358963, + "learning_rate": 0.00011282000000000002, + "loss": 0.0344, + "step": 4360 + }, + { + "epoch": 1.5155193341425353, + "grad_norm": 0.08992986381053925, + "learning_rate": 0.00011262000000000002, + "loss": 0.0282, + "step": 4370 + }, + { + "epoch": 1.518987341772152, + "grad_norm": 0.10055962204933167, + "learning_rate": 0.00011242000000000001, + "loss": 0.0368, + "step": 4380 + }, + { + "epoch": 1.5224553494017687, + "grad_norm": 0.06706701964139938, + "learning_rate": 0.00011222000000000002, + "loss": 0.0348, + "step": 4390 + }, + { + "epoch": 1.5259233570313855, + "grad_norm": 0.07412678748369217, + "learning_rate": 0.00011202000000000002, + "loss": 0.0322, + "step": 4400 + }, + { + "epoch": 1.5293913646610022, + "grad_norm": 0.0761900544166565, + "learning_rate": 0.00011182000000000001, + "loss": 0.0334, + "step": 4410 + }, + { + "epoch": 1.532859372290619, + "grad_norm": 0.06172578036785126, + "learning_rate": 0.00011162000000000002, + "loss": 0.0326, + "step": 4420 + }, + { + "epoch": 1.5363273799202357, + "grad_norm": 0.06953331083059311, + "learning_rate": 0.00011142000000000001, + "loss": 0.0298, + "step": 4430 + }, + { + "epoch": 1.5397953875498525, + "grad_norm": 0.07618329674005508, + "learning_rate": 0.00011122000000000001, + "loss": 0.0283, + "step": 4440 + }, + { + "epoch": 1.5432633951794694, + "grad_norm": 0.07265307009220123, + "learning_rate": 0.00011102000000000002, + "loss": 0.032, + "step": 4450 + }, + { + "epoch": 1.5467314028090862, + "grad_norm": 0.07409724593162537, + "learning_rate": 0.00011082000000000001, + "loss": 0.0303, + "step": 4460 + }, + { + "epoch": 1.5501994104387031, + "grad_norm": 0.05352557823061943, + "learning_rate": 0.00011062000000000001, + "loss": 0.0304, + "step": 4470 + }, + { + "epoch": 1.5536674180683199, + "grad_norm": 0.10606401413679123, + "learning_rate": 0.00011042000000000002, + "loss": 0.0338, + "step": 4480 + }, + { + "epoch": 1.5571354256979366, + "grad_norm": 0.07364092022180557, + "learning_rate": 0.00011022000000000001, + "loss": 0.0308, + "step": 4490 + }, + { + "epoch": 1.5606034333275534, + "grad_norm": 0.08737417310476303, + "learning_rate": 0.00011002000000000001, + "loss": 0.0299, + "step": 4500 + }, + { + "epoch": 1.56407144095717, + "grad_norm": 0.07062090188264847, + "learning_rate": 0.00010982000000000002, + "loss": 0.0314, + "step": 4510 + }, + { + "epoch": 1.5675394485867868, + "grad_norm": 0.0711718276143074, + "learning_rate": 0.00010962000000000001, + "loss": 0.0322, + "step": 4520 + }, + { + "epoch": 1.5710074562164036, + "grad_norm": 0.08718711882829666, + "learning_rate": 0.00010942, + "loss": 0.0297, + "step": 4530 + }, + { + "epoch": 1.5744754638460203, + "grad_norm": 0.06502439081668854, + "learning_rate": 0.00010922000000000001, + "loss": 0.0385, + "step": 4540 + }, + { + "epoch": 1.5779434714756373, + "grad_norm": 0.07162761688232422, + "learning_rate": 0.00010902000000000001, + "loss": 0.0301, + "step": 4550 + }, + { + "epoch": 1.581411479105254, + "grad_norm": 0.07721313089132309, + "learning_rate": 0.00010882, + "loss": 0.0284, + "step": 4560 + }, + { + "epoch": 1.5848794867348708, + "grad_norm": 0.07071566581726074, + "learning_rate": 0.00010862000000000001, + "loss": 0.0352, + "step": 4570 + }, + { + "epoch": 1.5883474943644877, + "grad_norm": 0.1029210165143013, + "learning_rate": 0.00010842000000000001, + "loss": 0.0319, + "step": 4580 + }, + { + "epoch": 1.5918155019941045, + "grad_norm": 0.05683687701821327, + "learning_rate": 0.00010822, + "loss": 0.0343, + "step": 4590 + }, + { + "epoch": 1.5952835096237212, + "grad_norm": 0.05821290984749794, + "learning_rate": 0.00010802000000000001, + "loss": 0.039, + "step": 4600 + }, + { + "epoch": 1.598751517253338, + "grad_norm": 0.07931312173604965, + "learning_rate": 0.00010782000000000001, + "loss": 0.0294, + "step": 4610 + }, + { + "epoch": 1.6022195248829547, + "grad_norm": 0.06197603419423103, + "learning_rate": 0.00010762, + "loss": 0.0322, + "step": 4620 + }, + { + "epoch": 1.6056875325125715, + "grad_norm": 0.06312838196754456, + "learning_rate": 0.00010742000000000001, + "loss": 0.0298, + "step": 4630 + }, + { + "epoch": 1.6091555401421882, + "grad_norm": 0.07855828106403351, + "learning_rate": 0.00010722000000000001, + "loss": 0.0303, + "step": 4640 + }, + { + "epoch": 1.612623547771805, + "grad_norm": 0.055718790739774704, + "learning_rate": 0.00010702000000000002, + "loss": 0.0311, + "step": 4650 + }, + { + "epoch": 1.616091555401422, + "grad_norm": 0.07305306941270828, + "learning_rate": 0.00010682000000000001, + "loss": 0.0303, + "step": 4660 + }, + { + "epoch": 1.6195595630310387, + "grad_norm": 0.07300154864788055, + "learning_rate": 0.00010662, + "loss": 0.0301, + "step": 4670 + }, + { + "epoch": 1.6230275706606556, + "grad_norm": 0.06121309846639633, + "learning_rate": 0.00010642000000000001, + "loss": 0.029, + "step": 4680 + }, + { + "epoch": 1.6264955782902724, + "grad_norm": 0.060993146151304245, + "learning_rate": 0.00010622000000000001, + "loss": 0.0297, + "step": 4690 + }, + { + "epoch": 1.629963585919889, + "grad_norm": 0.07691816985607147, + "learning_rate": 0.00010602, + "loss": 0.0339, + "step": 4700 + }, + { + "epoch": 1.6334315935495058, + "grad_norm": 0.07278670370578766, + "learning_rate": 0.00010582000000000001, + "loss": 0.0262, + "step": 4710 + }, + { + "epoch": 1.6368996011791226, + "grad_norm": 0.05704551190137863, + "learning_rate": 0.00010562000000000001, + "loss": 0.028, + "step": 4720 + }, + { + "epoch": 1.6403676088087393, + "grad_norm": 0.07973553240299225, + "learning_rate": 0.00010542, + "loss": 0.0328, + "step": 4730 + }, + { + "epoch": 1.643835616438356, + "grad_norm": 0.05720138177275658, + "learning_rate": 0.00010522000000000001, + "loss": 0.0309, + "step": 4740 + }, + { + "epoch": 1.6473036240679728, + "grad_norm": 0.07787197083234787, + "learning_rate": 0.00010502000000000001, + "loss": 0.0294, + "step": 4750 + }, + { + "epoch": 1.6507716316975898, + "grad_norm": 0.08834118396043777, + "learning_rate": 0.00010482, + "loss": 0.0301, + "step": 4760 + }, + { + "epoch": 1.6542396393272065, + "grad_norm": 0.08383214473724365, + "learning_rate": 0.00010462000000000001, + "loss": 0.0322, + "step": 4770 + }, + { + "epoch": 1.6577076469568233, + "grad_norm": 0.08266714960336685, + "learning_rate": 0.00010442, + "loss": 0.0323, + "step": 4780 + }, + { + "epoch": 1.6611756545864402, + "grad_norm": 0.06535809487104416, + "learning_rate": 0.00010422, + "loss": 0.0334, + "step": 4790 + }, + { + "epoch": 1.664643662216057, + "grad_norm": 0.07224865257740021, + "learning_rate": 0.00010402000000000001, + "loss": 0.0312, + "step": 4800 + }, + { + "epoch": 1.6681116698456737, + "grad_norm": 0.07816470414400101, + "learning_rate": 0.00010382, + "loss": 0.0342, + "step": 4810 + }, + { + "epoch": 1.6715796774752905, + "grad_norm": 0.48284128308296204, + "learning_rate": 0.00010362, + "loss": 0.0387, + "step": 4820 + }, + { + "epoch": 1.6750476851049072, + "grad_norm": 0.15331751108169556, + "learning_rate": 0.00010342000000000001, + "loss": 0.0289, + "step": 4830 + }, + { + "epoch": 1.678515692734524, + "grad_norm": 0.08506326377391815, + "learning_rate": 0.00010322, + "loss": 0.0299, + "step": 4840 + }, + { + "epoch": 1.6819837003641407, + "grad_norm": 0.06808125227689743, + "learning_rate": 0.00010302, + "loss": 0.0266, + "step": 4850 + }, + { + "epoch": 1.6854517079937574, + "grad_norm": 0.07349207252264023, + "learning_rate": 0.00010282000000000001, + "loss": 0.0297, + "step": 4860 + }, + { + "epoch": 1.6889197156233744, + "grad_norm": 0.3372306525707245, + "learning_rate": 0.00010262, + "loss": 0.0342, + "step": 4870 + }, + { + "epoch": 1.6923877232529911, + "grad_norm": 0.10550106316804886, + "learning_rate": 0.00010242, + "loss": 0.0338, + "step": 4880 + }, + { + "epoch": 1.695855730882608, + "grad_norm": 0.0708487331867218, + "learning_rate": 0.00010222000000000001, + "loss": 0.0294, + "step": 4890 + }, + { + "epoch": 1.6993237385122248, + "grad_norm": 0.5124090313911438, + "learning_rate": 0.00010202, + "loss": 0.0321, + "step": 4900 + }, + { + "epoch": 1.7027917461418416, + "grad_norm": 0.08598774671554565, + "learning_rate": 0.00010182, + "loss": 0.0326, + "step": 4910 + }, + { + "epoch": 1.7062597537714583, + "grad_norm": 0.06315886229276657, + "learning_rate": 0.00010162, + "loss": 0.0358, + "step": 4920 + }, + { + "epoch": 1.709727761401075, + "grad_norm": 0.2513497769832611, + "learning_rate": 0.00010142, + "loss": 0.0352, + "step": 4930 + }, + { + "epoch": 1.7131957690306918, + "grad_norm": 0.06378067284822464, + "learning_rate": 0.00010122000000000001, + "loss": 0.0277, + "step": 4940 + }, + { + "epoch": 1.7166637766603086, + "grad_norm": 0.08322855085134506, + "learning_rate": 0.00010102, + "loss": 0.0303, + "step": 4950 + }, + { + "epoch": 1.7201317842899253, + "grad_norm": 0.1097235381603241, + "learning_rate": 0.00010082, + "loss": 0.0328, + "step": 4960 + }, + { + "epoch": 1.7235997919195423, + "grad_norm": 0.06668414920568466, + "learning_rate": 0.00010062000000000001, + "loss": 0.034, + "step": 4970 + }, + { + "epoch": 1.727067799549159, + "grad_norm": 0.08046268671751022, + "learning_rate": 0.00010042, + "loss": 0.0315, + "step": 4980 + }, + { + "epoch": 1.7305358071787758, + "grad_norm": 0.06834772229194641, + "learning_rate": 0.00010022, + "loss": 0.0304, + "step": 4990 + }, + { + "epoch": 1.7340038148083927, + "grad_norm": 0.06567910313606262, + "learning_rate": 0.00010002000000000001, + "loss": 0.0338, + "step": 5000 + }, + { + "epoch": 1.7340038148083927, + "eval_loss": 0.059129249304533005, + "eval_runtime": 713.2468, + "eval_samples_per_second": 12.874, + "eval_steps_per_second": 1.61, + "step": 5000 + }, + { + "epoch": 1.7374718224380095, + "grad_norm": 0.0854811891913414, + "learning_rate": 9.982e-05, + "loss": 0.0308, + "step": 5010 + }, + { + "epoch": 1.7409398300676262, + "grad_norm": 0.07272527366876602, + "learning_rate": 9.962e-05, + "loss": 0.0333, + "step": 5020 + }, + { + "epoch": 1.744407837697243, + "grad_norm": 0.0846826583147049, + "learning_rate": 9.942000000000001e-05, + "loss": 0.0342, + "step": 5030 + }, + { + "epoch": 1.7478758453268597, + "grad_norm": 0.06776320934295654, + "learning_rate": 9.922e-05, + "loss": 0.0333, + "step": 5040 + }, + { + "epoch": 1.7513438529564764, + "grad_norm": 0.07815729081630707, + "learning_rate": 9.902e-05, + "loss": 0.0355, + "step": 5050 + }, + { + "epoch": 1.7548118605860932, + "grad_norm": 0.13363681733608246, + "learning_rate": 9.882e-05, + "loss": 0.032, + "step": 5060 + }, + { + "epoch": 1.75827986821571, + "grad_norm": 0.05876624956727028, + "learning_rate": 9.862e-05, + "loss": 0.033, + "step": 5070 + }, + { + "epoch": 1.7617478758453269, + "grad_norm": 0.10773160308599472, + "learning_rate": 9.842e-05, + "loss": 0.0309, + "step": 5080 + }, + { + "epoch": 1.7652158834749436, + "grad_norm": 0.05928561091423035, + "learning_rate": 9.822e-05, + "loss": 0.0288, + "step": 5090 + }, + { + "epoch": 1.7686838911045606, + "grad_norm": 0.058999065309762955, + "learning_rate": 9.802e-05, + "loss": 0.0251, + "step": 5100 + }, + { + "epoch": 1.7721518987341773, + "grad_norm": 0.0768052414059639, + "learning_rate": 9.782e-05, + "loss": 0.0292, + "step": 5110 + }, + { + "epoch": 1.775619906363794, + "grad_norm": 0.1691245585680008, + "learning_rate": 9.762e-05, + "loss": 0.0327, + "step": 5120 + }, + { + "epoch": 1.7790879139934108, + "grad_norm": 0.08563978224992752, + "learning_rate": 9.742e-05, + "loss": 0.0321, + "step": 5130 + }, + { + "epoch": 1.7825559216230276, + "grad_norm": 0.07125357538461685, + "learning_rate": 9.722e-05, + "loss": 0.0281, + "step": 5140 + }, + { + "epoch": 1.7860239292526443, + "grad_norm": 0.10519967973232269, + "learning_rate": 9.702e-05, + "loss": 0.03, + "step": 5150 + }, + { + "epoch": 1.789491936882261, + "grad_norm": 0.08440076559782028, + "learning_rate": 9.682e-05, + "loss": 0.0309, + "step": 5160 + }, + { + "epoch": 1.7929599445118778, + "grad_norm": 0.09944937378168106, + "learning_rate": 9.661999999999999e-05, + "loss": 0.0323, + "step": 5170 + }, + { + "epoch": 1.7964279521414948, + "grad_norm": 0.07523104548454285, + "learning_rate": 9.642e-05, + "loss": 0.026, + "step": 5180 + }, + { + "epoch": 1.7998959597711115, + "grad_norm": 0.09478747099637985, + "learning_rate": 9.622000000000001e-05, + "loss": 0.0304, + "step": 5190 + }, + { + "epoch": 1.8033639674007282, + "grad_norm": 0.0627417042851448, + "learning_rate": 9.602e-05, + "loss": 0.0353, + "step": 5200 + }, + { + "epoch": 1.8068319750303452, + "grad_norm": 0.06294772773981094, + "learning_rate": 9.582000000000001e-05, + "loss": 0.0301, + "step": 5210 + }, + { + "epoch": 1.810299982659962, + "grad_norm": 0.06453125923871994, + "learning_rate": 9.562000000000001e-05, + "loss": 0.0372, + "step": 5220 + }, + { + "epoch": 1.8137679902895787, + "grad_norm": 0.08603645861148834, + "learning_rate": 9.542e-05, + "loss": 0.0313, + "step": 5230 + }, + { + "epoch": 1.8172359979191954, + "grad_norm": 0.0609930120408535, + "learning_rate": 9.522000000000001e-05, + "loss": 0.0337, + "step": 5240 + }, + { + "epoch": 1.8207040055488122, + "grad_norm": 0.06789145618677139, + "learning_rate": 9.502000000000001e-05, + "loss": 0.0291, + "step": 5250 + }, + { + "epoch": 1.824172013178429, + "grad_norm": 0.062086399644613266, + "learning_rate": 9.482e-05, + "loss": 0.0284, + "step": 5260 + }, + { + "epoch": 1.8276400208080457, + "grad_norm": 0.0755184143781662, + "learning_rate": 9.462000000000001e-05, + "loss": 0.0311, + "step": 5270 + }, + { + "epoch": 1.8311080284376624, + "grad_norm": 0.06724268943071365, + "learning_rate": 9.442000000000001e-05, + "loss": 0.0288, + "step": 5280 + }, + { + "epoch": 1.8345760360672794, + "grad_norm": 0.08856779336929321, + "learning_rate": 9.422e-05, + "loss": 0.0313, + "step": 5290 + }, + { + "epoch": 1.8380440436968961, + "grad_norm": 0.08593250811100006, + "learning_rate": 9.402000000000001e-05, + "loss": 0.0299, + "step": 5300 + }, + { + "epoch": 1.841512051326513, + "grad_norm": 0.07944291085004807, + "learning_rate": 9.382e-05, + "loss": 0.0309, + "step": 5310 + }, + { + "epoch": 1.8449800589561298, + "grad_norm": 0.06297358870506287, + "learning_rate": 9.362e-05, + "loss": 0.0322, + "step": 5320 + }, + { + "epoch": 1.8484480665857466, + "grad_norm": 0.0893145203590393, + "learning_rate": 9.342000000000001e-05, + "loss": 0.0297, + "step": 5330 + }, + { + "epoch": 1.8519160742153633, + "grad_norm": 0.06626788526773453, + "learning_rate": 9.322e-05, + "loss": 0.0359, + "step": 5340 + }, + { + "epoch": 1.85538408184498, + "grad_norm": 0.10941380262374878, + "learning_rate": 9.302e-05, + "loss": 0.0322, + "step": 5350 + }, + { + "epoch": 1.8588520894745968, + "grad_norm": 0.08435889333486557, + "learning_rate": 9.282000000000001e-05, + "loss": 0.0361, + "step": 5360 + }, + { + "epoch": 1.8623200971042135, + "grad_norm": 0.08986232429742813, + "learning_rate": 9.262e-05, + "loss": 0.0288, + "step": 5370 + }, + { + "epoch": 1.8657881047338303, + "grad_norm": 0.0800371989607811, + "learning_rate": 9.242000000000001e-05, + "loss": 0.0339, + "step": 5380 + }, + { + "epoch": 1.8692561123634472, + "grad_norm": 0.08191009610891342, + "learning_rate": 9.222000000000001e-05, + "loss": 0.0284, + "step": 5390 + }, + { + "epoch": 1.872724119993064, + "grad_norm": 0.10277281701564789, + "learning_rate": 9.202e-05, + "loss": 0.0307, + "step": 5400 + }, + { + "epoch": 1.8761921276226807, + "grad_norm": 0.08141244202852249, + "learning_rate": 9.182000000000001e-05, + "loss": 0.0327, + "step": 5410 + }, + { + "epoch": 1.8796601352522977, + "grad_norm": 0.07554444670677185, + "learning_rate": 9.162000000000001e-05, + "loss": 0.031, + "step": 5420 + }, + { + "epoch": 1.8831281428819144, + "grad_norm": 0.0898871198296547, + "learning_rate": 9.142e-05, + "loss": 0.0306, + "step": 5430 + }, + { + "epoch": 1.8865961505115312, + "grad_norm": 0.06314833462238312, + "learning_rate": 9.122000000000001e-05, + "loss": 0.0348, + "step": 5440 + }, + { + "epoch": 1.890064158141148, + "grad_norm": 0.06320305913686752, + "learning_rate": 9.102e-05, + "loss": 0.0277, + "step": 5450 + }, + { + "epoch": 1.8935321657707647, + "grad_norm": 0.09586924314498901, + "learning_rate": 9.082e-05, + "loss": 0.0361, + "step": 5460 + }, + { + "epoch": 1.8970001734003814, + "grad_norm": 0.0577247217297554, + "learning_rate": 9.062000000000001e-05, + "loss": 0.0295, + "step": 5470 + }, + { + "epoch": 1.9004681810299981, + "grad_norm": 0.06483156979084015, + "learning_rate": 9.042e-05, + "loss": 0.0311, + "step": 5480 + }, + { + "epoch": 1.903936188659615, + "grad_norm": 0.0748353898525238, + "learning_rate": 9.022e-05, + "loss": 0.0324, + "step": 5490 + }, + { + "epoch": 1.9074041962892319, + "grad_norm": 0.09499184042215347, + "learning_rate": 9.002000000000001e-05, + "loss": 0.0307, + "step": 5500 + }, + { + "epoch": 1.9108722039188486, + "grad_norm": 0.07709678262472153, + "learning_rate": 8.982e-05, + "loss": 0.0337, + "step": 5510 + }, + { + "epoch": 1.9143402115484656, + "grad_norm": 0.06375749409198761, + "learning_rate": 8.962e-05, + "loss": 0.0231, + "step": 5520 + }, + { + "epoch": 1.9178082191780823, + "grad_norm": 0.04287609085440636, + "learning_rate": 8.942000000000001e-05, + "loss": 0.0243, + "step": 5530 + }, + { + "epoch": 1.921276226807699, + "grad_norm": 0.07707373052835464, + "learning_rate": 8.922e-05, + "loss": 0.0304, + "step": 5540 + }, + { + "epoch": 1.9247442344373158, + "grad_norm": 0.06724567711353302, + "learning_rate": 8.902e-05, + "loss": 0.0323, + "step": 5550 + }, + { + "epoch": 1.9282122420669325, + "grad_norm": 0.06895706802606583, + "learning_rate": 8.882000000000001e-05, + "loss": 0.0346, + "step": 5560 + }, + { + "epoch": 1.9316802496965493, + "grad_norm": 0.0870724767446518, + "learning_rate": 8.862e-05, + "loss": 0.0326, + "step": 5570 + }, + { + "epoch": 1.935148257326166, + "grad_norm": 0.061669524759054184, + "learning_rate": 8.842e-05, + "loss": 0.0287, + "step": 5580 + }, + { + "epoch": 1.9386162649557828, + "grad_norm": 0.06450454145669937, + "learning_rate": 8.822e-05, + "loss": 0.0295, + "step": 5590 + }, + { + "epoch": 1.9420842725853997, + "grad_norm": 0.07439760863780975, + "learning_rate": 8.802e-05, + "loss": 0.0325, + "step": 5600 + }, + { + "epoch": 1.9455522802150165, + "grad_norm": 0.08620608597993851, + "learning_rate": 8.782e-05, + "loss": 0.0306, + "step": 5610 + }, + { + "epoch": 1.9490202878446332, + "grad_norm": 0.08417027443647385, + "learning_rate": 8.762e-05, + "loss": 0.0339, + "step": 5620 + }, + { + "epoch": 1.9524882954742502, + "grad_norm": 0.0670836940407753, + "learning_rate": 8.742e-05, + "loss": 0.0318, + "step": 5630 + }, + { + "epoch": 1.955956303103867, + "grad_norm": 0.08096006512641907, + "learning_rate": 8.722e-05, + "loss": 0.0269, + "step": 5640 + }, + { + "epoch": 1.9594243107334837, + "grad_norm": 0.04559866338968277, + "learning_rate": 8.702e-05, + "loss": 0.0271, + "step": 5650 + }, + { + "epoch": 1.9628923183631004, + "grad_norm": 0.08112025260925293, + "learning_rate": 8.682e-05, + "loss": 0.036, + "step": 5660 + }, + { + "epoch": 1.9663603259927172, + "grad_norm": 0.07936326414346695, + "learning_rate": 8.662000000000001e-05, + "loss": 0.0323, + "step": 5670 + }, + { + "epoch": 1.969828333622334, + "grad_norm": 0.09269768744707108, + "learning_rate": 8.642e-05, + "loss": 0.0318, + "step": 5680 + }, + { + "epoch": 1.9732963412519506, + "grad_norm": 0.06654026359319687, + "learning_rate": 8.622e-05, + "loss": 0.0291, + "step": 5690 + }, + { + "epoch": 1.9767643488815674, + "grad_norm": 0.0971643254160881, + "learning_rate": 8.602e-05, + "loss": 0.031, + "step": 5700 + }, + { + "epoch": 1.9802323565111843, + "grad_norm": 0.06067187711596489, + "learning_rate": 8.582e-05, + "loss": 0.0306, + "step": 5710 + }, + { + "epoch": 1.983700364140801, + "grad_norm": 0.0945192202925682, + "learning_rate": 8.562e-05, + "loss": 0.0328, + "step": 5720 + }, + { + "epoch": 1.987168371770418, + "grad_norm": 0.07627417147159576, + "learning_rate": 8.542e-05, + "loss": 0.0352, + "step": 5730 + }, + { + "epoch": 1.9906363794000348, + "grad_norm": 0.08669853955507278, + "learning_rate": 8.522e-05, + "loss": 0.0337, + "step": 5740 + }, + { + "epoch": 1.9941043870296515, + "grad_norm": 0.06610149890184402, + "learning_rate": 8.502e-05, + "loss": 0.0288, + "step": 5750 + }, + { + "epoch": 1.9975723946592683, + "grad_norm": 0.06989070028066635, + "learning_rate": 8.482e-05, + "loss": 0.0306, + "step": 5760 + }, + { + "epoch": 2.001040402288885, + "grad_norm": 0.04675092548131943, + "learning_rate": 8.462e-05, + "loss": 0.0245, + "step": 5770 + }, + { + "epoch": 2.0045084099185018, + "grad_norm": 0.053521353751420975, + "learning_rate": 8.442e-05, + "loss": 0.0158, + "step": 5780 + }, + { + "epoch": 2.0079764175481185, + "grad_norm": 0.06453324854373932, + "learning_rate": 8.422e-05, + "loss": 0.0131, + "step": 5790 + }, + { + "epoch": 2.0114444251777353, + "grad_norm": 0.0763096958398819, + "learning_rate": 8.402e-05, + "loss": 0.0112, + "step": 5800 + }, + { + "epoch": 2.014912432807352, + "grad_norm": 0.0837683454155922, + "learning_rate": 8.382e-05, + "loss": 0.0135, + "step": 5810 + }, + { + "epoch": 2.018380440436969, + "grad_norm": 0.0868675634264946, + "learning_rate": 8.362000000000002e-05, + "loss": 0.0116, + "step": 5820 + }, + { + "epoch": 2.021848448066586, + "grad_norm": 0.06839966773986816, + "learning_rate": 8.342000000000001e-05, + "loss": 0.011, + "step": 5830 + }, + { + "epoch": 2.0253164556962027, + "grad_norm": 0.10859765112400055, + "learning_rate": 8.322e-05, + "loss": 0.0132, + "step": 5840 + }, + { + "epoch": 2.0287844633258194, + "grad_norm": 0.08824854344129562, + "learning_rate": 8.302000000000001e-05, + "loss": 0.0129, + "step": 5850 + }, + { + "epoch": 2.032252470955436, + "grad_norm": 0.0745101124048233, + "learning_rate": 8.282000000000001e-05, + "loss": 0.013, + "step": 5860 + }, + { + "epoch": 2.035720478585053, + "grad_norm": 0.06870684772729874, + "learning_rate": 8.262e-05, + "loss": 0.0126, + "step": 5870 + }, + { + "epoch": 2.0391884862146696, + "grad_norm": 0.08353777974843979, + "learning_rate": 8.242000000000001e-05, + "loss": 0.0127, + "step": 5880 + }, + { + "epoch": 2.0426564938442864, + "grad_norm": 0.07282493263483047, + "learning_rate": 8.222000000000001e-05, + "loss": 0.014, + "step": 5890 + }, + { + "epoch": 2.046124501473903, + "grad_norm": 0.07777410745620728, + "learning_rate": 8.202e-05, + "loss": 0.0108, + "step": 5900 + }, + { + "epoch": 2.04959250910352, + "grad_norm": 0.06361842900514603, + "learning_rate": 8.182000000000001e-05, + "loss": 0.0132, + "step": 5910 + }, + { + "epoch": 2.0530605167331366, + "grad_norm": 0.09724973887205124, + "learning_rate": 8.162000000000001e-05, + "loss": 0.0139, + "step": 5920 + }, + { + "epoch": 2.056528524362754, + "grad_norm": 0.09359490126371384, + "learning_rate": 8.142e-05, + "loss": 0.013, + "step": 5930 + }, + { + "epoch": 2.0599965319923705, + "grad_norm": 0.06116607412695885, + "learning_rate": 8.122000000000001e-05, + "loss": 0.0143, + "step": 5940 + }, + { + "epoch": 2.0634645396219873, + "grad_norm": 0.06787212938070297, + "learning_rate": 8.102000000000001e-05, + "loss": 0.0137, + "step": 5950 + }, + { + "epoch": 2.066932547251604, + "grad_norm": 0.06523068249225616, + "learning_rate": 8.082e-05, + "loss": 0.0116, + "step": 5960 + }, + { + "epoch": 2.0704005548812208, + "grad_norm": 0.09076279401779175, + "learning_rate": 8.062000000000001e-05, + "loss": 0.01, + "step": 5970 + }, + { + "epoch": 2.0738685625108375, + "grad_norm": 0.07997199892997742, + "learning_rate": 8.042e-05, + "loss": 0.0132, + "step": 5980 + }, + { + "epoch": 2.0773365701404543, + "grad_norm": 0.0561593659222126, + "learning_rate": 8.022e-05, + "loss": 0.0117, + "step": 5990 + }, + { + "epoch": 2.080804577770071, + "grad_norm": 0.08588268607854843, + "learning_rate": 8.002000000000001e-05, + "loss": 0.0125, + "step": 6000 + }, + { + "epoch": 2.080804577770071, + "eval_loss": 0.06557230651378632, + "eval_runtime": 707.733, + "eval_samples_per_second": 12.974, + "eval_steps_per_second": 1.622, + "step": 6000 + }, + { + "epoch": 2.0842725853996877, + "grad_norm": 0.09114642441272736, + "learning_rate": 7.982e-05, + "loss": 0.0107, + "step": 6010 + }, + { + "epoch": 2.0877405930293045, + "grad_norm": 0.06732100248336792, + "learning_rate": 7.962e-05, + "loss": 0.012, + "step": 6020 + }, + { + "epoch": 2.0912086006589217, + "grad_norm": 0.08021605014801025, + "learning_rate": 7.942000000000001e-05, + "loss": 0.0123, + "step": 6030 + }, + { + "epoch": 2.0946766082885384, + "grad_norm": 0.06761088222265244, + "learning_rate": 7.922e-05, + "loss": 0.0117, + "step": 6040 + }, + { + "epoch": 2.098144615918155, + "grad_norm": 0.11337202787399292, + "learning_rate": 7.902e-05, + "loss": 0.0131, + "step": 6050 + }, + { + "epoch": 2.101612623547772, + "grad_norm": 0.09859013557434082, + "learning_rate": 7.882000000000001e-05, + "loss": 0.0139, + "step": 6060 + }, + { + "epoch": 2.1050806311773886, + "grad_norm": 0.08745191991329193, + "learning_rate": 7.862e-05, + "loss": 0.0139, + "step": 6070 + }, + { + "epoch": 2.1085486388070054, + "grad_norm": 0.04903840273618698, + "learning_rate": 7.842e-05, + "loss": 0.0142, + "step": 6080 + }, + { + "epoch": 2.112016646436622, + "grad_norm": 0.07992135733366013, + "learning_rate": 7.822e-05, + "loss": 0.0123, + "step": 6090 + }, + { + "epoch": 2.115484654066239, + "grad_norm": 0.08518462628126144, + "learning_rate": 7.802e-05, + "loss": 0.0138, + "step": 6100 + }, + { + "epoch": 2.1189526616958556, + "grad_norm": 0.08434431999921799, + "learning_rate": 7.782000000000001e-05, + "loss": 0.0155, + "step": 6110 + }, + { + "epoch": 2.1224206693254724, + "grad_norm": 0.08734823763370514, + "learning_rate": 7.762e-05, + "loss": 0.0125, + "step": 6120 + }, + { + "epoch": 2.125888676955089, + "grad_norm": 0.08129347115755081, + "learning_rate": 7.742e-05, + "loss": 0.0118, + "step": 6130 + }, + { + "epoch": 2.1293566845847063, + "grad_norm": 0.09434698522090912, + "learning_rate": 7.722000000000001e-05, + "loss": 0.0128, + "step": 6140 + }, + { + "epoch": 2.132824692214323, + "grad_norm": 0.08226180821657181, + "learning_rate": 7.702e-05, + "loss": 0.0159, + "step": 6150 + }, + { + "epoch": 2.1362926998439398, + "grad_norm": 0.06351976841688156, + "learning_rate": 7.682e-05, + "loss": 0.0135, + "step": 6160 + }, + { + "epoch": 2.1397607074735565, + "grad_norm": 0.07568191736936569, + "learning_rate": 7.662000000000001e-05, + "loss": 0.0122, + "step": 6170 + }, + { + "epoch": 2.1432287151031733, + "grad_norm": 0.08052569627761841, + "learning_rate": 7.642e-05, + "loss": 0.0123, + "step": 6180 + }, + { + "epoch": 2.14669672273279, + "grad_norm": 0.08767958730459213, + "learning_rate": 7.622e-05, + "loss": 0.0145, + "step": 6190 + }, + { + "epoch": 2.1501647303624067, + "grad_norm": 0.09768091887235641, + "learning_rate": 7.602000000000001e-05, + "loss": 0.0113, + "step": 6200 + }, + { + "epoch": 2.1536327379920235, + "grad_norm": 0.09293138980865479, + "learning_rate": 7.582e-05, + "loss": 0.0111, + "step": 6210 + }, + { + "epoch": 2.1571007456216402, + "grad_norm": 0.07943341881036758, + "learning_rate": 7.562e-05, + "loss": 0.0133, + "step": 6220 + }, + { + "epoch": 2.160568753251257, + "grad_norm": 0.08316273242235184, + "learning_rate": 7.542e-05, + "loss": 0.0142, + "step": 6230 + }, + { + "epoch": 2.164036760880874, + "grad_norm": 0.05390129238367081, + "learning_rate": 7.522e-05, + "loss": 0.0128, + "step": 6240 + }, + { + "epoch": 2.167504768510491, + "grad_norm": 0.05505843088030815, + "learning_rate": 7.502e-05, + "loss": 0.0124, + "step": 6250 + }, + { + "epoch": 2.1709727761401076, + "grad_norm": 0.08427543193101883, + "learning_rate": 7.482e-05, + "loss": 0.0136, + "step": 6260 + }, + { + "epoch": 2.1744407837697244, + "grad_norm": 0.06911098957061768, + "learning_rate": 7.462e-05, + "loss": 0.0117, + "step": 6270 + }, + { + "epoch": 2.177908791399341, + "grad_norm": 0.08595360815525055, + "learning_rate": 7.442e-05, + "loss": 0.0155, + "step": 6280 + }, + { + "epoch": 2.181376799028958, + "grad_norm": 0.0827205702662468, + "learning_rate": 7.422e-05, + "loss": 0.0127, + "step": 6290 + }, + { + "epoch": 2.1848448066585746, + "grad_norm": 0.12057662755250931, + "learning_rate": 7.402e-05, + "loss": 0.0113, + "step": 6300 + }, + { + "epoch": 2.1883128142881914, + "grad_norm": 0.10641255229711533, + "learning_rate": 7.382e-05, + "loss": 0.0123, + "step": 6310 + }, + { + "epoch": 2.191780821917808, + "grad_norm": 0.11188361793756485, + "learning_rate": 7.362e-05, + "loss": 0.013, + "step": 6320 + }, + { + "epoch": 2.195248829547425, + "grad_norm": 0.06386271864175797, + "learning_rate": 7.342e-05, + "loss": 0.0114, + "step": 6330 + }, + { + "epoch": 2.1987168371770416, + "grad_norm": 0.09285665303468704, + "learning_rate": 7.322e-05, + "loss": 0.0156, + "step": 6340 + }, + { + "epoch": 2.2021848448066588, + "grad_norm": 0.0885651484131813, + "learning_rate": 7.302e-05, + "loss": 0.013, + "step": 6350 + }, + { + "epoch": 2.2056528524362755, + "grad_norm": 0.09753404557704926, + "learning_rate": 7.282e-05, + "loss": 0.0135, + "step": 6360 + }, + { + "epoch": 2.2091208600658923, + "grad_norm": 0.06870284676551819, + "learning_rate": 7.261999999999999e-05, + "loss": 0.0163, + "step": 6370 + }, + { + "epoch": 2.212588867695509, + "grad_norm": 0.07846927642822266, + "learning_rate": 7.242e-05, + "loss": 0.0147, + "step": 6380 + }, + { + "epoch": 2.2160568753251257, + "grad_norm": 0.06510089337825775, + "learning_rate": 7.222e-05, + "loss": 0.0129, + "step": 6390 + }, + { + "epoch": 2.2195248829547425, + "grad_norm": 0.13771073520183563, + "learning_rate": 7.202e-05, + "loss": 0.0151, + "step": 6400 + }, + { + "epoch": 2.2229928905843592, + "grad_norm": 0.08157498389482498, + "learning_rate": 7.182e-05, + "loss": 0.0136, + "step": 6410 + }, + { + "epoch": 2.226460898213976, + "grad_norm": 0.09444098174571991, + "learning_rate": 7.162e-05, + "loss": 0.016, + "step": 6420 + }, + { + "epoch": 2.2299289058435927, + "grad_norm": 0.0773581713438034, + "learning_rate": 7.142e-05, + "loss": 0.0132, + "step": 6430 + }, + { + "epoch": 2.2333969134732095, + "grad_norm": 0.10038639605045319, + "learning_rate": 7.122000000000001e-05, + "loss": 0.0142, + "step": 6440 + }, + { + "epoch": 2.2368649211028266, + "grad_norm": 0.10728047043085098, + "learning_rate": 7.102000000000001e-05, + "loss": 0.0152, + "step": 6450 + }, + { + "epoch": 2.2403329287324434, + "grad_norm": 0.10695026069879532, + "learning_rate": 7.082e-05, + "loss": 0.0163, + "step": 6460 + }, + { + "epoch": 2.24380093636206, + "grad_norm": 0.06996133178472519, + "learning_rate": 7.062000000000001e-05, + "loss": 0.0149, + "step": 6470 + }, + { + "epoch": 2.247268943991677, + "grad_norm": 0.04395058751106262, + "learning_rate": 7.042000000000001e-05, + "loss": 0.0127, + "step": 6480 + }, + { + "epoch": 2.2507369516212936, + "grad_norm": 0.06262753158807755, + "learning_rate": 7.022e-05, + "loss": 0.014, + "step": 6490 + }, + { + "epoch": 2.2542049592509104, + "grad_norm": 0.073272705078125, + "learning_rate": 7.002000000000001e-05, + "loss": 0.0121, + "step": 6500 + }, + { + "epoch": 2.257672966880527, + "grad_norm": 0.11415940523147583, + "learning_rate": 6.982e-05, + "loss": 0.0136, + "step": 6510 + }, + { + "epoch": 2.261140974510144, + "grad_norm": 0.09325289726257324, + "learning_rate": 6.962e-05, + "loss": 0.0121, + "step": 6520 + }, + { + "epoch": 2.2646089821397606, + "grad_norm": 0.07223498821258545, + "learning_rate": 6.942000000000001e-05, + "loss": 0.0143, + "step": 6530 + }, + { + "epoch": 2.2680769897693773, + "grad_norm": 0.08595094084739685, + "learning_rate": 6.922e-05, + "loss": 0.0137, + "step": 6540 + }, + { + "epoch": 2.271544997398994, + "grad_norm": 0.08120746910572052, + "learning_rate": 6.902000000000001e-05, + "loss": 0.0101, + "step": 6550 + }, + { + "epoch": 2.2750130050286113, + "grad_norm": 0.1085987389087677, + "learning_rate": 6.882000000000001e-05, + "loss": 0.0149, + "step": 6560 + }, + { + "epoch": 2.278481012658228, + "grad_norm": 0.07946083694696426, + "learning_rate": 6.862e-05, + "loss": 0.0133, + "step": 6570 + }, + { + "epoch": 2.2819490202878447, + "grad_norm": 0.05504854768514633, + "learning_rate": 6.842000000000001e-05, + "loss": 0.0139, + "step": 6580 + }, + { + "epoch": 2.2854170279174615, + "grad_norm": 0.07158561050891876, + "learning_rate": 6.822000000000001e-05, + "loss": 0.0135, + "step": 6590 + }, + { + "epoch": 2.2888850355470782, + "grad_norm": 0.06974880397319794, + "learning_rate": 6.802e-05, + "loss": 0.0123, + "step": 6600 + }, + { + "epoch": 2.292353043176695, + "grad_norm": 0.08541780710220337, + "learning_rate": 6.782000000000001e-05, + "loss": 0.0144, + "step": 6610 + }, + { + "epoch": 2.2958210508063117, + "grad_norm": 0.10203000158071518, + "learning_rate": 6.762e-05, + "loss": 0.0138, + "step": 6620 + }, + { + "epoch": 2.2992890584359285, + "grad_norm": 0.09905651211738586, + "learning_rate": 6.742e-05, + "loss": 0.0145, + "step": 6630 + }, + { + "epoch": 2.302757066065545, + "grad_norm": 0.04088059067726135, + "learning_rate": 6.722000000000001e-05, + "loss": 0.011, + "step": 6640 + }, + { + "epoch": 2.306225073695162, + "grad_norm": 0.10128574818372726, + "learning_rate": 6.702e-05, + "loss": 0.0159, + "step": 6650 + }, + { + "epoch": 2.309693081324779, + "grad_norm": 0.08455543220043182, + "learning_rate": 6.682e-05, + "loss": 0.0131, + "step": 6660 + }, + { + "epoch": 2.313161088954396, + "grad_norm": 0.1268063485622406, + "learning_rate": 6.662000000000001e-05, + "loss": 0.0129, + "step": 6670 + }, + { + "epoch": 2.3166290965840126, + "grad_norm": 0.11783897876739502, + "learning_rate": 6.642e-05, + "loss": 0.0106, + "step": 6680 + }, + { + "epoch": 2.3200971042136294, + "grad_norm": 0.09608971327543259, + "learning_rate": 6.622e-05, + "loss": 0.0137, + "step": 6690 + }, + { + "epoch": 2.323565111843246, + "grad_norm": 0.07495573163032532, + "learning_rate": 6.602000000000001e-05, + "loss": 0.0131, + "step": 6700 + }, + { + "epoch": 2.327033119472863, + "grad_norm": 0.07772886753082275, + "learning_rate": 6.582e-05, + "loss": 0.0124, + "step": 6710 + }, + { + "epoch": 2.3305011271024796, + "grad_norm": 0.07365009188652039, + "learning_rate": 6.562e-05, + "loss": 0.0112, + "step": 6720 + }, + { + "epoch": 2.3339691347320963, + "grad_norm": 0.0697893276810646, + "learning_rate": 6.542000000000001e-05, + "loss": 0.0105, + "step": 6730 + }, + { + "epoch": 2.337437142361713, + "grad_norm": 0.09057148545980453, + "learning_rate": 6.522e-05, + "loss": 0.0134, + "step": 6740 + }, + { + "epoch": 2.34090514999133, + "grad_norm": 0.09601489454507828, + "learning_rate": 6.502e-05, + "loss": 0.014, + "step": 6750 + }, + { + "epoch": 2.3443731576209466, + "grad_norm": 0.11969607323408127, + "learning_rate": 6.482e-05, + "loss": 0.0116, + "step": 6760 + }, + { + "epoch": 2.3478411652505637, + "grad_norm": 0.103757843375206, + "learning_rate": 6.462e-05, + "loss": 0.0088, + "step": 6770 + }, + { + "epoch": 2.3513091728801805, + "grad_norm": 0.09077152609825134, + "learning_rate": 6.442e-05, + "loss": 0.0111, + "step": 6780 + }, + { + "epoch": 2.3547771805097972, + "grad_norm": 0.06362780928611755, + "learning_rate": 6.422e-05, + "loss": 0.01, + "step": 6790 + }, + { + "epoch": 2.358245188139414, + "grad_norm": 0.10334885120391846, + "learning_rate": 6.402e-05, + "loss": 0.0144, + "step": 6800 + }, + { + "epoch": 2.3617131957690307, + "grad_norm": 0.0676029622554779, + "learning_rate": 6.382e-05, + "loss": 0.0131, + "step": 6810 + }, + { + "epoch": 2.3651812033986475, + "grad_norm": 0.06794017553329468, + "learning_rate": 6.362e-05, + "loss": 0.0136, + "step": 6820 + }, + { + "epoch": 2.368649211028264, + "grad_norm": 0.11005677282810211, + "learning_rate": 6.342e-05, + "loss": 0.0141, + "step": 6830 + }, + { + "epoch": 2.372117218657881, + "grad_norm": 0.07998326420783997, + "learning_rate": 6.322000000000001e-05, + "loss": 0.0119, + "step": 6840 + }, + { + "epoch": 2.3755852262874977, + "grad_norm": 0.07724535465240479, + "learning_rate": 6.302e-05, + "loss": 0.0149, + "step": 6850 + }, + { + "epoch": 2.3790532339171144, + "grad_norm": 0.07563537359237671, + "learning_rate": 6.282e-05, + "loss": 0.0145, + "step": 6860 + }, + { + "epoch": 2.3825212415467316, + "grad_norm": 0.048786722123622894, + "learning_rate": 6.262000000000001e-05, + "loss": 0.0122, + "step": 6870 + }, + { + "epoch": 2.3859892491763484, + "grad_norm": 0.09792380034923553, + "learning_rate": 6.242e-05, + "loss": 0.0133, + "step": 6880 + }, + { + "epoch": 2.389457256805965, + "grad_norm": 0.06414589285850525, + "learning_rate": 6.222e-05, + "loss": 0.0119, + "step": 6890 + }, + { + "epoch": 2.392925264435582, + "grad_norm": 0.0844031348824501, + "learning_rate": 6.202e-05, + "loss": 0.0139, + "step": 6900 + }, + { + "epoch": 2.3963932720651986, + "grad_norm": 0.0652434229850769, + "learning_rate": 6.182e-05, + "loss": 0.01, + "step": 6910 + }, + { + "epoch": 2.3998612796948153, + "grad_norm": 0.07981958985328674, + "learning_rate": 6.162e-05, + "loss": 0.0105, + "step": 6920 + }, + { + "epoch": 2.403329287324432, + "grad_norm": 0.064891017973423, + "learning_rate": 6.142e-05, + "loss": 0.0138, + "step": 6930 + }, + { + "epoch": 2.406797294954049, + "grad_norm": 0.07090698927640915, + "learning_rate": 6.122e-05, + "loss": 0.0103, + "step": 6940 + }, + { + "epoch": 2.4102653025836656, + "grad_norm": 0.09227363765239716, + "learning_rate": 6.102e-05, + "loss": 0.0146, + "step": 6950 + }, + { + "epoch": 2.4137333102132823, + "grad_norm": 0.056207556277513504, + "learning_rate": 6.082e-05, + "loss": 0.0143, + "step": 6960 + }, + { + "epoch": 2.417201317842899, + "grad_norm": 0.08221688121557236, + "learning_rate": 6.062e-05, + "loss": 0.0158, + "step": 6970 + }, + { + "epoch": 2.4206693254725162, + "grad_norm": 0.08280789107084274, + "learning_rate": 6.042e-05, + "loss": 0.0139, + "step": 6980 + }, + { + "epoch": 2.424137333102133, + "grad_norm": 0.06660090386867523, + "learning_rate": 6.0219999999999996e-05, + "loss": 0.013, + "step": 6990 + }, + { + "epoch": 2.4276053407317497, + "grad_norm": 0.10185576230287552, + "learning_rate": 6.002e-05, + "loss": 0.0153, + "step": 7000 + }, + { + "epoch": 2.4276053407317497, + "eval_loss": 0.0646032989025116, + "eval_runtime": 713.661, + "eval_samples_per_second": 12.866, + "eval_steps_per_second": 1.609, + "step": 7000 + }, + { + "epoch": 2.4310733483613665, + "grad_norm": 0.11863771080970764, + "learning_rate": 5.982e-05, + "loss": 0.0149, + "step": 7010 + }, + { + "epoch": 2.434541355990983, + "grad_norm": 0.08282611519098282, + "learning_rate": 5.9619999999999995e-05, + "loss": 0.0114, + "step": 7020 + }, + { + "epoch": 2.4380093636206, + "grad_norm": 0.10327205806970596, + "learning_rate": 5.942e-05, + "loss": 0.0113, + "step": 7030 + }, + { + "epoch": 2.4414773712502167, + "grad_norm": 0.07424825429916382, + "learning_rate": 5.922e-05, + "loss": 0.0149, + "step": 7040 + }, + { + "epoch": 2.4449453788798334, + "grad_norm": 0.07107970863580704, + "learning_rate": 5.902e-05, + "loss": 0.0133, + "step": 7050 + }, + { + "epoch": 2.44841338650945, + "grad_norm": 0.08729968965053558, + "learning_rate": 5.8819999999999996e-05, + "loss": 0.0136, + "step": 7060 + }, + { + "epoch": 2.451881394139067, + "grad_norm": 0.06300070136785507, + "learning_rate": 5.862000000000001e-05, + "loss": 0.01, + "step": 7070 + }, + { + "epoch": 2.455349401768684, + "grad_norm": 0.107483871281147, + "learning_rate": 5.8420000000000006e-05, + "loss": 0.0124, + "step": 7080 + }, + { + "epoch": 2.458817409398301, + "grad_norm": 0.06871318072080612, + "learning_rate": 5.822000000000001e-05, + "loss": 0.013, + "step": 7090 + }, + { + "epoch": 2.4622854170279176, + "grad_norm": 0.10566007345914841, + "learning_rate": 5.802000000000001e-05, + "loss": 0.0133, + "step": 7100 + }, + { + "epoch": 2.4657534246575343, + "grad_norm": 0.04511050507426262, + "learning_rate": 5.7820000000000005e-05, + "loss": 0.012, + "step": 7110 + }, + { + "epoch": 2.469221432287151, + "grad_norm": 0.09614109992980957, + "learning_rate": 5.762000000000001e-05, + "loss": 0.0115, + "step": 7120 + }, + { + "epoch": 2.472689439916768, + "grad_norm": 0.07863055169582367, + "learning_rate": 5.742000000000001e-05, + "loss": 0.0141, + "step": 7130 + }, + { + "epoch": 2.4761574475463846, + "grad_norm": 0.09857816249132156, + "learning_rate": 5.7220000000000004e-05, + "loss": 0.0131, + "step": 7140 + }, + { + "epoch": 2.4796254551760013, + "grad_norm": 0.11649773269891739, + "learning_rate": 5.7020000000000006e-05, + "loss": 0.014, + "step": 7150 + }, + { + "epoch": 2.483093462805618, + "grad_norm": 0.09917131811380386, + "learning_rate": 5.682000000000001e-05, + "loss": 0.0119, + "step": 7160 + }, + { + "epoch": 2.486561470435235, + "grad_norm": 0.090948186814785, + "learning_rate": 5.6620000000000003e-05, + "loss": 0.0131, + "step": 7170 + }, + { + "epoch": 2.4900294780648515, + "grad_norm": 0.10430017858743668, + "learning_rate": 5.6420000000000005e-05, + "loss": 0.0122, + "step": 7180 + }, + { + "epoch": 2.4934974856944687, + "grad_norm": 0.08023589104413986, + "learning_rate": 5.622000000000001e-05, + "loss": 0.0146, + "step": 7190 + }, + { + "epoch": 2.4969654933240855, + "grad_norm": 0.058440957218408585, + "learning_rate": 5.602000000000001e-05, + "loss": 0.0125, + "step": 7200 + }, + { + "epoch": 2.500433500953702, + "grad_norm": 0.07056768983602524, + "learning_rate": 5.5820000000000004e-05, + "loss": 0.0122, + "step": 7210 + }, + { + "epoch": 2.503901508583319, + "grad_norm": 0.09822002053260803, + "learning_rate": 5.5620000000000006e-05, + "loss": 0.0111, + "step": 7220 + }, + { + "epoch": 2.5073695162129357, + "grad_norm": 0.04473882168531418, + "learning_rate": 5.542000000000001e-05, + "loss": 0.0126, + "step": 7230 + }, + { + "epoch": 2.5108375238425524, + "grad_norm": 0.07200278341770172, + "learning_rate": 5.522e-05, + "loss": 0.0123, + "step": 7240 + }, + { + "epoch": 2.514305531472169, + "grad_norm": 0.054528553038835526, + "learning_rate": 5.5020000000000005e-05, + "loss": 0.0139, + "step": 7250 + }, + { + "epoch": 2.517773539101786, + "grad_norm": 0.06125443056225777, + "learning_rate": 5.482000000000001e-05, + "loss": 0.0136, + "step": 7260 + }, + { + "epoch": 2.5212415467314027, + "grad_norm": 0.0841405987739563, + "learning_rate": 5.462e-05, + "loss": 0.012, + "step": 7270 + }, + { + "epoch": 2.52470955436102, + "grad_norm": 0.05345413088798523, + "learning_rate": 5.4420000000000004e-05, + "loss": 0.0124, + "step": 7280 + }, + { + "epoch": 2.5281775619906366, + "grad_norm": 0.1266576498746872, + "learning_rate": 5.4220000000000006e-05, + "loss": 0.0115, + "step": 7290 + }, + { + "epoch": 2.5316455696202533, + "grad_norm": 0.06762862950563431, + "learning_rate": 5.402e-05, + "loss": 0.0144, + "step": 7300 + }, + { + "epoch": 2.53511357724987, + "grad_norm": 0.06965386867523193, + "learning_rate": 5.382e-05, + "loss": 0.0107, + "step": 7310 + }, + { + "epoch": 2.538581584879487, + "grad_norm": 0.13501428067684174, + "learning_rate": 5.3620000000000005e-05, + "loss": 0.0122, + "step": 7320 + }, + { + "epoch": 2.5420495925091036, + "grad_norm": 0.0995524674654007, + "learning_rate": 5.342e-05, + "loss": 0.0129, + "step": 7330 + }, + { + "epoch": 2.5455176001387203, + "grad_norm": 0.10781200975179672, + "learning_rate": 5.322e-05, + "loss": 0.0123, + "step": 7340 + }, + { + "epoch": 2.548985607768337, + "grad_norm": 0.08666220307350159, + "learning_rate": 5.3020000000000004e-05, + "loss": 0.0128, + "step": 7350 + }, + { + "epoch": 2.552453615397954, + "grad_norm": 0.0817803218960762, + "learning_rate": 5.2820000000000006e-05, + "loss": 0.0133, + "step": 7360 + }, + { + "epoch": 2.5559216230275705, + "grad_norm": 0.08346541970968246, + "learning_rate": 5.262e-05, + "loss": 0.0142, + "step": 7370 + }, + { + "epoch": 2.5593896306571873, + "grad_norm": 0.08062835782766342, + "learning_rate": 5.242e-05, + "loss": 0.0108, + "step": 7380 + }, + { + "epoch": 2.562857638286804, + "grad_norm": 0.07476343959569931, + "learning_rate": 5.2220000000000005e-05, + "loss": 0.0129, + "step": 7390 + }, + { + "epoch": 2.5663256459164208, + "grad_norm": 0.0925203412771225, + "learning_rate": 5.202e-05, + "loss": 0.0137, + "step": 7400 + }, + { + "epoch": 2.569793653546038, + "grad_norm": 0.06427571177482605, + "learning_rate": 5.182e-05, + "loss": 0.012, + "step": 7410 + }, + { + "epoch": 2.5732616611756547, + "grad_norm": 0.08569345623254776, + "learning_rate": 5.1620000000000004e-05, + "loss": 0.0103, + "step": 7420 + }, + { + "epoch": 2.5767296688052714, + "grad_norm": 0.09565524011850357, + "learning_rate": 5.142e-05, + "loss": 0.0099, + "step": 7430 + }, + { + "epoch": 2.580197676434888, + "grad_norm": 0.09814833849668503, + "learning_rate": 5.122e-05, + "loss": 0.0125, + "step": 7440 + }, + { + "epoch": 2.583665684064505, + "grad_norm": 0.09652476012706757, + "learning_rate": 5.102e-05, + "loss": 0.0139, + "step": 7450 + }, + { + "epoch": 2.5871336916941217, + "grad_norm": 0.09088584780693054, + "learning_rate": 5.082e-05, + "loss": 0.0129, + "step": 7460 + }, + { + "epoch": 2.5906016993237384, + "grad_norm": 0.0785548985004425, + "learning_rate": 5.062e-05, + "loss": 0.0119, + "step": 7470 + }, + { + "epoch": 2.594069706953355, + "grad_norm": 0.0708330050110817, + "learning_rate": 5.042e-05, + "loss": 0.0121, + "step": 7480 + }, + { + "epoch": 2.5975377145829723, + "grad_norm": 0.09456871449947357, + "learning_rate": 5.0220000000000004e-05, + "loss": 0.0127, + "step": 7490 + }, + { + "epoch": 2.601005722212589, + "grad_norm": 0.05804060399532318, + "learning_rate": 5.002e-05, + "loss": 0.012, + "step": 7500 + }, + { + "epoch": 2.604473729842206, + "grad_norm": 0.08347133547067642, + "learning_rate": 4.982e-05, + "loss": 0.0121, + "step": 7510 + }, + { + "epoch": 2.6079417374718226, + "grad_norm": 0.10097778588533401, + "learning_rate": 4.962e-05, + "loss": 0.0116, + "step": 7520 + }, + { + "epoch": 2.6114097451014393, + "grad_norm": 0.07667971402406693, + "learning_rate": 4.942e-05, + "loss": 0.0144, + "step": 7530 + }, + { + "epoch": 2.614877752731056, + "grad_norm": 0.09747060388326645, + "learning_rate": 4.9220000000000006e-05, + "loss": 0.0115, + "step": 7540 + }, + { + "epoch": 2.618345760360673, + "grad_norm": 0.07479839771986008, + "learning_rate": 4.902e-05, + "loss": 0.0127, + "step": 7550 + }, + { + "epoch": 2.6218137679902895, + "grad_norm": 0.11268935352563858, + "learning_rate": 4.8820000000000004e-05, + "loss": 0.0143, + "step": 7560 + }, + { + "epoch": 2.6252817756199063, + "grad_norm": 0.0846814289689064, + "learning_rate": 4.8620000000000005e-05, + "loss": 0.0133, + "step": 7570 + }, + { + "epoch": 2.628749783249523, + "grad_norm": 0.10621094703674316, + "learning_rate": 4.842000000000001e-05, + "loss": 0.013, + "step": 7580 + }, + { + "epoch": 2.6322177908791398, + "grad_norm": 0.07603222131729126, + "learning_rate": 4.822e-05, + "loss": 0.0124, + "step": 7590 + }, + { + "epoch": 2.6356857985087565, + "grad_norm": 0.08150995522737503, + "learning_rate": 4.8020000000000004e-05, + "loss": 0.013, + "step": 7600 + }, + { + "epoch": 2.6391538061383732, + "grad_norm": 0.08962032943964005, + "learning_rate": 4.7820000000000006e-05, + "loss": 0.0119, + "step": 7610 + }, + { + "epoch": 2.6426218137679904, + "grad_norm": 0.07465488463640213, + "learning_rate": 4.762e-05, + "loss": 0.0116, + "step": 7620 + }, + { + "epoch": 2.646089821397607, + "grad_norm": 0.1040426716208458, + "learning_rate": 4.742e-05, + "loss": 0.0115, + "step": 7630 + }, + { + "epoch": 2.649557829027224, + "grad_norm": 0.14476630091667175, + "learning_rate": 4.7220000000000005e-05, + "loss": 0.0114, + "step": 7640 + }, + { + "epoch": 2.6530258366568407, + "grad_norm": 0.07727625221014023, + "learning_rate": 4.702e-05, + "loss": 0.0122, + "step": 7650 + }, + { + "epoch": 2.6564938442864574, + "grad_norm": 0.0987255796790123, + "learning_rate": 4.682e-05, + "loss": 0.0132, + "step": 7660 + }, + { + "epoch": 2.659961851916074, + "grad_norm": 0.07335236668586731, + "learning_rate": 4.6620000000000004e-05, + "loss": 0.0126, + "step": 7670 + }, + { + "epoch": 2.663429859545691, + "grad_norm": 0.09524571150541306, + "learning_rate": 4.642e-05, + "loss": 0.0114, + "step": 7680 + }, + { + "epoch": 2.6668978671753076, + "grad_norm": 0.11405625939369202, + "learning_rate": 4.622e-05, + "loss": 0.0144, + "step": 7690 + }, + { + "epoch": 2.670365874804925, + "grad_norm": 0.09738898277282715, + "learning_rate": 4.602e-05, + "loss": 0.0139, + "step": 7700 + }, + { + "epoch": 2.6738338824345416, + "grad_norm": 0.08633222430944443, + "learning_rate": 4.5820000000000005e-05, + "loss": 0.0136, + "step": 7710 + }, + { + "epoch": 2.6773018900641583, + "grad_norm": 0.09597263485193253, + "learning_rate": 4.562e-05, + "loss": 0.015, + "step": 7720 + }, + { + "epoch": 2.680769897693775, + "grad_norm": 0.0945490151643753, + "learning_rate": 4.542e-05, + "loss": 0.012, + "step": 7730 + }, + { + "epoch": 2.684237905323392, + "grad_norm": 0.0953177884221077, + "learning_rate": 4.5220000000000004e-05, + "loss": 0.0121, + "step": 7740 + }, + { + "epoch": 2.6877059129530085, + "grad_norm": 0.09284798800945282, + "learning_rate": 4.502e-05, + "loss": 0.0128, + "step": 7750 + }, + { + "epoch": 2.6911739205826253, + "grad_norm": 0.10929513722658157, + "learning_rate": 4.482e-05, + "loss": 0.0122, + "step": 7760 + }, + { + "epoch": 2.694641928212242, + "grad_norm": 0.10034655779600143, + "learning_rate": 4.462e-05, + "loss": 0.0137, + "step": 7770 + }, + { + "epoch": 2.6981099358418588, + "grad_norm": 0.06497249752283096, + "learning_rate": 4.442e-05, + "loss": 0.0116, + "step": 7780 + }, + { + "epoch": 2.7015779434714755, + "grad_norm": 0.061711184680461884, + "learning_rate": 4.422e-05, + "loss": 0.0118, + "step": 7790 + }, + { + "epoch": 2.7050459511010922, + "grad_norm": 0.08474565297365189, + "learning_rate": 4.402e-05, + "loss": 0.0113, + "step": 7800 + }, + { + "epoch": 2.708513958730709, + "grad_norm": 0.06698320806026459, + "learning_rate": 4.382e-05, + "loss": 0.0116, + "step": 7810 + }, + { + "epoch": 2.7119819663603257, + "grad_norm": 0.11040494590997696, + "learning_rate": 4.362e-05, + "loss": 0.014, + "step": 7820 + }, + { + "epoch": 2.715449973989943, + "grad_norm": 0.077610544860363, + "learning_rate": 4.342e-05, + "loss": 0.0148, + "step": 7830 + }, + { + "epoch": 2.7189179816195597, + "grad_norm": 0.07504422217607498, + "learning_rate": 4.3219999999999996e-05, + "loss": 0.0109, + "step": 7840 + }, + { + "epoch": 2.7223859892491764, + "grad_norm": 0.083879753947258, + "learning_rate": 4.3020000000000005e-05, + "loss": 0.0136, + "step": 7850 + }, + { + "epoch": 2.725853996878793, + "grad_norm": 0.07689789682626724, + "learning_rate": 4.282000000000001e-05, + "loss": 0.013, + "step": 7860 + }, + { + "epoch": 2.72932200450841, + "grad_norm": 0.07872592657804489, + "learning_rate": 4.262e-05, + "loss": 0.0123, + "step": 7870 + }, + { + "epoch": 2.7327900121380266, + "grad_norm": 0.10761665552854538, + "learning_rate": 4.2420000000000004e-05, + "loss": 0.0141, + "step": 7880 + }, + { + "epoch": 2.7362580197676434, + "grad_norm": 0.07873303443193436, + "learning_rate": 4.2220000000000006e-05, + "loss": 0.0114, + "step": 7890 + }, + { + "epoch": 2.73972602739726, + "grad_norm": 0.05685030296444893, + "learning_rate": 4.202e-05, + "loss": 0.0102, + "step": 7900 + }, + { + "epoch": 2.7431940350268773, + "grad_norm": 0.10359474271535873, + "learning_rate": 4.182e-05, + "loss": 0.0131, + "step": 7910 + }, + { + "epoch": 2.746662042656494, + "grad_norm": 0.07630922645330429, + "learning_rate": 4.1620000000000005e-05, + "loss": 0.0115, + "step": 7920 + }, + { + "epoch": 2.750130050286111, + "grad_norm": 0.09103643894195557, + "learning_rate": 4.142000000000001e-05, + "loss": 0.012, + "step": 7930 + }, + { + "epoch": 2.7535980579157275, + "grad_norm": 0.08436847478151321, + "learning_rate": 4.122e-05, + "loss": 0.0115, + "step": 7940 + }, + { + "epoch": 2.7570660655453443, + "grad_norm": 0.09147916734218597, + "learning_rate": 4.1020000000000004e-05, + "loss": 0.0164, + "step": 7950 + }, + { + "epoch": 2.760534073174961, + "grad_norm": 0.1505836397409439, + "learning_rate": 4.0820000000000006e-05, + "loss": 0.0136, + "step": 7960 + }, + { + "epoch": 2.7640020808045778, + "grad_norm": 0.06953336298465729, + "learning_rate": 4.062e-05, + "loss": 0.0114, + "step": 7970 + }, + { + "epoch": 2.7674700884341945, + "grad_norm": 0.08808320015668869, + "learning_rate": 4.042e-05, + "loss": 0.0095, + "step": 7980 + }, + { + "epoch": 2.7709380960638113, + "grad_norm": 0.10280350595712662, + "learning_rate": 4.0220000000000005e-05, + "loss": 0.0095, + "step": 7990 + }, + { + "epoch": 2.774406103693428, + "grad_norm": 0.11173941940069199, + "learning_rate": 4.002e-05, + "loss": 0.0128, + "step": 8000 + }, + { + "epoch": 2.774406103693428, + "eval_loss": 0.06569719314575195, + "eval_runtime": 714.8383, + "eval_samples_per_second": 12.845, + "eval_steps_per_second": 1.606, + "step": 8000 + }, + { + "epoch": 2.7778741113230447, + "grad_norm": 0.0838409885764122, + "learning_rate": 3.982e-05, + "loss": 0.0132, + "step": 8010 + }, + { + "epoch": 2.7813421189526615, + "grad_norm": 0.04326736554503441, + "learning_rate": 3.9620000000000004e-05, + "loss": 0.011, + "step": 8020 + }, + { + "epoch": 2.7848101265822782, + "grad_norm": 0.054840609431266785, + "learning_rate": 3.942e-05, + "loss": 0.0104, + "step": 8030 + }, + { + "epoch": 2.7882781342118954, + "grad_norm": 0.055376045405864716, + "learning_rate": 3.922e-05, + "loss": 0.012, + "step": 8040 + }, + { + "epoch": 2.791746141841512, + "grad_norm": 0.10542810708284378, + "learning_rate": 3.902e-05, + "loss": 0.0143, + "step": 8050 + }, + { + "epoch": 2.795214149471129, + "grad_norm": 0.09201773256063461, + "learning_rate": 3.882e-05, + "loss": 0.0131, + "step": 8060 + }, + { + "epoch": 2.7986821571007456, + "grad_norm": 0.10081171244382858, + "learning_rate": 3.862e-05, + "loss": 0.0132, + "step": 8070 + }, + { + "epoch": 2.8021501647303624, + "grad_norm": 0.13332189619541168, + "learning_rate": 3.842e-05, + "loss": 0.0119, + "step": 8080 + }, + { + "epoch": 2.805618172359979, + "grad_norm": 0.060715481638908386, + "learning_rate": 3.822e-05, + "loss": 0.0132, + "step": 8090 + }, + { + "epoch": 2.809086179989596, + "grad_norm": 0.08016696572303772, + "learning_rate": 3.802e-05, + "loss": 0.0118, + "step": 8100 + }, + { + "epoch": 2.8125541876192126, + "grad_norm": 0.07697981595993042, + "learning_rate": 3.782e-05, + "loss": 0.0125, + "step": 8110 + }, + { + "epoch": 2.81602219524883, + "grad_norm": 0.10400475561618805, + "learning_rate": 3.762e-05, + "loss": 0.0096, + "step": 8120 + }, + { + "epoch": 2.8194902028784465, + "grad_norm": 0.09359320998191833, + "learning_rate": 3.742e-05, + "loss": 0.0124, + "step": 8130 + }, + { + "epoch": 2.8229582105080633, + "grad_norm": 0.08394026756286621, + "learning_rate": 3.722e-05, + "loss": 0.0101, + "step": 8140 + }, + { + "epoch": 2.82642621813768, + "grad_norm": 0.08613195270299911, + "learning_rate": 3.702e-05, + "loss": 0.0138, + "step": 8150 + }, + { + "epoch": 2.8298942257672968, + "grad_norm": 0.1419047713279724, + "learning_rate": 3.682e-05, + "loss": 0.0109, + "step": 8160 + }, + { + "epoch": 2.8333622333969135, + "grad_norm": 0.08159425854682922, + "learning_rate": 3.6620000000000005e-05, + "loss": 0.0117, + "step": 8170 + }, + { + "epoch": 2.8368302410265303, + "grad_norm": 0.07656218111515045, + "learning_rate": 3.642000000000001e-05, + "loss": 0.014, + "step": 8180 + }, + { + "epoch": 2.840298248656147, + "grad_norm": 0.07194588333368301, + "learning_rate": 3.622e-05, + "loss": 0.0097, + "step": 8190 + }, + { + "epoch": 2.8437662562857637, + "grad_norm": 0.09303127229213715, + "learning_rate": 3.6020000000000004e-05, + "loss": 0.0098, + "step": 8200 + }, + { + "epoch": 2.8472342639153805, + "grad_norm": 0.07557094097137451, + "learning_rate": 3.5820000000000006e-05, + "loss": 0.0105, + "step": 8210 + }, + { + "epoch": 2.8507022715449972, + "grad_norm": 0.07813210785388947, + "learning_rate": 3.562e-05, + "loss": 0.0108, + "step": 8220 + }, + { + "epoch": 2.854170279174614, + "grad_norm": 0.1083005964756012, + "learning_rate": 3.542e-05, + "loss": 0.0109, + "step": 8230 + }, + { + "epoch": 2.8576382868042307, + "grad_norm": 0.11126165091991425, + "learning_rate": 3.5220000000000005e-05, + "loss": 0.0117, + "step": 8240 + }, + { + "epoch": 2.861106294433848, + "grad_norm": 0.07165870070457458, + "learning_rate": 3.502e-05, + "loss": 0.0101, + "step": 8250 + }, + { + "epoch": 2.8645743020634646, + "grad_norm": 0.11130272597074509, + "learning_rate": 3.482e-05, + "loss": 0.0132, + "step": 8260 + }, + { + "epoch": 2.8680423096930814, + "grad_norm": 0.1115003302693367, + "learning_rate": 3.4620000000000004e-05, + "loss": 0.0157, + "step": 8270 + }, + { + "epoch": 2.871510317322698, + "grad_norm": 0.11023005098104477, + "learning_rate": 3.442e-05, + "loss": 0.0117, + "step": 8280 + }, + { + "epoch": 2.874978324952315, + "grad_norm": 0.11513439565896988, + "learning_rate": 3.422e-05, + "loss": 0.0132, + "step": 8290 + }, + { + "epoch": 2.8784463325819316, + "grad_norm": 0.07337312400341034, + "learning_rate": 3.402e-05, + "loss": 0.011, + "step": 8300 + }, + { + "epoch": 2.8819143402115484, + "grad_norm": 0.08071292191743851, + "learning_rate": 3.3820000000000005e-05, + "loss": 0.012, + "step": 8310 + }, + { + "epoch": 2.885382347841165, + "grad_norm": 0.09460859000682831, + "learning_rate": 3.362e-05, + "loss": 0.0118, + "step": 8320 + }, + { + "epoch": 2.8888503554707823, + "grad_norm": 0.08895740658044815, + "learning_rate": 3.342e-05, + "loss": 0.011, + "step": 8330 + }, + { + "epoch": 2.892318363100399, + "grad_norm": 0.08386581391096115, + "learning_rate": 3.3220000000000004e-05, + "loss": 0.0108, + "step": 8340 + }, + { + "epoch": 2.8957863707300158, + "grad_norm": 0.09314845502376556, + "learning_rate": 3.302e-05, + "loss": 0.0131, + "step": 8350 + }, + { + "epoch": 2.8992543783596325, + "grad_norm": 0.06327735632658005, + "learning_rate": 3.282e-05, + "loss": 0.0101, + "step": 8360 + }, + { + "epoch": 2.9027223859892493, + "grad_norm": 0.07077737152576447, + "learning_rate": 3.262e-05, + "loss": 0.0109, + "step": 8370 + }, + { + "epoch": 2.906190393618866, + "grad_norm": 0.07617966085672379, + "learning_rate": 3.242e-05, + "loss": 0.0104, + "step": 8380 + }, + { + "epoch": 2.9096584012484827, + "grad_norm": 0.10365297645330429, + "learning_rate": 3.222e-05, + "loss": 0.0114, + "step": 8390 + }, + { + "epoch": 2.9131264088780995, + "grad_norm": 0.07602348923683167, + "learning_rate": 3.202e-05, + "loss": 0.011, + "step": 8400 + }, + { + "epoch": 2.9165944165077162, + "grad_norm": 0.10498479008674622, + "learning_rate": 3.182e-05, + "loss": 0.0118, + "step": 8410 + }, + { + "epoch": 2.920062424137333, + "grad_norm": 0.08163636922836304, + "learning_rate": 3.162e-05, + "loss": 0.0121, + "step": 8420 + }, + { + "epoch": 2.9235304317669497, + "grad_norm": 0.09529927372932434, + "learning_rate": 3.142e-05, + "loss": 0.0128, + "step": 8430 + }, + { + "epoch": 2.9269984393965665, + "grad_norm": 0.10269023478031158, + "learning_rate": 3.122e-05, + "loss": 0.0102, + "step": 8440 + }, + { + "epoch": 2.930466447026183, + "grad_norm": 0.1097848117351532, + "learning_rate": 3.102e-05, + "loss": 0.0128, + "step": 8450 + }, + { + "epoch": 2.9339344546558004, + "grad_norm": 0.06423291563987732, + "learning_rate": 3.082e-05, + "loss": 0.0086, + "step": 8460 + }, + { + "epoch": 2.937402462285417, + "grad_norm": 0.10673278570175171, + "learning_rate": 3.062e-05, + "loss": 0.0128, + "step": 8470 + }, + { + "epoch": 2.940870469915034, + "grad_norm": 0.08257333934307098, + "learning_rate": 3.0420000000000004e-05, + "loss": 0.0122, + "step": 8480 + }, + { + "epoch": 2.9443384775446506, + "grad_norm": 0.08055904507637024, + "learning_rate": 3.0220000000000005e-05, + "loss": 0.0115, + "step": 8490 + }, + { + "epoch": 2.9478064851742674, + "grad_norm": 0.09221120178699493, + "learning_rate": 3.0020000000000004e-05, + "loss": 0.0117, + "step": 8500 + }, + { + "epoch": 2.951274492803884, + "grad_norm": 0.0831000804901123, + "learning_rate": 2.9820000000000002e-05, + "loss": 0.0108, + "step": 8510 + }, + { + "epoch": 2.954742500433501, + "grad_norm": 0.06930562108755112, + "learning_rate": 2.9620000000000004e-05, + "loss": 0.0117, + "step": 8520 + }, + { + "epoch": 2.9582105080631176, + "grad_norm": 0.07918361574411392, + "learning_rate": 2.9420000000000003e-05, + "loss": 0.0113, + "step": 8530 + }, + { + "epoch": 2.9616785156927348, + "grad_norm": 0.06908438354730606, + "learning_rate": 2.922e-05, + "loss": 0.0101, + "step": 8540 + }, + { + "epoch": 2.9651465233223515, + "grad_norm": 0.08593347668647766, + "learning_rate": 2.9020000000000003e-05, + "loss": 0.0127, + "step": 8550 + }, + { + "epoch": 2.9686145309519683, + "grad_norm": 0.09223376214504242, + "learning_rate": 2.8820000000000002e-05, + "loss": 0.0137, + "step": 8560 + }, + { + "epoch": 2.972082538581585, + "grad_norm": 0.09096557646989822, + "learning_rate": 2.8620000000000004e-05, + "loss": 0.0104, + "step": 8570 + }, + { + "epoch": 2.9755505462112017, + "grad_norm": 0.0837118849158287, + "learning_rate": 2.8420000000000002e-05, + "loss": 0.0113, + "step": 8580 + }, + { + "epoch": 2.9790185538408185, + "grad_norm": 0.060600802302360535, + "learning_rate": 2.822e-05, + "loss": 0.0119, + "step": 8590 + }, + { + "epoch": 2.9824865614704352, + "grad_norm": 0.09768325090408325, + "learning_rate": 2.8020000000000003e-05, + "loss": 0.0094, + "step": 8600 + }, + { + "epoch": 2.985954569100052, + "grad_norm": 0.10265090316534042, + "learning_rate": 2.782e-05, + "loss": 0.0117, + "step": 8610 + }, + { + "epoch": 2.9894225767296687, + "grad_norm": 0.11693336814641953, + "learning_rate": 2.762e-05, + "loss": 0.0126, + "step": 8620 + }, + { + "epoch": 2.9928905843592855, + "grad_norm": 0.06177123263478279, + "learning_rate": 2.7420000000000002e-05, + "loss": 0.0112, + "step": 8630 + }, + { + "epoch": 2.996358591988902, + "grad_norm": 0.07003209739923477, + "learning_rate": 2.722e-05, + "loss": 0.0111, + "step": 8640 + }, + { + "epoch": 2.999826599618519, + "grad_norm": 0.09619972854852676, + "learning_rate": 2.7020000000000002e-05, + "loss": 0.0113, + "step": 8650 + }, + { + "epoch": 3.003294607248136, + "grad_norm": 0.04991637170314789, + "learning_rate": 2.682e-05, + "loss": 0.0045, + "step": 8660 + }, + { + "epoch": 3.006762614877753, + "grad_norm": 0.054620299488306046, + "learning_rate": 2.662e-05, + "loss": 0.0029, + "step": 8670 + }, + { + "epoch": 3.0102306225073696, + "grad_norm": 0.031793851405382156, + "learning_rate": 2.642e-05, + "loss": 0.003, + "step": 8680 + }, + { + "epoch": 3.0136986301369864, + "grad_norm": 0.03903364762663841, + "learning_rate": 2.622e-05, + "loss": 0.0024, + "step": 8690 + }, + { + "epoch": 3.017166637766603, + "grad_norm": 0.05518382415175438, + "learning_rate": 2.602e-05, + "loss": 0.0028, + "step": 8700 + }, + { + "epoch": 3.02063464539622, + "grad_norm": 0.02921813167631626, + "learning_rate": 2.582e-05, + "loss": 0.0018, + "step": 8710 + }, + { + "epoch": 3.0241026530258366, + "grad_norm": 0.08628415316343307, + "learning_rate": 2.562e-05, + "loss": 0.0027, + "step": 8720 + }, + { + "epoch": 3.0275706606554533, + "grad_norm": 0.034173641353845596, + "learning_rate": 2.542e-05, + "loss": 0.002, + "step": 8730 + }, + { + "epoch": 3.03103866828507, + "grad_norm": 0.03343451768159866, + "learning_rate": 2.522e-05, + "loss": 0.0028, + "step": 8740 + }, + { + "epoch": 3.034506675914687, + "grad_norm": 0.0506439134478569, + "learning_rate": 2.5019999999999998e-05, + "loss": 0.002, + "step": 8750 + }, + { + "epoch": 3.037974683544304, + "grad_norm": 0.011204255744814873, + "learning_rate": 2.4820000000000003e-05, + "loss": 0.0026, + "step": 8760 + }, + { + "epoch": 3.0414426911739207, + "grad_norm": 0.05876510962843895, + "learning_rate": 2.462e-05, + "loss": 0.0023, + "step": 8770 + }, + { + "epoch": 3.0449106988035375, + "grad_norm": 0.017792142927646637, + "learning_rate": 2.442e-05, + "loss": 0.0014, + "step": 8780 + }, + { + "epoch": 3.0483787064331542, + "grad_norm": 0.0634002834558487, + "learning_rate": 2.4220000000000002e-05, + "loss": 0.002, + "step": 8790 + }, + { + "epoch": 3.051846714062771, + "grad_norm": 0.10669626295566559, + "learning_rate": 2.402e-05, + "loss": 0.002, + "step": 8800 + }, + { + "epoch": 3.0553147216923877, + "grad_norm": 0.02126486971974373, + "learning_rate": 2.3820000000000002e-05, + "loss": 0.002, + "step": 8810 + }, + { + "epoch": 3.0587827293220045, + "grad_norm": 0.13548636436462402, + "learning_rate": 2.362e-05, + "loss": 0.0034, + "step": 8820 + }, + { + "epoch": 3.062250736951621, + "grad_norm": 0.04129517078399658, + "learning_rate": 2.342e-05, + "loss": 0.0032, + "step": 8830 + }, + { + "epoch": 3.065718744581238, + "grad_norm": 0.0712270438671112, + "learning_rate": 2.322e-05, + "loss": 0.0025, + "step": 8840 + }, + { + "epoch": 3.0691867522108547, + "grad_norm": 0.03855830430984497, + "learning_rate": 2.302e-05, + "loss": 0.0022, + "step": 8850 + }, + { + "epoch": 3.0726547598404714, + "grad_norm": 0.06443187594413757, + "learning_rate": 2.282e-05, + "loss": 0.0028, + "step": 8860 + }, + { + "epoch": 3.0761227674700886, + "grad_norm": 0.09318159520626068, + "learning_rate": 2.2620000000000004e-05, + "loss": 0.0021, + "step": 8870 + }, + { + "epoch": 3.0795907750997054, + "grad_norm": 0.06871949136257172, + "learning_rate": 2.2420000000000002e-05, + "loss": 0.0033, + "step": 8880 + }, + { + "epoch": 3.083058782729322, + "grad_norm": 0.031738586723804474, + "learning_rate": 2.222e-05, + "loss": 0.0016, + "step": 8890 + }, + { + "epoch": 3.086526790358939, + "grad_norm": 0.028125595301389694, + "learning_rate": 2.2020000000000003e-05, + "loss": 0.0026, + "step": 8900 + }, + { + "epoch": 3.0899947979885556, + "grad_norm": 0.046880729496479034, + "learning_rate": 2.182e-05, + "loss": 0.0021, + "step": 8910 + }, + { + "epoch": 3.0934628056181723, + "grad_norm": 0.06030425429344177, + "learning_rate": 2.162e-05, + "loss": 0.0029, + "step": 8920 + }, + { + "epoch": 3.096930813247789, + "grad_norm": 0.028561096638441086, + "learning_rate": 2.142e-05, + "loss": 0.0023, + "step": 8930 + }, + { + "epoch": 3.100398820877406, + "grad_norm": 0.04128929600119591, + "learning_rate": 2.122e-05, + "loss": 0.0022, + "step": 8940 + }, + { + "epoch": 3.1038668285070226, + "grad_norm": 0.04722205549478531, + "learning_rate": 2.1020000000000002e-05, + "loss": 0.0027, + "step": 8950 + }, + { + "epoch": 3.1073348361366393, + "grad_norm": 0.057234570384025574, + "learning_rate": 2.082e-05, + "loss": 0.0016, + "step": 8960 + }, + { + "epoch": 3.1108028437662565, + "grad_norm": 0.010785787366330624, + "learning_rate": 2.062e-05, + "loss": 0.0016, + "step": 8970 + }, + { + "epoch": 3.1142708513958732, + "grad_norm": 0.08110074698925018, + "learning_rate": 2.042e-05, + "loss": 0.0022, + "step": 8980 + }, + { + "epoch": 3.11773885902549, + "grad_norm": 0.04953372851014137, + "learning_rate": 2.022e-05, + "loss": 0.0022, + "step": 8990 + }, + { + "epoch": 3.1212068666551067, + "grad_norm": 0.022535376250743866, + "learning_rate": 2.002e-05, + "loss": 0.0015, + "step": 9000 + }, + { + "epoch": 3.1212068666551067, + "eval_loss": 0.0743083506822586, + "eval_runtime": 714.2967, + "eval_samples_per_second": 12.855, + "eval_steps_per_second": 1.607, + "step": 9000 + } + ], + "logging_steps": 10, + "max_steps": 10000, + "num_input_tokens_seen": 0, + "num_train_epochs": 4, + "save_steps": 1000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 6.85989473816475e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +}