{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9999106584472438, "eval_steps": 500, "global_step": 5596, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0001786831055123738, "grad_norm": 0.677041232585907, "learning_rate": 3.5714285714285716e-07, "loss": 2.5858, "step": 1 }, { "epoch": 0.000893415527561869, "grad_norm": 0.87285977602005, "learning_rate": 1.7857142857142857e-06, "loss": 2.8564, "step": 5 }, { "epoch": 0.001786831055123738, "grad_norm": 0.91197270154953, "learning_rate": 3.5714285714285714e-06, "loss": 2.7743, "step": 10 }, { "epoch": 0.002680246582685607, "grad_norm": 0.8713774681091309, "learning_rate": 5.357142857142857e-06, "loss": 2.7189, "step": 15 }, { "epoch": 0.003573662110247476, "grad_norm": 0.7496689558029175, "learning_rate": 7.142857142857143e-06, "loss": 2.7557, "step": 20 }, { "epoch": 0.004467077637809345, "grad_norm": 0.7549262642860413, "learning_rate": 8.92857142857143e-06, "loss": 2.7256, "step": 25 }, { "epoch": 0.005360493165371214, "grad_norm": 0.8223145008087158, "learning_rate": 1.0714285714285714e-05, "loss": 2.7319, "step": 30 }, { "epoch": 0.006253908692933083, "grad_norm": 0.7578197121620178, "learning_rate": 1.25e-05, "loss": 2.661, "step": 35 }, { "epoch": 0.007147324220494952, "grad_norm": 0.6663674712181091, "learning_rate": 1.4285714285714285e-05, "loss": 2.6396, "step": 40 }, { "epoch": 0.008040739748056821, "grad_norm": 0.5226909518241882, "learning_rate": 1.6071428571428572e-05, "loss": 2.5739, "step": 45 }, { "epoch": 0.00893415527561869, "grad_norm": 0.4958447515964508, "learning_rate": 1.785714285714286e-05, "loss": 2.4765, "step": 50 }, { "epoch": 0.00982757080318056, "grad_norm": 0.40554603934288025, "learning_rate": 1.9642857142857145e-05, "loss": 2.4275, "step": 55 }, { "epoch": 0.010720986330742428, "grad_norm": 0.43960413336753845, "learning_rate": 2.1428571428571428e-05, "loss": 2.3822, "step": 60 }, { "epoch": 0.011614401858304297, "grad_norm": 0.40701353549957275, "learning_rate": 2.3214285714285715e-05, "loss": 2.3326, "step": 65 }, { "epoch": 0.012507817385866166, "grad_norm": 0.3899136483669281, "learning_rate": 2.5e-05, "loss": 2.3172, "step": 70 }, { "epoch": 0.013401232913428035, "grad_norm": 0.444366455078125, "learning_rate": 2.6785714285714288e-05, "loss": 2.2801, "step": 75 }, { "epoch": 0.014294648440989904, "grad_norm": 0.43577924370765686, "learning_rate": 2.857142857142857e-05, "loss": 2.23, "step": 80 }, { "epoch": 0.015188063968551773, "grad_norm": 0.42386895418167114, "learning_rate": 3.0357142857142857e-05, "loss": 2.2422, "step": 85 }, { "epoch": 0.016081479496113642, "grad_norm": 0.4131146967411041, "learning_rate": 3.2142857142857144e-05, "loss": 2.1822, "step": 90 }, { "epoch": 0.016974895023675513, "grad_norm": 0.43191808462142944, "learning_rate": 3.392857142857143e-05, "loss": 2.1113, "step": 95 }, { "epoch": 0.01786831055123738, "grad_norm": 0.49910837411880493, "learning_rate": 3.571428571428572e-05, "loss": 2.1751, "step": 100 }, { "epoch": 0.01876172607879925, "grad_norm": 0.4678761065006256, "learning_rate": 3.7500000000000003e-05, "loss": 2.1023, "step": 105 }, { "epoch": 0.01965514160636112, "grad_norm": 0.5106930136680603, "learning_rate": 3.928571428571429e-05, "loss": 2.0749, "step": 110 }, { "epoch": 0.02054855713392299, "grad_norm": 0.5173302292823792, "learning_rate": 4.107142857142857e-05, "loss": 2.0946, "step": 115 }, { "epoch": 0.021441972661484857, "grad_norm": 0.5672593712806702, "learning_rate": 4.2857142857142856e-05, "loss": 2.0089, "step": 120 }, { "epoch": 0.022335388189046727, "grad_norm": 0.5483328700065613, "learning_rate": 4.464285714285715e-05, "loss": 2.0303, "step": 125 }, { "epoch": 0.023228803716608595, "grad_norm": 0.5528935790061951, "learning_rate": 4.642857142857143e-05, "loss": 1.9952, "step": 130 }, { "epoch": 0.024122219244170465, "grad_norm": 0.6560689210891724, "learning_rate": 4.8214285714285716e-05, "loss": 1.9854, "step": 135 }, { "epoch": 0.025015634771732333, "grad_norm": 0.6699485182762146, "learning_rate": 5e-05, "loss": 2.0175, "step": 140 }, { "epoch": 0.025909050299294203, "grad_norm": 0.6667202711105347, "learning_rate": 5.1785714285714296e-05, "loss": 1.9973, "step": 145 }, { "epoch": 0.02680246582685607, "grad_norm": 0.624862790107727, "learning_rate": 5.3571428571428575e-05, "loss": 1.9596, "step": 150 }, { "epoch": 0.02769588135441794, "grad_norm": 0.7823134064674377, "learning_rate": 5.535714285714286e-05, "loss": 1.9863, "step": 155 }, { "epoch": 0.02858929688197981, "grad_norm": 0.7916033864021301, "learning_rate": 5.714285714285714e-05, "loss": 1.9652, "step": 160 }, { "epoch": 0.02948271240954168, "grad_norm": 0.8077523708343506, "learning_rate": 5.8928571428571435e-05, "loss": 1.961, "step": 165 }, { "epoch": 0.030376127937103547, "grad_norm": 0.7293629050254822, "learning_rate": 6.0714285714285715e-05, "loss": 1.9324, "step": 170 }, { "epoch": 0.031269543464665414, "grad_norm": 0.7660003304481506, "learning_rate": 6.25e-05, "loss": 1.8978, "step": 175 }, { "epoch": 0.032162958992227285, "grad_norm": 0.7684458494186401, "learning_rate": 6.428571428571429e-05, "loss": 1.8884, "step": 180 }, { "epoch": 0.033056374519789156, "grad_norm": 0.8633774518966675, "learning_rate": 6.607142857142857e-05, "loss": 1.8827, "step": 185 }, { "epoch": 0.033949790047351026, "grad_norm": 0.918566882610321, "learning_rate": 6.785714285714286e-05, "loss": 1.8765, "step": 190 }, { "epoch": 0.03484320557491289, "grad_norm": 0.8094245791435242, "learning_rate": 6.964285714285715e-05, "loss": 1.8944, "step": 195 }, { "epoch": 0.03573662110247476, "grad_norm": 0.8820126056671143, "learning_rate": 7.142857142857143e-05, "loss": 1.8934, "step": 200 }, { "epoch": 0.03663003663003663, "grad_norm": 0.9279444813728333, "learning_rate": 7.321428571428571e-05, "loss": 1.8604, "step": 205 }, { "epoch": 0.0375234521575985, "grad_norm": 0.8216668963432312, "learning_rate": 7.500000000000001e-05, "loss": 1.8466, "step": 210 }, { "epoch": 0.038416867685160366, "grad_norm": 0.9710394740104675, "learning_rate": 7.67857142857143e-05, "loss": 1.8802, "step": 215 }, { "epoch": 0.03931028321272224, "grad_norm": 1.00575852394104, "learning_rate": 7.857142857142858e-05, "loss": 1.8543, "step": 220 }, { "epoch": 0.04020369874028411, "grad_norm": 0.9068517088890076, "learning_rate": 8.035714285714287e-05, "loss": 1.8208, "step": 225 }, { "epoch": 0.04109711426784598, "grad_norm": 0.9185478091239929, "learning_rate": 8.214285714285714e-05, "loss": 1.8395, "step": 230 }, { "epoch": 0.04199052979540784, "grad_norm": 0.8614683747291565, "learning_rate": 8.392857142857144e-05, "loss": 1.7931, "step": 235 }, { "epoch": 0.04288394532296971, "grad_norm": 1.0321149826049805, "learning_rate": 8.571428571428571e-05, "loss": 1.8134, "step": 240 }, { "epoch": 0.043777360850531584, "grad_norm": 0.997043788433075, "learning_rate": 8.75e-05, "loss": 1.8195, "step": 245 }, { "epoch": 0.044670776378093455, "grad_norm": 0.9486415386199951, "learning_rate": 8.92857142857143e-05, "loss": 1.811, "step": 250 }, { "epoch": 0.04556419190565532, "grad_norm": 0.8874170780181885, "learning_rate": 9.107142857142857e-05, "loss": 1.808, "step": 255 }, { "epoch": 0.04645760743321719, "grad_norm": 0.9631446003913879, "learning_rate": 9.285714285714286e-05, "loss": 1.7976, "step": 260 }, { "epoch": 0.04735102296077906, "grad_norm": 0.9562223553657532, "learning_rate": 9.464285714285715e-05, "loss": 1.7817, "step": 265 }, { "epoch": 0.04824443848834093, "grad_norm": 0.8817927241325378, "learning_rate": 9.642857142857143e-05, "loss": 1.7646, "step": 270 }, { "epoch": 0.049137854015902795, "grad_norm": 0.9726014137268066, "learning_rate": 9.821428571428572e-05, "loss": 1.8197, "step": 275 }, { "epoch": 0.050031269543464665, "grad_norm": 0.914252758026123, "learning_rate": 0.0001, "loss": 1.7556, "step": 280 }, { "epoch": 0.050924685071026536, "grad_norm": 0.9449843764305115, "learning_rate": 0.00010178571428571428, "loss": 1.7112, "step": 285 }, { "epoch": 0.05181810059858841, "grad_norm": 0.9346893429756165, "learning_rate": 0.00010357142857142859, "loss": 1.7619, "step": 290 }, { "epoch": 0.05271151612615027, "grad_norm": 0.8253093957901001, "learning_rate": 0.00010535714285714286, "loss": 1.7342, "step": 295 }, { "epoch": 0.05360493165371214, "grad_norm": 0.8210521936416626, "learning_rate": 0.00010714285714285715, "loss": 1.7528, "step": 300 }, { "epoch": 0.05449834718127401, "grad_norm": 0.7791708707809448, "learning_rate": 0.00010892857142857142, "loss": 1.7881, "step": 305 }, { "epoch": 0.05539176270883588, "grad_norm": 0.9075311422348022, "learning_rate": 0.00011071428571428572, "loss": 1.7569, "step": 310 }, { "epoch": 0.05628517823639775, "grad_norm": 0.8817685842514038, "learning_rate": 0.00011250000000000001, "loss": 1.7444, "step": 315 }, { "epoch": 0.05717859376395962, "grad_norm": 0.9592775106430054, "learning_rate": 0.00011428571428571428, "loss": 1.7344, "step": 320 }, { "epoch": 0.05807200929152149, "grad_norm": 0.8894655704498291, "learning_rate": 0.00011607142857142858, "loss": 1.738, "step": 325 }, { "epoch": 0.05896542481908336, "grad_norm": 0.9590314030647278, "learning_rate": 0.00011785714285714287, "loss": 1.7336, "step": 330 }, { "epoch": 0.05985884034664522, "grad_norm": 0.9731721878051758, "learning_rate": 0.00011964285714285714, "loss": 1.7242, "step": 335 }, { "epoch": 0.060752255874207094, "grad_norm": 1.0074090957641602, "learning_rate": 0.00012142857142857143, "loss": 1.7719, "step": 340 }, { "epoch": 0.061645671401768964, "grad_norm": 0.9376741051673889, "learning_rate": 0.00012321428571428572, "loss": 1.7299, "step": 345 }, { "epoch": 0.06253908692933083, "grad_norm": 0.8597344756126404, "learning_rate": 0.000125, "loss": 1.7378, "step": 350 }, { "epoch": 0.0634325024568927, "grad_norm": 0.8836060762405396, "learning_rate": 0.0001267857142857143, "loss": 1.7379, "step": 355 }, { "epoch": 0.06432591798445457, "grad_norm": 1.0358092784881592, "learning_rate": 0.00012857142857142858, "loss": 1.7175, "step": 360 }, { "epoch": 0.06521933351201643, "grad_norm": 0.8426574468612671, "learning_rate": 0.00013035714285714286, "loss": 1.6632, "step": 365 }, { "epoch": 0.06611274903957831, "grad_norm": 0.9615597128868103, "learning_rate": 0.00013214285714285715, "loss": 1.6667, "step": 370 }, { "epoch": 0.06700616456714018, "grad_norm": 0.9134716391563416, "learning_rate": 0.00013392857142857144, "loss": 1.7246, "step": 375 }, { "epoch": 0.06789958009470205, "grad_norm": 0.8629338145256042, "learning_rate": 0.00013571428571428572, "loss": 1.6696, "step": 380 }, { "epoch": 0.06879299562226392, "grad_norm": 0.8393704295158386, "learning_rate": 0.0001375, "loss": 1.6838, "step": 385 }, { "epoch": 0.06968641114982578, "grad_norm": 0.9799148440361023, "learning_rate": 0.0001392857142857143, "loss": 1.7022, "step": 390 }, { "epoch": 0.07057982667738766, "grad_norm": 0.8769751191139221, "learning_rate": 0.00014107142857142858, "loss": 1.6783, "step": 395 }, { "epoch": 0.07147324220494952, "grad_norm": 0.8460381627082825, "learning_rate": 0.00014285714285714287, "loss": 1.6509, "step": 400 }, { "epoch": 0.07236665773251139, "grad_norm": 0.7672830820083618, "learning_rate": 0.00014464285714285715, "loss": 1.6485, "step": 405 }, { "epoch": 0.07326007326007326, "grad_norm": 0.8759172558784485, "learning_rate": 0.00014642857142857141, "loss": 1.6723, "step": 410 }, { "epoch": 0.07415348878763513, "grad_norm": 0.8765602707862854, "learning_rate": 0.00014821428571428573, "loss": 1.5925, "step": 415 }, { "epoch": 0.075046904315197, "grad_norm": 0.8540732264518738, "learning_rate": 0.00015000000000000001, "loss": 1.6583, "step": 420 }, { "epoch": 0.07594031984275887, "grad_norm": 0.7945284843444824, "learning_rate": 0.00015178571428571427, "loss": 1.6644, "step": 425 }, { "epoch": 0.07683373537032073, "grad_norm": 1.0458513498306274, "learning_rate": 0.0001535714285714286, "loss": 1.6946, "step": 430 }, { "epoch": 0.07772715089788261, "grad_norm": 0.8617852330207825, "learning_rate": 0.00015535714285714287, "loss": 1.6218, "step": 435 }, { "epoch": 0.07862056642544447, "grad_norm": 0.8525850176811218, "learning_rate": 0.00015714285714285716, "loss": 1.6579, "step": 440 }, { "epoch": 0.07951398195300634, "grad_norm": 0.7932422161102295, "learning_rate": 0.00015892857142857142, "loss": 1.6593, "step": 445 }, { "epoch": 0.08040739748056822, "grad_norm": 0.795437216758728, "learning_rate": 0.00016071428571428573, "loss": 1.6661, "step": 450 }, { "epoch": 0.08130081300813008, "grad_norm": 0.8602275848388672, "learning_rate": 0.00016250000000000002, "loss": 1.6667, "step": 455 }, { "epoch": 0.08219422853569196, "grad_norm": 0.7548096776008606, "learning_rate": 0.00016428571428571428, "loss": 1.6403, "step": 460 }, { "epoch": 0.08308764406325382, "grad_norm": 0.8513688445091248, "learning_rate": 0.0001660714285714286, "loss": 1.6543, "step": 465 }, { "epoch": 0.08398105959081568, "grad_norm": 0.836286187171936, "learning_rate": 0.00016785714285714288, "loss": 1.6227, "step": 470 }, { "epoch": 0.08487447511837756, "grad_norm": 0.7904291749000549, "learning_rate": 0.00016964285714285714, "loss": 1.6433, "step": 475 }, { "epoch": 0.08576789064593943, "grad_norm": 0.7850791215896606, "learning_rate": 0.00017142857142857143, "loss": 1.6113, "step": 480 }, { "epoch": 0.08666130617350129, "grad_norm": 0.6912005543708801, "learning_rate": 0.00017321428571428574, "loss": 1.6047, "step": 485 }, { "epoch": 0.08755472170106317, "grad_norm": 0.8217949867248535, "learning_rate": 0.000175, "loss": 1.6206, "step": 490 }, { "epoch": 0.08844813722862503, "grad_norm": 0.7975768446922302, "learning_rate": 0.00017678571428571428, "loss": 1.6711, "step": 495 }, { "epoch": 0.08934155275618691, "grad_norm": 0.7934837937355042, "learning_rate": 0.0001785714285714286, "loss": 1.6195, "step": 500 }, { "epoch": 0.09023496828374877, "grad_norm": 0.8391872048377991, "learning_rate": 0.00018035714285714286, "loss": 1.6068, "step": 505 }, { "epoch": 0.09112838381131064, "grad_norm": 0.8512648344039917, "learning_rate": 0.00018214285714285714, "loss": 1.6489, "step": 510 }, { "epoch": 0.09202179933887251, "grad_norm": 0.7523573637008667, "learning_rate": 0.00018392857142857143, "loss": 1.6035, "step": 515 }, { "epoch": 0.09291521486643438, "grad_norm": 0.6890606880187988, "learning_rate": 0.00018571428571428572, "loss": 1.6162, "step": 520 }, { "epoch": 0.09380863039399624, "grad_norm": 0.8024640679359436, "learning_rate": 0.0001875, "loss": 1.5836, "step": 525 }, { "epoch": 0.09470204592155812, "grad_norm": 0.778583824634552, "learning_rate": 0.0001892857142857143, "loss": 1.6539, "step": 530 }, { "epoch": 0.09559546144911998, "grad_norm": 0.6993560194969177, "learning_rate": 0.00019107142857142858, "loss": 1.5736, "step": 535 }, { "epoch": 0.09648887697668186, "grad_norm": 0.6990346908569336, "learning_rate": 0.00019285714285714286, "loss": 1.5987, "step": 540 }, { "epoch": 0.09738229250424373, "grad_norm": 0.7906347513198853, "learning_rate": 0.00019464285714285715, "loss": 1.625, "step": 545 }, { "epoch": 0.09827570803180559, "grad_norm": 0.8062006831169128, "learning_rate": 0.00019642857142857144, "loss": 1.5445, "step": 550 }, { "epoch": 0.09916912355936747, "grad_norm": 0.7047358751296997, "learning_rate": 0.00019821428571428572, "loss": 1.5868, "step": 555 }, { "epoch": 0.10006253908692933, "grad_norm": 0.7451456785202026, "learning_rate": 0.0002, "loss": 1.632, "step": 560 }, { "epoch": 0.1009559546144912, "grad_norm": 0.7321707606315613, "learning_rate": 0.00019999951355027364, "loss": 1.5762, "step": 565 }, { "epoch": 0.10184937014205307, "grad_norm": 0.790035605430603, "learning_rate": 0.00019999805420582728, "loss": 1.5721, "step": 570 }, { "epoch": 0.10274278566961494, "grad_norm": 0.7690011858940125, "learning_rate": 0.00019999562198085878, "loss": 1.5736, "step": 575 }, { "epoch": 0.10363620119717681, "grad_norm": 0.7484734058380127, "learning_rate": 0.00019999221689903133, "loss": 1.5685, "step": 580 }, { "epoch": 0.10452961672473868, "grad_norm": 0.6638728976249695, "learning_rate": 0.00019998783899347294, "loss": 1.5683, "step": 585 }, { "epoch": 0.10542303225230054, "grad_norm": 0.6009637117385864, "learning_rate": 0.0001999824883067762, "loss": 1.6239, "step": 590 }, { "epoch": 0.10631644777986242, "grad_norm": 0.6219035387039185, "learning_rate": 0.00019997616489099792, "loss": 1.5447, "step": 595 }, { "epoch": 0.10720986330742428, "grad_norm": 0.6466681957244873, "learning_rate": 0.00019996886880765854, "loss": 1.5165, "step": 600 }, { "epoch": 0.10810327883498615, "grad_norm": 0.7301130890846252, "learning_rate": 0.0001999606001277417, "loss": 1.5419, "step": 605 }, { "epoch": 0.10899669436254802, "grad_norm": 0.635531485080719, "learning_rate": 0.0001999513589316933, "loss": 1.5127, "step": 610 }, { "epoch": 0.10989010989010989, "grad_norm": 0.7266018986701965, "learning_rate": 0.00019994114530942088, "loss": 1.5785, "step": 615 }, { "epoch": 0.11078352541767177, "grad_norm": 0.7803109288215637, "learning_rate": 0.0001999299593602927, "loss": 1.5557, "step": 620 }, { "epoch": 0.11167694094523363, "grad_norm": 0.7546414136886597, "learning_rate": 0.00019991780119313682, "loss": 1.5858, "step": 625 }, { "epoch": 0.1125703564727955, "grad_norm": 0.7428862452507019, "learning_rate": 0.00019990467092623998, "loss": 1.5246, "step": 630 }, { "epoch": 0.11346377200035737, "grad_norm": 0.6253552436828613, "learning_rate": 0.00019989056868734647, "loss": 1.5436, "step": 635 }, { "epoch": 0.11435718752791924, "grad_norm": 0.6777628064155579, "learning_rate": 0.0001998754946136569, "loss": 1.5812, "step": 640 }, { "epoch": 0.1152506030554811, "grad_norm": 0.7340919971466064, "learning_rate": 0.00019985944885182687, "loss": 1.5125, "step": 645 }, { "epoch": 0.11614401858304298, "grad_norm": 0.6683153510093689, "learning_rate": 0.00019984243155796546, "loss": 1.5239, "step": 650 }, { "epoch": 0.11703743411060484, "grad_norm": 0.6631706357002258, "learning_rate": 0.00019982444289763388, "loss": 1.4991, "step": 655 }, { "epoch": 0.11793084963816672, "grad_norm": 0.6204883456230164, "learning_rate": 0.00019980548304584364, "loss": 1.4839, "step": 660 }, { "epoch": 0.11882426516572858, "grad_norm": 0.6456660032272339, "learning_rate": 0.00019978555218705513, "loss": 1.5401, "step": 665 }, { "epoch": 0.11971768069329045, "grad_norm": 0.6324872970581055, "learning_rate": 0.00019976465051517548, "loss": 1.5242, "step": 670 }, { "epoch": 0.12061109622085232, "grad_norm": 0.6695961952209473, "learning_rate": 0.00019974277823355698, "loss": 1.5427, "step": 675 }, { "epoch": 0.12150451174841419, "grad_norm": 0.6281846165657043, "learning_rate": 0.00019971993555499494, "loss": 1.5168, "step": 680 }, { "epoch": 0.12239792727597605, "grad_norm": 0.6930037140846252, "learning_rate": 0.00019969612270172567, "loss": 1.5093, "step": 685 }, { "epoch": 0.12329134280353793, "grad_norm": 0.7301099300384521, "learning_rate": 0.00019967133990542423, "loss": 1.5472, "step": 690 }, { "epoch": 0.12418475833109979, "grad_norm": 0.6746507287025452, "learning_rate": 0.0001996455874072024, "loss": 1.5237, "step": 695 }, { "epoch": 0.12507817385866166, "grad_norm": 0.7017915844917297, "learning_rate": 0.00019961886545760598, "loss": 1.5288, "step": 700 }, { "epoch": 0.12597158938622352, "grad_norm": 0.6702309846878052, "learning_rate": 0.00019959117431661273, "loss": 1.5484, "step": 705 }, { "epoch": 0.1268650049137854, "grad_norm": 0.7421427369117737, "learning_rate": 0.00019956251425362967, "loss": 1.4334, "step": 710 }, { "epoch": 0.12775842044134728, "grad_norm": 0.6117368340492249, "learning_rate": 0.0001995328855474903, "loss": 1.5339, "step": 715 }, { "epoch": 0.12865183596890914, "grad_norm": 0.6836646199226379, "learning_rate": 0.00019950228848645218, "loss": 1.5339, "step": 720 }, { "epoch": 0.129545251496471, "grad_norm": 0.6464391350746155, "learning_rate": 0.00019947072336819397, "loss": 1.5194, "step": 725 }, { "epoch": 0.13043866702403287, "grad_norm": 0.7467032670974731, "learning_rate": 0.00019943819049981248, "loss": 1.4732, "step": 730 }, { "epoch": 0.13133208255159476, "grad_norm": 0.6169785857200623, "learning_rate": 0.00019940469019781985, "loss": 1.503, "step": 735 }, { "epoch": 0.13222549807915662, "grad_norm": 0.6561746001243591, "learning_rate": 0.00019937022278814032, "loss": 1.5135, "step": 740 }, { "epoch": 0.1331189136067185, "grad_norm": 0.6765868663787842, "learning_rate": 0.00019933478860610713, "loss": 1.498, "step": 745 }, { "epoch": 0.13401232913428035, "grad_norm": 0.5688114762306213, "learning_rate": 0.00019929838799645925, "loss": 1.424, "step": 750 }, { "epoch": 0.13490574466184221, "grad_norm": 0.6881040334701538, "learning_rate": 0.00019926102131333803, "loss": 1.4588, "step": 755 }, { "epoch": 0.1357991601894041, "grad_norm": 0.7207677960395813, "learning_rate": 0.00019922268892028368, "loss": 1.4658, "step": 760 }, { "epoch": 0.13669257571696597, "grad_norm": 0.645916759967804, "learning_rate": 0.0001991833911902319, "loss": 1.4797, "step": 765 }, { "epoch": 0.13758599124452783, "grad_norm": 0.6428040266036987, "learning_rate": 0.00019914312850551, "loss": 1.5143, "step": 770 }, { "epoch": 0.1384794067720897, "grad_norm": 0.6166855096817017, "learning_rate": 0.0001991019012578335, "loss": 1.4823, "step": 775 }, { "epoch": 0.13937282229965156, "grad_norm": 0.6110146641731262, "learning_rate": 0.00019905970984830204, "loss": 1.4651, "step": 780 }, { "epoch": 0.14026623782721342, "grad_norm": 0.6262656450271606, "learning_rate": 0.00019901655468739562, "loss": 1.4983, "step": 785 }, { "epoch": 0.14115965335477532, "grad_norm": 0.6442315578460693, "learning_rate": 0.00019897243619497056, "loss": 1.4735, "step": 790 }, { "epoch": 0.14205306888233718, "grad_norm": 0.5954890251159668, "learning_rate": 0.00019892735480025545, "loss": 1.4986, "step": 795 }, { "epoch": 0.14294648440989904, "grad_norm": 0.656102180480957, "learning_rate": 0.0001988813109418469, "loss": 1.4934, "step": 800 }, { "epoch": 0.1438398999374609, "grad_norm": 0.630328893661499, "learning_rate": 0.00019883430506770536, "loss": 1.4769, "step": 805 }, { "epoch": 0.14473331546502277, "grad_norm": 0.6247830390930176, "learning_rate": 0.00019878633763515074, "loss": 1.4446, "step": 810 }, { "epoch": 0.14562673099258466, "grad_norm": 0.6472021341323853, "learning_rate": 0.00019873740911085792, "loss": 1.4711, "step": 815 }, { "epoch": 0.14652014652014653, "grad_norm": 0.6813840270042419, "learning_rate": 0.00019868751997085225, "loss": 1.4533, "step": 820 }, { "epoch": 0.1474135620477084, "grad_norm": 0.6286069750785828, "learning_rate": 0.0001986366707005049, "loss": 1.4806, "step": 825 }, { "epoch": 0.14830697757527025, "grad_norm": 0.6147023439407349, "learning_rate": 0.00019858486179452812, "loss": 1.4779, "step": 830 }, { "epoch": 0.14920039310283212, "grad_norm": 0.6876225471496582, "learning_rate": 0.0001985320937569705, "loss": 1.5072, "step": 835 }, { "epoch": 0.150093808630394, "grad_norm": 0.6138557195663452, "learning_rate": 0.00019847836710121198, "loss": 1.4605, "step": 840 }, { "epoch": 0.15098722415795587, "grad_norm": 0.642737865447998, "learning_rate": 0.0001984236823499589, "loss": 1.4692, "step": 845 }, { "epoch": 0.15188063968551774, "grad_norm": 0.6422574520111084, "learning_rate": 0.0001983680400352389, "loss": 1.4684, "step": 850 }, { "epoch": 0.1527740552130796, "grad_norm": 0.6301836967468262, "learning_rate": 0.00019831144069839578, "loss": 1.4375, "step": 855 }, { "epoch": 0.15366747074064147, "grad_norm": 0.5667446255683899, "learning_rate": 0.00019825388489008415, "loss": 1.427, "step": 860 }, { "epoch": 0.15456088626820333, "grad_norm": 0.6058502197265625, "learning_rate": 0.0001981953731702642, "loss": 1.4666, "step": 865 }, { "epoch": 0.15545430179576522, "grad_norm": 0.6004565358161926, "learning_rate": 0.00019813590610819604, "loss": 1.4691, "step": 870 }, { "epoch": 0.15634771732332708, "grad_norm": 0.6401507258415222, "learning_rate": 0.00019807548428243447, "loss": 1.4149, "step": 875 }, { "epoch": 0.15724113285088895, "grad_norm": 0.6585138440132141, "learning_rate": 0.00019801410828082307, "loss": 1.478, "step": 880 }, { "epoch": 0.1581345483784508, "grad_norm": 0.6155353784561157, "learning_rate": 0.00019795177870048864, "loss": 1.446, "step": 885 }, { "epoch": 0.15902796390601268, "grad_norm": 0.6294763684272766, "learning_rate": 0.00019788849614783534, "loss": 1.4208, "step": 890 }, { "epoch": 0.15992137943357457, "grad_norm": 0.5941749811172485, "learning_rate": 0.00019782426123853873, "loss": 1.4909, "step": 895 }, { "epoch": 0.16081479496113643, "grad_norm": 0.6547942161560059, "learning_rate": 0.0001977590745975399, "loss": 1.3606, "step": 900 }, { "epoch": 0.1617082104886983, "grad_norm": 0.6711387634277344, "learning_rate": 0.00019769293685903937, "loss": 1.4263, "step": 905 }, { "epoch": 0.16260162601626016, "grad_norm": 0.6281640529632568, "learning_rate": 0.0001976258486664908, "loss": 1.4387, "step": 910 }, { "epoch": 0.16349504154382202, "grad_norm": 0.5495440363883972, "learning_rate": 0.00019755781067259487, "loss": 1.4227, "step": 915 }, { "epoch": 0.16438845707138391, "grad_norm": 0.6133942008018494, "learning_rate": 0.00019748882353929283, "loss": 1.3777, "step": 920 }, { "epoch": 0.16528187259894578, "grad_norm": 0.6217373013496399, "learning_rate": 0.00019741888793776012, "loss": 1.4337, "step": 925 }, { "epoch": 0.16617528812650764, "grad_norm": 0.5960021018981934, "learning_rate": 0.00019734800454839985, "loss": 1.4001, "step": 930 }, { "epoch": 0.1670687036540695, "grad_norm": 0.6111594438552856, "learning_rate": 0.00019727617406083608, "loss": 1.4485, "step": 935 }, { "epoch": 0.16796211918163137, "grad_norm": 0.6061040163040161, "learning_rate": 0.00019720339717390725, "loss": 1.4341, "step": 940 }, { "epoch": 0.16885553470919323, "grad_norm": 0.5827285051345825, "learning_rate": 0.00019712967459565935, "loss": 1.4212, "step": 945 }, { "epoch": 0.16974895023675513, "grad_norm": 0.625501275062561, "learning_rate": 0.00019705500704333888, "loss": 1.3902, "step": 950 }, { "epoch": 0.170642365764317, "grad_norm": 0.6008924245834351, "learning_rate": 0.00019697939524338605, "loss": 1.4401, "step": 955 }, { "epoch": 0.17153578129187885, "grad_norm": 0.5836713314056396, "learning_rate": 0.00019690283993142768, "loss": 1.4285, "step": 960 }, { "epoch": 0.17242919681944072, "grad_norm": 0.6134169101715088, "learning_rate": 0.00019682534185226996, "loss": 1.4527, "step": 965 }, { "epoch": 0.17332261234700258, "grad_norm": 0.593845784664154, "learning_rate": 0.0001967469017598913, "loss": 1.4293, "step": 970 }, { "epoch": 0.17421602787456447, "grad_norm": 0.6239144802093506, "learning_rate": 0.00019666752041743485, "loss": 1.394, "step": 975 }, { "epoch": 0.17510944340212634, "grad_norm": 0.5358985066413879, "learning_rate": 0.00019658719859720137, "loss": 1.4162, "step": 980 }, { "epoch": 0.1760028589296882, "grad_norm": 0.5680092573165894, "learning_rate": 0.00019650593708064133, "loss": 1.4292, "step": 985 }, { "epoch": 0.17689627445725006, "grad_norm": 0.6288170218467712, "learning_rate": 0.0001964237366583476, "loss": 1.4065, "step": 990 }, { "epoch": 0.17778968998481193, "grad_norm": 0.6214203238487244, "learning_rate": 0.00019634059813004767, "loss": 1.3958, "step": 995 }, { "epoch": 0.17868310551237382, "grad_norm": 0.5966234803199768, "learning_rate": 0.00019625652230459577, "loss": 1.4245, "step": 1000 }, { "epoch": 0.17957652103993568, "grad_norm": 0.5608589053153992, "learning_rate": 0.00019617150999996522, "loss": 1.3759, "step": 1005 }, { "epoch": 0.18046993656749755, "grad_norm": 0.6149086952209473, "learning_rate": 0.00019608556204324016, "loss": 1.3928, "step": 1010 }, { "epoch": 0.1813633520950594, "grad_norm": 0.5998888611793518, "learning_rate": 0.00019599867927060788, "loss": 1.4535, "step": 1015 }, { "epoch": 0.18225676762262127, "grad_norm": 0.6067023873329163, "learning_rate": 0.0001959108625273504, "loss": 1.4224, "step": 1020 }, { "epoch": 0.18315018315018314, "grad_norm": 0.6208594441413879, "learning_rate": 0.0001958221126678363, "loss": 1.3772, "step": 1025 }, { "epoch": 0.18404359867774503, "grad_norm": 0.555916428565979, "learning_rate": 0.00019573243055551247, "loss": 1.4087, "step": 1030 }, { "epoch": 0.1849370142053069, "grad_norm": 0.5708112716674805, "learning_rate": 0.0001956418170628957, "loss": 1.3879, "step": 1035 }, { "epoch": 0.18583042973286876, "grad_norm": 0.6426623463630676, "learning_rate": 0.0001955502730715642, "loss": 1.4281, "step": 1040 }, { "epoch": 0.18672384526043062, "grad_norm": 0.6203565001487732, "learning_rate": 0.0001954577994721489, "loss": 1.457, "step": 1045 }, { "epoch": 0.18761726078799248, "grad_norm": 0.5595524907112122, "learning_rate": 0.00019536439716432496, "loss": 1.3567, "step": 1050 }, { "epoch": 0.18851067631555438, "grad_norm": 0.5930672287940979, "learning_rate": 0.00019527006705680297, "loss": 1.3994, "step": 1055 }, { "epoch": 0.18940409184311624, "grad_norm": 0.5650060176849365, "learning_rate": 0.00019517481006731997, "loss": 1.4198, "step": 1060 }, { "epoch": 0.1902975073706781, "grad_norm": 0.558860719203949, "learning_rate": 0.0001950786271226307, "loss": 1.399, "step": 1065 }, { "epoch": 0.19119092289823997, "grad_norm": 0.6348270177841187, "learning_rate": 0.00019498151915849855, "loss": 1.4218, "step": 1070 }, { "epoch": 0.19208433842580183, "grad_norm": 0.63995760679245, "learning_rate": 0.00019488348711968633, "loss": 1.3706, "step": 1075 }, { "epoch": 0.19297775395336372, "grad_norm": 0.5952381491661072, "learning_rate": 0.00019478453195994719, "loss": 1.4318, "step": 1080 }, { "epoch": 0.1938711694809256, "grad_norm": 0.6112159490585327, "learning_rate": 0.0001946846546420154, "loss": 1.3999, "step": 1085 }, { "epoch": 0.19476458500848745, "grad_norm": 0.5743675231933594, "learning_rate": 0.0001945838561375968, "loss": 1.4095, "step": 1090 }, { "epoch": 0.19565800053604931, "grad_norm": 0.542060911655426, "learning_rate": 0.00019448213742735942, "loss": 1.3946, "step": 1095 }, { "epoch": 0.19655141606361118, "grad_norm": 0.5920471549034119, "learning_rate": 0.0001943794995009242, "loss": 1.3708, "step": 1100 }, { "epoch": 0.19744483159117304, "grad_norm": 0.5597534775733948, "learning_rate": 0.00019427594335685478, "loss": 1.35, "step": 1105 }, { "epoch": 0.19833824711873493, "grad_norm": 0.531980037689209, "learning_rate": 0.00019417147000264852, "loss": 1.3503, "step": 1110 }, { "epoch": 0.1992316626462968, "grad_norm": 0.5867334604263306, "learning_rate": 0.0001940660804547259, "loss": 1.3673, "step": 1115 }, { "epoch": 0.20012507817385866, "grad_norm": 0.5700910091400146, "learning_rate": 0.00019395977573842142, "loss": 1.3664, "step": 1120 }, { "epoch": 0.20101849370142053, "grad_norm": 0.5407332181930542, "learning_rate": 0.000193852556887973, "loss": 1.3677, "step": 1125 }, { "epoch": 0.2019119092289824, "grad_norm": 0.5347940921783447, "learning_rate": 0.00019374442494651223, "loss": 1.3819, "step": 1130 }, { "epoch": 0.20280532475654428, "grad_norm": 0.5672542452812195, "learning_rate": 0.00019363538096605427, "loss": 1.3838, "step": 1135 }, { "epoch": 0.20369874028410614, "grad_norm": 0.5784164667129517, "learning_rate": 0.00019352542600748734, "loss": 1.3848, "step": 1140 }, { "epoch": 0.204592155811668, "grad_norm": 0.5457838177680969, "learning_rate": 0.00019341456114056263, "loss": 1.3567, "step": 1145 }, { "epoch": 0.20548557133922987, "grad_norm": 0.6002548336982727, "learning_rate": 0.00019330278744388385, "loss": 1.357, "step": 1150 }, { "epoch": 0.20637898686679174, "grad_norm": 0.58780437707901, "learning_rate": 0.00019319010600489663, "loss": 1.3618, "step": 1155 }, { "epoch": 0.20727240239435363, "grad_norm": 0.5515660047531128, "learning_rate": 0.00019307651791987816, "loss": 1.344, "step": 1160 }, { "epoch": 0.2081658179219155, "grad_norm": 0.5853777527809143, "learning_rate": 0.00019296202429392622, "loss": 1.3699, "step": 1165 }, { "epoch": 0.20905923344947736, "grad_norm": 0.558180034160614, "learning_rate": 0.00019284662624094874, "loss": 1.325, "step": 1170 }, { "epoch": 0.20995264897703922, "grad_norm": 0.5694774389266968, "learning_rate": 0.00019273032488365267, "loss": 1.3608, "step": 1175 }, { "epoch": 0.21084606450460108, "grad_norm": 0.5699239373207092, "learning_rate": 0.00019261312135353332, "loss": 1.3384, "step": 1180 }, { "epoch": 0.21173948003216295, "grad_norm": 0.5678359270095825, "learning_rate": 0.0001924950167908632, "loss": 1.3271, "step": 1185 }, { "epoch": 0.21263289555972484, "grad_norm": 0.560370147228241, "learning_rate": 0.00019237601234468096, "loss": 1.3563, "step": 1190 }, { "epoch": 0.2135263110872867, "grad_norm": 0.5844460129737854, "learning_rate": 0.0001922561091727802, "loss": 1.3256, "step": 1195 }, { "epoch": 0.21441972661484857, "grad_norm": 0.5936894416809082, "learning_rate": 0.00019213530844169817, "loss": 1.3581, "step": 1200 }, { "epoch": 0.21531314214241043, "grad_norm": 0.6150529384613037, "learning_rate": 0.00019201361132670456, "loss": 1.3306, "step": 1205 }, { "epoch": 0.2162065576699723, "grad_norm": 0.5778645873069763, "learning_rate": 0.00019189101901178997, "loss": 1.381, "step": 1210 }, { "epoch": 0.21709997319753419, "grad_norm": 0.5521702170372009, "learning_rate": 0.00019176753268965432, "loss": 1.3191, "step": 1215 }, { "epoch": 0.21799338872509605, "grad_norm": 0.550817608833313, "learning_rate": 0.00019164315356169536, "loss": 1.3381, "step": 1220 }, { "epoch": 0.2188868042526579, "grad_norm": 0.6147415041923523, "learning_rate": 0.00019151788283799698, "loss": 1.3591, "step": 1225 }, { "epoch": 0.21978021978021978, "grad_norm": 0.5430126786231995, "learning_rate": 0.00019139172173731733, "loss": 1.344, "step": 1230 }, { "epoch": 0.22067363530778164, "grad_norm": 0.5844172835350037, "learning_rate": 0.0001912646714870771, "loss": 1.371, "step": 1235 }, { "epoch": 0.22156705083534353, "grad_norm": 0.5882902145385742, "learning_rate": 0.0001911367333233474, "loss": 1.3712, "step": 1240 }, { "epoch": 0.2224604663629054, "grad_norm": 0.5897734761238098, "learning_rate": 0.00019100790849083804, "loss": 1.3004, "step": 1245 }, { "epoch": 0.22335388189046726, "grad_norm": 0.554033637046814, "learning_rate": 0.00019087819824288504, "loss": 1.3803, "step": 1250 }, { "epoch": 0.22424729741802912, "grad_norm": 0.5678560733795166, "learning_rate": 0.0001907476038414387, "loss": 1.3817, "step": 1255 }, { "epoch": 0.225140712945591, "grad_norm": 0.5871914029121399, "learning_rate": 0.00019061612655705128, "loss": 1.3456, "step": 1260 }, { "epoch": 0.22603412847315285, "grad_norm": 0.587515115737915, "learning_rate": 0.00019048376766886448, "loss": 1.3484, "step": 1265 }, { "epoch": 0.22692754400071474, "grad_norm": 0.5953535437583923, "learning_rate": 0.00019035052846459727, "loss": 1.3351, "step": 1270 }, { "epoch": 0.2278209595282766, "grad_norm": 0.5097110271453857, "learning_rate": 0.00019021641024053308, "loss": 1.3122, "step": 1275 }, { "epoch": 0.22871437505583847, "grad_norm": 0.563342273235321, "learning_rate": 0.00019008141430150745, "loss": 1.3345, "step": 1280 }, { "epoch": 0.22960779058340033, "grad_norm": 0.5420668721199036, "learning_rate": 0.00018994554196089506, "loss": 1.3379, "step": 1285 }, { "epoch": 0.2305012061109622, "grad_norm": 0.5748648047447205, "learning_rate": 0.0001898087945405972, "loss": 1.2791, "step": 1290 }, { "epoch": 0.2313946216385241, "grad_norm": 0.5751796364784241, "learning_rate": 0.00018967117337102883, "loss": 1.3039, "step": 1295 }, { "epoch": 0.23228803716608595, "grad_norm": 0.5670933127403259, "learning_rate": 0.00018953267979110545, "loss": 1.358, "step": 1300 }, { "epoch": 0.23318145269364782, "grad_norm": 0.6038489937782288, "learning_rate": 0.0001893933151482304, "loss": 1.3747, "step": 1305 }, { "epoch": 0.23407486822120968, "grad_norm": 0.567458987236023, "learning_rate": 0.00018925308079828152, "loss": 1.3369, "step": 1310 }, { "epoch": 0.23496828374877154, "grad_norm": 0.5277537107467651, "learning_rate": 0.00018911197810559803, "loss": 1.3464, "step": 1315 }, { "epoch": 0.23586169927633344, "grad_norm": 0.5678889155387878, "learning_rate": 0.00018897000844296727, "loss": 1.3414, "step": 1320 }, { "epoch": 0.2367551148038953, "grad_norm": 0.548362135887146, "learning_rate": 0.00018882717319161128, "loss": 1.3431, "step": 1325 }, { "epoch": 0.23764853033145716, "grad_norm": 0.5988355278968811, "learning_rate": 0.00018868347374117344, "loss": 1.287, "step": 1330 }, { "epoch": 0.23854194585901903, "grad_norm": 0.6163780093193054, "learning_rate": 0.00018853891148970498, "loss": 1.3503, "step": 1335 }, { "epoch": 0.2394353613865809, "grad_norm": 0.5658857822418213, "learning_rate": 0.00018839348784365116, "loss": 1.3291, "step": 1340 }, { "epoch": 0.24032877691414276, "grad_norm": 0.5733643770217896, "learning_rate": 0.0001882472042178379, "loss": 1.3533, "step": 1345 }, { "epoch": 0.24122219244170465, "grad_norm": 0.5487879514694214, "learning_rate": 0.0001881000620354578, "loss": 1.3396, "step": 1350 }, { "epoch": 0.2421156079692665, "grad_norm": 0.5239369869232178, "learning_rate": 0.0001879520627280563, "loss": 1.3193, "step": 1355 }, { "epoch": 0.24300902349682837, "grad_norm": 0.5878532528877258, "learning_rate": 0.0001878032077355179, "loss": 1.3203, "step": 1360 }, { "epoch": 0.24390243902439024, "grad_norm": 0.5400727391242981, "learning_rate": 0.00018765349850605195, "loss": 1.3079, "step": 1365 }, { "epoch": 0.2447958545519521, "grad_norm": 0.5751279592514038, "learning_rate": 0.0001875029364961788, "loss": 1.2999, "step": 1370 }, { "epoch": 0.245689270079514, "grad_norm": 0.5734187960624695, "learning_rate": 0.00018735152317071534, "loss": 1.3283, "step": 1375 }, { "epoch": 0.24658268560707586, "grad_norm": 0.5625242590904236, "learning_rate": 0.00018719926000276106, "loss": 1.3355, "step": 1380 }, { "epoch": 0.24747610113463772, "grad_norm": 0.5596960186958313, "learning_rate": 0.0001870461484736834, "loss": 1.3156, "step": 1385 }, { "epoch": 0.24836951666219959, "grad_norm": 0.5672683119773865, "learning_rate": 0.00018689219007310369, "loss": 1.342, "step": 1390 }, { "epoch": 0.24926293218976145, "grad_norm": 0.5605707764625549, "learning_rate": 0.00018673738629888226, "loss": 1.3166, "step": 1395 }, { "epoch": 0.2501563477173233, "grad_norm": 0.5812191367149353, "learning_rate": 0.0001865817386571043, "loss": 1.3484, "step": 1400 }, { "epoch": 0.2510497632448852, "grad_norm": 0.5138891935348511, "learning_rate": 0.00018642524866206475, "loss": 1.3201, "step": 1405 }, { "epoch": 0.25194317877244704, "grad_norm": 0.5659753084182739, "learning_rate": 0.000186267917836254, "loss": 1.3117, "step": 1410 }, { "epoch": 0.25283659430000893, "grad_norm": 0.5799670815467834, "learning_rate": 0.00018610974771034275, "loss": 1.3161, "step": 1415 }, { "epoch": 0.2537300098275708, "grad_norm": 0.537071943283081, "learning_rate": 0.00018595073982316732, "loss": 1.3183, "step": 1420 }, { "epoch": 0.25462342535513266, "grad_norm": 0.5561192631721497, "learning_rate": 0.00018579089572171454, "loss": 1.321, "step": 1425 }, { "epoch": 0.25551684088269455, "grad_norm": 0.5795418620109558, "learning_rate": 0.00018563021696110682, "loss": 1.3301, "step": 1430 }, { "epoch": 0.2564102564102564, "grad_norm": 0.537104070186615, "learning_rate": 0.0001854687051045869, "loss": 1.2937, "step": 1435 }, { "epoch": 0.2573036719378183, "grad_norm": 0.6000511050224304, "learning_rate": 0.00018530636172350287, "loss": 1.2987, "step": 1440 }, { "epoch": 0.25819708746538017, "grad_norm": 0.5180054903030396, "learning_rate": 0.00018514318839729242, "loss": 1.2903, "step": 1445 }, { "epoch": 0.259090502992942, "grad_norm": 0.5556187629699707, "learning_rate": 0.00018497918671346808, "loss": 1.3159, "step": 1450 }, { "epoch": 0.2599839185205039, "grad_norm": 0.5588898062705994, "learning_rate": 0.0001848143582676013, "loss": 1.295, "step": 1455 }, { "epoch": 0.26087733404806573, "grad_norm": 0.5521529316902161, "learning_rate": 0.0001846487046633071, "loss": 1.2851, "step": 1460 }, { "epoch": 0.2617707495756276, "grad_norm": 0.5741902589797974, "learning_rate": 0.0001844822275122285, "loss": 1.3095, "step": 1465 }, { "epoch": 0.2626641651031895, "grad_norm": 0.5776512026786804, "learning_rate": 0.00018431492843402084, "loss": 1.3194, "step": 1470 }, { "epoch": 0.26355758063075135, "grad_norm": 0.5787376165390015, "learning_rate": 0.00018414680905633586, "loss": 1.3132, "step": 1475 }, { "epoch": 0.26445099615831325, "grad_norm": 0.551239013671875, "learning_rate": 0.00018397787101480612, "loss": 1.2952, "step": 1480 }, { "epoch": 0.2653444116858751, "grad_norm": 0.535438060760498, "learning_rate": 0.0001838081159530289, "loss": 1.277, "step": 1485 }, { "epoch": 0.266237827213437, "grad_norm": 0.5279135704040527, "learning_rate": 0.0001836375455225502, "loss": 1.3103, "step": 1490 }, { "epoch": 0.26713124274099886, "grad_norm": 0.54332035779953, "learning_rate": 0.00018346616138284892, "loss": 1.3535, "step": 1495 }, { "epoch": 0.2680246582685607, "grad_norm": 0.519645631313324, "learning_rate": 0.0001832939652013203, "loss": 1.3083, "step": 1500 }, { "epoch": 0.2689180737961226, "grad_norm": 0.5469849705696106, "learning_rate": 0.00018312095865326012, "loss": 1.2726, "step": 1505 }, { "epoch": 0.26981148932368443, "grad_norm": 0.5668709874153137, "learning_rate": 0.0001829471434218481, "loss": 1.2965, "step": 1510 }, { "epoch": 0.2707049048512463, "grad_norm": 0.5535402894020081, "learning_rate": 0.00018277252119813176, "loss": 1.2742, "step": 1515 }, { "epoch": 0.2715983203788082, "grad_norm": 0.520740807056427, "learning_rate": 0.00018259709368100962, "loss": 1.2838, "step": 1520 }, { "epoch": 0.27249173590637005, "grad_norm": 0.5902200937271118, "learning_rate": 0.00018242086257721515, "loss": 1.278, "step": 1525 }, { "epoch": 0.27338515143393194, "grad_norm": 0.5141775608062744, "learning_rate": 0.00018224382960129972, "loss": 1.2815, "step": 1530 }, { "epoch": 0.2742785669614938, "grad_norm": 0.5365716218948364, "learning_rate": 0.00018206599647561627, "loss": 1.3087, "step": 1535 }, { "epoch": 0.27517198248905567, "grad_norm": 0.5323083400726318, "learning_rate": 0.00018188736493030222, "loss": 1.2783, "step": 1540 }, { "epoch": 0.2760653980166175, "grad_norm": 0.5199279189109802, "learning_rate": 0.00018170793670326292, "loss": 1.2883, "step": 1545 }, { "epoch": 0.2769588135441794, "grad_norm": 0.5308305025100708, "learning_rate": 0.0001815277135401546, "loss": 1.2868, "step": 1550 }, { "epoch": 0.2778522290717413, "grad_norm": 0.5134423971176147, "learning_rate": 0.00018134669719436749, "loss": 1.2758, "step": 1555 }, { "epoch": 0.2787456445993031, "grad_norm": 0.5420312881469727, "learning_rate": 0.00018116488942700857, "loss": 1.2882, "step": 1560 }, { "epoch": 0.279639060126865, "grad_norm": 0.5505488514900208, "learning_rate": 0.00018098229200688462, "loss": 1.3117, "step": 1565 }, { "epoch": 0.28053247565442685, "grad_norm": 0.5319713950157166, "learning_rate": 0.00018079890671048493, "loss": 1.2958, "step": 1570 }, { "epoch": 0.28142589118198874, "grad_norm": 0.5935749411582947, "learning_rate": 0.0001806147353219641, "loss": 1.2753, "step": 1575 }, { "epoch": 0.28231930670955063, "grad_norm": 0.5329294204711914, "learning_rate": 0.00018042977963312451, "loss": 1.3191, "step": 1580 }, { "epoch": 0.28321272223711247, "grad_norm": 0.5126677751541138, "learning_rate": 0.00018024404144339906, "loss": 1.2924, "step": 1585 }, { "epoch": 0.28410613776467436, "grad_norm": 0.554897129535675, "learning_rate": 0.00018005752255983355, "loss": 1.2715, "step": 1590 }, { "epoch": 0.2849995532922362, "grad_norm": 0.5345707535743713, "learning_rate": 0.00017987022479706922, "loss": 1.2999, "step": 1595 }, { "epoch": 0.2858929688197981, "grad_norm": 0.5505250096321106, "learning_rate": 0.000179682149977325, "loss": 1.3081, "step": 1600 }, { "epoch": 0.28678638434736, "grad_norm": 0.5407662987709045, "learning_rate": 0.0001794932999303797, "loss": 1.2921, "step": 1605 }, { "epoch": 0.2876797998749218, "grad_norm": 0.5246217846870422, "learning_rate": 0.00017930367649355447, "loss": 1.3035, "step": 1610 }, { "epoch": 0.2885732154024837, "grad_norm": 0.5454705953598022, "learning_rate": 0.00017911328151169466, "loss": 1.302, "step": 1615 }, { "epoch": 0.28946663093004554, "grad_norm": 0.571116030216217, "learning_rate": 0.000178922116837152, "loss": 1.2515, "step": 1620 }, { "epoch": 0.29036004645760743, "grad_norm": 0.5103644728660583, "learning_rate": 0.00017873018432976658, "loss": 1.2696, "step": 1625 }, { "epoch": 0.2912534619851693, "grad_norm": 0.5452694296836853, "learning_rate": 0.00017853748585684867, "loss": 1.2893, "step": 1630 }, { "epoch": 0.29214687751273116, "grad_norm": 0.5386608242988586, "learning_rate": 0.0001783440232931607, "loss": 1.2519, "step": 1635 }, { "epoch": 0.29304029304029305, "grad_norm": 0.5036607384681702, "learning_rate": 0.0001781497985208989, "loss": 1.2865, "step": 1640 }, { "epoch": 0.2939337085678549, "grad_norm": 0.4949742555618286, "learning_rate": 0.00017795481342967501, "loss": 1.218, "step": 1645 }, { "epoch": 0.2948271240954168, "grad_norm": 0.5089282989501953, "learning_rate": 0.0001777590699164979, "loss": 1.2993, "step": 1650 }, { "epoch": 0.2957205396229787, "grad_norm": 0.5262947678565979, "learning_rate": 0.00017756256988575513, "loss": 1.2875, "step": 1655 }, { "epoch": 0.2966139551505405, "grad_norm": 0.598348081111908, "learning_rate": 0.00017736531524919445, "loss": 1.3016, "step": 1660 }, { "epoch": 0.2975073706781024, "grad_norm": 0.5289075374603271, "learning_rate": 0.00017716730792590512, "loss": 1.3164, "step": 1665 }, { "epoch": 0.29840078620566424, "grad_norm": 0.5724149942398071, "learning_rate": 0.00017696854984229933, "loss": 1.299, "step": 1670 }, { "epoch": 0.29929420173322613, "grad_norm": 0.5685451626777649, "learning_rate": 0.00017676904293209336, "loss": 1.3032, "step": 1675 }, { "epoch": 0.300187617260788, "grad_norm": 0.5210792422294617, "learning_rate": 0.00017656878913628883, "loss": 1.2844, "step": 1680 }, { "epoch": 0.30108103278834986, "grad_norm": 0.5431951880455017, "learning_rate": 0.0001763677904031539, "loss": 1.274, "step": 1685 }, { "epoch": 0.30197444831591175, "grad_norm": 0.551347553730011, "learning_rate": 0.00017616604868820406, "loss": 1.3147, "step": 1690 }, { "epoch": 0.3028678638434736, "grad_norm": 0.54395991563797, "learning_rate": 0.0001759635659541834, "loss": 1.2935, "step": 1695 }, { "epoch": 0.3037612793710355, "grad_norm": 0.519129753112793, "learning_rate": 0.0001757603441710453, "loss": 1.2166, "step": 1700 }, { "epoch": 0.3046546948985973, "grad_norm": 0.5352473258972168, "learning_rate": 0.0001755563853159334, "loss": 1.2378, "step": 1705 }, { "epoch": 0.3055481104261592, "grad_norm": 0.5558743476867676, "learning_rate": 0.00017535169137316227, "loss": 1.2911, "step": 1710 }, { "epoch": 0.3064415259537211, "grad_norm": 0.5650675892829895, "learning_rate": 0.0001751462643341982, "loss": 1.254, "step": 1715 }, { "epoch": 0.30733494148128293, "grad_norm": 0.5582188367843628, "learning_rate": 0.0001749401061976397, "loss": 1.246, "step": 1720 }, { "epoch": 0.3082283570088448, "grad_norm": 0.5283738970756531, "learning_rate": 0.0001747332189691981, "loss": 1.246, "step": 1725 }, { "epoch": 0.30912177253640666, "grad_norm": 0.5312855839729309, "learning_rate": 0.00017452560466167818, "loss": 1.2898, "step": 1730 }, { "epoch": 0.31001518806396855, "grad_norm": 0.5232421159744263, "learning_rate": 0.00017431726529495837, "loss": 1.282, "step": 1735 }, { "epoch": 0.31090860359153044, "grad_norm": 0.5114389061927795, "learning_rate": 0.00017410820289597126, "loss": 1.2814, "step": 1740 }, { "epoch": 0.3118020191190923, "grad_norm": 0.49061334133148193, "learning_rate": 0.00017389841949868378, "loss": 1.2584, "step": 1745 }, { "epoch": 0.31269543464665417, "grad_norm": 0.531428873538971, "learning_rate": 0.0001736879171440774, "loss": 1.2596, "step": 1750 }, { "epoch": 0.313588850174216, "grad_norm": 0.5237516164779663, "learning_rate": 0.00017347669788012846, "loss": 1.2213, "step": 1755 }, { "epoch": 0.3144822657017779, "grad_norm": 0.4916597306728363, "learning_rate": 0.00017326476376178796, "loss": 1.2573, "step": 1760 }, { "epoch": 0.3153756812293398, "grad_norm": 0.5719464421272278, "learning_rate": 0.00017305211685096178, "loss": 1.2306, "step": 1765 }, { "epoch": 0.3162690967569016, "grad_norm": 0.5756252408027649, "learning_rate": 0.00017283875921649057, "loss": 1.2842, "step": 1770 }, { "epoch": 0.3171625122844635, "grad_norm": 0.544870913028717, "learning_rate": 0.0001726246929341296, "loss": 1.264, "step": 1775 }, { "epoch": 0.31805592781202535, "grad_norm": 0.4986020028591156, "learning_rate": 0.00017240992008652855, "loss": 1.2919, "step": 1780 }, { "epoch": 0.31894934333958724, "grad_norm": 0.5383627414703369, "learning_rate": 0.00017219444276321127, "loss": 1.2679, "step": 1785 }, { "epoch": 0.31984275886714914, "grad_norm": 0.5600801706314087, "learning_rate": 0.00017197826306055547, "loss": 1.2589, "step": 1790 }, { "epoch": 0.32073617439471097, "grad_norm": 0.5386105179786682, "learning_rate": 0.0001717613830817723, "loss": 1.2812, "step": 1795 }, { "epoch": 0.32162958992227286, "grad_norm": 0.5082394480705261, "learning_rate": 0.00017154380493688583, "loss": 1.2717, "step": 1800 }, { "epoch": 0.3225230054498347, "grad_norm": 0.5480278134346008, "learning_rate": 0.00017132553074271272, "loss": 1.2521, "step": 1805 }, { "epoch": 0.3234164209773966, "grad_norm": 0.522881269454956, "learning_rate": 0.00017110656262284135, "loss": 1.233, "step": 1810 }, { "epoch": 0.3243098365049585, "grad_norm": 0.5211268663406372, "learning_rate": 0.0001708869027076114, "loss": 1.2448, "step": 1815 }, { "epoch": 0.3252032520325203, "grad_norm": 0.5418733954429626, "learning_rate": 0.00017066655313409295, "loss": 1.2354, "step": 1820 }, { "epoch": 0.3260966675600822, "grad_norm": 0.524122416973114, "learning_rate": 0.00017044551604606585, "loss": 1.2641, "step": 1825 }, { "epoch": 0.32699008308764405, "grad_norm": 0.5205156207084656, "learning_rate": 0.00017022379359399868, "loss": 1.2312, "step": 1830 }, { "epoch": 0.32788349861520594, "grad_norm": 0.5240566730499268, "learning_rate": 0.00017000138793502796, "loss": 1.2603, "step": 1835 }, { "epoch": 0.32877691414276783, "grad_norm": 0.5625032186508179, "learning_rate": 0.00016977830123293713, "loss": 1.2308, "step": 1840 }, { "epoch": 0.32967032967032966, "grad_norm": 0.5633267164230347, "learning_rate": 0.00016955453565813548, "loss": 1.287, "step": 1845 }, { "epoch": 0.33056374519789156, "grad_norm": 0.5072932243347168, "learning_rate": 0.0001693300933876371, "loss": 1.259, "step": 1850 }, { "epoch": 0.3314571607254534, "grad_norm": 0.532630205154419, "learning_rate": 0.00016910497660503957, "loss": 1.2352, "step": 1855 }, { "epoch": 0.3323505762530153, "grad_norm": 0.5094119310379028, "learning_rate": 0.00016887918750050292, "loss": 1.2672, "step": 1860 }, { "epoch": 0.3332439917805771, "grad_norm": 0.5230539441108704, "learning_rate": 0.00016865272827072797, "loss": 1.2506, "step": 1865 }, { "epoch": 0.334137407308139, "grad_norm": 0.4917963445186615, "learning_rate": 0.00016842560111893543, "loss": 1.2685, "step": 1870 }, { "epoch": 0.3350308228357009, "grad_norm": 0.5256093144416809, "learning_rate": 0.00016819780825484413, "loss": 1.2344, "step": 1875 }, { "epoch": 0.33592423836326274, "grad_norm": 0.5623132586479187, "learning_rate": 0.00016796935189464956, "loss": 1.246, "step": 1880 }, { "epoch": 0.33681765389082463, "grad_norm": 0.5536134839057922, "learning_rate": 0.00016774023426100238, "loss": 1.2428, "step": 1885 }, { "epoch": 0.33771106941838647, "grad_norm": 0.4987887442111969, "learning_rate": 0.0001675104575829868, "loss": 1.2414, "step": 1890 }, { "epoch": 0.33860448494594836, "grad_norm": 0.5539235472679138, "learning_rate": 0.0001672800240960989, "loss": 1.262, "step": 1895 }, { "epoch": 0.33949790047351025, "grad_norm": 0.5383912920951843, "learning_rate": 0.00016704893604222476, "loss": 1.2957, "step": 1900 }, { "epoch": 0.3403913160010721, "grad_norm": 0.5472669005393982, "learning_rate": 0.00016681719566961887, "loss": 1.2723, "step": 1905 }, { "epoch": 0.341284731528634, "grad_norm": 0.5573126673698425, "learning_rate": 0.00016658480523288195, "loss": 1.214, "step": 1910 }, { "epoch": 0.3421781470561958, "grad_norm": 0.5227344632148743, "learning_rate": 0.00016635176699293934, "loss": 1.2613, "step": 1915 }, { "epoch": 0.3430715625837577, "grad_norm": 0.5032819509506226, "learning_rate": 0.00016611808321701882, "loss": 1.2575, "step": 1920 }, { "epoch": 0.3439649781113196, "grad_norm": 0.5216634273529053, "learning_rate": 0.00016588375617862858, "loss": 1.2253, "step": 1925 }, { "epoch": 0.34485839363888143, "grad_norm": 0.5421084761619568, "learning_rate": 0.000165648788157535, "loss": 1.2899, "step": 1930 }, { "epoch": 0.3457518091664433, "grad_norm": 0.5299046039581299, "learning_rate": 0.00016541318143974075, "loss": 1.263, "step": 1935 }, { "epoch": 0.34664522469400516, "grad_norm": 0.5358394980430603, "learning_rate": 0.00016517693831746225, "loss": 1.2614, "step": 1940 }, { "epoch": 0.34753864022156705, "grad_norm": 0.570563018321991, "learning_rate": 0.00016494006108910757, "loss": 1.257, "step": 1945 }, { "epoch": 0.34843205574912894, "grad_norm": 0.4944029450416565, "learning_rate": 0.000164702552059254, "loss": 1.2376, "step": 1950 }, { "epoch": 0.3493254712766908, "grad_norm": 0.5532763004302979, "learning_rate": 0.00016446441353862556, "loss": 1.3048, "step": 1955 }, { "epoch": 0.35021888680425267, "grad_norm": 0.5413596034049988, "learning_rate": 0.0001642256478440706, "loss": 1.2578, "step": 1960 }, { "epoch": 0.3511123023318145, "grad_norm": 0.5195985436439514, "learning_rate": 0.00016398625729853924, "loss": 1.2549, "step": 1965 }, { "epoch": 0.3520057178593764, "grad_norm": 0.5096262097358704, "learning_rate": 0.00016374624423106087, "loss": 1.2538, "step": 1970 }, { "epoch": 0.3528991333869383, "grad_norm": 0.5297495126724243, "learning_rate": 0.00016350561097672122, "loss": 1.2156, "step": 1975 }, { "epoch": 0.3537925489145001, "grad_norm": 0.523977518081665, "learning_rate": 0.00016326435987663995, "loss": 1.2471, "step": 1980 }, { "epoch": 0.354685964442062, "grad_norm": 0.5540353059768677, "learning_rate": 0.0001630224932779477, "loss": 1.232, "step": 1985 }, { "epoch": 0.35557937996962385, "grad_norm": 0.5836973190307617, "learning_rate": 0.00016278001353376323, "loss": 1.2102, "step": 1990 }, { "epoch": 0.35647279549718575, "grad_norm": 0.52564537525177, "learning_rate": 0.0001625369230031707, "loss": 1.2415, "step": 1995 }, { "epoch": 0.35736621102474764, "grad_norm": 0.512956976890564, "learning_rate": 0.00016229322405119655, "loss": 1.2297, "step": 2000 }, { "epoch": 0.3582596265523095, "grad_norm": 0.5465174317359924, "learning_rate": 0.00016204891904878657, "loss": 1.2394, "step": 2005 }, { "epoch": 0.35915304207987137, "grad_norm": 0.502322793006897, "learning_rate": 0.0001618040103727827, "loss": 1.2317, "step": 2010 }, { "epoch": 0.3600464576074332, "grad_norm": 0.5260788202285767, "learning_rate": 0.00016155850040590016, "loss": 1.2446, "step": 2015 }, { "epoch": 0.3609398731349951, "grad_norm": 0.5212415456771851, "learning_rate": 0.0001613123915367041, "loss": 1.2422, "step": 2020 }, { "epoch": 0.36183328866255693, "grad_norm": 0.5056548714637756, "learning_rate": 0.00016106568615958632, "loss": 1.1996, "step": 2025 }, { "epoch": 0.3627267041901188, "grad_norm": 0.52322918176651, "learning_rate": 0.00016081838667474213, "loss": 1.2173, "step": 2030 }, { "epoch": 0.3636201197176807, "grad_norm": 0.5279656052589417, "learning_rate": 0.0001605704954881468, "loss": 1.2358, "step": 2035 }, { "epoch": 0.36451353524524255, "grad_norm": 0.5513635277748108, "learning_rate": 0.00016032201501153242, "loss": 1.2278, "step": 2040 }, { "epoch": 0.36540695077280444, "grad_norm": 0.5040432214736938, "learning_rate": 0.00016007294766236406, "loss": 1.2253, "step": 2045 }, { "epoch": 0.3663003663003663, "grad_norm": 0.4764319062232971, "learning_rate": 0.00015982329586381675, "loss": 1.2304, "step": 2050 }, { "epoch": 0.36719378182792817, "grad_norm": 0.4786640703678131, "learning_rate": 0.00015957306204475132, "loss": 1.2555, "step": 2055 }, { "epoch": 0.36808719735549006, "grad_norm": 0.5668711066246033, "learning_rate": 0.00015932224863969135, "loss": 1.2294, "step": 2060 }, { "epoch": 0.3689806128830519, "grad_norm": 0.5136781930923462, "learning_rate": 0.000159070858088799, "loss": 1.1752, "step": 2065 }, { "epoch": 0.3698740284106138, "grad_norm": 0.5189619660377502, "learning_rate": 0.0001588188928378516, "loss": 1.2351, "step": 2070 }, { "epoch": 0.3707674439381756, "grad_norm": 0.5765734910964966, "learning_rate": 0.00015856635533821774, "loss": 1.2, "step": 2075 }, { "epoch": 0.3716608594657375, "grad_norm": 0.5305128693580627, "learning_rate": 0.00015831324804683328, "loss": 1.241, "step": 2080 }, { "epoch": 0.3725542749932994, "grad_norm": 0.5297000408172607, "learning_rate": 0.0001580595734261777, "loss": 1.1959, "step": 2085 }, { "epoch": 0.37344769052086124, "grad_norm": 0.5325464010238647, "learning_rate": 0.00015780533394425006, "loss": 1.2221, "step": 2090 }, { "epoch": 0.37434110604842313, "grad_norm": 0.5105712413787842, "learning_rate": 0.00015755053207454483, "loss": 1.2422, "step": 2095 }, { "epoch": 0.37523452157598497, "grad_norm": 0.5255334973335266, "learning_rate": 0.00015729517029602802, "loss": 1.2326, "step": 2100 }, { "epoch": 0.37612793710354686, "grad_norm": 0.5344464778900146, "learning_rate": 0.00015703925109311295, "loss": 1.2188, "step": 2105 }, { "epoch": 0.37702135263110875, "grad_norm": 0.5103366374969482, "learning_rate": 0.00015678277695563617, "loss": 1.2354, "step": 2110 }, { "epoch": 0.3779147681586706, "grad_norm": 0.5386320948600769, "learning_rate": 0.00015652575037883318, "loss": 1.1967, "step": 2115 }, { "epoch": 0.3788081836862325, "grad_norm": 0.5370935201644897, "learning_rate": 0.0001562681738633141, "loss": 1.2594, "step": 2120 }, { "epoch": 0.3797015992137943, "grad_norm": 0.5193549990653992, "learning_rate": 0.00015601004991503946, "loss": 1.2144, "step": 2125 }, { "epoch": 0.3805950147413562, "grad_norm": 0.5521916151046753, "learning_rate": 0.00015575138104529577, "loss": 1.2009, "step": 2130 }, { "epoch": 0.3814884302689181, "grad_norm": 0.4977095425128937, "learning_rate": 0.00015549216977067099, "loss": 1.2222, "step": 2135 }, { "epoch": 0.38238184579647994, "grad_norm": 0.4930327832698822, "learning_rate": 0.0001552324186130302, "loss": 1.2569, "step": 2140 }, { "epoch": 0.3832752613240418, "grad_norm": 0.5264571905136108, "learning_rate": 0.00015497213009949104, "loss": 1.2016, "step": 2145 }, { "epoch": 0.38416867685160366, "grad_norm": 0.505418598651886, "learning_rate": 0.000154711306762399, "loss": 1.2115, "step": 2150 }, { "epoch": 0.38506209237916555, "grad_norm": 0.5414779186248779, "learning_rate": 0.00015444995113930287, "loss": 1.2452, "step": 2155 }, { "epoch": 0.38595550790672745, "grad_norm": 0.534766435623169, "learning_rate": 0.00015418806577293013, "loss": 1.2098, "step": 2160 }, { "epoch": 0.3868489234342893, "grad_norm": 0.5061149001121521, "learning_rate": 0.00015392565321116207, "loss": 1.2156, "step": 2165 }, { "epoch": 0.3877423389618512, "grad_norm": 0.5177295207977295, "learning_rate": 0.00015366271600700902, "loss": 1.2279, "step": 2170 }, { "epoch": 0.388635754489413, "grad_norm": 0.5544948577880859, "learning_rate": 0.00015339925671858563, "loss": 1.2138, "step": 2175 }, { "epoch": 0.3895291700169749, "grad_norm": 0.531894862651825, "learning_rate": 0.0001531352779090859, "loss": 1.1939, "step": 2180 }, { "epoch": 0.39042258554453674, "grad_norm": 0.5233749151229858, "learning_rate": 0.00015287078214675819, "loss": 1.2357, "step": 2185 }, { "epoch": 0.39131600107209863, "grad_norm": 0.5275238156318665, "learning_rate": 0.00015260577200488034, "loss": 1.2322, "step": 2190 }, { "epoch": 0.3922094165996605, "grad_norm": 0.5474215745925903, "learning_rate": 0.00015234025006173452, "loss": 1.2265, "step": 2195 }, { "epoch": 0.39310283212722236, "grad_norm": 0.5120342373847961, "learning_rate": 0.00015207421890058237, "loss": 1.216, "step": 2200 }, { "epoch": 0.39399624765478425, "grad_norm": 0.5480685830116272, "learning_rate": 0.0001518076811096395, "loss": 1.2065, "step": 2205 }, { "epoch": 0.3948896631823461, "grad_norm": 0.47449925541877747, "learning_rate": 0.00015154063928205067, "loss": 1.1511, "step": 2210 }, { "epoch": 0.395783078709908, "grad_norm": 0.49476173520088196, "learning_rate": 0.00015127309601586434, "loss": 1.1736, "step": 2215 }, { "epoch": 0.39667649423746987, "grad_norm": 0.513217568397522, "learning_rate": 0.0001510050539140075, "loss": 1.2213, "step": 2220 }, { "epoch": 0.3975699097650317, "grad_norm": 0.5108741521835327, "learning_rate": 0.00015073651558426026, "loss": 1.2367, "step": 2225 }, { "epoch": 0.3984633252925936, "grad_norm": 0.5116496682167053, "learning_rate": 0.00015046748363923057, "loss": 1.2457, "step": 2230 }, { "epoch": 0.39935674082015543, "grad_norm": 0.5216470956802368, "learning_rate": 0.00015019796069632878, "loss": 1.2356, "step": 2235 }, { "epoch": 0.4002501563477173, "grad_norm": 0.5476000308990479, "learning_rate": 0.00014992794937774211, "loss": 1.2221, "step": 2240 }, { "epoch": 0.4011435718752792, "grad_norm": 0.5119903683662415, "learning_rate": 0.00014965745231040919, "loss": 1.2492, "step": 2245 }, { "epoch": 0.40203698740284105, "grad_norm": 0.5329665541648865, "learning_rate": 0.00014938647212599452, "loss": 1.2128, "step": 2250 }, { "epoch": 0.40293040293040294, "grad_norm": 0.5195185542106628, "learning_rate": 0.00014911501146086281, "loss": 1.2172, "step": 2255 }, { "epoch": 0.4038238184579648, "grad_norm": 0.4988158345222473, "learning_rate": 0.00014884307295605343, "loss": 1.2229, "step": 2260 }, { "epoch": 0.40471723398552667, "grad_norm": 0.5173177719116211, "learning_rate": 0.00014857065925725452, "loss": 1.1982, "step": 2265 }, { "epoch": 0.40561064951308856, "grad_norm": 0.5261069536209106, "learning_rate": 0.0001482977730147776, "loss": 1.2049, "step": 2270 }, { "epoch": 0.4065040650406504, "grad_norm": 0.5446577072143555, "learning_rate": 0.00014802441688353127, "loss": 1.2377, "step": 2275 }, { "epoch": 0.4073974805682123, "grad_norm": 0.5223033428192139, "learning_rate": 0.00014775059352299598, "loss": 1.1983, "step": 2280 }, { "epoch": 0.4082908960957741, "grad_norm": 0.5107192397117615, "learning_rate": 0.00014747630559719762, "loss": 1.2022, "step": 2285 }, { "epoch": 0.409184311623336, "grad_norm": 0.5348856449127197, "learning_rate": 0.00014720155577468193, "loss": 1.1886, "step": 2290 }, { "epoch": 0.4100777271508979, "grad_norm": 0.5229494571685791, "learning_rate": 0.00014692634672848847, "loss": 1.2079, "step": 2295 }, { "epoch": 0.41097114267845974, "grad_norm": 0.49456527829170227, "learning_rate": 0.00014665068113612449, "loss": 1.2135, "step": 2300 }, { "epoch": 0.41186455820602164, "grad_norm": 0.5290020108222961, "learning_rate": 0.00014637456167953907, "loss": 1.2236, "step": 2305 }, { "epoch": 0.41275797373358347, "grad_norm": 0.4855465888977051, "learning_rate": 0.00014609799104509685, "loss": 1.2053, "step": 2310 }, { "epoch": 0.41365138926114536, "grad_norm": 0.5055922269821167, "learning_rate": 0.00014582097192355207, "loss": 1.1958, "step": 2315 }, { "epoch": 0.41454480478870726, "grad_norm": 0.5334281325340271, "learning_rate": 0.00014554350701002222, "loss": 1.1807, "step": 2320 }, { "epoch": 0.4154382203162691, "grad_norm": 0.5472596287727356, "learning_rate": 0.00014526559900396188, "loss": 1.1889, "step": 2325 }, { "epoch": 0.416331635843831, "grad_norm": 0.5031034350395203, "learning_rate": 0.00014498725060913662, "loss": 1.2334, "step": 2330 }, { "epoch": 0.4172250513713928, "grad_norm": 0.5575581789016724, "learning_rate": 0.00014470846453359636, "loss": 1.199, "step": 2335 }, { "epoch": 0.4181184668989547, "grad_norm": 0.5209872126579285, "learning_rate": 0.00014442924348964938, "loss": 1.2144, "step": 2340 }, { "epoch": 0.41901188242651655, "grad_norm": 0.5038667321205139, "learning_rate": 0.00014414959019383564, "loss": 1.2162, "step": 2345 }, { "epoch": 0.41990529795407844, "grad_norm": 0.5263863801956177, "learning_rate": 0.00014386950736690053, "loss": 1.2165, "step": 2350 }, { "epoch": 0.42079871348164033, "grad_norm": 0.5653106570243835, "learning_rate": 0.00014358899773376832, "loss": 1.1906, "step": 2355 }, { "epoch": 0.42169212900920217, "grad_norm": 0.49144411087036133, "learning_rate": 0.00014330806402351574, "loss": 1.1886, "step": 2360 }, { "epoch": 0.42258554453676406, "grad_norm": 0.5124874114990234, "learning_rate": 0.00014302670896934532, "loss": 1.2159, "step": 2365 }, { "epoch": 0.4234789600643259, "grad_norm": 0.5377556085586548, "learning_rate": 0.00014274493530855878, "loss": 1.2051, "step": 2370 }, { "epoch": 0.4243723755918878, "grad_norm": 0.5524542331695557, "learning_rate": 0.00014246274578253059, "loss": 1.2096, "step": 2375 }, { "epoch": 0.4252657911194497, "grad_norm": 0.5271286368370056, "learning_rate": 0.00014218014313668105, "loss": 1.222, "step": 2380 }, { "epoch": 0.4261592066470115, "grad_norm": 0.5286676287651062, "learning_rate": 0.00014189713012044977, "loss": 1.2163, "step": 2385 }, { "epoch": 0.4270526221745734, "grad_norm": 0.5305603742599487, "learning_rate": 0.00014161370948726894, "loss": 1.1911, "step": 2390 }, { "epoch": 0.42794603770213524, "grad_norm": 0.5002890825271606, "learning_rate": 0.00014132988399453617, "loss": 1.242, "step": 2395 }, { "epoch": 0.42883945322969713, "grad_norm": 0.5160467028617859, "learning_rate": 0.00014104565640358824, "loss": 1.1924, "step": 2400 }, { "epoch": 0.429732868757259, "grad_norm": 0.5422557592391968, "learning_rate": 0.0001407610294796738, "loss": 1.193, "step": 2405 }, { "epoch": 0.43062628428482086, "grad_norm": 0.5061042904853821, "learning_rate": 0.00014047600599192666, "loss": 1.1835, "step": 2410 }, { "epoch": 0.43151969981238275, "grad_norm": 0.507078230381012, "learning_rate": 0.0001401905887133387, "loss": 1.1788, "step": 2415 }, { "epoch": 0.4324131153399446, "grad_norm": 0.5291483402252197, "learning_rate": 0.00013990478042073313, "loss": 1.1941, "step": 2420 }, { "epoch": 0.4333065308675065, "grad_norm": 0.5093267560005188, "learning_rate": 0.00013961858389473727, "loss": 1.1938, "step": 2425 }, { "epoch": 0.43419994639506837, "grad_norm": 0.5296744108200073, "learning_rate": 0.0001393320019197555, "loss": 1.1978, "step": 2430 }, { "epoch": 0.4350933619226302, "grad_norm": 0.49533629417419434, "learning_rate": 0.00013904503728394234, "loss": 1.2015, "step": 2435 }, { "epoch": 0.4359867774501921, "grad_norm": 0.5509293079376221, "learning_rate": 0.00013875769277917513, "loss": 1.1963, "step": 2440 }, { "epoch": 0.43688019297775393, "grad_norm": 0.49859216809272766, "learning_rate": 0.000138469971201027, "loss": 1.1748, "step": 2445 }, { "epoch": 0.4377736085053158, "grad_norm": 0.5663615465164185, "learning_rate": 0.00013818187534873954, "loss": 1.2001, "step": 2450 }, { "epoch": 0.4386670240328777, "grad_norm": 0.5102608799934387, "learning_rate": 0.00013789340802519581, "loss": 1.2063, "step": 2455 }, { "epoch": 0.43956043956043955, "grad_norm": 0.5029070973396301, "learning_rate": 0.0001376045720368928, "loss": 1.2071, "step": 2460 }, { "epoch": 0.44045385508800144, "grad_norm": 0.5075359344482422, "learning_rate": 0.00013731537019391428, "loss": 1.152, "step": 2465 }, { "epoch": 0.4413472706155633, "grad_norm": 0.5264108777046204, "learning_rate": 0.00013702580530990335, "loss": 1.2019, "step": 2470 }, { "epoch": 0.4422406861431252, "grad_norm": 0.5055813789367676, "learning_rate": 0.00013673588020203517, "loss": 1.1786, "step": 2475 }, { "epoch": 0.44313410167068706, "grad_norm": 0.5127353668212891, "learning_rate": 0.0001364455976909896, "loss": 1.1649, "step": 2480 }, { "epoch": 0.4440275171982489, "grad_norm": 0.5117409825325012, "learning_rate": 0.00013615496060092355, "loss": 1.1762, "step": 2485 }, { "epoch": 0.4449209327258108, "grad_norm": 0.5126619935035706, "learning_rate": 0.00013586397175944368, "loss": 1.2026, "step": 2490 }, { "epoch": 0.4458143482533726, "grad_norm": 0.5334925055503845, "learning_rate": 0.0001355726339975788, "loss": 1.2209, "step": 2495 }, { "epoch": 0.4467077637809345, "grad_norm": 0.5059501528739929, "learning_rate": 0.00013528095014975252, "loss": 1.1809, "step": 2500 }, { "epoch": 0.44760117930849636, "grad_norm": 0.5418409705162048, "learning_rate": 0.0001349889230537553, "loss": 1.2321, "step": 2505 }, { "epoch": 0.44849459483605825, "grad_norm": 0.5434291362762451, "learning_rate": 0.00013469655555071715, "loss": 1.1511, "step": 2510 }, { "epoch": 0.44938801036362014, "grad_norm": 0.523363471031189, "learning_rate": 0.00013440385048507997, "loss": 1.2086, "step": 2515 }, { "epoch": 0.450281425891182, "grad_norm": 0.5296551585197449, "learning_rate": 0.0001341108107045697, "loss": 1.1834, "step": 2520 }, { "epoch": 0.45117484141874387, "grad_norm": 0.5411925911903381, "learning_rate": 0.00013381743906016878, "loss": 1.173, "step": 2525 }, { "epoch": 0.4520682569463057, "grad_norm": 0.524066150188446, "learning_rate": 0.00013352373840608834, "loss": 1.2259, "step": 2530 }, { "epoch": 0.4529616724738676, "grad_norm": 0.5258145928382874, "learning_rate": 0.00013322971159974043, "loss": 1.1724, "step": 2535 }, { "epoch": 0.4538550880014295, "grad_norm": 0.5472233295440674, "learning_rate": 0.0001329353615017102, "loss": 1.1523, "step": 2540 }, { "epoch": 0.4547485035289913, "grad_norm": 0.5172765254974365, "learning_rate": 0.00013264069097572816, "loss": 1.1904, "step": 2545 }, { "epoch": 0.4556419190565532, "grad_norm": 0.49758392572402954, "learning_rate": 0.00013234570288864228, "loss": 1.1737, "step": 2550 }, { "epoch": 0.45653533458411505, "grad_norm": 0.5258619785308838, "learning_rate": 0.00013205040011039004, "loss": 1.1746, "step": 2555 }, { "epoch": 0.45742875011167694, "grad_norm": 0.4878065586090088, "learning_rate": 0.0001317547855139705, "loss": 1.1987, "step": 2560 }, { "epoch": 0.45832216563923883, "grad_norm": 0.4854235053062439, "learning_rate": 0.00013145886197541651, "loss": 1.1605, "step": 2565 }, { "epoch": 0.45921558116680067, "grad_norm": 0.48126521706581116, "learning_rate": 0.0001311626323737665, "loss": 1.2038, "step": 2570 }, { "epoch": 0.46010899669436256, "grad_norm": 0.5124824643135071, "learning_rate": 0.00013086609959103672, "loss": 1.1791, "step": 2575 }, { "epoch": 0.4610024122219244, "grad_norm": 0.5470382571220398, "learning_rate": 0.00013056926651219293, "loss": 1.2034, "step": 2580 }, { "epoch": 0.4618958277494863, "grad_norm": 0.5247116684913635, "learning_rate": 0.00013027213602512258, "loss": 1.1979, "step": 2585 }, { "epoch": 0.4627892432770482, "grad_norm": 0.5126326680183411, "learning_rate": 0.00012997471102060647, "loss": 1.2037, "step": 2590 }, { "epoch": 0.46368265880461, "grad_norm": 0.5170397162437439, "learning_rate": 0.00012967699439229093, "loss": 1.186, "step": 2595 }, { "epoch": 0.4645760743321719, "grad_norm": 0.5169094800949097, "learning_rate": 0.00012937898903665935, "loss": 1.1941, "step": 2600 }, { "epoch": 0.46546948985973374, "grad_norm": 0.537554144859314, "learning_rate": 0.0001290806978530042, "loss": 1.1847, "step": 2605 }, { "epoch": 0.46636290538729563, "grad_norm": 0.49902471899986267, "learning_rate": 0.00012878212374339883, "loss": 1.1983, "step": 2610 }, { "epoch": 0.4672563209148575, "grad_norm": 0.5386156439781189, "learning_rate": 0.0001284832696126691, "loss": 1.1829, "step": 2615 }, { "epoch": 0.46814973644241936, "grad_norm": 0.5420889854431152, "learning_rate": 0.00012818413836836515, "loss": 1.1821, "step": 2620 }, { "epoch": 0.46904315196998125, "grad_norm": 0.5098642110824585, "learning_rate": 0.00012788473292073328, "loss": 1.1982, "step": 2625 }, { "epoch": 0.4699365674975431, "grad_norm": 0.4834945499897003, "learning_rate": 0.00012758505618268743, "loss": 1.1951, "step": 2630 }, { "epoch": 0.470829983025105, "grad_norm": 0.5107752680778503, "learning_rate": 0.000127285111069781, "loss": 1.1451, "step": 2635 }, { "epoch": 0.4717233985526669, "grad_norm": 0.5363967418670654, "learning_rate": 0.00012698490050017824, "loss": 1.1832, "step": 2640 }, { "epoch": 0.4726168140802287, "grad_norm": 0.469596266746521, "learning_rate": 0.0001266844273946262, "loss": 1.1397, "step": 2645 }, { "epoch": 0.4735102296077906, "grad_norm": 0.5100436210632324, "learning_rate": 0.0001263836946764261, "loss": 1.1823, "step": 2650 }, { "epoch": 0.47440364513535244, "grad_norm": 0.5279127359390259, "learning_rate": 0.0001260827052714049, "loss": 1.1407, "step": 2655 }, { "epoch": 0.47529706066291433, "grad_norm": 0.49943119287490845, "learning_rate": 0.00012578146210788686, "loss": 1.16, "step": 2660 }, { "epoch": 0.47619047619047616, "grad_norm": 0.5424295663833618, "learning_rate": 0.0001254799681166651, "loss": 1.1971, "step": 2665 }, { "epoch": 0.47708389171803806, "grad_norm": 0.5269612669944763, "learning_rate": 0.00012517822623097296, "loss": 1.1883, "step": 2670 }, { "epoch": 0.47797730724559995, "grad_norm": 0.4970255196094513, "learning_rate": 0.0001248762393864556, "loss": 1.169, "step": 2675 }, { "epoch": 0.4788707227731618, "grad_norm": 0.46272140741348267, "learning_rate": 0.0001245740105211414, "loss": 1.1619, "step": 2680 }, { "epoch": 0.4797641383007237, "grad_norm": 0.5113199949264526, "learning_rate": 0.00012427154257541333, "loss": 1.183, "step": 2685 }, { "epoch": 0.4806575538282855, "grad_norm": 0.5340253114700317, "learning_rate": 0.0001239688384919804, "loss": 1.1534, "step": 2690 }, { "epoch": 0.4815509693558474, "grad_norm": 0.5490040183067322, "learning_rate": 0.00012366590121584895, "loss": 1.1496, "step": 2695 }, { "epoch": 0.4824443848834093, "grad_norm": 0.5425564646720886, "learning_rate": 0.0001233627336942941, "loss": 1.2009, "step": 2700 }, { "epoch": 0.48333780041097113, "grad_norm": 0.4985598027706146, "learning_rate": 0.00012305933887683102, "loss": 1.1716, "step": 2705 }, { "epoch": 0.484231215938533, "grad_norm": 0.48820316791534424, "learning_rate": 0.00012275571971518616, "loss": 1.189, "step": 2710 }, { "epoch": 0.48512463146609486, "grad_norm": 0.5365601181983948, "learning_rate": 0.00012245187916326878, "loss": 1.1969, "step": 2715 }, { "epoch": 0.48601804699365675, "grad_norm": 0.5123448371887207, "learning_rate": 0.00012214782017714185, "loss": 1.1424, "step": 2720 }, { "epoch": 0.48691146252121864, "grad_norm": 0.5060442686080933, "learning_rate": 0.00012184354571499365, "loss": 1.1682, "step": 2725 }, { "epoch": 0.4878048780487805, "grad_norm": 0.5641975998878479, "learning_rate": 0.00012153905873710878, "loss": 1.1952, "step": 2730 }, { "epoch": 0.48869829357634237, "grad_norm": 0.4496929347515106, "learning_rate": 0.00012123436220583931, "loss": 1.1788, "step": 2735 }, { "epoch": 0.4895917091039042, "grad_norm": 0.5671601295471191, "learning_rate": 0.00012092945908557616, "loss": 1.1792, "step": 2740 }, { "epoch": 0.4904851246314661, "grad_norm": 0.5188336968421936, "learning_rate": 0.00012062435234272007, "loss": 1.1921, "step": 2745 }, { "epoch": 0.491378540159028, "grad_norm": 0.5338768362998962, "learning_rate": 0.00012031904494565296, "loss": 1.1879, "step": 2750 }, { "epoch": 0.4922719556865898, "grad_norm": 0.5573875308036804, "learning_rate": 0.00012001353986470878, "loss": 1.2, "step": 2755 }, { "epoch": 0.4931653712141517, "grad_norm": 0.523671567440033, "learning_rate": 0.00011970784007214477, "loss": 1.1887, "step": 2760 }, { "epoch": 0.49405878674171355, "grad_norm": 0.5382546186447144, "learning_rate": 0.00011940194854211258, "loss": 1.1676, "step": 2765 }, { "epoch": 0.49495220226927544, "grad_norm": 0.5233120918273926, "learning_rate": 0.00011909586825062917, "loss": 1.178, "step": 2770 }, { "epoch": 0.49584561779683733, "grad_norm": 0.5520097017288208, "learning_rate": 0.00011878960217554809, "loss": 1.1757, "step": 2775 }, { "epoch": 0.49673903332439917, "grad_norm": 0.5060712695121765, "learning_rate": 0.00011848315329653028, "loss": 1.169, "step": 2780 }, { "epoch": 0.49763244885196106, "grad_norm": 0.5251619815826416, "learning_rate": 0.0001181765245950152, "loss": 1.1813, "step": 2785 }, { "epoch": 0.4985258643795229, "grad_norm": 0.48611682653427124, "learning_rate": 0.00011786971905419179, "loss": 1.1437, "step": 2790 }, { "epoch": 0.4994192799070848, "grad_norm": 0.5000553727149963, "learning_rate": 0.00011756273965896953, "loss": 1.1668, "step": 2795 }, { "epoch": 0.5003126954346466, "grad_norm": 0.548265278339386, "learning_rate": 0.00011725558939594924, "loss": 1.1859, "step": 2800 }, { "epoch": 0.5012061109622086, "grad_norm": 0.5142680406570435, "learning_rate": 0.00011694827125339418, "loss": 1.176, "step": 2805 }, { "epoch": 0.5020995264897704, "grad_norm": 0.5385056138038635, "learning_rate": 0.00011664078822120084, "loss": 1.195, "step": 2810 }, { "epoch": 0.5029929420173322, "grad_norm": 0.4849853217601776, "learning_rate": 0.00011633314329086993, "loss": 1.1613, "step": 2815 }, { "epoch": 0.5038863575448941, "grad_norm": 0.5373867154121399, "learning_rate": 0.00011602533945547737, "loss": 1.1389, "step": 2820 }, { "epoch": 0.504779773072456, "grad_norm": 0.5037809014320374, "learning_rate": 0.00011571737970964496, "loss": 1.1717, "step": 2825 }, { "epoch": 0.5056731886000179, "grad_norm": 0.4938678741455078, "learning_rate": 0.00011540926704951136, "loss": 1.1625, "step": 2830 }, { "epoch": 0.5065666041275797, "grad_norm": 0.47168177366256714, "learning_rate": 0.000115101004472703, "loss": 1.1373, "step": 2835 }, { "epoch": 0.5074600196551416, "grad_norm": 0.4792795479297638, "learning_rate": 0.00011479259497830472, "loss": 1.1635, "step": 2840 }, { "epoch": 0.5083534351827035, "grad_norm": 0.5276928544044495, "learning_rate": 0.00011448404156683088, "loss": 1.1809, "step": 2845 }, { "epoch": 0.5092468507102653, "grad_norm": 0.4980674684047699, "learning_rate": 0.00011417534724019592, "loss": 1.1669, "step": 2850 }, { "epoch": 0.5101402662378273, "grad_norm": 0.5144380927085876, "learning_rate": 0.00011386651500168524, "loss": 1.1944, "step": 2855 }, { "epoch": 0.5110336817653891, "grad_norm": 0.5162400007247925, "learning_rate": 0.00011355754785592596, "loss": 1.1526, "step": 2860 }, { "epoch": 0.5119270972929509, "grad_norm": 0.5225709080696106, "learning_rate": 0.00011324844880885783, "loss": 1.1731, "step": 2865 }, { "epoch": 0.5128205128205128, "grad_norm": 0.5327043533325195, "learning_rate": 0.00011293922086770375, "loss": 1.1483, "step": 2870 }, { "epoch": 0.5137139283480747, "grad_norm": 0.60153728723526, "learning_rate": 0.00011262986704094065, "loss": 1.1987, "step": 2875 }, { "epoch": 0.5146073438756366, "grad_norm": 0.5078974962234497, "learning_rate": 0.00011232039033827025, "loss": 1.15, "step": 2880 }, { "epoch": 0.5155007594031984, "grad_norm": 0.5076168775558472, "learning_rate": 0.00011201079377058963, "loss": 1.14, "step": 2885 }, { "epoch": 0.5163941749307603, "grad_norm": 0.5268626809120178, "learning_rate": 0.00011170108034996218, "loss": 1.1686, "step": 2890 }, { "epoch": 0.5172875904583222, "grad_norm": 0.5654718279838562, "learning_rate": 0.00011139125308958804, "loss": 1.1656, "step": 2895 }, { "epoch": 0.518181005985884, "grad_norm": 0.505577027797699, "learning_rate": 0.00011108131500377494, "loss": 1.1724, "step": 2900 }, { "epoch": 0.5190744215134458, "grad_norm": 0.50611811876297, "learning_rate": 0.00011077126910790882, "loss": 1.1562, "step": 2905 }, { "epoch": 0.5199678370410078, "grad_norm": 0.5131402015686035, "learning_rate": 0.0001104611184184245, "loss": 1.1485, "step": 2910 }, { "epoch": 0.5208612525685696, "grad_norm": 0.498565137386322, "learning_rate": 0.00011015086595277633, "loss": 1.1614, "step": 2915 }, { "epoch": 0.5217546680961315, "grad_norm": 0.4939316511154175, "learning_rate": 0.00010984051472940885, "loss": 1.1401, "step": 2920 }, { "epoch": 0.5226480836236934, "grad_norm": 0.5283201336860657, "learning_rate": 0.00010953006776772747, "loss": 1.1557, "step": 2925 }, { "epoch": 0.5235414991512553, "grad_norm": 0.492447167634964, "learning_rate": 0.00010921952808806888, "loss": 1.1519, "step": 2930 }, { "epoch": 0.5244349146788171, "grad_norm": 0.5183736681938171, "learning_rate": 0.00010890889871167203, "loss": 1.1543, "step": 2935 }, { "epoch": 0.525328330206379, "grad_norm": 0.5172650218009949, "learning_rate": 0.00010859818266064835, "loss": 1.1731, "step": 2940 }, { "epoch": 0.5262217457339409, "grad_norm": 0.5232465267181396, "learning_rate": 0.00010828738295795262, "loss": 1.1628, "step": 2945 }, { "epoch": 0.5271151612615027, "grad_norm": 0.5208895206451416, "learning_rate": 0.00010797650262735346, "loss": 1.1682, "step": 2950 }, { "epoch": 0.5280085767890645, "grad_norm": 0.5442891716957092, "learning_rate": 0.00010766554469340386, "loss": 1.1796, "step": 2955 }, { "epoch": 0.5289019923166265, "grad_norm": 0.5013934373855591, "learning_rate": 0.00010735451218141191, "loss": 1.1758, "step": 2960 }, { "epoch": 0.5297954078441883, "grad_norm": 0.560184895992279, "learning_rate": 0.0001070434081174112, "loss": 1.1602, "step": 2965 }, { "epoch": 0.5306888233717502, "grad_norm": 0.4728075861930847, "learning_rate": 0.00010673223552813147, "loss": 1.1579, "step": 2970 }, { "epoch": 0.5315822388993121, "grad_norm": 0.547065794467926, "learning_rate": 0.00010642099744096914, "loss": 1.1606, "step": 2975 }, { "epoch": 0.532475654426874, "grad_norm": 0.5507860779762268, "learning_rate": 0.00010610969688395782, "loss": 1.1524, "step": 2980 }, { "epoch": 0.5333690699544358, "grad_norm": 0.5366077423095703, "learning_rate": 0.00010579833688573897, "loss": 1.1668, "step": 2985 }, { "epoch": 0.5342624854819977, "grad_norm": 0.49021753668785095, "learning_rate": 0.00010548692047553227, "loss": 1.1554, "step": 2990 }, { "epoch": 0.5351559010095596, "grad_norm": 0.5239299535751343, "learning_rate": 0.00010517545068310635, "loss": 1.1501, "step": 2995 }, { "epoch": 0.5360493165371214, "grad_norm": 0.4920014441013336, "learning_rate": 0.00010486393053874902, "loss": 1.1705, "step": 3000 }, { "epoch": 0.5369427320646832, "grad_norm": 0.5581287741661072, "learning_rate": 0.0001045523630732381, "loss": 1.1742, "step": 3005 }, { "epoch": 0.5378361475922452, "grad_norm": 0.5146217346191406, "learning_rate": 0.00010424075131781178, "loss": 1.185, "step": 3010 }, { "epoch": 0.538729563119807, "grad_norm": 0.5112408995628357, "learning_rate": 0.00010392909830413904, "loss": 1.162, "step": 3015 }, { "epoch": 0.5396229786473689, "grad_norm": 0.526857316493988, "learning_rate": 0.00010361740706429046, "loss": 1.164, "step": 3020 }, { "epoch": 0.5405163941749308, "grad_norm": 0.5244843363761902, "learning_rate": 0.00010330568063070832, "loss": 1.1951, "step": 3025 }, { "epoch": 0.5414098097024926, "grad_norm": 0.5005354881286621, "learning_rate": 0.00010299392203617744, "loss": 1.1552, "step": 3030 }, { "epoch": 0.5423032252300545, "grad_norm": 0.5191918015480042, "learning_rate": 0.00010268213431379543, "loss": 1.1868, "step": 3035 }, { "epoch": 0.5431966407576164, "grad_norm": 0.4955528974533081, "learning_rate": 0.00010237032049694335, "loss": 1.1584, "step": 3040 }, { "epoch": 0.5440900562851783, "grad_norm": 0.4983779489994049, "learning_rate": 0.00010205848361925618, "loss": 1.1532, "step": 3045 }, { "epoch": 0.5449834718127401, "grad_norm": 0.5516743659973145, "learning_rate": 0.0001017466267145931, "loss": 1.1702, "step": 3050 }, { "epoch": 0.5458768873403019, "grad_norm": 0.5391404032707214, "learning_rate": 0.0001014347528170083, "loss": 1.1573, "step": 3055 }, { "epoch": 0.5467703028678639, "grad_norm": 0.5334770083427429, "learning_rate": 0.00010112286496072117, "loss": 1.1847, "step": 3060 }, { "epoch": 0.5476637183954257, "grad_norm": 0.5390617251396179, "learning_rate": 0.00010081096618008699, "loss": 1.1428, "step": 3065 }, { "epoch": 0.5485571339229875, "grad_norm": 0.5461019277572632, "learning_rate": 0.00010049905950956728, "loss": 1.1484, "step": 3070 }, { "epoch": 0.5494505494505495, "grad_norm": 0.5262424945831299, "learning_rate": 0.00010018714798370035, "loss": 1.1679, "step": 3075 }, { "epoch": 0.5503439649781113, "grad_norm": 0.4764571189880371, "learning_rate": 9.98752346370717e-05, "loss": 1.1585, "step": 3080 }, { "epoch": 0.5512373805056732, "grad_norm": 0.49840331077575684, "learning_rate": 9.956332250428457e-05, "loss": 1.1572, "step": 3085 }, { "epoch": 0.552130796033235, "grad_norm": 0.48860788345336914, "learning_rate": 9.925141461993043e-05, "loss": 1.126, "step": 3090 }, { "epoch": 0.553024211560797, "grad_norm": 0.5242288708686829, "learning_rate": 9.893951401855932e-05, "loss": 1.1558, "step": 3095 }, { "epoch": 0.5539176270883588, "grad_norm": 0.49476343393325806, "learning_rate": 9.862762373465055e-05, "loss": 1.1653, "step": 3100 }, { "epoch": 0.5548110426159206, "grad_norm": 0.583516538143158, "learning_rate": 9.831574680258297e-05, "loss": 1.1412, "step": 3105 }, { "epoch": 0.5557044581434826, "grad_norm": 0.5565661191940308, "learning_rate": 9.800388625660553e-05, "loss": 1.1802, "step": 3110 }, { "epoch": 0.5565978736710444, "grad_norm": 0.5394544005393982, "learning_rate": 9.769204513080775e-05, "loss": 1.1222, "step": 3115 }, { "epoch": 0.5574912891986062, "grad_norm": 0.5191746354103088, "learning_rate": 9.738022645909026e-05, "loss": 1.2078, "step": 3120 }, { "epoch": 0.5583847047261682, "grad_norm": 0.5176223516464233, "learning_rate": 9.706843327513521e-05, "loss": 1.1232, "step": 3125 }, { "epoch": 0.55927812025373, "grad_norm": 0.512442409992218, "learning_rate": 9.675666861237677e-05, "loss": 1.1489, "step": 3130 }, { "epoch": 0.5601715357812919, "grad_norm": 0.5076617002487183, "learning_rate": 9.644493550397168e-05, "loss": 1.1471, "step": 3135 }, { "epoch": 0.5610649513088537, "grad_norm": 0.5024195909500122, "learning_rate": 9.61332369827696e-05, "loss": 1.1461, "step": 3140 }, { "epoch": 0.5619583668364156, "grad_norm": 0.5137572884559631, "learning_rate": 9.582157608128374e-05, "loss": 1.1283, "step": 3145 }, { "epoch": 0.5628517823639775, "grad_norm": 0.5018545985221863, "learning_rate": 9.550995583166133e-05, "loss": 1.1518, "step": 3150 }, { "epoch": 0.5637451978915393, "grad_norm": 0.4772183299064636, "learning_rate": 9.519837926565409e-05, "loss": 1.1379, "step": 3155 }, { "epoch": 0.5646386134191013, "grad_norm": 0.5322695970535278, "learning_rate": 9.488684941458867e-05, "loss": 1.1748, "step": 3160 }, { "epoch": 0.5655320289466631, "grad_norm": 0.5145288109779358, "learning_rate": 9.45753693093373e-05, "loss": 1.1556, "step": 3165 }, { "epoch": 0.5664254444742249, "grad_norm": 0.5218089818954468, "learning_rate": 9.426394198028823e-05, "loss": 1.1221, "step": 3170 }, { "epoch": 0.5673188600017869, "grad_norm": 0.49951624870300293, "learning_rate": 9.395257045731627e-05, "loss": 1.1255, "step": 3175 }, { "epoch": 0.5682122755293487, "grad_norm": 0.5349987745285034, "learning_rate": 9.364125776975318e-05, "loss": 1.1647, "step": 3180 }, { "epoch": 0.5691056910569106, "grad_norm": 0.5180804133415222, "learning_rate": 9.333000694635849e-05, "loss": 1.1653, "step": 3185 }, { "epoch": 0.5699991065844724, "grad_norm": 0.49906909465789795, "learning_rate": 9.30188210152897e-05, "loss": 1.1138, "step": 3190 }, { "epoch": 0.5708925221120343, "grad_norm": 0.5072324872016907, "learning_rate": 9.270770300407305e-05, "loss": 1.1542, "step": 3195 }, { "epoch": 0.5717859376395962, "grad_norm": 0.52315354347229, "learning_rate": 9.239665593957398e-05, "loss": 1.1425, "step": 3200 }, { "epoch": 0.572679353167158, "grad_norm": 0.5507954359054565, "learning_rate": 9.208568284796766e-05, "loss": 1.1462, "step": 3205 }, { "epoch": 0.57357276869472, "grad_norm": 0.5008817911148071, "learning_rate": 9.177478675470956e-05, "loss": 1.1307, "step": 3210 }, { "epoch": 0.5744661842222818, "grad_norm": 0.47194069623947144, "learning_rate": 9.146397068450612e-05, "loss": 1.13, "step": 3215 }, { "epoch": 0.5753595997498436, "grad_norm": 0.5233185887336731, "learning_rate": 9.11532376612852e-05, "loss": 1.1596, "step": 3220 }, { "epoch": 0.5762530152774055, "grad_norm": 0.5047646164894104, "learning_rate": 9.084259070816662e-05, "loss": 1.1768, "step": 3225 }, { "epoch": 0.5771464308049674, "grad_norm": 0.5104073286056519, "learning_rate": 9.053203284743294e-05, "loss": 1.1276, "step": 3230 }, { "epoch": 0.5780398463325293, "grad_norm": 0.48400700092315674, "learning_rate": 9.022156710049985e-05, "loss": 1.1409, "step": 3235 }, { "epoch": 0.5789332618600911, "grad_norm": 0.5238465666770935, "learning_rate": 8.991119648788696e-05, "loss": 1.1316, "step": 3240 }, { "epoch": 0.579826677387653, "grad_norm": 0.49904054403305054, "learning_rate": 8.960092402918819e-05, "loss": 1.1225, "step": 3245 }, { "epoch": 0.5807200929152149, "grad_norm": 0.5016510486602783, "learning_rate": 8.929075274304267e-05, "loss": 1.1262, "step": 3250 }, { "epoch": 0.5816135084427767, "grad_norm": 0.5361919403076172, "learning_rate": 8.898068564710508e-05, "loss": 1.1453, "step": 3255 }, { "epoch": 0.5825069239703387, "grad_norm": 0.5483645796775818, "learning_rate": 8.86707257580165e-05, "loss": 1.1445, "step": 3260 }, { "epoch": 0.5834003394979005, "grad_norm": 0.49957215785980225, "learning_rate": 8.836087609137502e-05, "loss": 1.1917, "step": 3265 }, { "epoch": 0.5842937550254623, "grad_norm": 0.5269734859466553, "learning_rate": 8.805113966170635e-05, "loss": 1.1277, "step": 3270 }, { "epoch": 0.5851871705530242, "grad_norm": 0.5271036028862, "learning_rate": 8.774151948243453e-05, "loss": 1.1263, "step": 3275 }, { "epoch": 0.5860805860805861, "grad_norm": 0.5119196772575378, "learning_rate": 8.743201856585256e-05, "loss": 1.1596, "step": 3280 }, { "epoch": 0.5869740016081479, "grad_norm": 0.5136052370071411, "learning_rate": 8.712263992309318e-05, "loss": 1.1191, "step": 3285 }, { "epoch": 0.5878674171357098, "grad_norm": 0.500639796257019, "learning_rate": 8.681338656409951e-05, "loss": 1.1134, "step": 3290 }, { "epoch": 0.5887608326632717, "grad_norm": 0.5073699951171875, "learning_rate": 8.650426149759578e-05, "loss": 1.1346, "step": 3295 }, { "epoch": 0.5896542481908336, "grad_norm": 0.5222281813621521, "learning_rate": 8.619526773105813e-05, "loss": 1.1575, "step": 3300 }, { "epoch": 0.5905476637183954, "grad_norm": 0.5326797366142273, "learning_rate": 8.588640827068512e-05, "loss": 1.1509, "step": 3305 }, { "epoch": 0.5914410792459573, "grad_norm": 0.49078214168548584, "learning_rate": 8.557768612136882e-05, "loss": 1.0998, "step": 3310 }, { "epoch": 0.5923344947735192, "grad_norm": 0.5205763578414917, "learning_rate": 8.52691042866653e-05, "loss": 1.1247, "step": 3315 }, { "epoch": 0.593227910301081, "grad_norm": 0.49795159697532654, "learning_rate": 8.496066576876556e-05, "loss": 1.1234, "step": 3320 }, { "epoch": 0.5941213258286429, "grad_norm": 0.5376335382461548, "learning_rate": 8.465237356846624e-05, "loss": 1.1647, "step": 3325 }, { "epoch": 0.5950147413562048, "grad_norm": 0.5440948009490967, "learning_rate": 8.434423068514048e-05, "loss": 1.1377, "step": 3330 }, { "epoch": 0.5959081568837666, "grad_norm": 0.5351749062538147, "learning_rate": 8.403624011670871e-05, "loss": 1.1429, "step": 3335 }, { "epoch": 0.5968015724113285, "grad_norm": 0.5545377731323242, "learning_rate": 8.372840485960947e-05, "loss": 1.1869, "step": 3340 }, { "epoch": 0.5976949879388904, "grad_norm": 0.5247533917427063, "learning_rate": 8.342072790877033e-05, "loss": 1.138, "step": 3345 }, { "epoch": 0.5985884034664523, "grad_norm": 0.5115054845809937, "learning_rate": 8.311321225757858e-05, "loss": 1.1147, "step": 3350 }, { "epoch": 0.5994818189940141, "grad_norm": 0.49062785506248474, "learning_rate": 8.280586089785236e-05, "loss": 1.1498, "step": 3355 }, { "epoch": 0.600375234521576, "grad_norm": 0.5327452421188354, "learning_rate": 8.249867681981139e-05, "loss": 1.113, "step": 3360 }, { "epoch": 0.6012686500491379, "grad_norm": 0.5428333878517151, "learning_rate": 8.219166301204781e-05, "loss": 1.1632, "step": 3365 }, { "epoch": 0.6021620655766997, "grad_norm": 0.5033100247383118, "learning_rate": 8.188482246149736e-05, "loss": 1.1484, "step": 3370 }, { "epoch": 0.6030554811042615, "grad_norm": 0.5400857329368591, "learning_rate": 8.157815815341002e-05, "loss": 1.1556, "step": 3375 }, { "epoch": 0.6039488966318235, "grad_norm": 0.5341722369194031, "learning_rate": 8.127167307132119e-05, "loss": 1.1387, "step": 3380 }, { "epoch": 0.6048423121593853, "grad_norm": 0.5450566411018372, "learning_rate": 8.096537019702255e-05, "loss": 1.1598, "step": 3385 }, { "epoch": 0.6057357276869472, "grad_norm": 0.496113121509552, "learning_rate": 8.065925251053307e-05, "loss": 1.1664, "step": 3390 }, { "epoch": 0.6066291432145091, "grad_norm": 0.5090356469154358, "learning_rate": 8.035332299007014e-05, "loss": 1.1338, "step": 3395 }, { "epoch": 0.607522558742071, "grad_norm": 0.48691728711128235, "learning_rate": 8.004758461202023e-05, "loss": 1.1481, "step": 3400 }, { "epoch": 0.6084159742696328, "grad_norm": 0.4871600270271301, "learning_rate": 7.974204035091046e-05, "loss": 1.1227, "step": 3405 }, { "epoch": 0.6093093897971946, "grad_norm": 0.5205426216125488, "learning_rate": 7.943669317937923e-05, "loss": 1.1338, "step": 3410 }, { "epoch": 0.6102028053247566, "grad_norm": 0.5116109848022461, "learning_rate": 7.913154606814753e-05, "loss": 1.1329, "step": 3415 }, { "epoch": 0.6110962208523184, "grad_norm": 0.5520155429840088, "learning_rate": 7.882660198598993e-05, "loss": 1.1526, "step": 3420 }, { "epoch": 0.6119896363798802, "grad_norm": 0.5377830266952515, "learning_rate": 7.852186389970571e-05, "loss": 1.1676, "step": 3425 }, { "epoch": 0.6128830519074422, "grad_norm": 0.5186334252357483, "learning_rate": 7.821733477409005e-05, "loss": 1.1049, "step": 3430 }, { "epoch": 0.613776467435004, "grad_norm": 0.5580330491065979, "learning_rate": 7.791301757190516e-05, "loss": 1.1103, "step": 3435 }, { "epoch": 0.6146698829625659, "grad_norm": 0.5414636135101318, "learning_rate": 7.760891525385146e-05, "loss": 1.1243, "step": 3440 }, { "epoch": 0.6155632984901278, "grad_norm": 0.494886577129364, "learning_rate": 7.730503077853873e-05, "loss": 1.1476, "step": 3445 }, { "epoch": 0.6164567140176896, "grad_norm": 0.4783135950565338, "learning_rate": 7.700136710245731e-05, "loss": 1.0874, "step": 3450 }, { "epoch": 0.6173501295452515, "grad_norm": 0.5152401924133301, "learning_rate": 7.669792717994946e-05, "loss": 1.1166, "step": 3455 }, { "epoch": 0.6182435450728133, "grad_norm": 0.5359469056129456, "learning_rate": 7.639471396318057e-05, "loss": 1.1519, "step": 3460 }, { "epoch": 0.6191369606003753, "grad_norm": 0.5105694532394409, "learning_rate": 7.609173040211035e-05, "loss": 1.1516, "step": 3465 }, { "epoch": 0.6200303761279371, "grad_norm": 0.5673738718032837, "learning_rate": 7.578897944446417e-05, "loss": 1.168, "step": 3470 }, { "epoch": 0.6209237916554989, "grad_norm": 0.5316708087921143, "learning_rate": 7.548646403570449e-05, "loss": 1.1237, "step": 3475 }, { "epoch": 0.6218172071830609, "grad_norm": 0.5048865079879761, "learning_rate": 7.518418711900206e-05, "loss": 1.1546, "step": 3480 }, { "epoch": 0.6227106227106227, "grad_norm": 0.5034053921699524, "learning_rate": 7.488215163520729e-05, "loss": 1.114, "step": 3485 }, { "epoch": 0.6236040382381846, "grad_norm": 0.5447829365730286, "learning_rate": 7.458036052282185e-05, "loss": 1.0924, "step": 3490 }, { "epoch": 0.6244974537657465, "grad_norm": 0.5631914138793945, "learning_rate": 7.427881671796973e-05, "loss": 1.1281, "step": 3495 }, { "epoch": 0.6253908692933083, "grad_norm": 0.5400162935256958, "learning_rate": 7.3977523154369e-05, "loss": 1.1529, "step": 3500 }, { "epoch": 0.6262842848208702, "grad_norm": 0.5325872302055359, "learning_rate": 7.367648276330305e-05, "loss": 1.1596, "step": 3505 }, { "epoch": 0.627177700348432, "grad_norm": 0.5262747406959534, "learning_rate": 7.337569847359226e-05, "loss": 1.1311, "step": 3510 }, { "epoch": 0.628071115875994, "grad_norm": 0.530774712562561, "learning_rate": 7.307517321156528e-05, "loss": 1.1147, "step": 3515 }, { "epoch": 0.6289645314035558, "grad_norm": 0.5109870433807373, "learning_rate": 7.277490990103079e-05, "loss": 1.132, "step": 3520 }, { "epoch": 0.6298579469311176, "grad_norm": 0.5231294631958008, "learning_rate": 7.247491146324887e-05, "loss": 1.1106, "step": 3525 }, { "epoch": 0.6307513624586796, "grad_norm": 0.5209240913391113, "learning_rate": 7.217518081690265e-05, "loss": 1.1614, "step": 3530 }, { "epoch": 0.6316447779862414, "grad_norm": 0.5387827157974243, "learning_rate": 7.187572087807e-05, "loss": 1.1656, "step": 3535 }, { "epoch": 0.6325381935138032, "grad_norm": 0.5082910060882568, "learning_rate": 7.157653456019504e-05, "loss": 1.1274, "step": 3540 }, { "epoch": 0.6334316090413651, "grad_norm": 0.5118305683135986, "learning_rate": 7.127762477405976e-05, "loss": 1.1485, "step": 3545 }, { "epoch": 0.634325024568927, "grad_norm": 0.5459609031677246, "learning_rate": 7.097899442775584e-05, "loss": 1.1253, "step": 3550 }, { "epoch": 0.6352184400964889, "grad_norm": 0.5178401470184326, "learning_rate": 7.068064642665631e-05, "loss": 1.1465, "step": 3555 }, { "epoch": 0.6361118556240507, "grad_norm": 0.5017228722572327, "learning_rate": 7.038258367338723e-05, "loss": 1.1346, "step": 3560 }, { "epoch": 0.6370052711516127, "grad_norm": 0.5513466000556946, "learning_rate": 7.008480906779948e-05, "loss": 1.1423, "step": 3565 }, { "epoch": 0.6378986866791745, "grad_norm": 0.49276652932167053, "learning_rate": 6.97873255069406e-05, "loss": 1.1522, "step": 3570 }, { "epoch": 0.6387921022067363, "grad_norm": 0.5311774015426636, "learning_rate": 6.949013588502651e-05, "loss": 1.1291, "step": 3575 }, { "epoch": 0.6396855177342983, "grad_norm": 0.4871313273906708, "learning_rate": 6.919324309341341e-05, "loss": 1.1199, "step": 3580 }, { "epoch": 0.6405789332618601, "grad_norm": 0.5568504333496094, "learning_rate": 6.889665002056966e-05, "loss": 1.1184, "step": 3585 }, { "epoch": 0.6414723487894219, "grad_norm": 0.4711260199546814, "learning_rate": 6.860035955204767e-05, "loss": 1.1204, "step": 3590 }, { "epoch": 0.6423657643169838, "grad_norm": 0.5173743963241577, "learning_rate": 6.830437457045568e-05, "loss": 1.1491, "step": 3595 }, { "epoch": 0.6432591798445457, "grad_norm": 0.5250765085220337, "learning_rate": 6.800869795543007e-05, "loss": 1.1086, "step": 3600 }, { "epoch": 0.6441525953721076, "grad_norm": 0.5202131867408752, "learning_rate": 6.771333258360694e-05, "loss": 1.1355, "step": 3605 }, { "epoch": 0.6450460108996694, "grad_norm": 0.5153080821037292, "learning_rate": 6.74182813285944e-05, "loss": 1.1181, "step": 3610 }, { "epoch": 0.6459394264272313, "grad_norm": 0.5040249228477478, "learning_rate": 6.712354706094452e-05, "loss": 1.1061, "step": 3615 }, { "epoch": 0.6468328419547932, "grad_norm": 0.511741042137146, "learning_rate": 6.682913264812533e-05, "loss": 1.14, "step": 3620 }, { "epoch": 0.647726257482355, "grad_norm": 0.5170619487762451, "learning_rate": 6.653504095449305e-05, "loss": 1.1019, "step": 3625 }, { "epoch": 0.648619673009917, "grad_norm": 0.5282120704650879, "learning_rate": 6.624127484126421e-05, "loss": 1.1406, "step": 3630 }, { "epoch": 0.6495130885374788, "grad_norm": 0.5090754628181458, "learning_rate": 6.594783716648769e-05, "loss": 1.1396, "step": 3635 }, { "epoch": 0.6504065040650406, "grad_norm": 0.48920685052871704, "learning_rate": 6.565473078501698e-05, "loss": 1.1186, "step": 3640 }, { "epoch": 0.6512999195926025, "grad_norm": 0.5252284407615662, "learning_rate": 6.536195854848248e-05, "loss": 1.0744, "step": 3645 }, { "epoch": 0.6521933351201644, "grad_norm": 0.5125118494033813, "learning_rate": 6.506952330526373e-05, "loss": 1.1227, "step": 3650 }, { "epoch": 0.6530867506477263, "grad_norm": 0.5054253339767456, "learning_rate": 6.477742790046152e-05, "loss": 1.1191, "step": 3655 }, { "epoch": 0.6539801661752881, "grad_norm": 0.5262144804000854, "learning_rate": 6.448567517587052e-05, "loss": 1.1542, "step": 3660 }, { "epoch": 0.65487358170285, "grad_norm": 0.5490496158599854, "learning_rate": 6.419426796995137e-05, "loss": 1.1468, "step": 3665 }, { "epoch": 0.6557669972304119, "grad_norm": 0.5373784303665161, "learning_rate": 6.390320911780317e-05, "loss": 1.1393, "step": 3670 }, { "epoch": 0.6566604127579737, "grad_norm": 0.5126230716705322, "learning_rate": 6.36125014511359e-05, "loss": 1.1385, "step": 3675 }, { "epoch": 0.6575538282855357, "grad_norm": 0.5112528204917908, "learning_rate": 6.332214779824288e-05, "loss": 1.1505, "step": 3680 }, { "epoch": 0.6584472438130975, "grad_norm": 0.539935827255249, "learning_rate": 6.303215098397321e-05, "loss": 1.1044, "step": 3685 }, { "epoch": 0.6593406593406593, "grad_norm": 0.5397330522537231, "learning_rate": 6.274251382970427e-05, "loss": 1.1336, "step": 3690 }, { "epoch": 0.6602340748682212, "grad_norm": 0.5695128440856934, "learning_rate": 6.245323915331439e-05, "loss": 1.1468, "step": 3695 }, { "epoch": 0.6611274903957831, "grad_norm": 0.5581777095794678, "learning_rate": 6.216432976915527e-05, "loss": 1.1658, "step": 3700 }, { "epoch": 0.662020905923345, "grad_norm": 0.5400391817092896, "learning_rate": 6.187578848802475e-05, "loss": 1.1089, "step": 3705 }, { "epoch": 0.6629143214509068, "grad_norm": 0.5398588180541992, "learning_rate": 6.15876181171394e-05, "loss": 1.1412, "step": 3710 }, { "epoch": 0.6638077369784687, "grad_norm": 0.5350549817085266, "learning_rate": 6.129982146010713e-05, "loss": 1.1489, "step": 3715 }, { "epoch": 0.6647011525060306, "grad_norm": 0.49134427309036255, "learning_rate": 6.101240131690009e-05, "loss": 1.1069, "step": 3720 }, { "epoch": 0.6655945680335924, "grad_norm": 0.526871383190155, "learning_rate": 6.072536048382726e-05, "loss": 1.1264, "step": 3725 }, { "epoch": 0.6664879835611542, "grad_norm": 0.539218008518219, "learning_rate": 6.043870175350732e-05, "loss": 1.1582, "step": 3730 }, { "epoch": 0.6673813990887162, "grad_norm": 0.5474951267242432, "learning_rate": 6.0152427914841544e-05, "loss": 1.112, "step": 3735 }, { "epoch": 0.668274814616278, "grad_norm": 0.49627193808555603, "learning_rate": 5.9866541752986485e-05, "loss": 1.1059, "step": 3740 }, { "epoch": 0.6691682301438399, "grad_norm": 0.5210728645324707, "learning_rate": 5.958104604932706e-05, "loss": 1.1543, "step": 3745 }, { "epoch": 0.6700616456714018, "grad_norm": 0.510959267616272, "learning_rate": 5.9295943581449385e-05, "loss": 1.1369, "step": 3750 }, { "epoch": 0.6709550611989636, "grad_norm": 0.5323443412780762, "learning_rate": 5.901123712311385e-05, "loss": 1.1186, "step": 3755 }, { "epoch": 0.6718484767265255, "grad_norm": 0.5052646398544312, "learning_rate": 5.8726929444228016e-05, "loss": 1.1084, "step": 3760 }, { "epoch": 0.6727418922540874, "grad_norm": 0.5128947496414185, "learning_rate": 5.844302331081972e-05, "loss": 1.1218, "step": 3765 }, { "epoch": 0.6736353077816493, "grad_norm": 0.48404374718666077, "learning_rate": 5.8159521485010214e-05, "loss": 1.1055, "step": 3770 }, { "epoch": 0.6745287233092111, "grad_norm": 0.5284962058067322, "learning_rate": 5.787642672498719e-05, "loss": 1.1089, "step": 3775 }, { "epoch": 0.6754221388367729, "grad_norm": 0.49954989552497864, "learning_rate": 5.759374178497801e-05, "loss": 1.0823, "step": 3780 }, { "epoch": 0.6763155543643349, "grad_norm": 0.5159726142883301, "learning_rate": 5.731146941522292e-05, "loss": 1.1137, "step": 3785 }, { "epoch": 0.6772089698918967, "grad_norm": 0.5402414202690125, "learning_rate": 5.702961236194826e-05, "loss": 1.1067, "step": 3790 }, { "epoch": 0.6781023854194586, "grad_norm": 0.5399303436279297, "learning_rate": 5.674817336733975e-05, "loss": 1.106, "step": 3795 }, { "epoch": 0.6789958009470205, "grad_norm": 0.557214081287384, "learning_rate": 5.646715516951584e-05, "loss": 1.111, "step": 3800 }, { "epoch": 0.6798892164745823, "grad_norm": 0.5638577342033386, "learning_rate": 5.618656050250099e-05, "loss": 1.1255, "step": 3805 }, { "epoch": 0.6807826320021442, "grad_norm": 0.5255394577980042, "learning_rate": 5.5906392096199255e-05, "loss": 1.0976, "step": 3810 }, { "epoch": 0.6816760475297061, "grad_norm": 0.5245288610458374, "learning_rate": 5.562665267636751e-05, "loss": 1.1709, "step": 3815 }, { "epoch": 0.682569463057268, "grad_norm": 0.5460274815559387, "learning_rate": 5.5347344964588996e-05, "loss": 1.0736, "step": 3820 }, { "epoch": 0.6834628785848298, "grad_norm": 0.5235292911529541, "learning_rate": 5.506847167824696e-05, "loss": 1.1055, "step": 3825 }, { "epoch": 0.6843562941123916, "grad_norm": 0.5296208262443542, "learning_rate": 5.479003553049806e-05, "loss": 1.1081, "step": 3830 }, { "epoch": 0.6852497096399536, "grad_norm": 0.48945724964141846, "learning_rate": 5.4512039230246035e-05, "loss": 1.0935, "step": 3835 }, { "epoch": 0.6861431251675154, "grad_norm": 0.5146022439002991, "learning_rate": 5.42344854821154e-05, "loss": 1.128, "step": 3840 }, { "epoch": 0.6870365406950772, "grad_norm": 0.5302464962005615, "learning_rate": 5.395737698642503e-05, "loss": 1.1157, "step": 3845 }, { "epoch": 0.6879299562226392, "grad_norm": 0.502776026725769, "learning_rate": 5.368071643916194e-05, "loss": 1.1372, "step": 3850 }, { "epoch": 0.688823371750201, "grad_norm": 0.47847309708595276, "learning_rate": 5.3404506531955146e-05, "loss": 1.0927, "step": 3855 }, { "epoch": 0.6897167872777629, "grad_norm": 0.5378836989402771, "learning_rate": 5.3128749952049284e-05, "loss": 1.1137, "step": 3860 }, { "epoch": 0.6906102028053247, "grad_norm": 0.5127781629562378, "learning_rate": 5.2853449382278605e-05, "loss": 1.1158, "step": 3865 }, { "epoch": 0.6915036183328866, "grad_norm": 0.5223949551582336, "learning_rate": 5.2578607501040863e-05, "loss": 1.1043, "step": 3870 }, { "epoch": 0.6923970338604485, "grad_norm": 0.5131104588508606, "learning_rate": 5.2304226982271174e-05, "loss": 1.1444, "step": 3875 }, { "epoch": 0.6932904493880103, "grad_norm": 0.5290147662162781, "learning_rate": 5.203031049541621e-05, "loss": 1.1286, "step": 3880 }, { "epoch": 0.6941838649155723, "grad_norm": 0.5196212530136108, "learning_rate": 5.175686070540786e-05, "loss": 1.1189, "step": 3885 }, { "epoch": 0.6950772804431341, "grad_norm": 0.517926037311554, "learning_rate": 5.148388027263769e-05, "loss": 1.1648, "step": 3890 }, { "epoch": 0.6959706959706959, "grad_norm": 0.5101890563964844, "learning_rate": 5.1211371852930766e-05, "loss": 1.0711, "step": 3895 }, { "epoch": 0.6968641114982579, "grad_norm": 0.5416128039360046, "learning_rate": 5.0939338097520095e-05, "loss": 1.1141, "step": 3900 }, { "epoch": 0.6977575270258197, "grad_norm": 0.5422283411026001, "learning_rate": 5.0667781653020584e-05, "loss": 1.0783, "step": 3905 }, { "epoch": 0.6986509425533816, "grad_norm": 0.5199434161186218, "learning_rate": 5.039670516140338e-05, "loss": 1.1424, "step": 3910 }, { "epoch": 0.6995443580809434, "grad_norm": 0.5065299868583679, "learning_rate": 5.012611125997018e-05, "loss": 1.0881, "step": 3915 }, { "epoch": 0.7004377736085053, "grad_norm": 0.49652373790740967, "learning_rate": 4.9856002581327565e-05, "loss": 1.1281, "step": 3920 }, { "epoch": 0.7013311891360672, "grad_norm": 0.5640977621078491, "learning_rate": 4.958638175336137e-05, "loss": 1.1016, "step": 3925 }, { "epoch": 0.702224604663629, "grad_norm": 0.5382977724075317, "learning_rate": 4.931725139921126e-05, "loss": 1.1271, "step": 3930 }, { "epoch": 0.703118020191191, "grad_norm": 0.5227957963943481, "learning_rate": 4.9048614137244865e-05, "loss": 1.0931, "step": 3935 }, { "epoch": 0.7040114357187528, "grad_norm": 0.5395195484161377, "learning_rate": 4.878047258103267e-05, "loss": 1.1264, "step": 3940 }, { "epoch": 0.7049048512463146, "grad_norm": 0.5258668065071106, "learning_rate": 4.8512829339322375e-05, "loss": 1.1525, "step": 3945 }, { "epoch": 0.7057982667738766, "grad_norm": 0.5186535716056824, "learning_rate": 4.8245687016013696e-05, "loss": 1.119, "step": 3950 }, { "epoch": 0.7066916823014384, "grad_norm": 0.4964883327484131, "learning_rate": 4.797904821013278e-05, "loss": 1.1118, "step": 3955 }, { "epoch": 0.7075850978290003, "grad_norm": 0.4993223547935486, "learning_rate": 4.771291551580712e-05, "loss": 1.1115, "step": 3960 }, { "epoch": 0.7084785133565621, "grad_norm": 0.5137820243835449, "learning_rate": 4.744729152224024e-05, "loss": 1.0969, "step": 3965 }, { "epoch": 0.709371928884124, "grad_norm": 0.49753323197364807, "learning_rate": 4.71821788136865e-05, "loss": 1.0846, "step": 3970 }, { "epoch": 0.7102653444116859, "grad_norm": 0.5026699900627136, "learning_rate": 4.691757996942607e-05, "loss": 1.1105, "step": 3975 }, { "epoch": 0.7111587599392477, "grad_norm": 0.5345607995986938, "learning_rate": 4.665349756373957e-05, "loss": 1.1079, "step": 3980 }, { "epoch": 0.7120521754668097, "grad_norm": 0.5438992977142334, "learning_rate": 4.6389934165883306e-05, "loss": 1.1048, "step": 3985 }, { "epoch": 0.7129455909943715, "grad_norm": 0.48879095911979675, "learning_rate": 4.6126892340064096e-05, "loss": 1.1043, "step": 3990 }, { "epoch": 0.7138390065219333, "grad_norm": 0.5040327310562134, "learning_rate": 4.586437464541451e-05, "loss": 1.1026, "step": 3995 }, { "epoch": 0.7147324220494953, "grad_norm": 0.5185865759849548, "learning_rate": 4.560238363596771e-05, "loss": 1.0629, "step": 4000 }, { "epoch": 0.7156258375770571, "grad_norm": 0.4688703119754791, "learning_rate": 4.53409218606328e-05, "loss": 1.1158, "step": 4005 }, { "epoch": 0.716519253104619, "grad_norm": 0.5067079663276672, "learning_rate": 4.5079991863169966e-05, "loss": 1.0912, "step": 4010 }, { "epoch": 0.7174126686321808, "grad_norm": 0.5316655039787292, "learning_rate": 4.481959618216568e-05, "loss": 1.0901, "step": 4015 }, { "epoch": 0.7183060841597427, "grad_norm": 0.5138748288154602, "learning_rate": 4.455973735100818e-05, "loss": 1.1262, "step": 4020 }, { "epoch": 0.7191994996873046, "grad_norm": 0.5119633674621582, "learning_rate": 4.43004178978626e-05, "loss": 1.125, "step": 4025 }, { "epoch": 0.7200929152148664, "grad_norm": 0.5808505415916443, "learning_rate": 4.404164034564641e-05, "loss": 1.1447, "step": 4030 }, { "epoch": 0.7209863307424283, "grad_norm": 0.5414739847183228, "learning_rate": 4.378340721200501e-05, "loss": 1.1063, "step": 4035 }, { "epoch": 0.7218797462699902, "grad_norm": 0.5619556903839111, "learning_rate": 4.3525721009287215e-05, "loss": 1.1123, "step": 4040 }, { "epoch": 0.722773161797552, "grad_norm": 0.5197097063064575, "learning_rate": 4.326858424452063e-05, "loss": 1.0999, "step": 4045 }, { "epoch": 0.7236665773251139, "grad_norm": 0.5409570336341858, "learning_rate": 4.301199941938744e-05, "loss": 1.0848, "step": 4050 }, { "epoch": 0.7245599928526758, "grad_norm": 0.580436110496521, "learning_rate": 4.275596903020001e-05, "loss": 1.1184, "step": 4055 }, { "epoch": 0.7254534083802376, "grad_norm": 0.5301364660263062, "learning_rate": 4.250049556787655e-05, "loss": 1.1118, "step": 4060 }, { "epoch": 0.7263468239077995, "grad_norm": 0.5562632083892822, "learning_rate": 4.2245581517917065e-05, "loss": 1.0929, "step": 4065 }, { "epoch": 0.7272402394353614, "grad_norm": 0.5643694996833801, "learning_rate": 4.199122936037889e-05, "loss": 1.128, "step": 4070 }, { "epoch": 0.7281336549629233, "grad_norm": 0.5907604098320007, "learning_rate": 4.173744156985283e-05, "loss": 1.1354, "step": 4075 }, { "epoch": 0.7290270704904851, "grad_norm": 0.5513320565223694, "learning_rate": 4.148422061543884e-05, "loss": 1.1238, "step": 4080 }, { "epoch": 0.729920486018047, "grad_norm": 0.5430623292922974, "learning_rate": 4.123156896072219e-05, "loss": 1.0839, "step": 4085 }, { "epoch": 0.7308139015456089, "grad_norm": 0.5164535045623779, "learning_rate": 4.097948906374951e-05, "loss": 1.1311, "step": 4090 }, { "epoch": 0.7317073170731707, "grad_norm": 0.5333124399185181, "learning_rate": 4.0727983377004716e-05, "loss": 1.1088, "step": 4095 }, { "epoch": 0.7326007326007326, "grad_norm": 0.5494021773338318, "learning_rate": 4.047705434738527e-05, "loss": 1.0739, "step": 4100 }, { "epoch": 0.7334941481282945, "grad_norm": 0.5382459759712219, "learning_rate": 4.02267044161783e-05, "loss": 1.102, "step": 4105 }, { "epoch": 0.7343875636558563, "grad_norm": 0.5144065022468567, "learning_rate": 3.997693601903688e-05, "loss": 1.1283, "step": 4110 }, { "epoch": 0.7352809791834182, "grad_norm": 0.5298936367034912, "learning_rate": 3.9727751585956477e-05, "loss": 1.1328, "step": 4115 }, { "epoch": 0.7361743947109801, "grad_norm": 0.5441993474960327, "learning_rate": 3.9479153541251056e-05, "loss": 1.1399, "step": 4120 }, { "epoch": 0.737067810238542, "grad_norm": 0.5025814175605774, "learning_rate": 3.923114430352958e-05, "loss": 1.126, "step": 4125 }, { "epoch": 0.7379612257661038, "grad_norm": 0.512550950050354, "learning_rate": 3.8983726285672536e-05, "loss": 1.1322, "step": 4130 }, { "epoch": 0.7388546412936657, "grad_norm": 0.48689234256744385, "learning_rate": 3.8736901894808575e-05, "loss": 1.0553, "step": 4135 }, { "epoch": 0.7397480568212276, "grad_norm": 0.5119283199310303, "learning_rate": 3.849067353229078e-05, "loss": 1.1224, "step": 4140 }, { "epoch": 0.7406414723487894, "grad_norm": 0.5378916263580322, "learning_rate": 3.824504359367355e-05, "loss": 1.1229, "step": 4145 }, { "epoch": 0.7415348878763512, "grad_norm": 0.524563193321228, "learning_rate": 3.80000144686892e-05, "loss": 1.1329, "step": 4150 }, { "epoch": 0.7424283034039132, "grad_norm": 0.5638777017593384, "learning_rate": 3.775558854122475e-05, "loss": 1.0796, "step": 4155 }, { "epoch": 0.743321718931475, "grad_norm": 0.5309896469116211, "learning_rate": 3.7511768189298746e-05, "loss": 1.088, "step": 4160 }, { "epoch": 0.7442151344590369, "grad_norm": 0.5442638397216797, "learning_rate": 3.726855578503804e-05, "loss": 1.0958, "step": 4165 }, { "epoch": 0.7451085499865988, "grad_norm": 0.5414631962776184, "learning_rate": 3.70259536946548e-05, "loss": 1.103, "step": 4170 }, { "epoch": 0.7460019655141606, "grad_norm": 0.538829505443573, "learning_rate": 3.678396427842334e-05, "loss": 1.1218, "step": 4175 }, { "epoch": 0.7468953810417225, "grad_norm": 0.516124963760376, "learning_rate": 3.6542589890657476e-05, "loss": 1.0967, "step": 4180 }, { "epoch": 0.7477887965692843, "grad_norm": 0.5191131234169006, "learning_rate": 3.630183287968727e-05, "loss": 1.1197, "step": 4185 }, { "epoch": 0.7486822120968463, "grad_norm": 0.5277869701385498, "learning_rate": 3.606169558783635e-05, "loss": 1.0792, "step": 4190 }, { "epoch": 0.7495756276244081, "grad_norm": 0.5471330285072327, "learning_rate": 3.5822180351399136e-05, "loss": 1.0985, "step": 4195 }, { "epoch": 0.7504690431519699, "grad_norm": 0.5002536773681641, "learning_rate": 3.5583289500618e-05, "loss": 1.0913, "step": 4200 }, { "epoch": 0.7513624586795319, "grad_norm": 0.5431557297706604, "learning_rate": 3.53450253596608e-05, "loss": 1.0816, "step": 4205 }, { "epoch": 0.7522558742070937, "grad_norm": 0.5701403617858887, "learning_rate": 3.510739024659802e-05, "loss": 1.1358, "step": 4210 }, { "epoch": 0.7531492897346556, "grad_norm": 0.5729346871376038, "learning_rate": 3.487038647338038e-05, "loss": 1.0999, "step": 4215 }, { "epoch": 0.7540427052622175, "grad_norm": 0.5427317023277283, "learning_rate": 3.463401634581631e-05, "loss": 1.1184, "step": 4220 }, { "epoch": 0.7549361207897793, "grad_norm": 0.5459471344947815, "learning_rate": 3.4398282163549414e-05, "loss": 1.1179, "step": 4225 }, { "epoch": 0.7558295363173412, "grad_norm": 0.5358048677444458, "learning_rate": 3.416318622003634e-05, "loss": 1.0815, "step": 4230 }, { "epoch": 0.756722951844903, "grad_norm": 0.5081732869148254, "learning_rate": 3.39287308025242e-05, "loss": 1.1027, "step": 4235 }, { "epoch": 0.757616367372465, "grad_norm": 0.5418072938919067, "learning_rate": 3.369491819202849e-05, "loss": 1.112, "step": 4240 }, { "epoch": 0.7585097829000268, "grad_norm": 0.5014515519142151, "learning_rate": 3.34617506633108e-05, "loss": 1.1101, "step": 4245 }, { "epoch": 0.7594031984275886, "grad_norm": 0.5339967608451843, "learning_rate": 3.322923048485672e-05, "loss": 1.0815, "step": 4250 }, { "epoch": 0.7602966139551506, "grad_norm": 0.5515729188919067, "learning_rate": 3.2997359918853845e-05, "loss": 1.1077, "step": 4255 }, { "epoch": 0.7611900294827124, "grad_norm": 0.5655061602592468, "learning_rate": 3.276614122116962e-05, "loss": 1.0921, "step": 4260 }, { "epoch": 0.7620834450102743, "grad_norm": 0.5367369055747986, "learning_rate": 3.2535576641329514e-05, "loss": 1.091, "step": 4265 }, { "epoch": 0.7629768605378362, "grad_norm": 0.5201234817504883, "learning_rate": 3.230566842249497e-05, "loss": 1.0846, "step": 4270 }, { "epoch": 0.763870276065398, "grad_norm": 0.5395834445953369, "learning_rate": 3.2076418801441886e-05, "loss": 1.0972, "step": 4275 }, { "epoch": 0.7647636915929599, "grad_norm": 0.5469521284103394, "learning_rate": 3.1847830008538545e-05, "loss": 1.0966, "step": 4280 }, { "epoch": 0.7656571071205217, "grad_norm": 0.5283955931663513, "learning_rate": 3.1619904267724065e-05, "loss": 1.084, "step": 4285 }, { "epoch": 0.7665505226480837, "grad_norm": 0.5325857996940613, "learning_rate": 3.139264379648671e-05, "loss": 1.0797, "step": 4290 }, { "epoch": 0.7674439381756455, "grad_norm": 0.5412785410881042, "learning_rate": 3.116605080584235e-05, "loss": 1.0911, "step": 4295 }, { "epoch": 0.7683373537032073, "grad_norm": 0.5118827819824219, "learning_rate": 3.0940127500313e-05, "loss": 1.1005, "step": 4300 }, { "epoch": 0.7692307692307693, "grad_norm": 0.5167849063873291, "learning_rate": 3.071487607790524e-05, "loss": 1.143, "step": 4305 }, { "epoch": 0.7701241847583311, "grad_norm": 0.5016911625862122, "learning_rate": 3.049029873008893e-05, "loss": 1.0874, "step": 4310 }, { "epoch": 0.771017600285893, "grad_norm": 0.5407574772834778, "learning_rate": 3.0266397641775835e-05, "loss": 1.106, "step": 4315 }, { "epoch": 0.7719110158134549, "grad_norm": 0.5007250905036926, "learning_rate": 3.004317499129845e-05, "loss": 1.1162, "step": 4320 }, { "epoch": 0.7728044313410167, "grad_norm": 0.5197978019714355, "learning_rate": 2.9820632950388695e-05, "loss": 1.0865, "step": 4325 }, { "epoch": 0.7736978468685786, "grad_norm": 0.5517399311065674, "learning_rate": 2.9598773684156878e-05, "loss": 1.1008, "step": 4330 }, { "epoch": 0.7745912623961404, "grad_norm": 0.5117592215538025, "learning_rate": 2.9377599351070595e-05, "loss": 1.1244, "step": 4335 }, { "epoch": 0.7754846779237023, "grad_norm": 0.5437847971916199, "learning_rate": 2.915711210293367e-05, "loss": 1.1578, "step": 4340 }, { "epoch": 0.7763780934512642, "grad_norm": 0.5368949770927429, "learning_rate": 2.8937314084865407e-05, "loss": 1.1185, "step": 4345 }, { "epoch": 0.777271508978826, "grad_norm": 0.5332816243171692, "learning_rate": 2.8718207435279486e-05, "loss": 1.1008, "step": 4350 }, { "epoch": 0.778164924506388, "grad_norm": 0.5594480633735657, "learning_rate": 2.849979428586331e-05, "loss": 1.0855, "step": 4355 }, { "epoch": 0.7790583400339498, "grad_norm": 0.5220233201980591, "learning_rate": 2.828207676155722e-05, "loss": 1.0946, "step": 4360 }, { "epoch": 0.7799517555615116, "grad_norm": 0.5092322826385498, "learning_rate": 2.80650569805338e-05, "loss": 1.0925, "step": 4365 }, { "epoch": 0.7808451710890735, "grad_norm": 0.5742801427841187, "learning_rate": 2.784873705417731e-05, "loss": 1.0748, "step": 4370 }, { "epoch": 0.7817385866166354, "grad_norm": 0.5175203680992126, "learning_rate": 2.7633119087063152e-05, "loss": 1.1221, "step": 4375 }, { "epoch": 0.7826320021441973, "grad_norm": 0.5183749794960022, "learning_rate": 2.7418205176937327e-05, "loss": 1.1412, "step": 4380 }, { "epoch": 0.7835254176717591, "grad_norm": 0.5045965909957886, "learning_rate": 2.7203997414696104e-05, "loss": 1.1163, "step": 4385 }, { "epoch": 0.784418833199321, "grad_norm": 0.5237119793891907, "learning_rate": 2.6990497884365586e-05, "loss": 1.0593, "step": 4390 }, { "epoch": 0.7853122487268829, "grad_norm": 0.5403294563293457, "learning_rate": 2.67777086630816e-05, "loss": 1.1331, "step": 4395 }, { "epoch": 0.7862056642544447, "grad_norm": 0.5277557373046875, "learning_rate": 2.6565631821069304e-05, "loss": 1.0992, "step": 4400 }, { "epoch": 0.7870990797820067, "grad_norm": 0.5190049409866333, "learning_rate": 2.6354269421623112e-05, "loss": 1.131, "step": 4405 }, { "epoch": 0.7879924953095685, "grad_norm": 0.4882740080356598, "learning_rate": 2.6143623521086647e-05, "loss": 1.0654, "step": 4410 }, { "epoch": 0.7888859108371303, "grad_norm": 0.5139947533607483, "learning_rate": 2.5933696168832743e-05, "loss": 1.0843, "step": 4415 }, { "epoch": 0.7897793263646922, "grad_norm": 0.48566868901252747, "learning_rate": 2.5724489407243447e-05, "loss": 1.1022, "step": 4420 }, { "epoch": 0.7906727418922541, "grad_norm": 0.5145934224128723, "learning_rate": 2.5516005271690203e-05, "loss": 1.1237, "step": 4425 }, { "epoch": 0.791566157419816, "grad_norm": 0.51795893907547, "learning_rate": 2.530824579051403e-05, "loss": 1.0836, "step": 4430 }, { "epoch": 0.7924595729473778, "grad_norm": 0.5562942624092102, "learning_rate": 2.510121298500573e-05, "loss": 1.1093, "step": 4435 }, { "epoch": 0.7933529884749397, "grad_norm": 0.5125170350074768, "learning_rate": 2.4894908869386424e-05, "loss": 1.1064, "step": 4440 }, { "epoch": 0.7942464040025016, "grad_norm": 0.5201259851455688, "learning_rate": 2.4689335450787675e-05, "loss": 1.1108, "step": 4445 }, { "epoch": 0.7951398195300634, "grad_norm": 0.5387787222862244, "learning_rate": 2.4484494729232155e-05, "loss": 1.0988, "step": 4450 }, { "epoch": 0.7960332350576254, "grad_norm": 0.5109546184539795, "learning_rate": 2.428038869761412e-05, "loss": 1.0924, "step": 4455 }, { "epoch": 0.7969266505851872, "grad_norm": 0.524856448173523, "learning_rate": 2.4077019341680042e-05, "loss": 1.0735, "step": 4460 }, { "epoch": 0.797820066112749, "grad_norm": 0.5550493001937866, "learning_rate": 2.387438864000926e-05, "loss": 1.1041, "step": 4465 }, { "epoch": 0.7987134816403109, "grad_norm": 0.49559786915779114, "learning_rate": 2.3672498563994762e-05, "loss": 1.1026, "step": 4470 }, { "epoch": 0.7996068971678728, "grad_norm": 0.559374213218689, "learning_rate": 2.3471351077824e-05, "loss": 1.1167, "step": 4475 }, { "epoch": 0.8005003126954346, "grad_norm": 0.5809065699577332, "learning_rate": 2.3270948138459735e-05, "loss": 1.1228, "step": 4480 }, { "epoch": 0.8013937282229965, "grad_norm": 0.5337952971458435, "learning_rate": 2.3071291695621135e-05, "loss": 1.0787, "step": 4485 }, { "epoch": 0.8022871437505584, "grad_norm": 0.5353108048439026, "learning_rate": 2.2872383691764586e-05, "loss": 1.0598, "step": 4490 }, { "epoch": 0.8031805592781203, "grad_norm": 0.5580396056175232, "learning_rate": 2.2674226062064996e-05, "loss": 1.0763, "step": 4495 }, { "epoch": 0.8040739748056821, "grad_norm": 0.5546059608459473, "learning_rate": 2.2476820734396843e-05, "loss": 1.067, "step": 4500 }, { "epoch": 0.8049673903332439, "grad_norm": 0.5651270151138306, "learning_rate": 2.2280169629315484e-05, "loss": 1.134, "step": 4505 }, { "epoch": 0.8058608058608059, "grad_norm": 0.5143851637840271, "learning_rate": 2.208427466003844e-05, "loss": 1.0975, "step": 4510 }, { "epoch": 0.8067542213883677, "grad_norm": 0.5212395191192627, "learning_rate": 2.1889137732426802e-05, "loss": 1.1308, "step": 4515 }, { "epoch": 0.8076476369159296, "grad_norm": 0.5373924374580383, "learning_rate": 2.1694760744966668e-05, "loss": 1.134, "step": 4520 }, { "epoch": 0.8085410524434915, "grad_norm": 0.5924869179725647, "learning_rate": 2.1501145588750694e-05, "loss": 1.0626, "step": 4525 }, { "epoch": 0.8094344679710533, "grad_norm": 0.5431603193283081, "learning_rate": 2.1308294147459628e-05, "loss": 1.1114, "step": 4530 }, { "epoch": 0.8103278834986152, "grad_norm": 0.548326849937439, "learning_rate": 2.1116208297344155e-05, "loss": 1.1051, "step": 4535 }, { "epoch": 0.8112212990261771, "grad_norm": 0.5223492383956909, "learning_rate": 2.0924889907206425e-05, "loss": 1.071, "step": 4540 }, { "epoch": 0.812114714553739, "grad_norm": 0.5413901805877686, "learning_rate": 2.0734340838382015e-05, "loss": 1.0917, "step": 4545 }, { "epoch": 0.8130081300813008, "grad_norm": 0.5205683708190918, "learning_rate": 2.0544562944721778e-05, "loss": 1.0875, "step": 4550 }, { "epoch": 0.8139015456088626, "grad_norm": 0.6497357487678528, "learning_rate": 2.0355558072573787e-05, "loss": 1.1216, "step": 4555 }, { "epoch": 0.8147949611364246, "grad_norm": 0.5288578867912292, "learning_rate": 2.016732806076542e-05, "loss": 1.0919, "step": 4560 }, { "epoch": 0.8156883766639864, "grad_norm": 0.5458259582519531, "learning_rate": 1.9979874740585426e-05, "loss": 1.1187, "step": 4565 }, { "epoch": 0.8165817921915483, "grad_norm": 0.5482889413833618, "learning_rate": 1.979319993576614e-05, "loss": 1.083, "step": 4570 }, { "epoch": 0.8174752077191102, "grad_norm": 0.535325288772583, "learning_rate": 1.9607305462465686e-05, "loss": 1.0897, "step": 4575 }, { "epoch": 0.818368623246672, "grad_norm": 0.5331825017929077, "learning_rate": 1.942219312925042e-05, "loss": 1.1001, "step": 4580 }, { "epoch": 0.8192620387742339, "grad_norm": 0.541584312915802, "learning_rate": 1.9237864737077204e-05, "loss": 1.07, "step": 4585 }, { "epoch": 0.8201554543017958, "grad_norm": 0.49262598156929016, "learning_rate": 1.9054322079275953e-05, "loss": 1.1154, "step": 4590 }, { "epoch": 0.8210488698293577, "grad_norm": 0.510992705821991, "learning_rate": 1.8871566941532182e-05, "loss": 1.0934, "step": 4595 }, { "epoch": 0.8219422853569195, "grad_norm": 0.47639334201812744, "learning_rate": 1.8689601101869604e-05, "loss": 1.0779, "step": 4600 }, { "epoch": 0.8228357008844813, "grad_norm": 0.5555627346038818, "learning_rate": 1.8508426330632933e-05, "loss": 1.0919, "step": 4605 }, { "epoch": 0.8237291164120433, "grad_norm": 0.5235790014266968, "learning_rate": 1.8328044390470478e-05, "loss": 1.0678, "step": 4610 }, { "epoch": 0.8246225319396051, "grad_norm": 0.5292229652404785, "learning_rate": 1.8148457036317157e-05, "loss": 1.0999, "step": 4615 }, { "epoch": 0.8255159474671669, "grad_norm": 0.5767297744750977, "learning_rate": 1.796966601537734e-05, "loss": 1.0964, "step": 4620 }, { "epoch": 0.8264093629947289, "grad_norm": 0.5446600317955017, "learning_rate": 1.7791673067107927e-05, "loss": 1.1066, "step": 4625 }, { "epoch": 0.8273027785222907, "grad_norm": 0.48151037096977234, "learning_rate": 1.7614479923201333e-05, "loss": 1.0794, "step": 4630 }, { "epoch": 0.8281961940498526, "grad_norm": 0.5238900780677795, "learning_rate": 1.7438088307568667e-05, "loss": 1.073, "step": 4635 }, { "epoch": 0.8290896095774145, "grad_norm": 0.5254760384559631, "learning_rate": 1.7262499936322997e-05, "loss": 1.139, "step": 4640 }, { "epoch": 0.8299830251049763, "grad_norm": 0.5375151038169861, "learning_rate": 1.708771651776263e-05, "loss": 1.1143, "step": 4645 }, { "epoch": 0.8308764406325382, "grad_norm": 0.5599473118782043, "learning_rate": 1.6913739752354464e-05, "loss": 1.1043, "step": 4650 }, { "epoch": 0.8317698561601, "grad_norm": 0.5429250597953796, "learning_rate": 1.6740571332717558e-05, "loss": 1.096, "step": 4655 }, { "epoch": 0.832663271687662, "grad_norm": 0.5355926752090454, "learning_rate": 1.6568212943606465e-05, "loss": 1.103, "step": 4660 }, { "epoch": 0.8335566872152238, "grad_norm": 0.5205219388008118, "learning_rate": 1.6396666261895034e-05, "loss": 1.0765, "step": 4665 }, { "epoch": 0.8344501027427856, "grad_norm": 0.5437434315681458, "learning_rate": 1.6225932956559943e-05, "loss": 1.0793, "step": 4670 }, { "epoch": 0.8353435182703476, "grad_norm": 0.5111892223358154, "learning_rate": 1.6056014688664656e-05, "loss": 1.1069, "step": 4675 }, { "epoch": 0.8362369337979094, "grad_norm": 0.5251837968826294, "learning_rate": 1.5886913111343037e-05, "loss": 1.0945, "step": 4680 }, { "epoch": 0.8371303493254713, "grad_norm": 0.5311691164970398, "learning_rate": 1.571862986978342e-05, "loss": 1.1, "step": 4685 }, { "epoch": 0.8380237648530331, "grad_norm": 0.515058696269989, "learning_rate": 1.555116660121253e-05, "loss": 1.1141, "step": 4690 }, { "epoch": 0.838917180380595, "grad_norm": 0.5322168469429016, "learning_rate": 1.538452493487956e-05, "loss": 1.1025, "step": 4695 }, { "epoch": 0.8398105959081569, "grad_norm": 0.5564744472503662, "learning_rate": 1.5218706492040435e-05, "loss": 1.0679, "step": 4700 }, { "epoch": 0.8407040114357187, "grad_norm": 0.52479088306427, "learning_rate": 1.5053712885941862e-05, "loss": 1.0985, "step": 4705 }, { "epoch": 0.8415974269632807, "grad_norm": 0.5248088240623474, "learning_rate": 1.4889545721805687e-05, "loss": 1.0503, "step": 4710 }, { "epoch": 0.8424908424908425, "grad_norm": 0.525884747505188, "learning_rate": 1.4726206596813363e-05, "loss": 1.1073, "step": 4715 }, { "epoch": 0.8433842580184043, "grad_norm": 0.5724270343780518, "learning_rate": 1.456369710009038e-05, "loss": 1.126, "step": 4720 }, { "epoch": 0.8442776735459663, "grad_norm": 0.4863172173500061, "learning_rate": 1.4402018812690721e-05, "loss": 1.0695, "step": 4725 }, { "epoch": 0.8451710890735281, "grad_norm": 0.5609614849090576, "learning_rate": 1.4241173307581558e-05, "loss": 1.0684, "step": 4730 }, { "epoch": 0.84606450460109, "grad_norm": 0.5545142292976379, "learning_rate": 1.4081162149627936e-05, "loss": 1.0362, "step": 4735 }, { "epoch": 0.8469579201286518, "grad_norm": 0.5475736856460571, "learning_rate": 1.39219868955775e-05, "loss": 1.0926, "step": 4740 }, { "epoch": 0.8478513356562137, "grad_norm": 0.4896586537361145, "learning_rate": 1.3763649094045483e-05, "loss": 1.0675, "step": 4745 }, { "epoch": 0.8487447511837756, "grad_norm": 0.49385780096054077, "learning_rate": 1.3606150285499475e-05, "loss": 1.0739, "step": 4750 }, { "epoch": 0.8496381667113374, "grad_norm": 0.5389235019683838, "learning_rate": 1.3449492002244502e-05, "loss": 1.0808, "step": 4755 }, { "epoch": 0.8505315822388994, "grad_norm": 0.5244671702384949, "learning_rate": 1.329367576840812e-05, "loss": 1.0967, "step": 4760 }, { "epoch": 0.8514249977664612, "grad_norm": 0.5618093013763428, "learning_rate": 1.3138703099925676e-05, "loss": 1.0743, "step": 4765 }, { "epoch": 0.852318413294023, "grad_norm": 0.5471380949020386, "learning_rate": 1.2984575504525376e-05, "loss": 1.1035, "step": 4770 }, { "epoch": 0.853211828821585, "grad_norm": 0.5013295412063599, "learning_rate": 1.2831294481713763e-05, "loss": 1.0736, "step": 4775 }, { "epoch": 0.8541052443491468, "grad_norm": 0.5280576348304749, "learning_rate": 1.2678861522761066e-05, "loss": 1.0825, "step": 4780 }, { "epoch": 0.8549986598767086, "grad_norm": 0.5448872447013855, "learning_rate": 1.2527278110686712e-05, "loss": 1.0895, "step": 4785 }, { "epoch": 0.8558920754042705, "grad_norm": 0.5476821660995483, "learning_rate": 1.237654572024487e-05, "loss": 1.1082, "step": 4790 }, { "epoch": 0.8567854909318324, "grad_norm": 0.5306336283683777, "learning_rate": 1.2226665817910166e-05, "loss": 1.0715, "step": 4795 }, { "epoch": 0.8576789064593943, "grad_norm": 0.5052401423454285, "learning_rate": 1.2077639861863365e-05, "loss": 1.0731, "step": 4800 }, { "epoch": 0.8585723219869561, "grad_norm": 0.5403242707252502, "learning_rate": 1.1929469301977136e-05, "loss": 1.0941, "step": 4805 }, { "epoch": 0.859465737514518, "grad_norm": 0.5217272639274597, "learning_rate": 1.1782155579802034e-05, "loss": 1.0628, "step": 4810 }, { "epoch": 0.8603591530420799, "grad_norm": 0.570672869682312, "learning_rate": 1.1635700128552508e-05, "loss": 1.1157, "step": 4815 }, { "epoch": 0.8612525685696417, "grad_norm": 0.5487679839134216, "learning_rate": 1.1490104373092825e-05, "loss": 1.0981, "step": 4820 }, { "epoch": 0.8621459840972036, "grad_norm": 0.5410107970237732, "learning_rate": 1.1345369729923271e-05, "loss": 1.0917, "step": 4825 }, { "epoch": 0.8630393996247655, "grad_norm": 0.5344003438949585, "learning_rate": 1.1201497607166423e-05, "loss": 1.1058, "step": 4830 }, { "epoch": 0.8639328151523273, "grad_norm": 0.548433780670166, "learning_rate": 1.105848940455334e-05, "loss": 1.0724, "step": 4835 }, { "epoch": 0.8648262306798892, "grad_norm": 0.573233962059021, "learning_rate": 1.0916346513410081e-05, "loss": 1.0733, "step": 4840 }, { "epoch": 0.8657196462074511, "grad_norm": 0.5367952585220337, "learning_rate": 1.0775070316644042e-05, "loss": 1.0829, "step": 4845 }, { "epoch": 0.866613061735013, "grad_norm": 0.5424239635467529, "learning_rate": 1.0634662188730604e-05, "loss": 1.11, "step": 4850 }, { "epoch": 0.8675064772625748, "grad_norm": 0.5632114410400391, "learning_rate": 1.0495123495699588e-05, "loss": 1.0926, "step": 4855 }, { "epoch": 0.8683998927901367, "grad_norm": 0.517599880695343, "learning_rate": 1.0356455595122239e-05, "loss": 1.1214, "step": 4860 }, { "epoch": 0.8692933083176986, "grad_norm": 0.5182797908782959, "learning_rate": 1.02186598360978e-05, "loss": 1.0759, "step": 4865 }, { "epoch": 0.8701867238452604, "grad_norm": 0.5280758142471313, "learning_rate": 1.0081737559240445e-05, "loss": 1.1264, "step": 4870 }, { "epoch": 0.8710801393728222, "grad_norm": 0.534575343132019, "learning_rate": 9.945690096666249e-06, "loss": 1.0818, "step": 4875 }, { "epoch": 0.8719735549003842, "grad_norm": 0.5310872793197632, "learning_rate": 9.810518771980225e-06, "loss": 1.1092, "step": 4880 }, { "epoch": 0.872866970427946, "grad_norm": 0.5414004921913147, "learning_rate": 9.676224900263497e-06, "loss": 1.1235, "step": 4885 }, { "epoch": 0.8737603859555079, "grad_norm": 0.5505749583244324, "learning_rate": 9.542809788060358e-06, "loss": 1.1467, "step": 4890 }, { "epoch": 0.8746538014830698, "grad_norm": 0.507029116153717, "learning_rate": 9.410274733365753e-06, "loss": 1.0818, "step": 4895 }, { "epoch": 0.8755472170106317, "grad_norm": 0.5631771683692932, "learning_rate": 9.278621025612434e-06, "loss": 1.0975, "step": 4900 }, { "epoch": 0.8764406325381935, "grad_norm": 0.6328200697898865, "learning_rate": 9.147849945658648e-06, "loss": 1.0853, "step": 4905 }, { "epoch": 0.8773340480657554, "grad_norm": 0.5625194907188416, "learning_rate": 9.017962765775523e-06, "loss": 1.1083, "step": 4910 }, { "epoch": 0.8782274635933173, "grad_norm": 0.5809071063995361, "learning_rate": 8.888960749634712e-06, "loss": 1.0833, "step": 4915 }, { "epoch": 0.8791208791208791, "grad_norm": 0.5429468154907227, "learning_rate": 8.760845152296116e-06, "loss": 1.0814, "step": 4920 }, { "epoch": 0.8800142946484409, "grad_norm": 0.5104892253875732, "learning_rate": 8.63361722019569e-06, "loss": 1.0644, "step": 4925 }, { "epoch": 0.8809077101760029, "grad_norm": 0.5056647658348083, "learning_rate": 8.507278191133261e-06, "loss": 1.0679, "step": 4930 }, { "epoch": 0.8818011257035647, "grad_norm": 0.5692788362503052, "learning_rate": 8.38182929426059e-06, "loss": 1.112, "step": 4935 }, { "epoch": 0.8826945412311266, "grad_norm": 0.560130774974823, "learning_rate": 8.257271750069295e-06, "loss": 1.1309, "step": 4940 }, { "epoch": 0.8835879567586885, "grad_norm": 0.5144816040992737, "learning_rate": 8.133606770379055e-06, "loss": 1.0819, "step": 4945 }, { "epoch": 0.8844813722862503, "grad_norm": 0.484380304813385, "learning_rate": 8.010835558325735e-06, "loss": 1.0844, "step": 4950 }, { "epoch": 0.8853747878138122, "grad_norm": 0.5501832365989685, "learning_rate": 7.88895930834983e-06, "loss": 1.1013, "step": 4955 }, { "epoch": 0.8862682033413741, "grad_norm": 0.49795544147491455, "learning_rate": 7.767979206184694e-06, "loss": 1.0627, "step": 4960 }, { "epoch": 0.887161618868936, "grad_norm": 0.5382062792778015, "learning_rate": 7.64789642884508e-06, "loss": 1.0758, "step": 4965 }, { "epoch": 0.8880550343964978, "grad_norm": 0.5314489603042603, "learning_rate": 7.528712144615679e-06, "loss": 1.0501, "step": 4970 }, { "epoch": 0.8889484499240596, "grad_norm": 0.5101165175437927, "learning_rate": 7.4104275130397085e-06, "loss": 1.0807, "step": 4975 }, { "epoch": 0.8898418654516216, "grad_norm": 0.5354017019271851, "learning_rate": 7.29304368490773e-06, "loss": 1.0803, "step": 4980 }, { "epoch": 0.8907352809791834, "grad_norm": 0.5763533711433411, "learning_rate": 7.176561802246373e-06, "loss": 1.1026, "step": 4985 }, { "epoch": 0.8916286965067453, "grad_norm": 0.537377655506134, "learning_rate": 7.06098299830722e-06, "loss": 1.0888, "step": 4990 }, { "epoch": 0.8925221120343072, "grad_norm": 0.5555933713912964, "learning_rate": 6.946308397555823e-06, "loss": 1.1351, "step": 4995 }, { "epoch": 0.893415527561869, "grad_norm": 0.5403766632080078, "learning_rate": 6.832539115660752e-06, "loss": 1.0891, "step": 5000 }, { "epoch": 0.8943089430894309, "grad_norm": 0.517660915851593, "learning_rate": 6.719676259482721e-06, "loss": 1.0607, "step": 5005 }, { "epoch": 0.8952023586169927, "grad_norm": 0.5434561371803284, "learning_rate": 6.607720927063843e-06, "loss": 1.0982, "step": 5010 }, { "epoch": 0.8960957741445547, "grad_norm": 0.563258707523346, "learning_rate": 6.496674207616926e-06, "loss": 1.0701, "step": 5015 }, { "epoch": 0.8969891896721165, "grad_norm": 0.511574387550354, "learning_rate": 6.386537181514896e-06, "loss": 1.0808, "step": 5020 }, { "epoch": 0.8978826051996783, "grad_norm": 0.49810346961021423, "learning_rate": 6.277310920280299e-06, "loss": 1.0664, "step": 5025 }, { "epoch": 0.8987760207272403, "grad_norm": 0.526667058467865, "learning_rate": 6.1689964865748185e-06, "loss": 1.0714, "step": 5030 }, { "epoch": 0.8996694362548021, "grad_norm": 0.5600285530090332, "learning_rate": 6.061594934188985e-06, "loss": 1.0706, "step": 5035 }, { "epoch": 0.900562851782364, "grad_norm": 0.5716047883033752, "learning_rate": 5.955107308031915e-06, "loss": 1.0806, "step": 5040 }, { "epoch": 0.9014562673099259, "grad_norm": 0.5603742599487305, "learning_rate": 5.849534644121146e-06, "loss": 1.0755, "step": 5045 }, { "epoch": 0.9023496828374877, "grad_norm": 0.5195134878158569, "learning_rate": 5.744877969572537e-06, "loss": 1.0537, "step": 5050 }, { "epoch": 0.9032430983650496, "grad_norm": 0.5341048240661621, "learning_rate": 5.6411383025903205e-06, "loss": 1.0951, "step": 5055 }, { "epoch": 0.9041365138926114, "grad_norm": 0.55425626039505, "learning_rate": 5.538316652457121e-06, "loss": 1.0755, "step": 5060 }, { "epoch": 0.9050299294201734, "grad_norm": 0.520702600479126, "learning_rate": 5.436414019524216e-06, "loss": 1.1157, "step": 5065 }, { "epoch": 0.9059233449477352, "grad_norm": 0.5679965019226074, "learning_rate": 5.335431395201784e-06, "loss": 1.0801, "step": 5070 }, { "epoch": 0.906816760475297, "grad_norm": 0.5292661190032959, "learning_rate": 5.235369761949216e-06, "loss": 1.0908, "step": 5075 }, { "epoch": 0.907710176002859, "grad_norm": 0.5407772660255432, "learning_rate": 5.136230093265593e-06, "loss": 1.0639, "step": 5080 }, { "epoch": 0.9086035915304208, "grad_norm": 0.5404497981071472, "learning_rate": 5.038013353680204e-06, "loss": 1.1004, "step": 5085 }, { "epoch": 0.9094970070579826, "grad_norm": 0.5132922530174255, "learning_rate": 4.940720498743179e-06, "loss": 1.107, "step": 5090 }, { "epoch": 0.9103904225855446, "grad_norm": 0.49404093623161316, "learning_rate": 4.8443524750161676e-06, "loss": 1.1043, "step": 5095 }, { "epoch": 0.9112838381131064, "grad_norm": 0.5417544841766357, "learning_rate": 4.74891022006313e-06, "loss": 1.0917, "step": 5100 }, { "epoch": 0.9121772536406683, "grad_norm": 0.5480242371559143, "learning_rate": 4.654394662441264e-06, "loss": 1.0834, "step": 5105 }, { "epoch": 0.9130706691682301, "grad_norm": 0.555147647857666, "learning_rate": 4.560806721691913e-06, "loss": 1.0887, "step": 5110 }, { "epoch": 0.913964084695792, "grad_norm": 0.5397526621818542, "learning_rate": 4.468147308331605e-06, "loss": 1.1017, "step": 5115 }, { "epoch": 0.9148575002233539, "grad_norm": 0.510307252407074, "learning_rate": 4.376417323843318e-06, "loss": 1.0739, "step": 5120 }, { "epoch": 0.9157509157509157, "grad_norm": 0.5409055948257446, "learning_rate": 4.28561766066754e-06, "loss": 1.101, "step": 5125 }, { "epoch": 0.9166443312784777, "grad_norm": 0.552873969078064, "learning_rate": 4.195749202193699e-06, "loss": 1.0913, "step": 5130 }, { "epoch": 0.9175377468060395, "grad_norm": 0.5643073916435242, "learning_rate": 4.106812822751538e-06, "loss": 1.0779, "step": 5135 }, { "epoch": 0.9184311623336013, "grad_norm": 0.523567259311676, "learning_rate": 4.018809387602596e-06, "loss": 1.0872, "step": 5140 }, { "epoch": 0.9193245778611632, "grad_norm": 0.5306380391120911, "learning_rate": 3.931739752931829e-06, "loss": 1.0917, "step": 5145 }, { "epoch": 0.9202179933887251, "grad_norm": 0.5733050107955933, "learning_rate": 3.845604765839228e-06, "loss": 1.111, "step": 5150 }, { "epoch": 0.921111408916287, "grad_norm": 0.5619934797286987, "learning_rate": 3.760405264331612e-06, "loss": 1.065, "step": 5155 }, { "epoch": 0.9220048244438488, "grad_norm": 0.5529298186302185, "learning_rate": 3.676142077314448e-06, "loss": 1.0899, "step": 5160 }, { "epoch": 0.9228982399714107, "grad_norm": 0.5153440237045288, "learning_rate": 3.592816024583856e-06, "loss": 1.0995, "step": 5165 }, { "epoch": 0.9237916554989726, "grad_norm": 0.5353826880455017, "learning_rate": 3.510427916818526e-06, "loss": 1.0906, "step": 5170 }, { "epoch": 0.9246850710265344, "grad_norm": 0.5490701794624329, "learning_rate": 3.4289785555719157e-06, "loss": 1.1228, "step": 5175 }, { "epoch": 0.9255784865540964, "grad_norm": 0.5379487872123718, "learning_rate": 3.348468733264398e-06, "loss": 1.0582, "step": 5180 }, { "epoch": 0.9264719020816582, "grad_norm": 0.5375136733055115, "learning_rate": 3.268899233175604e-06, "loss": 1.0934, "step": 5185 }, { "epoch": 0.92736531760922, "grad_norm": 0.51712965965271, "learning_rate": 3.19027082943677e-06, "loss": 1.0704, "step": 5190 }, { "epoch": 0.9282587331367819, "grad_norm": 0.512758195400238, "learning_rate": 3.1125842870232014e-06, "loss": 1.124, "step": 5195 }, { "epoch": 0.9291521486643438, "grad_norm": 0.58779376745224, "learning_rate": 3.0358403617468446e-06, "loss": 1.1403, "step": 5200 }, { "epoch": 0.9300455641919056, "grad_norm": 0.5201031565666199, "learning_rate": 2.960039800248915e-06, "loss": 1.0889, "step": 5205 }, { "epoch": 0.9309389797194675, "grad_norm": 0.49719128012657166, "learning_rate": 2.885183339992692e-06, "loss": 1.0881, "step": 5210 }, { "epoch": 0.9318323952470294, "grad_norm": 0.5498881340026855, "learning_rate": 2.8112717092562358e-06, "loss": 1.0472, "step": 5215 }, { "epoch": 0.9327258107745913, "grad_norm": 0.5250979661941528, "learning_rate": 2.738305627125415e-06, "loss": 1.0962, "step": 5220 }, { "epoch": 0.9336192263021531, "grad_norm": 0.5712759494781494, "learning_rate": 2.6662858034868454e-06, "loss": 1.1149, "step": 5225 }, { "epoch": 0.934512641829715, "grad_norm": 0.5062239170074463, "learning_rate": 2.5952129390209854e-06, "loss": 1.0905, "step": 5230 }, { "epoch": 0.9354060573572769, "grad_norm": 0.5184369683265686, "learning_rate": 2.525087725195352e-06, "loss": 1.0709, "step": 5235 }, { "epoch": 0.9362994728848387, "grad_norm": 0.5225459933280945, "learning_rate": 2.4559108442577585e-06, "loss": 1.0738, "step": 5240 }, { "epoch": 0.9371928884124006, "grad_norm": 0.5459494590759277, "learning_rate": 2.38768296922971e-06, "loss": 1.064, "step": 5245 }, { "epoch": 0.9380863039399625, "grad_norm": 0.47445785999298096, "learning_rate": 2.3204047638998195e-06, "loss": 1.0625, "step": 5250 }, { "epoch": 0.9389797194675243, "grad_norm": 0.5076881051063538, "learning_rate": 2.2540768828173795e-06, "loss": 1.0736, "step": 5255 }, { "epoch": 0.9398731349950862, "grad_norm": 0.5733228921890259, "learning_rate": 2.1886999712860014e-06, "loss": 1.1301, "step": 5260 }, { "epoch": 0.9407665505226481, "grad_norm": 0.5537548065185547, "learning_rate": 2.1242746653572845e-06, "loss": 1.1, "step": 5265 }, { "epoch": 0.94165996605021, "grad_norm": 0.5685492753982544, "learning_rate": 2.060801591824668e-06, "loss": 1.1201, "step": 5270 }, { "epoch": 0.9425533815777718, "grad_norm": 0.5240213871002197, "learning_rate": 1.9982813682173586e-06, "loss": 1.0814, "step": 5275 }, { "epoch": 0.9434467971053337, "grad_norm": 0.5279719829559326, "learning_rate": 1.936714602794254e-06, "loss": 1.0812, "step": 5280 }, { "epoch": 0.9443402126328956, "grad_norm": 0.5382205247879028, "learning_rate": 1.8761018945380849e-06, "loss": 1.06, "step": 5285 }, { "epoch": 0.9452336281604574, "grad_norm": 0.503291666507721, "learning_rate": 1.8164438331495614e-06, "loss": 1.1014, "step": 5290 }, { "epoch": 0.9461270436880193, "grad_norm": 0.5597132444381714, "learning_rate": 1.7577409990416237e-06, "loss": 1.0768, "step": 5295 }, { "epoch": 0.9470204592155812, "grad_norm": 0.5632822513580322, "learning_rate": 1.6999939633338236e-06, "loss": 1.1121, "step": 5300 }, { "epoch": 0.947913874743143, "grad_norm": 0.5778818130493164, "learning_rate": 1.6432032878467729e-06, "loss": 1.0763, "step": 5305 }, { "epoch": 0.9488072902707049, "grad_norm": 0.5174503922462463, "learning_rate": 1.587369525096627e-06, "loss": 1.0799, "step": 5310 }, { "epoch": 0.9497007057982668, "grad_norm": 0.5436321496963501, "learning_rate": 1.5324932182897656e-06, "loss": 1.0975, "step": 5315 }, { "epoch": 0.9505941213258287, "grad_norm": 0.5373766422271729, "learning_rate": 1.4785749013174754e-06, "loss": 1.0808, "step": 5320 }, { "epoch": 0.9514875368533905, "grad_norm": 0.5411360859870911, "learning_rate": 1.4256150987507544e-06, "loss": 1.1184, "step": 5325 }, { "epoch": 0.9523809523809523, "grad_norm": 0.5489704608917236, "learning_rate": 1.3736143258352707e-06, "loss": 1.0733, "step": 5330 }, { "epoch": 0.9532743679085143, "grad_norm": 0.5345862507820129, "learning_rate": 1.322573088486212e-06, "loss": 1.1047, "step": 5335 }, { "epoch": 0.9541677834360761, "grad_norm": 0.4998050630092621, "learning_rate": 1.272491883283533e-06, "loss": 1.1046, "step": 5340 }, { "epoch": 0.955061198963638, "grad_norm": 0.5264286994934082, "learning_rate": 1.2233711974669714e-06, "loss": 1.0887, "step": 5345 }, { "epoch": 0.9559546144911999, "grad_norm": 0.5480632185935974, "learning_rate": 1.1752115089314398e-06, "loss": 1.1008, "step": 5350 }, { "epoch": 0.9568480300187617, "grad_norm": 0.5870484113693237, "learning_rate": 1.1280132862222737e-06, "loss": 1.0944, "step": 5355 }, { "epoch": 0.9577414455463236, "grad_norm": 0.5417635440826416, "learning_rate": 1.081776988530725e-06, "loss": 1.0934, "step": 5360 }, { "epoch": 0.9586348610738855, "grad_norm": 0.5396819114685059, "learning_rate": 1.0365030656894759e-06, "loss": 1.1019, "step": 5365 }, { "epoch": 0.9595282766014473, "grad_norm": 0.521092414855957, "learning_rate": 9.921919581682759e-07, "loss": 1.0899, "step": 5370 }, { "epoch": 0.9604216921290092, "grad_norm": 0.5292440056800842, "learning_rate": 9.488440970696566e-07, "loss": 1.1232, "step": 5375 }, { "epoch": 0.961315107656571, "grad_norm": 0.5165191292762756, "learning_rate": 9.064599041247124e-07, "loss": 1.1398, "step": 5380 }, { "epoch": 0.962208523184133, "grad_norm": 0.5134493708610535, "learning_rate": 8.650397916890263e-07, "loss": 1.0735, "step": 5385 }, { "epoch": 0.9631019387116948, "grad_norm": 0.5425078868865967, "learning_rate": 8.245841627386397e-07, "loss": 1.1054, "step": 5390 }, { "epoch": 0.9639953542392566, "grad_norm": 0.556138277053833, "learning_rate": 7.850934108661556e-07, "loss": 1.0938, "step": 5395 }, { "epoch": 0.9648887697668186, "grad_norm": 0.5389893054962158, "learning_rate": 7.465679202768749e-07, "loss": 1.1052, "step": 5400 }, { "epoch": 0.9657821852943804, "grad_norm": 0.47252729535102844, "learning_rate": 7.090080657850884e-07, "loss": 1.063, "step": 5405 }, { "epoch": 0.9666756008219423, "grad_norm": 0.4975879490375519, "learning_rate": 6.724142128104239e-07, "loss": 1.0983, "step": 5410 }, { "epoch": 0.9675690163495042, "grad_norm": 0.5652892589569092, "learning_rate": 6.367867173742603e-07, "loss": 1.075, "step": 5415 }, { "epoch": 0.968462431877066, "grad_norm": 0.5284938216209412, "learning_rate": 6.021259260963085e-07, "loss": 1.0788, "step": 5420 }, { "epoch": 0.9693558474046279, "grad_norm": 0.4806806445121765, "learning_rate": 5.684321761912247e-07, "loss": 1.0439, "step": 5425 }, { "epoch": 0.9702492629321897, "grad_norm": 0.5021379590034485, "learning_rate": 5.357057954653244e-07, "loss": 1.0704, "step": 5430 }, { "epoch": 0.9711426784597517, "grad_norm": 0.5142560005187988, "learning_rate": 5.039471023133624e-07, "loss": 1.1126, "step": 5435 }, { "epoch": 0.9720360939873135, "grad_norm": 0.5299791693687439, "learning_rate": 4.7315640571550246e-07, "loss": 1.1047, "step": 5440 }, { "epoch": 0.9729295095148753, "grad_norm": 0.49153584241867065, "learning_rate": 4.433340052342749e-07, "loss": 1.072, "step": 5445 }, { "epoch": 0.9738229250424373, "grad_norm": 0.5386480093002319, "learning_rate": 4.1448019101163473e-07, "loss": 1.0694, "step": 5450 }, { "epoch": 0.9747163405699991, "grad_norm": 0.5521134734153748, "learning_rate": 3.865952437661968e-07, "loss": 1.1142, "step": 5455 }, { "epoch": 0.975609756097561, "grad_norm": 0.5171558856964111, "learning_rate": 3.5967943479043867e-07, "loss": 1.1047, "step": 5460 }, { "epoch": 0.9765031716251228, "grad_norm": 0.5417544841766357, "learning_rate": 3.3373302594814637e-07, "loss": 1.0995, "step": 5465 }, { "epoch": 0.9773965871526847, "grad_norm": 0.5449820160865784, "learning_rate": 3.0875626967176165e-07, "loss": 1.1006, "step": 5470 }, { "epoch": 0.9782900026802466, "grad_norm": 0.5323817133903503, "learning_rate": 2.8474940896003887e-07, "loss": 1.139, "step": 5475 }, { "epoch": 0.9791834182078084, "grad_norm": 0.5092723369598389, "learning_rate": 2.617126773755696e-07, "loss": 1.0711, "step": 5480 }, { "epoch": 0.9800768337353704, "grad_norm": 0.5194780230522156, "learning_rate": 2.3964629904259514e-07, "loss": 1.1153, "step": 5485 }, { "epoch": 0.9809702492629322, "grad_norm": 0.5183728933334351, "learning_rate": 2.1855048864479754e-07, "loss": 1.0979, "step": 5490 }, { "epoch": 0.981863664790494, "grad_norm": 0.5222030878067017, "learning_rate": 1.984254514232009e-07, "loss": 1.0832, "step": 5495 }, { "epoch": 0.982757080318056, "grad_norm": 0.5462630987167358, "learning_rate": 1.7927138317417324e-07, "loss": 1.0618, "step": 5500 }, { "epoch": 0.9836504958456178, "grad_norm": 0.5219123959541321, "learning_rate": 1.6108847024755015e-07, "loss": 1.1152, "step": 5505 }, { "epoch": 0.9845439113731796, "grad_norm": 0.5546027421951294, "learning_rate": 1.4387688954478063e-07, "loss": 1.1397, "step": 5510 }, { "epoch": 0.9854373269007415, "grad_norm": 0.49054640531539917, "learning_rate": 1.276368085172397e-07, "loss": 1.0925, "step": 5515 }, { "epoch": 0.9863307424283034, "grad_norm": 0.5430747270584106, "learning_rate": 1.1236838516459625e-07, "loss": 1.091, "step": 5520 }, { "epoch": 0.9872241579558653, "grad_norm": 0.5430759191513062, "learning_rate": 9.807176803325879e-08, "loss": 1.1034, "step": 5525 }, { "epoch": 0.9881175734834271, "grad_norm": 0.554993212223053, "learning_rate": 8.474709621492105e-08, "loss": 1.105, "step": 5530 }, { "epoch": 0.989010989010989, "grad_norm": 0.5086091160774231, "learning_rate": 7.239449934525189e-08, "loss": 1.0637, "step": 5535 }, { "epoch": 0.9899044045385509, "grad_norm": 0.5667893290519714, "learning_rate": 6.101409760260746e-08, "loss": 1.0987, "step": 5540 }, { "epoch": 0.9907978200661127, "grad_norm": 0.5301327705383301, "learning_rate": 5.0606001706843264e-08, "loss": 1.0638, "step": 5545 }, { "epoch": 0.9916912355936747, "grad_norm": 0.5770794153213501, "learning_rate": 4.1170312918259456e-08, "loss": 1.089, "step": 5550 }, { "epoch": 0.9925846511212365, "grad_norm": 0.5558438897132874, "learning_rate": 3.2707123036646026e-08, "loss": 1.0793, "step": 5555 }, { "epoch": 0.9934780666487983, "grad_norm": 0.5180490016937256, "learning_rate": 2.5216514400305813e-08, "loss": 1.1154, "step": 5560 }, { "epoch": 0.9943714821763602, "grad_norm": 0.5610867142677307, "learning_rate": 1.869855988534397e-08, "loss": 1.0993, "step": 5565 }, { "epoch": 0.9952648977039221, "grad_norm": 0.5481888651847839, "learning_rate": 1.31533229049019e-08, "loss": 1.1059, "step": 5570 }, { "epoch": 0.996158313231484, "grad_norm": 0.5289728045463562, "learning_rate": 8.580857408546639e-09, "loss": 1.0708, "step": 5575 }, { "epoch": 0.9970517287590458, "grad_norm": 0.5380046963691711, "learning_rate": 4.9812078817934596e-09, "loss": 1.0486, "step": 5580 }, { "epoch": 0.9979451442866077, "grad_norm": 0.5321928858757019, "learning_rate": 2.3544093455951654e-09, "loss": 1.0629, "step": 5585 }, { "epoch": 0.9988385598141696, "grad_norm": 0.5315663814544678, "learning_rate": 7.004873560645387e-10, "loss": 1.0994, "step": 5590 }, { "epoch": 0.9997319753417314, "grad_norm": 0.5456522703170776, "learning_rate": 1.9458004196781787e-11, "loss": 1.0959, "step": 5595 }, { "epoch": 0.9999106584472438, "eval_loss": 1.0571272373199463, "eval_runtime": 870.817, "eval_samples_per_second": 5.123, "eval_steps_per_second": 0.641, "step": 5596 }, { "epoch": 0.9999106584472438, "step": 5596, "total_flos": 4.151601388859687e+18, "train_loss": 1.267410275018411, "train_runtime": 31869.1034, "train_samples_per_second": 1.405, "train_steps_per_second": 0.176 } ], "logging_steps": 5, "max_steps": 5596, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.151601388859687e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }