{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 3898, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 0.7940189116741517, "learning_rate": 5.128205128205128e-07, "loss": 1.2901, "step": 1 }, { "epoch": 0.0, "grad_norm": 0.8549670354423652, "learning_rate": 2.564102564102564e-06, "loss": 1.1983, "step": 5 }, { "epoch": 0.0, "grad_norm": 0.5576375802504129, "learning_rate": 5.128205128205128e-06, "loss": 1.0245, "step": 10 }, { "epoch": 0.0, "grad_norm": 0.659060642425238, "learning_rate": 7.692307692307694e-06, "loss": 1.059, "step": 15 }, { "epoch": 0.01, "grad_norm": 0.5974849589264373, "learning_rate": 1.0256410256410256e-05, "loss": 1.1426, "step": 20 }, { "epoch": 0.01, "grad_norm": 0.8598589383620435, "learning_rate": 1.282051282051282e-05, "loss": 1.0725, "step": 25 }, { "epoch": 0.01, "grad_norm": 0.7508232225220164, "learning_rate": 1.5384615384615387e-05, "loss": 1.1266, "step": 30 }, { "epoch": 0.01, "grad_norm": 0.7318286527682681, "learning_rate": 1.794871794871795e-05, "loss": 0.8731, "step": 35 }, { "epoch": 0.01, "grad_norm": 0.4400452199237334, "learning_rate": 2.0512820512820512e-05, "loss": 0.9001, "step": 40 }, { "epoch": 0.01, "grad_norm": 0.3581907713644131, "learning_rate": 2.307692307692308e-05, "loss": 0.8492, "step": 45 }, { "epoch": 0.01, "grad_norm": 0.4752977404545827, "learning_rate": 2.564102564102564e-05, "loss": 0.8143, "step": 50 }, { "epoch": 0.01, "grad_norm": 0.47540987111824407, "learning_rate": 2.8205128205128207e-05, "loss": 0.7851, "step": 55 }, { "epoch": 0.02, "grad_norm": 0.3486905588773928, "learning_rate": 3.0769230769230774e-05, "loss": 0.7796, "step": 60 }, { "epoch": 0.02, "grad_norm": 0.4379263022871022, "learning_rate": 3.3333333333333335e-05, "loss": 0.7896, "step": 65 }, { "epoch": 0.02, "grad_norm": 0.46208045268944636, "learning_rate": 3.58974358974359e-05, "loss": 0.8126, "step": 70 }, { "epoch": 0.02, "grad_norm": 0.38482216190878704, "learning_rate": 3.846153846153846e-05, "loss": 0.7794, "step": 75 }, { "epoch": 0.02, "grad_norm": 0.36177681119574645, "learning_rate": 4.1025641025641023e-05, "loss": 0.7081, "step": 80 }, { "epoch": 0.02, "grad_norm": 0.4222837002069195, "learning_rate": 4.358974358974359e-05, "loss": 0.7807, "step": 85 }, { "epoch": 0.02, "grad_norm": 0.37505732650970885, "learning_rate": 4.615384615384616e-05, "loss": 0.6835, "step": 90 }, { "epoch": 0.02, "grad_norm": 0.5376404476864999, "learning_rate": 4.871794871794872e-05, "loss": 0.6775, "step": 95 }, { "epoch": 0.03, "grad_norm": 0.33127308677570055, "learning_rate": 5.128205128205128e-05, "loss": 0.7702, "step": 100 }, { "epoch": 0.03, "grad_norm": 0.30805670386128153, "learning_rate": 5.384615384615385e-05, "loss": 0.7311, "step": 105 }, { "epoch": 0.03, "grad_norm": 0.3064832637009184, "learning_rate": 5.6410256410256414e-05, "loss": 0.7068, "step": 110 }, { "epoch": 0.03, "grad_norm": 0.36300239603665035, "learning_rate": 5.897435897435898e-05, "loss": 0.8024, "step": 115 }, { "epoch": 0.03, "grad_norm": 0.546235320721914, "learning_rate": 6.153846153846155e-05, "loss": 0.8347, "step": 120 }, { "epoch": 0.03, "grad_norm": 0.4262756501942376, "learning_rate": 6.410256410256412e-05, "loss": 0.7733, "step": 125 }, { "epoch": 0.03, "grad_norm": 0.31870027810040585, "learning_rate": 6.666666666666667e-05, "loss": 0.7072, "step": 130 }, { "epoch": 0.03, "grad_norm": 0.5080881100002653, "learning_rate": 6.923076923076924e-05, "loss": 0.7419, "step": 135 }, { "epoch": 0.04, "grad_norm": 0.45131101725332007, "learning_rate": 7.17948717948718e-05, "loss": 0.6721, "step": 140 }, { "epoch": 0.04, "grad_norm": 0.4015662904289768, "learning_rate": 7.435897435897436e-05, "loss": 0.8274, "step": 145 }, { "epoch": 0.04, "grad_norm": 0.30635979357766263, "learning_rate": 7.692307692307693e-05, "loss": 0.7501, "step": 150 }, { "epoch": 0.04, "grad_norm": 0.2563663658930907, "learning_rate": 7.948717948717948e-05, "loss": 0.6621, "step": 155 }, { "epoch": 0.04, "grad_norm": 0.3851611498390434, "learning_rate": 8.205128205128205e-05, "loss": 0.6981, "step": 160 }, { "epoch": 0.04, "grad_norm": 0.32054712548206815, "learning_rate": 8.461538461538461e-05, "loss": 0.7648, "step": 165 }, { "epoch": 0.04, "grad_norm": 0.3604034886379123, "learning_rate": 8.717948717948718e-05, "loss": 0.721, "step": 170 }, { "epoch": 0.04, "grad_norm": 0.2978365211793088, "learning_rate": 8.974358974358975e-05, "loss": 0.7665, "step": 175 }, { "epoch": 0.05, "grad_norm": 0.28010588660358143, "learning_rate": 9.230769230769232e-05, "loss": 0.6919, "step": 180 }, { "epoch": 0.05, "grad_norm": 0.3041954902881447, "learning_rate": 9.487179487179487e-05, "loss": 0.6963, "step": 185 }, { "epoch": 0.05, "grad_norm": 0.3624614794713746, "learning_rate": 9.743589743589744e-05, "loss": 0.7866, "step": 190 }, { "epoch": 0.05, "grad_norm": 0.29529594264100556, "learning_rate": 0.0001, "loss": 0.7542, "step": 195 }, { "epoch": 0.05, "grad_norm": 0.29313493488432457, "learning_rate": 0.00010256410256410256, "loss": 0.8368, "step": 200 }, { "epoch": 0.05, "grad_norm": 0.3309957222386564, "learning_rate": 0.00010512820512820514, "loss": 0.6755, "step": 205 }, { "epoch": 0.05, "grad_norm": 0.39864413284171824, "learning_rate": 0.0001076923076923077, "loss": 0.7935, "step": 210 }, { "epoch": 0.06, "grad_norm": 0.3022145052274676, "learning_rate": 0.00011025641025641027, "loss": 0.77, "step": 215 }, { "epoch": 0.06, "grad_norm": 0.3188687614270182, "learning_rate": 0.00011282051282051283, "loss": 0.764, "step": 220 }, { "epoch": 0.06, "grad_norm": 0.2771960353584736, "learning_rate": 0.00011538461538461538, "loss": 0.7025, "step": 225 }, { "epoch": 0.06, "grad_norm": 0.34181074966932906, "learning_rate": 0.00011794871794871796, "loss": 0.7206, "step": 230 }, { "epoch": 0.06, "grad_norm": 0.28461361041254846, "learning_rate": 0.00012051282051282052, "loss": 0.752, "step": 235 }, { "epoch": 0.06, "grad_norm": 0.7008172335009641, "learning_rate": 0.0001230769230769231, "loss": 0.8087, "step": 240 }, { "epoch": 0.06, "grad_norm": 0.21894206670691171, "learning_rate": 0.00012564102564102564, "loss": 0.7153, "step": 245 }, { "epoch": 0.06, "grad_norm": 0.29593725228870665, "learning_rate": 0.00012820512820512823, "loss": 0.7168, "step": 250 }, { "epoch": 0.07, "grad_norm": 0.2519334320641686, "learning_rate": 0.00013076923076923077, "loss": 0.7517, "step": 255 }, { "epoch": 0.07, "grad_norm": 0.32763786349227375, "learning_rate": 0.00013333333333333334, "loss": 0.6877, "step": 260 }, { "epoch": 0.07, "grad_norm": 0.3792543190404934, "learning_rate": 0.0001358974358974359, "loss": 0.7357, "step": 265 }, { "epoch": 0.07, "grad_norm": 0.21121164349273067, "learning_rate": 0.00013846153846153847, "loss": 0.7251, "step": 270 }, { "epoch": 0.07, "grad_norm": 0.25702919707855326, "learning_rate": 0.00014102564102564104, "loss": 0.728, "step": 275 }, { "epoch": 0.07, "grad_norm": 0.2523953056455061, "learning_rate": 0.0001435897435897436, "loss": 0.6635, "step": 280 }, { "epoch": 0.07, "grad_norm": 0.29400188903129665, "learning_rate": 0.00014615384615384615, "loss": 0.6737, "step": 285 }, { "epoch": 0.07, "grad_norm": 0.28480340987225256, "learning_rate": 0.00014871794871794872, "loss": 0.6627, "step": 290 }, { "epoch": 0.08, "grad_norm": 0.6318376196782489, "learning_rate": 0.00015128205128205128, "loss": 0.7294, "step": 295 }, { "epoch": 0.08, "grad_norm": 0.30039963893256977, "learning_rate": 0.00015384615384615385, "loss": 0.8311, "step": 300 }, { "epoch": 0.08, "grad_norm": 0.28222145819043015, "learning_rate": 0.00015641025641025642, "loss": 0.7052, "step": 305 }, { "epoch": 0.08, "grad_norm": 0.23714903546035276, "learning_rate": 0.00015897435897435896, "loss": 0.6598, "step": 310 }, { "epoch": 0.08, "grad_norm": 0.24757276162013622, "learning_rate": 0.00016153846153846155, "loss": 0.7401, "step": 315 }, { "epoch": 0.08, "grad_norm": 0.25140380741663676, "learning_rate": 0.0001641025641025641, "loss": 0.6531, "step": 320 }, { "epoch": 0.08, "grad_norm": 0.30309172865038203, "learning_rate": 0.0001666666666666667, "loss": 0.6746, "step": 325 }, { "epoch": 0.08, "grad_norm": 0.27958455276568867, "learning_rate": 0.00016923076923076923, "loss": 0.7966, "step": 330 }, { "epoch": 0.09, "grad_norm": 0.23922247511746272, "learning_rate": 0.0001717948717948718, "loss": 0.8222, "step": 335 }, { "epoch": 0.09, "grad_norm": 0.2525786116177449, "learning_rate": 0.00017435897435897436, "loss": 0.7323, "step": 340 }, { "epoch": 0.09, "grad_norm": 0.2657739010257245, "learning_rate": 0.00017692307692307693, "loss": 0.7402, "step": 345 }, { "epoch": 0.09, "grad_norm": 0.28236890518445834, "learning_rate": 0.0001794871794871795, "loss": 0.8582, "step": 350 }, { "epoch": 0.09, "grad_norm": 0.33868871224317904, "learning_rate": 0.00018205128205128207, "loss": 0.7021, "step": 355 }, { "epoch": 0.09, "grad_norm": 0.3770994222569293, "learning_rate": 0.00018461538461538463, "loss": 0.6761, "step": 360 }, { "epoch": 0.09, "grad_norm": 0.4018353610014071, "learning_rate": 0.0001871794871794872, "loss": 0.7644, "step": 365 }, { "epoch": 0.09, "grad_norm": 0.29301791588126386, "learning_rate": 0.00018974358974358974, "loss": 0.7497, "step": 370 }, { "epoch": 0.1, "grad_norm": 0.2525840897139569, "learning_rate": 0.00019230769230769233, "loss": 0.6905, "step": 375 }, { "epoch": 0.1, "grad_norm": 0.2282690464906889, "learning_rate": 0.00019487179487179487, "loss": 0.628, "step": 380 }, { "epoch": 0.1, "grad_norm": 0.29122371361878807, "learning_rate": 0.00019743589743589744, "loss": 0.7005, "step": 385 }, { "epoch": 0.1, "grad_norm": 0.2641474899247467, "learning_rate": 0.0002, "loss": 0.8084, "step": 390 }, { "epoch": 0.1, "grad_norm": 0.21454542770751026, "learning_rate": 0.00019999899748734544, "loss": 0.7702, "step": 395 }, { "epoch": 0.1, "grad_norm": 0.22924120222036806, "learning_rate": 0.00019999598996948235, "loss": 0.6556, "step": 400 }, { "epoch": 0.1, "grad_norm": 0.32163392243239314, "learning_rate": 0.00019999097750671223, "loss": 0.731, "step": 405 }, { "epoch": 0.11, "grad_norm": 0.25259388958326623, "learning_rate": 0.00019998396019953624, "loss": 0.6915, "step": 410 }, { "epoch": 0.11, "grad_norm": 0.258567962978177, "learning_rate": 0.00019997493818865318, "loss": 0.7234, "step": 415 }, { "epoch": 0.11, "grad_norm": 0.2743302123872774, "learning_rate": 0.0001999639116549566, "loss": 0.5963, "step": 420 }, { "epoch": 0.11, "grad_norm": 0.26201706655388224, "learning_rate": 0.00019995088081953136, "loss": 0.7192, "step": 425 }, { "epoch": 0.11, "grad_norm": 0.33838871762466366, "learning_rate": 0.00019993584594364894, "loss": 0.6471, "step": 430 }, { "epoch": 0.11, "grad_norm": 0.24855520937031114, "learning_rate": 0.00019991880732876246, "loss": 0.7001, "step": 435 }, { "epoch": 0.11, "grad_norm": 0.23979280174747938, "learning_rate": 0.0001998997653165004, "loss": 0.7427, "step": 440 }, { "epoch": 0.11, "grad_norm": 0.2491200817907366, "learning_rate": 0.00019987872028866003, "loss": 0.8069, "step": 445 }, { "epoch": 0.12, "grad_norm": 0.2799825613310986, "learning_rate": 0.00019985567266719934, "loss": 0.7842, "step": 450 }, { "epoch": 0.12, "grad_norm": 0.2171357139053752, "learning_rate": 0.00019983062291422908, "loss": 0.7458, "step": 455 }, { "epoch": 0.12, "grad_norm": 0.2571958093636779, "learning_rate": 0.00019980357153200315, "loss": 0.7086, "step": 460 }, { "epoch": 0.12, "grad_norm": 0.32814189406723127, "learning_rate": 0.00019977451906290854, "loss": 0.7168, "step": 465 }, { "epoch": 0.12, "grad_norm": 0.232636717309027, "learning_rate": 0.00019974346608945466, "loss": 0.7444, "step": 470 }, { "epoch": 0.12, "grad_norm": 0.2564556626752832, "learning_rate": 0.0001997104132342614, "loss": 0.6684, "step": 475 }, { "epoch": 0.12, "grad_norm": 0.2328343453656683, "learning_rate": 0.00019967536116004698, "loss": 0.814, "step": 480 }, { "epoch": 0.12, "grad_norm": 0.2743961290431054, "learning_rate": 0.00019963831056961433, "loss": 0.6833, "step": 485 }, { "epoch": 0.13, "grad_norm": 0.26151241758992244, "learning_rate": 0.00019959926220583713, "loss": 0.7213, "step": 490 }, { "epoch": 0.13, "grad_norm": 0.23479176176270686, "learning_rate": 0.000199558216851645, "loss": 0.7367, "step": 495 }, { "epoch": 0.13, "grad_norm": 0.2620759960364289, "learning_rate": 0.00019951517533000764, "loss": 0.6078, "step": 500 }, { "epoch": 0.13, "grad_norm": 0.2558927478335512, "learning_rate": 0.00019947013850391847, "loss": 0.712, "step": 505 }, { "epoch": 0.13, "grad_norm": 0.25988357901337633, "learning_rate": 0.00019942310727637724, "loss": 0.7728, "step": 510 }, { "epoch": 0.13, "grad_norm": 0.2750603881771427, "learning_rate": 0.000199374082590372, "loss": 0.7573, "step": 515 }, { "epoch": 0.13, "grad_norm": 0.22168568020157275, "learning_rate": 0.00019932306542886009, "loss": 0.741, "step": 520 }, { "epoch": 0.13, "grad_norm": 0.2776057859080138, "learning_rate": 0.0001992700568147485, "loss": 0.7221, "step": 525 }, { "epoch": 0.14, "grad_norm": 0.309528370273256, "learning_rate": 0.00019921505781087334, "loss": 0.7362, "step": 530 }, { "epoch": 0.14, "grad_norm": 0.3228391807985929, "learning_rate": 0.00019915806951997862, "loss": 0.7958, "step": 535 }, { "epoch": 0.14, "grad_norm": 0.23972118913650053, "learning_rate": 0.00019909909308469398, "loss": 0.76, "step": 540 }, { "epoch": 0.14, "grad_norm": 0.24351868304057644, "learning_rate": 0.0001990381296875118, "loss": 0.6869, "step": 545 }, { "epoch": 0.14, "grad_norm": 0.35582198947398647, "learning_rate": 0.0001989751805507637, "loss": 0.814, "step": 550 }, { "epoch": 0.14, "grad_norm": 0.2711433452256402, "learning_rate": 0.0001989102469365958, "loss": 0.8162, "step": 555 }, { "epoch": 0.14, "grad_norm": 0.23300078896983795, "learning_rate": 0.00019884333014694345, "loss": 0.7118, "step": 560 }, { "epoch": 0.14, "grad_norm": 0.2640371108479178, "learning_rate": 0.00019877443152350527, "loss": 0.7364, "step": 565 }, { "epoch": 0.15, "grad_norm": 0.24281805118776953, "learning_rate": 0.00019870355244771607, "loss": 0.6726, "step": 570 }, { "epoch": 0.15, "grad_norm": 0.2937389542412289, "learning_rate": 0.0001986306943407193, "loss": 0.7418, "step": 575 }, { "epoch": 0.15, "grad_norm": 0.244991044922138, "learning_rate": 0.00019855585866333835, "loss": 0.7476, "step": 580 }, { "epoch": 0.15, "grad_norm": 0.2799677386092935, "learning_rate": 0.00019847904691604757, "loss": 0.751, "step": 585 }, { "epoch": 0.15, "grad_norm": 0.28203702245279344, "learning_rate": 0.00019840026063894193, "loss": 0.6952, "step": 590 }, { "epoch": 0.15, "grad_norm": 0.32320414101450645, "learning_rate": 0.0001983195014117062, "loss": 0.7425, "step": 595 }, { "epoch": 0.15, "grad_norm": 0.2991698548751533, "learning_rate": 0.00019823677085358335, "loss": 0.6539, "step": 600 }, { "epoch": 0.16, "grad_norm": 0.279534327063564, "learning_rate": 0.00019815207062334197, "loss": 0.6844, "step": 605 }, { "epoch": 0.16, "grad_norm": 0.23855364154241135, "learning_rate": 0.00019806540241924317, "loss": 0.7484, "step": 610 }, { "epoch": 0.16, "grad_norm": 0.2750191162103587, "learning_rate": 0.00019797676797900633, "loss": 0.7655, "step": 615 }, { "epoch": 0.16, "grad_norm": 0.2592922229570989, "learning_rate": 0.00019788616907977441, "loss": 0.7368, "step": 620 }, { "epoch": 0.16, "grad_norm": 0.21539765246350112, "learning_rate": 0.0001977936075380783, "loss": 0.6314, "step": 625 }, { "epoch": 0.16, "grad_norm": 0.24593391755509492, "learning_rate": 0.00019769908520980034, "loss": 0.7463, "step": 630 }, { "epoch": 0.16, "grad_norm": 0.277301291124461, "learning_rate": 0.00019760260399013708, "loss": 0.721, "step": 635 }, { "epoch": 0.16, "grad_norm": 0.2676695353061291, "learning_rate": 0.00019750416581356146, "loss": 0.7112, "step": 640 }, { "epoch": 0.17, "grad_norm": 0.24096399922364115, "learning_rate": 0.0001974037726537838, "loss": 0.6304, "step": 645 }, { "epoch": 0.17, "grad_norm": 0.2973476162334585, "learning_rate": 0.00019730142652371236, "loss": 0.6467, "step": 650 }, { "epoch": 0.17, "grad_norm": 0.2766324556390413, "learning_rate": 0.00019719712947541295, "loss": 0.6812, "step": 655 }, { "epoch": 0.17, "grad_norm": 0.18389408950636116, "learning_rate": 0.0001970908836000678, "loss": 0.6861, "step": 660 }, { "epoch": 0.17, "grad_norm": 0.2410351531016274, "learning_rate": 0.00019698269102793358, "loss": 0.6532, "step": 665 }, { "epoch": 0.17, "grad_norm": 0.30102039950416376, "learning_rate": 0.00019687255392829877, "loss": 0.6568, "step": 670 }, { "epoch": 0.17, "grad_norm": 0.23183235360102647, "learning_rate": 0.00019676047450944008, "loss": 0.5751, "step": 675 }, { "epoch": 0.17, "grad_norm": 0.2933000436636493, "learning_rate": 0.0001966464550185782, "loss": 0.7648, "step": 680 }, { "epoch": 0.18, "grad_norm": 0.277707855647548, "learning_rate": 0.00019653049774183282, "loss": 0.5829, "step": 685 }, { "epoch": 0.18, "grad_norm": 0.2734745539147106, "learning_rate": 0.0001964126050041767, "loss": 0.7626, "step": 690 }, { "epoch": 0.18, "grad_norm": 0.27801067049837663, "learning_rate": 0.000196292779169389, "loss": 0.6745, "step": 695 }, { "epoch": 0.18, "grad_norm": 0.35804549859433416, "learning_rate": 0.0001961710226400081, "loss": 0.7154, "step": 700 }, { "epoch": 0.18, "grad_norm": 0.33687341140848004, "learning_rate": 0.00019604733785728317, "loss": 0.7147, "step": 705 }, { "epoch": 0.18, "grad_norm": 0.3002177976890752, "learning_rate": 0.00019592172730112544, "loss": 0.708, "step": 710 }, { "epoch": 0.18, "grad_norm": 0.29390663775324355, "learning_rate": 0.00019579419349005837, "loss": 0.7968, "step": 715 }, { "epoch": 0.18, "grad_norm": 0.2746665644289368, "learning_rate": 0.00019566473898116713, "loss": 0.7023, "step": 720 }, { "epoch": 0.19, "grad_norm": 0.30268203726113624, "learning_rate": 0.00019553336637004735, "loss": 0.742, "step": 725 }, { "epoch": 0.19, "grad_norm": 0.2203190362696549, "learning_rate": 0.0001954000782907532, "loss": 0.7494, "step": 730 }, { "epoch": 0.19, "grad_norm": 0.27657841960508833, "learning_rate": 0.00019526487741574437, "loss": 0.7819, "step": 735 }, { "epoch": 0.19, "grad_norm": 0.30505267558261334, "learning_rate": 0.00019512776645583263, "loss": 0.6908, "step": 740 }, { "epoch": 0.19, "grad_norm": 0.275012136618905, "learning_rate": 0.0001949887481601274, "loss": 0.812, "step": 745 }, { "epoch": 0.19, "grad_norm": 0.4270284365039847, "learning_rate": 0.00019484782531598073, "loss": 0.6919, "step": 750 }, { "epoch": 0.19, "grad_norm": 0.33434183829428676, "learning_rate": 0.0001947050007489313, "loss": 0.8373, "step": 755 }, { "epoch": 0.19, "grad_norm": 0.30427892181618366, "learning_rate": 0.00019456027732264784, "loss": 0.7148, "step": 760 }, { "epoch": 0.2, "grad_norm": 0.3415786735415135, "learning_rate": 0.00019441365793887162, "loss": 0.7583, "step": 765 }, { "epoch": 0.2, "grad_norm": 0.2507569242816778, "learning_rate": 0.00019426514553735848, "loss": 0.7327, "step": 770 }, { "epoch": 0.2, "grad_norm": 0.21839556494141277, "learning_rate": 0.00019411474309581958, "loss": 0.6872, "step": 775 }, { "epoch": 0.2, "grad_norm": 0.27502259439590165, "learning_rate": 0.00019396245362986197, "loss": 0.7212, "step": 780 }, { "epoch": 0.2, "grad_norm": 0.31223346554540926, "learning_rate": 0.00019380828019292798, "loss": 0.6689, "step": 785 }, { "epoch": 0.2, "grad_norm": 0.36757959748248226, "learning_rate": 0.00019365222587623405, "loss": 0.6952, "step": 790 }, { "epoch": 0.2, "grad_norm": 0.22435035120320537, "learning_rate": 0.00019349429380870873, "loss": 0.7388, "step": 795 }, { "epoch": 0.21, "grad_norm": 0.2829985423238805, "learning_rate": 0.00019333448715692995, "loss": 0.7512, "step": 800 }, { "epoch": 0.21, "grad_norm": 0.2496925408376058, "learning_rate": 0.0001931728091250615, "loss": 0.785, "step": 805 }, { "epoch": 0.21, "grad_norm": 0.25741128000037666, "learning_rate": 0.00019300926295478884, "loss": 0.769, "step": 810 }, { "epoch": 0.21, "grad_norm": 0.27594077518066623, "learning_rate": 0.00019284385192525405, "loss": 0.692, "step": 815 }, { "epoch": 0.21, "grad_norm": 0.2853537905685808, "learning_rate": 0.0001926765793529902, "loss": 0.7143, "step": 820 }, { "epoch": 0.21, "grad_norm": 0.2651467635782436, "learning_rate": 0.00019250744859185468, "loss": 0.7936, "step": 825 }, { "epoch": 0.21, "grad_norm": 0.3346954952733162, "learning_rate": 0.00019233646303296205, "loss": 0.7081, "step": 830 }, { "epoch": 0.21, "grad_norm": 0.26867358358733323, "learning_rate": 0.000192163626104616, "loss": 0.6809, "step": 835 }, { "epoch": 0.22, "grad_norm": 0.28504787352387484, "learning_rate": 0.00019198894127224074, "loss": 0.7529, "step": 840 }, { "epoch": 0.22, "grad_norm": 0.3899181097549748, "learning_rate": 0.00019181241203831137, "loss": 0.7733, "step": 845 }, { "epoch": 0.22, "grad_norm": 0.3386788353875776, "learning_rate": 0.0001916340419422837, "loss": 0.7467, "step": 850 }, { "epoch": 0.22, "grad_norm": 0.27850551310804966, "learning_rate": 0.00019145383456052327, "loss": 0.6916, "step": 855 }, { "epoch": 0.22, "grad_norm": 0.2671031051978256, "learning_rate": 0.00019127179350623372, "loss": 0.6951, "step": 860 }, { "epoch": 0.22, "grad_norm": 0.3758548589898052, "learning_rate": 0.00019108792242938425, "loss": 0.6892, "step": 865 }, { "epoch": 0.22, "grad_norm": 0.301048604600236, "learning_rate": 0.0001909022250166365, "loss": 0.6979, "step": 870 }, { "epoch": 0.22, "grad_norm": 0.23229573279853194, "learning_rate": 0.00019071470499127058, "loss": 0.6935, "step": 875 }, { "epoch": 0.23, "grad_norm": 0.30646168773183613, "learning_rate": 0.00019052536611311046, "loss": 0.7406, "step": 880 }, { "epoch": 0.23, "grad_norm": 0.23840081090154241, "learning_rate": 0.0001903342121784486, "loss": 0.6804, "step": 885 }, { "epoch": 0.23, "grad_norm": 0.3415488557571022, "learning_rate": 0.00019014124701996973, "loss": 0.7208, "step": 890 }, { "epoch": 0.23, "grad_norm": 0.2437048041319469, "learning_rate": 0.00018994647450667413, "loss": 0.7033, "step": 895 }, { "epoch": 0.23, "grad_norm": 0.27194596378878544, "learning_rate": 0.00018974989854379996, "loss": 0.7523, "step": 900 }, { "epoch": 0.23, "grad_norm": 0.29587923128889887, "learning_rate": 0.00018955152307274507, "loss": 0.687, "step": 905 }, { "epoch": 0.23, "grad_norm": 0.2461081590732809, "learning_rate": 0.00018935135207098785, "loss": 0.6655, "step": 910 }, { "epoch": 0.23, "grad_norm": 0.2924444087440772, "learning_rate": 0.00018914938955200754, "loss": 0.6819, "step": 915 }, { "epoch": 0.24, "grad_norm": 0.2553944811817433, "learning_rate": 0.00018894563956520374, "loss": 0.7684, "step": 920 }, { "epoch": 0.24, "grad_norm": 0.28957181182369396, "learning_rate": 0.0001887401061958153, "loss": 0.7766, "step": 925 }, { "epoch": 0.24, "grad_norm": 0.2785032642860862, "learning_rate": 0.00018853279356483826, "loss": 0.7712, "step": 930 }, { "epoch": 0.24, "grad_norm": 0.28527351035777204, "learning_rate": 0.00018832370582894334, "loss": 0.6425, "step": 935 }, { "epoch": 0.24, "grad_norm": 0.3575295869608585, "learning_rate": 0.00018811284718039256, "loss": 0.6777, "step": 940 }, { "epoch": 0.24, "grad_norm": 0.27729585332023104, "learning_rate": 0.00018790022184695523, "loss": 0.6822, "step": 945 }, { "epoch": 0.24, "grad_norm": 0.35797548004437607, "learning_rate": 0.00018768583409182305, "loss": 0.7349, "step": 950 }, { "epoch": 0.24, "grad_norm": 0.23438312320664967, "learning_rate": 0.00018746968821352483, "loss": 0.7424, "step": 955 }, { "epoch": 0.25, "grad_norm": 0.30004449608881456, "learning_rate": 0.00018725178854584007, "loss": 0.7366, "step": 960 }, { "epoch": 0.25, "grad_norm": 0.25679879736213285, "learning_rate": 0.00018703213945771229, "loss": 0.7473, "step": 965 }, { "epoch": 0.25, "grad_norm": 0.28382114645109796, "learning_rate": 0.00018681074535316125, "loss": 0.8144, "step": 970 }, { "epoch": 0.25, "grad_norm": 0.24972781946965186, "learning_rate": 0.00018658761067119484, "loss": 0.6927, "step": 975 }, { "epoch": 0.25, "grad_norm": 0.30834044136784544, "learning_rate": 0.00018636273988571991, "loss": 0.6803, "step": 980 }, { "epoch": 0.25, "grad_norm": 0.2649505204628836, "learning_rate": 0.00018613613750545255, "loss": 0.6411, "step": 985 }, { "epoch": 0.25, "grad_norm": 0.27400500427151636, "learning_rate": 0.0001859078080738279, "loss": 0.7043, "step": 990 }, { "epoch": 0.26, "grad_norm": 0.2723459029072118, "learning_rate": 0.00018567775616890888, "loss": 0.7234, "step": 995 }, { "epoch": 0.26, "grad_norm": 0.2635191155504064, "learning_rate": 0.00018544598640329432, "loss": 0.7131, "step": 1000 }, { "epoch": 0.26, "grad_norm": 0.24356420411141796, "learning_rate": 0.00018521250342402672, "loss": 0.7638, "step": 1005 }, { "epoch": 0.26, "grad_norm": 0.21340372473955896, "learning_rate": 0.00018497731191249894, "loss": 0.7671, "step": 1010 }, { "epoch": 0.26, "grad_norm": 0.24554004517500114, "learning_rate": 0.00018474041658436027, "loss": 0.7679, "step": 1015 }, { "epoch": 0.26, "grad_norm": 0.2556998456537391, "learning_rate": 0.000184501822189422, "loss": 0.7323, "step": 1020 }, { "epoch": 0.26, "grad_norm": 0.25911000225159375, "learning_rate": 0.0001842615335115621, "loss": 0.7126, "step": 1025 }, { "epoch": 0.26, "grad_norm": 0.29529555659956847, "learning_rate": 0.00018401955536862948, "loss": 0.8061, "step": 1030 }, { "epoch": 0.27, "grad_norm": 0.2923465430082093, "learning_rate": 0.00018377589261234705, "loss": 0.7151, "step": 1035 }, { "epoch": 0.27, "grad_norm": 0.2186889691059259, "learning_rate": 0.0001835305501282148, "loss": 0.8274, "step": 1040 }, { "epoch": 0.27, "grad_norm": 0.2841289282189221, "learning_rate": 0.00018328353283541158, "loss": 0.724, "step": 1045 }, { "epoch": 0.27, "grad_norm": 0.2951337244107648, "learning_rate": 0.00018303484568669667, "loss": 0.6899, "step": 1050 }, { "epoch": 0.27, "grad_norm": 0.15330703509842647, "learning_rate": 0.00018278449366831035, "loss": 0.5995, "step": 1055 }, { "epoch": 0.27, "grad_norm": 0.24116956612307694, "learning_rate": 0.00018253248179987388, "loss": 0.7076, "step": 1060 }, { "epoch": 0.27, "grad_norm": 0.2871300642732821, "learning_rate": 0.00018227881513428908, "loss": 0.7362, "step": 1065 }, { "epoch": 0.27, "grad_norm": 0.2979215272527789, "learning_rate": 0.0001820234987576368, "loss": 0.7023, "step": 1070 }, { "epoch": 0.28, "grad_norm": 0.2663915798849715, "learning_rate": 0.00018176653778907492, "loss": 0.7347, "step": 1075 }, { "epoch": 0.28, "grad_norm": 0.29904674261298314, "learning_rate": 0.00018150793738073602, "loss": 0.6495, "step": 1080 }, { "epoch": 0.28, "grad_norm": 0.3165528665189004, "learning_rate": 0.00018124770271762364, "loss": 0.6942, "step": 1085 }, { "epoch": 0.28, "grad_norm": 0.3378794686368825, "learning_rate": 0.00018098583901750867, "loss": 0.7095, "step": 1090 }, { "epoch": 0.28, "grad_norm": 1.1302992690531124, "learning_rate": 0.00018072235153082455, "loss": 0.6416, "step": 1095 }, { "epoch": 0.28, "grad_norm": 0.254510249703615, "learning_rate": 0.00018045724554056214, "loss": 0.7611, "step": 1100 }, { "epoch": 0.28, "grad_norm": 0.32229289030294994, "learning_rate": 0.0001801905263621636, "loss": 0.7632, "step": 1105 }, { "epoch": 0.28, "grad_norm": 0.2671186610260709, "learning_rate": 0.0001799221993434159, "loss": 0.7229, "step": 1110 }, { "epoch": 0.29, "grad_norm": 0.21433416331211266, "learning_rate": 0.00017965226986434377, "loss": 0.6888, "step": 1115 }, { "epoch": 0.29, "grad_norm": 0.39600504574420103, "learning_rate": 0.00017938074333710157, "loss": 0.7202, "step": 1120 }, { "epoch": 0.29, "grad_norm": 0.3134397942298124, "learning_rate": 0.00017910762520586485, "loss": 0.7254, "step": 1125 }, { "epoch": 0.29, "grad_norm": 0.24898380650617621, "learning_rate": 0.00017883292094672128, "loss": 0.7194, "step": 1130 }, { "epoch": 0.29, "grad_norm": 0.32622878287643925, "learning_rate": 0.00017855663606756078, "loss": 0.6838, "step": 1135 }, { "epoch": 0.29, "grad_norm": 0.2301148369047203, "learning_rate": 0.00017827877610796514, "loss": 0.6755, "step": 1140 }, { "epoch": 0.29, "grad_norm": 0.35924771537726957, "learning_rate": 0.00017799934663909682, "loss": 0.7154, "step": 1145 }, { "epoch": 0.3, "grad_norm": 0.3190069733998683, "learning_rate": 0.00017771835326358743, "loss": 0.6468, "step": 1150 }, { "epoch": 0.3, "grad_norm": 0.26523634849673267, "learning_rate": 0.00017743580161542525, "loss": 0.656, "step": 1155 }, { "epoch": 0.3, "grad_norm": 0.2073552819004321, "learning_rate": 0.00017715169735984233, "loss": 0.662, "step": 1160 }, { "epoch": 0.3, "grad_norm": 0.29144492490212454, "learning_rate": 0.00017686604619320093, "loss": 0.7378, "step": 1165 }, { "epoch": 0.3, "grad_norm": 0.2826085444372445, "learning_rate": 0.0001765788538428792, "loss": 0.8061, "step": 1170 }, { "epoch": 0.3, "grad_norm": 0.3503084495094886, "learning_rate": 0.00017629012606715648, "loss": 0.7476, "step": 1175 }, { "epoch": 0.3, "grad_norm": 0.40190773179750555, "learning_rate": 0.00017599986865509767, "loss": 0.7002, "step": 1180 }, { "epoch": 0.3, "grad_norm": 0.3512537138147811, "learning_rate": 0.00017570808742643746, "loss": 0.7955, "step": 1185 }, { "epoch": 0.31, "grad_norm": 0.2686040145129207, "learning_rate": 0.00017541478823146327, "loss": 0.7205, "step": 1190 }, { "epoch": 0.31, "grad_norm": 0.2572903388079848, "learning_rate": 0.00017511997695089822, "loss": 0.7258, "step": 1195 }, { "epoch": 0.31, "grad_norm": 0.2746446577869629, "learning_rate": 0.00017482365949578302, "loss": 0.6587, "step": 1200 }, { "epoch": 0.31, "grad_norm": 0.2656858612857696, "learning_rate": 0.00017452584180735774, "loss": 0.7334, "step": 1205 }, { "epoch": 0.31, "grad_norm": 0.25317024575228997, "learning_rate": 0.00017422652985694237, "loss": 0.7051, "step": 1210 }, { "epoch": 0.31, "grad_norm": 0.2309813519294251, "learning_rate": 0.00017392572964581725, "loss": 0.7013, "step": 1215 }, { "epoch": 0.31, "grad_norm": 0.23745896843674344, "learning_rate": 0.00017362344720510278, "loss": 0.706, "step": 1220 }, { "epoch": 0.31, "grad_norm": 0.2965283402007185, "learning_rate": 0.00017331968859563834, "loss": 0.7749, "step": 1225 }, { "epoch": 0.32, "grad_norm": 0.26530828241172705, "learning_rate": 0.00017301445990786102, "loss": 0.7566, "step": 1230 }, { "epoch": 0.32, "grad_norm": 0.3188702889740104, "learning_rate": 0.00017270776726168317, "loss": 0.7122, "step": 1235 }, { "epoch": 0.32, "grad_norm": 0.269581496964367, "learning_rate": 0.00017239961680637, "loss": 0.7457, "step": 1240 }, { "epoch": 0.32, "grad_norm": 0.2706727342486633, "learning_rate": 0.00017209001472041617, "loss": 0.7937, "step": 1245 }, { "epoch": 0.32, "grad_norm": 0.25517714775768086, "learning_rate": 0.0001717789672114218, "loss": 0.7613, "step": 1250 }, { "epoch": 0.32, "grad_norm": 0.2636311888681917, "learning_rate": 0.00017146648051596822, "loss": 0.7079, "step": 1255 }, { "epoch": 0.32, "grad_norm": 0.2989402934427053, "learning_rate": 0.0001711525608994927, "loss": 0.6716, "step": 1260 }, { "epoch": 0.32, "grad_norm": 0.22104379693485937, "learning_rate": 0.00017083721465616306, "loss": 0.6682, "step": 1265 }, { "epoch": 0.33, "grad_norm": 0.330168955232277, "learning_rate": 0.00017052044810875126, "loss": 0.8137, "step": 1270 }, { "epoch": 0.33, "grad_norm": 0.31150819405882774, "learning_rate": 0.00017020226760850677, "loss": 0.6959, "step": 1275 }, { "epoch": 0.33, "grad_norm": 0.32633550990131655, "learning_rate": 0.00016988267953502913, "loss": 0.6908, "step": 1280 }, { "epoch": 0.33, "grad_norm": 0.2247055690731712, "learning_rate": 0.0001695616902961401, "loss": 0.6666, "step": 1285 }, { "epoch": 0.33, "grad_norm": 0.2810455775039889, "learning_rate": 0.00016923930632775516, "loss": 0.7025, "step": 1290 }, { "epoch": 0.33, "grad_norm": 0.25483406919968543, "learning_rate": 0.00016891553409375444, "loss": 0.7234, "step": 1295 }, { "epoch": 0.33, "grad_norm": 0.28464930239979624, "learning_rate": 0.00016859038008585326, "loss": 0.7341, "step": 1300 }, { "epoch": 0.33, "grad_norm": 0.32647501669318885, "learning_rate": 0.0001682638508234717, "loss": 0.6706, "step": 1305 }, { "epoch": 0.34, "grad_norm": 0.2092534288356594, "learning_rate": 0.0001679359528536041, "loss": 0.6493, "step": 1310 }, { "epoch": 0.34, "grad_norm": 0.3422837500211391, "learning_rate": 0.0001676066927506878, "loss": 0.7394, "step": 1315 }, { "epoch": 0.34, "grad_norm": 0.29359271678876697, "learning_rate": 0.00016727607711647114, "loss": 0.6866, "step": 1320 }, { "epoch": 0.34, "grad_norm": 0.24439616326346505, "learning_rate": 0.00016694411257988135, "loss": 0.6546, "step": 1325 }, { "epoch": 0.34, "grad_norm": 0.25431678595144386, "learning_rate": 0.00016661080579689132, "loss": 0.7069, "step": 1330 }, { "epoch": 0.34, "grad_norm": 0.2643304489008311, "learning_rate": 0.00016627616345038642, "loss": 0.6616, "step": 1335 }, { "epoch": 0.34, "grad_norm": 0.3025866711501466, "learning_rate": 0.0001659401922500304, "loss": 0.7082, "step": 1340 }, { "epoch": 0.35, "grad_norm": 0.2576690605267584, "learning_rate": 0.0001656028989321309, "loss": 0.6431, "step": 1345 }, { "epoch": 0.35, "grad_norm": 0.23465616583898175, "learning_rate": 0.00016526429025950424, "loss": 0.6826, "step": 1350 }, { "epoch": 0.35, "grad_norm": 0.265312873732132, "learning_rate": 0.00016492437302134008, "loss": 0.7189, "step": 1355 }, { "epoch": 0.35, "grad_norm": 0.24774713265915735, "learning_rate": 0.00016458315403306502, "loss": 0.7884, "step": 1360 }, { "epoch": 0.35, "grad_norm": 0.30882736567653557, "learning_rate": 0.00016424064013620615, "loss": 0.6989, "step": 1365 }, { "epoch": 0.35, "grad_norm": 0.29435047655097046, "learning_rate": 0.0001638968381982538, "loss": 0.6983, "step": 1370 }, { "epoch": 0.35, "grad_norm": 0.24898245263549598, "learning_rate": 0.0001635517551125238, "loss": 0.8087, "step": 1375 }, { "epoch": 0.35, "grad_norm": 0.3361644987929908, "learning_rate": 0.0001632053977980194, "loss": 0.7269, "step": 1380 }, { "epoch": 0.36, "grad_norm": 0.3313137963733492, "learning_rate": 0.0001628577731992924, "loss": 0.7212, "step": 1385 }, { "epoch": 0.36, "grad_norm": 0.40222894480073607, "learning_rate": 0.000162508888286304, "loss": 0.7047, "step": 1390 }, { "epoch": 0.36, "grad_norm": 0.28365087568805586, "learning_rate": 0.00016215875005428499, "loss": 0.7654, "step": 1395 }, { "epoch": 0.36, "grad_norm": 0.22437081577512294, "learning_rate": 0.00016180736552359553, "loss": 0.7497, "step": 1400 }, { "epoch": 0.36, "grad_norm": 0.2713094825399269, "learning_rate": 0.0001614547417395844, "loss": 0.772, "step": 1405 }, { "epoch": 0.36, "grad_norm": 0.31175625126445866, "learning_rate": 0.00016110088577244773, "loss": 0.7136, "step": 1410 }, { "epoch": 0.36, "grad_norm": 0.295775151866562, "learning_rate": 0.0001607458047170872, "loss": 0.7448, "step": 1415 }, { "epoch": 0.36, "grad_norm": 0.29743619662352133, "learning_rate": 0.00016038950569296785, "loss": 0.6254, "step": 1420 }, { "epoch": 0.37, "grad_norm": 0.23830246086659249, "learning_rate": 0.00016003199584397528, "loss": 0.7976, "step": 1425 }, { "epoch": 0.37, "grad_norm": 0.24738045684528134, "learning_rate": 0.00015967328233827249, "loss": 0.6983, "step": 1430 }, { "epoch": 0.37, "grad_norm": 0.27326756366363736, "learning_rate": 0.000159313372368156, "loss": 0.6441, "step": 1435 }, { "epoch": 0.37, "grad_norm": 0.3475531782988717, "learning_rate": 0.00015895227314991178, "loss": 0.6814, "step": 1440 }, { "epoch": 0.37, "grad_norm": 0.3148530822785684, "learning_rate": 0.0001585899919236706, "loss": 0.6948, "step": 1445 }, { "epoch": 0.37, "grad_norm": 0.25377488886498006, "learning_rate": 0.00015822653595326275, "loss": 0.7238, "step": 1450 }, { "epoch": 0.37, "grad_norm": 0.3013200425652317, "learning_rate": 0.0001578619125260724, "loss": 0.75, "step": 1455 }, { "epoch": 0.37, "grad_norm": 0.2801858745291455, "learning_rate": 0.00015749612895289152, "loss": 0.6473, "step": 1460 }, { "epoch": 0.38, "grad_norm": 0.2643097355421632, "learning_rate": 0.00015712919256777335, "loss": 0.732, "step": 1465 }, { "epoch": 0.38, "grad_norm": 0.24859906355405817, "learning_rate": 0.00015676111072788527, "loss": 0.684, "step": 1470 }, { "epoch": 0.38, "grad_norm": 0.2978599865430522, "learning_rate": 0.0001563918908133614, "loss": 0.718, "step": 1475 }, { "epoch": 0.38, "grad_norm": 0.3175233807927027, "learning_rate": 0.00015602154022715435, "loss": 0.7452, "step": 1480 }, { "epoch": 0.38, "grad_norm": 0.29411432367231416, "learning_rate": 0.00015565006639488722, "loss": 0.6951, "step": 1485 }, { "epoch": 0.38, "grad_norm": 0.24712225993420803, "learning_rate": 0.0001552774767647043, "loss": 0.6921, "step": 1490 }, { "epoch": 0.38, "grad_norm": 0.296954935837503, "learning_rate": 0.000154903778807122, "loss": 0.695, "step": 1495 }, { "epoch": 0.38, "grad_norm": 0.3090817515220463, "learning_rate": 0.0001545289800148789, "loss": 0.725, "step": 1500 }, { "epoch": 0.39, "grad_norm": 0.25680112732997773, "learning_rate": 0.00015415308790278572, "loss": 0.7326, "step": 1505 }, { "epoch": 0.39, "grad_norm": 0.2713659444254739, "learning_rate": 0.0001537761100075744, "loss": 0.7369, "step": 1510 }, { "epoch": 0.39, "grad_norm": 0.3498089133387774, "learning_rate": 0.00015339805388774714, "loss": 0.7819, "step": 1515 }, { "epoch": 0.39, "grad_norm": 0.2657852286056311, "learning_rate": 0.00015301892712342482, "loss": 0.7273, "step": 1520 }, { "epoch": 0.39, "grad_norm": 0.25285783621953456, "learning_rate": 0.00015263873731619508, "loss": 0.766, "step": 1525 }, { "epoch": 0.39, "grad_norm": 0.3072970711277045, "learning_rate": 0.00015225749208895968, "loss": 0.7195, "step": 1530 }, { "epoch": 0.39, "grad_norm": 0.3660032567021492, "learning_rate": 0.000151875199085782, "loss": 0.7699, "step": 1535 }, { "epoch": 0.4, "grad_norm": 0.2924264482609829, "learning_rate": 0.0001514918659717335, "loss": 0.7105, "step": 1540 }, { "epoch": 0.4, "grad_norm": 0.3309598652416674, "learning_rate": 0.00015110750043274008, "loss": 0.732, "step": 1545 }, { "epoch": 0.4, "grad_norm": 0.25120067042745076, "learning_rate": 0.00015072211017542813, "loss": 0.7447, "step": 1550 }, { "epoch": 0.4, "grad_norm": 0.29548371062233497, "learning_rate": 0.0001503357029269698, "loss": 0.6245, "step": 1555 }, { "epoch": 0.4, "grad_norm": 0.2882495017189674, "learning_rate": 0.00014994828643492827, "loss": 0.7354, "step": 1560 }, { "epoch": 0.4, "grad_norm": 0.2926891269519082, "learning_rate": 0.00014955986846710222, "loss": 0.7234, "step": 1565 }, { "epoch": 0.4, "grad_norm": 0.3208205900575705, "learning_rate": 0.00014917045681137026, "loss": 0.6859, "step": 1570 }, { "epoch": 0.4, "grad_norm": 0.28730436475223414, "learning_rate": 0.00014878005927553456, "loss": 0.655, "step": 1575 }, { "epoch": 0.41, "grad_norm": 0.31375609878578203, "learning_rate": 0.0001483886836871646, "loss": 0.7068, "step": 1580 }, { "epoch": 0.41, "grad_norm": 0.28634685391130577, "learning_rate": 0.00014799633789343994, "loss": 0.7873, "step": 1585 }, { "epoch": 0.41, "grad_norm": 0.29219569906271287, "learning_rate": 0.00014760302976099304, "loss": 0.6414, "step": 1590 }, { "epoch": 0.41, "grad_norm": 0.2809961899361123, "learning_rate": 0.00014720876717575155, "loss": 0.689, "step": 1595 }, { "epoch": 0.41, "grad_norm": 0.2966023851758411, "learning_rate": 0.00014681355804278001, "loss": 0.6939, "step": 1600 }, { "epoch": 0.41, "grad_norm": 0.22135705676789802, "learning_rate": 0.00014641741028612162, "loss": 0.6388, "step": 1605 }, { "epoch": 0.41, "grad_norm": 0.2017382590189762, "learning_rate": 0.00014602033184863913, "loss": 0.7172, "step": 1610 }, { "epoch": 0.41, "grad_norm": 0.29520895754468235, "learning_rate": 0.00014562233069185572, "loss": 0.8331, "step": 1615 }, { "epoch": 0.42, "grad_norm": 0.25912176225264727, "learning_rate": 0.00014522341479579533, "loss": 0.7583, "step": 1620 }, { "epoch": 0.42, "grad_norm": 0.27070806912751083, "learning_rate": 0.0001448235921588226, "loss": 0.7706, "step": 1625 }, { "epoch": 0.42, "grad_norm": 0.22470926163816077, "learning_rate": 0.00014442287079748263, "loss": 0.701, "step": 1630 }, { "epoch": 0.42, "grad_norm": 0.26012437368269403, "learning_rate": 0.00014402125874634012, "loss": 0.6856, "step": 1635 }, { "epoch": 0.42, "grad_norm": 0.27655512567748813, "learning_rate": 0.00014361876405781832, "loss": 0.7092, "step": 1640 }, { "epoch": 0.42, "grad_norm": 0.22547503156029328, "learning_rate": 0.00014321539480203764, "loss": 0.7723, "step": 1645 }, { "epoch": 0.42, "grad_norm": 0.35807881396099706, "learning_rate": 0.00014281115906665374, "loss": 0.699, "step": 1650 }, { "epoch": 0.42, "grad_norm": 0.25312726731980933, "learning_rate": 0.00014240606495669538, "loss": 0.694, "step": 1655 }, { "epoch": 0.43, "grad_norm": 0.3021591327476025, "learning_rate": 0.00014200012059440207, "loss": 0.6978, "step": 1660 }, { "epoch": 0.43, "grad_norm": 0.3776238352797525, "learning_rate": 0.00014159333411906095, "loss": 0.7518, "step": 1665 }, { "epoch": 0.43, "grad_norm": 0.24540414704743882, "learning_rate": 0.00014118571368684383, "loss": 0.7891, "step": 1670 }, { "epoch": 0.43, "grad_norm": 0.3138641993671755, "learning_rate": 0.00014077726747064353, "loss": 0.6213, "step": 1675 }, { "epoch": 0.43, "grad_norm": 0.24104497594679253, "learning_rate": 0.00014036800365991008, "loss": 0.6773, "step": 1680 }, { "epoch": 0.43, "grad_norm": 0.2712106009342219, "learning_rate": 0.00013995793046048643, "loss": 0.6977, "step": 1685 }, { "epoch": 0.43, "grad_norm": 0.2847114290804009, "learning_rate": 0.00013954705609444404, "loss": 0.6903, "step": 1690 }, { "epoch": 0.43, "grad_norm": 0.31132954968304133, "learning_rate": 0.0001391353887999179, "loss": 0.7666, "step": 1695 }, { "epoch": 0.44, "grad_norm": 0.2648924104898787, "learning_rate": 0.00013872293683094152, "loss": 0.6656, "step": 1700 }, { "epoch": 0.44, "grad_norm": 0.2628528269004981, "learning_rate": 0.00013830970845728126, "loss": 0.7704, "step": 1705 }, { "epoch": 0.44, "grad_norm": 0.21693389371822336, "learning_rate": 0.00013789571196427055, "loss": 0.6493, "step": 1710 }, { "epoch": 0.44, "grad_norm": 0.2447657421175684, "learning_rate": 0.00013748095565264384, "loss": 0.6459, "step": 1715 }, { "epoch": 0.44, "grad_norm": 0.3180367525541911, "learning_rate": 0.00013706544783837022, "loss": 0.7049, "step": 1720 }, { "epoch": 0.44, "grad_norm": 0.2972710048973474, "learning_rate": 0.00013664919685248649, "loss": 0.7238, "step": 1725 }, { "epoch": 0.44, "grad_norm": 0.29131857236762787, "learning_rate": 0.00013623221104093025, "loss": 0.8293, "step": 1730 }, { "epoch": 0.45, "grad_norm": 0.27510973199735633, "learning_rate": 0.0001358144987643726, "loss": 0.7285, "step": 1735 }, { "epoch": 0.45, "grad_norm": 0.23692910598713893, "learning_rate": 0.00013539606839805036, "loss": 0.7199, "step": 1740 }, { "epoch": 0.45, "grad_norm": 0.29598035037077913, "learning_rate": 0.00013497692833159828, "loss": 0.7042, "step": 1745 }, { "epoch": 0.45, "grad_norm": 0.25149104979862424, "learning_rate": 0.00013455708696888085, "loss": 0.6942, "step": 1750 }, { "epoch": 0.45, "grad_norm": 0.25496421670247854, "learning_rate": 0.00013413655272782356, "loss": 0.7256, "step": 1755 }, { "epoch": 0.45, "grad_norm": 0.2554748374618532, "learning_rate": 0.00013371533404024438, "loss": 0.7607, "step": 1760 }, { "epoch": 0.45, "grad_norm": 0.2627306476634245, "learning_rate": 0.00013329343935168466, "loss": 0.7101, "step": 1765 }, { "epoch": 0.45, "grad_norm": 0.24657317727636707, "learning_rate": 0.00013287087712123962, "loss": 0.7726, "step": 1770 }, { "epoch": 0.46, "grad_norm": 0.2559430395751914, "learning_rate": 0.000132447655821389, "loss": 0.6601, "step": 1775 }, { "epoch": 0.46, "grad_norm": 0.3049767818100991, "learning_rate": 0.00013202378393782692, "loss": 0.7961, "step": 1780 }, { "epoch": 0.46, "grad_norm": 0.28360752211200474, "learning_rate": 0.00013159926996929192, "loss": 0.6639, "step": 1785 }, { "epoch": 0.46, "grad_norm": 0.26804633258042826, "learning_rate": 0.00013117412242739655, "loss": 0.6503, "step": 1790 }, { "epoch": 0.46, "grad_norm": 0.2411104477481089, "learning_rate": 0.00013074834983645657, "loss": 0.7532, "step": 1795 }, { "epoch": 0.46, "grad_norm": 0.2900355787539025, "learning_rate": 0.00013032196073332027, "loss": 0.7479, "step": 1800 }, { "epoch": 0.46, "grad_norm": 0.2568819906683512, "learning_rate": 0.000129894963667197, "loss": 0.7433, "step": 1805 }, { "epoch": 0.46, "grad_norm": 0.22709405983327266, "learning_rate": 0.00012946736719948607, "loss": 0.7163, "step": 1810 }, { "epoch": 0.47, "grad_norm": 0.28704276569845133, "learning_rate": 0.00012903917990360485, "loss": 0.6965, "step": 1815 }, { "epoch": 0.47, "grad_norm": 0.2573742688385154, "learning_rate": 0.000128610410364817, "loss": 0.6919, "step": 1820 }, { "epoch": 0.47, "grad_norm": 0.3570766420018621, "learning_rate": 0.00012818106718006023, "loss": 0.6073, "step": 1825 }, { "epoch": 0.47, "grad_norm": 0.2471571809588589, "learning_rate": 0.00012775115895777417, "loss": 0.753, "step": 1830 }, { "epoch": 0.47, "grad_norm": 0.25821147959621743, "learning_rate": 0.0001273206943177274, "loss": 0.6319, "step": 1835 }, { "epoch": 0.47, "grad_norm": 0.30013480040772167, "learning_rate": 0.00012688968189084493, "loss": 0.7263, "step": 1840 }, { "epoch": 0.47, "grad_norm": 0.28654821268777825, "learning_rate": 0.000126458130319035, "loss": 0.7259, "step": 1845 }, { "epoch": 0.47, "grad_norm": 0.29336529390659893, "learning_rate": 0.00012602604825501587, "loss": 0.7032, "step": 1850 }, { "epoch": 0.48, "grad_norm": 0.20273978059145287, "learning_rate": 0.00012559344436214223, "loss": 0.6322, "step": 1855 }, { "epoch": 0.48, "grad_norm": 0.2483147925181707, "learning_rate": 0.00012516032731423165, "loss": 0.7185, "step": 1860 }, { "epoch": 0.48, "grad_norm": 0.28968579800640504, "learning_rate": 0.00012472670579539055, "loss": 0.7028, "step": 1865 }, { "epoch": 0.48, "grad_norm": 0.24871377665254096, "learning_rate": 0.00012429258849984014, "loss": 0.7093, "step": 1870 }, { "epoch": 0.48, "grad_norm": 0.23758441118744317, "learning_rate": 0.00012385798413174206, "loss": 0.7267, "step": 1875 }, { "epoch": 0.48, "grad_norm": 0.29112304576955467, "learning_rate": 0.00012342290140502388, "loss": 0.7714, "step": 1880 }, { "epoch": 0.48, "grad_norm": 0.246124461049581, "learning_rate": 0.00012298734904320438, "loss": 0.6727, "step": 1885 }, { "epoch": 0.48, "grad_norm": 0.25856322825399286, "learning_rate": 0.00012255133577921868, "loss": 0.7769, "step": 1890 }, { "epoch": 0.49, "grad_norm": 0.22410369058426746, "learning_rate": 0.00012211487035524305, "loss": 0.6866, "step": 1895 }, { "epoch": 0.49, "grad_norm": 0.28079722454610034, "learning_rate": 0.0001216779615225197, "loss": 0.6963, "step": 1900 }, { "epoch": 0.49, "grad_norm": 0.30696044399665856, "learning_rate": 0.00012124061804118137, "loss": 0.7051, "step": 1905 }, { "epoch": 0.49, "grad_norm": 0.247967331116475, "learning_rate": 0.00012080284868007541, "loss": 0.6944, "step": 1910 }, { "epoch": 0.49, "grad_norm": 0.21128171750447838, "learning_rate": 0.00012036466221658847, "loss": 0.7199, "step": 1915 }, { "epoch": 0.49, "grad_norm": 0.31901890228430696, "learning_rate": 0.0001199260674364699, "loss": 0.7786, "step": 1920 }, { "epoch": 0.49, "grad_norm": 0.2755016732959131, "learning_rate": 0.00011948707313365614, "loss": 0.8069, "step": 1925 }, { "epoch": 0.5, "grad_norm": 0.28668018656021443, "learning_rate": 0.00011904768811009405, "loss": 0.7157, "step": 1930 }, { "epoch": 0.5, "grad_norm": 0.3328138328405363, "learning_rate": 0.00011860792117556454, "loss": 0.7099, "step": 1935 }, { "epoch": 0.5, "grad_norm": 0.21788073287798893, "learning_rate": 0.00011816778114750593, "loss": 0.7644, "step": 1940 }, { "epoch": 0.5, "grad_norm": 0.23062823471771338, "learning_rate": 0.00011772727685083724, "loss": 0.7716, "step": 1945 }, { "epoch": 0.5, "grad_norm": 0.26093096066749466, "learning_rate": 0.00011728641711778103, "loss": 0.7229, "step": 1950 }, { "epoch": 0.5, "grad_norm": 0.23801319602695262, "learning_rate": 0.00011684521078768658, "loss": 0.6845, "step": 1955 }, { "epoch": 0.5, "grad_norm": 0.18154354473533865, "learning_rate": 0.00011640366670685248, "loss": 0.6315, "step": 1960 }, { "epoch": 0.5, "grad_norm": 0.29572089884772323, "learning_rate": 0.00011596179372834924, "loss": 0.8437, "step": 1965 }, { "epoch": 0.51, "grad_norm": 0.27149388513396633, "learning_rate": 0.00011551960071184195, "loss": 0.8232, "step": 1970 }, { "epoch": 0.51, "grad_norm": 0.24540172557967463, "learning_rate": 0.00011507709652341256, "loss": 0.7132, "step": 1975 }, { "epoch": 0.51, "grad_norm": 0.26964683600534267, "learning_rate": 0.00011463429003538196, "loss": 0.7275, "step": 1980 }, { "epoch": 0.51, "grad_norm": 0.287051761373442, "learning_rate": 0.00011419119012613233, "loss": 0.7184, "step": 1985 }, { "epoch": 0.51, "grad_norm": 0.2663459541020728, "learning_rate": 0.000113747805679929, "loss": 0.7964, "step": 1990 }, { "epoch": 0.51, "grad_norm": 0.28638851869565946, "learning_rate": 0.00011330414558674234, "loss": 0.7129, "step": 1995 }, { "epoch": 0.51, "grad_norm": 0.3786862886137478, "learning_rate": 0.00011286021874206952, "loss": 0.6706, "step": 2000 }, { "epoch": 0.51, "grad_norm": 0.34276302477862775, "learning_rate": 0.00011241603404675609, "loss": 0.7856, "step": 2005 }, { "epoch": 0.52, "grad_norm": 0.3366461922626376, "learning_rate": 0.00011197160040681762, "loss": 0.8178, "step": 2010 }, { "epoch": 0.52, "grad_norm": 0.25846779036672146, "learning_rate": 0.00011152692673326111, "loss": 0.7387, "step": 2015 }, { "epoch": 0.52, "grad_norm": 0.2364492132114722, "learning_rate": 0.0001110820219419062, "loss": 0.7434, "step": 2020 }, { "epoch": 0.52, "grad_norm": 0.2699184257070918, "learning_rate": 0.00011063689495320661, "loss": 0.7102, "step": 2025 }, { "epoch": 0.52, "grad_norm": 0.2506412498240298, "learning_rate": 0.0001101915546920711, "loss": 0.7451, "step": 2030 }, { "epoch": 0.52, "grad_norm": 0.4833851120529976, "learning_rate": 0.00010974601008768461, "loss": 0.6558, "step": 2035 }, { "epoch": 0.52, "grad_norm": 0.25503131460937667, "learning_rate": 0.00010930027007332923, "loss": 0.6642, "step": 2040 }, { "epoch": 0.52, "grad_norm": 0.24985640167049433, "learning_rate": 0.00010885434358620507, "loss": 0.737, "step": 2045 }, { "epoch": 0.53, "grad_norm": 0.28451521701088694, "learning_rate": 0.00010840823956725103, "loss": 0.7406, "step": 2050 }, { "epoch": 0.53, "grad_norm": 0.29244380187254954, "learning_rate": 0.00010796196696096564, "loss": 0.7047, "step": 2055 }, { "epoch": 0.53, "grad_norm": 0.2236043726332343, "learning_rate": 0.00010751553471522757, "loss": 0.7312, "step": 2060 }, { "epoch": 0.53, "grad_norm": 0.27232955976439666, "learning_rate": 0.00010706895178111634, "loss": 0.7284, "step": 2065 }, { "epoch": 0.53, "grad_norm": 0.30610542432506455, "learning_rate": 0.00010662222711273279, "loss": 0.6335, "step": 2070 }, { "epoch": 0.53, "grad_norm": 0.23424590700585504, "learning_rate": 0.00010617536966701962, "loss": 0.7269, "step": 2075 }, { "epoch": 0.53, "grad_norm": 0.25989610387137024, "learning_rate": 0.00010572838840358168, "loss": 0.6304, "step": 2080 }, { "epoch": 0.53, "grad_norm": 0.2830081164838504, "learning_rate": 0.00010528129228450645, "loss": 0.6597, "step": 2085 }, { "epoch": 0.54, "grad_norm": 0.2603374987935435, "learning_rate": 0.00010483409027418425, "loss": 0.7489, "step": 2090 }, { "epoch": 0.54, "grad_norm": 0.24810278191222906, "learning_rate": 0.00010438679133912858, "loss": 0.6502, "step": 2095 }, { "epoch": 0.54, "grad_norm": 0.2528579675694808, "learning_rate": 0.00010393940444779635, "loss": 0.6969, "step": 2100 }, { "epoch": 0.54, "grad_norm": 0.28413657506453693, "learning_rate": 0.00010349193857040787, "loss": 0.7726, "step": 2105 }, { "epoch": 0.54, "grad_norm": 0.3144890009802897, "learning_rate": 0.00010304440267876727, "loss": 0.7856, "step": 2110 }, { "epoch": 0.54, "grad_norm": 0.2397459238431048, "learning_rate": 0.00010259680574608248, "loss": 0.764, "step": 2115 }, { "epoch": 0.54, "grad_norm": 0.2868615633571584, "learning_rate": 0.00010214915674678523, "loss": 0.7677, "step": 2120 }, { "epoch": 0.55, "grad_norm": 0.2497448183867612, "learning_rate": 0.00010170146465635126, "loss": 0.7783, "step": 2125 }, { "epoch": 0.55, "grad_norm": 0.253908792089182, "learning_rate": 0.00010125373845112034, "loss": 0.7461, "step": 2130 }, { "epoch": 0.55, "grad_norm": 0.23687933586461904, "learning_rate": 0.00010080598710811615, "loss": 0.69, "step": 2135 }, { "epoch": 0.55, "grad_norm": 0.3734210753745544, "learning_rate": 0.00010035821960486643, "loss": 0.77, "step": 2140 }, { "epoch": 0.55, "grad_norm": 0.2788998422640762, "learning_rate": 9.9910444919223e-05, "loss": 0.7058, "step": 2145 }, { "epoch": 0.55, "grad_norm": 0.3178188169556896, "learning_rate": 9.946267202918157e-05, "loss": 0.6965, "step": 2150 }, { "epoch": 0.55, "grad_norm": 0.26945241065973596, "learning_rate": 9.901490991270201e-05, "loss": 0.702, "step": 2155 }, { "epoch": 0.55, "grad_norm": 0.5207203316817794, "learning_rate": 9.856716754752796e-05, "loss": 0.6478, "step": 2160 }, { "epoch": 0.56, "grad_norm": 0.29256712809053254, "learning_rate": 9.811945391100724e-05, "loss": 0.7372, "step": 2165 }, { "epoch": 0.56, "grad_norm": 0.25601912122707865, "learning_rate": 9.767177797991155e-05, "loss": 0.7444, "step": 2170 }, { "epoch": 0.56, "grad_norm": 0.23247510741792463, "learning_rate": 9.722414873025664e-05, "loss": 0.6836, "step": 2175 }, { "epoch": 0.56, "grad_norm": 0.2460904591621441, "learning_rate": 9.677657513712221e-05, "loss": 0.7268, "step": 2180 }, { "epoch": 0.56, "grad_norm": 0.26208399329605536, "learning_rate": 9.632906617447212e-05, "loss": 0.6924, "step": 2185 }, { "epoch": 0.56, "grad_norm": 0.22915215031643407, "learning_rate": 9.588163081497427e-05, "loss": 0.7864, "step": 2190 }, { "epoch": 0.56, "grad_norm": 0.3119119948293986, "learning_rate": 9.543427802982095e-05, "loss": 0.768, "step": 2195 }, { "epoch": 0.56, "grad_norm": 0.2973539056568546, "learning_rate": 9.498701678854865e-05, "loss": 0.6836, "step": 2200 }, { "epoch": 0.57, "grad_norm": 0.24915238391515174, "learning_rate": 9.453985605885851e-05, "loss": 0.7307, "step": 2205 }, { "epoch": 0.57, "grad_norm": 0.29466985701358944, "learning_rate": 9.409280480643628e-05, "loss": 0.7751, "step": 2210 }, { "epoch": 0.57, "grad_norm": 0.31502684795729613, "learning_rate": 9.364587199477276e-05, "loss": 0.7183, "step": 2215 }, { "epoch": 0.57, "grad_norm": 0.23712422669700867, "learning_rate": 9.319906658498389e-05, "loss": 0.7522, "step": 2220 }, { "epoch": 0.57, "grad_norm": 0.2570700308864798, "learning_rate": 9.275239753563128e-05, "loss": 0.7171, "step": 2225 }, { "epoch": 0.57, "grad_norm": 0.23813307578157708, "learning_rate": 9.230587380254237e-05, "loss": 0.689, "step": 2230 }, { "epoch": 0.57, "grad_norm": 0.2656059785729245, "learning_rate": 9.185950433863107e-05, "loss": 0.6694, "step": 2235 }, { "epoch": 0.57, "grad_norm": 0.1857069050131154, "learning_rate": 9.141329809371803e-05, "loss": 0.7626, "step": 2240 }, { "epoch": 0.58, "grad_norm": 0.25034208426960974, "learning_rate": 9.096726401435146e-05, "loss": 0.7116, "step": 2245 }, { "epoch": 0.58, "grad_norm": 0.22067489301839913, "learning_rate": 9.052141104362748e-05, "loss": 0.6499, "step": 2250 }, { "epoch": 0.58, "grad_norm": 0.23199516425281616, "learning_rate": 9.007574812101107e-05, "loss": 0.7501, "step": 2255 }, { "epoch": 0.58, "grad_norm": 0.25421299632271965, "learning_rate": 8.963028418215653e-05, "loss": 0.8147, "step": 2260 }, { "epoch": 0.58, "grad_norm": 0.260651537286194, "learning_rate": 8.918502815872865e-05, "loss": 0.7485, "step": 2265 }, { "epoch": 0.58, "grad_norm": 0.2655797468541603, "learning_rate": 8.873998897822336e-05, "loss": 0.7505, "step": 2270 }, { "epoch": 0.58, "grad_norm": 0.2917164025171058, "learning_rate": 8.829517556378888e-05, "loss": 0.6755, "step": 2275 }, { "epoch": 0.58, "grad_norm": 0.3350035349023889, "learning_rate": 8.785059683404672e-05, "loss": 0.7015, "step": 2280 }, { "epoch": 0.59, "grad_norm": 0.3253925397932177, "learning_rate": 8.740626170291297e-05, "loss": 0.7551, "step": 2285 }, { "epoch": 0.59, "grad_norm": 0.2899593792802913, "learning_rate": 8.696217907941941e-05, "loss": 0.7185, "step": 2290 }, { "epoch": 0.59, "grad_norm": 0.274238966245006, "learning_rate": 8.65183578675351e-05, "loss": 0.7031, "step": 2295 }, { "epoch": 0.59, "grad_norm": 0.27886346870120726, "learning_rate": 8.607480696598762e-05, "loss": 0.61, "step": 2300 }, { "epoch": 0.59, "grad_norm": 0.3552726057492985, "learning_rate": 8.563153526808484e-05, "loss": 0.7706, "step": 2305 }, { "epoch": 0.59, "grad_norm": 0.26275345892575847, "learning_rate": 8.518855166153644e-05, "loss": 0.6925, "step": 2310 }, { "epoch": 0.59, "grad_norm": 0.3537795284485823, "learning_rate": 8.474586502827591e-05, "loss": 0.7005, "step": 2315 }, { "epoch": 0.6, "grad_norm": 0.2113496203404715, "learning_rate": 8.43034842442822e-05, "loss": 0.6709, "step": 2320 }, { "epoch": 0.6, "grad_norm": 0.46932602124140005, "learning_rate": 8.386141817940213e-05, "loss": 0.7019, "step": 2325 }, { "epoch": 0.6, "grad_norm": 0.2647970131377573, "learning_rate": 8.341967569717202e-05, "loss": 0.6522, "step": 2330 }, { "epoch": 0.6, "grad_norm": 0.2495289774517795, "learning_rate": 8.297826565464057e-05, "loss": 0.6333, "step": 2335 }, { "epoch": 0.6, "grad_norm": 0.317401195281625, "learning_rate": 8.253719690219079e-05, "loss": 0.7297, "step": 2340 }, { "epoch": 0.6, "grad_norm": 0.2722505836316711, "learning_rate": 8.20964782833628e-05, "loss": 0.6206, "step": 2345 }, { "epoch": 0.6, "grad_norm": 0.3347413066633305, "learning_rate": 8.165611863467644e-05, "loss": 0.7258, "step": 2350 }, { "epoch": 0.6, "grad_norm": 0.23149723477405734, "learning_rate": 8.12161267854542e-05, "loss": 0.6872, "step": 2355 }, { "epoch": 0.61, "grad_norm": 0.21852198583337765, "learning_rate": 8.077651155764387e-05, "loss": 0.6529, "step": 2360 }, { "epoch": 0.61, "grad_norm": 0.23663927702457488, "learning_rate": 8.033728176564216e-05, "loss": 0.7468, "step": 2365 }, { "epoch": 0.61, "grad_norm": 0.28901081689808944, "learning_rate": 7.98984462161175e-05, "loss": 0.7436, "step": 2370 }, { "epoch": 0.61, "grad_norm": 0.25802965477339335, "learning_rate": 7.946001370783375e-05, "loss": 0.7899, "step": 2375 }, { "epoch": 0.61, "grad_norm": 0.22951991679268438, "learning_rate": 7.902199303147363e-05, "loss": 0.6861, "step": 2380 }, { "epoch": 0.61, "grad_norm": 0.3057181147340439, "learning_rate": 7.858439296946263e-05, "loss": 0.7769, "step": 2385 }, { "epoch": 0.61, "grad_norm": 0.3159742724328365, "learning_rate": 7.814722229579264e-05, "loss": 0.6756, "step": 2390 }, { "epoch": 0.61, "grad_norm": 0.27245093189237785, "learning_rate": 7.771048977584641e-05, "loss": 0.6128, "step": 2395 }, { "epoch": 0.62, "grad_norm": 0.324925192823721, "learning_rate": 7.727420416622144e-05, "loss": 0.8306, "step": 2400 }, { "epoch": 0.62, "grad_norm": 0.2999192623424217, "learning_rate": 7.683837421455466e-05, "loss": 0.6782, "step": 2405 }, { "epoch": 0.62, "grad_norm": 0.25980740031389876, "learning_rate": 7.640300865934687e-05, "loss": 0.6913, "step": 2410 }, { "epoch": 0.62, "grad_norm": 0.25802354488510615, "learning_rate": 7.596811622978765e-05, "loss": 0.7178, "step": 2415 }, { "epoch": 0.62, "grad_norm": 0.26975633006496086, "learning_rate": 7.553370564558032e-05, "loss": 0.7667, "step": 2420 }, { "epoch": 0.62, "grad_norm": 0.3248726081338585, "learning_rate": 7.509978561676699e-05, "loss": 0.7559, "step": 2425 }, { "epoch": 0.62, "grad_norm": 0.27315448139602455, "learning_rate": 7.46663648435541e-05, "loss": 0.6903, "step": 2430 }, { "epoch": 0.62, "grad_norm": 0.2950802099290335, "learning_rate": 7.423345201613778e-05, "loss": 0.6731, "step": 2435 }, { "epoch": 0.63, "grad_norm": 0.35343151310449955, "learning_rate": 7.380105581452987e-05, "loss": 0.735, "step": 2440 }, { "epoch": 0.63, "grad_norm": 0.2766967860281418, "learning_rate": 7.336918490838356e-05, "loss": 0.7779, "step": 2445 }, { "epoch": 0.63, "grad_norm": 0.255190126277391, "learning_rate": 7.293784795681994e-05, "loss": 0.7213, "step": 2450 }, { "epoch": 0.63, "grad_norm": 0.23722517953499153, "learning_rate": 7.250705360825392e-05, "loss": 0.6734, "step": 2455 }, { "epoch": 0.63, "grad_norm": 0.30683380454221565, "learning_rate": 7.207681050022132e-05, "loss": 0.7439, "step": 2460 }, { "epoch": 0.63, "grad_norm": 0.3072464317121022, "learning_rate": 7.16471272592053e-05, "loss": 0.7804, "step": 2465 }, { "epoch": 0.63, "grad_norm": 0.2565247543734341, "learning_rate": 7.121801250046363e-05, "loss": 0.7678, "step": 2470 }, { "epoch": 0.63, "grad_norm": 0.2106442564735487, "learning_rate": 7.078947482785576e-05, "loss": 0.7063, "step": 2475 }, { "epoch": 0.64, "grad_norm": 0.2476449919758559, "learning_rate": 7.036152283367056e-05, "loss": 0.6918, "step": 2480 }, { "epoch": 0.64, "grad_norm": 0.247411584313298, "learning_rate": 6.993416509845376e-05, "loss": 0.7182, "step": 2485 }, { "epoch": 0.64, "grad_norm": 0.280770588683194, "learning_rate": 6.950741019083617e-05, "loss": 0.7192, "step": 2490 }, { "epoch": 0.64, "grad_norm": 0.28453079995263136, "learning_rate": 6.908126666736165e-05, "loss": 0.6912, "step": 2495 }, { "epoch": 0.64, "grad_norm": 0.34116367134660075, "learning_rate": 6.865574307231575e-05, "loss": 0.7064, "step": 2500 }, { "epoch": 0.64, "grad_norm": 0.20407453115123925, "learning_rate": 6.823084793755417e-05, "loss": 0.6835, "step": 2505 }, { "epoch": 0.64, "grad_norm": 0.34227090827399664, "learning_rate": 6.780658978233199e-05, "loss": 0.7364, "step": 2510 }, { "epoch": 0.65, "grad_norm": 0.20382515840687232, "learning_rate": 6.73829771131325e-05, "loss": 0.7278, "step": 2515 }, { "epoch": 0.65, "grad_norm": 0.32066035575970564, "learning_rate": 6.696001842349702e-05, "loss": 0.6735, "step": 2520 }, { "epoch": 0.65, "grad_norm": 0.30087509915880645, "learning_rate": 6.653772219385427e-05, "loss": 0.7563, "step": 2525 }, { "epoch": 0.65, "grad_norm": 0.2609826902639363, "learning_rate": 6.611609689135056e-05, "loss": 0.7865, "step": 2530 }, { "epoch": 0.65, "grad_norm": 0.2541688409079705, "learning_rate": 6.569515096967991e-05, "loss": 0.645, "step": 2535 }, { "epoch": 0.65, "grad_norm": 0.18975728074739295, "learning_rate": 6.527489286891459e-05, "loss": 0.5808, "step": 2540 }, { "epoch": 0.65, "grad_norm": 0.2395876243767242, "learning_rate": 6.485533101533583e-05, "loss": 0.6973, "step": 2545 }, { "epoch": 0.65, "grad_norm": 0.3063762562351843, "learning_rate": 6.443647382126509e-05, "loss": 0.7221, "step": 2550 }, { "epoch": 0.66, "grad_norm": 0.2674321599248453, "learning_rate": 6.401832968489501e-05, "loss": 0.6818, "step": 2555 }, { "epoch": 0.66, "grad_norm": 0.2231267411467715, "learning_rate": 6.360090699012145e-05, "loss": 0.7404, "step": 2560 }, { "epoch": 0.66, "grad_norm": 0.27582828393948144, "learning_rate": 6.3184214106375e-05, "loss": 0.6741, "step": 2565 }, { "epoch": 0.66, "grad_norm": 0.27355370388789374, "learning_rate": 6.27682593884535e-05, "loss": 0.6353, "step": 2570 }, { "epoch": 0.66, "grad_norm": 0.24340603859531101, "learning_rate": 6.235305117635428e-05, "loss": 0.7348, "step": 2575 }, { "epoch": 0.66, "grad_norm": 0.24374512948837185, "learning_rate": 6.193859779510712e-05, "loss": 0.6906, "step": 2580 }, { "epoch": 0.66, "grad_norm": 0.2617782514018871, "learning_rate": 6.152490755460715e-05, "loss": 0.6707, "step": 2585 }, { "epoch": 0.66, "grad_norm": 0.25140263838285776, "learning_rate": 6.111198874944845e-05, "loss": 0.6892, "step": 2590 }, { "epoch": 0.67, "grad_norm": 0.2382806536370909, "learning_rate": 6.0699849658757545e-05, "loss": 0.7549, "step": 2595 }, { "epoch": 0.67, "grad_norm": 0.2176694237385903, "learning_rate": 6.0288498546027536e-05, "loss": 0.7753, "step": 2600 }, { "epoch": 0.67, "grad_norm": 0.23853235406621046, "learning_rate": 5.987794365895229e-05, "loss": 0.6735, "step": 2605 }, { "epoch": 0.67, "grad_norm": 0.33567228876946426, "learning_rate": 5.946819322926127e-05, "loss": 0.6817, "step": 2610 }, { "epoch": 0.67, "grad_norm": 0.2947916364685711, "learning_rate": 5.9059255472554195e-05, "loss": 0.6897, "step": 2615 }, { "epoch": 0.67, "grad_norm": 0.25054528848541685, "learning_rate": 5.865113858813673e-05, "loss": 0.6654, "step": 2620 }, { "epoch": 0.67, "grad_norm": 0.295602257498029, "learning_rate": 5.824385075885557e-05, "loss": 0.6837, "step": 2625 }, { "epoch": 0.67, "grad_norm": 0.22560608641117053, "learning_rate": 5.783740015093484e-05, "loss": 0.7137, "step": 2630 }, { "epoch": 0.68, "grad_norm": 0.2710250240814673, "learning_rate": 5.743179491381204e-05, "loss": 0.7015, "step": 2635 }, { "epoch": 0.68, "grad_norm": 0.25266821650481586, "learning_rate": 5.702704317997492e-05, "loss": 0.6724, "step": 2640 }, { "epoch": 0.68, "grad_norm": 0.23287193866674372, "learning_rate": 5.662315306479806e-05, "loss": 0.7023, "step": 2645 }, { "epoch": 0.68, "grad_norm": 0.3278896346775186, "learning_rate": 5.6220132666380635e-05, "loss": 0.6209, "step": 2650 }, { "epoch": 0.68, "grad_norm": 0.3090011108241083, "learning_rate": 5.581799006538354e-05, "loss": 0.7575, "step": 2655 }, { "epoch": 0.68, "grad_norm": 0.30486291365985196, "learning_rate": 5.541673332486773e-05, "loss": 0.741, "step": 2660 }, { "epoch": 0.68, "grad_norm": 0.29222447612374014, "learning_rate": 5.5016370490132364e-05, "loss": 0.8456, "step": 2665 }, { "epoch": 0.68, "grad_norm": 0.30928681296284377, "learning_rate": 5.4616909588553674e-05, "loss": 0.785, "step": 2670 }, { "epoch": 0.69, "grad_norm": 0.2751241257699663, "learning_rate": 5.4218358629423794e-05, "loss": 0.797, "step": 2675 }, { "epoch": 0.69, "grad_norm": 0.3020155622331884, "learning_rate": 5.3820725603790346e-05, "loss": 0.7335, "step": 2680 }, { "epoch": 0.69, "grad_norm": 0.3132294366433861, "learning_rate": 5.342401848429615e-05, "loss": 0.7751, "step": 2685 }, { "epoch": 0.69, "grad_norm": 0.2617050345460146, "learning_rate": 5.30282452250193e-05, "loss": 0.7865, "step": 2690 }, { "epoch": 0.69, "grad_norm": 0.3088811287968099, "learning_rate": 5.263341376131382e-05, "loss": 0.7012, "step": 2695 }, { "epoch": 0.69, "grad_norm": 0.3053801166515793, "learning_rate": 5.223953200965055e-05, "loss": 0.733, "step": 2700 }, { "epoch": 0.69, "grad_norm": 0.23354984138597276, "learning_rate": 5.1846607867458196e-05, "loss": 0.8136, "step": 2705 }, { "epoch": 0.7, "grad_norm": 0.2958514046814973, "learning_rate": 5.145464921296537e-05, "loss": 0.7738, "step": 2710 }, { "epoch": 0.7, "grad_norm": 0.30080922548213934, "learning_rate": 5.1063663905042225e-05, "loss": 0.6893, "step": 2715 }, { "epoch": 0.7, "grad_norm": 0.3034194981171166, "learning_rate": 5.067365978304315e-05, "loss": 0.6922, "step": 2720 }, { "epoch": 0.7, "grad_norm": 0.33312809082970085, "learning_rate": 5.0284644666649485e-05, "loss": 0.6818, "step": 2725 }, { "epoch": 0.7, "grad_norm": 0.2615876279834696, "learning_rate": 4.9896626355712805e-05, "loss": 0.6837, "step": 2730 }, { "epoch": 0.7, "grad_norm": 0.23784349313669656, "learning_rate": 4.95096126300984e-05, "loss": 0.6899, "step": 2735 }, { "epoch": 0.7, "grad_norm": 0.2103897994837222, "learning_rate": 4.912361124952948e-05, "loss": 0.7308, "step": 2740 }, { "epoch": 0.7, "grad_norm": 0.23734181510531857, "learning_rate": 4.873862995343139e-05, "loss": 0.6635, "step": 2745 }, { "epoch": 0.71, "grad_norm": 0.28754328918934463, "learning_rate": 4.835467646077656e-05, "loss": 0.7335, "step": 2750 }, { "epoch": 0.71, "grad_norm": 0.2801177269384902, "learning_rate": 4.797175846992964e-05, "loss": 0.6538, "step": 2755 }, { "epoch": 0.71, "grad_norm": 0.20377437169807378, "learning_rate": 4.7589883658493296e-05, "loss": 0.7343, "step": 2760 }, { "epoch": 0.71, "grad_norm": 0.2633015839389786, "learning_rate": 4.72090596831542e-05, "loss": 0.7279, "step": 2765 }, { "epoch": 0.71, "grad_norm": 0.3365363313211822, "learning_rate": 4.682929417952939e-05, "loss": 0.6493, "step": 2770 }, { "epoch": 0.71, "grad_norm": 0.26197423166598477, "learning_rate": 4.645059476201333e-05, "loss": 0.6823, "step": 2775 }, { "epoch": 0.71, "grad_norm": 0.22148899156479446, "learning_rate": 4.6072969023625165e-05, "loss": 0.7062, "step": 2780 }, { "epoch": 0.71, "grad_norm": 0.25093425817830756, "learning_rate": 4.5696424535856574e-05, "loss": 0.7058, "step": 2785 }, { "epoch": 0.72, "grad_norm": 0.29374695702116843, "learning_rate": 4.532096884851978e-05, "loss": 0.7307, "step": 2790 }, { "epoch": 0.72, "grad_norm": 0.24148367178848454, "learning_rate": 4.494660948959645e-05, "loss": 0.7908, "step": 2795 }, { "epoch": 0.72, "grad_norm": 0.29661735702318875, "learning_rate": 4.457335396508631e-05, "loss": 0.6051, "step": 2800 }, { "epoch": 0.72, "grad_norm": 0.2587416887380861, "learning_rate": 4.420120975885723e-05, "loss": 0.7074, "step": 2805 }, { "epoch": 0.72, "grad_norm": 0.25077785676951697, "learning_rate": 4.383018433249464e-05, "loss": 0.6776, "step": 2810 }, { "epoch": 0.72, "grad_norm": 0.26766730021230084, "learning_rate": 4.346028512515232e-05, "loss": 0.5915, "step": 2815 }, { "epoch": 0.72, "grad_norm": 0.26255599946956093, "learning_rate": 4.309151955340297e-05, "loss": 0.7202, "step": 2820 }, { "epoch": 0.72, "grad_norm": 0.23590120420333469, "learning_rate": 4.272389501108962e-05, "loss": 0.754, "step": 2825 }, { "epoch": 0.73, "grad_norm": 0.34961647734551377, "learning_rate": 4.2357418869177354e-05, "loss": 0.7484, "step": 2830 }, { "epoch": 0.73, "grad_norm": 0.2567914752516631, "learning_rate": 4.1992098475605654e-05, "loss": 0.7758, "step": 2835 }, { "epoch": 0.73, "grad_norm": 0.2859760124915357, "learning_rate": 4.162794115514078e-05, "loss": 0.7259, "step": 2840 }, { "epoch": 0.73, "grad_norm": 0.40354694842899336, "learning_rate": 4.1264954209229254e-05, "loss": 0.6554, "step": 2845 }, { "epoch": 0.73, "grad_norm": 0.31410918890445977, "learning_rate": 4.0903144915851174e-05, "loss": 0.6495, "step": 2850 }, { "epoch": 0.73, "grad_norm": 0.21920834371525927, "learning_rate": 4.054252052937444e-05, "loss": 0.6159, "step": 2855 }, { "epoch": 0.73, "grad_norm": 0.2919386673597319, "learning_rate": 4.018308828040924e-05, "loss": 0.7266, "step": 2860 }, { "epoch": 0.73, "grad_norm": 0.2907127546681081, "learning_rate": 3.982485537566321e-05, "loss": 0.7245, "step": 2865 }, { "epoch": 0.74, "grad_norm": 0.19919891868305908, "learning_rate": 3.946782899779667e-05, "loss": 0.6152, "step": 2870 }, { "epoch": 0.74, "grad_norm": 0.2874798645438607, "learning_rate": 3.911201630527894e-05, "loss": 0.7391, "step": 2875 }, { "epoch": 0.74, "grad_norm": 0.29325850253665986, "learning_rate": 3.875742443224451e-05, "loss": 0.6121, "step": 2880 }, { "epoch": 0.74, "grad_norm": 0.3084377618826001, "learning_rate": 3.84040604883502e-05, "loss": 0.7255, "step": 2885 }, { "epoch": 0.74, "grad_norm": 0.2536420954931237, "learning_rate": 3.805193155863247e-05, "loss": 0.636, "step": 2890 }, { "epoch": 0.74, "grad_norm": 0.22759460676784352, "learning_rate": 3.770104470336555e-05, "loss": 0.717, "step": 2895 }, { "epoch": 0.74, "grad_norm": 0.22465050243662746, "learning_rate": 3.7351406957919636e-05, "loss": 0.726, "step": 2900 }, { "epoch": 0.75, "grad_norm": 0.30700513035899746, "learning_rate": 3.700302533262007e-05, "loss": 0.7157, "step": 2905 }, { "epoch": 0.75, "grad_norm": 0.3126415326969147, "learning_rate": 3.665590681260658e-05, "loss": 0.6963, "step": 2910 }, { "epoch": 0.75, "grad_norm": 0.3014647601543792, "learning_rate": 3.631005835769334e-05, "loss": 0.7982, "step": 2915 }, { "epoch": 0.75, "grad_norm": 0.2519716019821263, "learning_rate": 3.59654869022294e-05, "loss": 0.7267, "step": 2920 }, { "epoch": 0.75, "grad_norm": 0.2454160383073075, "learning_rate": 3.5622199354959677e-05, "loss": 0.8258, "step": 2925 }, { "epoch": 0.75, "grad_norm": 0.22670327761402992, "learning_rate": 3.5280202598886324e-05, "loss": 0.6447, "step": 2930 }, { "epoch": 0.75, "grad_norm": 0.24579997279476928, "learning_rate": 3.4939503491130934e-05, "loss": 0.6665, "step": 2935 }, { "epoch": 0.75, "grad_norm": 0.2849552092195358, "learning_rate": 3.4600108862796796e-05, "loss": 0.7252, "step": 2940 }, { "epoch": 0.76, "grad_norm": 0.3060590068774777, "learning_rate": 3.426202551883213e-05, "loss": 0.7553, "step": 2945 }, { "epoch": 0.76, "grad_norm": 0.25467435210378453, "learning_rate": 3.392526023789349e-05, "loss": 0.7237, "step": 2950 }, { "epoch": 0.76, "grad_norm": 0.2276342572822446, "learning_rate": 3.358981977221006e-05, "loss": 0.6985, "step": 2955 }, { "epoch": 0.76, "grad_norm": 0.24334295151794627, "learning_rate": 3.325571084744803e-05, "loss": 0.6964, "step": 2960 }, { "epoch": 0.76, "grad_norm": 0.3063546229020857, "learning_rate": 3.292294016257598e-05, "loss": 0.69, "step": 2965 }, { "epoch": 0.76, "grad_norm": 0.26843353730082814, "learning_rate": 3.259151438973024e-05, "loss": 0.7961, "step": 2970 }, { "epoch": 0.76, "grad_norm": 0.28440009360240165, "learning_rate": 3.2261440174081525e-05, "loss": 0.7703, "step": 2975 }, { "epoch": 0.76, "grad_norm": 0.2307084225718035, "learning_rate": 3.1932724133701344e-05, "loss": 0.5392, "step": 2980 }, { "epoch": 0.77, "grad_norm": 0.29912287559451467, "learning_rate": 3.160537285942956e-05, "loss": 0.6836, "step": 2985 }, { "epoch": 0.77, "grad_norm": 0.278878639560016, "learning_rate": 3.1279392914742046e-05, "loss": 0.7267, "step": 2990 }, { "epoch": 0.77, "grad_norm": 0.32874234679202635, "learning_rate": 3.0954790835619195e-05, "loss": 0.6985, "step": 2995 }, { "epoch": 0.77, "grad_norm": 0.27469910090881833, "learning_rate": 3.06315731304148e-05, "loss": 0.7915, "step": 3000 }, { "epoch": 0.77, "grad_norm": 0.33032565430804056, "learning_rate": 3.0309746279725748e-05, "loss": 0.672, "step": 3005 }, { "epoch": 0.77, "grad_norm": 0.22509955106392437, "learning_rate": 2.998931673626175e-05, "loss": 0.6924, "step": 3010 }, { "epoch": 0.77, "grad_norm": 0.24954960719448557, "learning_rate": 2.9670290924716337e-05, "loss": 0.7648, "step": 3015 }, { "epoch": 0.77, "grad_norm": 0.18097577821872354, "learning_rate": 2.935267524163774e-05, "loss": 0.724, "step": 3020 }, { "epoch": 0.78, "grad_norm": 0.2531814760979886, "learning_rate": 2.9036476055300765e-05, "loss": 0.7565, "step": 3025 }, { "epoch": 0.78, "grad_norm": 0.26537867964143863, "learning_rate": 2.872169970557913e-05, "loss": 0.7587, "step": 3030 }, { "epoch": 0.78, "grad_norm": 0.2503354683898301, "learning_rate": 2.8408352503818357e-05, "loss": 0.7464, "step": 3035 }, { "epoch": 0.78, "grad_norm": 0.31309433063885866, "learning_rate": 2.8096440732709083e-05, "loss": 0.7215, "step": 3040 }, { "epoch": 0.78, "grad_norm": 0.273515150138006, "learning_rate": 2.778597064616133e-05, "loss": 0.7107, "step": 3045 }, { "epoch": 0.78, "grad_norm": 0.29261711206548846, "learning_rate": 2.7476948469178887e-05, "loss": 0.6895, "step": 3050 }, { "epoch": 0.78, "grad_norm": 0.2613608515124986, "learning_rate": 2.716938039773459e-05, "loss": 0.751, "step": 3055 }, { "epoch": 0.79, "grad_norm": 0.3275410166575591, "learning_rate": 2.6863272598646106e-05, "loss": 0.6056, "step": 3060 }, { "epoch": 0.79, "grad_norm": 0.243877381080221, "learning_rate": 2.6558631209452323e-05, "loss": 0.7083, "step": 3065 }, { "epoch": 0.79, "grad_norm": 0.2670921124404005, "learning_rate": 2.625546233829016e-05, "loss": 0.6769, "step": 3070 }, { "epoch": 0.79, "grad_norm": 0.26772171237531533, "learning_rate": 2.5953772063772252e-05, "loss": 0.7421, "step": 3075 }, { "epoch": 0.79, "grad_norm": 0.24929713338883455, "learning_rate": 2.5653566434864928e-05, "loss": 0.7006, "step": 3080 }, { "epoch": 0.79, "grad_norm": 0.27784311313283716, "learning_rate": 2.5354851470767037e-05, "loss": 0.6997, "step": 3085 }, { "epoch": 0.79, "grad_norm": 0.27642936475366325, "learning_rate": 2.5057633160789184e-05, "loss": 0.7078, "step": 3090 }, { "epoch": 0.79, "grad_norm": 0.2916050170142823, "learning_rate": 2.476191746423373e-05, "loss": 0.7769, "step": 3095 }, { "epoch": 0.8, "grad_norm": 0.29742464103075256, "learning_rate": 2.446771031027527e-05, "loss": 0.6983, "step": 3100 }, { "epoch": 0.8, "grad_norm": 0.2520881480340573, "learning_rate": 2.4175017597841666e-05, "loss": 0.7945, "step": 3105 }, { "epoch": 0.8, "grad_norm": 0.2456996066425862, "learning_rate": 2.3883845195495878e-05, "loss": 0.7445, "step": 3110 }, { "epoch": 0.8, "grad_norm": 0.20892213586106406, "learning_rate": 2.3594198941318248e-05, "loss": 0.6771, "step": 3115 }, { "epoch": 0.8, "grad_norm": 0.22864359852007016, "learning_rate": 2.330608464278953e-05, "loss": 0.7326, "step": 3120 }, { "epoch": 0.8, "grad_norm": 0.2630175043544034, "learning_rate": 2.3019508076674267e-05, "loss": 0.6733, "step": 3125 }, { "epoch": 0.8, "grad_norm": 0.25018036673355504, "learning_rate": 2.273447498890521e-05, "loss": 0.8375, "step": 3130 }, { "epoch": 0.8, "grad_norm": 0.2970455279248964, "learning_rate": 2.2450991094467865e-05, "loss": 0.665, "step": 3135 }, { "epoch": 0.81, "grad_norm": 0.24388424842231599, "learning_rate": 2.2169062077286075e-05, "loss": 0.6899, "step": 3140 }, { "epoch": 0.81, "grad_norm": 0.2620919423557573, "learning_rate": 2.1888693590107946e-05, "loss": 0.7008, "step": 3145 }, { "epoch": 0.81, "grad_norm": 0.1597077931299718, "learning_rate": 2.1609891254392678e-05, "loss": 0.6285, "step": 3150 }, { "epoch": 0.81, "grad_norm": 0.24978397247092607, "learning_rate": 2.1332660660197602e-05, "loss": 0.6707, "step": 3155 }, { "epoch": 0.81, "grad_norm": 0.32226574046620643, "learning_rate": 2.1057007366066373e-05, "loss": 0.7732, "step": 3160 }, { "epoch": 0.81, "grad_norm": 0.2633118389195218, "learning_rate": 2.0782936898917204e-05, "loss": 0.6722, "step": 3165 }, { "epoch": 0.81, "grad_norm": 0.2065809642257561, "learning_rate": 2.0510454753932395e-05, "loss": 0.6913, "step": 3170 }, { "epoch": 0.81, "grad_norm": 0.2502396270636488, "learning_rate": 2.0239566394447874e-05, "loss": 0.7196, "step": 3175 }, { "epoch": 0.82, "grad_norm": 0.2553706993409484, "learning_rate": 1.9970277251843862e-05, "loss": 0.7153, "step": 3180 }, { "epoch": 0.82, "grad_norm": 0.20930260480653445, "learning_rate": 1.9702592725435786e-05, "loss": 0.7252, "step": 3185 }, { "epoch": 0.82, "grad_norm": 0.2624209534259566, "learning_rate": 1.9436518182366158e-05, "loss": 0.7354, "step": 3190 }, { "epoch": 0.82, "grad_norm": 0.2835973619438766, "learning_rate": 1.9172058957496876e-05, "loss": 0.7456, "step": 3195 }, { "epoch": 0.82, "grad_norm": 0.20758735208627346, "learning_rate": 1.8909220353302392e-05, "loss": 0.6972, "step": 3200 }, { "epoch": 0.82, "grad_norm": 0.34020958211549873, "learning_rate": 1.8648007639763176e-05, "loss": 0.7108, "step": 3205 }, { "epoch": 0.82, "grad_norm": 0.25980592525235724, "learning_rate": 1.838842605426031e-05, "loss": 0.6254, "step": 3210 }, { "epoch": 0.82, "grad_norm": 0.23905673738763042, "learning_rate": 1.813048080147025e-05, "loss": 0.6463, "step": 3215 }, { "epoch": 0.83, "grad_norm": 0.21218256948907488, "learning_rate": 1.7874177053260598e-05, "loss": 0.6492, "step": 3220 }, { "epoch": 0.83, "grad_norm": 0.2000868258423336, "learning_rate": 1.7619519948586348e-05, "loss": 0.6001, "step": 3225 }, { "epoch": 0.83, "grad_norm": 1.103753197876994, "learning_rate": 1.736651459338695e-05, "loss": 0.6952, "step": 3230 }, { "epoch": 0.83, "grad_norm": 0.2682682663167215, "learning_rate": 1.711516606048377e-05, "loss": 0.7109, "step": 3235 }, { "epoch": 0.83, "grad_norm": 0.25063878481734475, "learning_rate": 1.6865479389478545e-05, "loss": 0.7225, "step": 3240 }, { "epoch": 0.83, "grad_norm": 0.2149654094754697, "learning_rate": 1.6617459586652196e-05, "loss": 0.6709, "step": 3245 }, { "epoch": 0.83, "grad_norm": 0.24092454032089566, "learning_rate": 1.6371111624864543e-05, "loss": 0.6764, "step": 3250 }, { "epoch": 0.84, "grad_norm": 0.2454121794760406, "learning_rate": 1.612644044345456e-05, "loss": 0.6982, "step": 3255 }, { "epoch": 0.84, "grad_norm": 0.3041665993053294, "learning_rate": 1.5883450948141377e-05, "loss": 0.658, "step": 3260 }, { "epoch": 0.84, "grad_norm": 0.2143063192080754, "learning_rate": 1.564214801092585e-05, "loss": 0.7606, "step": 3265 }, { "epoch": 0.84, "grad_norm": 0.23715373616102758, "learning_rate": 1.540253646999299e-05, "loss": 0.7262, "step": 3270 }, { "epoch": 0.84, "grad_norm": 0.266422273258976, "learning_rate": 1.5164621129614787e-05, "loss": 0.6795, "step": 3275 }, { "epoch": 0.84, "grad_norm": 0.2303123926465437, "learning_rate": 1.4928406760054059e-05, "loss": 0.6995, "step": 3280 }, { "epoch": 0.84, "grad_norm": 0.30575921163679864, "learning_rate": 1.4693898097468662e-05, "loss": 0.7238, "step": 3285 }, { "epoch": 0.84, "grad_norm": 0.2723263223045365, "learning_rate": 1.4461099843816684e-05, "loss": 0.694, "step": 3290 }, { "epoch": 0.85, "grad_norm": 0.2704567901239254, "learning_rate": 1.4230016666761981e-05, "loss": 0.7654, "step": 3295 }, { "epoch": 0.85, "grad_norm": 0.24548577793970916, "learning_rate": 1.4000653199580782e-05, "loss": 0.766, "step": 3300 }, { "epoch": 0.85, "grad_norm": 0.31075109788590083, "learning_rate": 1.3773014041068633e-05, "loss": 0.7156, "step": 3305 }, { "epoch": 0.85, "grad_norm": 0.23774900014129338, "learning_rate": 1.3547103755448287e-05, "loss": 0.7829, "step": 3310 }, { "epoch": 0.85, "grad_norm": 0.20200953268790334, "learning_rate": 1.3322926872278097e-05, "loss": 0.7726, "step": 3315 }, { "epoch": 0.85, "grad_norm": 0.26897850573503673, "learning_rate": 1.3100487886361379e-05, "loss": 0.7458, "step": 3320 }, { "epoch": 0.85, "grad_norm": 0.2397184808414706, "learning_rate": 1.2879791257656049e-05, "loss": 0.7193, "step": 3325 }, { "epoch": 0.85, "grad_norm": 0.20474419583059367, "learning_rate": 1.266084141118542e-05, "loss": 0.6816, "step": 3330 }, { "epoch": 0.86, "grad_norm": 0.22050960020614993, "learning_rate": 1.2443642736949258e-05, "loss": 0.7013, "step": 3335 }, { "epoch": 0.86, "grad_norm": 0.22350828170159398, "learning_rate": 1.2228199589835999e-05, "loss": 0.7583, "step": 3340 }, { "epoch": 0.86, "grad_norm": 0.27678043547723374, "learning_rate": 1.2014516289535249e-05, "loss": 0.7093, "step": 3345 }, { "epoch": 0.86, "grad_norm": 0.2641609501988626, "learning_rate": 1.1802597120451286e-05, "loss": 0.7644, "step": 3350 }, { "epoch": 0.86, "grad_norm": 0.29287525021588734, "learning_rate": 1.1592446331617068e-05, "loss": 0.6913, "step": 3355 }, { "epoch": 0.86, "grad_norm": 0.3132569389457329, "learning_rate": 1.1384068136609105e-05, "loss": 0.7836, "step": 3360 }, { "epoch": 0.86, "grad_norm": 0.3257533444384463, "learning_rate": 1.1177466713462926e-05, "loss": 0.6575, "step": 3365 }, { "epoch": 0.86, "grad_norm": 0.3658464833825346, "learning_rate": 1.0972646204589377e-05, "loss": 0.6571, "step": 3370 }, { "epoch": 0.87, "grad_norm": 0.24393335900845253, "learning_rate": 1.076961071669148e-05, "loss": 0.7282, "step": 3375 }, { "epoch": 0.87, "grad_norm": 0.2673318448109565, "learning_rate": 1.0568364320682178e-05, "loss": 0.7788, "step": 3380 }, { "epoch": 0.87, "grad_norm": 0.24802808118552172, "learning_rate": 1.0368911051602636e-05, "loss": 0.6588, "step": 3385 }, { "epoch": 0.87, "grad_norm": 0.2725189792906911, "learning_rate": 1.0171254908541372e-05, "loss": 0.6729, "step": 3390 }, { "epoch": 0.87, "grad_norm": 0.3108637636142955, "learning_rate": 9.975399854554068e-06, "loss": 0.6717, "step": 3395 }, { "epoch": 0.87, "grad_norm": 0.2770989947603813, "learning_rate": 9.781349816584162e-06, "loss": 0.7259, "step": 3400 }, { "epoch": 0.87, "grad_norm": 0.23614090997589118, "learning_rate": 9.589108685383975e-06, "loss": 0.7101, "step": 3405 }, { "epoch": 0.87, "grad_norm": 0.2808653787629317, "learning_rate": 9.398680315436903e-06, "loss": 0.6933, "step": 3410 }, { "epoch": 0.88, "grad_norm": 0.24876373568373047, "learning_rate": 9.210068524879923e-06, "loss": 0.6842, "step": 3415 }, { "epoch": 0.88, "grad_norm": 0.2678931806909137, "learning_rate": 9.023277095427173e-06, "loss": 0.7012, "step": 3420 }, { "epoch": 0.88, "grad_norm": 0.25424385975928604, "learning_rate": 8.838309772294085e-06, "loss": 0.7443, "step": 3425 }, { "epoch": 0.88, "grad_norm": 0.2402611926537631, "learning_rate": 8.655170264122303e-06, "loss": 0.7288, "step": 3430 }, { "epoch": 0.88, "grad_norm": 0.24815108653858006, "learning_rate": 8.473862242905339e-06, "loss": 0.7894, "step": 3435 }, { "epoch": 0.88, "grad_norm": 0.21869677597116047, "learning_rate": 8.294389343914899e-06, "loss": 0.6481, "step": 3440 }, { "epoch": 0.88, "grad_norm": 0.2515057283571701, "learning_rate": 8.11675516562802e-06, "loss": 0.7195, "step": 3445 }, { "epoch": 0.89, "grad_norm": 0.24394419541003282, "learning_rate": 7.940963269654922e-06, "loss": 0.8022, "step": 3450 }, { "epoch": 0.89, "grad_norm": 0.2745733416076033, "learning_rate": 7.767017180667645e-06, "loss": 0.7189, "step": 3455 }, { "epoch": 0.89, "grad_norm": 0.19526287456709943, "learning_rate": 7.594920386329252e-06, "loss": 0.6829, "step": 3460 }, { "epoch": 0.89, "grad_norm": 0.3408386827932112, "learning_rate": 7.4246763372240795e-06, "loss": 0.6084, "step": 3465 }, { "epoch": 0.89, "grad_norm": 0.22382037366933571, "learning_rate": 7.256288446788362e-06, "loss": 0.6839, "step": 3470 }, { "epoch": 0.89, "grad_norm": 0.25700615722112985, "learning_rate": 7.089760091241937e-06, "loss": 0.766, "step": 3475 }, { "epoch": 0.89, "grad_norm": 0.20934034960089873, "learning_rate": 6.925094609520455e-06, "loss": 0.6959, "step": 3480 }, { "epoch": 0.89, "grad_norm": 0.1887454649423391, "learning_rate": 6.762295303208532e-06, "loss": 0.6084, "step": 3485 }, { "epoch": 0.9, "grad_norm": 0.24866922377491815, "learning_rate": 6.601365436473439e-06, "loss": 0.76, "step": 3490 }, { "epoch": 0.9, "grad_norm": 0.26188268702284273, "learning_rate": 6.44230823599975e-06, "loss": 0.716, "step": 3495 }, { "epoch": 0.9, "grad_norm": 0.29822606608650276, "learning_rate": 6.2851268909245865e-06, "loss": 0.6727, "step": 3500 }, { "epoch": 0.9, "grad_norm": 0.2732588418921964, "learning_rate": 6.1298245527737e-06, "loss": 0.7553, "step": 3505 }, { "epoch": 0.9, "grad_norm": 0.24961501560552232, "learning_rate": 5.976404335398256e-06, "loss": 0.6262, "step": 3510 }, { "epoch": 0.9, "grad_norm": 0.36125761108834664, "learning_rate": 5.824869314912473e-06, "loss": 0.702, "step": 3515 }, { "epoch": 0.9, "grad_norm": 0.20211776842690213, "learning_rate": 5.675222529631841e-06, "loss": 0.7146, "step": 3520 }, { "epoch": 0.9, "grad_norm": 0.34319105583503823, "learning_rate": 5.527466980012297e-06, "loss": 0.7202, "step": 3525 }, { "epoch": 0.91, "grad_norm": 0.2551086799124065, "learning_rate": 5.381605628590003e-06, "loss": 0.7343, "step": 3530 }, { "epoch": 0.91, "grad_norm": 0.24191784730605104, "learning_rate": 5.237641399921955e-06, "loss": 0.676, "step": 3535 }, { "epoch": 0.91, "grad_norm": 0.23642128893948972, "learning_rate": 5.095577180527378e-06, "loss": 0.75, "step": 3540 }, { "epoch": 0.91, "grad_norm": 0.24384299779088495, "learning_rate": 4.9554158188298445e-06, "loss": 0.7689, "step": 3545 }, { "epoch": 0.91, "grad_norm": 0.2622807786449119, "learning_rate": 4.817160125100106e-06, "loss": 0.7586, "step": 3550 }, { "epoch": 0.91, "grad_norm": 0.2849253620392849, "learning_rate": 4.680812871399854e-06, "loss": 0.7066, "step": 3555 }, { "epoch": 0.91, "grad_norm": 0.24191852621283047, "learning_rate": 4.546376791525975e-06, "loss": 0.7219, "step": 3560 }, { "epoch": 0.91, "grad_norm": 0.2212643288267702, "learning_rate": 4.413854580955945e-06, "loss": 0.7128, "step": 3565 }, { "epoch": 0.92, "grad_norm": 0.23498346434116985, "learning_rate": 4.2832488967935795e-06, "loss": 0.7419, "step": 3570 }, { "epoch": 0.92, "grad_norm": 0.24375104729509095, "learning_rate": 4.154562357715952e-06, "loss": 0.7244, "step": 3575 }, { "epoch": 0.92, "grad_norm": 0.2349397771494649, "learning_rate": 4.02779754392072e-06, "loss": 0.686, "step": 3580 }, { "epoch": 0.92, "grad_norm": 0.2657929830383169, "learning_rate": 3.90295699707447e-06, "loss": 0.7006, "step": 3585 }, { "epoch": 0.92, "grad_norm": 0.313209415878294, "learning_rate": 3.780043220261764e-06, "loss": 0.7512, "step": 3590 }, { "epoch": 0.92, "grad_norm": 0.2555739248275014, "learning_rate": 3.659058677934957e-06, "loss": 0.7011, "step": 3595 }, { "epoch": 0.92, "grad_norm": 0.33802960858245146, "learning_rate": 3.540005795864709e-06, "loss": 0.6056, "step": 3600 }, { "epoch": 0.92, "grad_norm": 0.1922803471445042, "learning_rate": 3.422886961091476e-06, "loss": 0.6843, "step": 3605 }, { "epoch": 0.93, "grad_norm": 0.29061663111511976, "learning_rate": 3.3077045218775192e-06, "loss": 0.7806, "step": 3610 }, { "epoch": 0.93, "grad_norm": 0.247673796937873, "learning_rate": 3.194460787659892e-06, "loss": 0.7209, "step": 3615 }, { "epoch": 0.93, "grad_norm": 0.2560647317152861, "learning_rate": 3.0831580290041184e-06, "loss": 0.7605, "step": 3620 }, { "epoch": 0.93, "grad_norm": 0.3634088101447885, "learning_rate": 2.973798477558709e-06, "loss": 0.7043, "step": 3625 }, { "epoch": 0.93, "grad_norm": 0.18733318063360602, "learning_rate": 2.8663843260103074e-06, "loss": 0.6693, "step": 3630 }, { "epoch": 0.93, "grad_norm": 0.21584750971265176, "learning_rate": 2.76091772803988e-06, "loss": 0.7167, "step": 3635 }, { "epoch": 0.93, "grad_norm": 0.3177595541385448, "learning_rate": 2.6574007982793857e-06, "loss": 0.7402, "step": 3640 }, { "epoch": 0.94, "grad_norm": 0.28268884396101174, "learning_rate": 2.555835612269475e-06, "loss": 0.6604, "step": 3645 }, { "epoch": 0.94, "grad_norm": 0.23788632053961686, "learning_rate": 2.456224206417812e-06, "loss": 0.6957, "step": 3650 }, { "epoch": 0.94, "grad_norm": 0.26249888785251196, "learning_rate": 2.35856857795832e-06, "loss": 0.7676, "step": 3655 }, { "epoch": 0.94, "grad_norm": 0.4060765673216887, "learning_rate": 2.262870684911045e-06, "loss": 0.6793, "step": 3660 }, { "epoch": 0.94, "grad_norm": 0.3094366288959559, "learning_rate": 2.169132446042976e-06, "loss": 0.6742, "step": 3665 }, { "epoch": 0.94, "grad_norm": 0.23238542996653788, "learning_rate": 2.0773557408295343e-06, "loss": 0.7142, "step": 3670 }, { "epoch": 0.94, "grad_norm": 0.24241260570013387, "learning_rate": 1.9875424094168647e-06, "loss": 0.694, "step": 3675 }, { "epoch": 0.94, "grad_norm": 0.24945986800032627, "learning_rate": 1.8996942525850047e-06, "loss": 0.6959, "step": 3680 }, { "epoch": 0.95, "grad_norm": 0.2943157341191807, "learning_rate": 1.813813031711742e-06, "loss": 0.6987, "step": 3685 }, { "epoch": 0.95, "grad_norm": 0.2501518225722179, "learning_rate": 1.7299004687372665e-06, "loss": 0.6767, "step": 3690 }, { "epoch": 0.95, "grad_norm": 0.2773238988005913, "learning_rate": 1.6479582461297217e-06, "loss": 0.6753, "step": 3695 }, { "epoch": 0.95, "grad_norm": 0.24642550651253037, "learning_rate": 1.5679880068514174e-06, "loss": 0.6836, "step": 3700 }, { "epoch": 0.95, "grad_norm": 0.23944679732673346, "learning_rate": 1.4899913543258814e-06, "loss": 0.7099, "step": 3705 }, { "epoch": 0.95, "grad_norm": 0.321999340037235, "learning_rate": 1.4139698524057165e-06, "loss": 0.7798, "step": 3710 }, { "epoch": 0.95, "grad_norm": 0.3325659255065644, "learning_rate": 1.3399250253413154e-06, "loss": 0.767, "step": 3715 }, { "epoch": 0.95, "grad_norm": 0.30837054124513025, "learning_rate": 1.2678583577501624e-06, "loss": 0.6706, "step": 3720 }, { "epoch": 0.96, "grad_norm": 0.31468665328396034, "learning_rate": 1.1977712945872243e-06, "loss": 0.7187, "step": 3725 }, { "epoch": 0.96, "grad_norm": 0.2858601801052754, "learning_rate": 1.1296652411158182e-06, "loss": 0.7022, "step": 3730 }, { "epoch": 0.96, "grad_norm": 0.24729340626419855, "learning_rate": 1.0635415628795665e-06, "loss": 0.691, "step": 3735 }, { "epoch": 0.96, "grad_norm": 0.23960085243766105, "learning_rate": 9.994015856749527e-07, "loss": 0.7091, "step": 3740 }, { "epoch": 0.96, "grad_norm": 0.20680354782081353, "learning_rate": 9.372465955247544e-07, "loss": 0.6042, "step": 3745 }, { "epoch": 0.96, "grad_norm": 0.2644774077674383, "learning_rate": 8.770778386522627e-07, "loss": 0.7627, "step": 3750 }, { "epoch": 0.96, "grad_norm": 0.2557259933660033, "learning_rate": 8.188965214562804e-07, "loss": 0.7372, "step": 3755 }, { "epoch": 0.96, "grad_norm": 0.2637639760583473, "learning_rate": 7.627038104869199e-07, "loss": 0.6702, "step": 3760 }, { "epoch": 0.97, "grad_norm": 0.28884424646282286, "learning_rate": 7.085008324222763e-07, "loss": 0.6375, "step": 3765 }, { "epoch": 0.97, "grad_norm": 0.27748515379743166, "learning_rate": 6.562886740457797e-07, "loss": 0.7361, "step": 3770 }, { "epoch": 0.97, "grad_norm": 0.2323425377751082, "learning_rate": 6.060683822244117e-07, "loss": 0.6933, "step": 3775 }, { "epoch": 0.97, "grad_norm": 0.357455400684556, "learning_rate": 5.578409638877457e-07, "loss": 0.5931, "step": 3780 }, { "epoch": 0.97, "grad_norm": 0.2724599339735716, "learning_rate": 5.116073860077064e-07, "loss": 0.6707, "step": 3785 }, { "epoch": 0.97, "grad_norm": 0.30393724062041544, "learning_rate": 4.6736857557925227e-07, "loss": 0.6907, "step": 3790 }, { "epoch": 0.97, "grad_norm": 0.30371804708625333, "learning_rate": 4.2512541960171294e-07, "loss": 0.7718, "step": 3795 }, { "epoch": 0.97, "grad_norm": 0.35894189655237235, "learning_rate": 3.8487876506106966e-07, "loss": 0.7522, "step": 3800 }, { "epoch": 0.98, "grad_norm": 0.24596908002441525, "learning_rate": 3.466294189129249e-07, "loss": 0.6583, "step": 3805 }, { "epoch": 0.98, "grad_norm": 0.2920176199203356, "learning_rate": 3.1037814806634815e-07, "loss": 0.707, "step": 3810 }, { "epoch": 0.98, "grad_norm": 0.2180144616080907, "learning_rate": 2.7612567936849964e-07, "loss": 0.6582, "step": 3815 }, { "epoch": 0.98, "grad_norm": 0.26484981601847507, "learning_rate": 2.43872699590042e-07, "loss": 0.6977, "step": 3820 }, { "epoch": 0.98, "grad_norm": 0.2626803702380842, "learning_rate": 2.136198554113844e-07, "loss": 0.7858, "step": 3825 }, { "epoch": 0.98, "grad_norm": 0.2744308629524476, "learning_rate": 1.8536775340970425e-07, "loss": 0.7359, "step": 3830 }, { "epoch": 0.98, "grad_norm": 0.2397873340818912, "learning_rate": 1.591169600468123e-07, "loss": 0.7332, "step": 3835 }, { "epoch": 0.99, "grad_norm": 0.261016022377339, "learning_rate": 1.348680016577397e-07, "loss": 0.7853, "step": 3840 }, { "epoch": 0.99, "grad_norm": 0.26956288380911886, "learning_rate": 1.126213644402463e-07, "loss": 0.7287, "step": 3845 }, { "epoch": 0.99, "grad_norm": 0.250467321318094, "learning_rate": 9.237749444505062e-08, "loss": 0.6572, "step": 3850 }, { "epoch": 0.99, "grad_norm": 0.2693162150713498, "learning_rate": 7.413679756684832e-08, "loss": 0.7318, "step": 3855 }, { "epoch": 0.99, "grad_norm": 0.29918573890964684, "learning_rate": 5.7899639536251883e-08, "loss": 0.707, "step": 3860 }, { "epoch": 0.99, "grad_norm": 0.27475327570345426, "learning_rate": 4.366634591237428e-08, "loss": 0.7348, "step": 3865 }, { "epoch": 0.99, "grad_norm": 0.21723287925397622, "learning_rate": 3.143720207635648e-08, "loss": 0.6721, "step": 3870 }, { "epoch": 0.99, "grad_norm": 0.35560080211348427, "learning_rate": 2.1212453225627482e-08, "loss": 0.7261, "step": 3875 }, { "epoch": 1.0, "grad_norm": 0.30524103845820444, "learning_rate": 1.299230436898613e-08, "loss": 0.7778, "step": 3880 }, { "epoch": 1.0, "grad_norm": 0.2662759700171976, "learning_rate": 6.776920322515423e-09, "loss": 0.6589, "step": 3885 }, { "epoch": 1.0, "grad_norm": 0.2828428158719728, "learning_rate": 2.566425706218567e-09, "loss": 0.6907, "step": 3890 }, { "epoch": 1.0, "grad_norm": 0.23136076886647983, "learning_rate": 3.609049415764787e-10, "loss": 0.7324, "step": 3895 }, { "epoch": 1.0, "eval_loss": NaN, "eval_runtime": 1502.4585, "eval_samples_per_second": 1.538, "eval_steps_per_second": 0.385, "step": 3898 }, { "epoch": 1.0, "step": 3898, "total_flos": 1.0023612803710976e+16, "train_loss": 0.45687119700224843, "train_runtime": 26657.4924, "train_samples_per_second": 2.339, "train_steps_per_second": 0.146 } ], "logging_steps": 5, "max_steps": 3898, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "total_flos": 1.0023612803710976e+16, "train_batch_size": 4, "trial_name": null, "trial_params": null }