{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 19419, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.007724393635099644, "grad_norm": 15.9375, "learning_rate": 2.5746652935118437e-06, "loss": 2.0265, "step": 50 }, { "epoch": 0.015448787270199289, "grad_norm": 11.0, "learning_rate": 5.149330587023687e-06, "loss": 1.6081, "step": 100 }, { "epoch": 0.023173180905298935, "grad_norm": 10.375, "learning_rate": 7.723995880535532e-06, "loss": 0.9443, "step": 150 }, { "epoch": 0.030897574540398577, "grad_norm": 7.40625, "learning_rate": 1.0298661174047375e-05, "loss": 0.6997, "step": 200 }, { "epoch": 0.03862196817549823, "grad_norm": 9.8125, "learning_rate": 1.287332646755922e-05, "loss": 0.6121, "step": 250 }, { "epoch": 0.04634636181059787, "grad_norm": 4.90625, "learning_rate": 1.5447991761071064e-05, "loss": 0.6053, "step": 300 }, { "epoch": 0.05407075544569751, "grad_norm": 4.84375, "learning_rate": 1.8022657054582907e-05, "loss": 0.4967, "step": 350 }, { "epoch": 0.061795149080797154, "grad_norm": 2.578125, "learning_rate": 2.059732234809475e-05, "loss": 0.466, "step": 400 }, { "epoch": 0.0695195427158968, "grad_norm": 3.515625, "learning_rate": 2.3171987641606592e-05, "loss": 0.4914, "step": 450 }, { "epoch": 0.07724393635099645, "grad_norm": 1.78125, "learning_rate": 2.574665293511844e-05, "loss": 0.4543, "step": 500 }, { "epoch": 0.08496832998609609, "grad_norm": 2.078125, "learning_rate": 2.8321318228630278e-05, "loss": 0.4135, "step": 550 }, { "epoch": 0.09269272362119574, "grad_norm": 3.734375, "learning_rate": 3.089598352214213e-05, "loss": 0.4241, "step": 600 }, { "epoch": 0.10041711725629537, "grad_norm": 2.671875, "learning_rate": 3.3470648815653964e-05, "loss": 0.4221, "step": 650 }, { "epoch": 0.10814151089139502, "grad_norm": 3.953125, "learning_rate": 3.604531410916581e-05, "loss": 0.4454, "step": 700 }, { "epoch": 0.11586590452649467, "grad_norm": 3.4375, "learning_rate": 3.8619979402677656e-05, "loss": 0.4157, "step": 750 }, { "epoch": 0.12359029816159431, "grad_norm": 2.125, "learning_rate": 4.11946446961895e-05, "loss": 0.4039, "step": 800 }, { "epoch": 0.13131469179669397, "grad_norm": 2.375, "learning_rate": 4.376930998970134e-05, "loss": 0.3939, "step": 850 }, { "epoch": 0.1390390854317936, "grad_norm": 1.078125, "learning_rate": 4.6343975283213184e-05, "loss": 0.4059, "step": 900 }, { "epoch": 0.14676347906689324, "grad_norm": 1.75, "learning_rate": 4.891864057672503e-05, "loss": 0.3746, "step": 950 }, { "epoch": 0.1544878727019929, "grad_norm": 1.3671875, "learning_rate": 5.149330587023688e-05, "loss": 0.3894, "step": 1000 }, { "epoch": 0.16221226633709254, "grad_norm": 3.734375, "learning_rate": 5.406797116374872e-05, "loss": 0.3776, "step": 1050 }, { "epoch": 0.16993665997219218, "grad_norm": 5.21875, "learning_rate": 5.6642636457260556e-05, "loss": 0.3761, "step": 1100 }, { "epoch": 0.17766105360729184, "grad_norm": 3.046875, "learning_rate": 5.9217301750772405e-05, "loss": 0.3685, "step": 1150 }, { "epoch": 0.18538544724239148, "grad_norm": 2.921875, "learning_rate": 6.179196704428426e-05, "loss": 0.3576, "step": 1200 }, { "epoch": 0.1931098408774911, "grad_norm": 1.765625, "learning_rate": 6.436663233779608e-05, "loss": 0.3846, "step": 1250 }, { "epoch": 0.20083423451259075, "grad_norm": 2.296875, "learning_rate": 6.694129763130793e-05, "loss": 0.3429, "step": 1300 }, { "epoch": 0.2085586281476904, "grad_norm": 1.953125, "learning_rate": 6.951596292481978e-05, "loss": 0.337, "step": 1350 }, { "epoch": 0.21628302178279005, "grad_norm": 2.484375, "learning_rate": 7.209062821833163e-05, "loss": 0.385, "step": 1400 }, { "epoch": 0.22400741541788968, "grad_norm": 1.3203125, "learning_rate": 7.466529351184346e-05, "loss": 0.3183, "step": 1450 }, { "epoch": 0.23173180905298935, "grad_norm": 2.046875, "learning_rate": 7.723995880535531e-05, "loss": 0.3743, "step": 1500 }, { "epoch": 0.23945620268808898, "grad_norm": 2.03125, "learning_rate": 7.981462409886715e-05, "loss": 0.3434, "step": 1550 }, { "epoch": 0.24718059632318862, "grad_norm": 2.578125, "learning_rate": 8.2389289392379e-05, "loss": 0.311, "step": 1600 }, { "epoch": 0.25490498995828825, "grad_norm": 3.765625, "learning_rate": 8.496395468589084e-05, "loss": 0.3344, "step": 1650 }, { "epoch": 0.26262938359338794, "grad_norm": 1.8046875, "learning_rate": 8.753861997940268e-05, "loss": 0.3135, "step": 1700 }, { "epoch": 0.2703537772284876, "grad_norm": 2.234375, "learning_rate": 9.011328527291453e-05, "loss": 0.3097, "step": 1750 }, { "epoch": 0.2780781708635872, "grad_norm": 1.859375, "learning_rate": 9.268795056642637e-05, "loss": 0.3259, "step": 1800 }, { "epoch": 0.28580256449868685, "grad_norm": 2.671875, "learning_rate": 9.526261585993821e-05, "loss": 0.3601, "step": 1850 }, { "epoch": 0.2935269581337865, "grad_norm": 1.9296875, "learning_rate": 9.783728115345005e-05, "loss": 0.308, "step": 1900 }, { "epoch": 0.3012513517688861, "grad_norm": 2.671875, "learning_rate": 0.0001004119464469619, "loss": 0.3259, "step": 1950 }, { "epoch": 0.3089757454039858, "grad_norm": 1.125, "learning_rate": 0.00010298661174047375, "loss": 0.2879, "step": 2000 }, { "epoch": 0.31670013903908545, "grad_norm": 2.484375, "learning_rate": 0.0001055612770339856, "loss": 0.3252, "step": 2050 }, { "epoch": 0.3244245326741851, "grad_norm": 1.09375, "learning_rate": 0.00010813594232749744, "loss": 0.3302, "step": 2100 }, { "epoch": 0.3321489263092847, "grad_norm": 1.21875, "learning_rate": 0.00011071060762100927, "loss": 0.313, "step": 2150 }, { "epoch": 0.33987331994438436, "grad_norm": 2.125, "learning_rate": 0.00011328527291452111, "loss": 0.3088, "step": 2200 }, { "epoch": 0.347597713579484, "grad_norm": 2.46875, "learning_rate": 0.00011585993820803295, "loss": 0.2967, "step": 2250 }, { "epoch": 0.3553221072145837, "grad_norm": 2.375, "learning_rate": 0.00011843460350154481, "loss": 0.2821, "step": 2300 }, { "epoch": 0.3630465008496833, "grad_norm": 3.40625, "learning_rate": 0.00012100926879505665, "loss": 0.3559, "step": 2350 }, { "epoch": 0.37077089448478295, "grad_norm": 0.984375, "learning_rate": 0.0001235839340885685, "loss": 0.295, "step": 2400 }, { "epoch": 0.3784952881198826, "grad_norm": 1.9296875, "learning_rate": 0.00012615859938208034, "loss": 0.3016, "step": 2450 }, { "epoch": 0.3862196817549822, "grad_norm": 2.703125, "learning_rate": 0.00012873326467559217, "loss": 0.2975, "step": 2500 }, { "epoch": 0.39394407539008186, "grad_norm": 1.5390625, "learning_rate": 0.00013130792996910402, "loss": 0.3172, "step": 2550 }, { "epoch": 0.4016684690251815, "grad_norm": 1.7265625, "learning_rate": 0.00013388259526261585, "loss": 0.2923, "step": 2600 }, { "epoch": 0.4093928626602812, "grad_norm": 1.8125, "learning_rate": 0.0001364572605561277, "loss": 0.2763, "step": 2650 }, { "epoch": 0.4171172562953808, "grad_norm": 2.0625, "learning_rate": 0.00013903192584963957, "loss": 0.3312, "step": 2700 }, { "epoch": 0.42484164993048046, "grad_norm": 1.765625, "learning_rate": 0.0001416065911431514, "loss": 0.3032, "step": 2750 }, { "epoch": 0.4325660435655801, "grad_norm": 2.578125, "learning_rate": 0.00014418125643666325, "loss": 0.3055, "step": 2800 }, { "epoch": 0.44029043720067973, "grad_norm": 3.28125, "learning_rate": 0.00014675592173017508, "loss": 0.3208, "step": 2850 }, { "epoch": 0.44801483083577937, "grad_norm": 3.328125, "learning_rate": 0.0001493305870236869, "loss": 0.2917, "step": 2900 }, { "epoch": 0.45573922447087906, "grad_norm": 1.6015625, "learning_rate": 0.00015190525231719877, "loss": 0.2998, "step": 2950 }, { "epoch": 0.4634636181059787, "grad_norm": 1.2109375, "learning_rate": 0.00015447991761071062, "loss": 0.2823, "step": 3000 }, { "epoch": 0.4711880117410783, "grad_norm": 2.265625, "learning_rate": 0.00015705458290422245, "loss": 0.3304, "step": 3050 }, { "epoch": 0.47891240537617796, "grad_norm": 2.34375, "learning_rate": 0.0001596292481977343, "loss": 0.2956, "step": 3100 }, { "epoch": 0.4866367990112776, "grad_norm": 2.03125, "learning_rate": 0.00016220391349124614, "loss": 0.2934, "step": 3150 }, { "epoch": 0.49436119264637723, "grad_norm": 0.93359375, "learning_rate": 0.000164778578784758, "loss": 0.3031, "step": 3200 }, { "epoch": 0.5020855862814769, "grad_norm": 3.046875, "learning_rate": 0.00016735324407826982, "loss": 0.2679, "step": 3250 }, { "epoch": 0.5098099799165765, "grad_norm": 3.203125, "learning_rate": 0.00016992790937178168, "loss": 0.3182, "step": 3300 }, { "epoch": 0.5175343735516762, "grad_norm": 0.94921875, "learning_rate": 0.0001725025746652935, "loss": 0.2888, "step": 3350 }, { "epoch": 0.5252587671867759, "grad_norm": 4.1875, "learning_rate": 0.00017507723995880537, "loss": 0.326, "step": 3400 }, { "epoch": 0.5329831608218755, "grad_norm": 3.09375, "learning_rate": 0.0001776519052523172, "loss": 0.3184, "step": 3450 }, { "epoch": 0.5407075544569752, "grad_norm": 2.0625, "learning_rate": 0.00018022657054582905, "loss": 0.3011, "step": 3500 }, { "epoch": 0.5484319480920747, "grad_norm": 0.796875, "learning_rate": 0.0001828012358393409, "loss": 0.309, "step": 3550 }, { "epoch": 0.5561563417271744, "grad_norm": 1.546875, "learning_rate": 0.00018537590113285274, "loss": 0.2876, "step": 3600 }, { "epoch": 0.563880735362274, "grad_norm": 2.828125, "learning_rate": 0.00018795056642636457, "loss": 0.3071, "step": 3650 }, { "epoch": 0.5716051289973737, "grad_norm": 2.296875, "learning_rate": 0.00019052523171987642, "loss": 0.3237, "step": 3700 }, { "epoch": 0.5793295226324734, "grad_norm": 0.77734375, "learning_rate": 0.00019309989701338825, "loss": 0.2923, "step": 3750 }, { "epoch": 0.587053916267573, "grad_norm": 2.015625, "learning_rate": 0.0001956745623069001, "loss": 0.296, "step": 3800 }, { "epoch": 0.5947783099026727, "grad_norm": 1.8828125, "learning_rate": 0.00019824922760041197, "loss": 0.2936, "step": 3850 }, { "epoch": 0.6025027035377722, "grad_norm": 1.6484375, "learning_rate": 0.0001999994765359982, "loss": 0.3222, "step": 3900 }, { "epoch": 0.6102270971728719, "grad_norm": 4.65625, "learning_rate": 0.00019999109305729934, "loss": 0.2641, "step": 3950 }, { "epoch": 0.6179514908079716, "grad_norm": 3.1875, "learning_rate": 0.0001999724866611328, "loss": 0.2733, "step": 4000 }, { "epoch": 0.6256758844430712, "grad_norm": 1.390625, "learning_rate": 0.00019994365924978462, "loss": 0.2864, "step": 4050 }, { "epoch": 0.6334002780781709, "grad_norm": 3.15625, "learning_rate": 0.00019990461377051967, "loss": 0.2958, "step": 4100 }, { "epoch": 0.6411246717132705, "grad_norm": 1.9609375, "learning_rate": 0.0001998553542152807, "loss": 0.2515, "step": 4150 }, { "epoch": 0.6488490653483702, "grad_norm": 1.9609375, "learning_rate": 0.00019979588562027995, "loss": 0.3032, "step": 4200 }, { "epoch": 0.6565734589834697, "grad_norm": 0.6484375, "learning_rate": 0.0001997262140654843, "loss": 0.2912, "step": 4250 }, { "epoch": 0.6642978526185694, "grad_norm": 2.375, "learning_rate": 0.0001996463466739938, "loss": 0.2887, "step": 4300 }, { "epoch": 0.6720222462536691, "grad_norm": 1.921875, "learning_rate": 0.0001995562916113133, "loss": 0.3176, "step": 4350 }, { "epoch": 0.6797466398887687, "grad_norm": 1.1171875, "learning_rate": 0.0001994560580845175, "loss": 0.2614, "step": 4400 }, { "epoch": 0.6874710335238684, "grad_norm": 1.1015625, "learning_rate": 0.0001993456563413099, "loss": 0.2594, "step": 4450 }, { "epoch": 0.695195427158968, "grad_norm": 1.515625, "learning_rate": 0.00019922509766897502, "loss": 0.2836, "step": 4500 }, { "epoch": 0.7029198207940677, "grad_norm": 3.125, "learning_rate": 0.0001990943943932242, "loss": 0.305, "step": 4550 }, { "epoch": 0.7106442144291674, "grad_norm": 5.71875, "learning_rate": 0.0001989535598769357, "loss": 0.2652, "step": 4600 }, { "epoch": 0.718368608064267, "grad_norm": 1.625, "learning_rate": 0.00019880260851878835, "loss": 0.2911, "step": 4650 }, { "epoch": 0.7260930016993666, "grad_norm": 2.96875, "learning_rate": 0.00019864155575178953, "loss": 0.2663, "step": 4700 }, { "epoch": 0.7338173953344662, "grad_norm": 4.40625, "learning_rate": 0.0001984704180416973, "loss": 0.2744, "step": 4750 }, { "epoch": 0.7415417889695659, "grad_norm": 2.640625, "learning_rate": 0.00019828921288533695, "loss": 0.2956, "step": 4800 }, { "epoch": 0.7492661826046655, "grad_norm": 2.828125, "learning_rate": 0.00019809795880881216, "loss": 0.2574, "step": 4850 }, { "epoch": 0.7569905762397652, "grad_norm": 0.734375, "learning_rate": 0.0001978966753656109, "loss": 0.3144, "step": 4900 }, { "epoch": 0.7647149698748649, "grad_norm": 2.09375, "learning_rate": 0.00019768538313460647, "loss": 0.2984, "step": 4950 }, { "epoch": 0.7724393635099644, "grad_norm": 1.1328125, "learning_rate": 0.00019746410371795334, "loss": 0.2347, "step": 5000 }, { "epoch": 0.7801637571450641, "grad_norm": 2.890625, "learning_rate": 0.00019723285973887865, "loss": 0.246, "step": 5050 }, { "epoch": 0.7878881507801637, "grad_norm": 2.046875, "learning_rate": 0.0001969916748393693, "loss": 0.2651, "step": 5100 }, { "epoch": 0.7956125444152634, "grad_norm": 1.03125, "learning_rate": 0.00019674057367775487, "loss": 0.2604, "step": 5150 }, { "epoch": 0.803336938050363, "grad_norm": 3.296875, "learning_rate": 0.00019647958192618644, "loss": 0.2768, "step": 5200 }, { "epoch": 0.8110613316854627, "grad_norm": 5.09375, "learning_rate": 0.00019620872626801203, "loss": 0.253, "step": 5250 }, { "epoch": 0.8187857253205624, "grad_norm": 1.8359375, "learning_rate": 0.00019592803439504855, "loss": 0.259, "step": 5300 }, { "epoch": 0.826510118955662, "grad_norm": 4.125, "learning_rate": 0.00019563753500475054, "loss": 0.2703, "step": 5350 }, { "epoch": 0.8342345125907616, "grad_norm": 1.25, "learning_rate": 0.00019533725779727635, "loss": 0.2941, "step": 5400 }, { "epoch": 0.8419589062258612, "grad_norm": 2.609375, "learning_rate": 0.00019502723347245143, "loss": 0.2706, "step": 5450 }, { "epoch": 0.8496832998609609, "grad_norm": 3.453125, "learning_rate": 0.0001947074937266299, "loss": 0.3306, "step": 5500 }, { "epoch": 0.8574076934960606, "grad_norm": 3.421875, "learning_rate": 0.00019437807124945362, "loss": 0.2738, "step": 5550 }, { "epoch": 0.8651320871311602, "grad_norm": 2.46875, "learning_rate": 0.00019403899972051045, "loss": 0.2659, "step": 5600 }, { "epoch": 0.8728564807662599, "grad_norm": 1.5234375, "learning_rate": 0.0001936903138058906, "loss": 0.2569, "step": 5650 }, { "epoch": 0.8805808744013595, "grad_norm": 4.09375, "learning_rate": 0.00019333204915464265, "loss": 0.2517, "step": 5700 }, { "epoch": 0.8883052680364592, "grad_norm": 2.703125, "learning_rate": 0.0001929642423951286, "loss": 0.2738, "step": 5750 }, { "epoch": 0.8960296616715587, "grad_norm": 0.78515625, "learning_rate": 0.0001925869311312793, "loss": 0.2498, "step": 5800 }, { "epoch": 0.9037540553066584, "grad_norm": 2.125, "learning_rate": 0.00019220015393874978, "loss": 0.231, "step": 5850 }, { "epoch": 0.9114784489417581, "grad_norm": 1.171875, "learning_rate": 0.0001918039503609753, "loss": 0.23, "step": 5900 }, { "epoch": 0.9192028425768577, "grad_norm": 1.546875, "learning_rate": 0.0001913983609051286, "loss": 0.2203, "step": 5950 }, { "epoch": 0.9269272362119574, "grad_norm": 4.0625, "learning_rate": 0.0001909834270379784, "loss": 0.2902, "step": 6000 }, { "epoch": 0.934651629847057, "grad_norm": 0.6171875, "learning_rate": 0.00019055919118165002, "loss": 0.2639, "step": 6050 }, { "epoch": 0.9423760234821567, "grad_norm": 2.65625, "learning_rate": 0.00019012569670928813, "loss": 0.2503, "step": 6100 }, { "epoch": 0.9501004171172563, "grad_norm": 1.1875, "learning_rate": 0.00018968298794062242, "loss": 0.2697, "step": 6150 }, { "epoch": 0.9578248107523559, "grad_norm": 2.28125, "learning_rate": 0.00018923111013743635, "loss": 0.2265, "step": 6200 }, { "epoch": 0.9655492043874556, "grad_norm": 2.109375, "learning_rate": 0.00018877010949893975, "loss": 0.2583, "step": 6250 }, { "epoch": 0.9732735980225552, "grad_norm": 2.703125, "learning_rate": 0.00018830003315704544, "loss": 0.24, "step": 6300 }, { "epoch": 0.9809979916576549, "grad_norm": 2.046875, "learning_rate": 0.00018782092917155047, "loss": 0.2824, "step": 6350 }, { "epoch": 0.9887223852927545, "grad_norm": 2.34375, "learning_rate": 0.00018733284652522285, "loss": 0.2911, "step": 6400 }, { "epoch": 0.9964467789278542, "grad_norm": 2.15625, "learning_rate": 0.0001868358351187932, "loss": 0.2468, "step": 6450 }, { "epoch": 1.0041711725629539, "grad_norm": 1.375, "learning_rate": 0.00018632994576585343, "loss": 0.2142, "step": 6500 }, { "epoch": 1.0118955661980535, "grad_norm": 1.1875, "learning_rate": 0.0001858152301876613, "loss": 0.2236, "step": 6550 }, { "epoch": 1.019619959833153, "grad_norm": 1.0, "learning_rate": 0.00018529174100785276, "loss": 0.1858, "step": 6600 }, { "epoch": 1.0273443534682527, "grad_norm": 1.0625, "learning_rate": 0.0001847595317470616, "loss": 0.1911, "step": 6650 }, { "epoch": 1.0350687471033524, "grad_norm": 1.4296875, "learning_rate": 0.00018421865681744786, "loss": 0.2166, "step": 6700 }, { "epoch": 1.042793140738452, "grad_norm": 2.0, "learning_rate": 0.0001836691715171345, "loss": 0.2184, "step": 6750 }, { "epoch": 1.0505175343735518, "grad_norm": 0.8828125, "learning_rate": 0.00018311113202455407, "loss": 0.2082, "step": 6800 }, { "epoch": 1.0582419280086512, "grad_norm": 1.125, "learning_rate": 0.00018254459539270495, "loss": 0.2122, "step": 6850 }, { "epoch": 1.065966321643751, "grad_norm": 2.8125, "learning_rate": 0.00018196961954331862, "loss": 0.2062, "step": 6900 }, { "epoch": 1.0736907152788506, "grad_norm": 2.046875, "learning_rate": 0.00018138626326093742, "loss": 0.203, "step": 6950 }, { "epoch": 1.0814151089139503, "grad_norm": 1.8046875, "learning_rate": 0.00018079458618690491, "loss": 0.1923, "step": 7000 }, { "epoch": 1.0891395025490498, "grad_norm": 1.390625, "learning_rate": 0.00018019464881326803, "loss": 0.1923, "step": 7050 }, { "epoch": 1.0968638961841495, "grad_norm": 1.2734375, "learning_rate": 0.00017958651247659257, "loss": 0.1892, "step": 7100 }, { "epoch": 1.1045882898192492, "grad_norm": 0.83984375, "learning_rate": 0.00017897023935169214, "loss": 0.2116, "step": 7150 }, { "epoch": 1.1123126834543489, "grad_norm": 0.376953125, "learning_rate": 0.0001783458924452717, "loss": 0.1857, "step": 7200 }, { "epoch": 1.1200370770894486, "grad_norm": 1.40625, "learning_rate": 0.00017771353558948568, "loss": 0.2102, "step": 7250 }, { "epoch": 1.1277614707245482, "grad_norm": 1.203125, "learning_rate": 0.00017707323343541198, "loss": 0.2191, "step": 7300 }, { "epoch": 1.1354858643596477, "grad_norm": 2.234375, "learning_rate": 0.00017642505144644212, "loss": 0.203, "step": 7350 }, { "epoch": 1.1432102579947474, "grad_norm": 1.4140625, "learning_rate": 0.00017576905589158844, "loss": 0.2028, "step": 7400 }, { "epoch": 1.150934651629847, "grad_norm": 1.203125, "learning_rate": 0.00017510531383870882, "loss": 0.2027, "step": 7450 }, { "epoch": 1.1586590452649468, "grad_norm": 1.8984375, "learning_rate": 0.00017443389314764977, "loss": 0.2239, "step": 7500 }, { "epoch": 1.1663834389000463, "grad_norm": 2.3125, "learning_rate": 0.00017375486246330848, "loss": 0.1946, "step": 7550 }, { "epoch": 1.174107832535146, "grad_norm": 0.609375, "learning_rate": 0.0001730682912086149, "loss": 0.2018, "step": 7600 }, { "epoch": 1.1818322261702456, "grad_norm": 1.3671875, "learning_rate": 0.0001723742495774339, "loss": 0.189, "step": 7650 }, { "epoch": 1.1895566198053453, "grad_norm": 4.375, "learning_rate": 0.00017167280852738878, "loss": 0.2031, "step": 7700 }, { "epoch": 1.197281013440445, "grad_norm": 3.453125, "learning_rate": 0.00017096403977260666, "loss": 0.2191, "step": 7750 }, { "epoch": 1.2050054070755445, "grad_norm": 1.5390625, "learning_rate": 0.0001702480157763867, "loss": 0.211, "step": 7800 }, { "epoch": 1.2127298007106442, "grad_norm": 0.51171875, "learning_rate": 0.00016952480974379136, "loss": 0.2126, "step": 7850 }, { "epoch": 1.2204541943457439, "grad_norm": 1.0703125, "learning_rate": 0.00016879449561416228, "loss": 0.1914, "step": 7900 }, { "epoch": 1.2281785879808436, "grad_norm": 0.8671875, "learning_rate": 0.00016805714805356056, "loss": 0.1786, "step": 7950 }, { "epoch": 1.235902981615943, "grad_norm": 1.421875, "learning_rate": 0.00016731284244713335, "loss": 0.2282, "step": 8000 }, { "epoch": 1.2436273752510427, "grad_norm": 2.953125, "learning_rate": 0.00016656165489140632, "loss": 0.1904, "step": 8050 }, { "epoch": 1.2513517688861424, "grad_norm": 1.9140625, "learning_rate": 0.0001658036621865039, "loss": 0.1763, "step": 8100 }, { "epoch": 1.259076162521242, "grad_norm": 1.515625, "learning_rate": 0.0001650389418282973, "loss": 0.2155, "step": 8150 }, { "epoch": 1.2668005561563418, "grad_norm": 2.484375, "learning_rate": 0.00016426757200048125, "loss": 0.2114, "step": 8200 }, { "epoch": 1.2745249497914415, "grad_norm": 2.71875, "learning_rate": 0.00016348963156658105, "loss": 0.1681, "step": 8250 }, { "epoch": 1.282249343426541, "grad_norm": 1.4921875, "learning_rate": 0.00016270520006188925, "loss": 0.1818, "step": 8300 }, { "epoch": 1.2899737370616406, "grad_norm": 1.734375, "learning_rate": 0.00016191435768533457, "loss": 0.1826, "step": 8350 }, { "epoch": 1.2976981306967403, "grad_norm": 0.74609375, "learning_rate": 0.000161117185291282, "loss": 0.1847, "step": 8400 }, { "epoch": 1.30542252433184, "grad_norm": 1.265625, "learning_rate": 0.00016031376438126687, "loss": 0.2101, "step": 8450 }, { "epoch": 1.3131469179669395, "grad_norm": 1.0625, "learning_rate": 0.00015950417709566182, "loss": 0.193, "step": 8500 }, { "epoch": 1.3208713116020392, "grad_norm": 1.0234375, "learning_rate": 0.0001586885062052793, "loss": 0.2035, "step": 8550 }, { "epoch": 1.3285957052371389, "grad_norm": 1.2421875, "learning_rate": 0.000157866835102909, "loss": 0.1661, "step": 8600 }, { "epoch": 1.3363200988722386, "grad_norm": 4.03125, "learning_rate": 0.00015703924779479186, "loss": 0.1882, "step": 8650 }, { "epoch": 1.3440444925073383, "grad_norm": 0.7734375, "learning_rate": 0.00015620582889203175, "loss": 0.1959, "step": 8700 }, { "epoch": 1.351768886142438, "grad_norm": 1.578125, "learning_rate": 0.00015536666360194454, "loss": 0.1594, "step": 8750 }, { "epoch": 1.3594932797775374, "grad_norm": 1.890625, "learning_rate": 0.00015452183771934696, "loss": 0.1855, "step": 8800 }, { "epoch": 1.3672176734126371, "grad_norm": 1.0078125, "learning_rate": 0.00015367143761778503, "loss": 0.1806, "step": 8850 }, { "epoch": 1.3749420670477368, "grad_norm": 1.6796875, "learning_rate": 0.00015281555024070323, "loss": 0.1944, "step": 8900 }, { "epoch": 1.3826664606828363, "grad_norm": 1.6171875, "learning_rate": 0.0001519542630925557, "loss": 0.1872, "step": 8950 }, { "epoch": 1.390390854317936, "grad_norm": 1.1875, "learning_rate": 0.00015108766422985994, "loss": 0.1587, "step": 9000 }, { "epoch": 1.3981152479530357, "grad_norm": 1.703125, "learning_rate": 0.00015021584225219394, "loss": 0.1686, "step": 9050 }, { "epoch": 1.4058396415881353, "grad_norm": 0.63671875, "learning_rate": 0.00014933888629313803, "loss": 0.1766, "step": 9100 }, { "epoch": 1.413564035223235, "grad_norm": 2.0, "learning_rate": 0.00014845688601116194, "loss": 0.1688, "step": 9150 }, { "epoch": 1.4212884288583347, "grad_norm": 1.2109375, "learning_rate": 0.00014756993158045837, "loss": 0.175, "step": 9200 }, { "epoch": 1.4290128224934342, "grad_norm": 1.578125, "learning_rate": 0.0001466781136817236, "loss": 0.1549, "step": 9250 }, { "epoch": 1.436737216128534, "grad_norm": 1.546875, "learning_rate": 0.00014578152349288648, "loss": 0.1776, "step": 9300 }, { "epoch": 1.4444616097636336, "grad_norm": 0.953125, "learning_rate": 0.00014488025267978683, "loss": 0.1887, "step": 9350 }, { "epoch": 1.4521860033987333, "grad_norm": 0.94921875, "learning_rate": 0.0001439743933868031, "loss": 0.1854, "step": 9400 }, { "epoch": 1.4599103970338327, "grad_norm": 0.79296875, "learning_rate": 0.0001430640382274323, "loss": 0.16, "step": 9450 }, { "epoch": 1.4676347906689324, "grad_norm": 0.921875, "learning_rate": 0.000142149280274821, "loss": 0.1752, "step": 9500 }, { "epoch": 1.4753591843040321, "grad_norm": 1.1171875, "learning_rate": 0.00014123021305224978, "loss": 0.1845, "step": 9550 }, { "epoch": 1.4830835779391318, "grad_norm": 1.1640625, "learning_rate": 0.00014030693052357164, "loss": 0.1675, "step": 9600 }, { "epoch": 1.4908079715742315, "grad_norm": 0.9453125, "learning_rate": 0.00013937952708360508, "loss": 0.1677, "step": 9650 }, { "epoch": 1.4985323652093312, "grad_norm": 1.640625, "learning_rate": 0.00013844809754848361, "loss": 0.1817, "step": 9700 }, { "epoch": 1.5062567588444307, "grad_norm": 1.5703125, "learning_rate": 0.0001375127371459618, "loss": 0.1588, "step": 9750 }, { "epoch": 1.5139811524795304, "grad_norm": 3.328125, "learning_rate": 0.00013657354150567924, "loss": 0.1769, "step": 9800 }, { "epoch": 1.52170554611463, "grad_norm": 1.7734375, "learning_rate": 0.00013563060664938375, "loss": 0.1604, "step": 9850 }, { "epoch": 1.5294299397497295, "grad_norm": 1.1484375, "learning_rate": 0.00013468402898111407, "loss": 0.1549, "step": 9900 }, { "epoch": 1.5371543333848292, "grad_norm": 1.8125, "learning_rate": 0.00013373390527734387, "loss": 0.1789, "step": 9950 }, { "epoch": 1.544878727019929, "grad_norm": 0.5859375, "learning_rate": 0.0001327803326770873, "loss": 0.1651, "step": 10000 }, { "epoch": 1.5526031206550286, "grad_norm": 1.4765625, "learning_rate": 0.0001318234086719678, "loss": 0.1789, "step": 10050 }, { "epoch": 1.5603275142901283, "grad_norm": 2.125, "learning_rate": 0.00013086323109625072, "loss": 0.1774, "step": 10100 }, { "epoch": 1.568051907925228, "grad_norm": 1.953125, "learning_rate": 0.0001298998981168408, "loss": 0.1604, "step": 10150 }, { "epoch": 1.5757763015603277, "grad_norm": 2.890625, "learning_rate": 0.0001289335082232459, "loss": 0.1755, "step": 10200 }, { "epoch": 1.5835006951954271, "grad_norm": 0.7578125, "learning_rate": 0.0001279641602175075, "loss": 0.1554, "step": 10250 }, { "epoch": 1.5912250888305268, "grad_norm": 0.4609375, "learning_rate": 0.00012699195320409938, "loss": 0.1518, "step": 10300 }, { "epoch": 1.5989494824656263, "grad_norm": 0.6484375, "learning_rate": 0.00012601698657979527, "loss": 0.1689, "step": 10350 }, { "epoch": 1.606673876100726, "grad_norm": 0.6953125, "learning_rate": 0.00012503936002350699, "loss": 0.139, "step": 10400 }, { "epoch": 1.6143982697358257, "grad_norm": 1.171875, "learning_rate": 0.00012405917348609296, "loss": 0.1465, "step": 10450 }, { "epoch": 1.6221226633709254, "grad_norm": 1.6875, "learning_rate": 0.0001230765271801399, "loss": 0.1446, "step": 10500 }, { "epoch": 1.629847057006025, "grad_norm": 0.9921875, "learning_rate": 0.00012209152156971692, "loss": 0.1393, "step": 10550 }, { "epoch": 1.6375714506411247, "grad_norm": 1.1640625, "learning_rate": 0.00012110425736010438, "loss": 0.1418, "step": 10600 }, { "epoch": 1.6452958442762244, "grad_norm": 0.52734375, "learning_rate": 0.00012011483548749793, "loss": 0.1552, "step": 10650 }, { "epoch": 1.653020237911324, "grad_norm": 0.44140625, "learning_rate": 0.00011912335710868905, "loss": 0.1533, "step": 10700 }, { "epoch": 1.6607446315464236, "grad_norm": 0.5625, "learning_rate": 0.00011812992359072275, "loss": 0.1625, "step": 10750 }, { "epoch": 1.6684690251815233, "grad_norm": 2.6875, "learning_rate": 0.00011713463650053423, "loss": 0.1542, "step": 10800 }, { "epoch": 1.6761934188166228, "grad_norm": 1.453125, "learning_rate": 0.00011613759759456462, "loss": 0.1306, "step": 10850 }, { "epoch": 1.6839178124517225, "grad_norm": 0.89453125, "learning_rate": 0.00011513890880835776, "loss": 0.1784, "step": 10900 }, { "epoch": 1.6916422060868221, "grad_norm": 2.46875, "learning_rate": 0.00011413867224613843, "loss": 0.1384, "step": 10950 }, { "epoch": 1.6993665997219218, "grad_norm": 2.0, "learning_rate": 0.00011313699017037327, "loss": 0.1279, "step": 11000 }, { "epoch": 1.7070909933570215, "grad_norm": 0.90234375, "learning_rate": 0.00011213396499131581, "loss": 0.1615, "step": 11050 }, { "epoch": 1.7148153869921212, "grad_norm": 1.0, "learning_rate": 0.00011112969925653618, "loss": 0.1344, "step": 11100 }, { "epoch": 1.722539780627221, "grad_norm": 0.50390625, "learning_rate": 0.00011012429564043667, "loss": 0.155, "step": 11150 }, { "epoch": 1.7302641742623204, "grad_norm": 0.51953125, "learning_rate": 0.00010911785693375478, "loss": 0.1394, "step": 11200 }, { "epoch": 1.73798856789742, "grad_norm": 0.88671875, "learning_rate": 0.00010811048603305371, "loss": 0.1499, "step": 11250 }, { "epoch": 1.7457129615325195, "grad_norm": 1.2890625, "learning_rate": 0.00010710228593020271, "loss": 0.1247, "step": 11300 }, { "epoch": 1.7534373551676192, "grad_norm": 4.0, "learning_rate": 0.00010609335970184715, "loss": 0.1259, "step": 11350 }, { "epoch": 1.761161748802719, "grad_norm": 1.7421875, "learning_rate": 0.00010508381049887029, "loss": 0.1631, "step": 11400 }, { "epoch": 1.7688861424378186, "grad_norm": 1.0390625, "learning_rate": 0.00010407374153584716, "loss": 0.1414, "step": 11450 }, { "epoch": 1.7766105360729183, "grad_norm": 1.3125, "learning_rate": 0.00010306325608049227, "loss": 0.1276, "step": 11500 }, { "epoch": 1.784334929708018, "grad_norm": 1.5625, "learning_rate": 0.0001020524574431015, "loss": 0.1586, "step": 11550 }, { "epoch": 1.7920593233431177, "grad_norm": 2.734375, "learning_rate": 0.00010104144896598997, "loss": 0.1461, "step": 11600 }, { "epoch": 1.7997837169782174, "grad_norm": 0.7734375, "learning_rate": 0.00010003033401292642, "loss": 0.1247, "step": 11650 }, { "epoch": 1.8075081106133168, "grad_norm": 3.03125, "learning_rate": 9.901921595856549e-05, "loss": 0.1482, "step": 11700 }, { "epoch": 1.8152325042484165, "grad_norm": 1.1796875, "learning_rate": 9.800819817787906e-05, "loss": 0.1543, "step": 11750 }, { "epoch": 1.822956897883516, "grad_norm": 0.3828125, "learning_rate": 9.699738403558688e-05, "loss": 0.1474, "step": 11800 }, { "epoch": 1.8306812915186157, "grad_norm": 2.46875, "learning_rate": 9.59868768755894e-05, "loss": 0.139, "step": 11850 }, { "epoch": 1.8384056851537154, "grad_norm": 0.9921875, "learning_rate": 9.497678001040152e-05, "loss": 0.1368, "step": 11900 }, { "epoch": 1.846130078788815, "grad_norm": 1.6640625, "learning_rate": 9.396719671059039e-05, "loss": 0.1191, "step": 11950 }, { "epoch": 1.8538544724239148, "grad_norm": 1.0859375, "learning_rate": 9.295823019421707e-05, "loss": 0.1345, "step": 12000 }, { "epoch": 1.8615788660590145, "grad_norm": 1.0, "learning_rate": 9.194998361628373e-05, "loss": 0.1333, "step": 12050 }, { "epoch": 1.8693032596941141, "grad_norm": 2.40625, "learning_rate": 9.094256005818736e-05, "loss": 0.1419, "step": 12100 }, { "epoch": 1.8770276533292136, "grad_norm": 2.765625, "learning_rate": 8.993606251718072e-05, "loss": 0.1546, "step": 12150 }, { "epoch": 1.8847520469643133, "grad_norm": 3.109375, "learning_rate": 8.893059389584226e-05, "loss": 0.1493, "step": 12200 }, { "epoch": 1.892476440599413, "grad_norm": 3.515625, "learning_rate": 8.792625699155531e-05, "loss": 0.1421, "step": 12250 }, { "epoch": 1.9002008342345125, "grad_norm": 1.2421875, "learning_rate": 8.692315448599864e-05, "loss": 0.1322, "step": 12300 }, { "epoch": 1.9079252278696122, "grad_norm": 0.8046875, "learning_rate": 8.592138893464792e-05, "loss": 0.136, "step": 12350 }, { "epoch": 1.9156496215047119, "grad_norm": 2.21875, "learning_rate": 8.492106275629117e-05, "loss": 0.1334, "step": 12400 }, { "epoch": 1.9233740151398115, "grad_norm": 1.15625, "learning_rate": 8.392227822255723e-05, "loss": 0.1314, "step": 12450 }, { "epoch": 1.9310984087749112, "grad_norm": 0.390625, "learning_rate": 8.292513744745992e-05, "loss": 0.1123, "step": 12500 }, { "epoch": 1.938822802410011, "grad_norm": 3.921875, "learning_rate": 8.192974237695798e-05, "loss": 0.1132, "step": 12550 }, { "epoch": 1.9465471960451106, "grad_norm": 1.3984375, "learning_rate": 8.093619477853233e-05, "loss": 0.112, "step": 12600 }, { "epoch": 1.95427158968021, "grad_norm": 1.4609375, "learning_rate": 7.99445962307815e-05, "loss": 0.1074, "step": 12650 }, { "epoch": 1.9619959833153098, "grad_norm": 1.359375, "learning_rate": 7.89550481130364e-05, "loss": 0.1303, "step": 12700 }, { "epoch": 1.9697203769504092, "grad_norm": 0.39453125, "learning_rate": 7.796765159499558e-05, "loss": 0.1272, "step": 12750 }, { "epoch": 1.977444770585509, "grad_norm": 1.078125, "learning_rate": 7.698250762638167e-05, "loss": 0.1154, "step": 12800 }, { "epoch": 1.9851691642206086, "grad_norm": 3.734375, "learning_rate": 7.599971692662056e-05, "loss": 0.1245, "step": 12850 }, { "epoch": 1.9928935578557083, "grad_norm": 0.447265625, "learning_rate": 7.501937997454392e-05, "loss": 0.1121, "step": 12900 }, { "epoch": 2.000617951490808, "grad_norm": 1.578125, "learning_rate": 7.404159699811656e-05, "loss": 0.1051, "step": 12950 }, { "epoch": 2.0083423451259077, "grad_norm": 0.87109375, "learning_rate": 7.306646796418912e-05, "loss": 0.0843, "step": 13000 }, { "epoch": 2.0160667387610074, "grad_norm": 0.8359375, "learning_rate": 7.209409256827782e-05, "loss": 0.0768, "step": 13050 }, { "epoch": 2.023791132396107, "grad_norm": 0.98828125, "learning_rate": 7.112457022437155e-05, "loss": 0.0752, "step": 13100 }, { "epoch": 2.0315155260312068, "grad_norm": 1.2734375, "learning_rate": 7.015800005476818e-05, "loss": 0.0776, "step": 13150 }, { "epoch": 2.039239919666306, "grad_norm": 1.1015625, "learning_rate": 6.919448087994023e-05, "loss": 0.0857, "step": 13200 }, { "epoch": 2.0469643133014057, "grad_norm": 2.609375, "learning_rate": 6.823411120843188e-05, "loss": 0.0769, "step": 13250 }, { "epoch": 2.0546887069365054, "grad_norm": 0.80078125, "learning_rate": 6.727698922678745e-05, "loss": 0.0902, "step": 13300 }, { "epoch": 2.062413100571605, "grad_norm": 0.796875, "learning_rate": 6.632321278951299e-05, "loss": 0.0841, "step": 13350 }, { "epoch": 2.070137494206705, "grad_norm": 1.0546875, "learning_rate": 6.537287940907194e-05, "loss": 0.0771, "step": 13400 }, { "epoch": 2.0778618878418045, "grad_norm": 0.62890625, "learning_rate": 6.442608624591541e-05, "loss": 0.0785, "step": 13450 }, { "epoch": 2.085586281476904, "grad_norm": 1.609375, "learning_rate": 6.348293009854896e-05, "loss": 0.0795, "step": 13500 }, { "epoch": 2.093310675112004, "grad_norm": 3.1875, "learning_rate": 6.254350739363564e-05, "loss": 0.0742, "step": 13550 }, { "epoch": 2.1010350687471036, "grad_norm": 0.373046875, "learning_rate": 6.160791417613805e-05, "loss": 0.079, "step": 13600 }, { "epoch": 2.108759462382203, "grad_norm": 0.74609375, "learning_rate": 6.067624609949837e-05, "loss": 0.0717, "step": 13650 }, { "epoch": 2.1164838560173025, "grad_norm": 0.61328125, "learning_rate": 5.974859841585925e-05, "loss": 0.0743, "step": 13700 }, { "epoch": 2.124208249652402, "grad_norm": 2.171875, "learning_rate": 5.882506596632517e-05, "loss": 0.0726, "step": 13750 }, { "epoch": 2.131932643287502, "grad_norm": 0.94140625, "learning_rate": 5.790574317126625e-05, "loss": 0.0848, "step": 13800 }, { "epoch": 2.1396570369226016, "grad_norm": 0.58203125, "learning_rate": 5.699072402066468e-05, "loss": 0.0738, "step": 13850 }, { "epoch": 2.1473814305577013, "grad_norm": 1.3671875, "learning_rate": 5.608010206450549e-05, "loss": 0.0675, "step": 13900 }, { "epoch": 2.155105824192801, "grad_norm": 2.5, "learning_rate": 5.5173970403212174e-05, "loss": 0.0886, "step": 13950 }, { "epoch": 2.1628302178279006, "grad_norm": 1.2734375, "learning_rate": 5.427242167812803e-05, "loss": 0.0796, "step": 14000 }, { "epoch": 2.1705546114630003, "grad_norm": 0.70703125, "learning_rate": 5.337554806204497e-05, "loss": 0.0626, "step": 14050 }, { "epoch": 2.1782790050980996, "grad_norm": 1.3359375, "learning_rate": 5.2483441249779755e-05, "loss": 0.0686, "step": 14100 }, { "epoch": 2.1860033987331993, "grad_norm": 0.6796875, "learning_rate": 5.159619244879936e-05, "loss": 0.07, "step": 14150 }, { "epoch": 2.193727792368299, "grad_norm": 0.6640625, "learning_rate": 5.07138923698959e-05, "loss": 0.0739, "step": 14200 }, { "epoch": 2.2014521860033986, "grad_norm": 1.0, "learning_rate": 4.983663121791289e-05, "loss": 0.0724, "step": 14250 }, { "epoch": 2.2091765796384983, "grad_norm": 1.7578125, "learning_rate": 4.8964498682522374e-05, "loss": 0.071, "step": 14300 }, { "epoch": 2.216900973273598, "grad_norm": 0.9296875, "learning_rate": 4.80975839290556e-05, "loss": 0.0729, "step": 14350 }, { "epoch": 2.2246253669086977, "grad_norm": 2.453125, "learning_rate": 4.723597558938672e-05, "loss": 0.0758, "step": 14400 }, { "epoch": 2.2323497605437974, "grad_norm": 1.0234375, "learning_rate": 4.637976175287129e-05, "loss": 0.0639, "step": 14450 }, { "epoch": 2.240074154178897, "grad_norm": 1.90625, "learning_rate": 4.552902995734003e-05, "loss": 0.0739, "step": 14500 }, { "epoch": 2.247798547813997, "grad_norm": 0.357421875, "learning_rate": 4.468386718014934e-05, "loss": 0.0634, "step": 14550 }, { "epoch": 2.2555229414490965, "grad_norm": 1.296875, "learning_rate": 4.384435982928871e-05, "loss": 0.0658, "step": 14600 }, { "epoch": 2.2632473350841957, "grad_norm": 2.59375, "learning_rate": 4.30105937345466e-05, "loss": 0.07, "step": 14650 }, { "epoch": 2.2709717287192954, "grad_norm": 1.7578125, "learning_rate": 4.218265413873519e-05, "loss": 0.0671, "step": 14700 }, { "epoch": 2.278696122354395, "grad_norm": 0.9765625, "learning_rate": 4.136062568897556e-05, "loss": 0.081, "step": 14750 }, { "epoch": 2.286420515989495, "grad_norm": 0.50390625, "learning_rate": 4.054459242804338e-05, "loss": 0.0746, "step": 14800 }, { "epoch": 2.2941449096245945, "grad_norm": 0.5703125, "learning_rate": 3.973463778577634e-05, "loss": 0.0755, "step": 14850 }, { "epoch": 2.301869303259694, "grad_norm": 0.484375, "learning_rate": 3.89308445705449e-05, "loss": 0.0626, "step": 14900 }, { "epoch": 2.309593696894794, "grad_norm": 0.60546875, "learning_rate": 3.813329496078553e-05, "loss": 0.0772, "step": 14950 }, { "epoch": 2.3173180905298936, "grad_norm": 0.3984375, "learning_rate": 3.734207049659937e-05, "loss": 0.0705, "step": 15000 }, { "epoch": 2.3250424841649933, "grad_norm": 0.482421875, "learning_rate": 3.65572520714155e-05, "loss": 0.0748, "step": 15050 }, { "epoch": 2.3327668778000925, "grad_norm": 0.67578125, "learning_rate": 3.577891992372066e-05, "loss": 0.0687, "step": 15100 }, { "epoch": 2.340491271435192, "grad_norm": 1.46875, "learning_rate": 3.5007153628855546e-05, "loss": 0.063, "step": 15150 }, { "epoch": 2.348215665070292, "grad_norm": 0.73046875, "learning_rate": 3.4242032090879536e-05, "loss": 0.0662, "step": 15200 }, { "epoch": 2.3559400587053916, "grad_norm": 0.5625, "learning_rate": 3.3483633534503413e-05, "loss": 0.0731, "step": 15250 }, { "epoch": 2.3636644523404913, "grad_norm": 1.34375, "learning_rate": 3.27320354970919e-05, "loss": 0.0733, "step": 15300 }, { "epoch": 2.371388845975591, "grad_norm": 0.77734375, "learning_rate": 3.1987314820736257e-05, "loss": 0.0645, "step": 15350 }, { "epoch": 2.3791132396106907, "grad_norm": 0.515625, "learning_rate": 3.12495476443983e-05, "loss": 0.0822, "step": 15400 }, { "epoch": 2.3868376332457903, "grad_norm": 0.486328125, "learning_rate": 3.051880939612592e-05, "loss": 0.0621, "step": 15450 }, { "epoch": 2.39456202688089, "grad_norm": 1.4296875, "learning_rate": 2.9795174785341384e-05, "loss": 0.0744, "step": 15500 }, { "epoch": 2.4022864205159893, "grad_norm": 0.4453125, "learning_rate": 2.9078717795203482e-05, "loss": 0.0648, "step": 15550 }, { "epoch": 2.410010814151089, "grad_norm": 0.6640625, "learning_rate": 2.836951167504317e-05, "loss": 0.0699, "step": 15600 }, { "epoch": 2.4177352077861887, "grad_norm": 1.1875, "learning_rate": 2.766762893287509e-05, "loss": 0.0666, "step": 15650 }, { "epoch": 2.4254596014212884, "grad_norm": 1.015625, "learning_rate": 2.6973141327984232e-05, "loss": 0.066, "step": 15700 }, { "epoch": 2.433183995056388, "grad_norm": 0.671875, "learning_rate": 2.6286119863589532e-05, "loss": 0.0709, "step": 15750 }, { "epoch": 2.4409083886914877, "grad_norm": 0.8515625, "learning_rate": 2.5606634779584414e-05, "loss": 0.0598, "step": 15800 }, { "epoch": 2.4486327823265874, "grad_norm": 1.078125, "learning_rate": 2.4934755545355926e-05, "loss": 0.0824, "step": 15850 }, { "epoch": 2.456357175961687, "grad_norm": 0.49609375, "learning_rate": 2.4270550852681916e-05, "loss": 0.067, "step": 15900 }, { "epoch": 2.464081569596787, "grad_norm": 0.59375, "learning_rate": 2.361408860870843e-05, "loss": 0.0591, "step": 15950 }, { "epoch": 2.471805963231886, "grad_norm": 1.09375, "learning_rate": 2.296543592900683e-05, "loss": 0.0662, "step": 16000 }, { "epoch": 2.4795303568669858, "grad_norm": 2.265625, "learning_rate": 2.2324659130712e-05, "loss": 0.0623, "step": 16050 }, { "epoch": 2.4872547505020854, "grad_norm": 0.84375, "learning_rate": 2.169182372574231e-05, "loss": 0.0692, "step": 16100 }, { "epoch": 2.494979144137185, "grad_norm": 0.3828125, "learning_rate": 2.1066994414101703e-05, "loss": 0.0577, "step": 16150 }, { "epoch": 2.502703537772285, "grad_norm": 2.140625, "learning_rate": 2.0450235077264955e-05, "loss": 0.0685, "step": 16200 }, { "epoch": 2.5104279314073845, "grad_norm": 0.85546875, "learning_rate": 1.984160877164637e-05, "loss": 0.0647, "step": 16250 }, { "epoch": 2.518152325042484, "grad_norm": 0.48046875, "learning_rate": 1.9241177722153235e-05, "loss": 0.0687, "step": 16300 }, { "epoch": 2.525876718677584, "grad_norm": 1.640625, "learning_rate": 1.8649003315823887e-05, "loss": 0.068, "step": 16350 }, { "epoch": 2.5336011123126836, "grad_norm": 1.09375, "learning_rate": 1.80651460955517e-05, "loss": 0.0532, "step": 16400 }, { "epoch": 2.541325505947783, "grad_norm": 0.380859375, "learning_rate": 1.7489665753895146e-05, "loss": 0.0681, "step": 16450 }, { "epoch": 2.549049899582883, "grad_norm": 0.9140625, "learning_rate": 1.6922621126975193e-05, "loss": 0.063, "step": 16500 }, { "epoch": 2.556774293217982, "grad_norm": 0.5546875, "learning_rate": 1.6364070188459646e-05, "loss": 0.07, "step": 16550 }, { "epoch": 2.564498686853082, "grad_norm": 0.384765625, "learning_rate": 1.5814070043636342e-05, "loss": 0.0799, "step": 16600 }, { "epoch": 2.5722230804881816, "grad_norm": 0.62890625, "learning_rate": 1.5272676923574637e-05, "loss": 0.0697, "step": 16650 }, { "epoch": 2.5799474741232813, "grad_norm": 0.3671875, "learning_rate": 1.473994617937643e-05, "loss": 0.0637, "step": 16700 }, { "epoch": 2.587671867758381, "grad_norm": 0.5546875, "learning_rate": 1.4215932276517253e-05, "loss": 0.0626, "step": 16750 }, { "epoch": 2.5953962613934807, "grad_norm": 0.7421875, "learning_rate": 1.3700688789277738e-05, "loss": 0.0735, "step": 16800 }, { "epoch": 2.6031206550285804, "grad_norm": 0.97265625, "learning_rate": 1.319426839526633e-05, "loss": 0.062, "step": 16850 }, { "epoch": 2.61084504866368, "grad_norm": 0.578125, "learning_rate": 1.2696722870033529e-05, "loss": 0.0632, "step": 16900 }, { "epoch": 2.6185694422987797, "grad_norm": 1.09375, "learning_rate": 1.2208103081778555e-05, "loss": 0.0677, "step": 16950 }, { "epoch": 2.626293835933879, "grad_norm": 1.3046875, "learning_rate": 1.1728458986148627e-05, "loss": 0.0605, "step": 17000 }, { "epoch": 2.6340182295689787, "grad_norm": 1.0234375, "learning_rate": 1.1257839621131584e-05, "loss": 0.0606, "step": 17050 }, { "epoch": 2.6417426232040784, "grad_norm": 0.8046875, "learning_rate": 1.0796293102042243e-05, "loss": 0.0641, "step": 17100 }, { "epoch": 2.649467016839178, "grad_norm": 0.64453125, "learning_rate": 1.034386661660337e-05, "loss": 0.0632, "step": 17150 }, { "epoch": 2.6571914104742778, "grad_norm": 1.21875, "learning_rate": 9.900606420121062e-06, "loss": 0.0671, "step": 17200 }, { "epoch": 2.6649158041093775, "grad_norm": 0.81640625, "learning_rate": 9.466557830755862e-06, "loss": 0.0646, "step": 17250 }, { "epoch": 2.672640197744477, "grad_norm": 0.91015625, "learning_rate": 9.041765224889399e-06, "loss": 0.0674, "step": 17300 }, { "epoch": 2.680364591379577, "grad_norm": 1.1015625, "learning_rate": 8.62627203258748e-06, "loss": 0.058, "step": 17350 }, { "epoch": 2.6880889850146765, "grad_norm": 1.09375, "learning_rate": 8.220120733159831e-06, "loss": 0.0563, "step": 17400 }, { "epoch": 2.6958133786497758, "grad_norm": 0.91796875, "learning_rate": 7.823352850817124e-06, "loss": 0.0673, "step": 17450 }, { "epoch": 2.703537772284876, "grad_norm": 3.546875, "learning_rate": 7.4360089504256215e-06, "loss": 0.0664, "step": 17500 }, { "epoch": 2.711262165919975, "grad_norm": 0.78125, "learning_rate": 7.058128633359806e-06, "loss": 0.0691, "step": 17550 }, { "epoch": 2.718986559555075, "grad_norm": 1.6484375, "learning_rate": 6.689750533453765e-06, "loss": 0.0606, "step": 17600 }, { "epoch": 2.7267109531901745, "grad_norm": 0.609375, "learning_rate": 6.33091231305114e-06, "loss": 0.0665, "step": 17650 }, { "epoch": 2.7344353468252742, "grad_norm": 0.67578125, "learning_rate": 5.9816506591547475e-06, "loss": 0.0623, "step": 17700 }, { "epoch": 2.742159740460374, "grad_norm": 0.72265625, "learning_rate": 5.6420012796757015e-06, "loss": 0.0665, "step": 17750 }, { "epoch": 2.7498841340954736, "grad_norm": 0.625, "learning_rate": 5.311998899782711e-06, "loss": 0.057, "step": 17800 }, { "epoch": 2.7576085277305733, "grad_norm": 0.392578125, "learning_rate": 4.99167725835179e-06, "loss": 0.0732, "step": 17850 }, { "epoch": 2.7653329213656725, "grad_norm": 0.8671875, "learning_rate": 4.681069104516922e-06, "loss": 0.0653, "step": 17900 }, { "epoch": 2.7730573150007727, "grad_norm": 1.3046875, "learning_rate": 4.380206194321812e-06, "loss": 0.0636, "step": 17950 }, { "epoch": 2.780781708635872, "grad_norm": 1.6875, "learning_rate": 4.089119287473242e-06, "loss": 0.0631, "step": 18000 }, { "epoch": 2.7885061022709716, "grad_norm": 0.59375, "learning_rate": 3.8078381441961407e-06, "loss": 0.063, "step": 18050 }, { "epoch": 2.7962304959060713, "grad_norm": 0.796875, "learning_rate": 3.536391522191096e-06, "loss": 0.0797, "step": 18100 }, { "epoch": 2.803954889541171, "grad_norm": 1.234375, "learning_rate": 3.2748071736941164e-06, "loss": 0.0595, "step": 18150 }, { "epoch": 2.8116792831762707, "grad_norm": 0.6015625, "learning_rate": 3.0231118426393035e-06, "loss": 0.0596, "step": 18200 }, { "epoch": 2.8194036768113704, "grad_norm": 1.828125, "learning_rate": 2.78133126192468e-06, "loss": 0.0642, "step": 18250 }, { "epoch": 2.82712807044647, "grad_norm": 0.96484375, "learning_rate": 2.5494901507811418e-06, "loss": 0.0549, "step": 18300 }, { "epoch": 2.8348524640815693, "grad_norm": 1.8359375, "learning_rate": 2.327612212245389e-06, "loss": 0.0636, "step": 18350 }, { "epoch": 2.8425768577166695, "grad_norm": 0.89453125, "learning_rate": 2.1157201307364315e-06, "loss": 0.0667, "step": 18400 }, { "epoch": 2.8503012513517687, "grad_norm": 0.49609375, "learning_rate": 1.913835569736444e-06, "loss": 0.0613, "step": 18450 }, { "epoch": 2.8580256449868684, "grad_norm": 1.0703125, "learning_rate": 1.7219791695758824e-06, "loss": 0.071, "step": 18500 }, { "epoch": 2.865750038621968, "grad_norm": 0.51171875, "learning_rate": 1.5401705453233051e-06, "loss": 0.076, "step": 18550 }, { "epoch": 2.873474432257068, "grad_norm": 0.39453125, "learning_rate": 1.3684282847799213e-06, "loss": 0.0667, "step": 18600 }, { "epoch": 2.8811988258921675, "grad_norm": 0.703125, "learning_rate": 1.2067699465792336e-06, "loss": 0.0651, "step": 18650 }, { "epoch": 2.888923219527267, "grad_norm": 0.455078125, "learning_rate": 1.0552120583918412e-06, "loss": 0.0575, "step": 18700 }, { "epoch": 2.896647613162367, "grad_norm": 1.2734375, "learning_rate": 9.137701152356792e-07, "loss": 0.0564, "step": 18750 }, { "epoch": 2.9043720067974665, "grad_norm": 0.52734375, "learning_rate": 7.824585778919091e-07, "loss": 0.0699, "step": 18800 }, { "epoch": 2.9120964004325662, "grad_norm": 0.67578125, "learning_rate": 6.612908714263455e-07, "loss": 0.0628, "step": 18850 }, { "epoch": 2.9198207940676655, "grad_norm": 0.443359375, "learning_rate": 5.502793838170428e-07, "loss": 0.0642, "step": 18900 }, { "epoch": 2.927545187702765, "grad_norm": 0.435546875, "learning_rate": 4.49435464687642e-07, "loss": 0.0642, "step": 18950 }, { "epoch": 2.935269581337865, "grad_norm": 0.5625, "learning_rate": 3.587694241470874e-07, "loss": 0.0614, "step": 19000 }, { "epoch": 2.9429939749729646, "grad_norm": 0.412109375, "learning_rate": 2.7829053173550333e-07, "loss": 0.0658, "step": 19050 }, { "epoch": 2.9507183686080642, "grad_norm": 0.419921875, "learning_rate": 2.0800701547651902e-07, "loss": 0.0675, "step": 19100 }, { "epoch": 2.958442762243164, "grad_norm": 1.3359375, "learning_rate": 1.479260610360078e-07, "loss": 0.0703, "step": 19150 }, { "epoch": 2.9661671558782636, "grad_norm": 0.39453125, "learning_rate": 9.805381098746402e-08, "loss": 0.0703, "step": 19200 }, { "epoch": 2.9738915495133633, "grad_norm": 1.171875, "learning_rate": 5.839536418401625e-08, "loss": 0.0633, "step": 19250 }, { "epoch": 2.981615943148463, "grad_norm": 2.765625, "learning_rate": 2.8954775237088894e-08, "loss": 0.0634, "step": 19300 }, { "epoch": 2.9893403367835623, "grad_norm": 1.6328125, "learning_rate": 9.735054101878138e-09, "loss": 0.058, "step": 19350 }, { "epoch": 2.9970647304186624, "grad_norm": 0.9296875, "learning_rate": 7.381657696425848e-10, "loss": 0.0586, "step": 19400 } ], "logging_steps": 50, "max_steps": 19419, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 0, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.3806574237646848e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }