|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 5.103125664469488, |
|
"global_step": 24000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.771793054571226e-06, |
|
"loss": 11.8316, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.543586109142452e-06, |
|
"loss": 8.5077, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.3153791637136785e-06, |
|
"loss": 6.6343, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7.087172218284904e-06, |
|
"loss": 6.0709, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.85896527285613e-06, |
|
"loss": 5.5047, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.0630758327427357e-05, |
|
"loss": 4.9657, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.2402551381998583e-05, |
|
"loss": 4.5669, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.4174344436569808e-05, |
|
"loss": 4.0142, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.5946137491141037e-05, |
|
"loss": 3.5509, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.771793054571226e-05, |
|
"loss": 2.9313, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.948972360028349e-05, |
|
"loss": 2.3465, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 2.1261516654854714e-05, |
|
"loss": 1.7314, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 2.303330970942594e-05, |
|
"loss": 1.2661, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 2.4805102763997165e-05, |
|
"loss": 1.0417, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 2.6576895818568394e-05, |
|
"loss": 0.9334, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 2.8348688873139616e-05, |
|
"loss": 0.8874, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 3.012048192771085e-05, |
|
"loss": 0.8745, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 3.1892274982282074e-05, |
|
"loss": 0.8545, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 3.36640680368533e-05, |
|
"loss": 0.8327, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 3.543586109142452e-05, |
|
"loss": 0.7961, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 3.720765414599575e-05, |
|
"loss": 0.7932, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 3.897944720056698e-05, |
|
"loss": 0.7677, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.07512402551382e-05, |
|
"loss": 0.7398, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.252303330970943e-05, |
|
"loss": 0.719, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.4294826364280654e-05, |
|
"loss": 0.7188, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.606661941885188e-05, |
|
"loss": 0.7027, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.7838412473423105e-05, |
|
"loss": 0.688, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.961020552799433e-05, |
|
"loss": 0.6751, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.992725780579699e-05, |
|
"loss": 0.6552, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.9833998582459806e-05, |
|
"loss": 0.6749, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.974073935912262e-05, |
|
"loss": 0.6607, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.9647480135785426e-05, |
|
"loss": 0.6949, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.955422091244825e-05, |
|
"loss": 0.6694, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.946096168911106e-05, |
|
"loss": 0.6725, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.936770246577387e-05, |
|
"loss": 0.6419, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.927444324243668e-05, |
|
"loss": 0.6336, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.9181184019099494e-05, |
|
"loss": 0.6478, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.90879247957623e-05, |
|
"loss": 0.6395, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.8994665572425114e-05, |
|
"loss": 0.6237, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.890140634908793e-05, |
|
"loss": 0.623, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.880814712575074e-05, |
|
"loss": 0.6267, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.8714887902413555e-05, |
|
"loss": 0.6271, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.862162867907636e-05, |
|
"loss": 0.6038, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.8528369455739175e-05, |
|
"loss": 0.5901, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.843511023240199e-05, |
|
"loss": 0.6193, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.8341851009064796e-05, |
|
"loss": 0.5823, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.824859178572761e-05, |
|
"loss": 0.5939, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.815533256239042e-05, |
|
"loss": 0.5954, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.8062073339053236e-05, |
|
"loss": 0.5818, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.796881411571605e-05, |
|
"loss": 0.5803, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.787555489237886e-05, |
|
"loss": 0.5652, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.778229566904167e-05, |
|
"loss": 0.5845, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.7689036445704483e-05, |
|
"loss": 0.5669, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.75957772223673e-05, |
|
"loss": 0.5835, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.7502517999030104e-05, |
|
"loss": 0.5977, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.740925877569292e-05, |
|
"loss": 0.5749, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 4.731599955235573e-05, |
|
"loss": 0.5512, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 4.7222740329018544e-05, |
|
"loss": 0.5496, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 4.712948110568136e-05, |
|
"loss": 0.5624, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 4.7036221882344165e-05, |
|
"loss": 0.5474, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 4.694296265900698e-05, |
|
"loss": 0.5573, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 4.684970343566979e-05, |
|
"loss": 0.5258, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 4.67564442123326e-05, |
|
"loss": 0.5418, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.666318498899541e-05, |
|
"loss": 0.5513, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.6569925765658226e-05, |
|
"loss": 0.5413, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.647666654232104e-05, |
|
"loss": 0.557, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.638340731898385e-05, |
|
"loss": 0.5425, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.629014809564666e-05, |
|
"loss": 0.5292, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.619688887230947e-05, |
|
"loss": 0.5297, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 4.6103629648972286e-05, |
|
"loss": 0.5469, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 4.601037042563509e-05, |
|
"loss": 0.5485, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.591711120229791e-05, |
|
"loss": 0.5543, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.582385197896072e-05, |
|
"loss": 0.5603, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.5730592755623534e-05, |
|
"loss": 0.5547, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 4.563733353228635e-05, |
|
"loss": 0.5245, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 4.554407430894916e-05, |
|
"loss": 0.5307, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 4.545081508561197e-05, |
|
"loss": 0.5329, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 4.535755586227478e-05, |
|
"loss": 0.5214, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 4.5264296638937595e-05, |
|
"loss": 0.5135, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 4.51710374156004e-05, |
|
"loss": 0.5261, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.5077778192263215e-05, |
|
"loss": 0.5193, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.498451896892603e-05, |
|
"loss": 0.5196, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 4.489125974558884e-05, |
|
"loss": 0.5249, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 4.4798000522251656e-05, |
|
"loss": 0.5147, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 4.470474129891446e-05, |
|
"loss": 0.5277, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.4611482075577276e-05, |
|
"loss": 0.509, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 4.451822285224009e-05, |
|
"loss": 0.5099, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 4.4424963628902896e-05, |
|
"loss": 0.5178, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 4.433170440556571e-05, |
|
"loss": 0.49, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 4.423844518222852e-05, |
|
"loss": 0.4989, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.414518595889134e-05, |
|
"loss": 0.5077, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 4.405192673555415e-05, |
|
"loss": 0.4866, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 4.3958667512216964e-05, |
|
"loss": 0.5038, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 4.386540828887977e-05, |
|
"loss": 0.4938, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 4.3772149065542584e-05, |
|
"loss": 0.516, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 4.36788898422054e-05, |
|
"loss": 0.4992, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 4.3585630618868204e-05, |
|
"loss": 0.4833, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 4.349237139553102e-05, |
|
"loss": 0.5047, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 4.339911217219383e-05, |
|
"loss": 0.4872, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 4.3305852948856645e-05, |
|
"loss": 0.5096, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 4.321259372551946e-05, |
|
"loss": 0.4837, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 4.3119334502182265e-05, |
|
"loss": 0.5084, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 4.302607527884508e-05, |
|
"loss": 0.5009, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 4.293281605550789e-05, |
|
"loss": 0.4796, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 4.28395568321707e-05, |
|
"loss": 0.4956, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 4.274629760883351e-05, |
|
"loss": 0.495, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 4.265303838549633e-05, |
|
"loss": 0.4797, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 4.255977916215914e-05, |
|
"loss": 0.5058, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 4.246651993882195e-05, |
|
"loss": 0.5026, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 4.237326071548477e-05, |
|
"loss": 0.4923, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 4.2280001492147573e-05, |
|
"loss": 0.48, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 4.218674226881039e-05, |
|
"loss": 0.4825, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 4.20934830454732e-05, |
|
"loss": 0.4714, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 4.200022382213601e-05, |
|
"loss": 0.4785, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 4.190696459879883e-05, |
|
"loss": 0.4794, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 4.1813705375461634e-05, |
|
"loss": 0.4608, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 4.172044615212445e-05, |
|
"loss": 0.4848, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 4.162718692878726e-05, |
|
"loss": 0.4667, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 4.153392770545007e-05, |
|
"loss": 0.4831, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 4.144066848211288e-05, |
|
"loss": 0.4827, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 4.1347409258775695e-05, |
|
"loss": 0.4781, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 4.12541500354385e-05, |
|
"loss": 0.4832, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 4.116089081210132e-05, |
|
"loss": 0.4759, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 4.1067631588764136e-05, |
|
"loss": 0.4902, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 4.097437236542694e-05, |
|
"loss": 0.4718, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 4.0881113142089756e-05, |
|
"loss": 0.4789, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 4.078785391875257e-05, |
|
"loss": 0.4927, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 4.0694594695415376e-05, |
|
"loss": 0.4849, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 4.060133547207819e-05, |
|
"loss": 0.4829, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 4.0508076248741003e-05, |
|
"loss": 0.4608, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 4.041481702540381e-05, |
|
"loss": 0.4457, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 4.032155780206663e-05, |
|
"loss": 0.4656, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 4.022829857872944e-05, |
|
"loss": 0.4575, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 4.013503935539225e-05, |
|
"loss": 0.4783, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 4.0041780132055064e-05, |
|
"loss": 0.4552, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 3.994852090871787e-05, |
|
"loss": 0.471, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 3.9855261685380685e-05, |
|
"loss": 0.4664, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 3.97620024620435e-05, |
|
"loss": 0.4632, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 3.9668743238706305e-05, |
|
"loss": 0.4694, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 3.9575484015369125e-05, |
|
"loss": 0.467, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 3.948222479203194e-05, |
|
"loss": 0.4647, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 3.9388965568694745e-05, |
|
"loss": 0.4709, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 3.929570634535756e-05, |
|
"loss": 0.4626, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 3.920244712202037e-05, |
|
"loss": 0.4739, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 3.910918789868318e-05, |
|
"loss": 0.4663, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 3.901592867534599e-05, |
|
"loss": 0.4595, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 3.8922669452008806e-05, |
|
"loss": 0.4559, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 3.882941022867162e-05, |
|
"loss": 0.4445, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 3.8736151005334433e-05, |
|
"loss": 0.4566, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 3.864289178199724e-05, |
|
"loss": 0.463, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 3.8549632558660054e-05, |
|
"loss": 0.4581, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 3.845637333532287e-05, |
|
"loss": 0.4501, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 3.8363114111985674e-05, |
|
"loss": 0.4532, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 3.826985488864849e-05, |
|
"loss": 0.4509, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 3.81765956653113e-05, |
|
"loss": 0.4495, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 3.8083336441974115e-05, |
|
"loss": 0.456, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 3.799007721863693e-05, |
|
"loss": 0.4626, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 3.7896817995299735e-05, |
|
"loss": 0.4543, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.780355877196255e-05, |
|
"loss": 0.4517, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 3.771029954862536e-05, |
|
"loss": 0.4389, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 3.761704032528817e-05, |
|
"loss": 0.4449, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 3.752378110195098e-05, |
|
"loss": 0.449, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 3.7430521878613796e-05, |
|
"loss": 0.4399, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.733726265527661e-05, |
|
"loss": 0.4434, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.724400343193942e-05, |
|
"loss": 0.4484, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.7150744208602236e-05, |
|
"loss": 0.4711, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.705748498526504e-05, |
|
"loss": 0.4502, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.696422576192786e-05, |
|
"loss": 0.4454, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.687096653859067e-05, |
|
"loss": 0.4556, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 3.677770731525348e-05, |
|
"loss": 0.4426, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 3.668444809191629e-05, |
|
"loss": 0.4454, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 3.6591188868579104e-05, |
|
"loss": 0.4474, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 3.649792964524192e-05, |
|
"loss": 0.4523, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 3.640467042190473e-05, |
|
"loss": 0.4426, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 3.631141119856754e-05, |
|
"loss": 0.4329, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 3.621815197523035e-05, |
|
"loss": 0.4364, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 3.6124892751893165e-05, |
|
"loss": 0.4405, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 3.603163352855597e-05, |
|
"loss": 0.4351, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 3.5938374305218785e-05, |
|
"loss": 0.4426, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 3.58451150818816e-05, |
|
"loss": 0.4389, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 3.575185585854441e-05, |
|
"loss": 0.433, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 3.5658596635207226e-05, |
|
"loss": 0.4389, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 3.556533741187004e-05, |
|
"loss": 0.4375, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 3.5472078188532846e-05, |
|
"loss": 0.4467, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 3.537881896519566e-05, |
|
"loss": 0.4402, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 3.528555974185847e-05, |
|
"loss": 0.4522, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 3.519230051852128e-05, |
|
"loss": 0.4408, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 3.5099041295184093e-05, |
|
"loss": 0.4273, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 3.500578207184691e-05, |
|
"loss": 0.4579, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 3.491252284850972e-05, |
|
"loss": 0.4349, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 3.4819263625172534e-05, |
|
"loss": 0.434, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 3.472600440183534e-05, |
|
"loss": 0.4396, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 3.4632745178498154e-05, |
|
"loss": 0.4251, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 3.453948595516097e-05, |
|
"loss": 0.4495, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 3.4446226731823775e-05, |
|
"loss": 0.4418, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 3.435296750848659e-05, |
|
"loss": 0.4367, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 3.425970828514941e-05, |
|
"loss": 0.4233, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 3.4166449061812215e-05, |
|
"loss": 0.4383, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 3.407318983847503e-05, |
|
"loss": 0.4368, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 3.397993061513784e-05, |
|
"loss": 0.4389, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 3.388667139180065e-05, |
|
"loss": 0.4226, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 3.379341216846346e-05, |
|
"loss": 0.4251, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 3.3700152945126276e-05, |
|
"loss": 0.4266, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 3.360689372178908e-05, |
|
"loss": 0.418, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 3.3513634498451896e-05, |
|
"loss": 0.4367, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 3.342037527511471e-05, |
|
"loss": 0.4311, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 3.3327116051777523e-05, |
|
"loss": 0.4308, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 3.323385682844034e-05, |
|
"loss": 0.4272, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 3.3140597605103144e-05, |
|
"loss": 0.4381, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 3.304733838176596e-05, |
|
"loss": 0.4253, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 3.295407915842877e-05, |
|
"loss": 0.4238, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 3.286081993509158e-05, |
|
"loss": 0.4157, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 3.276756071175439e-05, |
|
"loss": 0.4425, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 3.267430148841721e-05, |
|
"loss": 0.4352, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 3.258104226508002e-05, |
|
"loss": 0.4191, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 3.248778304174283e-05, |
|
"loss": 0.4153, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 3.2394523818405645e-05, |
|
"loss": 0.4139, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 3.230126459506845e-05, |
|
"loss": 0.4284, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 3.2208005371731265e-05, |
|
"loss": 0.4046, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 3.211474614839408e-05, |
|
"loss": 0.4049, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 3.2021486925056886e-05, |
|
"loss": 0.4325, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 3.1928227701719706e-05, |
|
"loss": 0.4308, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 3.183496847838251e-05, |
|
"loss": 0.4257, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 3.1741709255045326e-05, |
|
"loss": 0.4305, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 3.164845003170814e-05, |
|
"loss": 0.4296, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 3.155519080837095e-05, |
|
"loss": 0.4145, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 3.146193158503376e-05, |
|
"loss": 0.4175, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 3.1368672361696574e-05, |
|
"loss": 0.4275, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 3.127541313835938e-05, |
|
"loss": 0.4268, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 3.11821539150222e-05, |
|
"loss": 0.4368, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 3.1088894691685014e-05, |
|
"loss": 0.4141, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 3.099563546834782e-05, |
|
"loss": 0.4194, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 3.0902376245010635e-05, |
|
"loss": 0.4042, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 3.080911702167345e-05, |
|
"loss": 0.4298, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 3.0715857798336255e-05, |
|
"loss": 0.4079, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 3.062259857499907e-05, |
|
"loss": 0.3971, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 3.052933935166188e-05, |
|
"loss": 0.408, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 3.0436080128324695e-05, |
|
"loss": 0.4218, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 3.0342820904987506e-05, |
|
"loss": 0.416, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 3.024956168165032e-05, |
|
"loss": 0.4176, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 3.015630245831313e-05, |
|
"loss": 0.4255, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 3.006304323497594e-05, |
|
"loss": 0.4243, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 2.9969784011638753e-05, |
|
"loss": 0.4186, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 2.9876524788301563e-05, |
|
"loss": 0.4252, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 2.9783265564964373e-05, |
|
"loss": 0.4235, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 2.9690006341627187e-05, |
|
"loss": 0.4192, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 2.9596747118290004e-05, |
|
"loss": 0.4281, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 2.9503487894952814e-05, |
|
"loss": 0.4263, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 2.9410228671615624e-05, |
|
"loss": 0.416, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 2.9316969448278437e-05, |
|
"loss": 0.4239, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 2.9223710224941248e-05, |
|
"loss": 0.4011, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.9130451001604058e-05, |
|
"loss": 0.4067, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.903719177826687e-05, |
|
"loss": 0.4131, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.894393255492968e-05, |
|
"loss": 0.4166, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.88506733315925e-05, |
|
"loss": 0.3927, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 2.875741410825531e-05, |
|
"loss": 0.4247, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 2.8664154884918122e-05, |
|
"loss": 0.4248, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 2.8570895661580932e-05, |
|
"loss": 0.4083, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 2.8477636438243742e-05, |
|
"loss": 0.4248, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 2.8384377214906556e-05, |
|
"loss": 0.4268, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 2.8291117991569366e-05, |
|
"loss": 0.4194, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 2.8197858768232176e-05, |
|
"loss": 0.43, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 2.8104599544894993e-05, |
|
"loss": 0.4088, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 2.8011340321557807e-05, |
|
"loss": 0.4109, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 2.7918081098220617e-05, |
|
"loss": 0.4114, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 2.7824821874883427e-05, |
|
"loss": 0.4075, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 2.773156265154624e-05, |
|
"loss": 0.4106, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 2.763830342820905e-05, |
|
"loss": 0.4024, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 2.754504420487186e-05, |
|
"loss": 0.4268, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 2.7451784981534674e-05, |
|
"loss": 0.4031, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 2.735852575819749e-05, |
|
"loss": 0.4139, |
|
"step": 13550 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 2.72652665348603e-05, |
|
"loss": 0.4121, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 2.717200731152311e-05, |
|
"loss": 0.4, |
|
"step": 13650 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 2.7078748088185925e-05, |
|
"loss": 0.3982, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 2.6985488864848735e-05, |
|
"loss": 0.4128, |
|
"step": 13750 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 2.6892229641511545e-05, |
|
"loss": 0.4064, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 2.679897041817436e-05, |
|
"loss": 0.4245, |
|
"step": 13850 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 2.670571119483717e-05, |
|
"loss": 0.4022, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 2.6612451971499986e-05, |
|
"loss": 0.4249, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 2.6519192748162796e-05, |
|
"loss": 0.4185, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 2.642593352482561e-05, |
|
"loss": 0.3955, |
|
"step": 14050 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 2.633267430148842e-05, |
|
"loss": 0.4135, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 2.623941507815123e-05, |
|
"loss": 0.4045, |
|
"step": 14150 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 2.6146155854814043e-05, |
|
"loss": 0.4041, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 2.6052896631476854e-05, |
|
"loss": 0.4, |
|
"step": 14250 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 2.5959637408139664e-05, |
|
"loss": 0.4053, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 2.5866378184802477e-05, |
|
"loss": 0.3975, |
|
"step": 14350 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 2.577311896146529e-05, |
|
"loss": 0.4006, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 2.5679859738128104e-05, |
|
"loss": 0.407, |
|
"step": 14450 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 2.5586600514790914e-05, |
|
"loss": 0.4176, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 2.5493341291453725e-05, |
|
"loss": 0.4014, |
|
"step": 14550 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 2.5400082068116538e-05, |
|
"loss": 0.4111, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 2.5306822844779348e-05, |
|
"loss": 0.3951, |
|
"step": 14650 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 2.521356362144216e-05, |
|
"loss": 0.403, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 2.5120304398104972e-05, |
|
"loss": 0.4006, |
|
"step": 14750 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 2.502704517476779e-05, |
|
"loss": 0.4062, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 2.4933785951430596e-05, |
|
"loss": 0.4242, |
|
"step": 14850 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 2.484052672809341e-05, |
|
"loss": 0.3955, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 2.4747267504756223e-05, |
|
"loss": 0.4014, |
|
"step": 14950 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 2.4654008281419033e-05, |
|
"loss": 0.4073, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 2.4560749058081843e-05, |
|
"loss": 0.4224, |
|
"step": 15050 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 2.446748983474466e-05, |
|
"loss": 0.396, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 2.437423061140747e-05, |
|
"loss": 0.3879, |
|
"step": 15150 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 2.428097138807028e-05, |
|
"loss": 0.3986, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 2.4187712164733094e-05, |
|
"loss": 0.4013, |
|
"step": 15250 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 2.4094452941395907e-05, |
|
"loss": 0.4129, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 2.4001193718058717e-05, |
|
"loss": 0.385, |
|
"step": 15350 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 2.3907934494721527e-05, |
|
"loss": 0.3973, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 2.381467527138434e-05, |
|
"loss": 0.4077, |
|
"step": 15450 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 2.3721416048047155e-05, |
|
"loss": 0.3829, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 2.3628156824709965e-05, |
|
"loss": 0.3952, |
|
"step": 15550 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 2.3534897601372778e-05, |
|
"loss": 0.4166, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 2.344163837803559e-05, |
|
"loss": 0.3802, |
|
"step": 15650 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 2.3348379154698402e-05, |
|
"loss": 0.402, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 2.3255119931361212e-05, |
|
"loss": 0.3891, |
|
"step": 15750 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 2.3161860708024026e-05, |
|
"loss": 0.3961, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 2.3068601484686836e-05, |
|
"loss": 0.4022, |
|
"step": 15850 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 2.297534226134965e-05, |
|
"loss": 0.3854, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 2.2882083038012463e-05, |
|
"loss": 0.3939, |
|
"step": 15950 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 2.2788823814675273e-05, |
|
"loss": 0.4001, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 2.2695564591338083e-05, |
|
"loss": 0.3946, |
|
"step": 16050 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 2.2602305368000897e-05, |
|
"loss": 0.4052, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 2.250904614466371e-05, |
|
"loss": 0.3869, |
|
"step": 16150 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 2.241578692132652e-05, |
|
"loss": 0.4003, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 2.232252769798933e-05, |
|
"loss": 0.3748, |
|
"step": 16250 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 2.2229268474652144e-05, |
|
"loss": 0.4055, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 2.2136009251314957e-05, |
|
"loss": 0.4013, |
|
"step": 16350 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 2.2042750027977768e-05, |
|
"loss": 0.4041, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 2.194949080464058e-05, |
|
"loss": 0.394, |
|
"step": 16450 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 2.185623158130339e-05, |
|
"loss": 0.4041, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 2.1762972357966205e-05, |
|
"loss": 0.3931, |
|
"step": 16550 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 2.1669713134629015e-05, |
|
"loss": 0.3857, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 2.157645391129183e-05, |
|
"loss": 0.3884, |
|
"step": 16650 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 2.148319468795464e-05, |
|
"loss": 0.3984, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 2.1389935464617452e-05, |
|
"loss": 0.3918, |
|
"step": 16750 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 2.1296676241280262e-05, |
|
"loss": 0.4041, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"learning_rate": 2.1203417017943076e-05, |
|
"loss": 0.3781, |
|
"step": 16850 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 2.1110157794605886e-05, |
|
"loss": 0.3917, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 2.10168985712687e-05, |
|
"loss": 0.4061, |
|
"step": 16950 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 2.0923639347931513e-05, |
|
"loss": 0.3916, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 2.0830380124594323e-05, |
|
"loss": 0.3893, |
|
"step": 17050 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 2.0737120901257133e-05, |
|
"loss": 0.3926, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 2.0643861677919947e-05, |
|
"loss": 0.391, |
|
"step": 17150 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 2.055060245458276e-05, |
|
"loss": 0.4, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 2.045734323124557e-05, |
|
"loss": 0.3907, |
|
"step": 17250 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 2.036408400790838e-05, |
|
"loss": 0.4068, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 2.0270824784571198e-05, |
|
"loss": 0.3918, |
|
"step": 17350 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 2.0177565561234008e-05, |
|
"loss": 0.3924, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 2.0084306337896818e-05, |
|
"loss": 0.3948, |
|
"step": 17450 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"learning_rate": 1.999104711455963e-05, |
|
"loss": 0.3926, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 1.9897787891222445e-05, |
|
"loss": 0.4048, |
|
"step": 17550 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"learning_rate": 1.9804528667885255e-05, |
|
"loss": 0.3878, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 1.9711269444548065e-05, |
|
"loss": 0.3833, |
|
"step": 17650 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"learning_rate": 1.961801022121088e-05, |
|
"loss": 0.385, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 1.9524750997873692e-05, |
|
"loss": 0.3816, |
|
"step": 17750 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 1.9431491774536502e-05, |
|
"loss": 0.389, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 1.9338232551199316e-05, |
|
"loss": 0.3926, |
|
"step": 17850 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 1.9244973327862126e-05, |
|
"loss": 0.3939, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 1.9151714104524936e-05, |
|
"loss": 0.3958, |
|
"step": 17950 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 1.905845488118775e-05, |
|
"loss": 0.3962, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 1.8965195657850563e-05, |
|
"loss": 0.4012, |
|
"step": 18050 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 1.8871936434513373e-05, |
|
"loss": 0.4, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 1.8778677211176184e-05, |
|
"loss": 0.3816, |
|
"step": 18150 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"learning_rate": 1.8685417987839e-05, |
|
"loss": 0.4022, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 1.859215876450181e-05, |
|
"loss": 0.3865, |
|
"step": 18250 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 1.849889954116462e-05, |
|
"loss": 0.3896, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 1.8405640317827434e-05, |
|
"loss": 0.4095, |
|
"step": 18350 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 1.8312381094490248e-05, |
|
"loss": 0.3945, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"learning_rate": 1.8219121871153058e-05, |
|
"loss": 0.4106, |
|
"step": 18450 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 1.8125862647815868e-05, |
|
"loss": 0.3846, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 1.803260342447868e-05, |
|
"loss": 0.3833, |
|
"step": 18550 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 1.7939344201141495e-05, |
|
"loss": 0.3827, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 1.7846084977804305e-05, |
|
"loss": 0.3851, |
|
"step": 18650 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"learning_rate": 1.775282575446712e-05, |
|
"loss": 0.3882, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 1.765956653112993e-05, |
|
"loss": 0.3743, |
|
"step": 18750 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 1.7566307307792743e-05, |
|
"loss": 0.3919, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 1.7473048084455553e-05, |
|
"loss": 0.3906, |
|
"step": 18850 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 1.7379788861118366e-05, |
|
"loss": 0.3767, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 1.7286529637781176e-05, |
|
"loss": 0.3705, |
|
"step": 18950 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 1.719327041444399e-05, |
|
"loss": 0.3727, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"learning_rate": 1.71000111911068e-05, |
|
"loss": 0.3846, |
|
"step": 19050 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"learning_rate": 1.7006751967769614e-05, |
|
"loss": 0.3694, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 1.6913492744432424e-05, |
|
"loss": 0.3776, |
|
"step": 19150 |
|
}, |
|
{ |
|
"epoch": 4.08, |
|
"learning_rate": 1.6820233521095237e-05, |
|
"loss": 0.3787, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"learning_rate": 1.672697429775805e-05, |
|
"loss": 0.3818, |
|
"step": 19250 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 1.663371507442086e-05, |
|
"loss": 0.3862, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"learning_rate": 1.654045585108367e-05, |
|
"loss": 0.3697, |
|
"step": 19350 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"learning_rate": 1.6447196627746485e-05, |
|
"loss": 0.3797, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 1.6353937404409298e-05, |
|
"loss": 0.3881, |
|
"step": 19450 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"learning_rate": 1.6260678181072108e-05, |
|
"loss": 0.3924, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"learning_rate": 1.616741895773492e-05, |
|
"loss": 0.3804, |
|
"step": 19550 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 1.6074159734397735e-05, |
|
"loss": 0.3899, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 4.18, |
|
"learning_rate": 1.5980900511060546e-05, |
|
"loss": 0.3864, |
|
"step": 19650 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 1.5887641287723356e-05, |
|
"loss": 0.3918, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"learning_rate": 1.579438206438617e-05, |
|
"loss": 0.387, |
|
"step": 19750 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 1.570112284104898e-05, |
|
"loss": 0.3936, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 1.5607863617711793e-05, |
|
"loss": 0.3697, |
|
"step": 19850 |
|
}, |
|
{ |
|
"epoch": 4.23, |
|
"learning_rate": 1.5514604394374603e-05, |
|
"loss": 0.381, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"learning_rate": 1.5421345171037417e-05, |
|
"loss": 0.3829, |
|
"step": 19950 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 1.5328085947700227e-05, |
|
"loss": 0.3891, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 4.26, |
|
"learning_rate": 1.523482672436304e-05, |
|
"loss": 0.3946, |
|
"step": 20050 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"learning_rate": 1.5141567501025852e-05, |
|
"loss": 0.3913, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 1.5048308277688664e-05, |
|
"loss": 0.3778, |
|
"step": 20150 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 1.4955049054351474e-05, |
|
"loss": 0.3867, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 1.486178983101429e-05, |
|
"loss": 0.4, |
|
"step": 20250 |
|
}, |
|
{ |
|
"epoch": 4.32, |
|
"learning_rate": 1.47685306076771e-05, |
|
"loss": 0.3632, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 1.4675271384339911e-05, |
|
"loss": 0.3735, |
|
"step": 20350 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 1.4582012161002723e-05, |
|
"loss": 0.3829, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 4.35, |
|
"learning_rate": 1.4488752937665537e-05, |
|
"loss": 0.3725, |
|
"step": 20450 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 1.4395493714328348e-05, |
|
"loss": 0.4004, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 1.4302234490991159e-05, |
|
"loss": 0.3915, |
|
"step": 20550 |
|
}, |
|
{ |
|
"epoch": 4.38, |
|
"learning_rate": 1.420897526765397e-05, |
|
"loss": 0.3941, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"learning_rate": 1.4115716044316784e-05, |
|
"loss": 0.3845, |
|
"step": 20650 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 1.4022456820979596e-05, |
|
"loss": 0.3779, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 1.3929197597642408e-05, |
|
"loss": 0.3843, |
|
"step": 20750 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"learning_rate": 1.3835938374305218e-05, |
|
"loss": 0.4076, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 4.43, |
|
"learning_rate": 1.3742679150968033e-05, |
|
"loss": 0.3701, |
|
"step": 20850 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 1.3649419927630843e-05, |
|
"loss": 0.3765, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 4.45, |
|
"learning_rate": 1.3556160704293655e-05, |
|
"loss": 0.386, |
|
"step": 20950 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 1.3462901480956467e-05, |
|
"loss": 0.3738, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 4.48, |
|
"learning_rate": 1.336964225761928e-05, |
|
"loss": 0.3785, |
|
"step": 21050 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"learning_rate": 1.3276383034282092e-05, |
|
"loss": 0.3901, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"learning_rate": 1.3183123810944902e-05, |
|
"loss": 0.3876, |
|
"step": 21150 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"learning_rate": 1.3089864587607714e-05, |
|
"loss": 0.3769, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 1.2996605364270528e-05, |
|
"loss": 0.3808, |
|
"step": 21250 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 1.290334614093334e-05, |
|
"loss": 0.3727, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 1.2810086917596151e-05, |
|
"loss": 0.3604, |
|
"step": 21350 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"learning_rate": 1.2716827694258962e-05, |
|
"loss": 0.3852, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 4.56, |
|
"learning_rate": 1.2623568470921777e-05, |
|
"loss": 0.3731, |
|
"step": 21450 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 1.2530309247584587e-05, |
|
"loss": 0.3604, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"learning_rate": 1.2437050024247399e-05, |
|
"loss": 0.3891, |
|
"step": 21550 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"learning_rate": 1.234379080091021e-05, |
|
"loss": 0.3605, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 1.2250531577573022e-05, |
|
"loss": 0.3812, |
|
"step": 21650 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 1.2157272354235834e-05, |
|
"loss": 0.3714, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"learning_rate": 1.2064013130898646e-05, |
|
"loss": 0.3836, |
|
"step": 21750 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 1.1970753907561458e-05, |
|
"loss": 0.3876, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"learning_rate": 1.187749468422427e-05, |
|
"loss": 0.3899, |
|
"step": 21850 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 1.1784235460887082e-05, |
|
"loss": 0.3924, |
|
"step": 21900 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"learning_rate": 1.1690976237549895e-05, |
|
"loss": 0.3722, |
|
"step": 21950 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 1.1597717014212705e-05, |
|
"loss": 0.3681, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 4.69, |
|
"learning_rate": 1.1504457790875519e-05, |
|
"loss": 0.3682, |
|
"step": 22050 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"learning_rate": 1.1411198567538329e-05, |
|
"loss": 0.3911, |
|
"step": 22100 |
|
}, |
|
{ |
|
"epoch": 4.71, |
|
"learning_rate": 1.1317939344201142e-05, |
|
"loss": 0.3843, |
|
"step": 22150 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 1.1224680120863954e-05, |
|
"loss": 0.3684, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 1.1131420897526766e-05, |
|
"loss": 0.369, |
|
"step": 22250 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"learning_rate": 1.1038161674189578e-05, |
|
"loss": 0.3831, |
|
"step": 22300 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 1.094490245085239e-05, |
|
"loss": 0.3821, |
|
"step": 22350 |
|
}, |
|
{ |
|
"epoch": 4.76, |
|
"learning_rate": 1.0851643227515202e-05, |
|
"loss": 0.386, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"learning_rate": 1.0758384004178013e-05, |
|
"loss": 0.375, |
|
"step": 22450 |
|
}, |
|
{ |
|
"epoch": 4.78, |
|
"learning_rate": 1.0665124780840825e-05, |
|
"loss": 0.3873, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 1.0571865557503637e-05, |
|
"loss": 0.3766, |
|
"step": 22550 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 1.0478606334166449e-05, |
|
"loss": 0.3804, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"learning_rate": 1.0385347110829263e-05, |
|
"loss": 0.3724, |
|
"step": 22650 |
|
}, |
|
{ |
|
"epoch": 4.83, |
|
"learning_rate": 1.0292087887492073e-05, |
|
"loss": 0.3775, |
|
"step": 22700 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"learning_rate": 1.0198828664154886e-05, |
|
"loss": 0.3718, |
|
"step": 22750 |
|
}, |
|
{ |
|
"epoch": 4.85, |
|
"learning_rate": 1.0105569440817696e-05, |
|
"loss": 0.3664, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 1.001231021748051e-05, |
|
"loss": 0.3841, |
|
"step": 22850 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 9.919050994143322e-06, |
|
"loss": 0.3722, |
|
"step": 22900 |
|
}, |
|
{ |
|
"epoch": 4.88, |
|
"learning_rate": 9.825791770806134e-06, |
|
"loss": 0.3717, |
|
"step": 22950 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 9.732532547468945e-06, |
|
"loss": 0.3665, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"learning_rate": 9.639273324131757e-06, |
|
"loss": 0.3687, |
|
"step": 23050 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"learning_rate": 9.546014100794569e-06, |
|
"loss": 0.3823, |
|
"step": 23100 |
|
}, |
|
{ |
|
"epoch": 4.92, |
|
"learning_rate": 9.452754877457381e-06, |
|
"loss": 0.3726, |
|
"step": 23150 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 9.359495654120193e-06, |
|
"loss": 0.359, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 4.94, |
|
"learning_rate": 9.266236430783006e-06, |
|
"loss": 0.371, |
|
"step": 23250 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 9.172977207445816e-06, |
|
"loss": 0.377, |
|
"step": 23300 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 9.07971798410863e-06, |
|
"loss": 0.3793, |
|
"step": 23350 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"learning_rate": 8.98645876077144e-06, |
|
"loss": 0.379, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"learning_rate": 8.893199537434254e-06, |
|
"loss": 0.3718, |
|
"step": 23450 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 8.799940314097064e-06, |
|
"loss": 0.3749, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 8.706681090759877e-06, |
|
"loss": 0.3659, |
|
"step": 23550 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 8.613421867422689e-06, |
|
"loss": 0.3753, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 8.5201626440855e-06, |
|
"loss": 0.3629, |
|
"step": 23650 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 8.426903420748313e-06, |
|
"loss": 0.3733, |
|
"step": 23700 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"learning_rate": 8.333644197411123e-06, |
|
"loss": 0.3886, |
|
"step": 23750 |
|
}, |
|
{ |
|
"epoch": 5.06, |
|
"learning_rate": 8.240384974073936e-06, |
|
"loss": 0.3641, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"learning_rate": 8.147125750736748e-06, |
|
"loss": 0.3761, |
|
"step": 23850 |
|
}, |
|
{ |
|
"epoch": 5.08, |
|
"learning_rate": 8.05386652739956e-06, |
|
"loss": 0.3617, |
|
"step": 23900 |
|
}, |
|
{ |
|
"epoch": 5.09, |
|
"learning_rate": 7.960607304062372e-06, |
|
"loss": 0.3726, |
|
"step": 23950 |
|
}, |
|
{ |
|
"epoch": 5.1, |
|
"learning_rate": 7.867348080725184e-06, |
|
"loss": 0.3749, |
|
"step": 24000 |
|
} |
|
], |
|
"max_steps": 28218, |
|
"num_train_epochs": 6, |
|
"total_flos": 9.599266478715476e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|