|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 20.0, |
|
"eval_steps": 500, |
|
"global_step": 100280, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0997207818109294, |
|
"grad_norm": 0.9261826276779175, |
|
"learning_rate": 4.975069804547268e-05, |
|
"loss": 9.3488, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.1994415636218588, |
|
"grad_norm": 0.5372676253318787, |
|
"learning_rate": 4.9501396090945354e-05, |
|
"loss": 8.0556, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.2991623454327882, |
|
"grad_norm": 0.5985580086708069, |
|
"learning_rate": 4.925209413641803e-05, |
|
"loss": 7.7905, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.3988831272437176, |
|
"grad_norm": 0.7738422751426697, |
|
"learning_rate": 4.900279218189071e-05, |
|
"loss": 7.5996, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.49860390905464697, |
|
"grad_norm": 0.8470643162727356, |
|
"learning_rate": 4.875349022736339e-05, |
|
"loss": 7.4439, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.5983246908655764, |
|
"grad_norm": 0.9632411003112793, |
|
"learning_rate": 4.850418827283606e-05, |
|
"loss": 7.2826, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.6980454726765057, |
|
"grad_norm": 0.9540568590164185, |
|
"learning_rate": 4.825488631830874e-05, |
|
"loss": 7.1712, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.7977662544874352, |
|
"grad_norm": 0.9605555534362793, |
|
"learning_rate": 4.8005584363781416e-05, |
|
"loss": 7.0881, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.8974870362983646, |
|
"grad_norm": 1.101860523223877, |
|
"learning_rate": 4.775678101316314e-05, |
|
"loss": 7.0052, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.9972078181092939, |
|
"grad_norm": 1.8117992877960205, |
|
"learning_rate": 4.750747905863582e-05, |
|
"loss": 6.9103, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.0969285999202234, |
|
"grad_norm": 1.2837625741958618, |
|
"learning_rate": 4.72581771041085e-05, |
|
"loss": 6.8231, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 1.1966493817311528, |
|
"grad_norm": 1.3095475435256958, |
|
"learning_rate": 4.7008875149581175e-05, |
|
"loss": 6.7796, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 1.2963701635420821, |
|
"grad_norm": 1.4652228355407715, |
|
"learning_rate": 4.6760071798962906e-05, |
|
"loss": 6.707, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 1.3960909453530115, |
|
"grad_norm": 1.357987403869629, |
|
"learning_rate": 4.651076984443558e-05, |
|
"loss": 6.6568, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 1.4958117271639408, |
|
"grad_norm": 1.31229829788208, |
|
"learning_rate": 4.626146788990826e-05, |
|
"loss": 6.6144, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 1.5955325089748702, |
|
"grad_norm": 1.4246258735656738, |
|
"learning_rate": 4.6012165935380934e-05, |
|
"loss": 6.5669, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 1.6952532907857998, |
|
"grad_norm": 1.3963290452957153, |
|
"learning_rate": 4.576286398085361e-05, |
|
"loss": 6.508, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 1.7949740725967291, |
|
"grad_norm": 1.5991405248641968, |
|
"learning_rate": 4.551406063023534e-05, |
|
"loss": 6.4651, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 1.8946948544076585, |
|
"grad_norm": 1.6727421283721924, |
|
"learning_rate": 4.526475867570802e-05, |
|
"loss": 6.4129, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 1.994415636218588, |
|
"grad_norm": 1.5657908916473389, |
|
"learning_rate": 4.50154567211807e-05, |
|
"loss": 6.3859, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 2.0941364180295174, |
|
"grad_norm": 1.4079279899597168, |
|
"learning_rate": 4.4766154766653375e-05, |
|
"loss": 6.3136, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 2.193857199840447, |
|
"grad_norm": 1.4724199771881104, |
|
"learning_rate": 4.45173514160351e-05, |
|
"loss": 6.2935, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 2.293577981651376, |
|
"grad_norm": 1.7075014114379883, |
|
"learning_rate": 4.426804946150778e-05, |
|
"loss": 6.2724, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 2.3932987634623055, |
|
"grad_norm": 1.550031304359436, |
|
"learning_rate": 4.401874750698046e-05, |
|
"loss": 6.2526, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 2.493019545273235, |
|
"grad_norm": 1.5938421487808228, |
|
"learning_rate": 4.376944555245313e-05, |
|
"loss": 6.2218, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 2.5927403270841642, |
|
"grad_norm": 1.853521704673767, |
|
"learning_rate": 4.3520642201834866e-05, |
|
"loss": 6.1679, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 2.6924611088950936, |
|
"grad_norm": 1.7296770811080933, |
|
"learning_rate": 4.327134024730754e-05, |
|
"loss": 6.1442, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 2.792181890706023, |
|
"grad_norm": 1.752852201461792, |
|
"learning_rate": 4.302203829278022e-05, |
|
"loss": 6.1241, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 2.8919026725169523, |
|
"grad_norm": 1.6265596151351929, |
|
"learning_rate": 4.2772736338252893e-05, |
|
"loss": 6.0973, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 2.9916234543278817, |
|
"grad_norm": 1.7537871599197388, |
|
"learning_rate": 4.2523932987634625e-05, |
|
"loss": 6.0858, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 3.0913442361388115, |
|
"grad_norm": 1.5949361324310303, |
|
"learning_rate": 4.22746310331073e-05, |
|
"loss": 6.0188, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 3.191065017949741, |
|
"grad_norm": 1.907575011253357, |
|
"learning_rate": 4.202532907857998e-05, |
|
"loss": 6.0118, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 3.29078579976067, |
|
"grad_norm": 1.8919939994812012, |
|
"learning_rate": 4.177602712405265e-05, |
|
"loss": 5.9883, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 3.3905065815715996, |
|
"grad_norm": 1.8701094388961792, |
|
"learning_rate": 4.1527223773434384e-05, |
|
"loss": 5.9725, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 3.490227363382529, |
|
"grad_norm": 2.043443202972412, |
|
"learning_rate": 4.127792181890706e-05, |
|
"loss": 5.9474, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 3.5899481451934583, |
|
"grad_norm": 1.852910041809082, |
|
"learning_rate": 4.102861986437974e-05, |
|
"loss": 5.9103, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 3.6896689270043876, |
|
"grad_norm": 1.7579346895217896, |
|
"learning_rate": 4.077931790985242e-05, |
|
"loss": 5.9047, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 3.789389708815317, |
|
"grad_norm": 1.8655468225479126, |
|
"learning_rate": 4.053051455923414e-05, |
|
"loss": 5.9043, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 3.8891104906262464, |
|
"grad_norm": 2.0585408210754395, |
|
"learning_rate": 4.0281212604706826e-05, |
|
"loss": 5.8705, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 3.988831272437176, |
|
"grad_norm": 2.088595151901245, |
|
"learning_rate": 4.00319106501795e-05, |
|
"loss": 5.8608, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 4.0885520542481055, |
|
"grad_norm": 1.8989760875701904, |
|
"learning_rate": 3.978260869565217e-05, |
|
"loss": 5.8265, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 4.188272836059035, |
|
"grad_norm": 1.8559260368347168, |
|
"learning_rate": 3.953330674112485e-05, |
|
"loss": 5.8052, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 4.287993617869964, |
|
"grad_norm": 2.0016090869903564, |
|
"learning_rate": 3.9284503390506585e-05, |
|
"loss": 5.7947, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 4.387714399680894, |
|
"grad_norm": 2.0224192142486572, |
|
"learning_rate": 3.903520143597926e-05, |
|
"loss": 5.7605, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 4.487435181491823, |
|
"grad_norm": 1.9289922714233398, |
|
"learning_rate": 3.8785899481451936e-05, |
|
"loss": 5.7315, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 4.587155963302752, |
|
"grad_norm": 2.1070337295532227, |
|
"learning_rate": 3.853659752692461e-05, |
|
"loss": 5.7517, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 4.686876745113682, |
|
"grad_norm": 2.091681718826294, |
|
"learning_rate": 3.8287794176306343e-05, |
|
"loss": 5.7206, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 4.786597526924611, |
|
"grad_norm": 2.0619523525238037, |
|
"learning_rate": 3.803849222177902e-05, |
|
"loss": 5.6829, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 4.88631830873554, |
|
"grad_norm": 2.1252663135528564, |
|
"learning_rate": 3.7789190267251695e-05, |
|
"loss": 5.7028, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 4.98603909054647, |
|
"grad_norm": 2.179452657699585, |
|
"learning_rate": 3.753988831272438e-05, |
|
"loss": 5.6975, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 5.085759872357399, |
|
"grad_norm": 2.054488182067871, |
|
"learning_rate": 3.72910849621061e-05, |
|
"loss": 5.6527, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 5.1854806541683285, |
|
"grad_norm": 2.3839542865753174, |
|
"learning_rate": 3.704178300757878e-05, |
|
"loss": 5.6498, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 5.285201435979258, |
|
"grad_norm": 1.9893797636032104, |
|
"learning_rate": 3.679248105305146e-05, |
|
"loss": 5.6103, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 5.384922217790187, |
|
"grad_norm": 2.089535713195801, |
|
"learning_rate": 3.654317909852413e-05, |
|
"loss": 5.6011, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 5.484642999601117, |
|
"grad_norm": 1.9748643636703491, |
|
"learning_rate": 3.629387714399681e-05, |
|
"loss": 5.6206, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 5.584363781412046, |
|
"grad_norm": 2.1696887016296387, |
|
"learning_rate": 3.6045073793378544e-05, |
|
"loss": 5.607, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 5.684084563222975, |
|
"grad_norm": 2.2261533737182617, |
|
"learning_rate": 3.579577183885121e-05, |
|
"loss": 5.5895, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 5.783805345033905, |
|
"grad_norm": 1.9739435911178589, |
|
"learning_rate": 3.5546469884323896e-05, |
|
"loss": 5.5654, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 5.883526126844835, |
|
"grad_norm": 2.3373613357543945, |
|
"learning_rate": 3.529716792979657e-05, |
|
"loss": 5.554, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 5.983246908655763, |
|
"grad_norm": 2.0227203369140625, |
|
"learning_rate": 3.50483645791783e-05, |
|
"loss": 5.569, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 6.082967690466694, |
|
"grad_norm": 2.1894445419311523, |
|
"learning_rate": 3.479906262465098e-05, |
|
"loss": 5.5268, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 6.182688472277623, |
|
"grad_norm": 2.3545119762420654, |
|
"learning_rate": 3.4549760670123655e-05, |
|
"loss": 5.5102, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 6.282409254088552, |
|
"grad_norm": 2.380277156829834, |
|
"learning_rate": 3.430045871559634e-05, |
|
"loss": 5.5301, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 6.382130035899482, |
|
"grad_norm": 2.288188934326172, |
|
"learning_rate": 3.405165536497806e-05, |
|
"loss": 5.4927, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 6.481850817710411, |
|
"grad_norm": 2.2211456298828125, |
|
"learning_rate": 3.380235341045074e-05, |
|
"loss": 5.4786, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 6.58157159952134, |
|
"grad_norm": 2.5629711151123047, |
|
"learning_rate": 3.355305145592342e-05, |
|
"loss": 5.4932, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 6.68129238133227, |
|
"grad_norm": 2.385563611984253, |
|
"learning_rate": 3.330374950139609e-05, |
|
"loss": 5.4715, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 6.781013163143199, |
|
"grad_norm": 2.284985303878784, |
|
"learning_rate": 3.305494615077782e-05, |
|
"loss": 5.4599, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 6.8807339449541285, |
|
"grad_norm": 2.3653366565704346, |
|
"learning_rate": 3.2805644196250504e-05, |
|
"loss": 5.4498, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 6.980454726765058, |
|
"grad_norm": 2.311102867126465, |
|
"learning_rate": 3.255634224172317e-05, |
|
"loss": 5.4315, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 7.080175508575987, |
|
"grad_norm": 2.4815216064453125, |
|
"learning_rate": 3.2307040287195855e-05, |
|
"loss": 5.4282, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 7.179896290386917, |
|
"grad_norm": 2.228046178817749, |
|
"learning_rate": 3.205823693657759e-05, |
|
"loss": 5.4321, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 7.279617072197846, |
|
"grad_norm": 2.459022283554077, |
|
"learning_rate": 3.1808934982050256e-05, |
|
"loss": 5.3886, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 7.379337854008775, |
|
"grad_norm": 2.217167615890503, |
|
"learning_rate": 3.155963302752294e-05, |
|
"loss": 5.3924, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 7.479058635819705, |
|
"grad_norm": 2.3231680393218994, |
|
"learning_rate": 3.1310331072995614e-05, |
|
"loss": 5.389, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 7.578779417630634, |
|
"grad_norm": 2.220628261566162, |
|
"learning_rate": 3.1061527722377346e-05, |
|
"loss": 5.3856, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 7.678500199441563, |
|
"grad_norm": 2.612741708755493, |
|
"learning_rate": 3.081222576785002e-05, |
|
"loss": 5.3851, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 7.778220981252493, |
|
"grad_norm": 2.194031000137329, |
|
"learning_rate": 3.05629238133227e-05, |
|
"loss": 5.3744, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 7.877941763063422, |
|
"grad_norm": 2.342750310897827, |
|
"learning_rate": 3.0313621858795377e-05, |
|
"loss": 5.3644, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 7.9776625448743514, |
|
"grad_norm": 2.273401975631714, |
|
"learning_rate": 3.0064818508177105e-05, |
|
"loss": 5.3828, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 8.07738332668528, |
|
"grad_norm": 2.5998456478118896, |
|
"learning_rate": 2.981551655364978e-05, |
|
"loss": 5.3399, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 8.177104108496211, |
|
"grad_norm": 2.4312164783477783, |
|
"learning_rate": 2.956621459912246e-05, |
|
"loss": 5.3491, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 8.27682489030714, |
|
"grad_norm": 2.1767194271087646, |
|
"learning_rate": 2.9316912644595136e-05, |
|
"loss": 5.318, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 8.37654567211807, |
|
"grad_norm": 2.546261787414551, |
|
"learning_rate": 2.9068109293976864e-05, |
|
"loss": 5.3169, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 8.476266453928998, |
|
"grad_norm": 2.5187346935272217, |
|
"learning_rate": 2.8818807339449543e-05, |
|
"loss": 5.3051, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 8.575987235739928, |
|
"grad_norm": 2.4358792304992676, |
|
"learning_rate": 2.856950538492222e-05, |
|
"loss": 5.3063, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 8.675708017550857, |
|
"grad_norm": 2.22619891166687, |
|
"learning_rate": 2.8320203430394898e-05, |
|
"loss": 5.3241, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 8.775428799361787, |
|
"grad_norm": 2.6035451889038086, |
|
"learning_rate": 2.8071400079776626e-05, |
|
"loss": 5.2964, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 8.875149581172716, |
|
"grad_norm": 2.5391156673431396, |
|
"learning_rate": 2.7822098125249302e-05, |
|
"loss": 5.2914, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 8.974870362983646, |
|
"grad_norm": 2.4130935668945312, |
|
"learning_rate": 2.757279617072198e-05, |
|
"loss": 5.2883, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 9.074591144794574, |
|
"grad_norm": 2.411205530166626, |
|
"learning_rate": 2.7323494216194657e-05, |
|
"loss": 5.2893, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 9.174311926605505, |
|
"grad_norm": 2.484266757965088, |
|
"learning_rate": 2.7074690865576385e-05, |
|
"loss": 5.2649, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 9.274032708416435, |
|
"grad_norm": 2.446840524673462, |
|
"learning_rate": 2.6825388911049064e-05, |
|
"loss": 5.2579, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 9.373753490227363, |
|
"grad_norm": 2.2476446628570557, |
|
"learning_rate": 2.657608695652174e-05, |
|
"loss": 5.2616, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 9.473474272038294, |
|
"grad_norm": 2.36161732673645, |
|
"learning_rate": 2.632678500199442e-05, |
|
"loss": 5.2458, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 9.573195053849222, |
|
"grad_norm": 2.4564807415008545, |
|
"learning_rate": 2.6077981651376147e-05, |
|
"loss": 5.2522, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 9.672915835660152, |
|
"grad_norm": 2.477536678314209, |
|
"learning_rate": 2.5828679696848823e-05, |
|
"loss": 5.249, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 9.77263661747108, |
|
"grad_norm": 2.8510327339172363, |
|
"learning_rate": 2.5579377742321503e-05, |
|
"loss": 5.2397, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 9.872357399282011, |
|
"grad_norm": 2.4770243167877197, |
|
"learning_rate": 2.533007578779418e-05, |
|
"loss": 5.2172, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 9.97207818109294, |
|
"grad_norm": 2.492191791534424, |
|
"learning_rate": 2.5081272437175906e-05, |
|
"loss": 5.236, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 10.07179896290387, |
|
"grad_norm": 2.5560014247894287, |
|
"learning_rate": 2.4831970482648582e-05, |
|
"loss": 5.2159, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 10.171519744714798, |
|
"grad_norm": 2.550168752670288, |
|
"learning_rate": 2.458266852812126e-05, |
|
"loss": 5.2026, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 10.271240526525728, |
|
"grad_norm": 2.562626600265503, |
|
"learning_rate": 2.4333366573593937e-05, |
|
"loss": 5.2121, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 10.370961308336657, |
|
"grad_norm": 2.389833927154541, |
|
"learning_rate": 2.408456322297567e-05, |
|
"loss": 5.1963, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 10.470682090147587, |
|
"grad_norm": 2.617138385772705, |
|
"learning_rate": 2.3835261268448345e-05, |
|
"loss": 5.1838, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 10.570402871958516, |
|
"grad_norm": 2.6732029914855957, |
|
"learning_rate": 2.358595931392102e-05, |
|
"loss": 5.2135, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 10.670123653769446, |
|
"grad_norm": 2.509752035140991, |
|
"learning_rate": 2.33366573593937e-05, |
|
"loss": 5.1937, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 10.769844435580374, |
|
"grad_norm": 2.732623815536499, |
|
"learning_rate": 2.3087355404866376e-05, |
|
"loss": 5.2013, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 10.869565217391305, |
|
"grad_norm": 2.7967655658721924, |
|
"learning_rate": 2.2838552054248104e-05, |
|
"loss": 5.1751, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 10.969285999202233, |
|
"grad_norm": 2.6768581867218018, |
|
"learning_rate": 2.2589250099720783e-05, |
|
"loss": 5.1728, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 11.069006781013163, |
|
"grad_norm": 2.3465123176574707, |
|
"learning_rate": 2.233994814519346e-05, |
|
"loss": 5.1869, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 11.168727562824092, |
|
"grad_norm": 2.3460209369659424, |
|
"learning_rate": 2.2090646190666138e-05, |
|
"loss": 5.1704, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 11.268448344635022, |
|
"grad_norm": 2.7022573947906494, |
|
"learning_rate": 2.1841842840047866e-05, |
|
"loss": 5.1508, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 11.36816912644595, |
|
"grad_norm": 2.5259013175964355, |
|
"learning_rate": 2.1592540885520542e-05, |
|
"loss": 5.1441, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 11.46788990825688, |
|
"grad_norm": 2.6938321590423584, |
|
"learning_rate": 2.134323893099322e-05, |
|
"loss": 5.1628, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 11.56761069006781, |
|
"grad_norm": 2.874973773956299, |
|
"learning_rate": 2.1093936976465897e-05, |
|
"loss": 5.1405, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 11.66733147187874, |
|
"grad_norm": 2.762739896774292, |
|
"learning_rate": 2.0845133625847625e-05, |
|
"loss": 5.1264, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 11.76705225368967, |
|
"grad_norm": 2.617100954055786, |
|
"learning_rate": 2.0595831671320304e-05, |
|
"loss": 5.1297, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 11.866773035500598, |
|
"grad_norm": 2.754258632659912, |
|
"learning_rate": 2.034652971679298e-05, |
|
"loss": 5.1349, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 11.966493817311529, |
|
"grad_norm": 2.57446551322937, |
|
"learning_rate": 2.009722776226566e-05, |
|
"loss": 5.1401, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 12.066214599122457, |
|
"grad_norm": 2.7069363594055176, |
|
"learning_rate": 1.9848424411647387e-05, |
|
"loss": 5.1093, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 12.165935380933387, |
|
"grad_norm": 2.5617587566375732, |
|
"learning_rate": 1.9599122457120063e-05, |
|
"loss": 5.1071, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 12.265656162744316, |
|
"grad_norm": 2.7849984169006348, |
|
"learning_rate": 1.9349820502592742e-05, |
|
"loss": 5.1298, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 12.365376944555246, |
|
"grad_norm": 2.6736953258514404, |
|
"learning_rate": 1.910051854806542e-05, |
|
"loss": 5.0828, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 12.465097726366174, |
|
"grad_norm": 2.623760938644409, |
|
"learning_rate": 1.8851715197447146e-05, |
|
"loss": 5.1166, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 12.564818508177105, |
|
"grad_norm": 3.202988624572754, |
|
"learning_rate": 1.8602413242919826e-05, |
|
"loss": 5.1128, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 12.664539289988033, |
|
"grad_norm": 2.4918911457061768, |
|
"learning_rate": 1.83531112883925e-05, |
|
"loss": 5.1181, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 12.764260071798963, |
|
"grad_norm": 2.7274303436279297, |
|
"learning_rate": 1.810380933386518e-05, |
|
"loss": 5.1089, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 12.863980853609892, |
|
"grad_norm": 2.760390520095825, |
|
"learning_rate": 1.785500598324691e-05, |
|
"loss": 5.1175, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 12.963701635420822, |
|
"grad_norm": 2.7950050830841064, |
|
"learning_rate": 1.7605704028719585e-05, |
|
"loss": 5.1173, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 13.06342241723175, |
|
"grad_norm": 2.5247349739074707, |
|
"learning_rate": 1.7356402074192264e-05, |
|
"loss": 5.0765, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 13.16314319904268, |
|
"grad_norm": 2.5236001014709473, |
|
"learning_rate": 1.710710011966494e-05, |
|
"loss": 5.0699, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 13.26286398085361, |
|
"grad_norm": 2.407404661178589, |
|
"learning_rate": 1.6858296769046668e-05, |
|
"loss": 5.0999, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 13.36258476266454, |
|
"grad_norm": 2.665024518966675, |
|
"learning_rate": 1.6608994814519347e-05, |
|
"loss": 5.0852, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 13.462305544475468, |
|
"grad_norm": 2.7694313526153564, |
|
"learning_rate": 1.6359692859992023e-05, |
|
"loss": 5.0785, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 13.562026326286398, |
|
"grad_norm": 2.839297294616699, |
|
"learning_rate": 1.6110390905464702e-05, |
|
"loss": 5.0826, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 13.661747108097327, |
|
"grad_norm": 2.831908941268921, |
|
"learning_rate": 1.586158755484643e-05, |
|
"loss": 5.0877, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 13.761467889908257, |
|
"grad_norm": 2.9526407718658447, |
|
"learning_rate": 1.5612285600319106e-05, |
|
"loss": 5.0692, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 13.861188671719185, |
|
"grad_norm": 2.832224130630493, |
|
"learning_rate": 1.5362983645791785e-05, |
|
"loss": 5.0687, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 13.960909453530116, |
|
"grad_norm": 2.622544050216675, |
|
"learning_rate": 1.5113681691264461e-05, |
|
"loss": 5.065, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 14.060630235341046, |
|
"grad_norm": 2.9850549697875977, |
|
"learning_rate": 1.486487834064619e-05, |
|
"loss": 5.0665, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 14.160351017151974, |
|
"grad_norm": 2.7051777839660645, |
|
"learning_rate": 1.4615576386118868e-05, |
|
"loss": 5.0509, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 14.260071798962905, |
|
"grad_norm": 2.6535110473632812, |
|
"learning_rate": 1.4366274431591544e-05, |
|
"loss": 5.0426, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 14.359792580773833, |
|
"grad_norm": 2.8298914432525635, |
|
"learning_rate": 1.4116972477064222e-05, |
|
"loss": 5.0348, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 14.459513362584763, |
|
"grad_norm": 2.776466131210327, |
|
"learning_rate": 1.3868169126445951e-05, |
|
"loss": 5.0486, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 14.559234144395692, |
|
"grad_norm": 2.762392044067383, |
|
"learning_rate": 1.3618867171918629e-05, |
|
"loss": 5.0478, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 14.658954926206622, |
|
"grad_norm": 2.7879889011383057, |
|
"learning_rate": 1.3369565217391305e-05, |
|
"loss": 5.0453, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 14.75867570801755, |
|
"grad_norm": 2.7736377716064453, |
|
"learning_rate": 1.3120263262863982e-05, |
|
"loss": 5.0538, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 14.85839648982848, |
|
"grad_norm": 3.0703177452087402, |
|
"learning_rate": 1.2871459912245712e-05, |
|
"loss": 5.0649, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 14.95811727163941, |
|
"grad_norm": 2.8057234287261963, |
|
"learning_rate": 1.262215795771839e-05, |
|
"loss": 5.0558, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 15.05783805345034, |
|
"grad_norm": 2.9843761920928955, |
|
"learning_rate": 1.2372856003191066e-05, |
|
"loss": 5.0367, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 15.157558835261268, |
|
"grad_norm": 2.723043918609619, |
|
"learning_rate": 1.2123554048663741e-05, |
|
"loss": 5.0407, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 15.257279617072198, |
|
"grad_norm": 2.729093551635742, |
|
"learning_rate": 1.1874750698045473e-05, |
|
"loss": 5.0261, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 15.357000398883127, |
|
"grad_norm": 2.7686829566955566, |
|
"learning_rate": 1.162544874351815e-05, |
|
"loss": 5.0189, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 15.456721180694057, |
|
"grad_norm": 2.7151553630828857, |
|
"learning_rate": 1.1376146788990826e-05, |
|
"loss": 5.0186, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 15.556441962504985, |
|
"grad_norm": 2.6453444957733154, |
|
"learning_rate": 1.1126844834463502e-05, |
|
"loss": 5.0252, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 15.656162744315916, |
|
"grad_norm": 2.6896181106567383, |
|
"learning_rate": 1.0878041483845234e-05, |
|
"loss": 5.0129, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 15.755883526126844, |
|
"grad_norm": 2.6944217681884766, |
|
"learning_rate": 1.0628739529317911e-05, |
|
"loss": 5.0443, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 15.855604307937774, |
|
"grad_norm": 2.782818555831909, |
|
"learning_rate": 1.0379437574790587e-05, |
|
"loss": 5.0325, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 15.955325089748703, |
|
"grad_norm": 2.8319053649902344, |
|
"learning_rate": 1.0130135620263263e-05, |
|
"loss": 5.0277, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 16.05504587155963, |
|
"grad_norm": 2.616605758666992, |
|
"learning_rate": 9.881332269644994e-06, |
|
"loss": 5.0127, |
|
"step": 80500 |
|
}, |
|
{ |
|
"epoch": 16.15476665337056, |
|
"grad_norm": 2.8563239574432373, |
|
"learning_rate": 9.632030315117672e-06, |
|
"loss": 5.0094, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 16.254487435181492, |
|
"grad_norm": 2.536868095397949, |
|
"learning_rate": 9.382728360590348e-06, |
|
"loss": 4.9974, |
|
"step": 81500 |
|
}, |
|
{ |
|
"epoch": 16.354208216992422, |
|
"grad_norm": 2.8117527961730957, |
|
"learning_rate": 9.133426406063023e-06, |
|
"loss": 5.0044, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 16.453928998803352, |
|
"grad_norm": 2.776693344116211, |
|
"learning_rate": 8.884623055444755e-06, |
|
"loss": 4.994, |
|
"step": 82500 |
|
}, |
|
{ |
|
"epoch": 16.55364978061428, |
|
"grad_norm": 2.7791900634765625, |
|
"learning_rate": 8.635321100917432e-06, |
|
"loss": 5.0239, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 16.65337056242521, |
|
"grad_norm": 2.8289597034454346, |
|
"learning_rate": 8.386019146390108e-06, |
|
"loss": 5.0059, |
|
"step": 83500 |
|
}, |
|
{ |
|
"epoch": 16.75309134423614, |
|
"grad_norm": 2.89103102684021, |
|
"learning_rate": 8.136717191862784e-06, |
|
"loss": 5.0237, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 16.85281212604707, |
|
"grad_norm": 2.7034354209899902, |
|
"learning_rate": 7.887913841244516e-06, |
|
"loss": 5.0107, |
|
"step": 84500 |
|
}, |
|
{ |
|
"epoch": 16.952532907857996, |
|
"grad_norm": 3.348228693008423, |
|
"learning_rate": 7.638611886717193e-06, |
|
"loss": 5.0128, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 17.052253689668927, |
|
"grad_norm": 2.9482028484344482, |
|
"learning_rate": 7.389309932189868e-06, |
|
"loss": 5.0143, |
|
"step": 85500 |
|
}, |
|
{ |
|
"epoch": 17.151974471479857, |
|
"grad_norm": 2.8442418575286865, |
|
"learning_rate": 7.140007977662546e-06, |
|
"loss": 4.9986, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 17.251695253290787, |
|
"grad_norm": 2.7483792304992676, |
|
"learning_rate": 6.891204627044276e-06, |
|
"loss": 5.0003, |
|
"step": 86500 |
|
}, |
|
{ |
|
"epoch": 17.351416035101714, |
|
"grad_norm": 2.8354785442352295, |
|
"learning_rate": 6.641902672516953e-06, |
|
"loss": 5.0106, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 17.451136816912644, |
|
"grad_norm": 3.0782690048217773, |
|
"learning_rate": 6.392600717989629e-06, |
|
"loss": 4.9894, |
|
"step": 87500 |
|
}, |
|
{ |
|
"epoch": 17.550857598723574, |
|
"grad_norm": 2.572624444961548, |
|
"learning_rate": 6.143298763462306e-06, |
|
"loss": 4.9901, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 17.650578380534505, |
|
"grad_norm": 2.726686477661133, |
|
"learning_rate": 5.894495412844037e-06, |
|
"loss": 5.0045, |
|
"step": 88500 |
|
}, |
|
{ |
|
"epoch": 17.75029916234543, |
|
"grad_norm": 2.79811429977417, |
|
"learning_rate": 5.645193458316714e-06, |
|
"loss": 4.9927, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 17.85001994415636, |
|
"grad_norm": 2.7528791427612305, |
|
"learning_rate": 5.39589150378939e-06, |
|
"loss": 4.9834, |
|
"step": 89500 |
|
}, |
|
{ |
|
"epoch": 17.949740725967292, |
|
"grad_norm": 2.85099196434021, |
|
"learning_rate": 5.146589549262067e-06, |
|
"loss": 4.9967, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 18.049461507778222, |
|
"grad_norm": 2.686501979827881, |
|
"learning_rate": 4.897786198643798e-06, |
|
"loss": 4.9948, |
|
"step": 90500 |
|
}, |
|
{ |
|
"epoch": 18.14918228958915, |
|
"grad_norm": 3.057145357131958, |
|
"learning_rate": 4.648484244116474e-06, |
|
"loss": 4.9885, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 18.24890307140008, |
|
"grad_norm": 2.915149211883545, |
|
"learning_rate": 4.399182289589151e-06, |
|
"loss": 4.9876, |
|
"step": 91500 |
|
}, |
|
{ |
|
"epoch": 18.34862385321101, |
|
"grad_norm": 2.907449960708618, |
|
"learning_rate": 4.149880335061828e-06, |
|
"loss": 5.0049, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 18.44834463502194, |
|
"grad_norm": 2.798488140106201, |
|
"learning_rate": 3.901076984443558e-06, |
|
"loss": 4.9643, |
|
"step": 92500 |
|
}, |
|
{ |
|
"epoch": 18.54806541683287, |
|
"grad_norm": 2.8539681434631348, |
|
"learning_rate": 3.6517750299162346e-06, |
|
"loss": 4.9917, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 18.647786198643796, |
|
"grad_norm": 2.7598862648010254, |
|
"learning_rate": 3.4024730753889117e-06, |
|
"loss": 4.9894, |
|
"step": 93500 |
|
}, |
|
{ |
|
"epoch": 18.747506980454727, |
|
"grad_norm": 2.577714443206787, |
|
"learning_rate": 3.153171120861588e-06, |
|
"loss": 4.9607, |
|
"step": 94000 |
|
}, |
|
{ |
|
"epoch": 18.847227762265657, |
|
"grad_norm": 2.8153915405273438, |
|
"learning_rate": 2.904367770243319e-06, |
|
"loss": 4.9905, |
|
"step": 94500 |
|
}, |
|
{ |
|
"epoch": 18.946948544076587, |
|
"grad_norm": 2.7925517559051514, |
|
"learning_rate": 2.6550658157159952e-06, |
|
"loss": 5.0036, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 19.046669325887514, |
|
"grad_norm": 2.60158371925354, |
|
"learning_rate": 2.405763861188672e-06, |
|
"loss": 4.9877, |
|
"step": 95500 |
|
}, |
|
{ |
|
"epoch": 19.146390107698444, |
|
"grad_norm": 2.7144620418548584, |
|
"learning_rate": 2.156461906661348e-06, |
|
"loss": 5.0011, |
|
"step": 96000 |
|
}, |
|
{ |
|
"epoch": 19.246110889509374, |
|
"grad_norm": 2.779771327972412, |
|
"learning_rate": 1.9076585560430792e-06, |
|
"loss": 4.9869, |
|
"step": 96500 |
|
}, |
|
{ |
|
"epoch": 19.345831671320305, |
|
"grad_norm": 2.8586442470550537, |
|
"learning_rate": 1.6583566015157561e-06, |
|
"loss": 4.9858, |
|
"step": 97000 |
|
}, |
|
{ |
|
"epoch": 19.44555245313123, |
|
"grad_norm": 2.572789430618286, |
|
"learning_rate": 1.4090546469884324e-06, |
|
"loss": 4.9934, |
|
"step": 97500 |
|
}, |
|
{ |
|
"epoch": 19.54527323494216, |
|
"grad_norm": 2.7267684936523438, |
|
"learning_rate": 1.159752692461109e-06, |
|
"loss": 4.9714, |
|
"step": 98000 |
|
}, |
|
{ |
|
"epoch": 19.644994016753092, |
|
"grad_norm": 2.7596216201782227, |
|
"learning_rate": 9.109493418428401e-07, |
|
"loss": 4.9698, |
|
"step": 98500 |
|
}, |
|
{ |
|
"epoch": 19.744714798564022, |
|
"grad_norm": 2.850459575653076, |
|
"learning_rate": 6.616473873155166e-07, |
|
"loss": 4.9653, |
|
"step": 99000 |
|
}, |
|
{ |
|
"epoch": 19.84443558037495, |
|
"grad_norm": 3.2066328525543213, |
|
"learning_rate": 4.1234543278819307e-07, |
|
"loss": 4.964, |
|
"step": 99500 |
|
}, |
|
{ |
|
"epoch": 19.94415636218588, |
|
"grad_norm": 3.170915365219116, |
|
"learning_rate": 1.6304347826086955e-07, |
|
"loss": 4.9803, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"step": 100280, |
|
"total_flos": 488922611712000.0, |
|
"train_loss": 5.48259629872295, |
|
"train_runtime": 4495.3437, |
|
"train_samples_per_second": 356.898, |
|
"train_steps_per_second": 22.308 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 100280, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 20, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 488922611712000.0, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|