|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.6642427632498952, |
|
"eval_steps": 5000, |
|
"global_step": 9500, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.006992029086841001, |
|
"grad_norm": 75.90129852294922, |
|
"learning_rate": 6.780845858091577e-07, |
|
"loss": 16.865, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.013984058173682002, |
|
"grad_norm": 71.16612243652344, |
|
"learning_rate": 1.3771408598392172e-06, |
|
"loss": 16.1556, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.020976087260523003, |
|
"grad_norm": 130.67613220214844, |
|
"learning_rate": 2.0761971338692767e-06, |
|
"loss": 14.8008, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.027968116347364003, |
|
"grad_norm": 36.4179801940918, |
|
"learning_rate": 2.7682628451590356e-06, |
|
"loss": 12.4025, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.034960145434205, |
|
"grad_norm": 34.748844146728516, |
|
"learning_rate": 3.467319119189095e-06, |
|
"loss": 9.7465, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.04195217452104601, |
|
"grad_norm": 8.039146423339844, |
|
"learning_rate": 4.166375393219155e-06, |
|
"loss": 8.448, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.04894420360788701, |
|
"grad_norm": 10.288451194763184, |
|
"learning_rate": 4.865431667249214e-06, |
|
"loss": 8.1951, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.05593623269472801, |
|
"grad_norm": 6.3236308097839355, |
|
"learning_rate": 5.564487941279273e-06, |
|
"loss": 8.1093, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.06292826178156902, |
|
"grad_norm": 15.230368614196777, |
|
"learning_rate": 6.263544215309333e-06, |
|
"loss": 8.0567, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.06992029086841, |
|
"grad_norm": 5.573168754577637, |
|
"learning_rate": 6.962600489339393e-06, |
|
"loss": 8.0401, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.07691231995525101, |
|
"grad_norm": 6.7740325927734375, |
|
"learning_rate": 7.661656763369451e-06, |
|
"loss": 7.9491, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.08390434904209201, |
|
"grad_norm": 6.167610168457031, |
|
"learning_rate": 8.360713037399512e-06, |
|
"loss": 7.9494, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.09089637812893302, |
|
"grad_norm": 5.53561544418335, |
|
"learning_rate": 9.05976931142957e-06, |
|
"loss": 7.9386, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.09788840721577402, |
|
"grad_norm": 6.056953430175781, |
|
"learning_rate": 9.758825585459631e-06, |
|
"loss": 7.9033, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.10488043630261502, |
|
"grad_norm": 5.160706043243408, |
|
"learning_rate": 1.045788185948969e-05, |
|
"loss": 7.9055, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.11187246538945601, |
|
"grad_norm": 5.525616645812988, |
|
"learning_rate": 1.115693813351975e-05, |
|
"loss": 7.9203, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.11886449447629702, |
|
"grad_norm": 5.553597927093506, |
|
"learning_rate": 1.185599440754981e-05, |
|
"loss": 7.8381, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.12585652356313803, |
|
"grad_norm": 4.8832478523254395, |
|
"learning_rate": 1.2555050681579868e-05, |
|
"loss": 7.8679, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.13284855264997902, |
|
"grad_norm": 5.308164596557617, |
|
"learning_rate": 1.3254106955609927e-05, |
|
"loss": 7.8686, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.13984058173682, |
|
"grad_norm": 6.790646553039551, |
|
"learning_rate": 1.3953163229639988e-05, |
|
"loss": 7.8252, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.14683261082366103, |
|
"grad_norm": 5.003779888153076, |
|
"learning_rate": 1.4652219503670046e-05, |
|
"loss": 7.856, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.15382463991050202, |
|
"grad_norm": 5.0926642417907715, |
|
"learning_rate": 1.5351275777700107e-05, |
|
"loss": 7.8301, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.16081666899734304, |
|
"grad_norm": 4.154193878173828, |
|
"learning_rate": 1.6050332051730168e-05, |
|
"loss": 7.8595, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.16780869808418403, |
|
"grad_norm": 4.168135166168213, |
|
"learning_rate": 1.6749388325760225e-05, |
|
"loss": 7.8138, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.17480072717102504, |
|
"grad_norm": 4.69076681137085, |
|
"learning_rate": 1.7448444599790285e-05, |
|
"loss": 7.812, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.18179275625786603, |
|
"grad_norm": 4.386991500854492, |
|
"learning_rate": 1.8147500873820346e-05, |
|
"loss": 7.8261, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.18878478534470702, |
|
"grad_norm": 5.777139186859131, |
|
"learning_rate": 1.8846557147850403e-05, |
|
"loss": 7.7988, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.19577681443154804, |
|
"grad_norm": 18.3277645111084, |
|
"learning_rate": 1.9545613421880464e-05, |
|
"loss": 7.7965, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.20276884351838903, |
|
"grad_norm": 4.5356669425964355, |
|
"learning_rate": 1.997280814201919e-05, |
|
"loss": 7.783, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.20976087260523005, |
|
"grad_norm": 5.4392476081848145, |
|
"learning_rate": 1.9895117119216877e-05, |
|
"loss": 7.7752, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.21675290169207104, |
|
"grad_norm": 5.365427017211914, |
|
"learning_rate": 1.981742609641456e-05, |
|
"loss": 7.7715, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.22374493077891203, |
|
"grad_norm": 4.111672401428223, |
|
"learning_rate": 1.9739735073612244e-05, |
|
"loss": 7.7903, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.23073695986575304, |
|
"grad_norm": 4.107003211975098, |
|
"learning_rate": 1.9662044050809928e-05, |
|
"loss": 7.7656, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.23772898895259403, |
|
"grad_norm": 4.667971611022949, |
|
"learning_rate": 1.9584353028007615e-05, |
|
"loss": 7.749, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.24472101803943505, |
|
"grad_norm": 4.010958194732666, |
|
"learning_rate": 1.95066620052053e-05, |
|
"loss": 7.7662, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.25171304712627607, |
|
"grad_norm": 4.07546329498291, |
|
"learning_rate": 1.9428970982402986e-05, |
|
"loss": 7.7492, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.25870507621311706, |
|
"grad_norm": 5.414000988006592, |
|
"learning_rate": 1.935127995960067e-05, |
|
"loss": 7.737, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.26569710529995805, |
|
"grad_norm": 4.4744648933410645, |
|
"learning_rate": 1.9273588936798357e-05, |
|
"loss": 7.7232, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.27268913438679904, |
|
"grad_norm": 4.141202926635742, |
|
"learning_rate": 1.919589791399604e-05, |
|
"loss": 7.7616, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.27968116347364, |
|
"grad_norm": 4.493043899536133, |
|
"learning_rate": 1.9118206891193724e-05, |
|
"loss": 7.7391, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.28667319256048107, |
|
"grad_norm": 4.083165168762207, |
|
"learning_rate": 1.904051586839141e-05, |
|
"loss": 7.7552, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.29366522164732206, |
|
"grad_norm": 8.303427696228027, |
|
"learning_rate": 1.8962824845589095e-05, |
|
"loss": 7.7273, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.30065725073416305, |
|
"grad_norm": 4.003914833068848, |
|
"learning_rate": 1.888513382278678e-05, |
|
"loss": 7.7216, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.30764927982100404, |
|
"grad_norm": 4.859315395355225, |
|
"learning_rate": 1.8807442799984462e-05, |
|
"loss": 7.7371, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.31464130890784503, |
|
"grad_norm": 4.693440914154053, |
|
"learning_rate": 1.872975177718215e-05, |
|
"loss": 7.7426, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.3216333379946861, |
|
"grad_norm": 4.81652307510376, |
|
"learning_rate": 1.8652060754379833e-05, |
|
"loss": 7.7406, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.32862536708152706, |
|
"grad_norm": 3.861663341522217, |
|
"learning_rate": 1.8574369731577516e-05, |
|
"loss": 7.712, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.33561739616836805, |
|
"grad_norm": 4.218888282775879, |
|
"learning_rate": 1.8496678708775203e-05, |
|
"loss": 7.7466, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.34260942525520904, |
|
"grad_norm": 8.032899856567383, |
|
"learning_rate": 1.8418987685972887e-05, |
|
"loss": 7.7058, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.3496014543420501, |
|
"grad_norm": 4.292692184448242, |
|
"learning_rate": 1.8342073573398596e-05, |
|
"loss": 7.7139, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.3496014543420501, |
|
"eval_loss": 7.6895527839660645, |
|
"eval_runtime": 270.2677, |
|
"eval_samples_per_second": 1696.499, |
|
"eval_steps_per_second": 13.257, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.3565934834288911, |
|
"grad_norm": 4.507449626922607, |
|
"learning_rate": 1.826438255059628e-05, |
|
"loss": 7.7457, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.36358551251573207, |
|
"grad_norm": 10.12509536743164, |
|
"learning_rate": 1.8186691527793966e-05, |
|
"loss": 7.7172, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.37057754160257306, |
|
"grad_norm": 4.254384517669678, |
|
"learning_rate": 1.810900050499165e-05, |
|
"loss": 7.739, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.37756957068941405, |
|
"grad_norm": 6.059901237487793, |
|
"learning_rate": 1.8031309482189334e-05, |
|
"loss": 7.7259, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.3845615997762551, |
|
"grad_norm": 5.770705223083496, |
|
"learning_rate": 1.7953618459387017e-05, |
|
"loss": 7.6977, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.3915536288630961, |
|
"grad_norm": 6.927788257598877, |
|
"learning_rate": 1.7875927436584705e-05, |
|
"loss": 7.7237, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.39854565794993707, |
|
"grad_norm": 4.323265552520752, |
|
"learning_rate": 1.7798236413782388e-05, |
|
"loss": 7.7118, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.40553768703677806, |
|
"grad_norm": 4.448848247528076, |
|
"learning_rate": 1.7720545390980072e-05, |
|
"loss": 7.7099, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.41252971612361905, |
|
"grad_norm": 66.31098175048828, |
|
"learning_rate": 1.764285436817776e-05, |
|
"loss": 7.7142, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.4195217452104601, |
|
"grad_norm": 6.65512228012085, |
|
"learning_rate": 1.7565163345375443e-05, |
|
"loss": 7.6885, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.4265137742973011, |
|
"grad_norm": 4.743120193481445, |
|
"learning_rate": 1.748747232257313e-05, |
|
"loss": 7.6799, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.4335058033841421, |
|
"grad_norm": 4.90665864944458, |
|
"learning_rate": 1.7409781299770813e-05, |
|
"loss": 7.7039, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.44049783247098306, |
|
"grad_norm": 6.465912342071533, |
|
"learning_rate": 1.7332090276968497e-05, |
|
"loss": 7.6825, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.44748986155782405, |
|
"grad_norm": 3.6341958045959473, |
|
"learning_rate": 1.7254399254166184e-05, |
|
"loss": 7.6846, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.4544818906446651, |
|
"grad_norm": 3.931257486343384, |
|
"learning_rate": 1.7176708231363868e-05, |
|
"loss": 7.7078, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.4614739197315061, |
|
"grad_norm": 4.2776665687561035, |
|
"learning_rate": 1.709901720856155e-05, |
|
"loss": 7.6945, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.4684659488183471, |
|
"grad_norm": 4.510386943817139, |
|
"learning_rate": 1.702132618575924e-05, |
|
"loss": 7.7017, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.47545797790518807, |
|
"grad_norm": 5.780529975891113, |
|
"learning_rate": 1.6943635162956922e-05, |
|
"loss": 7.6781, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.4824500069920291, |
|
"grad_norm": 5.802525997161865, |
|
"learning_rate": 1.6865944140154606e-05, |
|
"loss": 7.6885, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.4894420360788701, |
|
"grad_norm": 5.534703254699707, |
|
"learning_rate": 1.6789030027580314e-05, |
|
"loss": 7.7426, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.4964340651657111, |
|
"grad_norm": 5.897773742675781, |
|
"learning_rate": 1.6711339004778e-05, |
|
"loss": 7.6809, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.5034260942525521, |
|
"grad_norm": 5.115645408630371, |
|
"learning_rate": 1.6633647981975685e-05, |
|
"loss": 7.6977, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.5104181233393931, |
|
"grad_norm": 7.200738430023193, |
|
"learning_rate": 1.655595695917337e-05, |
|
"loss": 7.6964, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.5174101524262341, |
|
"grad_norm": 5.526965618133545, |
|
"learning_rate": 1.6478265936371052e-05, |
|
"loss": 7.6834, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.524402181513075, |
|
"grad_norm": 8.350557327270508, |
|
"learning_rate": 1.640057491356874e-05, |
|
"loss": 7.6593, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.5313942105999161, |
|
"grad_norm": 4.367435455322266, |
|
"learning_rate": 1.6322883890766423e-05, |
|
"loss": 7.6745, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.5383862396867571, |
|
"grad_norm": 7.726882457733154, |
|
"learning_rate": 1.6245192867964107e-05, |
|
"loss": 7.6587, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.5453782687735981, |
|
"grad_norm": 5.289243221282959, |
|
"learning_rate": 1.6167501845161794e-05, |
|
"loss": 7.6389, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.5523702978604391, |
|
"grad_norm": 4.716782093048096, |
|
"learning_rate": 1.6089810822359478e-05, |
|
"loss": 7.6298, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.55936232694728, |
|
"grad_norm": 4.2459797859191895, |
|
"learning_rate": 1.601211979955716e-05, |
|
"loss": 7.6693, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.5663543560341211, |
|
"grad_norm": 4.4402594566345215, |
|
"learning_rate": 1.5934428776754848e-05, |
|
"loss": 7.6454, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.5733463851209621, |
|
"grad_norm": 20.979795455932617, |
|
"learning_rate": 1.5856737753952532e-05, |
|
"loss": 7.6491, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.5803384142078031, |
|
"grad_norm": 4.494659900665283, |
|
"learning_rate": 1.5779046731150216e-05, |
|
"loss": 7.661, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.5873304432946441, |
|
"grad_norm": 4.788604736328125, |
|
"learning_rate": 1.57013557083479e-05, |
|
"loss": 7.6525, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.5943224723814851, |
|
"grad_norm": 4.576944351196289, |
|
"learning_rate": 1.5623664685545586e-05, |
|
"loss": 7.6669, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.6013145014683261, |
|
"grad_norm": 5.762516498565674, |
|
"learning_rate": 1.5545973662743273e-05, |
|
"loss": 7.6379, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.6083065305551671, |
|
"grad_norm": 4.864853858947754, |
|
"learning_rate": 1.5468282639940957e-05, |
|
"loss": 7.6706, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.6152985596420081, |
|
"grad_norm": 11.886726379394531, |
|
"learning_rate": 1.539059161713864e-05, |
|
"loss": 7.6487, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.6222905887288491, |
|
"grad_norm": 4.3859052658081055, |
|
"learning_rate": 1.5312900594336328e-05, |
|
"loss": 7.6607, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.6292826178156901, |
|
"grad_norm": 4.877004146575928, |
|
"learning_rate": 1.5235209571534011e-05, |
|
"loss": 7.6334, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.6362746469025311, |
|
"grad_norm": 10.190913200378418, |
|
"learning_rate": 1.5158295458959717e-05, |
|
"loss": 7.6891, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.6432666759893721, |
|
"grad_norm": 4.916624546051025, |
|
"learning_rate": 1.5080604436157402e-05, |
|
"loss": 7.734, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.6502587050762131, |
|
"grad_norm": 5.250345230102539, |
|
"learning_rate": 1.5002913413355089e-05, |
|
"loss": 7.6283, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.6572507341630541, |
|
"grad_norm": 4.70583963394165, |
|
"learning_rate": 1.4925222390552775e-05, |
|
"loss": 7.6461, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.6642427632498952, |
|
"grad_norm": 4.943760871887207, |
|
"learning_rate": 1.4847531367750458e-05, |
|
"loss": 7.623, |
|
"step": 9500 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 28604, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 128, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|