|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.999792917788362, |
|
"eval_steps": 500, |
|
"global_step": 1207, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0008283288465520812, |
|
"grad_norm": 24.23784679529398, |
|
"learning_rate": 8.264462809917357e-08, |
|
"loss": 1.4304, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0041416442327604054, |
|
"grad_norm": 23.234915420644615, |
|
"learning_rate": 4.132231404958678e-07, |
|
"loss": 1.4144, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.008283288465520811, |
|
"grad_norm": 8.222025109804092, |
|
"learning_rate": 8.264462809917356e-07, |
|
"loss": 1.307, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.012424932698281217, |
|
"grad_norm": 10.133789775285283, |
|
"learning_rate": 1.2396694214876035e-06, |
|
"loss": 1.1636, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.016566576931041622, |
|
"grad_norm": 2.9246570997342918, |
|
"learning_rate": 1.6528925619834712e-06, |
|
"loss": 1.0151, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.02070822116380203, |
|
"grad_norm": 2.7388624531591717, |
|
"learning_rate": 2.066115702479339e-06, |
|
"loss": 0.9484, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.024849865396562434, |
|
"grad_norm": 2.365875950285307, |
|
"learning_rate": 2.479338842975207e-06, |
|
"loss": 0.9332, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.028991509629322842, |
|
"grad_norm": 2.374922722027774, |
|
"learning_rate": 2.8925619834710743e-06, |
|
"loss": 0.9044, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.033133153862083244, |
|
"grad_norm": 2.238010214125612, |
|
"learning_rate": 3.3057851239669424e-06, |
|
"loss": 0.8913, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.03727479809484365, |
|
"grad_norm": 2.4490973058865078, |
|
"learning_rate": 3.71900826446281e-06, |
|
"loss": 0.8833, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.04141644232760406, |
|
"grad_norm": 2.469516335084805, |
|
"learning_rate": 4.132231404958678e-06, |
|
"loss": 0.8813, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.04555808656036447, |
|
"grad_norm": 2.2371905595594495, |
|
"learning_rate": 4.5454545454545455e-06, |
|
"loss": 0.8533, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.04969973079312487, |
|
"grad_norm": 2.409737572391393, |
|
"learning_rate": 4.958677685950414e-06, |
|
"loss": 0.8581, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.05384137502588528, |
|
"grad_norm": 2.2291530452515977, |
|
"learning_rate": 5.371900826446281e-06, |
|
"loss": 0.8519, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.057983019258645685, |
|
"grad_norm": 2.35453266908653, |
|
"learning_rate": 5.785123966942149e-06, |
|
"loss": 0.8515, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.062124663491406086, |
|
"grad_norm": 2.336463988259032, |
|
"learning_rate": 6.198347107438017e-06, |
|
"loss": 0.8295, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.06626630772416649, |
|
"grad_norm": 2.4980329883385624, |
|
"learning_rate": 6.611570247933885e-06, |
|
"loss": 0.832, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.0704079519569269, |
|
"grad_norm": 2.6275854974083384, |
|
"learning_rate": 7.0247933884297525e-06, |
|
"loss": 0.816, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.0745495961896873, |
|
"grad_norm": 2.684536070785031, |
|
"learning_rate": 7.43801652892562e-06, |
|
"loss": 0.8162, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.07869124042244771, |
|
"grad_norm": 2.805145187507444, |
|
"learning_rate": 7.851239669421489e-06, |
|
"loss": 0.8309, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.08283288465520812, |
|
"grad_norm": 2.4685576686167874, |
|
"learning_rate": 8.264462809917356e-06, |
|
"loss": 0.8155, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.08697452888796853, |
|
"grad_norm": 2.540099419306748, |
|
"learning_rate": 8.677685950413224e-06, |
|
"loss": 0.8071, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.09111617312072894, |
|
"grad_norm": 2.562872480103855, |
|
"learning_rate": 9.090909090909091e-06, |
|
"loss": 0.8029, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.09525781735348933, |
|
"grad_norm": 2.5970244377170313, |
|
"learning_rate": 9.50413223140496e-06, |
|
"loss": 0.7873, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.09939946158624974, |
|
"grad_norm": 2.4308464508943346, |
|
"learning_rate": 9.917355371900828e-06, |
|
"loss": 0.7915, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.10354110581901015, |
|
"grad_norm": 2.6905207268547158, |
|
"learning_rate": 9.999665269535307e-06, |
|
"loss": 0.7993, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.10768275005177055, |
|
"grad_norm": 34.133455991097065, |
|
"learning_rate": 9.998305503833872e-06, |
|
"loss": 0.8009, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.11182439428453096, |
|
"grad_norm": 2.3166119518346866, |
|
"learning_rate": 9.995900066492902e-06, |
|
"loss": 0.785, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.11596603851729137, |
|
"grad_norm": 2.5556998820575076, |
|
"learning_rate": 9.992449460742464e-06, |
|
"loss": 0.8134, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.12010768275005176, |
|
"grad_norm": 2.615095977165256, |
|
"learning_rate": 9.98795440846732e-06, |
|
"loss": 0.7898, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.12424932698281217, |
|
"grad_norm": 2.729536225853297, |
|
"learning_rate": 9.982415850055902e-06, |
|
"loss": 0.7681, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.12839097121557258, |
|
"grad_norm": 3.895339838479258, |
|
"learning_rate": 9.975834944203581e-06, |
|
"loss": 0.7503, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.13253261544833297, |
|
"grad_norm": 2.4875917416718187, |
|
"learning_rate": 9.968213067670265e-06, |
|
"loss": 0.7632, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.1366742596810934, |
|
"grad_norm": 2.5249141397756847, |
|
"learning_rate": 9.959551814992364e-06, |
|
"loss": 0.7444, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.1408159039138538, |
|
"grad_norm": 2.6613194516011656, |
|
"learning_rate": 9.949852998149217e-06, |
|
"loss": 0.7592, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.1449575481466142, |
|
"grad_norm": 2.535010612771426, |
|
"learning_rate": 9.939118646184007e-06, |
|
"loss": 0.7625, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.1490991923793746, |
|
"grad_norm": 2.658211254916015, |
|
"learning_rate": 9.927351004779275e-06, |
|
"loss": 0.7405, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.15324083661213503, |
|
"grad_norm": 2.343186435782198, |
|
"learning_rate": 9.914552535787122e-06, |
|
"loss": 0.7285, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.15738248084489542, |
|
"grad_norm": 2.4661024566837653, |
|
"learning_rate": 9.900725916714157e-06, |
|
"loss": 0.7405, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.16152412507765582, |
|
"grad_norm": 2.7732706905434497, |
|
"learning_rate": 9.885874040161373e-06, |
|
"loss": 0.7373, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.16566576931041624, |
|
"grad_norm": 2.523332136572105, |
|
"learning_rate": 9.87000001321898e-06, |
|
"loss": 0.7261, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.16980741354317663, |
|
"grad_norm": 2.2311284634557937, |
|
"learning_rate": 9.853107156816393e-06, |
|
"loss": 0.7184, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.17394905777593705, |
|
"grad_norm": 2.379531410344561, |
|
"learning_rate": 9.835199005027477e-06, |
|
"loss": 0.7144, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.17809070200869745, |
|
"grad_norm": 2.3058139982021215, |
|
"learning_rate": 9.816279304331202e-06, |
|
"loss": 0.7252, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.18223234624145787, |
|
"grad_norm": 2.2361882677362974, |
|
"learning_rate": 9.79635201282785e-06, |
|
"loss": 0.7151, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.18637399047421827, |
|
"grad_norm": 2.7702057401479006, |
|
"learning_rate": 9.775421299410977e-06, |
|
"loss": 0.7045, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.19051563470697866, |
|
"grad_norm": 2.5524695305725404, |
|
"learning_rate": 9.753491542895237e-06, |
|
"loss": 0.715, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.19465727893973908, |
|
"grad_norm": 2.4487445215481305, |
|
"learning_rate": 9.730567331100333e-06, |
|
"loss": 0.7082, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.19879892317249948, |
|
"grad_norm": 2.2482641063346116, |
|
"learning_rate": 9.706653459891207e-06, |
|
"loss": 0.699, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.2029405674052599, |
|
"grad_norm": 2.451295868720965, |
|
"learning_rate": 9.681754932174719e-06, |
|
"loss": 0.6913, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.2070822116380203, |
|
"grad_norm": 2.56626439770444, |
|
"learning_rate": 9.655876956853025e-06, |
|
"loss": 0.6819, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.2112238558707807, |
|
"grad_norm": 2.490881018154125, |
|
"learning_rate": 9.629024947733836e-06, |
|
"loss": 0.6843, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.2153655001035411, |
|
"grad_norm": 2.4655094247796283, |
|
"learning_rate": 9.601204522397826e-06, |
|
"loss": 0.6874, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.2195071443363015, |
|
"grad_norm": 2.6029112443690625, |
|
"learning_rate": 9.572421501023403e-06, |
|
"loss": 0.6773, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.22364878856906192, |
|
"grad_norm": 2.303292362769761, |
|
"learning_rate": 9.5426819051691e-06, |
|
"loss": 0.6669, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.22779043280182232, |
|
"grad_norm": 2.2947735744081568, |
|
"learning_rate": 9.511991956513828e-06, |
|
"loss": 0.6636, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.23193207703458274, |
|
"grad_norm": 2.3561731246407045, |
|
"learning_rate": 9.480358075555278e-06, |
|
"loss": 0.6671, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.23607372126734313, |
|
"grad_norm": 2.4666706522590442, |
|
"learning_rate": 9.447786880266706e-06, |
|
"loss": 0.6618, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.24021536550010353, |
|
"grad_norm": 2.186824941231579, |
|
"learning_rate": 9.414285184712432e-06, |
|
"loss": 0.6619, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.24435700973286395, |
|
"grad_norm": 2.1988625035941047, |
|
"learning_rate": 9.37985999762229e-06, |
|
"loss": 0.6473, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.24849865396562434, |
|
"grad_norm": 2.43262245251786, |
|
"learning_rate": 9.344518520925377e-06, |
|
"loss": 0.6534, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.25264029819838474, |
|
"grad_norm": 2.3140467166469527, |
|
"learning_rate": 9.308268148243355e-06, |
|
"loss": 0.637, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.25678194243114516, |
|
"grad_norm": 2.5057247056088996, |
|
"learning_rate": 9.271116463343692e-06, |
|
"loss": 0.6417, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.2609235866639056, |
|
"grad_norm": 2.288973054378257, |
|
"learning_rate": 9.23307123855307e-06, |
|
"loss": 0.663, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.26506523089666595, |
|
"grad_norm": 2.416778922999839, |
|
"learning_rate": 9.194140433131397e-06, |
|
"loss": 0.6552, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.26920687512942637, |
|
"grad_norm": 2.2651292744956244, |
|
"learning_rate": 9.154332191606671e-06, |
|
"loss": 0.6267, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.2733485193621868, |
|
"grad_norm": 2.3801866420145954, |
|
"learning_rate": 9.113654842071114e-06, |
|
"loss": 0.6306, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.2774901635949472, |
|
"grad_norm": 2.5824252886751964, |
|
"learning_rate": 9.072116894438885e-06, |
|
"loss": 0.6369, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.2816318078277076, |
|
"grad_norm": 2.3287880261931875, |
|
"learning_rate": 9.029727038665765e-06, |
|
"loss": 0.6252, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.285773452060468, |
|
"grad_norm": 2.3253013276198358, |
|
"learning_rate": 8.986494142931168e-06, |
|
"loss": 0.6165, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.2899150962932284, |
|
"grad_norm": 2.3141576363210086, |
|
"learning_rate": 8.94242725178288e-06, |
|
"loss": 0.6003, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.2940567405259888, |
|
"grad_norm": 2.2453531818296595, |
|
"learning_rate": 8.89753558424488e-06, |
|
"loss": 0.6, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.2981983847587492, |
|
"grad_norm": 2.224520930970933, |
|
"learning_rate": 8.851828531888692e-06, |
|
"loss": 0.6117, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.30234002899150964, |
|
"grad_norm": 2.312078602771557, |
|
"learning_rate": 8.805315656868587e-06, |
|
"loss": 0.6067, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.30648167322427006, |
|
"grad_norm": 2.287152946702887, |
|
"learning_rate": 8.75800668992117e-06, |
|
"loss": 0.5979, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.3106233174570304, |
|
"grad_norm": 2.0335651194660596, |
|
"learning_rate": 8.709911528329623e-06, |
|
"loss": 0.5911, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.31476496168979085, |
|
"grad_norm": 2.190338892871162, |
|
"learning_rate": 8.661040233853166e-06, |
|
"loss": 0.5884, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.31890660592255127, |
|
"grad_norm": 2.198737460879657, |
|
"learning_rate": 8.611403030622074e-06, |
|
"loss": 0.578, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.32304825015531163, |
|
"grad_norm": 2.320069378855735, |
|
"learning_rate": 8.561010302998734e-06, |
|
"loss": 0.5809, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.32718989438807206, |
|
"grad_norm": 2.1718618035260793, |
|
"learning_rate": 8.509872593405189e-06, |
|
"loss": 0.5834, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.3313315386208325, |
|
"grad_norm": 2.306249404711425, |
|
"learning_rate": 8.458000600117604e-06, |
|
"loss": 0.5817, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.3354731828535929, |
|
"grad_norm": 2.2320669776692377, |
|
"learning_rate": 8.40540517502813e-06, |
|
"loss": 0.579, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.33961482708635327, |
|
"grad_norm": 2.380541232391175, |
|
"learning_rate": 8.35209732137463e-06, |
|
"loss": 0.5657, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.3437564713191137, |
|
"grad_norm": 2.270445952132619, |
|
"learning_rate": 8.298088191438753e-06, |
|
"loss": 0.5569, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.3478981155518741, |
|
"grad_norm": 2.353449916475927, |
|
"learning_rate": 8.243389084212808e-06, |
|
"loss": 0.5642, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.3520397597846345, |
|
"grad_norm": 2.321086875711968, |
|
"learning_rate": 8.188011443035962e-06, |
|
"loss": 0.5519, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.3561814040173949, |
|
"grad_norm": 2.112886796528374, |
|
"learning_rate": 8.131966853200226e-06, |
|
"loss": 0.561, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.3603230482501553, |
|
"grad_norm": 2.354896774745211, |
|
"learning_rate": 8.075267039526764e-06, |
|
"loss": 0.5586, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.36446469248291574, |
|
"grad_norm": 2.291143610369071, |
|
"learning_rate": 8.017923863912989e-06, |
|
"loss": 0.5491, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.3686063367156761, |
|
"grad_norm": 2.2300308493171115, |
|
"learning_rate": 7.959949322850994e-06, |
|
"loss": 0.5562, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.37274798094843653, |
|
"grad_norm": 2.28359859552775, |
|
"learning_rate": 7.901355544917827e-06, |
|
"loss": 0.5494, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.37688962518119695, |
|
"grad_norm": 2.251157158354555, |
|
"learning_rate": 7.842154788238124e-06, |
|
"loss": 0.5424, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.3810312694139573, |
|
"grad_norm": 2.285389132836112, |
|
"learning_rate": 7.782359437919644e-06, |
|
"loss": 0.5423, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.38517291364671774, |
|
"grad_norm": 2.578954308129232, |
|
"learning_rate": 7.721982003462255e-06, |
|
"loss": 0.5335, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.38931455787947816, |
|
"grad_norm": 2.1334475270465774, |
|
"learning_rate": 7.661035116140856e-06, |
|
"loss": 0.5342, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.3934562021122386, |
|
"grad_norm": 2.3233034598991797, |
|
"learning_rate": 7.599531526362873e-06, |
|
"loss": 0.5358, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.39759784634499895, |
|
"grad_norm": 2.2758511904313785, |
|
"learning_rate": 7.537484101000787e-06, |
|
"loss": 0.532, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.4017394905777594, |
|
"grad_norm": 2.151297761784936, |
|
"learning_rate": 7.474905820700334e-06, |
|
"loss": 0.5135, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.4058811348105198, |
|
"grad_norm": 2.3747508627507066, |
|
"learning_rate": 7.411809777164873e-06, |
|
"loss": 0.5306, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.41002277904328016, |
|
"grad_norm": 2.202301523886571, |
|
"learning_rate": 7.3482091704165405e-06, |
|
"loss": 0.5247, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.4141644232760406, |
|
"grad_norm": 2.153444078468182, |
|
"learning_rate": 7.284117306034733e-06, |
|
"loss": 0.5243, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.418306067508801, |
|
"grad_norm": 2.171064236273241, |
|
"learning_rate": 7.219547592372512e-06, |
|
"loss": 0.5187, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.4224477117415614, |
|
"grad_norm": 2.2667901238143586, |
|
"learning_rate": 7.15451353775151e-06, |
|
"loss": 0.5126, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.4265893559743218, |
|
"grad_norm": 2.4217664685492997, |
|
"learning_rate": 7.089028747635908e-06, |
|
"loss": 0.5166, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.4307310002070822, |
|
"grad_norm": 2.269198948091921, |
|
"learning_rate": 7.023106921786118e-06, |
|
"loss": 0.5102, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.43487264443984264, |
|
"grad_norm": 2.3389291631103717, |
|
"learning_rate": 6.956761851392706e-06, |
|
"loss": 0.5147, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.439014288672603, |
|
"grad_norm": 2.130307708383959, |
|
"learning_rate": 6.890007416191209e-06, |
|
"loss": 0.5, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.4431559329053634, |
|
"grad_norm": 2.2319755819000133, |
|
"learning_rate": 6.822857581558423e-06, |
|
"loss": 0.5031, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.44729757713812385, |
|
"grad_norm": 2.201034276249066, |
|
"learning_rate": 6.7553263955907755e-06, |
|
"loss": 0.5003, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.4514392213708842, |
|
"grad_norm": 2.1906710526713877, |
|
"learning_rate": 6.687427986165379e-06, |
|
"loss": 0.498, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.45558086560364464, |
|
"grad_norm": 2.421545219253794, |
|
"learning_rate": 6.6191765579844205e-06, |
|
"loss": 0.4996, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.45972250983640506, |
|
"grad_norm": 2.2909926333329977, |
|
"learning_rate": 6.550586389603451e-06, |
|
"loss": 0.4969, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.4638641540691655, |
|
"grad_norm": 2.3444032170833116, |
|
"learning_rate": 6.481671830444243e-06, |
|
"loss": 0.4945, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.46800579830192585, |
|
"grad_norm": 2.1725157981177916, |
|
"learning_rate": 6.412447297792818e-06, |
|
"loss": 0.4863, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.47214744253468627, |
|
"grad_norm": 2.1573058335782753, |
|
"learning_rate": 6.3429272737832726e-06, |
|
"loss": 0.4891, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.4762890867674467, |
|
"grad_norm": 2.2084194407586373, |
|
"learning_rate": 6.273126302368037e-06, |
|
"loss": 0.487, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.48043073100020706, |
|
"grad_norm": 2.21344464284995, |
|
"learning_rate": 6.203058986275207e-06, |
|
"loss": 0.4857, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.4845723752329675, |
|
"grad_norm": 2.2203608524225666, |
|
"learning_rate": 6.132739983953579e-06, |
|
"loss": 0.4828, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.4887140194657279, |
|
"grad_norm": 2.113371175419138, |
|
"learning_rate": 6.062184006506027e-06, |
|
"loss": 0.4826, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.4928556636984883, |
|
"grad_norm": 2.3454680885393464, |
|
"learning_rate": 5.991405814611855e-06, |
|
"loss": 0.4676, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.4969973079312487, |
|
"grad_norm": 2.321964160860843, |
|
"learning_rate": 5.920420215438794e-06, |
|
"loss": 0.4737, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.5011389521640092, |
|
"grad_norm": 2.1991634408721943, |
|
"learning_rate": 5.849242059545259e-06, |
|
"loss": 0.465, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.5052805963967695, |
|
"grad_norm": 2.165800122945675, |
|
"learning_rate": 5.777886237773542e-06, |
|
"loss": 0.4623, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.5094222406295299, |
|
"grad_norm": 2.1403595970614133, |
|
"learning_rate": 5.706367678134562e-06, |
|
"loss": 0.4767, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.5135638848622903, |
|
"grad_norm": 2.1427662247598986, |
|
"learning_rate": 5.634701342684852e-06, |
|
"loss": 0.4607, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.5177055290950507, |
|
"grad_norm": 2.0626908104196673, |
|
"learning_rate": 5.562902224396416e-06, |
|
"loss": 0.4617, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.5218471733278112, |
|
"grad_norm": 2.0098208614700326, |
|
"learning_rate": 5.49098534402012e-06, |
|
"loss": 0.4618, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.5259888175605716, |
|
"grad_norm": 2.0735743487283607, |
|
"learning_rate": 5.418965746943281e-06, |
|
"loss": 0.459, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.5301304617933319, |
|
"grad_norm": 2.024087007627894, |
|
"learning_rate": 5.34685850004208e-06, |
|
"loss": 0.4539, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.5342721060260923, |
|
"grad_norm": 2.130234594696483, |
|
"learning_rate": 5.2746786885295034e-06, |
|
"loss": 0.453, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.5384137502588527, |
|
"grad_norm": 2.0986269383387706, |
|
"learning_rate": 5.2024414127994325e-06, |
|
"loss": 0.4538, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.5425553944916132, |
|
"grad_norm": 2.209414073449144, |
|
"learning_rate": 5.13016178526756e-06, |
|
"loss": 0.4448, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.5466970387243736, |
|
"grad_norm": 2.15118663280229, |
|
"learning_rate": 5.057854927209804e-06, |
|
"loss": 0.4446, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.550838682957134, |
|
"grad_norm": 2.033860154274491, |
|
"learning_rate": 4.985535965598843e-06, |
|
"loss": 0.4447, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.5549803271898944, |
|
"grad_norm": 2.129135189760016, |
|
"learning_rate": 4.913220029939491e-06, |
|
"loss": 0.4512, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.5591219714226547, |
|
"grad_norm": 2.090833400754654, |
|
"learning_rate": 4.840922249103506e-06, |
|
"loss": 0.4467, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.5632636156554152, |
|
"grad_norm": 2.0680502896348836, |
|
"learning_rate": 4.7686577481645745e-06, |
|
"loss": 0.4316, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.5674052598881756, |
|
"grad_norm": 2.07449578249477, |
|
"learning_rate": 4.696441645234042e-06, |
|
"loss": 0.4421, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.571546904120936, |
|
"grad_norm": 2.2932786860060235, |
|
"learning_rate": 4.624289048298147e-06, |
|
"loss": 0.4433, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.5756885483536964, |
|
"grad_norm": 2.058591317854592, |
|
"learning_rate": 4.55221505205734e-06, |
|
"loss": 0.4298, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.5798301925864568, |
|
"grad_norm": 2.3196625048015704, |
|
"learning_rate": 4.480234734768393e-06, |
|
"loss": 0.4326, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.5839718368192173, |
|
"grad_norm": 2.073473068971661, |
|
"learning_rate": 4.408363155089952e-06, |
|
"loss": 0.4335, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.5881134810519776, |
|
"grad_norm": 2.029766952429309, |
|
"learning_rate": 4.3366153489321855e-06, |
|
"loss": 0.4273, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.592255125284738, |
|
"grad_norm": 2.1078664280012007, |
|
"learning_rate": 4.265006326311199e-06, |
|
"loss": 0.415, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.5963967695174984, |
|
"grad_norm": 2.0038145759132915, |
|
"learning_rate": 4.1935510682088545e-06, |
|
"loss": 0.4244, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.6005384137502588, |
|
"grad_norm": 2.0468827081169976, |
|
"learning_rate": 4.122264523438668e-06, |
|
"loss": 0.4226, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.6046800579830193, |
|
"grad_norm": 2.0660110564277923, |
|
"learning_rate": 4.051161605518453e-06, |
|
"loss": 0.4222, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.6088217022157797, |
|
"grad_norm": 2.088150514221739, |
|
"learning_rate": 3.980257189550316e-06, |
|
"loss": 0.433, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.6129633464485401, |
|
"grad_norm": 2.1225320313052447, |
|
"learning_rate": 3.909566109108727e-06, |
|
"loss": 0.4161, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.6171049906813004, |
|
"grad_norm": 2.0990451349998227, |
|
"learning_rate": 3.839103153137247e-06, |
|
"loss": 0.417, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.6212466349140608, |
|
"grad_norm": 2.104242942119707, |
|
"learning_rate": 3.768883062854598e-06, |
|
"loss": 0.4081, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.6253882791468213, |
|
"grad_norm": 2.051137757693821, |
|
"learning_rate": 3.6989205286707398e-06, |
|
"loss": 0.4108, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.6295299233795817, |
|
"grad_norm": 2.099437836109555, |
|
"learning_rate": 3.6292301871135425e-06, |
|
"loss": 0.411, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.6336715676123421, |
|
"grad_norm": 2.056296390597689, |
|
"learning_rate": 3.55982661776676e-06, |
|
"loss": 0.4047, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.6378132118451025, |
|
"grad_norm": 2.0454981613636503, |
|
"learning_rate": 3.4907243402199013e-06, |
|
"loss": 0.4044, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.641954856077863, |
|
"grad_norm": 2.0625058430327092, |
|
"learning_rate": 3.4219378110306523e-06, |
|
"loss": 0.4103, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.6460965003106233, |
|
"grad_norm": 2.0640184519207123, |
|
"learning_rate": 3.353481420700495e-06, |
|
"loss": 0.4109, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.6502381445433837, |
|
"grad_norm": 2.096435391565135, |
|
"learning_rate": 3.285369490664133e-06, |
|
"loss": 0.4103, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.6543797887761441, |
|
"grad_norm": 2.0511959776297983, |
|
"learning_rate": 3.2176162702933816e-06, |
|
"loss": 0.3991, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.6585214330089045, |
|
"grad_norm": 2.004915794597741, |
|
"learning_rate": 3.150235933916115e-06, |
|
"loss": 0.401, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.662663077241665, |
|
"grad_norm": 1.989931694202458, |
|
"learning_rate": 3.0832425778509235e-06, |
|
"loss": 0.4015, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.6668047214744254, |
|
"grad_norm": 2.0927029677784383, |
|
"learning_rate": 3.0166502174581012e-06, |
|
"loss": 0.3904, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.6709463657071858, |
|
"grad_norm": 2.05569943435716, |
|
"learning_rate": 2.950472784207544e-06, |
|
"loss": 0.3976, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.6750880099399461, |
|
"grad_norm": 2.0070990032518456, |
|
"learning_rate": 2.8847241227642255e-06, |
|
"loss": 0.3855, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.6792296541727065, |
|
"grad_norm": 1.9791880977464777, |
|
"learning_rate": 2.819417988091814e-06, |
|
"loss": 0.3831, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.683371298405467, |
|
"grad_norm": 1.9874217780709027, |
|
"learning_rate": 2.754568042575061e-06, |
|
"loss": 0.3928, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.6875129426382274, |
|
"grad_norm": 1.9444889966862584, |
|
"learning_rate": 2.6901878531615677e-06, |
|
"loss": 0.3967, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.6916545868709878, |
|
"grad_norm": 2.190695001941064, |
|
"learning_rate": 2.6262908885235046e-06, |
|
"loss": 0.384, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.6957962311037482, |
|
"grad_norm": 1.991552664612379, |
|
"learning_rate": 2.5628905162398797e-06, |
|
"loss": 0.3831, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.6999378753365086, |
|
"grad_norm": 2.017973650880143, |
|
"learning_rate": 2.5000000000000015e-06, |
|
"loss": 0.3851, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.704079519569269, |
|
"grad_norm": 2.138390804530181, |
|
"learning_rate": 2.4376324968286154e-06, |
|
"loss": 0.3777, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.7082211638020294, |
|
"grad_norm": 2.0021049336250814, |
|
"learning_rate": 2.375801054333409e-06, |
|
"loss": 0.3891, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.7123628080347898, |
|
"grad_norm": 2.1027439407928505, |
|
"learning_rate": 2.3145186079753685e-06, |
|
"loss": 0.381, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.7165044522675502, |
|
"grad_norm": 2.0200252919367823, |
|
"learning_rate": 2.253797978362617e-06, |
|
"loss": 0.3754, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.7206460965003106, |
|
"grad_norm": 2.022944794755911, |
|
"learning_rate": 2.193651868568285e-06, |
|
"loss": 0.3719, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.7247877407330711, |
|
"grad_norm": 2.011383665562108, |
|
"learning_rate": 2.1340928614729445e-06, |
|
"loss": 0.3716, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.7289293849658315, |
|
"grad_norm": 2.1000401745759767, |
|
"learning_rate": 2.075133417132223e-06, |
|
"loss": 0.3773, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.7330710291985918, |
|
"grad_norm": 2.0322415551222277, |
|
"learning_rate": 2.016785870170079e-06, |
|
"loss": 0.3755, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.7372126734313522, |
|
"grad_norm": 2.0586335139183327, |
|
"learning_rate": 1.9590624271983406e-06, |
|
"loss": 0.3749, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.7413543176641126, |
|
"grad_norm": 2.057098686991852, |
|
"learning_rate": 1.9019751642630252e-06, |
|
"loss": 0.3733, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.7454959618968731, |
|
"grad_norm": 1.9856940958346814, |
|
"learning_rate": 1.8455360243179537e-06, |
|
"loss": 0.3737, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.7496376061296335, |
|
"grad_norm": 2.049797779671496, |
|
"learning_rate": 1.7897568147262323e-06, |
|
"loss": 0.3678, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.7537792503623939, |
|
"grad_norm": 2.1869162271482083, |
|
"learning_rate": 1.7346492047900897e-06, |
|
"loss": 0.3769, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.7579208945951543, |
|
"grad_norm": 2.0166177063427444, |
|
"learning_rate": 1.6802247233095914e-06, |
|
"loss": 0.3722, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.7620625388279146, |
|
"grad_norm": 2.0284630775550094, |
|
"learning_rate": 1.626494756170765e-06, |
|
"loss": 0.3562, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.7662041830606751, |
|
"grad_norm": 1.9844703515401159, |
|
"learning_rate": 1.5734705439636017e-06, |
|
"loss": 0.3641, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.7703458272934355, |
|
"grad_norm": 2.026880974919187, |
|
"learning_rate": 1.5211631796304721e-06, |
|
"loss": 0.3671, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.7744874715261959, |
|
"grad_norm": 2.2261033018640775, |
|
"learning_rate": 1.46958360614543e-06, |
|
"loss": 0.3677, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.7786291157589563, |
|
"grad_norm": 2.0419572811826527, |
|
"learning_rate": 1.4187426142248723e-06, |
|
"loss": 0.3567, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.7827707599917167, |
|
"grad_norm": 1.9436995231419443, |
|
"learning_rate": 1.3686508400700787e-06, |
|
"loss": 0.3659, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.7869124042244772, |
|
"grad_norm": 2.118970998919544, |
|
"learning_rate": 1.3193187631420462e-06, |
|
"loss": 0.3621, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.7910540484572375, |
|
"grad_norm": 2.059747456229496, |
|
"learning_rate": 1.2707567039691505e-06, |
|
"loss": 0.3565, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.7951956926899979, |
|
"grad_norm": 2.090689117796637, |
|
"learning_rate": 1.222974821988024e-06, |
|
"loss": 0.3583, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.7993373369227583, |
|
"grad_norm": 2.059076495278081, |
|
"learning_rate": 1.1759831134181504e-06, |
|
"loss": 0.3622, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.8034789811555187, |
|
"grad_norm": 1.9884532847703864, |
|
"learning_rate": 1.1297914091706086e-06, |
|
"loss": 0.3541, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.8076206253882792, |
|
"grad_norm": 1.8905068597211632, |
|
"learning_rate": 1.0844093727913868e-06, |
|
"loss": 0.3578, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.8117622696210396, |
|
"grad_norm": 2.0536970678158206, |
|
"learning_rate": 1.039846498439727e-06, |
|
"loss": 0.353, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.8159039138538, |
|
"grad_norm": 1.9394493842379006, |
|
"learning_rate": 9.961121089018933e-07, |
|
"loss": 0.3552, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.8200455580865603, |
|
"grad_norm": 1.9161234213176144, |
|
"learning_rate": 9.532153536407923e-07, |
|
"loss": 0.3572, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.8241872023193207, |
|
"grad_norm": 2.0679533219870394, |
|
"learning_rate": 9.111652068818621e-07, |
|
"loss": 0.3499, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.8283288465520812, |
|
"grad_norm": 2.007021263318756, |
|
"learning_rate": 8.699704657356195e-07, |
|
"loss": 0.3503, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.8324704907848416, |
|
"grad_norm": 1.940869076922602, |
|
"learning_rate": 8.296397483572515e-07, |
|
"loss": 0.3588, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.836612135017602, |
|
"grad_norm": 2.0578809556631774, |
|
"learning_rate": 7.901814921436624e-07, |
|
"loss": 0.3497, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.8407537792503624, |
|
"grad_norm": 1.9963820836617243, |
|
"learning_rate": 7.516039519683105e-07, |
|
"loss": 0.3459, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.8448954234831229, |
|
"grad_norm": 2.1312384802093707, |
|
"learning_rate": 7.139151984542636e-07, |
|
"loss": 0.3515, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.8490370677158832, |
|
"grad_norm": 2.051969582922354, |
|
"learning_rate": 6.771231162857722e-07, |
|
"loss": 0.3497, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.8531787119486436, |
|
"grad_norm": 1.944237335752014, |
|
"learning_rate": 6.412354025587509e-07, |
|
"loss": 0.3454, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.857320356181404, |
|
"grad_norm": 2.0013927970500824, |
|
"learning_rate": 6.062595651705111e-07, |
|
"loss": 0.3484, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.8614620004141644, |
|
"grad_norm": 1.9953241481292785, |
|
"learning_rate": 5.722029212490666e-07, |
|
"loss": 0.3467, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.8656036446469249, |
|
"grad_norm": 1.8562753423514788, |
|
"learning_rate": 5.390725956223531e-07, |
|
"loss": 0.3439, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.8697452888796853, |
|
"grad_norm": 2.014412721333329, |
|
"learning_rate": 5.068755193276798e-07, |
|
"loss": 0.3475, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.8738869331124457, |
|
"grad_norm": 2.0883606449971013, |
|
"learning_rate": 4.756184281617121e-07, |
|
"loss": 0.3442, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.878028577345206, |
|
"grad_norm": 1.9479502623498623, |
|
"learning_rate": 4.4530786127131575e-07, |
|
"loss": 0.3516, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.8821702215779664, |
|
"grad_norm": 1.9878832197736611, |
|
"learning_rate": 4.159501597855287e-07, |
|
"loss": 0.3468, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.8863118658107269, |
|
"grad_norm": 1.981044415868457, |
|
"learning_rate": 3.8755146548896784e-07, |
|
"loss": 0.3442, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.8904535100434873, |
|
"grad_norm": 2.078622119500157, |
|
"learning_rate": 3.6011771953693044e-07, |
|
"loss": 0.3414, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.8945951542762477, |
|
"grad_norm": 1.9832065578238605, |
|
"learning_rate": 3.336546612124758e-07, |
|
"loss": 0.3462, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.8987367985090081, |
|
"grad_norm": 2.077188737775822, |
|
"learning_rate": 3.081678267257404e-07, |
|
"loss": 0.3445, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.9028784427417684, |
|
"grad_norm": 1.9680080545892447, |
|
"learning_rate": 2.836625480557265e-07, |
|
"loss": 0.3433, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.9070200869745288, |
|
"grad_norm": 2.0554539318334344, |
|
"learning_rate": 2.601439518348331e-07, |
|
"loss": 0.3411, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.9111617312072893, |
|
"grad_norm": 1.9738460247856697, |
|
"learning_rate": 2.376169582763288e-07, |
|
"loss": 0.3423, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.9153033754400497, |
|
"grad_norm": 2.100671455290073, |
|
"learning_rate": 2.1608628014502364e-07, |
|
"loss": 0.3412, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.9194450196728101, |
|
"grad_norm": 2.0790276531527865, |
|
"learning_rate": 1.955564217713335e-07, |
|
"loss": 0.3376, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.9235866639055705, |
|
"grad_norm": 2.039950191776352, |
|
"learning_rate": 1.7603167810894662e-07, |
|
"loss": 0.3406, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.927728308138331, |
|
"grad_norm": 2.0162487373459435, |
|
"learning_rate": 1.5751613383630128e-07, |
|
"loss": 0.3465, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.9318699523710913, |
|
"grad_norm": 2.092197022372298, |
|
"learning_rate": 1.4001366250204762e-07, |
|
"loss": 0.337, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.9360115966038517, |
|
"grad_norm": 1.944362051868436, |
|
"learning_rate": 1.235279257146804e-07, |
|
"loss": 0.3378, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.9401532408366121, |
|
"grad_norm": 2.06431956869675, |
|
"learning_rate": 1.080623723765134e-07, |
|
"loss": 0.3352, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.9442948850693725, |
|
"grad_norm": 2.067267199918096, |
|
"learning_rate": 9.362023796215036e-08, |
|
"loss": 0.3385, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.948436529302133, |
|
"grad_norm": 2.050374938357445, |
|
"learning_rate": 8.020454384160437e-08, |
|
"loss": 0.345, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 0.9525781735348934, |
|
"grad_norm": 2.058842049841172, |
|
"learning_rate": 6.78180966482156e-08, |
|
"loss": 0.3431, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.9567198177676538, |
|
"grad_norm": 2.1228820426864825, |
|
"learning_rate": 5.646348769148491e-08, |
|
"loss": 0.3415, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.9608614620004141, |
|
"grad_norm": 1.9574417210663868, |
|
"learning_rate": 4.6143092414961396e-08, |
|
"loss": 0.3346, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.9650031062331745, |
|
"grad_norm": 2.0568028289295293, |
|
"learning_rate": 3.685906989928656e-08, |
|
"loss": 0.3404, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 0.969144750465935, |
|
"grad_norm": 2.187483815397574, |
|
"learning_rate": 2.861336241050061e-08, |
|
"loss": 0.3366, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.9732863946986954, |
|
"grad_norm": 2.0602914938549626, |
|
"learning_rate": 2.1407694993714755e-08, |
|
"loss": 0.3419, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.9774280389314558, |
|
"grad_norm": 2.0097264769694174, |
|
"learning_rate": 1.5243575112218744e-08, |
|
"loss": 0.3391, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.9815696831642162, |
|
"grad_norm": 2.1300950765468456, |
|
"learning_rate": 1.0122292332114814e-08, |
|
"loss": 0.3479, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 0.9857113273969766, |
|
"grad_norm": 1.9814636182776308, |
|
"learning_rate": 6.044918052531268e-09, |
|
"loss": 0.3359, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.989852971629737, |
|
"grad_norm": 2.0905112637682053, |
|
"learning_rate": 3.0123052814812203e-09, |
|
"loss": 0.3383, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 0.9939946158624974, |
|
"grad_norm": 1.966590567136286, |
|
"learning_rate": 1.025088457409229e-09, |
|
"loss": 0.3325, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.9981362600952578, |
|
"grad_norm": 2.035603528671296, |
|
"learning_rate": 8.368331646302353e-11, |
|
"loss": 0.3416, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 0.999792917788362, |
|
"eval_loss": 0.3488326072692871, |
|
"eval_runtime": 0.9501, |
|
"eval_samples_per_second": 3.158, |
|
"eval_steps_per_second": 1.053, |
|
"step": 1207 |
|
}, |
|
{ |
|
"epoch": 0.999792917788362, |
|
"step": 1207, |
|
"total_flos": 252668899491840.0, |
|
"train_loss": 0.5272446889569172, |
|
"train_runtime": 29415.576, |
|
"train_samples_per_second": 1.313, |
|
"train_steps_per_second": 0.041 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 1207, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 252668899491840.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|