terry69's picture
Model save
cb3d171 verified
raw
history blame contribute delete
No virus
43.2 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.999793431109275,
"eval_steps": 500,
"global_step": 1210,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0008262755629002272,
"grad_norm": 23.554044233313164,
"learning_rate": 8.264462809917357e-08,
"loss": 1.4094,
"step": 1
},
{
"epoch": 0.004131377814501136,
"grad_norm": 23.487881173685214,
"learning_rate": 4.132231404958678e-07,
"loss": 1.44,
"step": 5
},
{
"epoch": 0.008262755629002272,
"grad_norm": 8.324804089201631,
"learning_rate": 8.264462809917356e-07,
"loss": 1.3352,
"step": 10
},
{
"epoch": 0.01239413344350341,
"grad_norm": 10.482872108095675,
"learning_rate": 1.2396694214876035e-06,
"loss": 1.1699,
"step": 15
},
{
"epoch": 0.016525511258004544,
"grad_norm": 3.0017247032336667,
"learning_rate": 1.6528925619834712e-06,
"loss": 1.0235,
"step": 20
},
{
"epoch": 0.02065688907250568,
"grad_norm": 2.760976256425166,
"learning_rate": 2.066115702479339e-06,
"loss": 0.9675,
"step": 25
},
{
"epoch": 0.02478826688700682,
"grad_norm": 2.376505582612684,
"learning_rate": 2.479338842975207e-06,
"loss": 0.9406,
"step": 30
},
{
"epoch": 0.028919644701507954,
"grad_norm": 2.2154227436349765,
"learning_rate": 2.8925619834710743e-06,
"loss": 0.9302,
"step": 35
},
{
"epoch": 0.03305102251600909,
"grad_norm": 2.3739150674432445,
"learning_rate": 3.3057851239669424e-06,
"loss": 0.8994,
"step": 40
},
{
"epoch": 0.03718240033051023,
"grad_norm": 2.2109809527940554,
"learning_rate": 3.71900826446281e-06,
"loss": 0.8982,
"step": 45
},
{
"epoch": 0.04131377814501136,
"grad_norm": 2.28822301545997,
"learning_rate": 4.132231404958678e-06,
"loss": 0.8837,
"step": 50
},
{
"epoch": 0.0454451559595125,
"grad_norm": 2.2559023546514343,
"learning_rate": 4.5454545454545455e-06,
"loss": 0.8754,
"step": 55
},
{
"epoch": 0.04957653377401364,
"grad_norm": 2.2137771037984684,
"learning_rate": 4.958677685950414e-06,
"loss": 0.8582,
"step": 60
},
{
"epoch": 0.05370791158851477,
"grad_norm": 2.373934998656087,
"learning_rate": 5.371900826446281e-06,
"loss": 0.8593,
"step": 65
},
{
"epoch": 0.05783928940301591,
"grad_norm": 2.467630003052232,
"learning_rate": 5.785123966942149e-06,
"loss": 0.856,
"step": 70
},
{
"epoch": 0.06197066721751704,
"grad_norm": 2.424099124838778,
"learning_rate": 6.198347107438017e-06,
"loss": 0.853,
"step": 75
},
{
"epoch": 0.06610204503201818,
"grad_norm": 2.362990499630114,
"learning_rate": 6.611570247933885e-06,
"loss": 0.8383,
"step": 80
},
{
"epoch": 0.07023342284651932,
"grad_norm": 2.399473277862941,
"learning_rate": 7.0247933884297525e-06,
"loss": 0.8318,
"step": 85
},
{
"epoch": 0.07436480066102046,
"grad_norm": 2.5700793481674755,
"learning_rate": 7.43801652892562e-06,
"loss": 0.8303,
"step": 90
},
{
"epoch": 0.07849617847552158,
"grad_norm": 2.3911560791916466,
"learning_rate": 7.851239669421489e-06,
"loss": 0.8283,
"step": 95
},
{
"epoch": 0.08262755629002272,
"grad_norm": 2.5970628827400057,
"learning_rate": 8.264462809917356e-06,
"loss": 0.8249,
"step": 100
},
{
"epoch": 0.08675893410452386,
"grad_norm": 2.3382293960099143,
"learning_rate": 8.677685950413224e-06,
"loss": 0.8186,
"step": 105
},
{
"epoch": 0.090890311919025,
"grad_norm": 2.423365953954326,
"learning_rate": 9.090909090909091e-06,
"loss": 0.8202,
"step": 110
},
{
"epoch": 0.09502168973352614,
"grad_norm": 2.5282725418752694,
"learning_rate": 9.50413223140496e-06,
"loss": 0.8122,
"step": 115
},
{
"epoch": 0.09915306754802727,
"grad_norm": 2.3639158309512522,
"learning_rate": 9.917355371900828e-06,
"loss": 0.8127,
"step": 120
},
{
"epoch": 0.1032844453625284,
"grad_norm": 2.844448120421469,
"learning_rate": 9.999667111219573e-06,
"loss": 0.7962,
"step": 125
},
{
"epoch": 0.10741582317702954,
"grad_norm": 3.443216907128371,
"learning_rate": 9.998314826517564e-06,
"loss": 0.8031,
"step": 130
},
{
"epoch": 0.11154720099153068,
"grad_norm": 2.681450835682187,
"learning_rate": 9.995922621477252e-06,
"loss": 0.805,
"step": 135
},
{
"epoch": 0.11567857880603181,
"grad_norm": 2.41859239043235,
"learning_rate": 9.99249099380692e-06,
"loss": 0.7891,
"step": 140
},
{
"epoch": 0.11980995662053295,
"grad_norm": 2.2320583844472375,
"learning_rate": 9.988020657471078e-06,
"loss": 0.7945,
"step": 145
},
{
"epoch": 0.12394133443503408,
"grad_norm": 2.6404700133557064,
"learning_rate": 9.98251254254193e-06,
"loss": 0.7789,
"step": 150
},
{
"epoch": 0.12807271224953523,
"grad_norm": 2.4700012252546855,
"learning_rate": 9.97596779500586e-06,
"loss": 0.7776,
"step": 155
},
{
"epoch": 0.13220409006403636,
"grad_norm": 2.534567680863635,
"learning_rate": 9.968387776525009e-06,
"loss": 0.7665,
"step": 160
},
{
"epoch": 0.13633546787853748,
"grad_norm": 2.3195785880081274,
"learning_rate": 9.959774064153977e-06,
"loss": 0.7633,
"step": 165
},
{
"epoch": 0.14046684569303863,
"grad_norm": 2.6303462252799124,
"learning_rate": 9.950128450011706e-06,
"loss": 0.7587,
"step": 170
},
{
"epoch": 0.14459822350753976,
"grad_norm": 2.4452240563950687,
"learning_rate": 9.939452940908627e-06,
"loss": 0.7627,
"step": 175
},
{
"epoch": 0.1487296013220409,
"grad_norm": 2.3656117475844294,
"learning_rate": 9.927749757929125e-06,
"loss": 0.7546,
"step": 180
},
{
"epoch": 0.15286097913654204,
"grad_norm": 2.307155404533025,
"learning_rate": 9.915021335969452e-06,
"loss": 0.744,
"step": 185
},
{
"epoch": 0.15699235695104316,
"grad_norm": 2.3398594948394127,
"learning_rate": 9.901270323231114e-06,
"loss": 0.7423,
"step": 190
},
{
"epoch": 0.1611237347655443,
"grad_norm": 2.3156292065612214,
"learning_rate": 9.886499580669917e-06,
"loss": 0.7396,
"step": 195
},
{
"epoch": 0.16525511258004544,
"grad_norm": 2.574743453191487,
"learning_rate": 9.870712181400726e-06,
"loss": 0.7335,
"step": 200
},
{
"epoch": 0.1693864903945466,
"grad_norm": 2.339577898310028,
"learning_rate": 9.853911410058097e-06,
"loss": 0.7398,
"step": 205
},
{
"epoch": 0.17351786820904772,
"grad_norm": 2.489900169912843,
"learning_rate": 9.836100762112887e-06,
"loss": 0.7423,
"step": 210
},
{
"epoch": 0.17764924602354884,
"grad_norm": 2.5114661447059365,
"learning_rate": 9.817283943145014e-06,
"loss": 0.7273,
"step": 215
},
{
"epoch": 0.18178062383805,
"grad_norm": 2.4765147278897097,
"learning_rate": 9.797464868072489e-06,
"loss": 0.7171,
"step": 220
},
{
"epoch": 0.18591200165255112,
"grad_norm": 2.35952144225102,
"learning_rate": 9.776647660336905e-06,
"loss": 0.7165,
"step": 225
},
{
"epoch": 0.19004337946705227,
"grad_norm": 2.3928640553968115,
"learning_rate": 9.754836651045538e-06,
"loss": 0.7228,
"step": 230
},
{
"epoch": 0.1941747572815534,
"grad_norm": 2.359655168950183,
"learning_rate": 9.732036378070243e-06,
"loss": 0.6999,
"step": 235
},
{
"epoch": 0.19830613509605455,
"grad_norm": 2.7166804592398854,
"learning_rate": 9.708251585103324e-06,
"loss": 0.6924,
"step": 240
},
{
"epoch": 0.20243751291055567,
"grad_norm": 2.6360301220005415,
"learning_rate": 9.683487220670595e-06,
"loss": 0.7038,
"step": 245
},
{
"epoch": 0.2065688907250568,
"grad_norm": 2.358814454829314,
"learning_rate": 9.657748437101819e-06,
"loss": 0.7076,
"step": 250
},
{
"epoch": 0.21070026853955795,
"grad_norm": 2.250240362766412,
"learning_rate": 9.631040589458742e-06,
"loss": 0.6833,
"step": 255
},
{
"epoch": 0.21483164635405907,
"grad_norm": 2.4709639248779083,
"learning_rate": 9.603369234420944e-06,
"loss": 0.6751,
"step": 260
},
{
"epoch": 0.21896302416856023,
"grad_norm": 2.315474134768072,
"learning_rate": 9.574740129129767e-06,
"loss": 0.6988,
"step": 265
},
{
"epoch": 0.22309440198306135,
"grad_norm": 2.2090103807004975,
"learning_rate": 9.545159229990493e-06,
"loss": 0.6854,
"step": 270
},
{
"epoch": 0.22722577979756248,
"grad_norm": 2.366702578463947,
"learning_rate": 9.514632691433108e-06,
"loss": 0.6657,
"step": 275
},
{
"epoch": 0.23135715761206363,
"grad_norm": 2.298899028385364,
"learning_rate": 9.483166864631837e-06,
"loss": 0.663,
"step": 280
},
{
"epoch": 0.23548853542656475,
"grad_norm": 2.2714246758736394,
"learning_rate": 9.450768296183764e-06,
"loss": 0.6655,
"step": 285
},
{
"epoch": 0.2396199132410659,
"grad_norm": 2.1782463852503438,
"learning_rate": 9.417443726746776e-06,
"loss": 0.6678,
"step": 290
},
{
"epoch": 0.24375129105556703,
"grad_norm": 2.206125943355958,
"learning_rate": 9.383200089637143e-06,
"loss": 0.6579,
"step": 295
},
{
"epoch": 0.24788266887006816,
"grad_norm": 2.4822099384962133,
"learning_rate": 9.348044509387022e-06,
"loss": 0.6688,
"step": 300
},
{
"epoch": 0.2520140466845693,
"grad_norm": 2.6347462586981827,
"learning_rate": 9.311984300262151e-06,
"loss": 0.6723,
"step": 305
},
{
"epoch": 0.25614542449907046,
"grad_norm": 2.3969886300206915,
"learning_rate": 9.275026964740101e-06,
"loss": 0.6495,
"step": 310
},
{
"epoch": 0.2602768023135716,
"grad_norm": 2.2255060581251733,
"learning_rate": 9.237180191949347e-06,
"loss": 0.6548,
"step": 315
},
{
"epoch": 0.2644081801280727,
"grad_norm": 2.4470079917886585,
"learning_rate": 9.198451856069514e-06,
"loss": 0.6402,
"step": 320
},
{
"epoch": 0.26853955794257384,
"grad_norm": 2.232598978818489,
"learning_rate": 9.158850014693123e-06,
"loss": 0.6428,
"step": 325
},
{
"epoch": 0.27267093575707496,
"grad_norm": 2.2780042831807807,
"learning_rate": 9.118382907149164e-06,
"loss": 0.6449,
"step": 330
},
{
"epoch": 0.27680231357157614,
"grad_norm": 2.250070488730356,
"learning_rate": 9.077058952788888e-06,
"loss": 0.6248,
"step": 335
},
{
"epoch": 0.28093369138607727,
"grad_norm": 2.257817990210915,
"learning_rate": 9.034886749234112e-06,
"loss": 0.6287,
"step": 340
},
{
"epoch": 0.2850650692005784,
"grad_norm": 2.2753575010897373,
"learning_rate": 8.991875070588449e-06,
"loss": 0.631,
"step": 345
},
{
"epoch": 0.2891964470150795,
"grad_norm": 2.5652909583991534,
"learning_rate": 8.948032865611823e-06,
"loss": 0.6232,
"step": 350
},
{
"epoch": 0.29332782482958064,
"grad_norm": 2.3365773173303843,
"learning_rate": 8.90336925585864e-06,
"loss": 0.608,
"step": 355
},
{
"epoch": 0.2974592026440818,
"grad_norm": 2.27141916228481,
"learning_rate": 8.857893533780015e-06,
"loss": 0.6048,
"step": 360
},
{
"epoch": 0.30159058045858295,
"grad_norm": 2.121899092032811,
"learning_rate": 8.811615160790428e-06,
"loss": 0.6192,
"step": 365
},
{
"epoch": 0.30572195827308407,
"grad_norm": 2.210463830947181,
"learning_rate": 8.764543765299245e-06,
"loss": 0.611,
"step": 370
},
{
"epoch": 0.3098533360875852,
"grad_norm": 2.2581243380183835,
"learning_rate": 8.716689140707488e-06,
"loss": 0.6116,
"step": 375
},
{
"epoch": 0.3139847139020863,
"grad_norm": 2.291459985719934,
"learning_rate": 8.668061243370273e-06,
"loss": 0.6094,
"step": 380
},
{
"epoch": 0.3181160917165875,
"grad_norm": 2.242064952311342,
"learning_rate": 8.61867019052535e-06,
"loss": 0.6076,
"step": 385
},
{
"epoch": 0.3222474695310886,
"grad_norm": 2.3770181909891304,
"learning_rate": 8.568526258188172e-06,
"loss": 0.6047,
"step": 390
},
{
"epoch": 0.32637884734558975,
"grad_norm": 2.4304957747194593,
"learning_rate": 8.517639879013918e-06,
"loss": 0.6037,
"step": 395
},
{
"epoch": 0.3305102251600909,
"grad_norm": 2.159929931604691,
"learning_rate": 8.466021640126946e-06,
"loss": 0.5924,
"step": 400
},
{
"epoch": 0.334641602974592,
"grad_norm": 2.181588993237923,
"learning_rate": 8.413682280918094e-06,
"loss": 0.5859,
"step": 405
},
{
"epoch": 0.3387729807890932,
"grad_norm": 2.239848195768108,
"learning_rate": 8.3606326908103e-06,
"loss": 0.5808,
"step": 410
},
{
"epoch": 0.3429043586035943,
"grad_norm": 2.2395305817254165,
"learning_rate": 8.306883906993022e-06,
"loss": 0.5805,
"step": 415
},
{
"epoch": 0.34703573641809543,
"grad_norm": 2.276771540411031,
"learning_rate": 8.25244711212589e-06,
"loss": 0.5771,
"step": 420
},
{
"epoch": 0.35116711423259656,
"grad_norm": 2.2443733616231674,
"learning_rate": 8.197333632012123e-06,
"loss": 0.5724,
"step": 425
},
{
"epoch": 0.3552984920470977,
"grad_norm": 2.266086543452868,
"learning_rate": 8.141554933242135e-06,
"loss": 0.5766,
"step": 430
},
{
"epoch": 0.35942986986159886,
"grad_norm": 2.3440174185212426,
"learning_rate": 8.08512262080787e-06,
"loss": 0.5704,
"step": 435
},
{
"epoch": 0.3635612476761,
"grad_norm": 2.231383713024198,
"learning_rate": 8.028048435688333e-06,
"loss": 0.5672,
"step": 440
},
{
"epoch": 0.3676926254906011,
"grad_norm": 2.210345864732212,
"learning_rate": 7.970344252406832e-06,
"loss": 0.5654,
"step": 445
},
{
"epoch": 0.37182400330510224,
"grad_norm": 2.201487752141703,
"learning_rate": 7.912022076560426e-06,
"loss": 0.5608,
"step": 450
},
{
"epoch": 0.37595538111960336,
"grad_norm": 2.2126765922615017,
"learning_rate": 7.853094042322121e-06,
"loss": 0.5676,
"step": 455
},
{
"epoch": 0.38008675893410454,
"grad_norm": 2.368648057415258,
"learning_rate": 7.7935724099163e-06,
"loss": 0.5611,
"step": 460
},
{
"epoch": 0.38421813674860567,
"grad_norm": 2.249897659892324,
"learning_rate": 7.733469563067928e-06,
"loss": 0.5632,
"step": 465
},
{
"epoch": 0.3883495145631068,
"grad_norm": 2.257196050360634,
"learning_rate": 7.67279800642607e-06,
"loss": 0.5414,
"step": 470
},
{
"epoch": 0.3924808923776079,
"grad_norm": 2.1223073675520125,
"learning_rate": 7.611570362962247e-06,
"loss": 0.5421,
"step": 475
},
{
"epoch": 0.3966122701921091,
"grad_norm": 2.209118984554094,
"learning_rate": 7.549799371344175e-06,
"loss": 0.5475,
"step": 480
},
{
"epoch": 0.4007436480066102,
"grad_norm": 2.070603624522537,
"learning_rate": 7.487497883285428e-06,
"loss": 0.5477,
"step": 485
},
{
"epoch": 0.40487502582111135,
"grad_norm": 2.2718749779596097,
"learning_rate": 7.424678860871584e-06,
"loss": 0.5384,
"step": 490
},
{
"epoch": 0.40900640363561247,
"grad_norm": 2.347289860613089,
"learning_rate": 7.361355373863415e-06,
"loss": 0.5407,
"step": 495
},
{
"epoch": 0.4131377814501136,
"grad_norm": 2.491186132317311,
"learning_rate": 7.297540596977663e-06,
"loss": 0.5393,
"step": 500
},
{
"epoch": 0.4172691592646148,
"grad_norm": 2.3320063032579177,
"learning_rate": 7.233247807145989e-06,
"loss": 0.5311,
"step": 505
},
{
"epoch": 0.4214005370791159,
"grad_norm": 2.122601227745636,
"learning_rate": 7.168490380752648e-06,
"loss": 0.537,
"step": 510
},
{
"epoch": 0.425531914893617,
"grad_norm": 2.2214499757842865,
"learning_rate": 7.103281790851483e-06,
"loss": 0.5338,
"step": 515
},
{
"epoch": 0.42966329270811815,
"grad_norm": 2.1654002355826494,
"learning_rate": 7.037635604362786e-06,
"loss": 0.5188,
"step": 520
},
{
"epoch": 0.4337946705226193,
"grad_norm": 2.1599888645905847,
"learning_rate": 6.971565479250659e-06,
"loss": 0.5321,
"step": 525
},
{
"epoch": 0.43792604833712045,
"grad_norm": 2.4521064066537,
"learning_rate": 6.905085161681408e-06,
"loss": 0.5137,
"step": 530
},
{
"epoch": 0.4420574261516216,
"grad_norm": 2.0735269808191563,
"learning_rate": 6.838208483163601e-06,
"loss": 0.5181,
"step": 535
},
{
"epoch": 0.4461888039661227,
"grad_norm": 2.1964532327451307,
"learning_rate": 6.770949357670358e-06,
"loss": 0.5079,
"step": 540
},
{
"epoch": 0.45032018178062383,
"grad_norm": 2.1721038727423996,
"learning_rate": 6.703321778744495e-06,
"loss": 0.5169,
"step": 545
},
{
"epoch": 0.45445155959512495,
"grad_norm": 2.083001885275941,
"learning_rate": 6.635339816587109e-06,
"loss": 0.5189,
"step": 550
},
{
"epoch": 0.45858293740962613,
"grad_norm": 2.2406805358253523,
"learning_rate": 6.5670176151302136e-06,
"loss": 0.5056,
"step": 555
},
{
"epoch": 0.46271431522412726,
"grad_norm": 2.362515256614418,
"learning_rate": 6.4983693890940335e-06,
"loss": 0.5068,
"step": 560
},
{
"epoch": 0.4668456930386284,
"grad_norm": 2.073329405987291,
"learning_rate": 6.4294094210295725e-06,
"loss": 0.5146,
"step": 565
},
{
"epoch": 0.4709770708531295,
"grad_norm": 2.0802359451245276,
"learning_rate": 6.360152058347068e-06,
"loss": 0.4993,
"step": 570
},
{
"epoch": 0.47510844866763063,
"grad_norm": 2.205382348988315,
"learning_rate": 6.290611710330957e-06,
"loss": 0.5056,
"step": 575
},
{
"epoch": 0.4792398264821318,
"grad_norm": 2.106317834822583,
"learning_rate": 6.2208028451419575e-06,
"loss": 0.4939,
"step": 580
},
{
"epoch": 0.48337120429663294,
"grad_norm": 2.0997265512356953,
"learning_rate": 6.150739986806911e-06,
"loss": 0.4904,
"step": 585
},
{
"epoch": 0.48750258211113406,
"grad_norm": 2.087397537314153,
"learning_rate": 6.0804377121969985e-06,
"loss": 0.4903,
"step": 590
},
{
"epoch": 0.4916339599256352,
"grad_norm": 2.1772365057315834,
"learning_rate": 6.009910647994956e-06,
"loss": 0.5043,
"step": 595
},
{
"epoch": 0.4957653377401363,
"grad_norm": 2.0357765671221872,
"learning_rate": 5.939173467651942e-06,
"loss": 0.4916,
"step": 600
},
{
"epoch": 0.4998967155546375,
"grad_norm": 2.140953972933528,
"learning_rate": 5.8682408883346535e-06,
"loss": 0.4881,
"step": 605
},
{
"epoch": 0.5040280933691386,
"grad_norm": 2.133259953600996,
"learning_rate": 5.7971276678633625e-06,
"loss": 0.4925,
"step": 610
},
{
"epoch": 0.5081594711836398,
"grad_norm": 2.0897840178036753,
"learning_rate": 5.725848601641492e-06,
"loss": 0.4909,
"step": 615
},
{
"epoch": 0.5122908489981409,
"grad_norm": 2.1033344267926553,
"learning_rate": 5.654418519577369e-06,
"loss": 0.4808,
"step": 620
},
{
"epoch": 0.516422226812642,
"grad_norm": 2.0931483592499927,
"learning_rate": 5.5828522829987965e-06,
"loss": 0.4697,
"step": 625
},
{
"epoch": 0.5205536046271432,
"grad_norm": 2.0194919084584444,
"learning_rate": 5.511164781561096e-06,
"loss": 0.479,
"step": 630
},
{
"epoch": 0.5246849824416443,
"grad_norm": 2.070247481858063,
"learning_rate": 5.439370930149252e-06,
"loss": 0.4705,
"step": 635
},
{
"epoch": 0.5288163602561454,
"grad_norm": 2.060542153916691,
"learning_rate": 5.367485665774802e-06,
"loss": 0.4689,
"step": 640
},
{
"epoch": 0.5329477380706465,
"grad_norm": 2.171769190139748,
"learning_rate": 5.295523944468137e-06,
"loss": 0.4678,
"step": 645
},
{
"epoch": 0.5370791158851477,
"grad_norm": 2.437387629266701,
"learning_rate": 5.223500738166837e-06,
"loss": 0.4723,
"step": 650
},
{
"epoch": 0.5412104936996488,
"grad_norm": 2.139213174417897,
"learning_rate": 5.1514310316006835e-06,
"loss": 0.472,
"step": 655
},
{
"epoch": 0.5453418715141499,
"grad_norm": 2.083008247659044,
"learning_rate": 5.07932981917404e-06,
"loss": 0.4684,
"step": 660
},
{
"epoch": 0.5494732493286512,
"grad_norm": 2.161242770803335,
"learning_rate": 5.007212101846194e-06,
"loss": 0.4587,
"step": 665
},
{
"epoch": 0.5536046271431523,
"grad_norm": 2.0500783532272604,
"learning_rate": 4.935092884010347e-06,
"loss": 0.4611,
"step": 670
},
{
"epoch": 0.5577360049576534,
"grad_norm": 2.1439435869325116,
"learning_rate": 4.8629871703718844e-06,
"loss": 0.4622,
"step": 675
},
{
"epoch": 0.5618673827721545,
"grad_norm": 2.066743306657874,
"learning_rate": 4.7909099628265946e-06,
"loss": 0.4672,
"step": 680
},
{
"epoch": 0.5659987605866557,
"grad_norm": 2.0461947863660574,
"learning_rate": 4.718876257339444e-06,
"loss": 0.4512,
"step": 685
},
{
"epoch": 0.5701301384011568,
"grad_norm": 2.13397891500544,
"learning_rate": 4.646901040824622e-06,
"loss": 0.4568,
"step": 690
},
{
"epoch": 0.5742615162156579,
"grad_norm": 2.038909744378958,
"learning_rate": 4.574999288027437e-06,
"loss": 0.4505,
"step": 695
},
{
"epoch": 0.578392894030159,
"grad_norm": 2.061119461177199,
"learning_rate": 4.503185958408767e-06,
"loss": 0.4536,
"step": 700
},
{
"epoch": 0.5825242718446602,
"grad_norm": 2.147709730965502,
"learning_rate": 4.431475993032673e-06,
"loss": 0.4567,
"step": 705
},
{
"epoch": 0.5866556496591613,
"grad_norm": 2.0471799173895078,
"learning_rate": 4.359884311457857e-06,
"loss": 0.452,
"step": 710
},
{
"epoch": 0.5907870274736625,
"grad_norm": 2.096762090340259,
"learning_rate": 4.2884258086335755e-06,
"loss": 0.4407,
"step": 715
},
{
"epoch": 0.5949184052881636,
"grad_norm": 2.1776523990414436,
"learning_rate": 4.217115351800693e-06,
"loss": 0.4364,
"step": 720
},
{
"epoch": 0.5990497831026648,
"grad_norm": 2.148545496231866,
"learning_rate": 4.145967777398481e-06,
"loss": 0.4407,
"step": 725
},
{
"epoch": 0.6031811609171659,
"grad_norm": 2.0382917606214326,
"learning_rate": 4.074997887977843e-06,
"loss": 0.4339,
"step": 730
},
{
"epoch": 0.607312538731667,
"grad_norm": 2.096587532898138,
"learning_rate": 4.004220449121574e-06,
"loss": 0.4337,
"step": 735
},
{
"epoch": 0.6114439165461681,
"grad_norm": 2.0260734306608272,
"learning_rate": 3.933650186372329e-06,
"loss": 0.4297,
"step": 740
},
{
"epoch": 0.6155752943606693,
"grad_norm": 2.0747974830236116,
"learning_rate": 3.863301782168896e-06,
"loss": 0.4362,
"step": 745
},
{
"epoch": 0.6197066721751704,
"grad_norm": 2.1542278120116154,
"learning_rate": 3.7931898727914723e-06,
"loss": 0.4413,
"step": 750
},
{
"epoch": 0.6238380499896715,
"grad_norm": 1.9775710411271903,
"learning_rate": 3.7233290453165127e-06,
"loss": 0.4273,
"step": 755
},
{
"epoch": 0.6279694278041726,
"grad_norm": 2.034389521374345,
"learning_rate": 3.6537338345818273e-06,
"loss": 0.4312,
"step": 760
},
{
"epoch": 0.6321008056186739,
"grad_norm": 2.059316170572532,
"learning_rate": 3.5844187201625567e-06,
"loss": 0.4301,
"step": 765
},
{
"epoch": 0.636232183433175,
"grad_norm": 2.0806349809281235,
"learning_rate": 3.5153981233586277e-06,
"loss": 0.4234,
"step": 770
},
{
"epoch": 0.6403635612476761,
"grad_norm": 2.1474742343641933,
"learning_rate": 3.446686404194337e-06,
"loss": 0.4228,
"step": 775
},
{
"epoch": 0.6444949390621773,
"grad_norm": 2.141080032663376,
"learning_rate": 3.3782978584307035e-06,
"loss": 0.4179,
"step": 780
},
{
"epoch": 0.6486263168766784,
"grad_norm": 2.0266388142571876,
"learning_rate": 3.310246714591162e-06,
"loss": 0.4197,
"step": 785
},
{
"epoch": 0.6527576946911795,
"grad_norm": 2.1015063906736873,
"learning_rate": 3.2425471310012645e-06,
"loss": 0.4224,
"step": 790
},
{
"epoch": 0.6568890725056806,
"grad_norm": 2.0688211913933126,
"learning_rate": 3.1752131928429787e-06,
"loss": 0.4148,
"step": 795
},
{
"epoch": 0.6610204503201818,
"grad_norm": 2.020342446426012,
"learning_rate": 3.1082589092242116e-06,
"loss": 0.4209,
"step": 800
},
{
"epoch": 0.6651518281346829,
"grad_norm": 2.050265450999902,
"learning_rate": 3.041698210264149e-06,
"loss": 0.4086,
"step": 805
},
{
"epoch": 0.669283205949184,
"grad_norm": 2.0414704146915117,
"learning_rate": 2.9755449441950434e-06,
"loss": 0.424,
"step": 810
},
{
"epoch": 0.6734145837636852,
"grad_norm": 2.0309738852976924,
"learning_rate": 2.9098128744810245e-06,
"loss": 0.4136,
"step": 815
},
{
"epoch": 0.6775459615781864,
"grad_norm": 2.123876449787772,
"learning_rate": 2.84451567695456e-06,
"loss": 0.4213,
"step": 820
},
{
"epoch": 0.6816773393926875,
"grad_norm": 2.1356187469686043,
"learning_rate": 2.7796669369711294e-06,
"loss": 0.4216,
"step": 825
},
{
"epoch": 0.6858087172071886,
"grad_norm": 2.06757675819696,
"learning_rate": 2.715280146582752e-06,
"loss": 0.4192,
"step": 830
},
{
"epoch": 0.6899400950216897,
"grad_norm": 2.0954350411098916,
"learning_rate": 2.651368701730889e-06,
"loss": 0.4077,
"step": 835
},
{
"epoch": 0.6940714728361909,
"grad_norm": 1.9490264079119275,
"learning_rate": 2.5879458994593786e-06,
"loss": 0.4018,
"step": 840
},
{
"epoch": 0.698202850650692,
"grad_norm": 1.9862703131789983,
"learning_rate": 2.5250249351479206e-06,
"loss": 0.4127,
"step": 845
},
{
"epoch": 0.7023342284651931,
"grad_norm": 1.9907644914298972,
"learning_rate": 2.4626188997667224e-06,
"loss": 0.4057,
"step": 850
},
{
"epoch": 0.7064656062796942,
"grad_norm": 2.0178385478283,
"learning_rate": 2.400740777152874e-06,
"loss": 0.405,
"step": 855
},
{
"epoch": 0.7105969840941954,
"grad_norm": 2.0040936929718627,
"learning_rate": 2.3394034413090015e-06,
"loss": 0.4054,
"step": 860
},
{
"epoch": 0.7147283619086966,
"grad_norm": 2.0032870121182516,
"learning_rate": 2.278619653724781e-06,
"loss": 0.4025,
"step": 865
},
{
"epoch": 0.7188597397231977,
"grad_norm": 2.0315969629289867,
"learning_rate": 2.218402060721845e-06,
"loss": 0.3929,
"step": 870
},
{
"epoch": 0.7229911175376988,
"grad_norm": 2.0727282353845626,
"learning_rate": 2.1587631908226812e-06,
"loss": 0.3986,
"step": 875
},
{
"epoch": 0.7271224953522,
"grad_norm": 1.9152876057461066,
"learning_rate": 2.09971545214401e-06,
"loss": 0.3915,
"step": 880
},
{
"epoch": 0.7312538731667011,
"grad_norm": 2.006764866181322,
"learning_rate": 2.04127112981522e-06,
"loss": 0.3972,
"step": 885
},
{
"epoch": 0.7353852509812022,
"grad_norm": 2.009533139724546,
"learning_rate": 1.9834423834224014e-06,
"loss": 0.3952,
"step": 890
},
{
"epoch": 0.7395166287957033,
"grad_norm": 1.9136982154678623,
"learning_rate": 1.926241244478496e-06,
"loss": 0.3894,
"step": 895
},
{
"epoch": 0.7436480066102045,
"grad_norm": 1.9940424884511188,
"learning_rate": 1.8696796139200852e-06,
"loss": 0.3964,
"step": 900
},
{
"epoch": 0.7477793844247056,
"grad_norm": 1.9232507318610614,
"learning_rate": 1.8137692596313528e-06,
"loss": 0.39,
"step": 905
},
{
"epoch": 0.7519107622392067,
"grad_norm": 1.9669358048444339,
"learning_rate": 1.7585218139957205e-06,
"loss": 0.4089,
"step": 910
},
{
"epoch": 0.756042140053708,
"grad_norm": 2.0275135426829007,
"learning_rate": 1.7039487714756953e-06,
"loss": 0.3897,
"step": 915
},
{
"epoch": 0.7601735178682091,
"grad_norm": 1.9145558326260135,
"learning_rate": 1.6500614862213866e-06,
"loss": 0.3845,
"step": 920
},
{
"epoch": 0.7643048956827102,
"grad_norm": 1.946476089351215,
"learning_rate": 1.596871169708235e-06,
"loss": 0.3827,
"step": 925
},
{
"epoch": 0.7684362734972113,
"grad_norm": 2.061234713938133,
"learning_rate": 1.54438888840441e-06,
"loss": 0.3865,
"step": 930
},
{
"epoch": 0.7725676513117125,
"grad_norm": 2.0046087896210993,
"learning_rate": 1.4926255614683931e-06,
"loss": 0.3919,
"step": 935
},
{
"epoch": 0.7766990291262136,
"grad_norm": 2.013100573926538,
"learning_rate": 1.4415919584771999e-06,
"loss": 0.3791,
"step": 940
},
{
"epoch": 0.7808304069407147,
"grad_norm": 2.052665994161596,
"learning_rate": 1.3912986971857168e-06,
"loss": 0.3813,
"step": 945
},
{
"epoch": 0.7849617847552158,
"grad_norm": 1.9665758334089818,
"learning_rate": 1.3417562413176405e-06,
"loss": 0.3743,
"step": 950
},
{
"epoch": 0.789093162569717,
"grad_norm": 1.9653082426188921,
"learning_rate": 1.292974898388456e-06,
"loss": 0.3884,
"step": 955
},
{
"epoch": 0.7932245403842182,
"grad_norm": 2.1523771874487707,
"learning_rate": 1.2449648175609115e-06,
"loss": 0.383,
"step": 960
},
{
"epoch": 0.7973559181987193,
"grad_norm": 1.8532885315782601,
"learning_rate": 1.1977359875334454e-06,
"loss": 0.3812,
"step": 965
},
{
"epoch": 0.8014872960132204,
"grad_norm": 2.0616460511941477,
"learning_rate": 1.1512982344619904e-06,
"loss": 0.3742,
"step": 970
},
{
"epoch": 0.8056186738277216,
"grad_norm": 2.000103259130858,
"learning_rate": 1.1056612199156093e-06,
"loss": 0.3795,
"step": 975
},
{
"epoch": 0.8097500516422227,
"grad_norm": 2.0764477975903404,
"learning_rate": 1.0608344388663576e-06,
"loss": 0.3732,
"step": 980
},
{
"epoch": 0.8138814294567238,
"grad_norm": 2.0844020934084315,
"learning_rate": 1.01682721771382e-06,
"loss": 0.3738,
"step": 985
},
{
"epoch": 0.8180128072712249,
"grad_norm": 2.050710610881507,
"learning_rate": 9.73648712344707e-07,
"loss": 0.373,
"step": 990
},
{
"epoch": 0.8221441850857261,
"grad_norm": 2.068871529971891,
"learning_rate": 9.313079062279429e-07,
"loss": 0.3852,
"step": 995
},
{
"epoch": 0.8262755629002272,
"grad_norm": 1.9622479084933468,
"learning_rate": 8.898136085456127e-07,
"loss": 0.3839,
"step": 1000
},
{
"epoch": 0.8304069407147283,
"grad_norm": 2.0489239754877913,
"learning_rate": 8.491744523601741e-07,
"loss": 0.3789,
"step": 1005
},
{
"epoch": 0.8345383185292296,
"grad_norm": 1.97405744731925,
"learning_rate": 8.093988928183127e-07,
"loss": 0.3659,
"step": 1010
},
{
"epoch": 0.8386696963437307,
"grad_norm": 2.0649700701955966,
"learning_rate": 7.70495205391818e-07,
"loss": 0.3684,
"step": 1015
},
{
"epoch": 0.8428010741582318,
"grad_norm": 1.9711762707705944,
"learning_rate": 7.324714841558322e-07,
"loss": 0.3728,
"step": 1020
},
{
"epoch": 0.8469324519727329,
"grad_norm": 2.089515078127623,
"learning_rate": 6.953356401048466e-07,
"loss": 0.361,
"step": 1025
},
{
"epoch": 0.851063829787234,
"grad_norm": 2.0969140211573536,
"learning_rate": 6.590953995067812e-07,
"loss": 0.3621,
"step": 1030
},
{
"epoch": 0.8551952076017352,
"grad_norm": 2.014897877209678,
"learning_rate": 6.237583022955079e-07,
"loss": 0.377,
"step": 1035
},
{
"epoch": 0.8593265854162363,
"grad_norm": 2.0616530107992617,
"learning_rate": 5.89331700502136e-07,
"loss": 0.373,
"step": 1040
},
{
"epoch": 0.8634579632307374,
"grad_norm": 2.0017552200191164,
"learning_rate": 5.558227567253832e-07,
"loss": 0.3738,
"step": 1045
},
{
"epoch": 0.8675893410452385,
"grad_norm": 2.0884355733341415,
"learning_rate": 5.232384426413772e-07,
"loss": 0.3625,
"step": 1050
},
{
"epoch": 0.8717207188597397,
"grad_norm": 2.111748293695631,
"learning_rate": 4.915855375531642e-07,
"loss": 0.3789,
"step": 1055
},
{
"epoch": 0.8758520966742409,
"grad_norm": 2.0363025293506305,
"learning_rate": 4.608706269802471e-07,
"loss": 0.3615,
"step": 1060
},
{
"epoch": 0.879983474488742,
"grad_norm": 2.0861363712497742,
"learning_rate": 4.3110010128843747e-07,
"loss": 0.3803,
"step": 1065
},
{
"epoch": 0.8841148523032432,
"grad_norm": 1.9794223337927541,
"learning_rate": 4.022801543603194e-07,
"loss": 0.37,
"step": 1070
},
{
"epoch": 0.8882462301177443,
"grad_norm": 1.976468279453303,
"learning_rate": 3.744167823065814e-07,
"loss": 0.3511,
"step": 1075
},
{
"epoch": 0.8923776079322454,
"grad_norm": 2.0814002530931583,
"learning_rate": 3.4751578221850377e-07,
"loss": 0.3611,
"step": 1080
},
{
"epoch": 0.8965089857467465,
"grad_norm": 2.0984819483954844,
"learning_rate": 3.2158275096184744e-07,
"loss": 0.3562,
"step": 1085
},
{
"epoch": 0.9006403635612477,
"grad_norm": 2.144295712901401,
"learning_rate": 2.966230840124007e-07,
"loss": 0.3674,
"step": 1090
},
{
"epoch": 0.9047717413757488,
"grad_norm": 2.118969417696939,
"learning_rate": 2.72641974333433e-07,
"loss": 0.3622,
"step": 1095
},
{
"epoch": 0.9089031191902499,
"grad_norm": 2.0003217874891694,
"learning_rate": 2.4964441129527337e-07,
"loss": 0.3697,
"step": 1100
},
{
"epoch": 0.913034497004751,
"grad_norm": 2.052789066548848,
"learning_rate": 2.2763517963725169e-07,
"loss": 0.3709,
"step": 1105
},
{
"epoch": 0.9171658748192523,
"grad_norm": 1.9424275069967558,
"learning_rate": 2.0661885847221398e-07,
"loss": 0.3672,
"step": 1110
},
{
"epoch": 0.9212972526337534,
"grad_norm": 2.030858945131136,
"learning_rate": 1.8659982033381928e-07,
"loss": 0.3675,
"step": 1115
},
{
"epoch": 0.9254286304482545,
"grad_norm": 2.0600092854113132,
"learning_rate": 1.6758223026681507e-07,
"loss": 0.3636,
"step": 1120
},
{
"epoch": 0.9295600082627556,
"grad_norm": 2.0750384756960374,
"learning_rate": 1.4957004496048256e-07,
"loss": 0.3656,
"step": 1125
},
{
"epoch": 0.9336913860772568,
"grad_norm": 2.006693676746082,
"learning_rate": 1.3256701192542853e-07,
"loss": 0.3676,
"step": 1130
},
{
"epoch": 0.9378227638917579,
"grad_norm": 2.114218265869402,
"learning_rate": 1.1657666871390471e-07,
"loss": 0.3658,
"step": 1135
},
{
"epoch": 0.941954141706259,
"grad_norm": 2.0597089907860537,
"learning_rate": 1.0160234218380305e-07,
"loss": 0.3573,
"step": 1140
},
{
"epoch": 0.9460855195207601,
"grad_norm": 2.0898079361017334,
"learning_rate": 8.76471478064872e-08,
"loss": 0.3724,
"step": 1145
},
{
"epoch": 0.9502168973352613,
"grad_norm": 2.0371025770189632,
"learning_rate": 7.471398901860772e-08,
"loss": 0.3625,
"step": 1150
},
{
"epoch": 0.9543482751497624,
"grad_norm": 1.938944230136909,
"learning_rate": 6.280555661802857e-08,
"loss": 0.3634,
"step": 1155
},
{
"epoch": 0.9584796529642636,
"grad_norm": 2.1345661558712967,
"learning_rate": 5.192432820399718e-08,
"loss": 0.3674,
"step": 1160
},
{
"epoch": 0.9626110307787648,
"grad_norm": 1.9759349002447268,
"learning_rate": 4.207256766166845e-08,
"loss": 0.3621,
"step": 1165
},
{
"epoch": 0.9667424085932659,
"grad_norm": 2.0037876312026213,
"learning_rate": 3.3252324691093185e-08,
"loss": 0.3645,
"step": 1170
},
{
"epoch": 0.970873786407767,
"grad_norm": 2.059268789116945,
"learning_rate": 2.546543438077087e-08,
"loss": 0.3569,
"step": 1175
},
{
"epoch": 0.9750051642222681,
"grad_norm": 2.1019547322975916,
"learning_rate": 1.8713516825851207e-08,
"loss": 0.3604,
"step": 1180
},
{
"epoch": 0.9791365420367693,
"grad_norm": 2.067476136669,
"learning_rate": 1.2997976791065403e-08,
"loss": 0.3611,
"step": 1185
},
{
"epoch": 0.9832679198512704,
"grad_norm": 1.970947575751658,
"learning_rate": 8.32000341846162e-09,
"loss": 0.3626,
"step": 1190
},
{
"epoch": 0.9873992976657715,
"grad_norm": 2.0213770050103816,
"learning_rate": 4.6805699799967744e-09,
"loss": 0.3577,
"step": 1195
},
{
"epoch": 0.9915306754802726,
"grad_norm": 2.0031463180853915,
"learning_rate": 2.0804336750429588e-09,
"loss": 0.3594,
"step": 1200
},
{
"epoch": 0.9956620532947738,
"grad_norm": 2.0391505951802165,
"learning_rate": 5.201354728517905e-10,
"loss": 0.3624,
"step": 1205
},
{
"epoch": 0.999793431109275,
"grad_norm": 2.0674337207917235,
"learning_rate": 0.0,
"loss": 0.3717,
"step": 1210
},
{
"epoch": 0.999793431109275,
"eval_loss": 0.3491726517677307,
"eval_runtime": 125.1809,
"eval_samples_per_second": 3.115,
"eval_steps_per_second": 0.783,
"step": 1210
},
{
"epoch": 0.999793431109275,
"step": 1210,
"total_flos": 253297038458880.0,
"train_loss": 0.5434080246066259,
"train_runtime": 27449.6475,
"train_samples_per_second": 1.411,
"train_steps_per_second": 0.044
}
],
"logging_steps": 5,
"max_steps": 1210,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 253297038458880.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}