Llama-8b-SimPO-Plus1 / trainer_state.json
Teng Xiao
TX
bf961d3
raw
history blame contribute delete
No virus
51.4 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.998691442030882,
"eval_steps": 500,
"global_step": 477,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.010468463752944255,
"grad_norm": 0.1797010649383585,
"learning_rate": 5.208333333333333e-08,
"logits/chosen": -1.448793649673462,
"logits/rejected": -1.450628638267517,
"logps/chosen": -7.982501983642578,
"logps/rejected": -8.156150817871094,
"loss": -0.0009,
"rewards/accuracies": 0.5062500238418579,
"rewards/chosen": -7.982501983642578,
"rewards/margins": 0.17364946007728577,
"rewards/rejected": -8.156150817871094,
"step": 5
},
{
"epoch": 0.02093692750588851,
"grad_norm": 0.07416876680191102,
"learning_rate": 1.0416666666666667e-07,
"logits/chosen": -1.4468624591827393,
"logits/rejected": -1.4378924369812012,
"logps/chosen": -8.046422958374023,
"logps/rejected": -7.960408687591553,
"loss": 0.0001,
"rewards/accuracies": 0.4937500059604645,
"rewards/chosen": -8.046422958374023,
"rewards/margins": -0.08601457625627518,
"rewards/rejected": -7.960408687591553,
"step": 10
},
{
"epoch": 0.031405391258832765,
"grad_norm": 0.1932776998667392,
"learning_rate": 1.5624999999999999e-07,
"logits/chosen": -1.4454925060272217,
"logits/rejected": -1.4418766498565674,
"logps/chosen": -7.853631496429443,
"logps/rejected": -7.868775844573975,
"loss": -0.0011,
"rewards/accuracies": 0.5,
"rewards/chosen": -7.853631496429443,
"rewards/margins": 0.01514405943453312,
"rewards/rejected": -7.868775844573975,
"step": 15
},
{
"epoch": 0.04187385501177702,
"grad_norm": 0.18172311042773612,
"learning_rate": 2.0833333333333333e-07,
"logits/chosen": -1.4386684894561768,
"logits/rejected": -1.4446001052856445,
"logps/chosen": -8.188841819763184,
"logps/rejected": -8.21079158782959,
"loss": 0.0005,
"rewards/accuracies": 0.5,
"rewards/chosen": -8.188841819763184,
"rewards/margins": 0.02194919064640999,
"rewards/rejected": -8.21079158782959,
"step": 20
},
{
"epoch": 0.05234231876472128,
"grad_norm": 0.07565588877379835,
"learning_rate": 2.604166666666667e-07,
"logits/chosen": -1.469273328781128,
"logits/rejected": -1.4595043659210205,
"logps/chosen": -8.117727279663086,
"logps/rejected": -8.016858100891113,
"loss": 0.0008,
"rewards/accuracies": 0.4375,
"rewards/chosen": -8.117727279663086,
"rewards/margins": -0.1008685976266861,
"rewards/rejected": -8.016858100891113,
"step": 25
},
{
"epoch": 0.06281078251766553,
"grad_norm": 0.15180659620060438,
"learning_rate": 3.1249999999999997e-07,
"logits/chosen": -1.4331070184707642,
"logits/rejected": -1.4219419956207275,
"logps/chosen": -7.856657981872559,
"logps/rejected": -7.774781703948975,
"loss": 0.001,
"rewards/accuracies": 0.48124998807907104,
"rewards/chosen": -7.856657981872559,
"rewards/margins": -0.08187668025493622,
"rewards/rejected": -7.774781703948975,
"step": 30
},
{
"epoch": 0.07327924627060979,
"grad_norm": 0.08490698432858583,
"learning_rate": 3.645833333333333e-07,
"logits/chosen": -1.4420219659805298,
"logits/rejected": -1.4201197624206543,
"logps/chosen": -8.129011154174805,
"logps/rejected": -7.994417667388916,
"loss": -0.0007,
"rewards/accuracies": 0.5062500238418579,
"rewards/chosen": -8.129011154174805,
"rewards/margins": -0.13459424674510956,
"rewards/rejected": -7.994417667388916,
"step": 35
},
{
"epoch": 0.08374771002355404,
"grad_norm": 0.18105138747703975,
"learning_rate": 4.1666666666666667e-07,
"logits/chosen": -1.420586109161377,
"logits/rejected": -1.4364769458770752,
"logps/chosen": -8.086874961853027,
"logps/rejected": -8.243904113769531,
"loss": -0.0007,
"rewards/accuracies": 0.46875,
"rewards/chosen": -8.086874961853027,
"rewards/margins": 0.15702828764915466,
"rewards/rejected": -8.243904113769531,
"step": 40
},
{
"epoch": 0.0942161737764983,
"grad_norm": 0.05754725019388417,
"learning_rate": 4.6874999999999996e-07,
"logits/chosen": -1.4414559602737427,
"logits/rejected": -1.4433170557022095,
"logps/chosen": -8.102398872375488,
"logps/rejected": -7.88956356048584,
"loss": -0.0001,
"rewards/accuracies": 0.4375,
"rewards/chosen": -8.102398872375488,
"rewards/margins": -0.21283535659313202,
"rewards/rejected": -7.88956356048584,
"step": 45
},
{
"epoch": 0.10468463752944256,
"grad_norm": 0.09776323510444927,
"learning_rate": 4.999731868769026e-07,
"logits/chosen": -1.4331969022750854,
"logits/rejected": -1.451460838317871,
"logps/chosen": -8.243871688842773,
"logps/rejected": -8.234269142150879,
"loss": 0.0006,
"rewards/accuracies": 0.48124998807907104,
"rewards/chosen": -8.243871688842773,
"rewards/margins": -0.009602278470993042,
"rewards/rejected": -8.234269142150879,
"step": 50
},
{
"epoch": 0.11515310128238682,
"grad_norm": 0.06802225811273538,
"learning_rate": 4.996716052911017e-07,
"logits/chosen": -1.4432759284973145,
"logits/rejected": -1.4394512176513672,
"logps/chosen": -8.15147876739502,
"logps/rejected": -8.12590217590332,
"loss": 0.0,
"rewards/accuracies": 0.42500001192092896,
"rewards/chosen": -8.15147876739502,
"rewards/margins": -0.025576595216989517,
"rewards/rejected": -8.12590217590332,
"step": 55
},
{
"epoch": 0.12562156503533106,
"grad_norm": 0.19637523644104168,
"learning_rate": 4.990353313429303e-07,
"logits/chosen": -1.4439882040023804,
"logits/rejected": -1.4562370777130127,
"logps/chosen": -8.153823852539062,
"logps/rejected": -8.392339706420898,
"loss": -0.0006,
"rewards/accuracies": 0.574999988079071,
"rewards/chosen": -8.153823852539062,
"rewards/margins": 0.23851680755615234,
"rewards/rejected": -8.392339706420898,
"step": 60
},
{
"epoch": 0.1360900287882753,
"grad_norm": 0.15858568302533665,
"learning_rate": 4.980652179769217e-07,
"logits/chosen": -1.3945095539093018,
"logits/rejected": -1.4153038263320923,
"logps/chosen": -7.983679294586182,
"logps/rejected": -8.382383346557617,
"loss": 0.0001,
"rewards/accuracies": 0.574999988079071,
"rewards/chosen": -7.983679294586182,
"rewards/margins": 0.3987043499946594,
"rewards/rejected": -8.382383346557617,
"step": 65
},
{
"epoch": 0.14655849254121958,
"grad_norm": 0.449952417527854,
"learning_rate": 4.967625656594781e-07,
"logits/chosen": -1.3980720043182373,
"logits/rejected": -1.3917102813720703,
"logps/chosen": -8.066298484802246,
"logps/rejected": -8.08208179473877,
"loss": 0.0015,
"rewards/accuracies": 0.5062500238418579,
"rewards/chosen": -8.066298484802246,
"rewards/margins": 0.015783464536070824,
"rewards/rejected": -8.08208179473877,
"step": 70
},
{
"epoch": 0.15702695629416383,
"grad_norm": 0.0694833097923889,
"learning_rate": 4.951291206355559e-07,
"logits/chosen": -1.4114166498184204,
"logits/rejected": -1.4182840585708618,
"logps/chosen": -8.104999542236328,
"logps/rejected": -8.062032699584961,
"loss": -0.0,
"rewards/accuracies": 0.45625001192092896,
"rewards/chosen": -8.104999542236328,
"rewards/margins": -0.042965833097696304,
"rewards/rejected": -8.062032699584961,
"step": 75
},
{
"epoch": 0.16749542004710807,
"grad_norm": 0.13732894835039064,
"learning_rate": 4.93167072587771e-07,
"logits/chosen": -1.416013479232788,
"logits/rejected": -1.4303759336471558,
"logps/chosen": -8.252126693725586,
"logps/rejected": -8.272165298461914,
"loss": -0.0007,
"rewards/accuracies": 0.4437499940395355,
"rewards/chosen": -8.252126693725586,
"rewards/margins": 0.020038936287164688,
"rewards/rejected": -8.272165298461914,
"step": 80
},
{
"epoch": 0.17796388380005235,
"grad_norm": 0.13841613125213173,
"learning_rate": 4.908790517010636e-07,
"logits/chosen": -1.4297425746917725,
"logits/rejected": -1.4293019771575928,
"logps/chosen": -8.420356750488281,
"logps/rejected": -8.493918418884277,
"loss": 0.0,
"rewards/accuracies": 0.48124998807907104,
"rewards/chosen": -8.420356750488281,
"rewards/margins": 0.07356245070695877,
"rewards/rejected": -8.493918418884277,
"step": 85
},
{
"epoch": 0.1884323475529966,
"grad_norm": 0.26065093480528173,
"learning_rate": 4.882681251368548e-07,
"logits/chosen": -1.411726713180542,
"logits/rejected": -1.4245941638946533,
"logps/chosen": -8.424099922180176,
"logps/rejected": -8.504185676574707,
"loss": -0.0001,
"rewards/accuracies": 0.48750001192092896,
"rewards/chosen": -8.424099922180176,
"rewards/margins": 0.08008553832769394,
"rewards/rejected": -8.504185676574707,
"step": 90
},
{
"epoch": 0.19890081130594087,
"grad_norm": 0.13775483175577152,
"learning_rate": 4.853377929214243e-07,
"logits/chosen": -1.4114550352096558,
"logits/rejected": -1.4067610502243042,
"logps/chosen": -8.460356712341309,
"logps/rejected": -8.213418960571289,
"loss": 0.0002,
"rewards/accuracies": 0.3812499940395355,
"rewards/chosen": -8.460356712341309,
"rewards/margins": -0.24693819880485535,
"rewards/rejected": -8.213418960571289,
"step": 95
},
{
"epoch": 0.2093692750588851,
"grad_norm": 0.0669599374264611,
"learning_rate": 4.820919832540181e-07,
"logits/chosen": -1.4511619806289673,
"logits/rejected": -1.4595460891723633,
"logps/chosen": -8.38967514038086,
"logps/rejected": -8.327149391174316,
"loss": 0.0005,
"rewards/accuracies": 0.4625000059604645,
"rewards/chosen": -8.38967514038086,
"rewards/margins": -0.06252529472112656,
"rewards/rejected": -8.327149391174316,
"step": 100
},
{
"epoch": 0.21983773881182936,
"grad_norm": 0.03744091461066133,
"learning_rate": 4.785350472409791e-07,
"logits/chosen": -1.4548447132110596,
"logits/rejected": -1.4479488134384155,
"logps/chosen": -8.238606452941895,
"logps/rejected": -8.342194557189941,
"loss": 0.0003,
"rewards/accuracies": 0.5375000238418579,
"rewards/chosen": -8.238606452941895,
"rewards/margins": 0.10358722507953644,
"rewards/rejected": -8.342194557189941,
"step": 105
},
{
"epoch": 0.23030620256477363,
"grad_norm": 0.2325346467677191,
"learning_rate": 4.7467175306295647e-07,
"logits/chosen": -1.4924818277359009,
"logits/rejected": -1.4839593172073364,
"logps/chosen": -8.65635871887207,
"logps/rejected": -8.75800895690918,
"loss": -0.0001,
"rewards/accuracies": 0.5562499761581421,
"rewards/chosen": -8.65635871887207,
"rewards/margins": 0.10164938122034073,
"rewards/rejected": -8.75800895690918,
"step": 110
},
{
"epoch": 0.24077466631771788,
"grad_norm": 0.07013048443895094,
"learning_rate": 4.70507279583015e-07,
"logits/chosen": -1.4494998455047607,
"logits/rejected": -1.4506022930145264,
"logps/chosen": -8.537219047546387,
"logps/rejected": -8.694546699523926,
"loss": 0.0002,
"rewards/accuracies": 0.53125,
"rewards/chosen": -8.537219047546387,
"rewards/margins": 0.1573261320590973,
"rewards/rejected": -8.694546699523926,
"step": 115
},
{
"epoch": 0.2512431300706621,
"grad_norm": 0.13941104681218558,
"learning_rate": 4.6604720940421207e-07,
"logits/chosen": -1.4503843784332275,
"logits/rejected": -1.4492030143737793,
"logps/chosen": -8.896661758422852,
"logps/rejected": -8.842092514038086,
"loss": 0.0006,
"rewards/accuracies": 0.5,
"rewards/chosen": -8.896661758422852,
"rewards/margins": -0.054569344967603683,
"rewards/rejected": -8.842092514038086,
"step": 120
},
{
"epoch": 0.26171159382360637,
"grad_norm": 0.01588215105344238,
"learning_rate": 4.612975213859487e-07,
"logits/chosen": -1.5376265048980713,
"logits/rejected": -1.5326207876205444,
"logps/chosen": -9.327282905578613,
"logps/rejected": -9.20351505279541,
"loss": 0.0004,
"rewards/accuracies": 0.46875,
"rewards/chosen": -9.327282905578613,
"rewards/margins": -0.12376775592565536,
"rewards/rejected": -9.20351505279541,
"step": 125
},
{
"epoch": 0.2721800575765506,
"grad_norm": 0.060048076521496456,
"learning_rate": 4.5626458262912735e-07,
"logits/chosen": -1.560910701751709,
"logits/rejected": -1.5665909051895142,
"logps/chosen": -9.537469863891602,
"logps/rejected": -9.556791305541992,
"loss": -0.0001,
"rewards/accuracies": 0.5,
"rewards/chosen": -9.537469863891602,
"rewards/margins": 0.019321158528327942,
"rewards/rejected": -9.556791305541992,
"step": 130
},
{
"epoch": 0.2826485213294949,
"grad_norm": 0.010696845716364521,
"learning_rate": 4.5095513994085974e-07,
"logits/chosen": -1.5800793170928955,
"logits/rejected": -1.582044243812561,
"logps/chosen": -9.630374908447266,
"logps/rejected": -9.887577056884766,
"loss": 0.0001,
"rewards/accuracies": 0.543749988079071,
"rewards/chosen": -9.630374908447266,
"rewards/margins": 0.2572011351585388,
"rewards/rejected": -9.887577056884766,
"step": 135
},
{
"epoch": 0.29311698508243916,
"grad_norm": 0.03319847693750231,
"learning_rate": 4.453763107901675e-07,
"logits/chosen": -1.622972846031189,
"logits/rejected": -1.6185715198516846,
"logps/chosen": -9.712789535522461,
"logps/rejected": -9.67750358581543,
"loss": 0.0,
"rewards/accuracies": 0.44999998807907104,
"rewards/chosen": -9.712789535522461,
"rewards/margins": -0.03528591990470886,
"rewards/rejected": -9.67750358581543,
"step": 140
},
{
"epoch": 0.3035854488353834,
"grad_norm": 0.03281163505651298,
"learning_rate": 4.395355737667985e-07,
"logits/chosen": -1.6432979106903076,
"logits/rejected": -1.6470788717269897,
"logps/chosen": -9.864433288574219,
"logps/rejected": -9.815264701843262,
"loss": 0.0001,
"rewards/accuracies": 0.518750011920929,
"rewards/chosen": -9.864433288574219,
"rewards/margins": -0.049169473350048065,
"rewards/rejected": -9.815264701843262,
"step": 145
},
{
"epoch": 0.31405391258832765,
"grad_norm": 0.023094845909487956,
"learning_rate": 4.3344075855595097e-07,
"logits/chosen": -1.6595284938812256,
"logits/rejected": -1.669385552406311,
"logps/chosen": -9.924379348754883,
"logps/rejected": -9.930808067321777,
"loss": -0.0003,
"rewards/accuracies": 0.41874998807907104,
"rewards/chosen": -9.924379348754883,
"rewards/margins": 0.006428956985473633,
"rewards/rejected": -9.930808067321777,
"step": 150
},
{
"epoch": 0.3245223763412719,
"grad_norm": 0.1908026987463115,
"learning_rate": 4.271000354423425e-07,
"logits/chosen": -1.6287696361541748,
"logits/rejected": -1.6056560277938843,
"logps/chosen": -9.800788879394531,
"logps/rejected": -9.611404418945312,
"loss": 0.0001,
"rewards/accuracies": 0.4312500059604645,
"rewards/chosen": -9.800788879394531,
"rewards/margins": -0.1893850713968277,
"rewards/rejected": -9.611404418945312,
"step": 155
},
{
"epoch": 0.33499084009421615,
"grad_norm": 0.03202775302677848,
"learning_rate": 4.2052190435769554e-07,
"logits/chosen": -1.6098477840423584,
"logits/rejected": -1.6122195720672607,
"logps/chosen": -9.803987503051758,
"logps/rejected": -9.737637519836426,
"loss": 0.0,
"rewards/accuracies": 0.48750001192092896,
"rewards/chosen": -9.803987503051758,
"rewards/margins": -0.06634993851184845,
"rewards/rejected": -9.737637519836426,
"step": 160
},
{
"epoch": 0.34545930384716045,
"grad_norm": 0.07232776561687705,
"learning_rate": 4.137151834863213e-07,
"logits/chosen": -1.65714430809021,
"logits/rejected": -1.6471843719482422,
"logps/chosen": -9.869422912597656,
"logps/rejected": -9.760724067687988,
"loss": 0.0003,
"rewards/accuracies": 0.4625000059604645,
"rewards/chosen": -9.869422912597656,
"rewards/margins": -0.10870015621185303,
"rewards/rejected": -9.760724067687988,
"step": 165
},
{
"epoch": 0.3559277676001047,
"grad_norm": 0.024726012832997274,
"learning_rate": 4.0668899744407567e-07,
"logits/chosen": -1.6514545679092407,
"logits/rejected": -1.6379756927490234,
"logps/chosen": -10.327203750610352,
"logps/rejected": -10.077325820922852,
"loss": 0.0,
"rewards/accuracies": 0.4312500059604645,
"rewards/chosen": -10.327203750610352,
"rewards/margins": -0.24987690150737762,
"rewards/rejected": -10.077325820922852,
"step": 170
},
{
"epoch": 0.36639623135304894,
"grad_norm": 0.15695534318195675,
"learning_rate": 3.994527650465352e-07,
"logits/chosen": -1.6916990280151367,
"logits/rejected": -1.6754090785980225,
"logps/chosen": -10.47531795501709,
"logps/rejected": -10.233348846435547,
"loss": -0.0001,
"rewards/accuracies": 0.4000000059604645,
"rewards/chosen": -10.47531795501709,
"rewards/margins": -0.24196887016296387,
"rewards/rejected": -10.233348846435547,
"step": 175
},
{
"epoch": 0.3768646951059932,
"grad_norm": 0.055831382164290344,
"learning_rate": 3.920161866827889e-07,
"logits/chosen": -1.7216554880142212,
"logits/rejected": -1.7197192907333374,
"logps/chosen": -10.651076316833496,
"logps/rejected": -10.578369140625,
"loss": -0.0,
"rewards/accuracies": 0.4625000059604645,
"rewards/chosen": -10.651076316833496,
"rewards/margins": -0.07270796597003937,
"rewards/rejected": -10.578369140625,
"step": 180
},
{
"epoch": 0.38733315885893743,
"grad_norm": 0.005192915052039036,
"learning_rate": 3.8438923131177237e-07,
"logits/chosen": -1.7085365056991577,
"logits/rejected": -1.695673942565918,
"logps/chosen": -10.687432289123535,
"logps/rejected": -10.502934455871582,
"loss": 0.0,
"rewards/accuracies": 0.45625001192092896,
"rewards/chosen": -10.687432289123535,
"rewards/margins": -0.18449781835079193,
"rewards/rejected": -10.502934455871582,
"step": 185
},
{
"epoch": 0.39780162261188173,
"grad_norm": 0.00386703700543382,
"learning_rate": 3.765821230985757e-07,
"logits/chosen": -1.7206989526748657,
"logits/rejected": -1.7167381048202515,
"logps/chosen": -10.415846824645996,
"logps/rejected": -10.514742851257324,
"loss": -0.0001,
"rewards/accuracies": 0.48124998807907104,
"rewards/chosen": -10.415846824645996,
"rewards/margins": 0.09889563173055649,
"rewards/rejected": -10.514742851257324,
"step": 190
},
{
"epoch": 0.408270086364826,
"grad_norm": 0.018906277546872338,
"learning_rate": 3.6860532770864005e-07,
"logits/chosen": -1.6922369003295898,
"logits/rejected": -1.681626558303833,
"logps/chosen": -10.679250717163086,
"logps/rejected": -10.37684154510498,
"loss": 0.0,
"rewards/accuracies": 0.44999998807907104,
"rewards/chosen": -10.679250717163086,
"rewards/margins": -0.3024083971977234,
"rewards/rejected": -10.37684154510498,
"step": 195
},
{
"epoch": 0.4187385501177702,
"grad_norm": 0.04569629831070709,
"learning_rate": 3.604695382782159e-07,
"logits/chosen": -1.7120405435562134,
"logits/rejected": -1.7134020328521729,
"logps/chosen": -10.352312088012695,
"logps/rejected": -10.341361999511719,
"loss": 0.0,
"rewards/accuracies": 0.44999998807907104,
"rewards/chosen": -10.352312088012695,
"rewards/margins": -0.010949483141303062,
"rewards/rejected": -10.341361999511719,
"step": 200
},
{
"epoch": 0.42920701387071447,
"grad_norm": 0.01589336857340258,
"learning_rate": 3.5218566107988867e-07,
"logits/chosen": -1.6780574321746826,
"logits/rejected": -1.6641277074813843,
"logps/chosen": -10.600610733032227,
"logps/rejected": -10.596931457519531,
"loss": -0.0,
"rewards/accuracies": 0.46875,
"rewards/chosen": -10.600610733032227,
"rewards/margins": -0.003679597284644842,
"rewards/rejected": -10.596931457519531,
"step": 205
},
{
"epoch": 0.4396754776236587,
"grad_norm": 0.017049353993959408,
"learning_rate": 3.4376480090239047e-07,
"logits/chosen": -1.6822540760040283,
"logits/rejected": -1.6784160137176514,
"logps/chosen": -10.497976303100586,
"logps/rejected": -10.553048133850098,
"loss": -0.0,
"rewards/accuracies": 0.44999998807907104,
"rewards/chosen": -10.497976303100586,
"rewards/margins": 0.0550723597407341,
"rewards/rejected": -10.553048133850098,
"step": 210
},
{
"epoch": 0.45014394137660296,
"grad_norm": 0.2539481968464961,
"learning_rate": 3.3521824616429284e-07,
"logits/chosen": -1.6765304803848267,
"logits/rejected": -1.68255615234375,
"logps/chosen": -10.369840621948242,
"logps/rejected": -10.547749519348145,
"loss": -0.0002,
"rewards/accuracies": 0.5249999761581421,
"rewards/chosen": -10.369840621948242,
"rewards/margins": 0.17790831625461578,
"rewards/rejected": -10.547749519348145,
"step": 215
},
{
"epoch": 0.46061240512954726,
"grad_norm": 0.03490868784910135,
"learning_rate": 3.265574537815398e-07,
"logits/chosen": -1.7063719034194946,
"logits/rejected": -1.702691674232483,
"logps/chosen": -10.227754592895508,
"logps/rejected": -10.175545692443848,
"loss": -0.0,
"rewards/accuracies": 0.4749999940395355,
"rewards/chosen": -10.227754592895508,
"rewards/margins": -0.05220963433384895,
"rewards/rejected": -10.175545692443848,
"step": 220
},
{
"epoch": 0.4710808688824915,
"grad_norm": 0.02089161978800978,
"learning_rate": 3.1779403380910425e-07,
"logits/chosen": -1.6891885995864868,
"logits/rejected": -1.6823409795761108,
"logps/chosen": -9.965743064880371,
"logps/rejected": -10.085700035095215,
"loss": -0.0001,
"rewards/accuracies": 0.512499988079071,
"rewards/chosen": -9.965743064880371,
"rewards/margins": 0.11995597183704376,
"rewards/rejected": -10.085700035095215,
"step": 225
},
{
"epoch": 0.48154933263543576,
"grad_norm": 0.042927151818770604,
"learning_rate": 3.0893973387735683e-07,
"logits/chosen": -1.6752849817276,
"logits/rejected": -1.6569786071777344,
"logps/chosen": -10.218798637390137,
"logps/rejected": -10.10997200012207,
"loss": -0.0004,
"rewards/accuracies": 0.44999998807907104,
"rewards/chosen": -10.218798637390137,
"rewards/margins": -0.1088268980383873,
"rewards/rejected": -10.10997200012207,
"step": 230
},
{
"epoch": 0.49201779638838,
"grad_norm": 0.055950224699650504,
"learning_rate": 3.000064234440111e-07,
"logits/chosen": -1.6750266551971436,
"logits/rejected": -1.6845659017562866,
"logps/chosen": -9.966501235961914,
"logps/rejected": -10.05066967010498,
"loss": -0.0,
"rewards/accuracies": 0.48124998807907104,
"rewards/chosen": -9.966501235961914,
"rewards/margins": 0.08416947722434998,
"rewards/rejected": -10.05066967010498,
"step": 235
},
{
"epoch": 0.5024862601413242,
"grad_norm": 0.10458741539165252,
"learning_rate": 2.910060778827554e-07,
"logits/chosen": -1.6746679544448853,
"logits/rejected": -1.6765581369400024,
"logps/chosen": -10.034064292907715,
"logps/rejected": -9.960668563842773,
"loss": -0.0003,
"rewards/accuracies": 0.4437499940395355,
"rewards/chosen": -10.034064292907715,
"rewards/margins": -0.07339614629745483,
"rewards/rejected": -9.960668563842773,
"step": 240
},
{
"epoch": 0.5129547238942685,
"grad_norm": 0.14020310226910793,
"learning_rate": 2.8195076242990116e-07,
"logits/chosen": -1.656341314315796,
"logits/rejected": -1.6607738733291626,
"logps/chosen": -9.828956604003906,
"logps/rejected": -9.777771949768066,
"loss": 0.0,
"rewards/accuracies": 0.5062500238418579,
"rewards/chosen": -9.828956604003906,
"rewards/margins": -0.05118449404835701,
"rewards/rejected": -9.777771949768066,
"step": 245
},
{
"epoch": 0.5234231876472127,
"grad_norm": 0.029712092586844915,
"learning_rate": 2.7285261601056697e-07,
"logits/chosen": -1.651460886001587,
"logits/rejected": -1.6511856317520142,
"logps/chosen": -9.723172187805176,
"logps/rejected": -9.84160327911377,
"loss": 0.0005,
"rewards/accuracies": 0.518750011920929,
"rewards/chosen": -9.723172187805176,
"rewards/margins": 0.11843116581439972,
"rewards/rejected": -9.84160327911377,
"step": 250
},
{
"epoch": 0.533891651400157,
"grad_norm": 0.03206118746185812,
"learning_rate": 2.6372383496608186e-07,
"logits/chosen": -1.6517183780670166,
"logits/rejected": -1.6432650089263916,
"logps/chosen": -9.812753677368164,
"logps/rejected": -9.902433395385742,
"loss": -0.0,
"rewards/accuracies": 0.518750011920929,
"rewards/chosen": -9.812753677368164,
"rewards/margins": 0.08968041837215424,
"rewards/rejected": -9.902433395385742,
"step": 255
},
{
"epoch": 0.5443601151531012,
"grad_norm": 0.020964013644010484,
"learning_rate": 2.5457665670441937e-07,
"logits/chosen": -1.592605710029602,
"logits/rejected": -1.5922209024429321,
"logps/chosen": -9.709136962890625,
"logps/rejected": -9.819464683532715,
"loss": 0.0002,
"rewards/accuracies": 0.4937500059604645,
"rewards/chosen": -9.709136962890625,
"rewards/margins": 0.11032851040363312,
"rewards/rejected": -9.819464683532715,
"step": 260
},
{
"epoch": 0.5548285789060455,
"grad_norm": 0.042268508150251446,
"learning_rate": 2.454233432955807e-07,
"logits/chosen": -1.6564003229141235,
"logits/rejected": -1.6685075759887695,
"logps/chosen": -9.937254905700684,
"logps/rejected": -10.057160377502441,
"loss": -0.0,
"rewards/accuracies": 0.5062500238418579,
"rewards/chosen": -9.937254905700684,
"rewards/margins": 0.11990515887737274,
"rewards/rejected": -10.057160377502441,
"step": 265
},
{
"epoch": 0.5652970426589898,
"grad_norm": 0.13512185096282692,
"learning_rate": 2.3627616503391812e-07,
"logits/chosen": -1.680994987487793,
"logits/rejected": -1.6468158960342407,
"logps/chosen": -10.322751998901367,
"logps/rejected": -10.258121490478516,
"loss": 0.0003,
"rewards/accuracies": 0.4937500059604645,
"rewards/chosen": -10.322751998901367,
"rewards/margins": -0.06462957710027695,
"rewards/rejected": -10.258121490478516,
"step": 270
},
{
"epoch": 0.575765506411934,
"grad_norm": 0.012066073597164235,
"learning_rate": 2.2714738398943308e-07,
"logits/chosen": -1.7274888753890991,
"logits/rejected": -1.7199790477752686,
"logps/chosen": -10.117765426635742,
"logps/rejected": -10.104223251342773,
"loss": 0.0001,
"rewards/accuracies": 0.4625000059604645,
"rewards/chosen": -10.117765426635742,
"rewards/margins": -0.013540792278945446,
"rewards/rejected": -10.104223251342773,
"step": 275
},
{
"epoch": 0.5862339701648783,
"grad_norm": 0.01636207441210717,
"learning_rate": 2.1804923757009882e-07,
"logits/chosen": -1.7347486019134521,
"logits/rejected": -1.7411794662475586,
"logps/chosen": -10.153875350952148,
"logps/rejected": -10.280826568603516,
"loss": -0.0,
"rewards/accuracies": 0.4749999940395355,
"rewards/chosen": -10.153875350952148,
"rewards/margins": 0.12695105373859406,
"rewards/rejected": -10.280826568603516,
"step": 280
},
{
"epoch": 0.5967024339178225,
"grad_norm": 0.007389012661156647,
"learning_rate": 2.089939221172446e-07,
"logits/chosen": -1.6744228601455688,
"logits/rejected": -1.6806707382202148,
"logps/chosen": -10.230897903442383,
"logps/rejected": -10.295170783996582,
"loss": -0.0,
"rewards/accuracies": 0.512499988079071,
"rewards/chosen": -10.230897903442383,
"rewards/margins": 0.06427150964736938,
"rewards/rejected": -10.295170783996582,
"step": 285
},
{
"epoch": 0.6071708976707668,
"grad_norm": 0.11180626971526854,
"learning_rate": 1.9999357655598891e-07,
"logits/chosen": -1.715934157371521,
"logits/rejected": -1.7268234491348267,
"logps/chosen": -10.12391471862793,
"logps/rejected": -10.262609481811523,
"loss": -0.0003,
"rewards/accuracies": 0.518750011920929,
"rewards/chosen": -10.12391471862793,
"rewards/margins": 0.13869477808475494,
"rewards/rejected": -10.262609481811523,
"step": 290
},
{
"epoch": 0.6176393614237111,
"grad_norm": 0.03266198194021972,
"learning_rate": 1.9106026612264315e-07,
"logits/chosen": -1.744179368019104,
"logits/rejected": -1.7408939599990845,
"logps/chosen": -10.210718154907227,
"logps/rejected": -10.17192554473877,
"loss": -0.0,
"rewards/accuracies": 0.5062500238418579,
"rewards/chosen": -10.210718154907227,
"rewards/margins": -0.0387921966612339,
"rewards/rejected": -10.17192554473877,
"step": 295
},
{
"epoch": 0.6281078251766553,
"grad_norm": 0.043605609853355175,
"learning_rate": 1.8220596619089573e-07,
"logits/chosen": -1.69484543800354,
"logits/rejected": -1.692705512046814,
"logps/chosen": -10.109774589538574,
"logps/rejected": -10.205177307128906,
"loss": 0.0001,
"rewards/accuracies": 0.4749999940395355,
"rewards/chosen": -10.109774589538574,
"rewards/margins": 0.09540363401174545,
"rewards/rejected": -10.205177307128906,
"step": 300
},
{
"epoch": 0.6385762889295996,
"grad_norm": 0.013464112956501361,
"learning_rate": 1.7344254621846017e-07,
"logits/chosen": -1.6985801458358765,
"logits/rejected": -1.7095019817352295,
"logps/chosen": -10.274127960205078,
"logps/rejected": -10.364879608154297,
"loss": 0.0,
"rewards/accuracies": 0.48124998807907104,
"rewards/chosen": -10.274127960205078,
"rewards/margins": 0.09075053036212921,
"rewards/rejected": -10.364879608154297,
"step": 305
},
{
"epoch": 0.6490447526825438,
"grad_norm": 0.030133040256006938,
"learning_rate": 1.647817538357072e-07,
"logits/chosen": -1.7417259216308594,
"logits/rejected": -1.740393877029419,
"logps/chosen": -10.28361701965332,
"logps/rejected": -10.132238388061523,
"loss": 0.0,
"rewards/accuracies": 0.46875,
"rewards/chosen": -10.28361701965332,
"rewards/margins": -0.15137743949890137,
"rewards/rejected": -10.132238388061523,
"step": 310
},
{
"epoch": 0.6595132164354881,
"grad_norm": 0.04002742762071502,
"learning_rate": 1.562351990976095e-07,
"logits/chosen": -1.69840407371521,
"logits/rejected": -1.7028906345367432,
"logps/chosen": -10.156418800354004,
"logps/rejected": -10.191594123840332,
"loss": -0.0007,
"rewards/accuracies": 0.48124998807907104,
"rewards/chosen": -10.156418800354004,
"rewards/margins": 0.03517543524503708,
"rewards/rejected": -10.191594123840332,
"step": 315
},
{
"epoch": 0.6699816801884323,
"grad_norm": 0.021724114416033388,
"learning_rate": 1.478143389201113e-07,
"logits/chosen": -1.7531684637069702,
"logits/rejected": -1.7422311305999756,
"logps/chosen": -10.416919708251953,
"logps/rejected": -10.20901107788086,
"loss": 0.0001,
"rewards/accuracies": 0.48124998807907104,
"rewards/chosen": -10.416919708251953,
"rewards/margins": -0.20790883898735046,
"rewards/rejected": -10.20901107788086,
"step": 320
},
{
"epoch": 0.6804501439413766,
"grad_norm": 0.08104613852953838,
"learning_rate": 1.3953046172178413e-07,
"logits/chosen": -1.725590467453003,
"logits/rejected": -1.7158119678497314,
"logps/chosen": -10.29697322845459,
"logps/rejected": -10.40684986114502,
"loss": -0.0003,
"rewards/accuracies": 0.4625000059604645,
"rewards/chosen": -10.29697322845459,
"rewards/margins": 0.10987655818462372,
"rewards/rejected": -10.40684986114502,
"step": 325
},
{
"epoch": 0.6909186076943209,
"grad_norm": 0.04339077417181125,
"learning_rate": 1.3139467229135998e-07,
"logits/chosen": -1.7010434865951538,
"logits/rejected": -1.6965417861938477,
"logps/chosen": -10.263284683227539,
"logps/rejected": -10.193758010864258,
"loss": -0.0,
"rewards/accuracies": 0.48124998807907104,
"rewards/chosen": -10.263284683227539,
"rewards/margins": -0.0695262923836708,
"rewards/rejected": -10.193758010864258,
"step": 330
},
{
"epoch": 0.7013870714472651,
"grad_norm": 0.019942381418589637,
"learning_rate": 1.2341787690142435e-07,
"logits/chosen": -1.7109758853912354,
"logits/rejected": -1.7116413116455078,
"logps/chosen": -10.025394439697266,
"logps/rejected": -10.113506317138672,
"loss": 0.0,
"rewards/accuracies": 0.5,
"rewards/chosen": -10.025394439697266,
"rewards/margins": 0.08811323344707489,
"rewards/rejected": -10.113506317138672,
"step": 335
},
{
"epoch": 0.7118555352002094,
"grad_norm": 0.030458935372565502,
"learning_rate": 1.1561076868822755e-07,
"logits/chosen": -1.7183024883270264,
"logits/rejected": -1.7164380550384521,
"logps/chosen": -10.216609954833984,
"logps/rejected": -10.225852966308594,
"loss": 0.0,
"rewards/accuracies": 0.4749999940395355,
"rewards/chosen": -10.216609954833984,
"rewards/margins": 0.00924335140734911,
"rewards/rejected": -10.225852966308594,
"step": 340
},
{
"epoch": 0.7223239989531536,
"grad_norm": 0.04400215150996701,
"learning_rate": 1.0798381331721107e-07,
"logits/chosen": -1.718132734298706,
"logits/rejected": -1.732581377029419,
"logps/chosen": -10.151108741760254,
"logps/rejected": -10.323533058166504,
"loss": -0.0,
"rewards/accuracies": 0.4937500059604645,
"rewards/chosen": -10.151108741760254,
"rewards/margins": 0.17242416739463806,
"rewards/rejected": -10.323533058166504,
"step": 345
},
{
"epoch": 0.7327924627060979,
"grad_norm": 0.03515217556644255,
"learning_rate": 1.0054723495346482e-07,
"logits/chosen": -1.6874383687973022,
"logits/rejected": -1.6842788457870483,
"logps/chosen": -10.151719093322754,
"logps/rejected": -10.054476737976074,
"loss": -0.0001,
"rewards/accuracies": 0.4312500059604645,
"rewards/chosen": -10.151719093322754,
"rewards/margins": -0.09724216163158417,
"rewards/rejected": -10.054476737976074,
"step": 350
},
{
"epoch": 0.7432609264590422,
"grad_norm": 0.021861424316864955,
"learning_rate": 9.331100255592436e-08,
"logits/chosen": -1.7121349573135376,
"logits/rejected": -1.6940410137176514,
"logps/chosen": -10.209026336669922,
"logps/rejected": -9.959612846374512,
"loss": 0.0002,
"rewards/accuracies": 0.4124999940395355,
"rewards/chosen": -10.209026336669922,
"rewards/margins": -0.24941349029541016,
"rewards/rejected": -9.959612846374512,
"step": 355
},
{
"epoch": 0.7537293902119864,
"grad_norm": 0.013506797744103302,
"learning_rate": 8.628481651367875e-08,
"logits/chosen": -1.6897398233413696,
"logits/rejected": -1.6821842193603516,
"logps/chosen": -9.989233016967773,
"logps/rejected": -10.048975944519043,
"loss": 0.0,
"rewards/accuracies": 0.518750011920929,
"rewards/chosen": -9.989233016967773,
"rewards/margins": 0.05974303558468819,
"rewards/rejected": -10.048975944519043,
"step": 360
},
{
"epoch": 0.7641978539649307,
"grad_norm": 0.07239679867383533,
"learning_rate": 7.947809564230445e-08,
"logits/chosen": -1.6575393676757812,
"logits/rejected": -1.6490436792373657,
"logps/chosen": -10.096014022827148,
"logps/rejected": -10.024383544921875,
"loss": 0.0,
"rewards/accuracies": 0.4937500059604645,
"rewards/chosen": -10.096014022827148,
"rewards/margins": -0.07163101434707642,
"rewards/rejected": -10.024383544921875,
"step": 365
},
{
"epoch": 0.7746663177178749,
"grad_norm": 0.08043566497276283,
"learning_rate": 7.289996455765748e-08,
"logits/chosen": -1.730115294456482,
"logits/rejected": -1.7217315435409546,
"logps/chosen": -10.174753189086914,
"logps/rejected": -10.153951644897461,
"loss": 0.0,
"rewards/accuracies": 0.5249999761581421,
"rewards/chosen": -10.174753189086914,
"rewards/margins": -0.02080065943300724,
"rewards/rejected": -10.153951644897461,
"step": 370
},
{
"epoch": 0.7851347814708192,
"grad_norm": 0.018294513390755808,
"learning_rate": 6.655924144404906e-08,
"logits/chosen": -1.7276290655136108,
"logits/rejected": -1.728005051612854,
"logps/chosen": -10.183911323547363,
"logps/rejected": -10.19709587097168,
"loss": 0.0,
"rewards/accuracies": 0.5,
"rewards/chosen": -10.183911323547363,
"rewards/margins": 0.01318590622395277,
"rewards/rejected": -10.19709587097168,
"step": 375
},
{
"epoch": 0.7956032452237635,
"grad_norm": 0.006815837156718109,
"learning_rate": 6.046442623320145e-08,
"logits/chosen": -1.7046406269073486,
"logits/rejected": -1.6621993780136108,
"logps/chosen": -10.309165954589844,
"logps/rejected": -10.365669250488281,
"loss": 0.0001,
"rewards/accuracies": 0.5375000238418579,
"rewards/chosen": -10.309165954589844,
"rewards/margins": 0.056504249572753906,
"rewards/rejected": -10.365669250488281,
"step": 380
},
{
"epoch": 0.8060717089767077,
"grad_norm": 0.030388876590904525,
"learning_rate": 5.4623689209832484e-08,
"logits/chosen": -1.7268683910369873,
"logits/rejected": -1.7090175151824951,
"logps/chosen": -10.306918144226074,
"logps/rejected": -10.260132789611816,
"loss": -0.0,
"rewards/accuracies": 0.4375,
"rewards/chosen": -10.306918144226074,
"rewards/margins": -0.04678579792380333,
"rewards/rejected": -10.260132789611816,
"step": 385
},
{
"epoch": 0.816540172729652,
"grad_norm": 0.04364016204439969,
"learning_rate": 4.904486005914027e-08,
"logits/chosen": -1.7389265298843384,
"logits/rejected": -1.7257276773452759,
"logps/chosen": -10.114287376403809,
"logps/rejected": -10.103799819946289,
"loss": 0.0001,
"rewards/accuracies": 0.44999998807907104,
"rewards/chosen": -10.114287376403809,
"rewards/margins": -0.010487285442650318,
"rewards/rejected": -10.103799819946289,
"step": 390
},
{
"epoch": 0.8270086364825961,
"grad_norm": 0.026690398874973438,
"learning_rate": 4.373541737087263e-08,
"logits/chosen": -1.689626693725586,
"logits/rejected": -1.6864984035491943,
"logps/chosen": -9.968095779418945,
"logps/rejected": -9.953658103942871,
"loss": -0.0005,
"rewards/accuracies": 0.5062500238418579,
"rewards/chosen": -9.968095779418945,
"rewards/margins": -0.014436977915465832,
"rewards/rejected": -9.953658103942871,
"step": 395
},
{
"epoch": 0.8374771002355405,
"grad_norm": 0.023648830133579133,
"learning_rate": 3.8702478614051345e-08,
"logits/chosen": -1.6882154941558838,
"logits/rejected": -1.672609567642212,
"logps/chosen": -10.05548095703125,
"logps/rejected": -9.970457077026367,
"loss": -0.0002,
"rewards/accuracies": 0.48124998807907104,
"rewards/chosen": -10.05548095703125,
"rewards/margins": -0.08502475917339325,
"rewards/rejected": -9.970457077026367,
"step": 400
},
{
"epoch": 0.8479455639884846,
"grad_norm": 0.11775947165792712,
"learning_rate": 3.3952790595787986e-08,
"logits/chosen": -1.7214962244033813,
"logits/rejected": -1.7271496057510376,
"logps/chosen": -10.209165573120117,
"logps/rejected": -10.225208282470703,
"loss": -0.0002,
"rewards/accuracies": 0.4749999940395355,
"rewards/chosen": -10.209165573120117,
"rewards/margins": 0.01604318618774414,
"rewards/rejected": -10.225208282470703,
"step": 405
},
{
"epoch": 0.8584140277414289,
"grad_norm": 0.016736815353629707,
"learning_rate": 2.9492720416985e-08,
"logits/chosen": -1.731414556503296,
"logits/rejected": -1.7266390323638916,
"logps/chosen": -10.035455703735352,
"logps/rejected": -9.934586524963379,
"loss": 0.0002,
"rewards/accuracies": 0.5062500238418579,
"rewards/chosen": -10.035455703735352,
"rewards/margins": -0.10086920112371445,
"rewards/rejected": -9.934586524963379,
"step": 410
},
{
"epoch": 0.8688824914943732,
"grad_norm": 0.010543453920296524,
"learning_rate": 2.5328246937043525e-08,
"logits/chosen": -1.7460575103759766,
"logits/rejected": -1.7472984790802002,
"logps/chosen": -10.30466079711914,
"logps/rejected": -10.271262168884277,
"loss": 0.0,
"rewards/accuracies": 0.44999998807907104,
"rewards/chosen": -10.30466079711914,
"rewards/margins": -0.033397965133190155,
"rewards/rejected": -10.271262168884277,
"step": 415
},
{
"epoch": 0.8793509552473174,
"grad_norm": 0.1619070822929026,
"learning_rate": 2.1464952759020856e-08,
"logits/chosen": -1.6861051321029663,
"logits/rejected": -1.6756645441055298,
"logps/chosen": -10.12368392944336,
"logps/rejected": -10.014533996582031,
"loss": 0.0,
"rewards/accuracies": 0.44999998807907104,
"rewards/chosen": -10.12368392944336,
"rewards/margins": -0.10914961993694305,
"rewards/rejected": -10.014533996582031,
"step": 420
},
{
"epoch": 0.8898194190002617,
"grad_norm": 0.009278538808287691,
"learning_rate": 1.7908016745981856e-08,
"logits/chosen": -1.7117290496826172,
"logits/rejected": -1.7069095373153687,
"logps/chosen": -10.257353782653809,
"logps/rejected": -10.189541816711426,
"loss": 0.0,
"rewards/accuracies": 0.4625000059604645,
"rewards/chosen": -10.257353782653809,
"rewards/margins": -0.06781347841024399,
"rewards/rejected": -10.189541816711426,
"step": 425
},
{
"epoch": 0.9002878827532059,
"grad_norm": 0.02721516902148881,
"learning_rate": 1.4662207078575684e-08,
"logits/chosen": -1.717145562171936,
"logits/rejected": -1.7124665975570679,
"logps/chosen": -10.076754570007324,
"logps/rejected": -10.053953170776367,
"loss": -0.0002,
"rewards/accuracies": 0.45625001192092896,
"rewards/chosen": -10.076754570007324,
"rewards/margins": -0.022800538688898087,
"rewards/rejected": -10.053953170776367,
"step": 430
},
{
"epoch": 0.9107563465061502,
"grad_norm": 0.17249350484919276,
"learning_rate": 1.1731874863145142e-08,
"logits/chosen": -1.74248468875885,
"logits/rejected": -1.7478322982788086,
"logps/chosen": -10.22671890258789,
"logps/rejected": -10.209540367126465,
"loss": 0.0007,
"rewards/accuracies": 0.4937500059604645,
"rewards/chosen": -10.22671890258789,
"rewards/margins": -0.017179716378450394,
"rewards/rejected": -10.209540367126465,
"step": 435
},
{
"epoch": 0.9212248102590945,
"grad_norm": 0.059902683583967914,
"learning_rate": 9.12094829893642e-09,
"logits/chosen": -1.6926848888397217,
"logits/rejected": -1.6890709400177002,
"logps/chosen": -10.345104217529297,
"logps/rejected": -10.262327194213867,
"loss": 0.0004,
"rewards/accuracies": 0.44999998807907104,
"rewards/chosen": -10.345104217529297,
"rewards/margins": -0.08277805149555206,
"rewards/rejected": -10.262327194213867,
"step": 440
},
{
"epoch": 0.9316932740120387,
"grad_norm": 0.10421328261403418,
"learning_rate": 6.832927412229017e-09,
"logits/chosen": -1.7495588064193726,
"logits/rejected": -1.7407792806625366,
"logps/chosen": -10.188596725463867,
"logps/rejected": -10.2090425491333,
"loss": 0.0001,
"rewards/accuracies": 0.4749999940395355,
"rewards/chosen": -10.188596725463867,
"rewards/margins": 0.02044558897614479,
"rewards/rejected": -10.2090425491333,
"step": 445
},
{
"epoch": 0.942161737764983,
"grad_norm": 0.033309080831972024,
"learning_rate": 4.8708793644441086e-09,
"logits/chosen": -1.6661462783813477,
"logits/rejected": -1.6469885110855103,
"logps/chosen": -10.289316177368164,
"logps/rejected": -10.164963722229004,
"loss": 0.0002,
"rewards/accuracies": 0.44999998807907104,
"rewards/chosen": -10.289316177368164,
"rewards/margins": -0.1243540421128273,
"rewards/rejected": -10.164963722229004,
"step": 450
},
{
"epoch": 0.9526302015179272,
"grad_norm": 0.1949041816950042,
"learning_rate": 3.2374343405217884e-09,
"logits/chosen": -1.7108314037322998,
"logits/rejected": -1.710608720779419,
"logps/chosen": -10.109529495239258,
"logps/rejected": -10.343156814575195,
"loss": 0.0002,
"rewards/accuracies": 0.574999988079071,
"rewards/chosen": -10.109529495239258,
"rewards/margins": 0.2336268424987793,
"rewards/rejected": -10.343156814575195,
"step": 455
},
{
"epoch": 0.9630986652708715,
"grad_norm": 0.01652601811230786,
"learning_rate": 1.9347820230782295e-09,
"logits/chosen": -1.7258888483047485,
"logits/rejected": -1.728637933731079,
"logps/chosen": -10.354926109313965,
"logps/rejected": -10.41981315612793,
"loss": -0.0,
"rewards/accuracies": 0.48124998807907104,
"rewards/chosen": -10.354926109313965,
"rewards/margins": 0.06488712131977081,
"rewards/rejected": -10.41981315612793,
"step": 460
},
{
"epoch": 0.9735671290238157,
"grad_norm": 0.12136785382119708,
"learning_rate": 9.64668657069706e-10,
"logits/chosen": -1.7371879816055298,
"logits/rejected": -1.7365795373916626,
"logps/chosen": -10.489349365234375,
"logps/rejected": -10.43460750579834,
"loss": -0.0003,
"rewards/accuracies": 0.4375,
"rewards/chosen": -10.489349365234375,
"rewards/margins": -0.0547415092587471,
"rewards/rejected": -10.43460750579834,
"step": 465
},
{
"epoch": 0.98403559277676,
"grad_norm": 0.0469433162766017,
"learning_rate": 3.2839470889836627e-10,
"logits/chosen": -1.7311279773712158,
"logits/rejected": -1.7309757471084595,
"logps/chosen": -10.06026554107666,
"logps/rejected": -10.253448486328125,
"loss": 0.0001,
"rewards/accuracies": 0.550000011920929,
"rewards/chosen": -10.06026554107666,
"rewards/margins": 0.19318366050720215,
"rewards/rejected": -10.253448486328125,
"step": 470
},
{
"epoch": 0.9945040565297043,
"grad_norm": 0.039418088806606334,
"learning_rate": 2.6813123097352287e-11,
"logits/chosen": -1.7103230953216553,
"logits/rejected": -1.6960973739624023,
"logps/chosen": -10.431560516357422,
"logps/rejected": -10.349306106567383,
"loss": -0.0,
"rewards/accuracies": 0.512499988079071,
"rewards/chosen": -10.431560516357422,
"rewards/margins": -0.08225315064191818,
"rewards/rejected": -10.349306106567383,
"step": 475
},
{
"epoch": 0.998691442030882,
"step": 477,
"total_flos": 0.0,
"train_loss": 1.550099917438415e-05,
"train_runtime": 8183.0083,
"train_samples_per_second": 7.471,
"train_steps_per_second": 0.058
}
],
"logging_steps": 5,
"max_steps": 477,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 1000000,
"total_flos": 0.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}