|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.9493670886075947, |
|
"eval_steps": 1, |
|
"global_step": 156, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02531645569620253, |
|
"grad_norm": 57.114834738699926, |
|
"learning_rate": 3.125e-08, |
|
"logits/chosen": 3.3286685943603516, |
|
"logits/rejected": 3.4021615982055664, |
|
"logps/chosen": -32.21625900268555, |
|
"logps/rejected": -38.764957427978516, |
|
"loss": 0.6931, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.0, |
|
"rewards/margins": 0.0, |
|
"rewards/rejected": 0.0, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.05063291139240506, |
|
"grad_norm": 55.56377943215762, |
|
"learning_rate": 6.25e-08, |
|
"logits/chosen": 2.647796630859375, |
|
"logits/rejected": 2.733036994934082, |
|
"logps/chosen": -24.994104385375977, |
|
"logps/rejected": -29.78559112548828, |
|
"loss": 0.6931, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.0, |
|
"rewards/margins": 0.0, |
|
"rewards/rejected": 0.0, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0759493670886076, |
|
"grad_norm": 61.65658878594894, |
|
"learning_rate": 9.375e-08, |
|
"logits/chosen": 2.33656644821167, |
|
"logits/rejected": 2.518265962600708, |
|
"logps/chosen": -28.91141128540039, |
|
"logps/rejected": -35.79951477050781, |
|
"loss": 0.7163, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": -0.046910468488931656, |
|
"rewards/margins": -0.0570392943918705, |
|
"rewards/rejected": 0.01012883335351944, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.10126582278481013, |
|
"grad_norm": 54.332498272912595, |
|
"learning_rate": 1.25e-07, |
|
"logits/chosen": 2.9658162593841553, |
|
"logits/rejected": 3.0012216567993164, |
|
"logps/chosen": -29.552040100097656, |
|
"logps/rejected": -34.8863639831543, |
|
"loss": 0.7065, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": -0.047019001096487045, |
|
"rewards/margins": -0.010130930691957474, |
|
"rewards/rejected": -0.03688807412981987, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.12658227848101267, |
|
"grad_norm": 55.132631513210136, |
|
"learning_rate": 1.5624999999999999e-07, |
|
"logits/chosen": 3.340583086013794, |
|
"logits/rejected": 3.325289487838745, |
|
"logps/chosen": -29.905162811279297, |
|
"logps/rejected": -38.20769500732422, |
|
"loss": 0.7069, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": -0.06609320640563965, |
|
"rewards/margins": -0.07786056399345398, |
|
"rewards/rejected": 0.01176736131310463, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.1518987341772152, |
|
"grad_norm": 57.59086219970169, |
|
"learning_rate": 1.875e-07, |
|
"logits/chosen": 2.2882869243621826, |
|
"logits/rejected": 2.7739901542663574, |
|
"logps/chosen": -23.262107849121094, |
|
"logps/rejected": -38.5238151550293, |
|
"loss": 0.7044, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": -0.004146900027990341, |
|
"rewards/margins": -0.010677304118871689, |
|
"rewards/rejected": 0.006530404556542635, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.17721518987341772, |
|
"grad_norm": 51.18146497947048, |
|
"learning_rate": 2.1875e-07, |
|
"logits/chosen": 2.7440547943115234, |
|
"logits/rejected": 2.5763754844665527, |
|
"logps/chosen": -29.953731536865234, |
|
"logps/rejected": -33.62974166870117, |
|
"loss": 0.7102, |
|
"rewards/accuracies": 0.375, |
|
"rewards/chosen": -0.038678932934999466, |
|
"rewards/margins": -0.042138125747442245, |
|
"rewards/rejected": 0.0034591909497976303, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.20253164556962025, |
|
"grad_norm": 53.68918501282459, |
|
"learning_rate": 2.5e-07, |
|
"logits/chosen": 3.2939491271972656, |
|
"logits/rejected": 3.003464698791504, |
|
"logps/chosen": -28.833343505859375, |
|
"logps/rejected": -23.717449188232422, |
|
"loss": 0.6967, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.013088438659906387, |
|
"rewards/margins": 0.03873240202665329, |
|
"rewards/rejected": -0.025643955916166306, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.22784810126582278, |
|
"grad_norm": 56.10725202078908, |
|
"learning_rate": 2.8125e-07, |
|
"logits/chosen": 2.243440866470337, |
|
"logits/rejected": 2.4725394248962402, |
|
"logps/chosen": -22.95738983154297, |
|
"logps/rejected": -27.134145736694336, |
|
"loss": 0.684, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.04334143549203873, |
|
"rewards/margins": 0.07354923337697983, |
|
"rewards/rejected": -0.03020780347287655, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.25316455696202533, |
|
"grad_norm": 52.307052055168896, |
|
"learning_rate": 3.1249999999999997e-07, |
|
"logits/chosen": 3.2282490730285645, |
|
"logits/rejected": 3.332897424697876, |
|
"logps/chosen": -26.25811004638672, |
|
"logps/rejected": -33.90528869628906, |
|
"loss": 0.69, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.05821816250681877, |
|
"rewards/margins": 0.11382907629013062, |
|
"rewards/rejected": -0.05561092123389244, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.27848101265822783, |
|
"grad_norm": 51.178561062412925, |
|
"learning_rate": 3.4375e-07, |
|
"logits/chosen": 3.0024123191833496, |
|
"logits/rejected": 2.8223249912261963, |
|
"logps/chosen": -25.597423553466797, |
|
"logps/rejected": -27.973861694335938, |
|
"loss": 0.6937, |
|
"rewards/accuracies": 0.375, |
|
"rewards/chosen": -0.006550817750394344, |
|
"rewards/margins": 0.020815890282392502, |
|
"rewards/rejected": -0.02736670896410942, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.3037974683544304, |
|
"grad_norm": 55.45441942378304, |
|
"learning_rate": 3.75e-07, |
|
"logits/chosen": 2.8794424533843994, |
|
"logits/rejected": 2.8852243423461914, |
|
"logps/chosen": -24.754030227661133, |
|
"logps/rejected": -37.79888916015625, |
|
"loss": 0.6747, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.028544973582029343, |
|
"rewards/margins": 0.1005675345659256, |
|
"rewards/rejected": -0.12911249697208405, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.3291139240506329, |
|
"grad_norm": 53.09392363916735, |
|
"learning_rate": 4.0625e-07, |
|
"logits/chosen": 2.4571850299835205, |
|
"logits/rejected": 2.484537363052368, |
|
"logps/chosen": -25.80780601501465, |
|
"logps/rejected": -29.53820037841797, |
|
"loss": 0.6911, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.03155774995684624, |
|
"rewards/margins": 0.07387223094701767, |
|
"rewards/rejected": -0.04231448099017143, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.35443037974683544, |
|
"grad_norm": 52.83620354330049, |
|
"learning_rate": 4.375e-07, |
|
"logits/chosen": 2.509875535964966, |
|
"logits/rejected": 2.7940757274627686, |
|
"logps/chosen": -23.630701065063477, |
|
"logps/rejected": -30.94167709350586, |
|
"loss": 0.6538, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.0901968702673912, |
|
"rewards/margins": 0.12045808881521225, |
|
"rewards/rejected": -0.030261218547821045, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.379746835443038, |
|
"grad_norm": 54.43440442749398, |
|
"learning_rate": 4.6874999999999996e-07, |
|
"logits/chosen": 2.544405460357666, |
|
"logits/rejected": 2.4906702041625977, |
|
"logps/chosen": -31.145732879638672, |
|
"logps/rejected": -35.83900451660156, |
|
"loss": 0.6699, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.015418028458952904, |
|
"rewards/margins": 0.1095210611820221, |
|
"rewards/rejected": -0.09410304576158524, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.4050632911392405, |
|
"grad_norm": 56.228300664828, |
|
"learning_rate": 5e-07, |
|
"logits/chosen": 2.400423049926758, |
|
"logits/rejected": 2.620924949645996, |
|
"logps/chosen": -19.128610610961914, |
|
"logps/rejected": -33.12566375732422, |
|
"loss": 0.6456, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.02211858704686165, |
|
"rewards/margins": 0.07731227576732635, |
|
"rewards/rejected": -0.0994308590888977, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.43037974683544306, |
|
"grad_norm": 55.89193538913802, |
|
"learning_rate": 4.999370587356267e-07, |
|
"logits/chosen": 2.4614853858947754, |
|
"logits/rejected": 2.2287514209747314, |
|
"logps/chosen": -30.090726852416992, |
|
"logps/rejected": -36.996307373046875, |
|
"loss": 0.6714, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": 0.010743052698671818, |
|
"rewards/margins": 0.19052858650684357, |
|
"rewards/rejected": -0.17978551983833313, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.45569620253164556, |
|
"grad_norm": 47.67611457430014, |
|
"learning_rate": 4.997482666353286e-07, |
|
"logits/chosen": 3.0267255306243896, |
|
"logits/rejected": 3.2108960151672363, |
|
"logps/chosen": -24.398426055908203, |
|
"logps/rejected": -36.53913879394531, |
|
"loss": 0.6385, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": 0.06351403892040253, |
|
"rewards/margins": 0.35269594192504883, |
|
"rewards/rejected": -0.2891818583011627, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.4810126582278481, |
|
"grad_norm": 52.87018041987955, |
|
"learning_rate": 4.99433718761614e-07, |
|
"logits/chosen": 1.927664041519165, |
|
"logits/rejected": 2.412555694580078, |
|
"logps/chosen": -23.279611587524414, |
|
"logps/rejected": -35.23127746582031, |
|
"loss": 0.6107, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.028891097754240036, |
|
"rewards/margins": 0.2483980357646942, |
|
"rewards/rejected": -0.21950694918632507, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.5063291139240507, |
|
"grad_norm": 48.602746524439674, |
|
"learning_rate": 4.989935734988097e-07, |
|
"logits/chosen": 2.6588048934936523, |
|
"logits/rejected": 2.663198232650757, |
|
"logps/chosen": -27.881851196289062, |
|
"logps/rejected": -40.07163619995117, |
|
"loss": 0.6026, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.008863937109708786, |
|
"rewards/margins": 0.33766672015190125, |
|
"rewards/rejected": -0.32880276441574097, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.5316455696202531, |
|
"grad_norm": 54.16428877770723, |
|
"learning_rate": 4.984280524733107e-07, |
|
"logits/chosen": 2.961024284362793, |
|
"logits/rejected": 2.982067584991455, |
|
"logps/chosen": -33.65000915527344, |
|
"logps/rejected": -44.283145904541016, |
|
"loss": 0.6255, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.0565052255988121, |
|
"rewards/margins": 0.3312607407569885, |
|
"rewards/rejected": -0.387766033411026, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.5569620253164557, |
|
"grad_norm": 45.46343620685162, |
|
"learning_rate": 4.977374404419837e-07, |
|
"logits/chosen": 2.731034278869629, |
|
"logits/rejected": 2.8560142517089844, |
|
"logps/chosen": -28.184062957763672, |
|
"logps/rejected": -29.667505264282227, |
|
"loss": 0.5798, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.03637414053082466, |
|
"rewards/margins": 0.30975764989852905, |
|
"rewards/rejected": -0.2733834683895111, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.5822784810126582, |
|
"grad_norm": 44.12365829762132, |
|
"learning_rate": 4.969220851487844e-07, |
|
"logits/chosen": 2.936208963394165, |
|
"logits/rejected": 2.651503086090088, |
|
"logps/chosen": -33.56488037109375, |
|
"logps/rejected": -41.19189453125, |
|
"loss": 0.5772, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.06098739802837372, |
|
"rewards/margins": 0.4615795612335205, |
|
"rewards/rejected": -0.4005921483039856, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.6075949367088608, |
|
"grad_norm": 45.017464152313536, |
|
"learning_rate": 4.959823971496574e-07, |
|
"logits/chosen": 2.6883764266967773, |
|
"logits/rejected": 2.8039445877075195, |
|
"logps/chosen": -25.476293563842773, |
|
"logps/rejected": -27.924108505249023, |
|
"loss": 0.5669, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.003826950676739216, |
|
"rewards/margins": 0.374502956867218, |
|
"rewards/rejected": -0.37067604064941406, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.6329113924050633, |
|
"grad_norm": 42.27141207917208, |
|
"learning_rate": 4.949188496058089e-07, |
|
"logits/chosen": 2.4697136878967285, |
|
"logits/rejected": 2.4843673706054688, |
|
"logps/chosen": -29.330278396606445, |
|
"logps/rejected": -34.346717834472656, |
|
"loss": 0.5643, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.015232307836413383, |
|
"rewards/margins": 0.2725878059864044, |
|
"rewards/rejected": -0.28782013058662415, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.6582278481012658, |
|
"grad_norm": 43.50860864037927, |
|
"learning_rate": 4.937319780454559e-07, |
|
"logits/chosen": 2.4729130268096924, |
|
"logits/rejected": 2.728221893310547, |
|
"logps/chosen": -15.063579559326172, |
|
"logps/rejected": -29.431615829467773, |
|
"loss": 0.5485, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.010220827534794807, |
|
"rewards/margins": 0.6472443342208862, |
|
"rewards/rejected": -0.6370234489440918, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.6835443037974683, |
|
"grad_norm": 44.10326676433438, |
|
"learning_rate": 4.924223800941717e-07, |
|
"logits/chosen": 2.506294012069702, |
|
"logits/rejected": 2.611067295074463, |
|
"logps/chosen": -27.957015991210938, |
|
"logps/rejected": -34.36131286621094, |
|
"loss": 0.5338, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": 0.09378986805677414, |
|
"rewards/margins": 0.7992368936538696, |
|
"rewards/rejected": -0.7054470777511597, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.7088607594936709, |
|
"grad_norm": 42.53652604653026, |
|
"learning_rate": 4.909907151739633e-07, |
|
"logits/chosen": 2.370758295059204, |
|
"logits/rejected": 2.3593924045562744, |
|
"logps/chosen": -24.23151969909668, |
|
"logps/rejected": -28.833358764648438, |
|
"loss": 0.5266, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.015818338841199875, |
|
"rewards/margins": 0.35423219203948975, |
|
"rewards/rejected": -0.3384138345718384, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.7341772151898734, |
|
"grad_norm": 43.92462963258578, |
|
"learning_rate": 4.894377041712326e-07, |
|
"logits/chosen": 2.85575532913208, |
|
"logits/rejected": 2.9239468574523926, |
|
"logps/chosen": -24.336238861083984, |
|
"logps/rejected": -32.19173049926758, |
|
"loss": 0.5788, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.1497105062007904, |
|
"rewards/margins": 0.41176801919937134, |
|
"rewards/rejected": -0.5614784955978394, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.759493670886076, |
|
"grad_norm": 45.52462540230979, |
|
"learning_rate": 4.877641290737883e-07, |
|
"logits/chosen": 2.577241897583008, |
|
"logits/rejected": 2.6502695083618164, |
|
"logps/chosen": -27.839168548583984, |
|
"logps/rejected": -36.543941497802734, |
|
"loss": 0.5261, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.1546507328748703, |
|
"rewards/margins": 1.282554268836975, |
|
"rewards/rejected": -1.4372049570083618, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.7848101265822784, |
|
"grad_norm": 40.84962297403327, |
|
"learning_rate": 4.859708325770919e-07, |
|
"logits/chosen": 2.821880340576172, |
|
"logits/rejected": 2.7289583683013916, |
|
"logps/chosen": -32.00937271118164, |
|
"logps/rejected": -32.9297981262207, |
|
"loss": 0.5013, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.05743463337421417, |
|
"rewards/margins": 0.7840094566345215, |
|
"rewards/rejected": -0.8414440155029297, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.810126582278481, |
|
"grad_norm": 42.08408504733762, |
|
"learning_rate": 4.840587176599343e-07, |
|
"logits/chosen": 2.684711456298828, |
|
"logits/rejected": 3.0184218883514404, |
|
"logps/chosen": -29.447864532470703, |
|
"logps/rejected": -45.021995544433594, |
|
"loss": 0.5126, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.03239244967699051, |
|
"rewards/margins": 1.4539411067962646, |
|
"rewards/rejected": -1.4215487241744995, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.8354430379746836, |
|
"grad_norm": 39.28461553560508, |
|
"learning_rate": 4.820287471297597e-07, |
|
"logits/chosen": 2.2539427280426025, |
|
"logits/rejected": 2.2139289379119873, |
|
"logps/chosen": -28.69852066040039, |
|
"logps/rejected": -31.549945831298828, |
|
"loss": 0.5182, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -0.13107189536094666, |
|
"rewards/margins": 1.036413311958313, |
|
"rewards/rejected": -1.167485237121582, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.8607594936708861, |
|
"grad_norm": 40.986833656434555, |
|
"learning_rate": 4.798819431378626e-07, |
|
"logits/chosen": 2.653698444366455, |
|
"logits/rejected": 2.910414218902588, |
|
"logps/chosen": -21.828556060791016, |
|
"logps/rejected": -38.98440170288086, |
|
"loss": 0.4764, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.08699348568916321, |
|
"rewards/margins": 0.9170618057250977, |
|
"rewards/rejected": -0.8300682306289673, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.8860759493670886, |
|
"grad_norm": 44.21046799920892, |
|
"learning_rate": 4.776193866647039e-07, |
|
"logits/chosen": 2.2878611087799072, |
|
"logits/rejected": 2.6529476642608643, |
|
"logps/chosen": -32.880462646484375, |
|
"logps/rejected": -48.58502197265625, |
|
"loss": 0.5087, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": -0.1624097228050232, |
|
"rewards/margins": 0.6843761205673218, |
|
"rewards/rejected": -0.8467859029769897, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.9113924050632911, |
|
"grad_norm": 37.027834114234714, |
|
"learning_rate": 4.752422169756047e-07, |
|
"logits/chosen": 1.9958323240280151, |
|
"logits/rejected": 2.1072514057159424, |
|
"logps/chosen": -25.151729583740234, |
|
"logps/rejected": -34.56462478637695, |
|
"loss": 0.448, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": 0.035374294966459274, |
|
"rewards/margins": 0.7525378465652466, |
|
"rewards/rejected": -0.7171635627746582, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.9367088607594937, |
|
"grad_norm": 42.723620077832024, |
|
"learning_rate": 4.7275163104709194e-07, |
|
"logits/chosen": 2.223971128463745, |
|
"logits/rejected": 2.2867026329040527, |
|
"logps/chosen": -28.47149658203125, |
|
"logps/rejected": -41.30598831176758, |
|
"loss": 0.4752, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.04645688831806183, |
|
"rewards/margins": 0.7158762812614441, |
|
"rewards/rejected": -0.7623331546783447, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.9620253164556962, |
|
"grad_norm": 37.84394079738309, |
|
"learning_rate": 4.7014888296418447e-07, |
|
"logits/chosen": 2.7294678688049316, |
|
"logits/rejected": 2.7926061153411865, |
|
"logps/chosen": -27.44596290588379, |
|
"logps/rejected": -37.265743255615234, |
|
"loss": 0.4448, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": 0.049863241612911224, |
|
"rewards/margins": 1.077908992767334, |
|
"rewards/rejected": -1.0280457735061646, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.9873417721518988, |
|
"grad_norm": 37.542515394049275, |
|
"learning_rate": 4.6743528328892384e-07, |
|
"logits/chosen": 2.9349310398101807, |
|
"logits/rejected": 2.7823617458343506, |
|
"logps/chosen": -28.974937438964844, |
|
"logps/rejected": -31.127620697021484, |
|
"loss": 0.4247, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.06356163322925568, |
|
"rewards/margins": 0.8182165622711182, |
|
"rewards/rejected": -0.7546550035476685, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 1.0126582278481013, |
|
"grad_norm": 34.306915155273884, |
|
"learning_rate": 4.646121984004665e-07, |
|
"logits/chosen": 2.93210768699646, |
|
"logits/rejected": 2.952573299407959, |
|
"logps/chosen": -22.667604446411133, |
|
"logps/rejected": -33.4359130859375, |
|
"loss": 0.4023, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.15648899972438812, |
|
"rewards/margins": 1.7330085039138794, |
|
"rewards/rejected": -1.5765196084976196, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 1.0379746835443038, |
|
"grad_norm": 34.583417240102314, |
|
"learning_rate": 4.6168104980707103e-07, |
|
"logits/chosen": 1.8353512287139893, |
|
"logits/rejected": 1.8640844821929932, |
|
"logps/chosen": -30.39456558227539, |
|
"logps/rejected": -39.79166793823242, |
|
"loss": 0.3791, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.04578995332121849, |
|
"rewards/margins": 1.216768741607666, |
|
"rewards/rejected": -1.2625586986541748, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 1.0632911392405062, |
|
"grad_norm": 34.082479241604844, |
|
"learning_rate": 4.5864331343032565e-07, |
|
"logits/chosen": 1.8534090518951416, |
|
"logits/rejected": 2.3674142360687256, |
|
"logps/chosen": -24.82038116455078, |
|
"logps/rejected": -36.29607009887695, |
|
"loss": 0.3826, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.017678476870059967, |
|
"rewards/margins": 1.1888235807418823, |
|
"rewards/rejected": -1.1711452007293701, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 1.0886075949367089, |
|
"grad_norm": 34.56856221568917, |
|
"learning_rate": 4.555005188619775e-07, |
|
"logits/chosen": 2.3347442150115967, |
|
"logits/rejected": 2.896552562713623, |
|
"logps/chosen": -27.94430923461914, |
|
"logps/rejected": -56.70208740234375, |
|
"loss": 0.3719, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.003501923754811287, |
|
"rewards/margins": 1.5741243362426758, |
|
"rewards/rejected": -1.5706223249435425, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 1.1139240506329113, |
|
"grad_norm": 31.370937274347394, |
|
"learning_rate": 4.5225424859373684e-07, |
|
"logits/chosen": 2.6383707523345947, |
|
"logits/rejected": 2.5485527515411377, |
|
"logps/chosen": -19.298486709594727, |
|
"logps/rejected": -25.37928009033203, |
|
"loss": 0.3744, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": 0.22318463027477264, |
|
"rewards/margins": 0.6971991658210754, |
|
"rewards/rejected": -0.474014550447464, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 1.139240506329114, |
|
"grad_norm": 34.10062660077868, |
|
"learning_rate": 4.489061372204452e-07, |
|
"logits/chosen": 2.376051664352417, |
|
"logits/rejected": 2.3887717723846436, |
|
"logps/chosen": -27.828815460205078, |
|
"logps/rejected": -35.97361755371094, |
|
"loss": 0.3561, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.13765862584114075, |
|
"rewards/margins": 0.8190426826477051, |
|
"rewards/rejected": -0.6813840866088867, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 1.1645569620253164, |
|
"grad_norm": 34.550138211560416, |
|
"learning_rate": 4.4545787061700746e-07, |
|
"logits/chosen": 1.5758471488952637, |
|
"logits/rejected": 1.6474454402923584, |
|
"logps/chosen": -26.35508155822754, |
|
"logps/rejected": -33.66876983642578, |
|
"loss": 0.3738, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.11953898519277573, |
|
"rewards/margins": 1.4261627197265625, |
|
"rewards/rejected": -1.3066238164901733, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 1.189873417721519, |
|
"grad_norm": 32.97226834077475, |
|
"learning_rate": 4.4191118508950277e-07, |
|
"logits/chosen": 2.352506160736084, |
|
"logits/rejected": 2.5829250812530518, |
|
"logps/chosen": -30.464948654174805, |
|
"logps/rejected": -48.602386474609375, |
|
"loss": 0.3773, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.28972989320755005, |
|
"rewards/margins": 1.533113956451416, |
|
"rewards/rejected": -1.2433841228485107, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 1.2151898734177216, |
|
"grad_norm": 31.173818641433147, |
|
"learning_rate": 4.3826786650090273e-07, |
|
"logits/chosen": 2.6230416297912598, |
|
"logits/rejected": 2.9404265880584717, |
|
"logps/chosen": -26.442947387695312, |
|
"logps/rejected": -40.41877746582031, |
|
"loss": 0.3218, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.13079282641410828, |
|
"rewards/margins": 1.8220040798187256, |
|
"rewards/rejected": -1.6912113428115845, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 1.240506329113924, |
|
"grad_norm": 31.65330353079338, |
|
"learning_rate": 4.345297493718352e-07, |
|
"logits/chosen": 2.0961148738861084, |
|
"logits/rejected": 2.31225848197937, |
|
"logps/chosen": -28.76830291748047, |
|
"logps/rejected": -41.471527099609375, |
|
"loss": 0.324, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.17320382595062256, |
|
"rewards/margins": 1.8003501892089844, |
|
"rewards/rejected": -1.6271462440490723, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 1.2658227848101267, |
|
"grad_norm": 32.0517210362914, |
|
"learning_rate": 4.3069871595684787e-07, |
|
"logits/chosen": 2.477304458618164, |
|
"logits/rejected": 2.607898473739624, |
|
"logps/chosen": -25.64235496520996, |
|
"logps/rejected": -30.789039611816406, |
|
"loss": 0.3365, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.04965633898973465, |
|
"rewards/margins": 1.5173263549804688, |
|
"rewards/rejected": -1.4676700830459595, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.2911392405063291, |
|
"grad_norm": 34.75297133689012, |
|
"learning_rate": 4.2677669529663686e-07, |
|
"logits/chosen": 2.4322314262390137, |
|
"logits/rejected": 2.3749780654907227, |
|
"logps/chosen": -28.2196102142334, |
|
"logps/rejected": -36.66654968261719, |
|
"loss": 0.376, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -0.014162863604724407, |
|
"rewards/margins": 1.2780933380126953, |
|
"rewards/rejected": -1.292256236076355, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 1.3164556962025316, |
|
"grad_norm": 30.679462903169394, |
|
"learning_rate": 4.227656622467162e-07, |
|
"logits/chosen": 2.16550350189209, |
|
"logits/rejected": 2.483671188354492, |
|
"logps/chosen": -27.87743377685547, |
|
"logps/rejected": -35.79537582397461, |
|
"loss": 0.328, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.016814690083265305, |
|
"rewards/margins": 0.7990709543228149, |
|
"rewards/rejected": -0.8158857226371765, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 1.3417721518987342, |
|
"grad_norm": 33.362419000257866, |
|
"learning_rate": 4.186676364830186e-07, |
|
"logits/chosen": 1.6940243244171143, |
|
"logits/rejected": 1.8324474096298218, |
|
"logps/chosen": -22.575485229492188, |
|
"logps/rejected": -37.813812255859375, |
|
"loss": 0.3899, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.18489199876785278, |
|
"rewards/margins": 1.3045916557312012, |
|
"rewards/rejected": -1.4894835948944092, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 1.3670886075949367, |
|
"grad_norm": 34.27040807040801, |
|
"learning_rate": 4.1448468148492814e-07, |
|
"logits/chosen": 2.4027748107910156, |
|
"logits/rejected": 2.6370065212249756, |
|
"logps/chosen": -25.746681213378906, |
|
"logps/rejected": -39.262149810791016, |
|
"loss": 0.3439, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.14556269347667694, |
|
"rewards/margins": 1.5188615322113037, |
|
"rewards/rejected": -1.664424180984497, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 1.3924050632911391, |
|
"grad_norm": 30.395014985650786, |
|
"learning_rate": 4.10218903496256e-07, |
|
"logits/chosen": 1.9827089309692383, |
|
"logits/rejected": 2.0387563705444336, |
|
"logps/chosen": -23.929346084594727, |
|
"logps/rejected": -31.56798553466797, |
|
"loss": 0.3347, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.012682847678661346, |
|
"rewards/margins": 1.1076587438583374, |
|
"rewards/rejected": -1.0949759483337402, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 1.4177215189873418, |
|
"grad_norm": 30.07232120551146, |
|
"learning_rate": 4.058724504646834e-07, |
|
"logits/chosen": 2.144460916519165, |
|
"logits/rejected": 2.3883056640625, |
|
"logps/chosen": -29.003910064697266, |
|
"logps/rejected": -48.30259704589844, |
|
"loss": 0.3069, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.12483976036310196, |
|
"rewards/margins": 2.0643203258514404, |
|
"rewards/rejected": -1.9394805431365967, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 1.4430379746835442, |
|
"grad_norm": 29.37804816237709, |
|
"learning_rate": 4.0144751096020497e-07, |
|
"logits/chosen": 2.4025626182556152, |
|
"logits/rejected": 2.6997711658477783, |
|
"logps/chosen": -20.084449768066406, |
|
"logps/rejected": -41.12477493286133, |
|
"loss": 0.2932, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": -0.05090276896953583, |
|
"rewards/margins": 1.5857845544815063, |
|
"rewards/rejected": -1.6366872787475586, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 1.4683544303797469, |
|
"grad_norm": 28.707949914480356, |
|
"learning_rate": 3.9694631307311825e-07, |
|
"logits/chosen": 1.7879891395568848, |
|
"logits/rejected": 1.9085421562194824, |
|
"logps/chosen": -24.52581214904785, |
|
"logps/rejected": -42.650333404541016, |
|
"loss": 0.2848, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -0.09958541393280029, |
|
"rewards/margins": 1.8922325372695923, |
|
"rewards/rejected": -1.991817831993103, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 1.4936708860759493, |
|
"grad_norm": 30.2249376141746, |
|
"learning_rate": 3.92371123292113e-07, |
|
"logits/chosen": 2.3484342098236084, |
|
"logits/rejected": 2.6477420330047607, |
|
"logps/chosen": -25.54952621459961, |
|
"logps/rejected": -42.36656951904297, |
|
"loss": 0.2884, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.025862978771328926, |
|
"rewards/margins": 2.4160983562469482, |
|
"rewards/rejected": -2.4419612884521484, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 1.518987341772152, |
|
"grad_norm": 29.710974044035055, |
|
"learning_rate": 3.877242453630256e-07, |
|
"logits/chosen": 2.434805154800415, |
|
"logits/rejected": 2.456228256225586, |
|
"logps/chosen": -20.521100997924805, |
|
"logps/rejected": -28.69353485107422, |
|
"loss": 0.3295, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.3444966673851013, |
|
"rewards/margins": 2.016552448272705, |
|
"rewards/rejected": -1.672055959701538, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.5443037974683544, |
|
"grad_norm": 33.402955887536194, |
|
"learning_rate": 3.8300801912883414e-07, |
|
"logits/chosen": 2.511925220489502, |
|
"logits/rejected": 2.43487286567688, |
|
"logps/chosen": -30.91568946838379, |
|
"logps/rejected": -38.48870086669922, |
|
"loss": 0.297, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.009598510339856148, |
|
"rewards/margins": 1.2743103504180908, |
|
"rewards/rejected": -1.2839089632034302, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 1.5696202531645569, |
|
"grad_norm": 34.221583251712744, |
|
"learning_rate": 3.7822481935147655e-07, |
|
"logits/chosen": 2.2030210494995117, |
|
"logits/rejected": 2.4195468425750732, |
|
"logps/chosen": -26.423166275024414, |
|
"logps/rejected": -36.79461669921875, |
|
"loss": 0.3281, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": -0.2358216643333435, |
|
"rewards/margins": 1.8056204319000244, |
|
"rewards/rejected": -2.0414421558380127, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 1.5949367088607596, |
|
"grad_norm": 27.39193509195394, |
|
"learning_rate": 3.7337705451608667e-07, |
|
"logits/chosen": 2.442267894744873, |
|
"logits/rejected": 2.6541049480438232, |
|
"logps/chosen": -24.891948699951172, |
|
"logps/rejected": -40.318904876708984, |
|
"loss": 0.2564, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.08293318003416061, |
|
"rewards/margins": 1.7536163330078125, |
|
"rewards/rejected": -1.8365494012832642, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 1.620253164556962, |
|
"grad_norm": 32.95811212676831, |
|
"learning_rate": 3.6846716561824967e-07, |
|
"logits/chosen": 2.404311180114746, |
|
"logits/rejected": 2.50986385345459, |
|
"logps/chosen": -24.17348861694336, |
|
"logps/rejected": -35.517120361328125, |
|
"loss": 0.3643, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.00808901246637106, |
|
"rewards/margins": 1.9171310663223267, |
|
"rewards/rejected": -1.925220251083374, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 1.6455696202531644, |
|
"grad_norm": 31.89102457068134, |
|
"learning_rate": 3.634976249348867e-07, |
|
"logits/chosen": 1.8697429895401, |
|
"logits/rejected": 1.9369537830352783, |
|
"logps/chosen": -26.27237319946289, |
|
"logps/rejected": -43.584659576416016, |
|
"loss": 0.3047, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.018543783575296402, |
|
"rewards/margins": 1.540452480316162, |
|
"rewards/rejected": -1.5219086408615112, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 1.6708860759493671, |
|
"grad_norm": 29.55969704854734, |
|
"learning_rate": 3.584709347793895e-07, |
|
"logits/chosen": 2.12164044380188, |
|
"logits/rejected": 2.12618350982666, |
|
"logps/chosen": -19.538429260253906, |
|
"logps/rejected": -34.374759674072266, |
|
"loss": 0.3159, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.1228712797164917, |
|
"rewards/margins": 1.97737455368042, |
|
"rewards/rejected": -1.8545031547546387, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 1.6962025316455698, |
|
"grad_norm": 29.63374295371732, |
|
"learning_rate": 3.5338962624163016e-07, |
|
"logits/chosen": 2.3299381732940674, |
|
"logits/rejected": 2.5256855487823486, |
|
"logps/chosen": -24.86815643310547, |
|
"logps/rejected": -32.63747024536133, |
|
"loss": 0.304, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.14013400673866272, |
|
"rewards/margins": 1.9178043603897095, |
|
"rewards/rejected": -2.05793833732605, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 1.721518987341772, |
|
"grad_norm": 29.816270030062956, |
|
"learning_rate": 3.482562579134809e-07, |
|
"logits/chosen": 2.947877883911133, |
|
"logits/rejected": 2.919872760772705, |
|
"logps/chosen": -24.169422149658203, |
|
"logps/rejected": -35.171504974365234, |
|
"loss": 0.2947, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.18604925274848938, |
|
"rewards/margins": 2.742734432220459, |
|
"rewards/rejected": -2.556685209274292, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 1.7468354430379747, |
|
"grad_norm": 28.204359911989563, |
|
"learning_rate": 3.430734146004863e-07, |
|
"logits/chosen": 2.15493106842041, |
|
"logits/rejected": 2.3170015811920166, |
|
"logps/chosen": -22.2800350189209, |
|
"logps/rejected": -35.93194580078125, |
|
"loss": 0.2844, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -0.17889335751533508, |
|
"rewards/margins": 2.3182311058044434, |
|
"rewards/rejected": -2.497124433517456, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 1.7721518987341773, |
|
"grad_norm": 26.787627596958465, |
|
"learning_rate": 3.378437060203357e-07, |
|
"logits/chosen": 2.5400497913360596, |
|
"logits/rejected": 2.5072226524353027, |
|
"logps/chosen": -21.618453979492188, |
|
"logps/rejected": -33.122589111328125, |
|
"loss": 0.2766, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.05331309512257576, |
|
"rewards/margins": 2.7737677097320557, |
|
"rewards/rejected": -2.720454692840576, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.7974683544303798, |
|
"grad_norm": 31.205788121930055, |
|
"learning_rate": 3.325697654887918e-07, |
|
"logits/chosen": 1.4359736442565918, |
|
"logits/rejected": 1.0823308229446411, |
|
"logps/chosen": -26.589662551879883, |
|
"logps/rejected": -38.18647766113281, |
|
"loss": 0.2822, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.2071526050567627, |
|
"rewards/margins": 2.4934191703796387, |
|
"rewards/rejected": -2.286266565322876, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 1.8227848101265822, |
|
"grad_norm": 29.788415172955876, |
|
"learning_rate": 3.272542485937368e-07, |
|
"logits/chosen": 2.800675868988037, |
|
"logits/rejected": 2.9707438945770264, |
|
"logps/chosen": -25.928173065185547, |
|
"logps/rejected": -38.25574493408203, |
|
"loss": 0.2955, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.0182593185454607, |
|
"rewards/margins": 2.1046228408813477, |
|
"rewards/rejected": -2.0863637924194336, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 1.8481012658227849, |
|
"grad_norm": 31.197362620805283, |
|
"learning_rate": 3.218998318580043e-07, |
|
"logits/chosen": 2.2215137481689453, |
|
"logits/rejected": 2.1676788330078125, |
|
"logps/chosen": -23.131969451904297, |
|
"logps/rejected": -34.484920501708984, |
|
"loss": 0.276, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.19679217040538788, |
|
"rewards/margins": 2.1039047241210938, |
|
"rewards/rejected": -1.9071124792099, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 1.8734177215189873, |
|
"grad_norm": 27.14206740038897, |
|
"learning_rate": 3.1650921139166874e-07, |
|
"logits/chosen": 2.0885462760925293, |
|
"logits/rejected": 2.1695892810821533, |
|
"logps/chosen": -24.332992553710938, |
|
"logps/rejected": -39.35213088989258, |
|
"loss": 0.2598, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.13339604437351227, |
|
"rewards/margins": 2.6932549476623535, |
|
"rewards/rejected": -2.559858798980713, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 1.8987341772151898, |
|
"grad_norm": 31.93987675909966, |
|
"learning_rate": 3.110851015344735e-07, |
|
"logits/chosen": 2.1219115257263184, |
|
"logits/rejected": 2.1965231895446777, |
|
"logps/chosen": -25.27581024169922, |
|
"logps/rejected": -34.003536224365234, |
|
"loss": 0.3142, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.1961415410041809, |
|
"rewards/margins": 2.4378857612609863, |
|
"rewards/rejected": -2.24174427986145, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.9240506329113924, |
|
"grad_norm": 29.021478328579693, |
|
"learning_rate": 3.056302334890786e-07, |
|
"logits/chosen": 1.8746654987335205, |
|
"logits/rejected": 2.389371871948242, |
|
"logps/chosen": -22.446157455444336, |
|
"logps/rejected": -39.833770751953125, |
|
"loss": 0.2671, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.18668973445892334, |
|
"rewards/margins": 3.256779909133911, |
|
"rewards/rejected": -3.0700900554656982, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 1.9493670886075949, |
|
"grad_norm": 31.861473137885586, |
|
"learning_rate": 3.001473539458182e-07, |
|
"logits/chosen": 2.0852279663085938, |
|
"logits/rejected": 2.2871861457824707, |
|
"logps/chosen": -27.48174285888672, |
|
"logps/rejected": -38.38823699951172, |
|
"loss": 0.2623, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.12376511096954346, |
|
"rewards/margins": 2.879195213317871, |
|
"rewards/rejected": -2.755429983139038, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 1.9746835443037973, |
|
"grad_norm": 30.580581160948984, |
|
"learning_rate": 2.9463922369965915e-07, |
|
"logits/chosen": 2.183192253112793, |
|
"logits/rejected": 1.8703649044036865, |
|
"logps/chosen": -25.529539108276367, |
|
"logps/rejected": -38.35371780395508, |
|
"loss": 0.2812, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -0.09511762112379074, |
|
"rewards/margins": 1.8313283920288086, |
|
"rewards/rejected": -1.9264459609985352, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 28.28932497860918, |
|
"learning_rate": 2.8910861626005773e-07, |
|
"logits/chosen": 2.1175851821899414, |
|
"logits/rejected": 2.7737534046173096, |
|
"logps/chosen": -22.948108673095703, |
|
"logps/rejected": -54.805484771728516, |
|
"loss": 0.2737, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.07266075909137726, |
|
"rewards/margins": 2.9845314025878906, |
|
"rewards/rejected": -3.057192087173462, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 2.0253164556962027, |
|
"grad_norm": 22.84435575128268, |
|
"learning_rate": 2.8355831645441387e-07, |
|
"logits/chosen": 2.129598617553711, |
|
"logits/rejected": 2.214559316635132, |
|
"logps/chosen": -33.85695266723633, |
|
"logps/rejected": -40.15086364746094, |
|
"loss": 0.2116, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.04292801395058632, |
|
"rewards/margins": 2.009500741958618, |
|
"rewards/rejected": -2.052428722381592, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 2.050632911392405, |
|
"grad_norm": 22.68444850198704, |
|
"learning_rate": 2.7799111902582693e-07, |
|
"logits/chosen": 2.512767791748047, |
|
"logits/rejected": 2.5582704544067383, |
|
"logps/chosen": -25.28483009338379, |
|
"logps/rejected": -36.232017517089844, |
|
"loss": 0.2011, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.14101354777812958, |
|
"rewards/margins": 2.3776185512542725, |
|
"rewards/rejected": -2.236604690551758, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 2.0759493670886076, |
|
"grad_norm": 23.440024916384186, |
|
"learning_rate": 2.7240982722585837e-07, |
|
"logits/chosen": 1.941485047340393, |
|
"logits/rejected": 2.076082944869995, |
|
"logps/chosen": -25.91452407836914, |
|
"logps/rejected": -34.40084457397461, |
|
"loss": 0.2306, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.1705533266067505, |
|
"rewards/margins": 1.7886393070220947, |
|
"rewards/rejected": -1.9591926336288452, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 2.1012658227848102, |
|
"grad_norm": 24.39463073368113, |
|
"learning_rate": 2.6681725140300995e-07, |
|
"logits/chosen": 1.5756129026412964, |
|
"logits/rejected": 1.9195802211761475, |
|
"logps/chosen": -21.2039737701416, |
|
"logps/rejected": -43.85890579223633, |
|
"loss": 0.2048, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.012569274753332138, |
|
"rewards/margins": 3.1985719203948975, |
|
"rewards/rejected": -3.2111411094665527, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 2.1265822784810124, |
|
"grad_norm": 21.14214956646525, |
|
"learning_rate": 2.6121620758762875e-07, |
|
"logits/chosen": 2.5736641883850098, |
|
"logits/rejected": 2.7494866847991943, |
|
"logps/chosen": -27.16839027404785, |
|
"logps/rejected": -37.205440521240234, |
|
"loss": 0.1933, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.010486233979463577, |
|
"rewards/margins": 2.571336269378662, |
|
"rewards/rejected": -2.560849905014038, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 2.151898734177215, |
|
"grad_norm": 21.59512651806526, |
|
"learning_rate": 2.5560951607395126e-07, |
|
"logits/chosen": 2.4533567428588867, |
|
"logits/rejected": 2.3579037189483643, |
|
"logps/chosen": -28.007740020751953, |
|
"logps/rejected": -36.33196258544922, |
|
"loss": 0.192, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.04641330614686012, |
|
"rewards/margins": 1.9481606483459473, |
|
"rewards/rejected": -1.901747465133667, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 2.1772151898734178, |
|
"grad_norm": 24.879344750048478, |
|
"learning_rate": 2.5e-07, |
|
"logits/chosen": 2.1627023220062256, |
|
"logits/rejected": 2.059532403945923, |
|
"logps/chosen": -32.905269622802734, |
|
"logps/rejected": -31.235652923583984, |
|
"loss": 0.2379, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.02136234939098358, |
|
"rewards/margins": 1.5867081880569458, |
|
"rewards/rejected": -1.6080706119537354, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 2.2025316455696204, |
|
"grad_norm": 22.69000503277292, |
|
"learning_rate": 2.4439048392604877e-07, |
|
"logits/chosen": 2.3072502613067627, |
|
"logits/rejected": 2.6580121517181396, |
|
"logps/chosen": -23.412647247314453, |
|
"logps/rejected": -42.0001106262207, |
|
"loss": 0.2005, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.16801214218139648, |
|
"rewards/margins": 2.332923412322998, |
|
"rewards/rejected": -2.5009355545043945, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 2.2278481012658227, |
|
"grad_norm": 24.693555115896494, |
|
"learning_rate": 2.3878379241237134e-07, |
|
"logits/chosen": 2.4124932289123535, |
|
"logits/rejected": 2.2819831371307373, |
|
"logps/chosen": -25.710378646850586, |
|
"logps/rejected": -32.972537994384766, |
|
"loss": 0.2211, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.0028405077755451202, |
|
"rewards/margins": 2.1923816204071045, |
|
"rewards/rejected": -2.1895413398742676, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 2.2531645569620253, |
|
"grad_norm": 22.755579101890042, |
|
"learning_rate": 2.3318274859699008e-07, |
|
"logits/chosen": 2.006254196166992, |
|
"logits/rejected": 2.4240238666534424, |
|
"logps/chosen": -26.35239028930664, |
|
"logps/rejected": -42.78788757324219, |
|
"loss": 0.2437, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -0.381151020526886, |
|
"rewards/margins": 2.395555019378662, |
|
"rewards/rejected": -2.7767062187194824, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 2.278481012658228, |
|
"grad_norm": 23.73790882415418, |
|
"learning_rate": 2.2759017277414164e-07, |
|
"logits/chosen": 2.482501745223999, |
|
"logits/rejected": 2.456974744796753, |
|
"logps/chosen": -32.156795501708984, |
|
"logps/rejected": -35.610713958740234, |
|
"loss": 0.2328, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.08202718943357468, |
|
"rewards/margins": 1.4416414499282837, |
|
"rewards/rejected": -1.523668646812439, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 2.3037974683544302, |
|
"grad_norm": 22.47212835812806, |
|
"learning_rate": 2.2200888097417302e-07, |
|
"logits/chosen": 2.5853986740112305, |
|
"logits/rejected": 2.6601390838623047, |
|
"logps/chosen": -33.39643859863281, |
|
"logps/rejected": -43.778709411621094, |
|
"loss": 0.1828, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.010579407215118408, |
|
"rewards/margins": 2.473262310028076, |
|
"rewards/rejected": -2.483841896057129, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 2.329113924050633, |
|
"grad_norm": 22.509363583762184, |
|
"learning_rate": 2.164416835455862e-07, |
|
"logits/chosen": 2.3972582817077637, |
|
"logits/rejected": 2.534813642501831, |
|
"logps/chosen": -22.962604522705078, |
|
"logps/rejected": -37.55415344238281, |
|
"loss": 0.2213, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.0858815610408783, |
|
"rewards/margins": 2.756964683532715, |
|
"rewards/rejected": -2.842846155166626, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 2.3544303797468356, |
|
"grad_norm": 22.253217260948635, |
|
"learning_rate": 2.1089138373994222e-07, |
|
"logits/chosen": 2.153351306915283, |
|
"logits/rejected": 2.3578603267669678, |
|
"logps/chosen": -22.811275482177734, |
|
"logps/rejected": -39.044639587402344, |
|
"loss": 0.1953, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.2400507628917694, |
|
"rewards/margins": 2.045787811279297, |
|
"rewards/rejected": -2.2858386039733887, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 2.379746835443038, |
|
"grad_norm": 23.13742394622883, |
|
"learning_rate": 2.0536077630034085e-07, |
|
"logits/chosen": 1.7368831634521484, |
|
"logits/rejected": 1.9436362981796265, |
|
"logps/chosen": -26.176815032958984, |
|
"logps/rejected": -40.83625411987305, |
|
"loss": 0.2061, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.39153048396110535, |
|
"rewards/margins": 2.8905537128448486, |
|
"rewards/rejected": -2.499023199081421, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 2.4050632911392404, |
|
"grad_norm": 21.94757471208971, |
|
"learning_rate": 1.998526460541818e-07, |
|
"logits/chosen": 2.3502449989318848, |
|
"logits/rejected": 2.3544671535491943, |
|
"logps/chosen": -26.833280563354492, |
|
"logps/rejected": -37.404605865478516, |
|
"loss": 0.1962, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.0008228495717048645, |
|
"rewards/margins": 2.2684872150421143, |
|
"rewards/rejected": -2.2676644325256348, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 2.430379746835443, |
|
"grad_norm": 20.124080148253146, |
|
"learning_rate": 1.9436976651092142e-07, |
|
"logits/chosen": 1.1564311981201172, |
|
"logits/rejected": 1.8700743913650513, |
|
"logps/chosen": -25.664215087890625, |
|
"logps/rejected": -42.25444030761719, |
|
"loss": 0.1803, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.013121239840984344, |
|
"rewards/margins": 2.5192627906799316, |
|
"rewards/rejected": -2.5061416625976562, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 2.4556962025316453, |
|
"grad_norm": 22.336736253170702, |
|
"learning_rate": 1.8891489846552644e-07, |
|
"logits/chosen": 2.3158211708068848, |
|
"logits/rejected": 2.718118667602539, |
|
"logps/chosen": -23.182619094848633, |
|
"logps/rejected": -41.77816390991211, |
|
"loss": 0.2237, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.010374765843153, |
|
"rewards/margins": 2.8450756072998047, |
|
"rewards/rejected": -2.8347012996673584, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 2.481012658227848, |
|
"grad_norm": 21.95104798778307, |
|
"learning_rate": 1.8349078860833124e-07, |
|
"logits/chosen": 2.2794296741485596, |
|
"logits/rejected": 2.176607847213745, |
|
"logps/chosen": -33.21621322631836, |
|
"logps/rejected": -38.940025329589844, |
|
"loss": 0.1762, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.013390496373176575, |
|
"rewards/margins": 2.8369154930114746, |
|
"rewards/rejected": -2.8503060340881348, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 2.5063291139240507, |
|
"grad_norm": 22.129805040414322, |
|
"learning_rate": 1.781001681419957e-07, |
|
"logits/chosen": 1.8799982070922852, |
|
"logits/rejected": 2.1679162979125977, |
|
"logps/chosen": -20.39841079711914, |
|
"logps/rejected": -40.38523864746094, |
|
"loss": 0.2032, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.04029037803411484, |
|
"rewards/margins": 2.2781383991241455, |
|
"rewards/rejected": -2.2378478050231934, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 2.5316455696202533, |
|
"grad_norm": 18.828604860640993, |
|
"learning_rate": 1.7274575140626315e-07, |
|
"logits/chosen": 2.4270401000976562, |
|
"logits/rejected": 2.768435478210449, |
|
"logps/chosen": -28.433048248291016, |
|
"logps/rejected": -47.266197204589844, |
|
"loss": 0.1502, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.2632801830768585, |
|
"rewards/margins": 2.7442736625671387, |
|
"rewards/rejected": -3.007554054260254, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 2.5569620253164556, |
|
"grad_norm": 20.880074717956166, |
|
"learning_rate": 1.674302345112083e-07, |
|
"logits/chosen": 1.9880956411361694, |
|
"logits/rejected": 2.2008445262908936, |
|
"logps/chosen": -24.262271881103516, |
|
"logps/rejected": -45.60854721069336, |
|
"loss": 0.2089, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.04222775995731354, |
|
"rewards/margins": 3.01908802986145, |
|
"rewards/rejected": -2.9768600463867188, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 2.5822784810126582, |
|
"grad_norm": 21.838535925363054, |
|
"learning_rate": 1.621562939796643e-07, |
|
"logits/chosen": 2.205430746078491, |
|
"logits/rejected": 2.4456026554107666, |
|
"logps/chosen": -24.633563995361328, |
|
"logps/rejected": -37.809478759765625, |
|
"loss": 0.2187, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.2103780210018158, |
|
"rewards/margins": 2.4951250553131104, |
|
"rewards/rejected": -2.2847468852996826, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 2.607594936708861, |
|
"grad_norm": 22.277589914673204, |
|
"learning_rate": 1.569265853995137e-07, |
|
"logits/chosen": 1.4377330541610718, |
|
"logits/rejected": 1.4871635437011719, |
|
"logps/chosen": -24.512638092041016, |
|
"logps/rejected": -37.595733642578125, |
|
"loss": 0.1964, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.3236038088798523, |
|
"rewards/margins": 2.621075391769409, |
|
"rewards/rejected": -2.9446792602539062, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 2.632911392405063, |
|
"grad_norm": 21.07481872439906, |
|
"learning_rate": 1.517437420865191e-07, |
|
"logits/chosen": 1.6525717973709106, |
|
"logits/rejected": 2.0419435501098633, |
|
"logps/chosen": -24.287813186645508, |
|
"logps/rejected": -43.26752853393555, |
|
"loss": 0.1873, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.2492070347070694, |
|
"rewards/margins": 3.5699872970581055, |
|
"rewards/rejected": -3.3207802772521973, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 2.6582278481012658, |
|
"grad_norm": 20.677444088444823, |
|
"learning_rate": 1.4661037375836987e-07, |
|
"logits/chosen": 2.6604502201080322, |
|
"logits/rejected": 2.587705373764038, |
|
"logps/chosen": -25.429487228393555, |
|
"logps/rejected": -36.87110137939453, |
|
"loss": 0.1855, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.07871632277965546, |
|
"rewards/margins": 2.1279850006103516, |
|
"rewards/rejected": -2.2067012786865234, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 2.6835443037974684, |
|
"grad_norm": 23.5751426913938, |
|
"learning_rate": 1.4152906522061047e-07, |
|
"logits/chosen": 2.1078615188598633, |
|
"logits/rejected": 2.089411735534668, |
|
"logps/chosen": -25.450345993041992, |
|
"logps/rejected": -33.47675323486328, |
|
"loss": 0.1944, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.11998379230499268, |
|
"rewards/margins": 2.037046432495117, |
|
"rewards/rejected": -1.917062759399414, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 2.708860759493671, |
|
"grad_norm": 22.767643882863975, |
|
"learning_rate": 1.365023750651133e-07, |
|
"logits/chosen": 1.7534555196762085, |
|
"logits/rejected": 2.2760560512542725, |
|
"logps/chosen": -27.695234298706055, |
|
"logps/rejected": -46.436767578125, |
|
"loss": 0.1638, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.030735721811652184, |
|
"rewards/margins": 3.136977434158325, |
|
"rewards/rejected": -3.1677134037017822, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 2.7341772151898733, |
|
"grad_norm": 20.58426575505033, |
|
"learning_rate": 1.3153283438175034e-07, |
|
"logits/chosen": 2.3597166538238525, |
|
"logits/rejected": 2.5851235389709473, |
|
"logps/chosen": -23.936431884765625, |
|
"logps/rejected": -47.1025390625, |
|
"loss": 0.1785, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.1647489368915558, |
|
"rewards/margins": 3.3675107955932617, |
|
"rewards/rejected": -3.532259941101074, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 2.759493670886076, |
|
"grad_norm": 21.349116101829402, |
|
"learning_rate": 1.2662294548391328e-07, |
|
"logits/chosen": 1.9581003189086914, |
|
"logits/rejected": 1.994449496269226, |
|
"logps/chosen": -28.851770401000977, |
|
"logps/rejected": -39.374019622802734, |
|
"loss": 0.1772, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.023888343945145607, |
|
"rewards/margins": 2.033907890319824, |
|
"rewards/rejected": -2.0100197792053223, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 2.7848101265822782, |
|
"grad_norm": 20.523600951939002, |
|
"learning_rate": 1.2177518064852348e-07, |
|
"logits/chosen": 2.149280071258545, |
|
"logits/rejected": 2.338773012161255, |
|
"logps/chosen": -24.151710510253906, |
|
"logps/rejected": -44.250370025634766, |
|
"loss": 0.197, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.008578717708587646, |
|
"rewards/margins": 2.208198308944702, |
|
"rewards/rejected": -2.1996195316314697, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 2.810126582278481, |
|
"grad_norm": 21.793053109415474, |
|
"learning_rate": 1.1699198087116588e-07, |
|
"logits/chosen": 1.9670238494873047, |
|
"logits/rejected": 2.2674317359924316, |
|
"logps/chosen": -19.423011779785156, |
|
"logps/rejected": -43.466705322265625, |
|
"loss": 0.1912, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.0942595824599266, |
|
"rewards/margins": 2.482712984085083, |
|
"rewards/rejected": -2.576972484588623, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 2.8354430379746836, |
|
"grad_norm": 18.7267582468967, |
|
"learning_rate": 1.1227575463697439e-07, |
|
"logits/chosen": 2.125753402709961, |
|
"logits/rejected": 2.3489747047424316, |
|
"logps/chosen": -24.716629028320312, |
|
"logps/rejected": -35.32937240600586, |
|
"loss": 0.1497, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.06388971209526062, |
|
"rewards/margins": 2.4911532402038574, |
|
"rewards/rejected": -2.5550432205200195, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 2.8607594936708862, |
|
"grad_norm": 21.953907512144703, |
|
"learning_rate": 1.0762887670788701e-07, |
|
"logits/chosen": 1.75847327709198, |
|
"logits/rejected": 1.766452431678772, |
|
"logps/chosen": -33.99777603149414, |
|
"logps/rejected": -49.58311080932617, |
|
"loss": 0.1887, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.09381230175495148, |
|
"rewards/margins": 3.5883729457855225, |
|
"rewards/rejected": -3.682185649871826, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 2.8860759493670884, |
|
"grad_norm": 21.263544038283367, |
|
"learning_rate": 1.0305368692688174e-07, |
|
"logits/chosen": 1.9139025211334229, |
|
"logits/rejected": 1.8650161027908325, |
|
"logps/chosen": -17.992523193359375, |
|
"logps/rejected": -38.388519287109375, |
|
"loss": 0.1854, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.09737363457679749, |
|
"rewards/margins": 3.4405698776245117, |
|
"rewards/rejected": -3.3431966304779053, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 2.911392405063291, |
|
"grad_norm": 19.070922554256914, |
|
"learning_rate": 9.855248903979505e-08, |
|
"logits/chosen": 2.087536096572876, |
|
"logits/rejected": 2.118912935256958, |
|
"logps/chosen": -30.79559326171875, |
|
"logps/rejected": -47.3471565246582, |
|
"loss": 0.1693, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.07975523173809052, |
|
"rewards/margins": 3.129413366317749, |
|
"rewards/rejected": -3.2091684341430664, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 2.9367088607594938, |
|
"grad_norm": 21.46961371171142, |
|
"learning_rate": 9.412754953531663e-08, |
|
"logits/chosen": 1.7869651317596436, |
|
"logits/rejected": 1.881705641746521, |
|
"logps/chosen": -24.043615341186523, |
|
"logps/rejected": -38.01569366455078, |
|
"loss": 0.1852, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.10924415290355682, |
|
"rewards/margins": 2.661862373352051, |
|
"rewards/rejected": -2.5526182651519775, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 2.962025316455696, |
|
"grad_norm": 19.929759669575105, |
|
"learning_rate": 8.978109650374396e-08, |
|
"logits/chosen": 1.7048578262329102, |
|
"logits/rejected": 1.9161901473999023, |
|
"logps/chosen": -26.87795639038086, |
|
"logps/rejected": -42.6606330871582, |
|
"loss": 0.1589, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.13115011155605316, |
|
"rewards/margins": 3.121182441711426, |
|
"rewards/rejected": -2.990032196044922, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 2.9873417721518987, |
|
"grad_norm": 25.118856703851232, |
|
"learning_rate": 8.551531851507185e-08, |
|
"logits/chosen": 2.3607137203216553, |
|
"logits/rejected": 2.5181257724761963, |
|
"logps/chosen": -30.176773071289062, |
|
"logps/rejected": -44.625003814697266, |
|
"loss": 0.224, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -0.30477192997932434, |
|
"rewards/margins": 2.4585647583007812, |
|
"rewards/rejected": -2.763336658477783, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 3.0126582278481013, |
|
"grad_norm": 24.215057475486347, |
|
"learning_rate": 8.133236351698142e-08, |
|
"logits/chosen": 1.6108707189559937, |
|
"logits/rejected": 1.7527028322219849, |
|
"logps/chosen": -18.345888137817383, |
|
"logps/rejected": -42.66941833496094, |
|
"loss": 0.2033, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.02492370456457138, |
|
"rewards/margins": 3.3315927982330322, |
|
"rewards/rejected": -3.306669235229492, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 3.037974683544304, |
|
"grad_norm": 18.940438234622814, |
|
"learning_rate": 7.723433775328384e-08, |
|
"logits/chosen": 2.0556139945983887, |
|
"logits/rejected": 2.3014981746673584, |
|
"logps/chosen": -21.665407180786133, |
|
"logps/rejected": -41.777896881103516, |
|
"loss": 0.1836, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.17341355979442596, |
|
"rewards/margins": 2.674577236175537, |
|
"rewards/rejected": -2.5011634826660156, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 3.0632911392405062, |
|
"grad_norm": 18.297883681094888, |
|
"learning_rate": 7.322330470336313e-08, |
|
"logits/chosen": 2.301410675048828, |
|
"logits/rejected": 2.2454984188079834, |
|
"logps/chosen": -23.815502166748047, |
|
"logps/rejected": -41.840091705322266, |
|
"loss": 0.1505, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.02547144889831543, |
|
"rewards/margins": 3.206540584564209, |
|
"rewards/rejected": -3.1810688972473145, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 3.088607594936709, |
|
"grad_norm": 18.507140827349733, |
|
"learning_rate": 6.930128404315214e-08, |
|
"logits/chosen": 2.0005404949188232, |
|
"logits/rejected": 2.2419371604919434, |
|
"logps/chosen": -23.189014434814453, |
|
"logps/rejected": -44.778133392333984, |
|
"loss": 0.1697, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.21837717294692993, |
|
"rewards/margins": 2.497182846069336, |
|
"rewards/rejected": -2.715559959411621, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 3.1139240506329116, |
|
"grad_norm": 20.458085524449526, |
|
"learning_rate": 6.547025062816486e-08, |
|
"logits/chosen": 2.5110156536102295, |
|
"logits/rejected": 2.835902452468872, |
|
"logps/chosen": -21.34715461730957, |
|
"logps/rejected": -42.06149673461914, |
|
"loss": 0.1769, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.2778359651565552, |
|
"rewards/margins": 3.1256825923919678, |
|
"rewards/rejected": -2.847846508026123, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 3.1392405063291138, |
|
"grad_norm": 19.657986671859284, |
|
"learning_rate": 6.173213349909728e-08, |
|
"logits/chosen": 2.4499258995056152, |
|
"logits/rejected": 2.4804461002349854, |
|
"logps/chosen": -24.912944793701172, |
|
"logps/rejected": -38.67371368408203, |
|
"loss": 0.172, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.06854649633169174, |
|
"rewards/margins": 2.4865663051605225, |
|
"rewards/rejected": -2.5551130771636963, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 3.1645569620253164, |
|
"grad_norm": 18.87674548749799, |
|
"learning_rate": 5.808881491049722e-08, |
|
"logits/chosen": 1.3738226890563965, |
|
"logits/rejected": 1.6148171424865723, |
|
"logps/chosen": -25.101566314697266, |
|
"logps/rejected": -40.51878356933594, |
|
"loss": 0.1786, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.33920496702194214, |
|
"rewards/margins": 2.1792898178100586, |
|
"rewards/rejected": -2.5184948444366455, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 3.189873417721519, |
|
"grad_norm": 20.113483471649342, |
|
"learning_rate": 5.454212938299255e-08, |
|
"logits/chosen": 2.0692553520202637, |
|
"logits/rejected": 2.1005959510803223, |
|
"logps/chosen": -26.763628005981445, |
|
"logps/rejected": -37.69628143310547, |
|
"loss": 0.1624, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.04393099620938301, |
|
"rewards/margins": 2.375842809677124, |
|
"rewards/rejected": -2.419773817062378, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 3.2151898734177213, |
|
"grad_norm": 19.33193488722329, |
|
"learning_rate": 5.109386277955477e-08, |
|
"logits/chosen": 1.9354777336120605, |
|
"logits/rejected": 2.3222734928131104, |
|
"logps/chosen": -22.444690704345703, |
|
"logps/rejected": -42.732322692871094, |
|
"loss": 0.1615, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.1692490577697754, |
|
"rewards/margins": 2.950117349624634, |
|
"rewards/rejected": -2.7808682918548584, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 3.240506329113924, |
|
"grad_norm": 17.657011863002797, |
|
"learning_rate": 4.774575140626316e-08, |
|
"logits/chosen": 1.7825671434402466, |
|
"logits/rejected": 2.0206806659698486, |
|
"logps/chosen": -26.314062118530273, |
|
"logps/rejected": -44.959896087646484, |
|
"loss": 0.1292, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.18020519614219666, |
|
"rewards/margins": 3.4240498542785645, |
|
"rewards/rejected": -3.243844747543335, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 3.2658227848101267, |
|
"grad_norm": 17.36688426753045, |
|
"learning_rate": 4.449948113802254e-08, |
|
"logits/chosen": 1.963316559791565, |
|
"logits/rejected": 2.073986530303955, |
|
"logps/chosen": -29.986591339111328, |
|
"logps/rejected": -46.73766326904297, |
|
"loss": 0.1476, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.1175435334444046, |
|
"rewards/margins": 3.3601157665252686, |
|
"rewards/rejected": -3.477659225463867, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 3.291139240506329, |
|
"grad_norm": 17.66376104354918, |
|
"learning_rate": 4.1356686569674335e-08, |
|
"logits/chosen": 1.4700018167495728, |
|
"logits/rejected": 1.7809115648269653, |
|
"logps/chosen": -27.53030014038086, |
|
"logps/rejected": -43.54850387573242, |
|
"loss": 0.1642, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.1137082502245903, |
|
"rewards/margins": 2.9248647689819336, |
|
"rewards/rejected": -2.811156749725342, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 3.3164556962025316, |
|
"grad_norm": 19.797305830292228, |
|
"learning_rate": 3.831895019292897e-08, |
|
"logits/chosen": 1.8914071321487427, |
|
"logits/rejected": 1.8628309965133667, |
|
"logps/chosen": -25.410337448120117, |
|
"logps/rejected": -42.06156921386719, |
|
"loss": 0.1525, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.044954825192689896, |
|
"rewards/margins": 2.8561882972717285, |
|
"rewards/rejected": -2.9011433124542236, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 3.3417721518987342, |
|
"grad_norm": 18.851413557062596, |
|
"learning_rate": 3.538780159953347e-08, |
|
"logits/chosen": 2.0251123905181885, |
|
"logits/rejected": 2.210965156555176, |
|
"logps/chosen": -20.657623291015625, |
|
"logps/rejected": -45.97637939453125, |
|
"loss": 0.1488, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.21129505336284637, |
|
"rewards/margins": 3.6020712852478027, |
|
"rewards/rejected": -3.3907763957977295, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 3.367088607594937, |
|
"grad_norm": 18.383406639630245, |
|
"learning_rate": 3.256471671107616e-08, |
|
"logits/chosen": 1.5468469858169556, |
|
"logits/rejected": 1.8942471742630005, |
|
"logps/chosen": -34.759246826171875, |
|
"logps/rejected": -54.94669723510742, |
|
"loss": 0.161, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.08364612609148026, |
|
"rewards/margins": 3.432290554046631, |
|
"rewards/rejected": -3.348644495010376, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 3.392405063291139, |
|
"grad_norm": 18.442661691122066, |
|
"learning_rate": 2.98511170358155e-08, |
|
"logits/chosen": 1.9410444498062134, |
|
"logits/rejected": 1.9713021516799927, |
|
"logps/chosen": -33.468505859375, |
|
"logps/rejected": -40.28277587890625, |
|
"loss": 0.1517, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.1791667342185974, |
|
"rewards/margins": 2.8069636821746826, |
|
"rewards/rejected": -2.986130475997925, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 3.4177215189873418, |
|
"grad_norm": 19.123016466489233, |
|
"learning_rate": 2.724836895290805e-08, |
|
"logits/chosen": 2.172494649887085, |
|
"logits/rejected": 2.3771188259124756, |
|
"logps/chosen": -18.799209594726562, |
|
"logps/rejected": -27.414329528808594, |
|
"loss": 0.1574, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.05184905230998993, |
|
"rewards/margins": 1.7008121013641357, |
|
"rewards/rejected": -1.648963212966919, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 3.4430379746835444, |
|
"grad_norm": 19.389448097902036, |
|
"learning_rate": 2.475778302439524e-08, |
|
"logits/chosen": 1.5435059070587158, |
|
"logits/rejected": 2.0342299938201904, |
|
"logps/chosen": -23.09898567199707, |
|
"logps/rejected": -41.35944366455078, |
|
"loss": 0.1673, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.01025979220867157, |
|
"rewards/margins": 3.0842089653015137, |
|
"rewards/rejected": -3.094468832015991, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 3.4683544303797467, |
|
"grad_norm": 18.168150870329598, |
|
"learning_rate": 2.2380613335296033e-08, |
|
"logits/chosen": 1.5661985874176025, |
|
"logits/rejected": 1.6321345567703247, |
|
"logps/chosen": -21.557811737060547, |
|
"logps/rejected": -37.444087982177734, |
|
"loss": 0.1457, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.32843151688575745, |
|
"rewards/margins": 3.217196464538574, |
|
"rewards/rejected": -2.888765335083008, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 3.4936708860759493, |
|
"grad_norm": 18.10048373008008, |
|
"learning_rate": 2.0118056862137354e-08, |
|
"logits/chosen": 2.2138497829437256, |
|
"logits/rejected": 2.3827362060546875, |
|
"logps/chosen": -25.050493240356445, |
|
"logps/rejected": -40.5790901184082, |
|
"loss": 0.1574, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.05445308983325958, |
|
"rewards/margins": 2.6866414546966553, |
|
"rewards/rejected": -2.7410943508148193, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 3.518987341772152, |
|
"grad_norm": 17.844608513376855, |
|
"learning_rate": 1.797125287024029e-08, |
|
"logits/chosen": 2.252528667449951, |
|
"logits/rejected": 2.517775297164917, |
|
"logps/chosen": -25.718944549560547, |
|
"logps/rejected": -40.703636169433594, |
|
"loss": 0.1633, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.23676270246505737, |
|
"rewards/margins": 2.2277328968048096, |
|
"rewards/rejected": -2.4644956588745117, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 3.5443037974683547, |
|
"grad_norm": 19.82721673855706, |
|
"learning_rate": 1.5941282340065697e-08, |
|
"logits/chosen": 2.1378254890441895, |
|
"logits/rejected": 2.3837673664093018, |
|
"logps/chosen": -21.759735107421875, |
|
"logps/rejected": -35.555789947509766, |
|
"loss": 0.1725, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.32154813408851624, |
|
"rewards/margins": 2.5137619972229004, |
|
"rewards/rejected": -2.192214012145996, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 3.569620253164557, |
|
"grad_norm": 19.687984330411137, |
|
"learning_rate": 1.4029167422908105e-08, |
|
"logits/chosen": 1.8691318035125732, |
|
"logits/rejected": 1.951183557510376, |
|
"logps/chosen": -22.21922492980957, |
|
"logps/rejected": -38.082176208496094, |
|
"loss": 0.1587, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.12926092743873596, |
|
"rewards/margins": 3.5831055641174316, |
|
"rewards/rejected": -3.4538450241088867, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 3.5949367088607596, |
|
"grad_norm": 19.210992875859024, |
|
"learning_rate": 1.2235870926211616e-08, |
|
"logits/chosen": 1.878167748451233, |
|
"logits/rejected": 2.1185381412506104, |
|
"logps/chosen": -25.917736053466797, |
|
"logps/rejected": -45.47472381591797, |
|
"loss": 0.1359, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.0160946324467659, |
|
"rewards/margins": 3.645671844482422, |
|
"rewards/rejected": -3.629577398300171, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 3.620253164556962, |
|
"grad_norm": 20.116883252259765, |
|
"learning_rate": 1.0562295828767387e-08, |
|
"logits/chosen": 1.2207794189453125, |
|
"logits/rejected": 1.1344801187515259, |
|
"logps/chosen": -23.36668586730957, |
|
"logps/rejected": -31.341829299926758, |
|
"loss": 0.1841, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.16667915880680084, |
|
"rewards/margins": 1.5721771717071533, |
|
"rewards/rejected": -1.7388561964035034, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 3.6455696202531644, |
|
"grad_norm": 19.720241233310634, |
|
"learning_rate": 9.009284826036689e-09, |
|
"logits/chosen": 2.4453952312469482, |
|
"logits/rejected": 2.662018299102783, |
|
"logps/chosen": -22.387048721313477, |
|
"logps/rejected": -36.92119216918945, |
|
"loss": 0.1617, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.031851306557655334, |
|
"rewards/margins": 2.5965640544891357, |
|
"rewards/rejected": -2.5647125244140625, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 3.670886075949367, |
|
"grad_norm": 17.86661138635714, |
|
"learning_rate": 7.577619905828281e-09, |
|
"logits/chosen": 2.745617628097534, |
|
"logits/rejected": 2.844001531600952, |
|
"logps/chosen": -26.597410202026367, |
|
"logps/rejected": -36.82072448730469, |
|
"loss": 0.142, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.17600667476654053, |
|
"rewards/margins": 2.553907632827759, |
|
"rewards/rejected": -2.7299141883850098, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 3.6962025316455698, |
|
"grad_norm": 16.0477696571449, |
|
"learning_rate": 6.268021954544095e-09, |
|
"logits/chosen": 2.1746973991394043, |
|
"logits/rejected": 2.2613565921783447, |
|
"logps/chosen": -32.3970947265625, |
|
"logps/rejected": -40.7056884765625, |
|
"loss": 0.1372, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.09012872725725174, |
|
"rewards/margins": 4.017242431640625, |
|
"rewards/rejected": -4.1073713302612305, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 3.721518987341772, |
|
"grad_norm": 18.112204362761403, |
|
"learning_rate": 5.08115039419113e-09, |
|
"logits/chosen": 2.2588939666748047, |
|
"logits/rejected": 2.217837333679199, |
|
"logps/chosen": -23.21999740600586, |
|
"logps/rejected": -38.22819519042969, |
|
"loss": 0.1574, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.07647465169429779, |
|
"rewards/margins": 3.384047031402588, |
|
"rewards/rejected": -3.460521697998047, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 3.7468354430379747, |
|
"grad_norm": 19.146595468250382, |
|
"learning_rate": 4.0176028503425826e-09, |
|
"logits/chosen": 2.2898073196411133, |
|
"logits/rejected": 2.612645149230957, |
|
"logps/chosen": -21.40774917602539, |
|
"logps/rejected": -39.644142150878906, |
|
"loss": 0.1575, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.14267057180404663, |
|
"rewards/margins": 2.3787121772766113, |
|
"rewards/rejected": -2.23604154586792, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 3.7721518987341773, |
|
"grad_norm": 16.066181981705437, |
|
"learning_rate": 3.077914851215585e-09, |
|
"logits/chosen": 2.0213255882263184, |
|
"logits/rejected": 2.22232723236084, |
|
"logps/chosen": -20.719486236572266, |
|
"logps/rejected": -33.82309341430664, |
|
"loss": 0.1448, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.04364713281393051, |
|
"rewards/margins": 2.438537836074829, |
|
"rewards/rejected": -2.482184886932373, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 3.7974683544303796, |
|
"grad_norm": 20.168906914662397, |
|
"learning_rate": 2.2625595580163247e-09, |
|
"logits/chosen": 1.8664175271987915, |
|
"logits/rejected": 2.135633945465088, |
|
"logps/chosen": -24.14510726928711, |
|
"logps/rejected": -39.59516906738281, |
|
"loss": 0.1844, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -0.004626031965017319, |
|
"rewards/margins": 2.253840446472168, |
|
"rewards/rejected": -2.2584664821624756, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 3.8227848101265822, |
|
"grad_norm": 21.962898348700495, |
|
"learning_rate": 1.5719475266893489e-09, |
|
"logits/chosen": 1.7901225090026855, |
|
"logits/rejected": 1.9218878746032715, |
|
"logps/chosen": -22.082956314086914, |
|
"logps/rejected": -43.000667572021484, |
|
"loss": 0.192, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.07910831272602081, |
|
"rewards/margins": 2.6802215576171875, |
|
"rewards/rejected": -2.6011130809783936, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 3.848101265822785, |
|
"grad_norm": 16.511033653339577, |
|
"learning_rate": 1.0064265011902328e-09, |
|
"logits/chosen": 2.4072799682617188, |
|
"logits/rejected": 2.4377200603485107, |
|
"logps/chosen": -25.803665161132812, |
|
"logps/rejected": -40.92607879638672, |
|
"loss": 0.1549, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.035745151340961456, |
|
"rewards/margins": 3.1956663131713867, |
|
"rewards/rejected": -3.159921169281006, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 3.8734177215189876, |
|
"grad_norm": 18.636452488465395, |
|
"learning_rate": 5.662812383859794e-10, |
|
"logits/chosen": 2.361091136932373, |
|
"logits/rejected": 2.3107473850250244, |
|
"logps/chosen": -27.91613006591797, |
|
"logps/rejected": -39.84597396850586, |
|
"loss": 0.1544, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.1522400677204132, |
|
"rewards/margins": 3.036912441253662, |
|
"rewards/rejected": -3.189152240753174, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 3.8987341772151898, |
|
"grad_norm": 17.763028400462222, |
|
"learning_rate": 2.5173336467135263e-10, |
|
"logits/chosen": 2.0824742317199707, |
|
"logits/rejected": 2.3250155448913574, |
|
"logps/chosen": -23.416114807128906, |
|
"logps/rejected": -38.985511779785156, |
|
"loss": 0.1542, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.06766586005687714, |
|
"rewards/margins": 2.1996846199035645, |
|
"rewards/rejected": -2.267350673675537, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 3.9240506329113924, |
|
"grad_norm": 19.457837238018815, |
|
"learning_rate": 6.294126437336733e-11, |
|
"logits/chosen": 2.00872540473938, |
|
"logits/rejected": 2.4769511222839355, |
|
"logps/chosen": -24.827199935913086, |
|
"logps/rejected": -44.62668991088867, |
|
"loss": 0.1614, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.19776426255702972, |
|
"rewards/margins": 2.720571756362915, |
|
"rewards/rejected": -2.9183361530303955, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 3.9493670886075947, |
|
"grad_norm": 19.565085246341297, |
|
"learning_rate": 0.0, |
|
"logits/chosen": 1.8402495384216309, |
|
"logits/rejected": 2.033839464187622, |
|
"logps/chosen": -28.689044952392578, |
|
"logps/rejected": -46.74669647216797, |
|
"loss": 0.1636, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.04416623339056969, |
|
"rewards/margins": 3.22959041595459, |
|
"rewards/rejected": -3.185424327850342, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 3.9493670886075947, |
|
"step": 156, |
|
"total_flos": 0.0, |
|
"train_loss": 0.3210896818110576, |
|
"train_runtime": 45702.9058, |
|
"train_samples_per_second": 0.442, |
|
"train_steps_per_second": 0.003 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 156, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 200, |
|
"total_flos": 0.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|