|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.998691442030882, |
|
"eval_steps": 500, |
|
"global_step": 477, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.002093692750588851, |
|
"grad_norm": 6813.720901215927, |
|
"learning_rate": 1.0416666666666666e-08, |
|
"logits/chosen": 5002.53564453125, |
|
"logits/rejected": 4591.72021484375, |
|
"logps/chosen": -265.6396789550781, |
|
"logps/rejected": -206.22401428222656, |
|
"loss": 503.0168, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.0, |
|
"rewards/margins": 0.0, |
|
"rewards/rejected": 0.0, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.02093692750588851, |
|
"grad_norm": 6935.582411893378, |
|
"learning_rate": 1.0416666666666667e-07, |
|
"logits/chosen": 5977.00244140625, |
|
"logits/rejected": 5015.09765625, |
|
"logps/chosen": -292.2739562988281, |
|
"logps/rejected": -259.0675048828125, |
|
"loss": 549.3314, |
|
"rewards/accuracies": 0.4583333432674408, |
|
"rewards/chosen": 0.0028684597928076982, |
|
"rewards/margins": 0.0003068610094487667, |
|
"rewards/rejected": 0.0025615987833589315, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.04187385501177702, |
|
"grad_norm": 3127.701327967131, |
|
"learning_rate": 2.0833333333333333e-07, |
|
"logits/chosen": 5864.72802734375, |
|
"logits/rejected": 4844.66650390625, |
|
"logps/chosen": -283.6185302734375, |
|
"logps/rejected": -243.01513671875, |
|
"loss": 523.6729, |
|
"rewards/accuracies": 0.5562499761581421, |
|
"rewards/chosen": 0.06974941492080688, |
|
"rewards/margins": 0.002566719427704811, |
|
"rewards/rejected": 0.06718270480632782, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.06281078251766553, |
|
"grad_norm": 2074.6624019206756, |
|
"learning_rate": 3.1249999999999997e-07, |
|
"logits/chosen": 5557.00439453125, |
|
"logits/rejected": 4840.8193359375, |
|
"logps/chosen": -256.04144287109375, |
|
"logps/rejected": -223.24319458007812, |
|
"loss": 494.8969, |
|
"rewards/accuracies": 0.5687500238418579, |
|
"rewards/chosen": 0.16196031868457794, |
|
"rewards/margins": 0.006370093673467636, |
|
"rewards/rejected": 0.1555902361869812, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.08374771002355404, |
|
"grad_norm": 1576.9313876726794, |
|
"learning_rate": 4.1666666666666667e-07, |
|
"logits/chosen": 5559.41650390625, |
|
"logits/rejected": 4830.68994140625, |
|
"logps/chosen": -251.85604858398438, |
|
"logps/rejected": -236.69235229492188, |
|
"loss": 496.3616, |
|
"rewards/accuracies": 0.518750011920929, |
|
"rewards/chosen": 0.22101028263568878, |
|
"rewards/margins": 0.0033943026792258024, |
|
"rewards/rejected": 0.217616006731987, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.10468463752944256, |
|
"grad_norm": 1640.8995064962066, |
|
"learning_rate": 4.999731868769026e-07, |
|
"logits/chosen": 5797.015625, |
|
"logits/rejected": 5035.59326171875, |
|
"logps/chosen": -253.31631469726562, |
|
"logps/rejected": -242.33651733398438, |
|
"loss": 492.3395, |
|
"rewards/accuracies": 0.5218750238418579, |
|
"rewards/chosen": 0.2530141770839691, |
|
"rewards/margins": 0.0011868119472637773, |
|
"rewards/rejected": 0.2518273890018463, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.12562156503533106, |
|
"grad_norm": 1480.537194221153, |
|
"learning_rate": 4.990353313429303e-07, |
|
"logits/chosen": 5540.62158203125, |
|
"logits/rejected": 4727.97607421875, |
|
"logps/chosen": -248.41848754882812, |
|
"logps/rejected": -221.98336791992188, |
|
"loss": 466.3484, |
|
"rewards/accuracies": 0.5218750238418579, |
|
"rewards/chosen": 0.30004000663757324, |
|
"rewards/margins": 0.005084425210952759, |
|
"rewards/rejected": 0.29495561122894287, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.14655849254121958, |
|
"grad_norm": 1439.2312712606115, |
|
"learning_rate": 4.967625656594781e-07, |
|
"logits/chosen": 5331.10791015625, |
|
"logits/rejected": 4925.798828125, |
|
"logps/chosen": -247.7162628173828, |
|
"logps/rejected": -237.12353515625, |
|
"loss": 465.379, |
|
"rewards/accuracies": 0.518750011920929, |
|
"rewards/chosen": 0.30591654777526855, |
|
"rewards/margins": 0.009119048714637756, |
|
"rewards/rejected": 0.2967974543571472, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.16749542004710807, |
|
"grad_norm": 1420.7895768869655, |
|
"learning_rate": 4.93167072587771e-07, |
|
"logits/chosen": 5414.23828125, |
|
"logits/rejected": 4648.93115234375, |
|
"logps/chosen": -242.97555541992188, |
|
"logps/rejected": -222.2198486328125, |
|
"loss": 464.9691, |
|
"rewards/accuracies": 0.5531250238418579, |
|
"rewards/chosen": 0.33111369609832764, |
|
"rewards/margins": 0.022587263956665993, |
|
"rewards/rejected": 0.3085264265537262, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.1884323475529966, |
|
"grad_norm": 1456.2100554164706, |
|
"learning_rate": 4.882681251368548e-07, |
|
"logits/chosen": 5535.8984375, |
|
"logits/rejected": 4687.62255859375, |
|
"logps/chosen": -242.0615234375, |
|
"logps/rejected": -220.8953399658203, |
|
"loss": 465.3252, |
|
"rewards/accuracies": 0.5406249761581421, |
|
"rewards/chosen": 0.35646066069602966, |
|
"rewards/margins": 0.015677783638238907, |
|
"rewards/rejected": 0.34078291058540344, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.2093692750588851, |
|
"grad_norm": 1536.760185739677, |
|
"learning_rate": 4.820919832540181e-07, |
|
"logits/chosen": 5591.662109375, |
|
"logits/rejected": 5172.16259765625, |
|
"logps/chosen": -243.97738647460938, |
|
"logps/rejected": -236.50595092773438, |
|
"loss": 460.8633, |
|
"rewards/accuracies": 0.48750001192092896, |
|
"rewards/chosen": 0.3589131832122803, |
|
"rewards/margins": -0.021181438118219376, |
|
"rewards/rejected": 0.38009461760520935, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.23030620256477363, |
|
"grad_norm": 1408.5395404483565, |
|
"learning_rate": 4.7467175306295647e-07, |
|
"logits/chosen": 5752.2939453125, |
|
"logits/rejected": 4664.9677734375, |
|
"logps/chosen": -236.39962768554688, |
|
"logps/rejected": -224.271484375, |
|
"loss": 463.9909, |
|
"rewards/accuracies": 0.5562499761581421, |
|
"rewards/chosen": 0.382229745388031, |
|
"rewards/margins": 0.03622853755950928, |
|
"rewards/rejected": 0.34600120782852173, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.2512431300706621, |
|
"grad_norm": 1430.8052441244079, |
|
"learning_rate": 4.6604720940421207e-07, |
|
"logits/chosen": 5345.5830078125, |
|
"logits/rejected": 4974.67724609375, |
|
"logps/chosen": -230.8021240234375, |
|
"logps/rejected": -229.9935302734375, |
|
"loss": 472.1473, |
|
"rewards/accuracies": 0.503125011920929, |
|
"rewards/chosen": 0.36376887559890747, |
|
"rewards/margins": -0.005053712520748377, |
|
"rewards/rejected": 0.36882254481315613, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.2721800575765506, |
|
"grad_norm": 1613.4321490419964, |
|
"learning_rate": 4.5626458262912735e-07, |
|
"logits/chosen": 5406.8876953125, |
|
"logits/rejected": 4819.064453125, |
|
"logps/chosen": -233.12362670898438, |
|
"logps/rejected": -232.2109375, |
|
"loss": 465.5601, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": 0.35393673181533813, |
|
"rewards/margins": -0.014040246605873108, |
|
"rewards/rejected": 0.36797699332237244, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.29311698508243916, |
|
"grad_norm": 1397.7266878069634, |
|
"learning_rate": 4.453763107901675e-07, |
|
"logits/chosen": 5204.806640625, |
|
"logits/rejected": 4634.70947265625, |
|
"logps/chosen": -235.88473510742188, |
|
"logps/rejected": -216.02548217773438, |
|
"loss": 475.0084, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": 0.36020082235336304, |
|
"rewards/margins": 0.009127211757004261, |
|
"rewards/rejected": 0.35107359290122986, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.31405391258832765, |
|
"grad_norm": 1771.9334786366776, |
|
"learning_rate": 4.3344075855595097e-07, |
|
"logits/chosen": 5751.51171875, |
|
"logits/rejected": 4749.1025390625, |
|
"logps/chosen": -253.08169555664062, |
|
"logps/rejected": -215.91110229492188, |
|
"loss": 453.9507, |
|
"rewards/accuracies": 0.5531250238418579, |
|
"rewards/chosen": 0.3603329658508301, |
|
"rewards/margins": 0.004536592401564121, |
|
"rewards/rejected": 0.35579636693000793, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.33499084009421615, |
|
"grad_norm": 1436.0408028073434, |
|
"learning_rate": 4.2052190435769554e-07, |
|
"logits/chosen": 5464.77099609375, |
|
"logits/rejected": 4553.0888671875, |
|
"logps/chosen": -241.3671112060547, |
|
"logps/rejected": -218.6112518310547, |
|
"loss": 470.9686, |
|
"rewards/accuracies": 0.550000011920929, |
|
"rewards/chosen": 0.4322701394557953, |
|
"rewards/margins": 0.054599881172180176, |
|
"rewards/rejected": 0.3776702284812927, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.3559277676001047, |
|
"grad_norm": 1452.849884308315, |
|
"learning_rate": 4.0668899744407567e-07, |
|
"logits/chosen": 5712.3564453125, |
|
"logits/rejected": 4878.48193359375, |
|
"logps/chosen": -248.1728515625, |
|
"logps/rejected": -206.45004272460938, |
|
"loss": 457.2006, |
|
"rewards/accuracies": 0.550000011920929, |
|
"rewards/chosen": 0.34142884612083435, |
|
"rewards/margins": -0.0005211926763877273, |
|
"rewards/rejected": 0.34195005893707275, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.3768646951059932, |
|
"grad_norm": 1345.993393031555, |
|
"learning_rate": 3.920161866827889e-07, |
|
"logits/chosen": 5387.048828125, |
|
"logits/rejected": 4639.03271484375, |
|
"logps/chosen": -250.1925811767578, |
|
"logps/rejected": -211.57275390625, |
|
"loss": 463.2538, |
|
"rewards/accuracies": 0.5218750238418579, |
|
"rewards/chosen": 0.37102702260017395, |
|
"rewards/margins": -0.013053232803940773, |
|
"rewards/rejected": 0.3840802311897278, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.39780162261188173, |
|
"grad_norm": 1364.64342210683, |
|
"learning_rate": 3.765821230985757e-07, |
|
"logits/chosen": 5516.2822265625, |
|
"logits/rejected": 4661.30517578125, |
|
"logps/chosen": -231.8093719482422, |
|
"logps/rejected": -209.94967651367188, |
|
"loss": 460.1928, |
|
"rewards/accuracies": 0.48750001192092896, |
|
"rewards/chosen": 0.37942126393318176, |
|
"rewards/margins": 8.549987978767604e-05, |
|
"rewards/rejected": 0.3793357312679291, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.4187385501177702, |
|
"grad_norm": 1501.9460217734675, |
|
"learning_rate": 3.604695382782159e-07, |
|
"logits/chosen": 5149.3212890625, |
|
"logits/rejected": 4487.83251953125, |
|
"logps/chosen": -217.900634765625, |
|
"logps/rejected": -206.22817993164062, |
|
"loss": 462.1984, |
|
"rewards/accuracies": 0.5406249761581421, |
|
"rewards/chosen": 0.3645021319389343, |
|
"rewards/margins": -0.0018228956032544374, |
|
"rewards/rejected": 0.3663250207901001, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.4396754776236587, |
|
"grad_norm": 1429.3016977538932, |
|
"learning_rate": 3.4376480090239047e-07, |
|
"logits/chosen": 5780.9111328125, |
|
"logits/rejected": 4931.05810546875, |
|
"logps/chosen": -248.72396850585938, |
|
"logps/rejected": -224.09872436523438, |
|
"loss": 466.6646, |
|
"rewards/accuracies": 0.5375000238418579, |
|
"rewards/chosen": 0.38043394684791565, |
|
"rewards/margins": 0.004385782405734062, |
|
"rewards/rejected": 0.37604817748069763, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.46061240512954726, |
|
"grad_norm": 1443.7077268756489, |
|
"learning_rate": 3.265574537815398e-07, |
|
"logits/chosen": 5190.0830078125, |
|
"logits/rejected": 4833.0302734375, |
|
"logps/chosen": -234.4996795654297, |
|
"logps/rejected": -221.47488403320312, |
|
"loss": 474.1076, |
|
"rewards/accuracies": 0.5062500238418579, |
|
"rewards/chosen": 0.36616024374961853, |
|
"rewards/margins": -0.022525813430547714, |
|
"rewards/rejected": 0.38868606090545654, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.48154933263543576, |
|
"grad_norm": 1524.9948933221324, |
|
"learning_rate": 3.0893973387735683e-07, |
|
"logits/chosen": 5389.8935546875, |
|
"logits/rejected": 4572.7626953125, |
|
"logps/chosen": -228.4185028076172, |
|
"logps/rejected": -209.15811157226562, |
|
"loss": 446.1821, |
|
"rewards/accuracies": 0.5531250238418579, |
|
"rewards/chosen": 0.39850276708602905, |
|
"rewards/margins": 0.035766832530498505, |
|
"rewards/rejected": 0.36273595690727234, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.5024862601413242, |
|
"grad_norm": 1425.7617156287813, |
|
"learning_rate": 2.910060778827554e-07, |
|
"logits/chosen": 5539.8349609375, |
|
"logits/rejected": 4716.81201171875, |
|
"logps/chosen": -227.1032257080078, |
|
"logps/rejected": -210.68612670898438, |
|
"loss": 466.6096, |
|
"rewards/accuracies": 0.534375011920929, |
|
"rewards/chosen": 0.38473668694496155, |
|
"rewards/margins": -0.006211507134139538, |
|
"rewards/rejected": 0.3909481465816498, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.5234231876472127, |
|
"grad_norm": 1407.421320970437, |
|
"learning_rate": 2.7285261601056697e-07, |
|
"logits/chosen": 5670.39453125, |
|
"logits/rejected": 4789.2001953125, |
|
"logps/chosen": -244.33895874023438, |
|
"logps/rejected": -223.86453247070312, |
|
"loss": 459.7925, |
|
"rewards/accuracies": 0.5531250238418579, |
|
"rewards/chosen": 0.4160972535610199, |
|
"rewards/margins": 0.03399023413658142, |
|
"rewards/rejected": 0.3821069598197937, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.5443601151531012, |
|
"grad_norm": 1368.2023715670355, |
|
"learning_rate": 2.5457665670441937e-07, |
|
"logits/chosen": 5489.0439453125, |
|
"logits/rejected": 4975.576171875, |
|
"logps/chosen": -245.1668243408203, |
|
"logps/rejected": -221.49661254882812, |
|
"loss": 458.9599, |
|
"rewards/accuracies": 0.5062500238418579, |
|
"rewards/chosen": 0.4090227484703064, |
|
"rewards/margins": 0.013166209682822227, |
|
"rewards/rejected": 0.3958565592765808, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.5652970426589898, |
|
"grad_norm": 1496.3498583187165, |
|
"learning_rate": 2.3627616503391812e-07, |
|
"logits/chosen": 5568.6240234375, |
|
"logits/rejected": 5110.2138671875, |
|
"logps/chosen": -233.6811065673828, |
|
"logps/rejected": -220.8893585205078, |
|
"loss": 447.9319, |
|
"rewards/accuracies": 0.5562499761581421, |
|
"rewards/chosen": 0.4093276560306549, |
|
"rewards/margins": 0.014256368391215801, |
|
"rewards/rejected": 0.3950712978839874, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.5862339701648783, |
|
"grad_norm": 1386.5146227495984, |
|
"learning_rate": 2.1804923757009882e-07, |
|
"logits/chosen": 5578.41455078125, |
|
"logits/rejected": 4647.2431640625, |
|
"logps/chosen": -241.32284545898438, |
|
"logps/rejected": -207.7650146484375, |
|
"loss": 448.7999, |
|
"rewards/accuracies": 0.534375011920929, |
|
"rewards/chosen": 0.3940281867980957, |
|
"rewards/margins": 0.003837400348857045, |
|
"rewards/rejected": 0.3901907801628113, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.6071708976707668, |
|
"grad_norm": 1565.300342790805, |
|
"learning_rate": 1.9999357655598891e-07, |
|
"logits/chosen": 5835.33984375, |
|
"logits/rejected": 5479.173828125, |
|
"logps/chosen": -249.04843139648438, |
|
"logps/rejected": -245.552734375, |
|
"loss": 468.733, |
|
"rewards/accuracies": 0.5062500238418579, |
|
"rewards/chosen": 0.4238561689853668, |
|
"rewards/margins": -0.09983213245868683, |
|
"rewards/rejected": 0.5236883163452148, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.6281078251766553, |
|
"grad_norm": 1477.060718071622, |
|
"learning_rate": 1.8220596619089573e-07, |
|
"logits/chosen": 5350.3173828125, |
|
"logits/rejected": 5039.15966796875, |
|
"logps/chosen": -233.1553192138672, |
|
"logps/rejected": -227.7581024169922, |
|
"loss": 463.6427, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.3819466233253479, |
|
"rewards/margins": -0.04966121166944504, |
|
"rewards/rejected": 0.43160781264305115, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.6490447526825438, |
|
"grad_norm": 1467.0428814727877, |
|
"learning_rate": 1.647817538357072e-07, |
|
"logits/chosen": 5974.5576171875, |
|
"logits/rejected": 5248.5712890625, |
|
"logps/chosen": -246.29739379882812, |
|
"logps/rejected": -230.0133819580078, |
|
"loss": 451.9249, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.4453219473361969, |
|
"rewards/margins": 0.049888670444488525, |
|
"rewards/rejected": 0.395433247089386, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.6699816801884323, |
|
"grad_norm": 1539.7938960746517, |
|
"learning_rate": 1.478143389201113e-07, |
|
"logits/chosen": 5526.9794921875, |
|
"logits/rejected": 4800.4169921875, |
|
"logps/chosen": -232.0591583251953, |
|
"logps/rejected": -218.22909545898438, |
|
"loss": 457.9831, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.3980428874492645, |
|
"rewards/margins": -0.017790189012885094, |
|
"rewards/rejected": 0.41583308577537537, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.6909186076943209, |
|
"grad_norm": 1416.4774770653923, |
|
"learning_rate": 1.3139467229135998e-07, |
|
"logits/chosen": 5347.95751953125, |
|
"logits/rejected": 4977.25830078125, |
|
"logps/chosen": -237.83120727539062, |
|
"logps/rejected": -223.06503295898438, |
|
"loss": 456.2857, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": 0.3895305097103119, |
|
"rewards/margins": -0.023113062605261803, |
|
"rewards/rejected": 0.41264358162879944, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.7118555352002094, |
|
"grad_norm": 1354.3420498241383, |
|
"learning_rate": 1.1561076868822755e-07, |
|
"logits/chosen": 5184.0341796875, |
|
"logits/rejected": 4512.36376953125, |
|
"logps/chosen": -215.9802703857422, |
|
"logps/rejected": -202.09982299804688, |
|
"loss": 441.2616, |
|
"rewards/accuracies": 0.5562499761581421, |
|
"rewards/chosen": 0.4158664345741272, |
|
"rewards/margins": 0.02483862265944481, |
|
"rewards/rejected": 0.3910277783870697, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.7327924627060979, |
|
"grad_norm": 1399.0816486347255, |
|
"learning_rate": 1.0054723495346482e-07, |
|
"logits/chosen": 5854.7001953125, |
|
"logits/rejected": 4766.1748046875, |
|
"logps/chosen": -257.1776428222656, |
|
"logps/rejected": -228.104736328125, |
|
"loss": 471.4627, |
|
"rewards/accuracies": 0.534375011920929, |
|
"rewards/chosen": 0.4373684823513031, |
|
"rewards/margins": 0.03503155708312988, |
|
"rewards/rejected": 0.402336984872818, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.7537293902119864, |
|
"grad_norm": 1352.127266112888, |
|
"learning_rate": 8.628481651367875e-08, |
|
"logits/chosen": 5555.2060546875, |
|
"logits/rejected": 4852.7666015625, |
|
"logps/chosen": -235.2597198486328, |
|
"logps/rejected": -216.91415405273438, |
|
"loss": 443.8227, |
|
"rewards/accuracies": 0.5249999761581421, |
|
"rewards/chosen": 0.42463842034339905, |
|
"rewards/margins": 0.01125773973762989, |
|
"rewards/rejected": 0.4133806824684143, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.7746663177178749, |
|
"grad_norm": 1283.6956906144683, |
|
"learning_rate": 7.289996455765748e-08, |
|
"logits/chosen": 5428.78271484375, |
|
"logits/rejected": 4644.25390625, |
|
"logps/chosen": -251.05245971679688, |
|
"logps/rejected": -221.3125, |
|
"loss": 453.7341, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": 0.3774058520793915, |
|
"rewards/margins": 0.02088645100593567, |
|
"rewards/rejected": 0.3565194010734558, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.7956032452237635, |
|
"grad_norm": 1389.3936014370584, |
|
"learning_rate": 6.046442623320145e-08, |
|
"logits/chosen": 5676.46826171875, |
|
"logits/rejected": 4961.1416015625, |
|
"logps/chosen": -245.52609252929688, |
|
"logps/rejected": -210.39529418945312, |
|
"loss": 472.2963, |
|
"rewards/accuracies": 0.5249999761581421, |
|
"rewards/chosen": 0.39175355434417725, |
|
"rewards/margins": -0.012869933620095253, |
|
"rewards/rejected": 0.40462350845336914, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.816540172729652, |
|
"grad_norm": 1904.7699931128577, |
|
"learning_rate": 4.904486005914027e-08, |
|
"logits/chosen": 5950.8603515625, |
|
"logits/rejected": 5257.07080078125, |
|
"logps/chosen": -246.93521118164062, |
|
"logps/rejected": -232.27798461914062, |
|
"loss": 459.8493, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": 0.44814223051071167, |
|
"rewards/margins": 0.023155368864536285, |
|
"rewards/rejected": 0.42498689889907837, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.8374771002355405, |
|
"grad_norm": 1314.6946141188957, |
|
"learning_rate": 3.8702478614051345e-08, |
|
"logits/chosen": 5522.93212890625, |
|
"logits/rejected": 4914.50146484375, |
|
"logps/chosen": -246.7040557861328, |
|
"logps/rejected": -212.8903045654297, |
|
"loss": 463.632, |
|
"rewards/accuracies": 0.528124988079071, |
|
"rewards/chosen": 0.4420008063316345, |
|
"rewards/margins": 0.01232604868710041, |
|
"rewards/rejected": 0.42967477440834045, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.8584140277414289, |
|
"grad_norm": 1386.6765646261215, |
|
"learning_rate": 2.9492720416985e-08, |
|
"logits/chosen": 5542.53125, |
|
"logits/rejected": 4905.96435546875, |
|
"logps/chosen": -233.2512664794922, |
|
"logps/rejected": -216.3258819580078, |
|
"loss": 459.2343, |
|
"rewards/accuracies": 0.550000011920929, |
|
"rewards/chosen": 0.4861605167388916, |
|
"rewards/margins": -0.16168272495269775, |
|
"rewards/rejected": 0.6478432416915894, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.8793509552473174, |
|
"grad_norm": 1421.0973915303414, |
|
"learning_rate": 2.1464952759020856e-08, |
|
"logits/chosen": 5875.4296875, |
|
"logits/rejected": 5138.45947265625, |
|
"logps/chosen": -249.4132843017578, |
|
"logps/rejected": -230.8553009033203, |
|
"loss": 457.3266, |
|
"rewards/accuracies": 0.48124998807907104, |
|
"rewards/chosen": 0.42565909028053284, |
|
"rewards/margins": 0.009328785352408886, |
|
"rewards/rejected": 0.4163302779197693, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.9002878827532059, |
|
"grad_norm": 2530.194729647411, |
|
"learning_rate": 1.4662207078575684e-08, |
|
"logits/chosen": 5782.8046875, |
|
"logits/rejected": 4674.46826171875, |
|
"logps/chosen": -244.35745239257812, |
|
"logps/rejected": -218.1392364501953, |
|
"loss": 449.309, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": 0.4324522614479065, |
|
"rewards/margins": 0.03925667330622673, |
|
"rewards/rejected": 0.39319556951522827, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.9212248102590945, |
|
"grad_norm": 1331.0653218757734, |
|
"learning_rate": 9.12094829893642e-09, |
|
"logits/chosen": 5607.314453125, |
|
"logits/rejected": 5351.4794921875, |
|
"logps/chosen": -240.36508178710938, |
|
"logps/rejected": -235.7848358154297, |
|
"loss": 447.1441, |
|
"rewards/accuracies": 0.47187501192092896, |
|
"rewards/chosen": 0.3817844092845917, |
|
"rewards/margins": -0.020826738327741623, |
|
"rewards/rejected": 0.4026111960411072, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.942161737764983, |
|
"grad_norm": 1400.3246977795131, |
|
"learning_rate": 4.8708793644441086e-09, |
|
"logits/chosen": 5760.21923828125, |
|
"logits/rejected": 4969.435546875, |
|
"logps/chosen": -248.47109985351562, |
|
"logps/rejected": -217.5637969970703, |
|
"loss": 450.0296, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.41864126920700073, |
|
"rewards/margins": -0.020225150510668755, |
|
"rewards/rejected": 0.43886643648147583, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.9630986652708715, |
|
"grad_norm": 1459.4316477205139, |
|
"learning_rate": 1.9347820230782295e-09, |
|
"logits/chosen": 5948.6728515625, |
|
"logits/rejected": 5038.36181640625, |
|
"logps/chosen": -268.30438232421875, |
|
"logps/rejected": -239.8975067138672, |
|
"loss": 486.2789, |
|
"rewards/accuracies": 0.5249999761581421, |
|
"rewards/chosen": 0.4469468593597412, |
|
"rewards/margins": -0.011204786598682404, |
|
"rewards/rejected": 0.4581516683101654, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.98403559277676, |
|
"grad_norm": 1500.642382585496, |
|
"learning_rate": 3.2839470889836627e-10, |
|
"logits/chosen": 5545.70947265625, |
|
"logits/rejected": 4929.12255859375, |
|
"logps/chosen": -237.06051635742188, |
|
"logps/rejected": -213.99795532226562, |
|
"loss": 457.8427, |
|
"rewards/accuracies": 0.528124988079071, |
|
"rewards/chosen": 0.4377865195274353, |
|
"rewards/margins": -0.0002623729524202645, |
|
"rewards/rejected": 0.4380488991737366, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.998691442030882, |
|
"step": 477, |
|
"total_flos": 0.0, |
|
"train_loss": 465.78390522123135, |
|
"train_runtime": 16522.4635, |
|
"train_samples_per_second": 3.7, |
|
"train_steps_per_second": 0.029 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 477, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 1000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": false, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|