{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.998691442030882, "eval_steps": 500, "global_step": 477, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.002093692750588851, "grad_norm": 4.60553352196104, "learning_rate": 1.0416666666666666e-08, "logits/chosen": 5002.53564453125, "logits/rejected": 4591.72021484375, "logps/chosen": -265.6396789550781, "logps/rejected": -206.22401428222656, "loss": 0.6931, "rewards/accuracies": 0.0, "rewards/chosen": 0.0, "rewards/margins": 0.0, "rewards/rejected": 0.0, "step": 1 }, { "epoch": 0.02093692750588851, "grad_norm": 3.906214184095967, "learning_rate": 1.0416666666666667e-07, "logits/chosen": 5977.31591796875, "logits/rejected": 5015.41943359375, "logps/chosen": -292.51190185546875, "logps/rejected": -259.33819580078125, "loss": 0.6931, "rewards/accuracies": 0.4791666567325592, "rewards/chosen": 0.0004885931266471744, "rewards/margins": 0.0006340253748930991, "rewards/rejected": -0.00014543221914209425, "step": 10 }, { "epoch": 0.04187385501177702, "grad_norm": 4.408104522173234, "learning_rate": 2.0833333333333333e-07, "logits/chosen": 5877.7744140625, "logits/rejected": 4856.02734375, "logps/chosen": -290.4228820800781, "logps/rejected": -249.76522827148438, "loss": 0.6927, "rewards/accuracies": 0.6000000238418579, "rewards/chosen": 0.0017061525722965598, "rewards/margins": 0.002024381887167692, "rewards/rejected": -0.00031822946039028466, "step": 20 }, { "epoch": 0.06281078251766553, "grad_norm": 4.1323992714408995, "learning_rate": 3.1249999999999997e-07, "logits/chosen": 5644.30419921875, "logits/rejected": 4922.60302734375, "logps/chosen": -270.80462646484375, "logps/rejected": -238.2168426513672, "loss": 0.6907, "rewards/accuracies": 0.5874999761581421, "rewards/chosen": 0.014328369870781898, "rewards/margins": 0.008474691770970821, "rewards/rejected": 0.00585367763414979, "step": 30 }, { "epoch": 0.08374771002355404, "grad_norm": 3.9871215863103235, "learning_rate": 4.1666666666666667e-07, "logits/chosen": 5738.802734375, "logits/rejected": 4992.0400390625, "logps/chosen": -269.14434814453125, "logps/rejected": -256.4861145019531, "loss": 0.6855, "rewards/accuracies": 0.6656249761581421, "rewards/chosen": 0.048127662390470505, "rewards/margins": 0.028449412435293198, "rewards/rejected": 0.019678251817822456, "step": 40 }, { "epoch": 0.10468463752944256, "grad_norm": 4.4563458428725395, "learning_rate": 4.999731868769026e-07, "logits/chosen": 6019.1103515625, "logits/rejected": 5225.02392578125, "logps/chosen": -276.1994323730469, "logps/rejected": -271.67559814453125, "loss": 0.679, "rewards/accuracies": 0.6812499761581421, "rewards/chosen": 0.024182714521884918, "rewards/margins": 0.0657457783818245, "rewards/rejected": -0.04156306013464928, "step": 50 }, { "epoch": 0.12562156503533106, "grad_norm": 5.034590628294388, "learning_rate": 4.990353313429303e-07, "logits/chosen": 5789.7138671875, "logits/rejected": 4927.72412109375, "logps/chosen": -283.95257568359375, "logps/rejected": -268.9764709472656, "loss": 0.6716, "rewards/accuracies": 0.6812499761581421, "rewards/chosen": -0.055300962179899216, "rewards/margins": 0.11967425048351288, "rewards/rejected": -0.1749752014875412, "step": 60 }, { "epoch": 0.14655849254121958, "grad_norm": 6.512738637492193, "learning_rate": 4.967625656594781e-07, "logits/chosen": 5590.95166015625, "logits/rejected": 5145.8408203125, "logps/chosen": -288.2838439941406, "logps/rejected": -293.23541259765625, "loss": 0.664, "rewards/accuracies": 0.6625000238418579, "rewards/chosen": -0.09975917637348175, "rewards/margins": 0.16456229984760284, "rewards/rejected": -0.2643214762210846, "step": 70 }, { "epoch": 0.16749542004710807, "grad_norm": 6.2250099209512655, "learning_rate": 4.93167072587771e-07, "logits/chosen": 5744.78662109375, "logits/rejected": 4929.46435546875, "logps/chosen": -284.8975524902344, "logps/rejected": -283.4288330078125, "loss": 0.6577, "rewards/accuracies": 0.6781250238418579, "rewards/chosen": -0.08810611069202423, "rewards/margins": 0.21545740962028503, "rewards/rejected": -0.30356353521347046, "step": 80 }, { "epoch": 0.1884323475529966, "grad_norm": 6.510058896775878, "learning_rate": 4.882681251368548e-07, "logits/chosen": 5883.4892578125, "logits/rejected": 4958.2470703125, "logps/chosen": -292.3514709472656, "logps/rejected": -294.89471435546875, "loss": 0.6536, "rewards/accuracies": 0.684374988079071, "rewards/chosen": -0.14643868803977966, "rewards/margins": 0.2527722418308258, "rewards/rejected": -0.39921098947525024, "step": 90 }, { "epoch": 0.2093692750588851, "grad_norm": 9.163808135102661, "learning_rate": 4.820919832540181e-07, "logits/chosen": 5974.72802734375, "logits/rejected": 5452.3408203125, "logps/chosen": -311.05804443359375, "logps/rejected": -347.23944091796875, "loss": 0.6466, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -0.3118935227394104, "rewards/margins": 0.41534680128097534, "rewards/rejected": -0.727240264415741, "step": 100 }, { "epoch": 0.23030620256477363, "grad_norm": 7.986681346223677, "learning_rate": 4.7467175306295647e-07, "logits/chosen": 6145.7939453125, "logits/rejected": 4966.3818359375, "logps/chosen": -318.4931640625, "logps/rejected": -344.3719177246094, "loss": 0.6384, "rewards/accuracies": 0.7093750238418579, "rewards/chosen": -0.4387056231498718, "rewards/margins": 0.4162977337837219, "rewards/rejected": -0.8550033569335938, "step": 110 }, { "epoch": 0.2512431300706621, "grad_norm": 9.318478390808158, "learning_rate": 4.6604720940421207e-07, "logits/chosen": 5758.07373046875, "logits/rejected": 5305.98193359375, "logps/chosen": -309.2460632324219, "logps/rejected": -346.93157958984375, "loss": 0.6446, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -0.42067036032676697, "rewards/margins": 0.3798871636390686, "rewards/rejected": -0.8005574941635132, "step": 120 }, { "epoch": 0.2721800575765506, "grad_norm": 12.596663584835428, "learning_rate": 4.5626458262912735e-07, "logits/chosen": 5819.57373046875, "logits/rejected": 5167.3623046875, "logps/chosen": -298.8730163574219, "logps/rejected": -338.01397705078125, "loss": 0.6418, "rewards/accuracies": 0.675000011920929, "rewards/chosen": -0.3035568296909332, "rewards/margins": 0.3864966928958893, "rewards/rejected": -0.6900535225868225, "step": 130 }, { "epoch": 0.29311698508243916, "grad_norm": 8.250031153064022, "learning_rate": 4.453763107901675e-07, "logits/chosen": 5601.96484375, "logits/rejected": 4937.931640625, "logps/chosen": -306.513427734375, "logps/rejected": -329.6960754394531, "loss": 0.6421, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -0.3460861146450043, "rewards/margins": 0.4395461976528168, "rewards/rejected": -0.785632312297821, "step": 140 }, { "epoch": 0.31405391258832765, "grad_norm": 10.215191441356916, "learning_rate": 4.3344075855595097e-07, "logits/chosen": 6188.27392578125, "logits/rejected": 5056.6064453125, "logps/chosen": -325.03436279296875, "logps/rejected": -337.5921630859375, "loss": 0.6367, "rewards/accuracies": 0.71875, "rewards/chosen": -0.3591935336589813, "rewards/margins": 0.5018205642700195, "rewards/rejected": -0.8610140681266785, "step": 150 }, { "epoch": 0.33499084009421615, "grad_norm": 11.588598413111276, "learning_rate": 4.2052190435769554e-07, "logits/chosen": 5836.8857421875, "logits/rejected": 4833.65673828125, "logps/chosen": -338.6114807128906, "logps/rejected": -368.7120056152344, "loss": 0.6254, "rewards/accuracies": 0.737500011920929, "rewards/chosen": -0.5401726961135864, "rewards/margins": 0.583164393901825, "rewards/rejected": -1.1233372688293457, "step": 160 }, { "epoch": 0.3559277676001047, "grad_norm": 16.573341127038365, "learning_rate": 4.0668899744407567e-07, "logits/chosen": 6162.7734375, "logits/rejected": 5209.9150390625, "logps/chosen": -329.6107482910156, "logps/rejected": -334.8266906738281, "loss": 0.6289, "rewards/accuracies": 0.737500011920929, "rewards/chosen": -0.47295036911964417, "rewards/margins": 0.46886569261550903, "rewards/rejected": -0.941815972328186, "step": 170 }, { "epoch": 0.3768646951059932, "grad_norm": 8.541127331717703, "learning_rate": 3.920161866827889e-07, "logits/chosen": 5801.9912109375, "logits/rejected": 4906.3310546875, "logps/chosen": -352.50616455078125, "logps/rejected": -369.984375, "loss": 0.6347, "rewards/accuracies": 0.721875011920929, "rewards/chosen": -0.6521086096763611, "rewards/margins": 0.5479272603988647, "rewards/rejected": -1.200035810470581, "step": 180 }, { "epoch": 0.39780162261188173, "grad_norm": 10.691736891405624, "learning_rate": 3.765821230985757e-07, "logits/chosen": 5942.0029296875, "logits/rejected": 4911.9306640625, "logps/chosen": -313.849853515625, "logps/rejected": -363.1335144042969, "loss": 0.6223, "rewards/accuracies": 0.75, "rewards/chosen": -0.44098320603370667, "rewards/margins": 0.7115194201469421, "rewards/rejected": -1.1525026559829712, "step": 190 }, { "epoch": 0.4187385501177702, "grad_norm": 12.414499403080086, "learning_rate": 3.604695382782159e-07, "logits/chosen": 5570.83056640625, "logits/rejected": 4781.8603515625, "logps/chosen": -307.8031921386719, "logps/rejected": -357.94049072265625, "loss": 0.6348, "rewards/accuracies": 0.71875, "rewards/chosen": -0.5345231294631958, "rewards/margins": 0.6162750720977783, "rewards/rejected": -1.1507983207702637, "step": 200 }, { "epoch": 0.4396754776236587, "grad_norm": 9.725871680307973, "learning_rate": 3.4376480090239047e-07, "logits/chosen": 6185.18505859375, "logits/rejected": 5210.94775390625, "logps/chosen": -344.7357177734375, "logps/rejected": -370.8301696777344, "loss": 0.6292, "rewards/accuracies": 0.7406250238418579, "rewards/chosen": -0.5796831846237183, "rewards/margins": 0.5115828514099121, "rewards/rejected": -1.0912659168243408, "step": 210 }, { "epoch": 0.46061240512954726, "grad_norm": 10.26939339999947, "learning_rate": 3.265574537815398e-07, "logits/chosen": 5568.31298828125, "logits/rejected": 5116.90087890625, "logps/chosen": -316.9978942871094, "logps/rejected": -368.52630615234375, "loss": 0.6218, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.45882177352905273, "rewards/margins": 0.6230064630508423, "rewards/rejected": -1.0818283557891846, "step": 220 }, { "epoch": 0.48154933263543576, "grad_norm": 11.837301535764077, "learning_rate": 3.0893973387735683e-07, "logits/chosen": 5724.6162109375, "logits/rejected": 4797.6650390625, "logps/chosen": -357.8770446777344, "logps/rejected": -405.49285888671875, "loss": 0.6198, "rewards/accuracies": 0.7718750238418579, "rewards/chosen": -0.8960827589035034, "rewards/margins": 0.7045291066169739, "rewards/rejected": -1.600611925125122, "step": 230 }, { "epoch": 0.5024862601413242, "grad_norm": 11.232499614895293, "learning_rate": 2.910060778827554e-07, "logits/chosen": 5925.3447265625, "logits/rejected": 4943.4423828125, "logps/chosen": -331.6304931640625, "logps/rejected": -372.5296325683594, "loss": 0.6256, "rewards/accuracies": 0.737500011920929, "rewards/chosen": -0.6605360507965088, "rewards/margins": 0.5669505000114441, "rewards/rejected": -1.2274866104125977, "step": 240 }, { "epoch": 0.5234231876472127, "grad_norm": 12.90825856853013, "learning_rate": 2.7285261601056697e-07, "logits/chosen": 6004.00732421875, "logits/rejected": 5001.3466796875, "logps/chosen": -353.7926330566406, "logps/rejected": -393.0778503417969, "loss": 0.6308, "rewards/accuracies": 0.734375, "rewards/chosen": -0.6784394383430481, "rewards/margins": 0.6315867900848389, "rewards/rejected": -1.3100261688232422, "step": 250 }, { "epoch": 0.5443601151531012, "grad_norm": 13.478020063581878, "learning_rate": 2.5457665670441937e-07, "logits/chosen": 5818.59130859375, "logits/rejected": 5201.59130859375, "logps/chosen": -355.6083068847656, "logps/rejected": -395.66754150390625, "loss": 0.6232, "rewards/accuracies": 0.78125, "rewards/chosen": -0.6953917145729065, "rewards/margins": 0.6504613757133484, "rewards/rejected": -1.3458532094955444, "step": 260 }, { "epoch": 0.5652970426589898, "grad_norm": 11.290710956607453, "learning_rate": 2.3627616503391812e-07, "logits/chosen": 5881.2919921875, "logits/rejected": 5333.2744140625, "logps/chosen": -335.8533935546875, "logps/rejected": -390.1767883300781, "loss": 0.6207, "rewards/accuracies": 0.793749988079071, "rewards/chosen": -0.612395167350769, "rewards/margins": 0.6854075193405151, "rewards/rejected": -1.2978026866912842, "step": 270 }, { "epoch": 0.5862339701648783, "grad_norm": 11.562696281009446, "learning_rate": 2.1804923757009882e-07, "logits/chosen": 5932.57861328125, "logits/rejected": 4862.0146484375, "logps/chosen": -341.2466735839844, "logps/rejected": -384.97857666015625, "loss": 0.6207, "rewards/accuracies": 0.7749999761581421, "rewards/chosen": -0.6052099466323853, "rewards/margins": 0.7767351269721985, "rewards/rejected": -1.3819451332092285, "step": 280 }, { "epoch": 0.6071708976707668, "grad_norm": 14.192713016769995, "learning_rate": 1.9999357655598891e-07, "logits/chosen": 6152.61669921875, "logits/rejected": 5673.72900390625, "logps/chosen": -372.6900329589844, "logps/rejected": -436.62286376953125, "loss": 0.6269, "rewards/accuracies": 0.737500011920929, "rewards/chosen": -0.8125594258308411, "rewards/margins": 0.5744532346725464, "rewards/rejected": -1.3870127201080322, "step": 290 }, { "epoch": 0.6281078251766553, "grad_norm": 13.845927183072918, "learning_rate": 1.8220596619089573e-07, "logits/chosen": 5640.8662109375, "logits/rejected": 5246.69091796875, "logps/chosen": -338.5563659667969, "logps/rejected": -399.28900146484375, "loss": 0.6223, "rewards/accuracies": 0.734375, "rewards/chosen": -0.6720638275146484, "rewards/margins": 0.61163729429245, "rewards/rejected": -1.2837010622024536, "step": 300 }, { "epoch": 0.6490447526825438, "grad_norm": 11.103620407034061, "learning_rate": 1.647817538357072e-07, "logits/chosen": 6317.97607421875, "logits/rejected": 5491.85498046875, "logps/chosen": -355.3973388671875, "logps/rejected": -393.3858947753906, "loss": 0.6185, "rewards/accuracies": 0.7749999761581421, "rewards/chosen": -0.6456772089004517, "rewards/margins": 0.5926147699356079, "rewards/rejected": -1.2382919788360596, "step": 310 }, { "epoch": 0.6699816801884323, "grad_norm": 12.452201295868143, "learning_rate": 1.478143389201113e-07, "logits/chosen": 5882.0078125, "logits/rejected": 5039.4091796875, "logps/chosen": -335.7853698730469, "logps/rejected": -396.05712890625, "loss": 0.6228, "rewards/accuracies": 0.71875, "rewards/chosen": -0.6392192840576172, "rewards/margins": 0.723227858543396, "rewards/rejected": -1.3624470233917236, "step": 320 }, { "epoch": 0.6909186076943209, "grad_norm": 11.031080841746133, "learning_rate": 1.3139467229135998e-07, "logits/chosen": 5675.63623046875, "logits/rejected": 5184.22607421875, "logps/chosen": -334.7772521972656, "logps/rejected": -389.0362243652344, "loss": 0.6271, "rewards/accuracies": 0.7562500238418579, "rewards/chosen": -0.5799298286437988, "rewards/margins": 0.6671385765075684, "rewards/rejected": -1.2470682859420776, "step": 330 }, { "epoch": 0.7118555352002094, "grad_norm": 10.777776010932476, "learning_rate": 1.1561076868822755e-07, "logits/chosen": 5497.7236328125, "logits/rejected": 4727.431640625, "logps/chosen": -319.00543212890625, "logps/rejected": -367.11663818359375, "loss": 0.6253, "rewards/accuracies": 0.7281249761581421, "rewards/chosen": -0.6143845915794373, "rewards/margins": 0.6447553038597107, "rewards/rejected": -1.2591397762298584, "step": 340 }, { "epoch": 0.7327924627060979, "grad_norm": 13.171378321554494, "learning_rate": 1.0054723495346482e-07, "logits/chosen": 6181.7666015625, "logits/rejected": 4961.35400390625, "logps/chosen": -359.21905517578125, "logps/rejected": -390.53350830078125, "loss": 0.6232, "rewards/accuracies": 0.7749999761581421, "rewards/chosen": -0.5830456018447876, "rewards/margins": 0.6389049291610718, "rewards/rejected": -1.2219504117965698, "step": 350 }, { "epoch": 0.7537293902119864, "grad_norm": 12.17139481883657, "learning_rate": 8.628481651367875e-08, "logits/chosen": 5867.166015625, "logits/rejected": 5056.07568359375, "logps/chosen": -332.9352722167969, "logps/rejected": -381.3428955078125, "loss": 0.6279, "rewards/accuracies": 0.7281249761581421, "rewards/chosen": -0.5521169900894165, "rewards/margins": 0.678789496421814, "rewards/rejected": -1.2309064865112305, "step": 360 }, { "epoch": 0.7746663177178749, "grad_norm": 10.377060003916368, "learning_rate": 7.289996455765748e-08, "logits/chosen": 5769.4111328125, "logits/rejected": 4844.27880859375, "logps/chosen": -343.3131103515625, "logps/rejected": -383.9873962402344, "loss": 0.6145, "rewards/accuracies": 0.765625, "rewards/chosen": -0.5452004671096802, "rewards/margins": 0.725028932094574, "rewards/rejected": -1.270229458808899, "step": 370 }, { "epoch": 0.7956032452237635, "grad_norm": 13.056548862753138, "learning_rate": 6.046442623320145e-08, "logits/chosen": 6045.7568359375, "logits/rejected": 5162.32958984375, "logps/chosen": -337.4881896972656, "logps/rejected": -362.61529541015625, "loss": 0.626, "rewards/accuracies": 0.737500011920929, "rewards/chosen": -0.527867317199707, "rewards/margins": 0.5897092223167419, "rewards/rejected": -1.1175765991210938, "step": 380 }, { "epoch": 0.816540172729652, "grad_norm": 9.602896358416052, "learning_rate": 4.904486005914027e-08, "logits/chosen": 6276.33837890625, "logits/rejected": 5479.82177734375, "logps/chosen": -344.8057861328125, "logps/rejected": -400.14154052734375, "loss": 0.6156, "rewards/accuracies": 0.762499988079071, "rewards/chosen": -0.530563473701477, "rewards/margins": 0.7230854034423828, "rewards/rejected": -1.2536487579345703, "step": 390 }, { "epoch": 0.8374771002355405, "grad_norm": 10.002000791834785, "learning_rate": 3.8702478614051345e-08, "logits/chosen": 5853.0087890625, "logits/rejected": 5118.06494140625, "logps/chosen": -347.6117248535156, "logps/rejected": -372.6192932128906, "loss": 0.6239, "rewards/accuracies": 0.753125011920929, "rewards/chosen": -0.5670760273933411, "rewards/margins": 0.6005390882492065, "rewards/rejected": -1.1676150560379028, "step": 400 }, { "epoch": 0.8584140277414289, "grad_norm": 10.39957529568141, "learning_rate": 2.9492720416985e-08, "logits/chosen": 5829.76318359375, "logits/rejected": 5050.9814453125, "logps/chosen": -343.9368591308594, "logps/rejected": -402.4644470214844, "loss": 0.6171, "rewards/accuracies": 0.715624988079071, "rewards/chosen": -0.6206948757171631, "rewards/margins": 0.592847466468811, "rewards/rejected": -1.2135423421859741, "step": 410 }, { "epoch": 0.8793509552473174, "grad_norm": 12.525714643168069, "learning_rate": 2.1464952759020856e-08, "logits/chosen": 6221.03662109375, "logits/rejected": 5368.93994140625, "logps/chosen": -350.761474609375, "logps/rejected": -397.9172058105469, "loss": 0.6175, "rewards/accuracies": 0.737500011920929, "rewards/chosen": -0.5878224968910217, "rewards/margins": 0.6664667129516602, "rewards/rejected": -1.254289150238037, "step": 420 }, { "epoch": 0.9002878827532059, "grad_norm": 11.787131409532831, "learning_rate": 1.4662207078575684e-08, "logits/chosen": 6147.80322265625, "logits/rejected": 4875.4560546875, "logps/chosen": -347.69866943359375, "logps/rejected": -387.70556640625, "loss": 0.612, "rewards/accuracies": 0.762499988079071, "rewards/chosen": -0.6009602546691895, "rewards/margins": 0.7015070915222168, "rewards/rejected": -1.3024674654006958, "step": 430 }, { "epoch": 0.9212248102590945, "grad_norm": 11.622178629621203, "learning_rate": 9.12094829893642e-09, "logits/chosen": 5975.74169921875, "logits/rejected": 5603.3095703125, "logps/chosen": -335.4075622558594, "logps/rejected": -398.03759765625, "loss": 0.619, "rewards/accuracies": 0.753125011920929, "rewards/chosen": -0.5686401724815369, "rewards/margins": 0.6512762308120728, "rewards/rejected": -1.2199163436889648, "step": 440 }, { "epoch": 0.942161737764983, "grad_norm": 12.334438799788524, "learning_rate": 4.8708793644441086e-09, "logits/chosen": 6104.8115234375, "logits/rejected": 5160.8291015625, "logps/chosen": -347.5029602050781, "logps/rejected": -385.18292236328125, "loss": 0.6176, "rewards/accuracies": 0.7749999761581421, "rewards/chosen": -0.5716773271560669, "rewards/margins": 0.6656473875045776, "rewards/rejected": -1.237324595451355, "step": 450 }, { "epoch": 0.9630986652708715, "grad_norm": 14.274087744528849, "learning_rate": 1.9347820230782295e-09, "logits/chosen": 6278.4931640625, "logits/rejected": 5222.73828125, "logps/chosen": -374.8132629394531, "logps/rejected": -419.890625, "loss": 0.6239, "rewards/accuracies": 0.753125011920929, "rewards/chosen": -0.6181422472000122, "rewards/margins": 0.7236369848251343, "rewards/rejected": -1.3417792320251465, "step": 460 }, { "epoch": 0.98403559277676, "grad_norm": 9.711310883436695, "learning_rate": 3.2839470889836627e-10, "logits/chosen": 5856.66259765625, "logits/rejected": 5124.48193359375, "logps/chosen": -340.29803466796875, "logps/rejected": -388.76165771484375, "loss": 0.6131, "rewards/accuracies": 0.7718750238418579, "rewards/chosen": -0.5945884585380554, "rewards/margins": 0.7149998545646667, "rewards/rejected": -1.3095881938934326, "step": 470 }, { "epoch": 0.998691442030882, "step": 477, "total_flos": 0.0, "train_loss": 0.6354130409048788, "train_runtime": 12872.7493, "train_samples_per_second": 4.749, "train_steps_per_second": 0.037 } ], "logging_steps": 10, "max_steps": 477, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }