{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.0837464826477288, "eval_steps": 500, "global_step": 2500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.001674929652954576, "grad_norm": 0.30847179889678955, "learning_rate": 2.9983250703470456e-05, "loss": 0.5044, "step": 50 }, { "epoch": 0.003349859305909152, "grad_norm": 0.04476890340447426, "learning_rate": 2.9966501406940907e-05, "loss": 0.0044, "step": 100 }, { "epoch": 0.005024788958863728, "grad_norm": 0.08374588936567307, "learning_rate": 2.9949752110411362e-05, "loss": 0.0026, "step": 150 }, { "epoch": 0.006699718611818304, "grad_norm": 0.0017046213615685701, "learning_rate": 2.9933002813881818e-05, "loss": 0.0023, "step": 200 }, { "epoch": 0.008374648264772879, "grad_norm": 0.019402090460062027, "learning_rate": 2.9916253517352273e-05, "loss": 0.0013, "step": 250 }, { "epoch": 0.010049577917727455, "grad_norm": 0.010655886493623257, "learning_rate": 2.9899504220822728e-05, "loss": 0.002, "step": 300 }, { "epoch": 0.011724507570682031, "grad_norm": 0.006348441354930401, "learning_rate": 2.9882754924293183e-05, "loss": 0.0017, "step": 350 }, { "epoch": 0.013399437223636608, "grad_norm": 0.06810770183801651, "learning_rate": 2.9866005627763634e-05, "loss": 0.0015, "step": 400 }, { "epoch": 0.015074366876591184, "grad_norm": 0.003874759189784527, "learning_rate": 2.984925633123409e-05, "loss": 0.0012, "step": 450 }, { "epoch": 0.016749296529545758, "grad_norm": 0.003985659219324589, "learning_rate": 2.9832507034704544e-05, "loss": 0.0012, "step": 500 }, { "epoch": 0.018424226182500336, "grad_norm": 0.01902610994875431, "learning_rate": 2.9815757738174996e-05, "loss": 0.0014, "step": 550 }, { "epoch": 0.02009915583545491, "grad_norm": 0.01615321636199951, "learning_rate": 2.979900844164545e-05, "loss": 0.0013, "step": 600 }, { "epoch": 0.02177408548840949, "grad_norm": 0.0055249775759875774, "learning_rate": 2.9782259145115903e-05, "loss": 0.0008, "step": 650 }, { "epoch": 0.023449015141364063, "grad_norm": 0.0019460869953036308, "learning_rate": 2.976550984858636e-05, "loss": 0.0011, "step": 700 }, { "epoch": 0.025123944794318637, "grad_norm": 0.012369350530207157, "learning_rate": 2.9748760552056816e-05, "loss": 0.0011, "step": 750 }, { "epoch": 0.026798874447273215, "grad_norm": 0.00836873333901167, "learning_rate": 2.9732011255527268e-05, "loss": 0.0013, "step": 800 }, { "epoch": 0.02847380410022779, "grad_norm": 0.0034731472842395306, "learning_rate": 2.9715261958997723e-05, "loss": 0.0016, "step": 850 }, { "epoch": 0.030148733753182368, "grad_norm": 0.015365710482001305, "learning_rate": 2.9698512662468178e-05, "loss": 0.0013, "step": 900 }, { "epoch": 0.031823663406136946, "grad_norm": 0.017056584358215332, "learning_rate": 2.968176336593863e-05, "loss": 0.0013, "step": 950 }, { "epoch": 0.033498593059091517, "grad_norm": 0.0038616659585386515, "learning_rate": 2.9665014069409085e-05, "loss": 0.0009, "step": 1000 }, { "epoch": 0.035173522712046094, "grad_norm": 0.0008767916006036103, "learning_rate": 2.964826477287954e-05, "loss": 0.0006, "step": 1050 }, { "epoch": 0.03684845236500067, "grad_norm": 0.050677187740802765, "learning_rate": 2.963151547634999e-05, "loss": 0.0014, "step": 1100 }, { "epoch": 0.03852338201795524, "grad_norm": 0.011451843194663525, "learning_rate": 2.961476617982045e-05, "loss": 0.0008, "step": 1150 }, { "epoch": 0.04019831167090982, "grad_norm": 0.00335301854647696, "learning_rate": 2.9598016883290905e-05, "loss": 0.0013, "step": 1200 }, { "epoch": 0.0418732413238644, "grad_norm": 0.01707889698445797, "learning_rate": 2.9581267586761356e-05, "loss": 0.0005, "step": 1250 }, { "epoch": 0.04354817097681898, "grad_norm": 0.0004460318305063993, "learning_rate": 2.956451829023181e-05, "loss": 0.0011, "step": 1300 }, { "epoch": 0.04522310062977355, "grad_norm": 0.00859643705189228, "learning_rate": 2.9547768993702266e-05, "loss": 0.0007, "step": 1350 }, { "epoch": 0.046898030282728126, "grad_norm": 0.012995535507798195, "learning_rate": 2.9531019697172718e-05, "loss": 0.0009, "step": 1400 }, { "epoch": 0.048572959935682704, "grad_norm": 0.004834771156311035, "learning_rate": 2.9514270400643173e-05, "loss": 0.0009, "step": 1450 }, { "epoch": 0.050247889588637275, "grad_norm": 0.006121751386672258, "learning_rate": 2.9497521104113628e-05, "loss": 0.0009, "step": 1500 }, { "epoch": 0.05192281924159185, "grad_norm": 0.004112472757697105, "learning_rate": 2.9480771807584083e-05, "loss": 0.0007, "step": 1550 }, { "epoch": 0.05359774889454643, "grad_norm": 0.0025941322091966867, "learning_rate": 2.9464022511054538e-05, "loss": 0.0008, "step": 1600 }, { "epoch": 0.055272678547501, "grad_norm": 0.0033354111947119236, "learning_rate": 2.9447273214524993e-05, "loss": 0.0014, "step": 1650 }, { "epoch": 0.05694760820045558, "grad_norm": 0.0006163300131447613, "learning_rate": 2.9430523917995445e-05, "loss": 0.0013, "step": 1700 }, { "epoch": 0.05862253785341016, "grad_norm": 0.03437214344739914, "learning_rate": 2.94137746214659e-05, "loss": 0.0011, "step": 1750 }, { "epoch": 0.060297467506364735, "grad_norm": 0.004870133940130472, "learning_rate": 2.939702532493635e-05, "loss": 0.0011, "step": 1800 }, { "epoch": 0.061972397159319306, "grad_norm": 0.011027672328054905, "learning_rate": 2.9380276028406807e-05, "loss": 0.0011, "step": 1850 }, { "epoch": 0.06364732681227389, "grad_norm": 0.005804801359772682, "learning_rate": 2.936352673187726e-05, "loss": 0.0009, "step": 1900 }, { "epoch": 0.06532225646522846, "grad_norm": 0.001421495107933879, "learning_rate": 2.9346777435347713e-05, "loss": 0.0007, "step": 1950 }, { "epoch": 0.06699718611818303, "grad_norm": 0.00775284506380558, "learning_rate": 2.9330028138818172e-05, "loss": 0.0008, "step": 2000 }, { "epoch": 0.06867211577113762, "grad_norm": 0.009989109821617603, "learning_rate": 2.9313278842288627e-05, "loss": 0.0013, "step": 2050 }, { "epoch": 0.07034704542409219, "grad_norm": 0.004933220334351063, "learning_rate": 2.929652954575908e-05, "loss": 0.0008, "step": 2100 }, { "epoch": 0.07202197507704676, "grad_norm": 0.0003998636966571212, "learning_rate": 2.9279780249229533e-05, "loss": 0.0005, "step": 2150 }, { "epoch": 0.07369690473000134, "grad_norm": 0.00017748262325767428, "learning_rate": 2.926303095269999e-05, "loss": 0.0006, "step": 2200 }, { "epoch": 0.07537183438295592, "grad_norm": 0.0009646079852245748, "learning_rate": 2.924628165617044e-05, "loss": 0.0007, "step": 2250 }, { "epoch": 0.07704676403591049, "grad_norm": 0.004842822439968586, "learning_rate": 2.9229532359640895e-05, "loss": 0.0011, "step": 2300 }, { "epoch": 0.07872169368886507, "grad_norm": 0.010380016639828682, "learning_rate": 2.921278306311135e-05, "loss": 0.001, "step": 2350 }, { "epoch": 0.08039662334181964, "grad_norm": 0.0021645210217684507, "learning_rate": 2.9196033766581802e-05, "loss": 0.0009, "step": 2400 }, { "epoch": 0.08207155299477421, "grad_norm": 0.0019599520601332188, "learning_rate": 2.917928447005226e-05, "loss": 0.0009, "step": 2450 }, { "epoch": 0.0837464826477288, "grad_norm": 0.03942473977804184, "learning_rate": 2.9162535173522715e-05, "loss": 0.0011, "step": 2500 } ], "logging_steps": 50, "max_steps": 89556, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "total_flos": 0.0, "train_batch_size": 64, "trial_name": null, "trial_params": null }