{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 430, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.29, "grad_norm": 0.3876499831676483, "learning_rate": 0.00011627906976744187, "loss": 1.3841, "step": 25 }, { "epoch": 0.58, "grad_norm": 0.35026389360427856, "learning_rate": 0.0001963824289405685, "loss": 0.8968, "step": 50 }, { "epoch": 0.87, "grad_norm": 0.317654550075531, "learning_rate": 0.00018346253229974162, "loss": 0.6679, "step": 75 }, { "epoch": 1.16, "grad_norm": 0.4258534610271454, "learning_rate": 0.00017054263565891473, "loss": 0.5714, "step": 100 }, { "epoch": 1.45, "grad_norm": 0.4060456156730652, "learning_rate": 0.00015762273901808786, "loss": 0.5099, "step": 125 }, { "epoch": 1.74, "grad_norm": 0.37451767921447754, "learning_rate": 0.000144702842377261, "loss": 0.5017, "step": 150 }, { "epoch": 2.03, "grad_norm": 0.36458975076675415, "learning_rate": 0.0001317829457364341, "loss": 0.4647, "step": 175 }, { "epoch": 2.33, "grad_norm": 0.3686697781085968, "learning_rate": 0.00011886304909560724, "loss": 0.4359, "step": 200 }, { "epoch": 2.62, "grad_norm": 0.37151384353637695, "learning_rate": 0.00010594315245478037, "loss": 0.4268, "step": 225 }, { "epoch": 2.91, "grad_norm": 0.45622408390045166, "learning_rate": 9.30232558139535e-05, "loss": 0.4423, "step": 250 }, { "epoch": 3.2, "grad_norm": 0.44779106974601746, "learning_rate": 8.010335917312663e-05, "loss": 0.3945, "step": 275 }, { "epoch": 3.49, "grad_norm": 0.4048546552658081, "learning_rate": 6.718346253229974e-05, "loss": 0.3765, "step": 300 }, { "epoch": 3.78, "grad_norm": 0.5087065696716309, "learning_rate": 5.426356589147287e-05, "loss": 0.3912, "step": 325 }, { "epoch": 4.07, "grad_norm": 0.5312517285346985, "learning_rate": 4.1343669250646e-05, "loss": 0.3972, "step": 350 }, { "epoch": 4.36, "grad_norm": 0.47890299558639526, "learning_rate": 2.842377260981912e-05, "loss": 0.3485, "step": 375 }, { "epoch": 4.65, "grad_norm": 0.5938854217529297, "learning_rate": 1.5503875968992248e-05, "loss": 0.3707, "step": 400 }, { "epoch": 4.94, "grad_norm": 0.5553755760192871, "learning_rate": 2.583979328165375e-06, "loss": 0.3336, "step": 425 } ], "logging_steps": 25, "max_steps": 430, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "total_flos": 6.991294110892032e+16, "train_batch_size": 4, "trial_name": null, "trial_params": null }