{ "best_metric": null, "best_model_checkpoint": null, "epoch": 10.285714285714286, "eval_steps": 500, "global_step": 18, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.57, "grad_norm": 0.2423473596572876, "learning_rate": 0.0001, "loss": 4.0857, "step": 1 }, { "epoch": 1.14, "grad_norm": 0.2876349985599518, "learning_rate": 0.0002, "loss": 4.396, "step": 2 }, { "epoch": 1.71, "grad_norm": 0.25892937183380127, "learning_rate": 0.0001875, "loss": 4.257, "step": 3 }, { "epoch": 2.29, "grad_norm": 0.24695058166980743, "learning_rate": 0.000175, "loss": 4.2171, "step": 4 }, { "epoch": 2.86, "grad_norm": 0.439170777797699, "learning_rate": 0.00016250000000000002, "loss": 4.0037, "step": 5 }, { "epoch": 3.43, "grad_norm": 0.3510642647743225, "learning_rate": 0.00015000000000000001, "loss": 4.1377, "step": 6 }, { "epoch": 4.0, "grad_norm": 0.3439651131629944, "learning_rate": 0.0001375, "loss": 4.0279, "step": 7 }, { "epoch": 4.57, "grad_norm": 0.3093964159488678, "learning_rate": 0.000125, "loss": 3.8389, "step": 8 }, { "epoch": 5.14, "grad_norm": 0.3536764979362488, "learning_rate": 0.00011250000000000001, "loss": 4.1566, "step": 9 }, { "epoch": 5.71, "grad_norm": 0.3965250849723816, "learning_rate": 0.0001, "loss": 3.969, "step": 10 }, { "epoch": 6.29, "grad_norm": 0.4037889838218689, "learning_rate": 8.75e-05, "loss": 3.9695, "step": 11 }, { "epoch": 6.86, "grad_norm": 0.3280973434448242, "learning_rate": 7.500000000000001e-05, "loss": 3.7841, "step": 12 }, { "epoch": 7.43, "grad_norm": 0.3607095777988434, "learning_rate": 6.25e-05, "loss": 3.8726, "step": 13 }, { "epoch": 8.0, "grad_norm": 0.3536747992038727, "learning_rate": 5e-05, "loss": 3.8848, "step": 14 }, { "epoch": 8.57, "grad_norm": 0.42837923765182495, "learning_rate": 3.7500000000000003e-05, "loss": 3.8312, "step": 15 }, { "epoch": 9.14, "grad_norm": 0.3861906826496124, "learning_rate": 2.5e-05, "loss": 3.9284, "step": 16 }, { "epoch": 9.71, "grad_norm": 0.3450121581554413, "learning_rate": 1.25e-05, "loss": 3.8218, "step": 17 }, { "epoch": 10.29, "grad_norm": 0.3705593943595886, "learning_rate": 0.0, "loss": 3.726, "step": 18 } ], "logging_steps": 1, "max_steps": 18, "num_input_tokens_seen": 0, "num_train_epochs": 18, "save_steps": 500, "total_flos": 5730050919936000.0, "train_batch_size": 5, "trial_name": null, "trial_params": null }