{ "best_metric": null, "best_model_checkpoint": null, "epoch": 8.227067050596462, "eval_steps": 500, "global_step": 5000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.16454134101192924, "grad_norm": 0.8724656105041504, "learning_rate": 1e-05, "loss": 0.9845, "step": 100 }, { "epoch": 0.3290826820238585, "grad_norm": 0.6713359951972961, "learning_rate": 2e-05, "loss": 0.5833, "step": 200 }, { "epoch": 0.49362402303578773, "grad_norm": 1.1457316875457764, "learning_rate": 3e-05, "loss": 0.4109, "step": 300 }, { "epoch": 0.658165364047717, "grad_norm": 1.878250241279602, "learning_rate": 4e-05, "loss": 0.3576, "step": 400 }, { "epoch": 0.8227067050596463, "grad_norm": 1.0137649774551392, "learning_rate": 5e-05, "loss": 0.3388, "step": 500 }, { "epoch": 0.8227067050596463, "eval_loss": 0.29997578263282776, "eval_runtime": 2.1518, "eval_samples_per_second": 4.647, "eval_steps_per_second": 2.324, "step": 500 }, { "epoch": 0.9872480460715755, "grad_norm": 1.1368335485458374, "learning_rate": 4.888888888888889e-05, "loss": 0.3053, "step": 600 }, { "epoch": 1.1517893870835048, "grad_norm": 0.9668510556221008, "learning_rate": 4.7777777777777784e-05, "loss": 0.2896, "step": 700 }, { "epoch": 1.316330728095434, "grad_norm": 1.2195199728012085, "learning_rate": 4.666666666666667e-05, "loss": 0.2852, "step": 800 }, { "epoch": 1.4808720691073631, "grad_norm": 1.9592208862304688, "learning_rate": 4.555555555555556e-05, "loss": 0.2706, "step": 900 }, { "epoch": 1.6454134101192923, "grad_norm": 2.3532769680023193, "learning_rate": 4.4444444444444447e-05, "loss": 0.2645, "step": 1000 }, { "epoch": 1.6454134101192923, "eval_loss": 0.25353601574897766, "eval_runtime": 2.1517, "eval_samples_per_second": 4.647, "eval_steps_per_second": 2.324, "step": 1000 }, { "epoch": 1.8099547511312217, "grad_norm": 0.9459941983222961, "learning_rate": 4.3333333333333334e-05, "loss": 0.2723, "step": 1100 }, { "epoch": 1.974496092143151, "grad_norm": 4.532841682434082, "learning_rate": 4.222222222222222e-05, "loss": 0.2589, "step": 1200 }, { "epoch": 2.1390374331550803, "grad_norm": 1.5837737321853638, "learning_rate": 4.111111111111111e-05, "loss": 0.2544, "step": 1300 }, { "epoch": 2.3035787741670095, "grad_norm": 2.3430395126342773, "learning_rate": 4e-05, "loss": 0.2377, "step": 1400 }, { "epoch": 2.4681201151789387, "grad_norm": 0.8787763118743896, "learning_rate": 3.888888888888889e-05, "loss": 0.2407, "step": 1500 }, { "epoch": 2.4681201151789387, "eval_loss": 0.2262120544910431, "eval_runtime": 2.1241, "eval_samples_per_second": 4.708, "eval_steps_per_second": 2.354, "step": 1500 }, { "epoch": 2.632661456190868, "grad_norm": 2.3523459434509277, "learning_rate": 3.777777777777778e-05, "loss": 0.2321, "step": 1600 }, { "epoch": 2.797202797202797, "grad_norm": 1.0582653284072876, "learning_rate": 3.6666666666666666e-05, "loss": 0.2333, "step": 1700 }, { "epoch": 2.9617441382147263, "grad_norm": 2.3663830757141113, "learning_rate": 3.555555555555556e-05, "loss": 0.2292, "step": 1800 }, { "epoch": 3.126285479226656, "grad_norm": 1.1538413763046265, "learning_rate": 3.444444444444445e-05, "loss": 0.2162, "step": 1900 }, { "epoch": 3.290826820238585, "grad_norm": 3.5856103897094727, "learning_rate": 3.3333333333333335e-05, "loss": 0.2151, "step": 2000 }, { "epoch": 3.290826820238585, "eval_loss": 0.1963958442211151, "eval_runtime": 2.1323, "eval_samples_per_second": 4.69, "eval_steps_per_second": 2.345, "step": 2000 }, { "epoch": 3.4553681612505143, "grad_norm": 2.666325092315674, "learning_rate": 3.222222222222223e-05, "loss": 0.212, "step": 2100 }, { "epoch": 3.6199095022624435, "grad_norm": 1.245647668838501, "learning_rate": 3.111111111111111e-05, "loss": 0.2157, "step": 2200 }, { "epoch": 3.7844508432743726, "grad_norm": 2.961127281188965, "learning_rate": 3e-05, "loss": 0.2123, "step": 2300 }, { "epoch": 3.948992184286302, "grad_norm": 1.1635862588882446, "learning_rate": 2.8888888888888888e-05, "loss": 0.2044, "step": 2400 }, { "epoch": 4.113533525298231, "grad_norm": 5.296552658081055, "learning_rate": 2.777777777777778e-05, "loss": 0.2003, "step": 2500 }, { "epoch": 4.113533525298231, "eval_loss": 0.20144304633140564, "eval_runtime": 2.1228, "eval_samples_per_second": 4.711, "eval_steps_per_second": 2.355, "step": 2500 }, { "epoch": 4.278074866310161, "grad_norm": 3.240604877471924, "learning_rate": 2.6666666666666667e-05, "loss": 0.1918, "step": 2600 }, { "epoch": 4.442616207322089, "grad_norm": 4.324797630310059, "learning_rate": 2.5555555555555554e-05, "loss": 0.1954, "step": 2700 }, { "epoch": 4.607157548334019, "grad_norm": 2.8329825401306152, "learning_rate": 2.4444444444444445e-05, "loss": 0.1942, "step": 2800 }, { "epoch": 4.771698889345949, "grad_norm": 1.8872101306915283, "learning_rate": 2.3333333333333336e-05, "loss": 0.188, "step": 2900 }, { "epoch": 4.936240230357877, "grad_norm": 6.444405555725098, "learning_rate": 2.2222222222222223e-05, "loss": 0.1918, "step": 3000 }, { "epoch": 4.936240230357877, "eval_loss": 0.2167702168226242, "eval_runtime": 2.1187, "eval_samples_per_second": 4.72, "eval_steps_per_second": 2.36, "step": 3000 }, { "epoch": 5.100781571369807, "grad_norm": 3.5297648906707764, "learning_rate": 2.111111111111111e-05, "loss": 0.1844, "step": 3100 }, { "epoch": 5.265322912381736, "grad_norm": 1.4769492149353027, "learning_rate": 2e-05, "loss": 0.174, "step": 3200 }, { "epoch": 5.429864253393665, "grad_norm": 5.627166748046875, "learning_rate": 1.888888888888889e-05, "loss": 0.1773, "step": 3300 }, { "epoch": 5.594405594405594, "grad_norm": 1.1272560358047485, "learning_rate": 1.777777777777778e-05, "loss": 0.1836, "step": 3400 }, { "epoch": 5.758946935417524, "grad_norm": 0.9033811092376709, "learning_rate": 1.6666666666666667e-05, "loss": 0.1823, "step": 3500 }, { "epoch": 5.758946935417524, "eval_loss": 0.2007795125246048, "eval_runtime": 2.1257, "eval_samples_per_second": 4.704, "eval_steps_per_second": 2.352, "step": 3500 }, { "epoch": 5.9234882764294525, "grad_norm": 1.0827577114105225, "learning_rate": 1.5555555555555555e-05, "loss": 0.1784, "step": 3600 }, { "epoch": 6.088029617441382, "grad_norm": 3.107731819152832, "learning_rate": 1.4444444444444444e-05, "loss": 0.1762, "step": 3700 }, { "epoch": 6.252570958453312, "grad_norm": 1.4182140827178955, "learning_rate": 1.3333333333333333e-05, "loss": 0.1711, "step": 3800 }, { "epoch": 6.4171122994652405, "grad_norm": 1.3971117734909058, "learning_rate": 1.2222222222222222e-05, "loss": 0.1679, "step": 3900 }, { "epoch": 6.58165364047717, "grad_norm": 3.196338415145874, "learning_rate": 1.1111111111111112e-05, "loss": 0.1663, "step": 4000 }, { "epoch": 6.58165364047717, "eval_loss": 0.19463543593883514, "eval_runtime": 2.135, "eval_samples_per_second": 4.684, "eval_steps_per_second": 2.342, "step": 4000 }, { "epoch": 6.746194981489099, "grad_norm": 1.7863503694534302, "learning_rate": 1e-05, "loss": 0.1672, "step": 4100 }, { "epoch": 6.9107363225010285, "grad_norm": 2.310948610305786, "learning_rate": 8.88888888888889e-06, "loss": 0.1704, "step": 4200 }, { "epoch": 7.075277663512957, "grad_norm": 2.464998722076416, "learning_rate": 7.777777777777777e-06, "loss": 0.1625, "step": 4300 }, { "epoch": 7.239819004524887, "grad_norm": 2.8195559978485107, "learning_rate": 6.666666666666667e-06, "loss": 0.1569, "step": 4400 }, { "epoch": 7.404360345536816, "grad_norm": 1.7469302415847778, "learning_rate": 5.555555555555556e-06, "loss": 0.1568, "step": 4500 }, { "epoch": 7.404360345536816, "eval_loss": 0.18890294432640076, "eval_runtime": 2.1046, "eval_samples_per_second": 4.752, "eval_steps_per_second": 2.376, "step": 4500 }, { "epoch": 7.568901686548745, "grad_norm": 1.7718433141708374, "learning_rate": 4.444444444444445e-06, "loss": 0.1559, "step": 4600 }, { "epoch": 7.733443027560675, "grad_norm": 0.978532075881958, "learning_rate": 3.3333333333333333e-06, "loss": 0.1604, "step": 4700 }, { "epoch": 7.897984368572604, "grad_norm": 1.3090927600860596, "learning_rate": 2.2222222222222225e-06, "loss": 0.1649, "step": 4800 }, { "epoch": 8.062525709584532, "grad_norm": 1.2589505910873413, "learning_rate": 1.1111111111111112e-06, "loss": 0.1581, "step": 4900 }, { "epoch": 8.227067050596462, "grad_norm": 1.417422890663147, "learning_rate": 0.0, "loss": 0.159, "step": 5000 }, { "epoch": 8.227067050596462, "eval_loss": 0.19616171717643738, "eval_runtime": 2.1191, "eval_samples_per_second": 4.719, "eval_steps_per_second": 2.359, "step": 5000 } ], "logging_steps": 100, "max_steps": 5000, "num_input_tokens_seen": 0, "num_train_epochs": 9, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.653061211756956e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }