{ "best_metric": 0.8550169467926025, "best_model_checkpoint": "./beans_outputs/checkpoint-1950", "epoch": 15.0, "eval_steps": 500, "global_step": 1950, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.07692307692307693, "grad_norm": 2.0955963134765625, "learning_rate": 1.9692307692307696e-05, "loss": 1.1239, "step": 10 }, { "epoch": 0.15384615384615385, "grad_norm": 1.8236017227172852, "learning_rate": 1.9384615384615386e-05, "loss": 1.1222, "step": 20 }, { "epoch": 0.23076923076923078, "grad_norm": 1.9866633415222168, "learning_rate": 1.907692307692308e-05, "loss": 1.1165, "step": 30 }, { "epoch": 0.3076923076923077, "grad_norm": 2.4556338787078857, "learning_rate": 1.876923076923077e-05, "loss": 1.1047, "step": 40 }, { "epoch": 0.38461538461538464, "grad_norm": 1.6562598943710327, "learning_rate": 1.8461538461538465e-05, "loss": 1.1085, "step": 50 }, { "epoch": 0.46153846153846156, "grad_norm": 1.501336932182312, "learning_rate": 1.8153846153846155e-05, "loss": 1.1048, "step": 60 }, { "epoch": 0.5384615384615384, "grad_norm": 2.606107234954834, "learning_rate": 1.784615384615385e-05, "loss": 1.1031, "step": 70 }, { "epoch": 0.6153846153846154, "grad_norm": 2.634584426879883, "learning_rate": 1.753846153846154e-05, "loss": 1.0916, "step": 80 }, { "epoch": 0.6923076923076923, "grad_norm": 2.039886236190796, "learning_rate": 1.7230769230769234e-05, "loss": 1.095, "step": 90 }, { "epoch": 0.7692307692307693, "grad_norm": 2.4800949096679688, "learning_rate": 1.6923076923076924e-05, "loss": 1.0916, "step": 100 }, { "epoch": 0.8461538461538461, "grad_norm": 1.6766040325164795, "learning_rate": 1.6615384615384618e-05, "loss": 1.0959, "step": 110 }, { "epoch": 0.9230769230769231, "grad_norm": 2.2848353385925293, "learning_rate": 1.630769230769231e-05, "loss": 1.1005, "step": 120 }, { "epoch": 1.0, "grad_norm": 4.07199239730835, "learning_rate": 1.6000000000000003e-05, "loss": 1.0881, "step": 130 }, { "epoch": 1.0, "eval_accuracy": 0.41353383458646614, "eval_loss": 1.0901767015457153, "eval_runtime": 0.7894, "eval_samples_per_second": 168.48, "eval_steps_per_second": 21.535, "step": 130 }, { "epoch": 1.0769230769230769, "grad_norm": 2.0711560249328613, "learning_rate": 1.5692307692307693e-05, "loss": 1.0794, "step": 140 }, { "epoch": 1.1538461538461537, "grad_norm": 2.499732494354248, "learning_rate": 1.5384615384615387e-05, "loss": 1.0755, "step": 150 }, { "epoch": 1.2307692307692308, "grad_norm": 1.4569323062896729, "learning_rate": 1.5076923076923078e-05, "loss": 1.0807, "step": 160 }, { "epoch": 1.3076923076923077, "grad_norm": 2.351478338241577, "learning_rate": 1.4769230769230772e-05, "loss": 1.0965, "step": 170 }, { "epoch": 1.3846153846153846, "grad_norm": 2.1514322757720947, "learning_rate": 1.4461538461538462e-05, "loss": 1.0839, "step": 180 }, { "epoch": 1.4615384615384617, "grad_norm": 2.151601791381836, "learning_rate": 1.4153846153846156e-05, "loss": 1.0837, "step": 190 }, { "epoch": 1.5384615384615383, "grad_norm": 1.797500729560852, "learning_rate": 1.3846153846153847e-05, "loss": 1.0807, "step": 200 }, { "epoch": 1.6153846153846154, "grad_norm": 1.7583892345428467, "learning_rate": 1.353846153846154e-05, "loss": 1.073, "step": 210 }, { "epoch": 1.6923076923076923, "grad_norm": 3.5006496906280518, "learning_rate": 1.3230769230769231e-05, "loss": 1.0741, "step": 220 }, { "epoch": 1.7692307692307692, "grad_norm": 2.0179672241210938, "learning_rate": 1.2923076923076925e-05, "loss": 1.077, "step": 230 }, { "epoch": 1.8461538461538463, "grad_norm": 2.057086944580078, "learning_rate": 1.2615384615384616e-05, "loss": 1.0925, "step": 240 }, { "epoch": 1.9230769230769231, "grad_norm": 1.9710103273391724, "learning_rate": 1.230769230769231e-05, "loss": 1.0757, "step": 250 }, { "epoch": 2.0, "grad_norm": 2.9315028190612793, "learning_rate": 1.2e-05, "loss": 1.0716, "step": 260 }, { "epoch": 2.0, "eval_accuracy": 0.5037593984962406, "eval_loss": 1.0685255527496338, "eval_runtime": 0.7711, "eval_samples_per_second": 172.471, "eval_steps_per_second": 22.045, "step": 260 }, { "epoch": 2.076923076923077, "grad_norm": 2.183527946472168, "learning_rate": 1.1692307692307694e-05, "loss": 1.0786, "step": 270 }, { "epoch": 2.1538461538461537, "grad_norm": 2.379652500152588, "learning_rate": 1.1384615384615385e-05, "loss": 1.0676, "step": 280 }, { "epoch": 2.230769230769231, "grad_norm": 2.387296438217163, "learning_rate": 1.1076923076923079e-05, "loss": 1.064, "step": 290 }, { "epoch": 2.3076923076923075, "grad_norm": 2.8164947032928467, "learning_rate": 1.076923076923077e-05, "loss": 1.0548, "step": 300 }, { "epoch": 2.3846153846153846, "grad_norm": 1.849363088607788, "learning_rate": 1.0461538461538463e-05, "loss": 1.0693, "step": 310 }, { "epoch": 2.4615384615384617, "grad_norm": 2.0274300575256348, "learning_rate": 1.0153846153846154e-05, "loss": 1.0604, "step": 320 }, { "epoch": 2.5384615384615383, "grad_norm": 1.4544223546981812, "learning_rate": 9.846153846153848e-06, "loss": 1.0458, "step": 330 }, { "epoch": 2.6153846153846154, "grad_norm": 2.1712183952331543, "learning_rate": 9.53846153846154e-06, "loss": 1.0609, "step": 340 }, { "epoch": 2.6923076923076925, "grad_norm": 1.922677993774414, "learning_rate": 9.230769230769232e-06, "loss": 1.0646, "step": 350 }, { "epoch": 2.769230769230769, "grad_norm": 1.8288860321044922, "learning_rate": 8.923076923076925e-06, "loss": 1.0597, "step": 360 }, { "epoch": 2.8461538461538463, "grad_norm": 2.123480796813965, "learning_rate": 8.615384615384617e-06, "loss": 1.052, "step": 370 }, { "epoch": 2.9230769230769234, "grad_norm": 1.820168137550354, "learning_rate": 8.307692307692309e-06, "loss": 1.0434, "step": 380 }, { "epoch": 3.0, "grad_norm": 4.973505020141602, "learning_rate": 8.000000000000001e-06, "loss": 1.061, "step": 390 }, { "epoch": 3.0, "eval_accuracy": 0.6240601503759399, "eval_loss": 1.0459144115447998, "eval_runtime": 0.7618, "eval_samples_per_second": 174.588, "eval_steps_per_second": 22.316, "step": 390 }, { "epoch": 3.076923076923077, "grad_norm": 2.1673223972320557, "learning_rate": 7.692307692307694e-06, "loss": 1.0616, "step": 400 }, { "epoch": 3.1538461538461537, "grad_norm": 1.8567888736724854, "learning_rate": 7.384615384615386e-06, "loss": 1.0501, "step": 410 }, { "epoch": 3.230769230769231, "grad_norm": 2.0640571117401123, "learning_rate": 7.076923076923078e-06, "loss": 1.0611, "step": 420 }, { "epoch": 3.3076923076923075, "grad_norm": 2.215384006500244, "learning_rate": 6.76923076923077e-06, "loss": 1.0469, "step": 430 }, { "epoch": 3.3846153846153846, "grad_norm": 2.1000049114227295, "learning_rate": 6.461538461538463e-06, "loss": 1.0499, "step": 440 }, { "epoch": 3.4615384615384617, "grad_norm": 1.7218382358551025, "learning_rate": 6.153846153846155e-06, "loss": 1.0564, "step": 450 }, { "epoch": 3.5384615384615383, "grad_norm": 2.3569300174713135, "learning_rate": 5.846153846153847e-06, "loss": 1.0599, "step": 460 }, { "epoch": 3.6153846153846154, "grad_norm": 1.5210909843444824, "learning_rate": 5.538461538461539e-06, "loss": 1.0367, "step": 470 }, { "epoch": 3.6923076923076925, "grad_norm": 2.7621657848358154, "learning_rate": 5.230769230769232e-06, "loss": 1.0421, "step": 480 }, { "epoch": 3.769230769230769, "grad_norm": 1.5097808837890625, "learning_rate": 4.923076923076924e-06, "loss": 1.0362, "step": 490 }, { "epoch": 3.8461538461538463, "grad_norm": 1.5118447542190552, "learning_rate": 4.615384615384616e-06, "loss": 1.0572, "step": 500 }, { "epoch": 3.9230769230769234, "grad_norm": 1.7513490915298462, "learning_rate": 4.307692307692308e-06, "loss": 1.0361, "step": 510 }, { "epoch": 4.0, "grad_norm": 5.398025035858154, "learning_rate": 4.000000000000001e-06, "loss": 1.0514, "step": 520 }, { "epoch": 4.0, "eval_accuracy": 0.6015037593984962, "eval_loss": 1.0407124757766724, "eval_runtime": 0.7726, "eval_samples_per_second": 172.155, "eval_steps_per_second": 22.005, "step": 520 }, { "epoch": 4.076923076923077, "grad_norm": 2.5516345500946045, "learning_rate": 3.692307692307693e-06, "loss": 1.0529, "step": 530 }, { "epoch": 4.153846153846154, "grad_norm": 1.6976008415222168, "learning_rate": 3.384615384615385e-06, "loss": 1.0472, "step": 540 }, { "epoch": 4.230769230769231, "grad_norm": 2.5672519207000732, "learning_rate": 3.0769230769230774e-06, "loss": 1.0565, "step": 550 }, { "epoch": 4.3076923076923075, "grad_norm": 2.166529655456543, "learning_rate": 2.7692307692307697e-06, "loss": 1.0619, "step": 560 }, { "epoch": 4.384615384615385, "grad_norm": 1.961472511291504, "learning_rate": 2.461538461538462e-06, "loss": 1.0322, "step": 570 }, { "epoch": 4.461538461538462, "grad_norm": 2.392319440841675, "learning_rate": 2.153846153846154e-06, "loss": 1.0388, "step": 580 }, { "epoch": 4.538461538461538, "grad_norm": 2.3034205436706543, "learning_rate": 1.8461538461538465e-06, "loss": 1.0358, "step": 590 }, { "epoch": 4.615384615384615, "grad_norm": 2.037050247192383, "learning_rate": 1.5384615384615387e-06, "loss": 1.0334, "step": 600 }, { "epoch": 4.6923076923076925, "grad_norm": 3.0737335681915283, "learning_rate": 1.230769230769231e-06, "loss": 1.0506, "step": 610 }, { "epoch": 4.769230769230769, "grad_norm": 2.1824796199798584, "learning_rate": 9.230769230769232e-07, "loss": 1.0516, "step": 620 }, { "epoch": 4.846153846153846, "grad_norm": 1.9239214658737183, "learning_rate": 6.153846153846155e-07, "loss": 1.0399, "step": 630 }, { "epoch": 4.923076923076923, "grad_norm": 2.267302989959717, "learning_rate": 3.0769230769230774e-07, "loss": 1.0374, "step": 640 }, { "epoch": 5.0, "grad_norm": 4.4957404136657715, "learning_rate": 0.0, "loss": 1.05, "step": 650 }, { "epoch": 5.0, "eval_accuracy": 0.6766917293233082, "eval_loss": 1.0332472324371338, "eval_runtime": 0.8208, "eval_samples_per_second": 162.035, "eval_steps_per_second": 20.711, "step": 650 }, { "epoch": 5.076923076923077, "grad_norm": 1.9310355186462402, "learning_rate": 1.3230769230769231e-05, "loss": 1.0281, "step": 660 }, { "epoch": 5.153846153846154, "grad_norm": 2.36603045463562, "learning_rate": 1.312820512820513e-05, "loss": 1.0587, "step": 670 }, { "epoch": 5.230769230769231, "grad_norm": 2.0758917331695557, "learning_rate": 1.3025641025641027e-05, "loss": 1.0534, "step": 680 }, { "epoch": 5.3076923076923075, "grad_norm": 2.549725294113159, "learning_rate": 1.2923076923076925e-05, "loss": 1.0421, "step": 690 }, { "epoch": 5.384615384615385, "grad_norm": 2.7183680534362793, "learning_rate": 1.2820512820512823e-05, "loss": 1.035, "step": 700 }, { "epoch": 5.461538461538462, "grad_norm": 1.7176955938339233, "learning_rate": 1.2717948717948718e-05, "loss": 1.0268, "step": 710 }, { "epoch": 5.538461538461538, "grad_norm": 1.8257861137390137, "learning_rate": 1.2615384615384616e-05, "loss": 1.0302, "step": 720 }, { "epoch": 5.615384615384615, "grad_norm": 2.50368595123291, "learning_rate": 1.2512820512820514e-05, "loss": 1.0331, "step": 730 }, { "epoch": 5.6923076923076925, "grad_norm": 2.4315121173858643, "learning_rate": 1.2410256410256412e-05, "loss": 1.0208, "step": 740 }, { "epoch": 5.769230769230769, "grad_norm": 2.043854236602783, "learning_rate": 1.230769230769231e-05, "loss": 1.0241, "step": 750 }, { "epoch": 5.846153846153846, "grad_norm": 2.0800740718841553, "learning_rate": 1.2205128205128208e-05, "loss": 1.0251, "step": 760 }, { "epoch": 5.923076923076923, "grad_norm": 3.4356396198272705, "learning_rate": 1.2102564102564102e-05, "loss": 1.0104, "step": 770 }, { "epoch": 6.0, "grad_norm": 4.49416971206665, "learning_rate": 1.2e-05, "loss": 1.0357, "step": 780 }, { "epoch": 6.0, "eval_accuracy": 0.6541353383458647, "eval_loss": 1.0109117031097412, "eval_runtime": 0.7801, "eval_samples_per_second": 170.483, "eval_steps_per_second": 21.791, "step": 780 }, { "epoch": 6.076923076923077, "grad_norm": 2.650513172149658, "learning_rate": 1.1897435897435898e-05, "loss": 1.0148, "step": 790 }, { "epoch": 6.153846153846154, "grad_norm": 2.3682632446289062, "learning_rate": 1.1794871794871796e-05, "loss": 1.0094, "step": 800 }, { "epoch": 6.230769230769231, "grad_norm": 1.6716077327728271, "learning_rate": 1.1692307692307694e-05, "loss": 1.0097, "step": 810 }, { "epoch": 6.3076923076923075, "grad_norm": 2.4839890003204346, "learning_rate": 1.1589743589743592e-05, "loss": 1.01, "step": 820 }, { "epoch": 6.384615384615385, "grad_norm": 2.4004769325256348, "learning_rate": 1.1487179487179487e-05, "loss": 1.0104, "step": 830 }, { "epoch": 6.461538461538462, "grad_norm": 2.9597084522247314, "learning_rate": 1.1384615384615385e-05, "loss": 1.0137, "step": 840 }, { "epoch": 6.538461538461538, "grad_norm": 2.680335760116577, "learning_rate": 1.1282051282051283e-05, "loss": 1.0147, "step": 850 }, { "epoch": 6.615384615384615, "grad_norm": 1.7677160501480103, "learning_rate": 1.117948717948718e-05, "loss": 0.9947, "step": 860 }, { "epoch": 6.6923076923076925, "grad_norm": 2.0484132766723633, "learning_rate": 1.1076923076923079e-05, "loss": 1.0169, "step": 870 }, { "epoch": 6.769230769230769, "grad_norm": 2.1910479068756104, "learning_rate": 1.0974358974358977e-05, "loss": 1.0024, "step": 880 }, { "epoch": 6.846153846153846, "grad_norm": 2.181236743927002, "learning_rate": 1.0871794871794871e-05, "loss": 0.9962, "step": 890 }, { "epoch": 6.923076923076923, "grad_norm": 2.898885488510132, "learning_rate": 1.076923076923077e-05, "loss": 1.012, "step": 900 }, { "epoch": 7.0, "grad_norm": 4.950052738189697, "learning_rate": 1.0666666666666667e-05, "loss": 1.0012, "step": 910 }, { "epoch": 7.0, "eval_accuracy": 0.7368421052631579, "eval_loss": 0.981479287147522, "eval_runtime": 0.7693, "eval_samples_per_second": 172.886, "eval_steps_per_second": 22.098, "step": 910 }, { "epoch": 7.076923076923077, "grad_norm": 2.692753553390503, "learning_rate": 1.0564102564102565e-05, "loss": 0.9889, "step": 920 }, { "epoch": 7.153846153846154, "grad_norm": 2.9175124168395996, "learning_rate": 1.0461538461538463e-05, "loss": 0.9911, "step": 930 }, { "epoch": 7.230769230769231, "grad_norm": 3.221527099609375, "learning_rate": 1.0358974358974361e-05, "loss": 0.9827, "step": 940 }, { "epoch": 7.3076923076923075, "grad_norm": 2.507923126220703, "learning_rate": 1.0256410256410256e-05, "loss": 0.9919, "step": 950 }, { "epoch": 7.384615384615385, "grad_norm": 2.4533870220184326, "learning_rate": 1.0153846153846154e-05, "loss": 0.9962, "step": 960 }, { "epoch": 7.461538461538462, "grad_norm": 2.1032631397247314, "learning_rate": 1.0051282051282052e-05, "loss": 0.9751, "step": 970 }, { "epoch": 7.538461538461538, "grad_norm": 2.5848186016082764, "learning_rate": 9.94871794871795e-06, "loss": 0.9939, "step": 980 }, { "epoch": 7.615384615384615, "grad_norm": 2.17742919921875, "learning_rate": 9.846153846153848e-06, "loss": 0.9745, "step": 990 }, { "epoch": 7.6923076923076925, "grad_norm": 1.9953967332839966, "learning_rate": 9.743589743589744e-06, "loss": 0.9665, "step": 1000 }, { "epoch": 7.769230769230769, "grad_norm": 3.0263218879699707, "learning_rate": 9.641025641025642e-06, "loss": 0.9646, "step": 1010 }, { "epoch": 7.846153846153846, "grad_norm": 2.3735406398773193, "learning_rate": 9.53846153846154e-06, "loss": 0.9836, "step": 1020 }, { "epoch": 7.923076923076923, "grad_norm": 2.548480272293091, "learning_rate": 9.435897435897436e-06, "loss": 0.9546, "step": 1030 }, { "epoch": 8.0, "grad_norm": 3.8450028896331787, "learning_rate": 9.333333333333334e-06, "loss": 0.9932, "step": 1040 }, { "epoch": 8.0, "eval_accuracy": 0.7669172932330827, "eval_loss": 0.9549766778945923, "eval_runtime": 0.7536, "eval_samples_per_second": 176.486, "eval_steps_per_second": 22.558, "step": 1040 }, { "epoch": 8.076923076923077, "grad_norm": 1.6870847940444946, "learning_rate": 9.230769230769232e-06, "loss": 0.9752, "step": 1050 }, { "epoch": 8.153846153846153, "grad_norm": 2.1122217178344727, "learning_rate": 9.128205128205129e-06, "loss": 0.9573, "step": 1060 }, { "epoch": 8.23076923076923, "grad_norm": 4.1552886962890625, "learning_rate": 9.025641025641027e-06, "loss": 0.9764, "step": 1070 }, { "epoch": 8.307692307692308, "grad_norm": 1.7864203453063965, "learning_rate": 8.923076923076925e-06, "loss": 0.9434, "step": 1080 }, { "epoch": 8.384615384615385, "grad_norm": 2.2091946601867676, "learning_rate": 8.820512820512821e-06, "loss": 0.974, "step": 1090 }, { "epoch": 8.461538461538462, "grad_norm": 2.4063644409179688, "learning_rate": 8.717948717948719e-06, "loss": 0.9576, "step": 1100 }, { "epoch": 8.538461538461538, "grad_norm": 1.6061931848526, "learning_rate": 8.615384615384617e-06, "loss": 0.9588, "step": 1110 }, { "epoch": 8.615384615384615, "grad_norm": 2.8999595642089844, "learning_rate": 8.512820512820513e-06, "loss": 0.9791, "step": 1120 }, { "epoch": 8.692307692307692, "grad_norm": 3.6554131507873535, "learning_rate": 8.410256410256411e-06, "loss": 0.9629, "step": 1130 }, { "epoch": 8.76923076923077, "grad_norm": 1.7246966361999512, "learning_rate": 8.307692307692309e-06, "loss": 0.9707, "step": 1140 }, { "epoch": 8.846153846153847, "grad_norm": 2.3160033226013184, "learning_rate": 8.205128205128205e-06, "loss": 0.9578, "step": 1150 }, { "epoch": 8.923076923076923, "grad_norm": 2.685718059539795, "learning_rate": 8.102564102564103e-06, "loss": 0.972, "step": 1160 }, { "epoch": 9.0, "grad_norm": 3.6465442180633545, "learning_rate": 8.000000000000001e-06, "loss": 0.9748, "step": 1170 }, { "epoch": 9.0, "eval_accuracy": 0.7669172932330827, "eval_loss": 0.9408761858940125, "eval_runtime": 0.7615, "eval_samples_per_second": 174.644, "eval_steps_per_second": 22.323, "step": 1170 }, { "epoch": 9.076923076923077, "grad_norm": 1.8668149709701538, "learning_rate": 7.897435897435898e-06, "loss": 0.9712, "step": 1180 }, { "epoch": 9.153846153846153, "grad_norm": 2.042644739151001, "learning_rate": 7.794871794871796e-06, "loss": 0.9407, "step": 1190 }, { "epoch": 9.23076923076923, "grad_norm": 1.967020869255066, "learning_rate": 7.692307692307694e-06, "loss": 0.9457, "step": 1200 }, { "epoch": 9.307692307692308, "grad_norm": 2.147862672805786, "learning_rate": 7.58974358974359e-06, "loss": 0.9442, "step": 1210 }, { "epoch": 9.384615384615385, "grad_norm": 1.8528053760528564, "learning_rate": 7.487179487179488e-06, "loss": 0.9526, "step": 1220 }, { "epoch": 9.461538461538462, "grad_norm": 3.2000551223754883, "learning_rate": 7.384615384615386e-06, "loss": 0.9465, "step": 1230 }, { "epoch": 9.538461538461538, "grad_norm": 2.259323835372925, "learning_rate": 7.282051282051282e-06, "loss": 0.9503, "step": 1240 }, { "epoch": 9.615384615384615, "grad_norm": 2.4054858684539795, "learning_rate": 7.17948717948718e-06, "loss": 0.9274, "step": 1250 }, { "epoch": 9.692307692307692, "grad_norm": 3.4811408519744873, "learning_rate": 7.076923076923078e-06, "loss": 0.943, "step": 1260 }, { "epoch": 9.76923076923077, "grad_norm": 1.7080141305923462, "learning_rate": 6.974358974358974e-06, "loss": 0.9247, "step": 1270 }, { "epoch": 9.846153846153847, "grad_norm": 2.0476508140563965, "learning_rate": 6.871794871794872e-06, "loss": 0.9194, "step": 1280 }, { "epoch": 9.923076923076923, "grad_norm": 2.149641990661621, "learning_rate": 6.76923076923077e-06, "loss": 0.9269, "step": 1290 }, { "epoch": 10.0, "grad_norm": 5.121323108673096, "learning_rate": 6.666666666666667e-06, "loss": 0.9113, "step": 1300 }, { "epoch": 10.0, "eval_accuracy": 0.7819548872180451, "eval_loss": 0.9149269461631775, "eval_runtime": 0.7775, "eval_samples_per_second": 171.06, "eval_steps_per_second": 21.865, "step": 1300 }, { "epoch": 10.076923076923077, "grad_norm": 2.06109881401062, "learning_rate": 6.564102564102565e-06, "loss": 0.925, "step": 1310 }, { "epoch": 10.153846153846153, "grad_norm": 1.9137018918991089, "learning_rate": 6.461538461538463e-06, "loss": 0.9657, "step": 1320 }, { "epoch": 10.23076923076923, "grad_norm": 2.0686280727386475, "learning_rate": 6.358974358974359e-06, "loss": 0.9565, "step": 1330 }, { "epoch": 10.307692307692308, "grad_norm": 2.046623945236206, "learning_rate": 6.256410256410257e-06, "loss": 0.918, "step": 1340 }, { "epoch": 10.384615384615385, "grad_norm": 2.281343936920166, "learning_rate": 6.153846153846155e-06, "loss": 0.9118, "step": 1350 }, { "epoch": 10.461538461538462, "grad_norm": 2.694427728652954, "learning_rate": 6.051282051282051e-06, "loss": 0.9377, "step": 1360 }, { "epoch": 10.538461538461538, "grad_norm": 2.3148765563964844, "learning_rate": 5.948717948717949e-06, "loss": 0.911, "step": 1370 }, { "epoch": 10.615384615384615, "grad_norm": 2.595669746398926, "learning_rate": 5.846153846153847e-06, "loss": 0.9146, "step": 1380 }, { "epoch": 10.692307692307692, "grad_norm": 2.136301279067993, "learning_rate": 5.743589743589743e-06, "loss": 0.9061, "step": 1390 }, { "epoch": 10.76923076923077, "grad_norm": 3.0159363746643066, "learning_rate": 5.641025641025641e-06, "loss": 0.9365, "step": 1400 }, { "epoch": 10.846153846153847, "grad_norm": 2.1136507987976074, "learning_rate": 5.538461538461539e-06, "loss": 0.9232, "step": 1410 }, { "epoch": 10.923076923076923, "grad_norm": 1.713663101196289, "learning_rate": 5.435897435897436e-06, "loss": 0.9344, "step": 1420 }, { "epoch": 11.0, "grad_norm": 4.04538631439209, "learning_rate": 5.333333333333334e-06, "loss": 0.9255, "step": 1430 }, { "epoch": 11.0, "eval_accuracy": 0.7894736842105263, "eval_loss": 0.8905543088912964, "eval_runtime": 0.7659, "eval_samples_per_second": 173.659, "eval_steps_per_second": 22.197, "step": 1430 }, { "epoch": 11.076923076923077, "grad_norm": 2.04194974899292, "learning_rate": 5.230769230769232e-06, "loss": 0.9333, "step": 1440 }, { "epoch": 11.153846153846153, "grad_norm": 3.108344554901123, "learning_rate": 5.128205128205128e-06, "loss": 0.9174, "step": 1450 }, { "epoch": 11.23076923076923, "grad_norm": 2.406233072280884, "learning_rate": 5.025641025641026e-06, "loss": 0.8948, "step": 1460 }, { "epoch": 11.307692307692308, "grad_norm": 2.4100501537323, "learning_rate": 4.923076923076924e-06, "loss": 0.9155, "step": 1470 }, { "epoch": 11.384615384615385, "grad_norm": 2.7117860317230225, "learning_rate": 4.820512820512821e-06, "loss": 0.9075, "step": 1480 }, { "epoch": 11.461538461538462, "grad_norm": 2.0159695148468018, "learning_rate": 4.717948717948718e-06, "loss": 0.9338, "step": 1490 }, { "epoch": 11.538461538461538, "grad_norm": 3.280245304107666, "learning_rate": 4.615384615384616e-06, "loss": 0.9243, "step": 1500 }, { "epoch": 11.615384615384615, "grad_norm": 3.1355690956115723, "learning_rate": 4.512820512820513e-06, "loss": 0.9185, "step": 1510 }, { "epoch": 11.692307692307692, "grad_norm": 3.0900094509124756, "learning_rate": 4.4102564102564104e-06, "loss": 0.937, "step": 1520 }, { "epoch": 11.76923076923077, "grad_norm": 1.8758033514022827, "learning_rate": 4.307692307692308e-06, "loss": 0.9052, "step": 1530 }, { "epoch": 11.846153846153847, "grad_norm": 2.0586955547332764, "learning_rate": 4.2051282051282055e-06, "loss": 0.8874, "step": 1540 }, { "epoch": 11.923076923076923, "grad_norm": 2.0720062255859375, "learning_rate": 4.102564102564103e-06, "loss": 0.9141, "step": 1550 }, { "epoch": 12.0, "grad_norm": 3.183523416519165, "learning_rate": 4.000000000000001e-06, "loss": 0.8877, "step": 1560 }, { "epoch": 12.0, "eval_accuracy": 0.7894736842105263, "eval_loss": 0.8749483823776245, "eval_runtime": 0.7374, "eval_samples_per_second": 180.372, "eval_steps_per_second": 23.055, "step": 1560 }, { "epoch": 12.076923076923077, "grad_norm": 2.0058720111846924, "learning_rate": 3.897435897435898e-06, "loss": 0.8829, "step": 1570 }, { "epoch": 12.153846153846153, "grad_norm": 2.2991676330566406, "learning_rate": 3.794871794871795e-06, "loss": 0.9152, "step": 1580 }, { "epoch": 12.23076923076923, "grad_norm": 1.5903538465499878, "learning_rate": 3.692307692307693e-06, "loss": 0.9149, "step": 1590 }, { "epoch": 12.307692307692308, "grad_norm": 1.7883615493774414, "learning_rate": 3.58974358974359e-06, "loss": 0.9163, "step": 1600 }, { "epoch": 12.384615384615385, "grad_norm": 2.2841601371765137, "learning_rate": 3.487179487179487e-06, "loss": 0.8958, "step": 1610 }, { "epoch": 12.461538461538462, "grad_norm": 2.3814501762390137, "learning_rate": 3.384615384615385e-06, "loss": 0.8918, "step": 1620 }, { "epoch": 12.538461538461538, "grad_norm": 1.9848734140396118, "learning_rate": 3.2820512820512823e-06, "loss": 0.889, "step": 1630 }, { "epoch": 12.615384615384615, "grad_norm": 1.7236778736114502, "learning_rate": 3.1794871794871795e-06, "loss": 0.8979, "step": 1640 }, { "epoch": 12.692307692307692, "grad_norm": 3.340665102005005, "learning_rate": 3.0769230769230774e-06, "loss": 0.8695, "step": 1650 }, { "epoch": 12.76923076923077, "grad_norm": 2.127927780151367, "learning_rate": 2.9743589743589746e-06, "loss": 0.9323, "step": 1660 }, { "epoch": 12.846153846153847, "grad_norm": 1.8213707208633423, "learning_rate": 2.8717948717948717e-06, "loss": 0.9178, "step": 1670 }, { "epoch": 12.923076923076923, "grad_norm": 2.0011963844299316, "learning_rate": 2.7692307692307697e-06, "loss": 0.8872, "step": 1680 }, { "epoch": 13.0, "grad_norm": 3.812871217727661, "learning_rate": 2.666666666666667e-06, "loss": 0.9032, "step": 1690 }, { "epoch": 13.0, "eval_accuracy": 0.7969924812030075, "eval_loss": 0.8698711395263672, "eval_runtime": 0.7423, "eval_samples_per_second": 179.165, "eval_steps_per_second": 22.901, "step": 1690 }, { "epoch": 13.076923076923077, "grad_norm": 2.8741540908813477, "learning_rate": 2.564102564102564e-06, "loss": 0.8842, "step": 1700 }, { "epoch": 13.153846153846153, "grad_norm": 2.3278818130493164, "learning_rate": 2.461538461538462e-06, "loss": 0.9131, "step": 1710 }, { "epoch": 13.23076923076923, "grad_norm": 2.8419501781463623, "learning_rate": 2.358974358974359e-06, "loss": 0.8965, "step": 1720 }, { "epoch": 13.307692307692308, "grad_norm": 1.8506221771240234, "learning_rate": 2.2564102564102566e-06, "loss": 0.8967, "step": 1730 }, { "epoch": 13.384615384615385, "grad_norm": 2.6166839599609375, "learning_rate": 2.153846153846154e-06, "loss": 0.8785, "step": 1740 }, { "epoch": 13.461538461538462, "grad_norm": 4.287515640258789, "learning_rate": 2.0512820512820513e-06, "loss": 0.914, "step": 1750 }, { "epoch": 13.538461538461538, "grad_norm": 2.516889810562134, "learning_rate": 1.948717948717949e-06, "loss": 0.9286, "step": 1760 }, { "epoch": 13.615384615384615, "grad_norm": 1.8332946300506592, "learning_rate": 1.8461538461538465e-06, "loss": 0.8995, "step": 1770 }, { "epoch": 13.692307692307692, "grad_norm": 2.2418551445007324, "learning_rate": 1.7435897435897436e-06, "loss": 0.8818, "step": 1780 }, { "epoch": 13.76923076923077, "grad_norm": 1.794832706451416, "learning_rate": 1.6410256410256412e-06, "loss": 0.9044, "step": 1790 }, { "epoch": 13.846153846153847, "grad_norm": 3.0142152309417725, "learning_rate": 1.5384615384615387e-06, "loss": 0.8826, "step": 1800 }, { "epoch": 13.923076923076923, "grad_norm": 2.5891315937042236, "learning_rate": 1.4358974358974359e-06, "loss": 0.8387, "step": 1810 }, { "epoch": 14.0, "grad_norm": 5.37412691116333, "learning_rate": 1.3333333333333334e-06, "loss": 0.9001, "step": 1820 }, { "epoch": 14.0, "eval_accuracy": 0.7819548872180451, "eval_loss": 0.8673797845840454, "eval_runtime": 0.7642, "eval_samples_per_second": 174.027, "eval_steps_per_second": 22.244, "step": 1820 }, { "epoch": 14.076923076923077, "grad_norm": 1.8213236331939697, "learning_rate": 1.230769230769231e-06, "loss": 0.9047, "step": 1830 }, { "epoch": 14.153846153846153, "grad_norm": 1.8006333112716675, "learning_rate": 1.1282051282051283e-06, "loss": 0.8768, "step": 1840 }, { "epoch": 14.23076923076923, "grad_norm": 2.691574811935425, "learning_rate": 1.0256410256410257e-06, "loss": 0.8757, "step": 1850 }, { "epoch": 14.307692307692308, "grad_norm": 5.015848636627197, "learning_rate": 9.230769230769232e-07, "loss": 0.8734, "step": 1860 }, { "epoch": 14.384615384615385, "grad_norm": 2.5233821868896484, "learning_rate": 8.205128205128206e-07, "loss": 0.8787, "step": 1870 }, { "epoch": 14.461538461538462, "grad_norm": 2.1718924045562744, "learning_rate": 7.179487179487179e-07, "loss": 0.8767, "step": 1880 }, { "epoch": 14.538461538461538, "grad_norm": 3.0364015102386475, "learning_rate": 6.153846153846155e-07, "loss": 0.873, "step": 1890 }, { "epoch": 14.615384615384615, "grad_norm": 2.5152034759521484, "learning_rate": 5.128205128205128e-07, "loss": 0.9096, "step": 1900 }, { "epoch": 14.692307692307692, "grad_norm": 1.819096565246582, "learning_rate": 4.102564102564103e-07, "loss": 0.892, "step": 1910 }, { "epoch": 14.76923076923077, "grad_norm": 3.512732982635498, "learning_rate": 3.0769230769230774e-07, "loss": 0.8937, "step": 1920 }, { "epoch": 14.846153846153847, "grad_norm": 2.917677879333496, "learning_rate": 2.0512820512820514e-07, "loss": 0.9231, "step": 1930 }, { "epoch": 14.923076923076923, "grad_norm": 2.0683395862579346, "learning_rate": 1.0256410256410257e-07, "loss": 0.8613, "step": 1940 }, { "epoch": 15.0, "grad_norm": 4.704519271850586, "learning_rate": 0.0, "loss": 0.8842, "step": 1950 }, { "epoch": 15.0, "eval_accuracy": 0.7894736842105263, "eval_loss": 0.8550169467926025, "eval_runtime": 0.7718, "eval_samples_per_second": 172.329, "eval_steps_per_second": 22.027, "step": 1950 }, { "epoch": 15.0, "step": 1950, "total_flos": 1.5658365504595968e+17, "train_loss": 0.6299933981284117, "train_runtime": 92.772, "train_samples_per_second": 167.184, "train_steps_per_second": 21.019 } ], "logging_steps": 10, "max_steps": 1950, "num_input_tokens_seen": 0, "num_train_epochs": 15, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.5658365504595968e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }