|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 735, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.04081632653061224, |
|
"grad_norm": 20286.314453125, |
|
"learning_rate": 1.2244897959183673e-06, |
|
"loss": 2.0797, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.08163265306122448, |
|
"grad_norm": 21226.833984375, |
|
"learning_rate": 2.4489795918367347e-06, |
|
"loss": 2.0794, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.12244897959183673, |
|
"grad_norm": 20473.34375, |
|
"learning_rate": 3.673469387755102e-06, |
|
"loss": 2.0791, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.16326530612244897, |
|
"grad_norm": 21295.623046875, |
|
"learning_rate": 4.897959183673469e-06, |
|
"loss": 2.0786, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.20408163265306123, |
|
"grad_norm": 22374.26953125, |
|
"learning_rate": 6.122448979591837e-06, |
|
"loss": 2.0775, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.24489795918367346, |
|
"grad_norm": 27922.837890625, |
|
"learning_rate": 7.346938775510204e-06, |
|
"loss": 2.0762, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.2857142857142857, |
|
"grad_norm": 21962.65625, |
|
"learning_rate": 8.571428571428571e-06, |
|
"loss": 2.0751, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.32653061224489793, |
|
"grad_norm": 25949.669921875, |
|
"learning_rate": 9.795918367346939e-06, |
|
"loss": 2.073, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.3673469387755102, |
|
"grad_norm": 21337.779296875, |
|
"learning_rate": 1.1020408163265306e-05, |
|
"loss": 2.0711, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.40816326530612246, |
|
"grad_norm": 21139.53125, |
|
"learning_rate": 1.2244897959183674e-05, |
|
"loss": 2.068, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.4489795918367347, |
|
"grad_norm": 18687.326171875, |
|
"learning_rate": 1.3469387755102042e-05, |
|
"loss": 2.0621, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.4897959183673469, |
|
"grad_norm": 23997.892578125, |
|
"learning_rate": 1.4693877551020408e-05, |
|
"loss": 2.0557, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.5306122448979592, |
|
"grad_norm": 25094.125, |
|
"learning_rate": 1.5918367346938776e-05, |
|
"loss": 2.0436, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.5714285714285714, |
|
"grad_norm": 35127.62890625, |
|
"learning_rate": 1.7142857142857142e-05, |
|
"loss": 2.0313, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.6122448979591837, |
|
"grad_norm": 38480.73046875, |
|
"learning_rate": 1.836734693877551e-05, |
|
"loss": 2.0124, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.6530612244897959, |
|
"grad_norm": 37025.25390625, |
|
"learning_rate": 1.9591836734693877e-05, |
|
"loss": 1.9832, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.6938775510204082, |
|
"grad_norm": 47326.75390625, |
|
"learning_rate": 2.0816326530612247e-05, |
|
"loss": 1.9503, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.7346938775510204, |
|
"grad_norm": 44386.79296875, |
|
"learning_rate": 2.2040816326530613e-05, |
|
"loss": 1.9033, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.7755102040816326, |
|
"grad_norm": 60427.28515625, |
|
"learning_rate": 2.326530612244898e-05, |
|
"loss": 1.8548, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.8163265306122449, |
|
"grad_norm": 60310.234375, |
|
"learning_rate": 2.448979591836735e-05, |
|
"loss": 1.7809, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.8571428571428571, |
|
"grad_norm": 73581.171875, |
|
"learning_rate": 2.5714285714285714e-05, |
|
"loss": 1.8435, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.8979591836734694, |
|
"grad_norm": 63305.5, |
|
"learning_rate": 2.6938775510204084e-05, |
|
"loss": 1.7222, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.9387755102040817, |
|
"grad_norm": 78695.546875, |
|
"learning_rate": 2.816326530612245e-05, |
|
"loss": 1.7823, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.9795918367346939, |
|
"grad_norm": 118529.84375, |
|
"learning_rate": 2.9387755102040816e-05, |
|
"loss": 1.7203, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.0204081632653061, |
|
"grad_norm": 39453.51953125, |
|
"learning_rate": 2.9931972789115647e-05, |
|
"loss": 1.7304, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.0612244897959184, |
|
"grad_norm": 58271.04296875, |
|
"learning_rate": 2.979591836734694e-05, |
|
"loss": 1.6894, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.1020408163265305, |
|
"grad_norm": 66804.171875, |
|
"learning_rate": 2.9659863945578233e-05, |
|
"loss": 1.7216, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.1428571428571428, |
|
"grad_norm": 85323.1328125, |
|
"learning_rate": 2.9523809523809523e-05, |
|
"loss": 1.6702, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.183673469387755, |
|
"grad_norm": 56704.1484375, |
|
"learning_rate": 2.9387755102040816e-05, |
|
"loss": 1.7761, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.2244897959183674, |
|
"grad_norm": 26519.884765625, |
|
"learning_rate": 2.925170068027211e-05, |
|
"loss": 1.7161, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.2653061224489797, |
|
"grad_norm": 30323.771484375, |
|
"learning_rate": 2.9115646258503402e-05, |
|
"loss": 1.7138, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.306122448979592, |
|
"grad_norm": 88774.5234375, |
|
"learning_rate": 2.8979591836734695e-05, |
|
"loss": 1.7125, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.346938775510204, |
|
"grad_norm": 81892.703125, |
|
"learning_rate": 2.8843537414965988e-05, |
|
"loss": 1.7273, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.3877551020408163, |
|
"grad_norm": 35876.6171875, |
|
"learning_rate": 2.870748299319728e-05, |
|
"loss": 1.7185, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.4285714285714286, |
|
"grad_norm": 91867.0390625, |
|
"learning_rate": 2.857142857142857e-05, |
|
"loss": 1.7057, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.469387755102041, |
|
"grad_norm": 47658.41796875, |
|
"learning_rate": 2.8435374149659864e-05, |
|
"loss": 1.7744, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.510204081632653, |
|
"grad_norm": 105770.484375, |
|
"learning_rate": 2.8299319727891157e-05, |
|
"loss": 1.6898, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.5510204081632653, |
|
"grad_norm": 104692.265625, |
|
"learning_rate": 2.816326530612245e-05, |
|
"loss": 1.7351, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.5918367346938775, |
|
"grad_norm": 45664.515625, |
|
"learning_rate": 2.802721088435374e-05, |
|
"loss": 1.6806, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.6326530612244898, |
|
"grad_norm": 45678.98828125, |
|
"learning_rate": 2.7891156462585036e-05, |
|
"loss": 1.7555, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.6734693877551021, |
|
"grad_norm": 108897.0078125, |
|
"learning_rate": 2.775510204081633e-05, |
|
"loss": 1.6663, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.7142857142857144, |
|
"grad_norm": 46971.11328125, |
|
"learning_rate": 2.761904761904762e-05, |
|
"loss": 1.7053, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.7551020408163265, |
|
"grad_norm": 69147.1171875, |
|
"learning_rate": 2.7482993197278912e-05, |
|
"loss": 1.7412, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.7959183673469388, |
|
"grad_norm": 89437.34375, |
|
"learning_rate": 2.7346938775510205e-05, |
|
"loss": 1.7245, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.836734693877551, |
|
"grad_norm": 40401.5625, |
|
"learning_rate": 2.7210884353741498e-05, |
|
"loss": 1.7614, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.8775510204081631, |
|
"grad_norm": 53271.234375, |
|
"learning_rate": 2.7074829931972787e-05, |
|
"loss": 1.7179, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.9183673469387754, |
|
"grad_norm": 69800.3203125, |
|
"learning_rate": 2.6938775510204084e-05, |
|
"loss": 1.7467, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.9591836734693877, |
|
"grad_norm": 86541.828125, |
|
"learning_rate": 2.6802721088435377e-05, |
|
"loss": 1.6975, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 100622.125, |
|
"learning_rate": 2.6666666666666667e-05, |
|
"loss": 1.6769, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 2.0408163265306123, |
|
"grad_norm": 140721.46875, |
|
"learning_rate": 2.653061224489796e-05, |
|
"loss": 1.7128, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.0408163265306123, |
|
"eval_accuracy": 0.1981891348088531, |
|
"eval_loss": 1.6279175281524658, |
|
"eval_runtime": 282.7617, |
|
"eval_samples_per_second": 10.546, |
|
"eval_steps_per_second": 0.223, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.0816326530612246, |
|
"grad_norm": 101114.078125, |
|
"learning_rate": 2.6394557823129253e-05, |
|
"loss": 1.7745, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 2.122448979591837, |
|
"grad_norm": 114329.1640625, |
|
"learning_rate": 2.6258503401360546e-05, |
|
"loss": 1.7301, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 2.163265306122449, |
|
"grad_norm": 102430.6328125, |
|
"learning_rate": 2.6122448979591835e-05, |
|
"loss": 1.7219, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 2.204081632653061, |
|
"grad_norm": 92331.4453125, |
|
"learning_rate": 2.598639455782313e-05, |
|
"loss": 1.6976, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 2.2448979591836733, |
|
"grad_norm": 37969.83203125, |
|
"learning_rate": 2.5850340136054425e-05, |
|
"loss": 1.6713, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.2857142857142856, |
|
"grad_norm": 88346.3046875, |
|
"learning_rate": 2.5714285714285714e-05, |
|
"loss": 1.7195, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.326530612244898, |
|
"grad_norm": 85409.15625, |
|
"learning_rate": 2.5578231292517007e-05, |
|
"loss": 1.7757, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.36734693877551, |
|
"grad_norm": 26777.841796875, |
|
"learning_rate": 2.54421768707483e-05, |
|
"loss": 1.7285, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 2.4081632653061225, |
|
"grad_norm": 75528.03125, |
|
"learning_rate": 2.5306122448979594e-05, |
|
"loss": 1.7239, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 2.4489795918367347, |
|
"grad_norm": 99135.8515625, |
|
"learning_rate": 2.5170068027210883e-05, |
|
"loss": 1.639, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.489795918367347, |
|
"grad_norm": 75343.3046875, |
|
"learning_rate": 2.5034013605442176e-05, |
|
"loss": 1.7259, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 2.5306122448979593, |
|
"grad_norm": 118070.5546875, |
|
"learning_rate": 2.4897959183673473e-05, |
|
"loss": 1.6815, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.571428571428571, |
|
"grad_norm": 154820.453125, |
|
"learning_rate": 2.4761904761904762e-05, |
|
"loss": 1.6935, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.612244897959184, |
|
"grad_norm": 95532.75, |
|
"learning_rate": 2.4625850340136055e-05, |
|
"loss": 1.6823, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.6530612244897958, |
|
"grad_norm": 86994.4921875, |
|
"learning_rate": 2.448979591836735e-05, |
|
"loss": 1.662, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.693877551020408, |
|
"grad_norm": 78671.8359375, |
|
"learning_rate": 2.435374149659864e-05, |
|
"loss": 1.6588, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.7346938775510203, |
|
"grad_norm": 96924.7265625, |
|
"learning_rate": 2.421768707482993e-05, |
|
"loss": 1.7981, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.7755102040816326, |
|
"grad_norm": 83239.3203125, |
|
"learning_rate": 2.4081632653061224e-05, |
|
"loss": 1.7216, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 2.816326530612245, |
|
"grad_norm": 88283.4765625, |
|
"learning_rate": 2.3945578231292517e-05, |
|
"loss": 1.6632, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.857142857142857, |
|
"grad_norm": 57946.15234375, |
|
"learning_rate": 2.380952380952381e-05, |
|
"loss": 1.6614, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.8979591836734695, |
|
"grad_norm": 69202.7109375, |
|
"learning_rate": 2.3673469387755103e-05, |
|
"loss": 1.753, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 2.938775510204082, |
|
"grad_norm": 65562.8046875, |
|
"learning_rate": 2.3537414965986396e-05, |
|
"loss": 1.7314, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.979591836734694, |
|
"grad_norm": 114870.2265625, |
|
"learning_rate": 2.340136054421769e-05, |
|
"loss": 1.714, |
|
"step": 730 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 2450, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 9.72903329888256e+18, |
|
"train_batch_size": 48, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|