hr16's picture
Training in progress, epoch 3, checkpoint
3b3a75c verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 735,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.04081632653061224,
"grad_norm": 20286.314453125,
"learning_rate": 1.2244897959183673e-06,
"loss": 2.0797,
"step": 10
},
{
"epoch": 0.08163265306122448,
"grad_norm": 21226.833984375,
"learning_rate": 2.4489795918367347e-06,
"loss": 2.0794,
"step": 20
},
{
"epoch": 0.12244897959183673,
"grad_norm": 20473.34375,
"learning_rate": 3.673469387755102e-06,
"loss": 2.0791,
"step": 30
},
{
"epoch": 0.16326530612244897,
"grad_norm": 21295.623046875,
"learning_rate": 4.897959183673469e-06,
"loss": 2.0786,
"step": 40
},
{
"epoch": 0.20408163265306123,
"grad_norm": 22374.26953125,
"learning_rate": 6.122448979591837e-06,
"loss": 2.0775,
"step": 50
},
{
"epoch": 0.24489795918367346,
"grad_norm": 27922.837890625,
"learning_rate": 7.346938775510204e-06,
"loss": 2.0762,
"step": 60
},
{
"epoch": 0.2857142857142857,
"grad_norm": 21962.65625,
"learning_rate": 8.571428571428571e-06,
"loss": 2.0751,
"step": 70
},
{
"epoch": 0.32653061224489793,
"grad_norm": 25949.669921875,
"learning_rate": 9.795918367346939e-06,
"loss": 2.073,
"step": 80
},
{
"epoch": 0.3673469387755102,
"grad_norm": 21337.779296875,
"learning_rate": 1.1020408163265306e-05,
"loss": 2.0711,
"step": 90
},
{
"epoch": 0.40816326530612246,
"grad_norm": 21139.53125,
"learning_rate": 1.2244897959183674e-05,
"loss": 2.068,
"step": 100
},
{
"epoch": 0.4489795918367347,
"grad_norm": 18687.326171875,
"learning_rate": 1.3469387755102042e-05,
"loss": 2.0621,
"step": 110
},
{
"epoch": 0.4897959183673469,
"grad_norm": 23997.892578125,
"learning_rate": 1.4693877551020408e-05,
"loss": 2.0557,
"step": 120
},
{
"epoch": 0.5306122448979592,
"grad_norm": 25094.125,
"learning_rate": 1.5918367346938776e-05,
"loss": 2.0436,
"step": 130
},
{
"epoch": 0.5714285714285714,
"grad_norm": 35127.62890625,
"learning_rate": 1.7142857142857142e-05,
"loss": 2.0313,
"step": 140
},
{
"epoch": 0.6122448979591837,
"grad_norm": 38480.73046875,
"learning_rate": 1.836734693877551e-05,
"loss": 2.0124,
"step": 150
},
{
"epoch": 0.6530612244897959,
"grad_norm": 37025.25390625,
"learning_rate": 1.9591836734693877e-05,
"loss": 1.9832,
"step": 160
},
{
"epoch": 0.6938775510204082,
"grad_norm": 47326.75390625,
"learning_rate": 2.0816326530612247e-05,
"loss": 1.9503,
"step": 170
},
{
"epoch": 0.7346938775510204,
"grad_norm": 44386.79296875,
"learning_rate": 2.2040816326530613e-05,
"loss": 1.9033,
"step": 180
},
{
"epoch": 0.7755102040816326,
"grad_norm": 60427.28515625,
"learning_rate": 2.326530612244898e-05,
"loss": 1.8548,
"step": 190
},
{
"epoch": 0.8163265306122449,
"grad_norm": 60310.234375,
"learning_rate": 2.448979591836735e-05,
"loss": 1.7809,
"step": 200
},
{
"epoch": 0.8571428571428571,
"grad_norm": 73581.171875,
"learning_rate": 2.5714285714285714e-05,
"loss": 1.8435,
"step": 210
},
{
"epoch": 0.8979591836734694,
"grad_norm": 63305.5,
"learning_rate": 2.6938775510204084e-05,
"loss": 1.7222,
"step": 220
},
{
"epoch": 0.9387755102040817,
"grad_norm": 78695.546875,
"learning_rate": 2.816326530612245e-05,
"loss": 1.7823,
"step": 230
},
{
"epoch": 0.9795918367346939,
"grad_norm": 118529.84375,
"learning_rate": 2.9387755102040816e-05,
"loss": 1.7203,
"step": 240
},
{
"epoch": 1.0204081632653061,
"grad_norm": 39453.51953125,
"learning_rate": 2.9931972789115647e-05,
"loss": 1.7304,
"step": 250
},
{
"epoch": 1.0612244897959184,
"grad_norm": 58271.04296875,
"learning_rate": 2.979591836734694e-05,
"loss": 1.6894,
"step": 260
},
{
"epoch": 1.1020408163265305,
"grad_norm": 66804.171875,
"learning_rate": 2.9659863945578233e-05,
"loss": 1.7216,
"step": 270
},
{
"epoch": 1.1428571428571428,
"grad_norm": 85323.1328125,
"learning_rate": 2.9523809523809523e-05,
"loss": 1.6702,
"step": 280
},
{
"epoch": 1.183673469387755,
"grad_norm": 56704.1484375,
"learning_rate": 2.9387755102040816e-05,
"loss": 1.7761,
"step": 290
},
{
"epoch": 1.2244897959183674,
"grad_norm": 26519.884765625,
"learning_rate": 2.925170068027211e-05,
"loss": 1.7161,
"step": 300
},
{
"epoch": 1.2653061224489797,
"grad_norm": 30323.771484375,
"learning_rate": 2.9115646258503402e-05,
"loss": 1.7138,
"step": 310
},
{
"epoch": 1.306122448979592,
"grad_norm": 88774.5234375,
"learning_rate": 2.8979591836734695e-05,
"loss": 1.7125,
"step": 320
},
{
"epoch": 1.346938775510204,
"grad_norm": 81892.703125,
"learning_rate": 2.8843537414965988e-05,
"loss": 1.7273,
"step": 330
},
{
"epoch": 1.3877551020408163,
"grad_norm": 35876.6171875,
"learning_rate": 2.870748299319728e-05,
"loss": 1.7185,
"step": 340
},
{
"epoch": 1.4285714285714286,
"grad_norm": 91867.0390625,
"learning_rate": 2.857142857142857e-05,
"loss": 1.7057,
"step": 350
},
{
"epoch": 1.469387755102041,
"grad_norm": 47658.41796875,
"learning_rate": 2.8435374149659864e-05,
"loss": 1.7744,
"step": 360
},
{
"epoch": 1.510204081632653,
"grad_norm": 105770.484375,
"learning_rate": 2.8299319727891157e-05,
"loss": 1.6898,
"step": 370
},
{
"epoch": 1.5510204081632653,
"grad_norm": 104692.265625,
"learning_rate": 2.816326530612245e-05,
"loss": 1.7351,
"step": 380
},
{
"epoch": 1.5918367346938775,
"grad_norm": 45664.515625,
"learning_rate": 2.802721088435374e-05,
"loss": 1.6806,
"step": 390
},
{
"epoch": 1.6326530612244898,
"grad_norm": 45678.98828125,
"learning_rate": 2.7891156462585036e-05,
"loss": 1.7555,
"step": 400
},
{
"epoch": 1.6734693877551021,
"grad_norm": 108897.0078125,
"learning_rate": 2.775510204081633e-05,
"loss": 1.6663,
"step": 410
},
{
"epoch": 1.7142857142857144,
"grad_norm": 46971.11328125,
"learning_rate": 2.761904761904762e-05,
"loss": 1.7053,
"step": 420
},
{
"epoch": 1.7551020408163265,
"grad_norm": 69147.1171875,
"learning_rate": 2.7482993197278912e-05,
"loss": 1.7412,
"step": 430
},
{
"epoch": 1.7959183673469388,
"grad_norm": 89437.34375,
"learning_rate": 2.7346938775510205e-05,
"loss": 1.7245,
"step": 440
},
{
"epoch": 1.836734693877551,
"grad_norm": 40401.5625,
"learning_rate": 2.7210884353741498e-05,
"loss": 1.7614,
"step": 450
},
{
"epoch": 1.8775510204081631,
"grad_norm": 53271.234375,
"learning_rate": 2.7074829931972787e-05,
"loss": 1.7179,
"step": 460
},
{
"epoch": 1.9183673469387754,
"grad_norm": 69800.3203125,
"learning_rate": 2.6938775510204084e-05,
"loss": 1.7467,
"step": 470
},
{
"epoch": 1.9591836734693877,
"grad_norm": 86541.828125,
"learning_rate": 2.6802721088435377e-05,
"loss": 1.6975,
"step": 480
},
{
"epoch": 2.0,
"grad_norm": 100622.125,
"learning_rate": 2.6666666666666667e-05,
"loss": 1.6769,
"step": 490
},
{
"epoch": 2.0408163265306123,
"grad_norm": 140721.46875,
"learning_rate": 2.653061224489796e-05,
"loss": 1.7128,
"step": 500
},
{
"epoch": 2.0408163265306123,
"eval_accuracy": 0.1981891348088531,
"eval_loss": 1.6279175281524658,
"eval_runtime": 282.7617,
"eval_samples_per_second": 10.546,
"eval_steps_per_second": 0.223,
"step": 500
},
{
"epoch": 2.0816326530612246,
"grad_norm": 101114.078125,
"learning_rate": 2.6394557823129253e-05,
"loss": 1.7745,
"step": 510
},
{
"epoch": 2.122448979591837,
"grad_norm": 114329.1640625,
"learning_rate": 2.6258503401360546e-05,
"loss": 1.7301,
"step": 520
},
{
"epoch": 2.163265306122449,
"grad_norm": 102430.6328125,
"learning_rate": 2.6122448979591835e-05,
"loss": 1.7219,
"step": 530
},
{
"epoch": 2.204081632653061,
"grad_norm": 92331.4453125,
"learning_rate": 2.598639455782313e-05,
"loss": 1.6976,
"step": 540
},
{
"epoch": 2.2448979591836733,
"grad_norm": 37969.83203125,
"learning_rate": 2.5850340136054425e-05,
"loss": 1.6713,
"step": 550
},
{
"epoch": 2.2857142857142856,
"grad_norm": 88346.3046875,
"learning_rate": 2.5714285714285714e-05,
"loss": 1.7195,
"step": 560
},
{
"epoch": 2.326530612244898,
"grad_norm": 85409.15625,
"learning_rate": 2.5578231292517007e-05,
"loss": 1.7757,
"step": 570
},
{
"epoch": 2.36734693877551,
"grad_norm": 26777.841796875,
"learning_rate": 2.54421768707483e-05,
"loss": 1.7285,
"step": 580
},
{
"epoch": 2.4081632653061225,
"grad_norm": 75528.03125,
"learning_rate": 2.5306122448979594e-05,
"loss": 1.7239,
"step": 590
},
{
"epoch": 2.4489795918367347,
"grad_norm": 99135.8515625,
"learning_rate": 2.5170068027210883e-05,
"loss": 1.639,
"step": 600
},
{
"epoch": 2.489795918367347,
"grad_norm": 75343.3046875,
"learning_rate": 2.5034013605442176e-05,
"loss": 1.7259,
"step": 610
},
{
"epoch": 2.5306122448979593,
"grad_norm": 118070.5546875,
"learning_rate": 2.4897959183673473e-05,
"loss": 1.6815,
"step": 620
},
{
"epoch": 2.571428571428571,
"grad_norm": 154820.453125,
"learning_rate": 2.4761904761904762e-05,
"loss": 1.6935,
"step": 630
},
{
"epoch": 2.612244897959184,
"grad_norm": 95532.75,
"learning_rate": 2.4625850340136055e-05,
"loss": 1.6823,
"step": 640
},
{
"epoch": 2.6530612244897958,
"grad_norm": 86994.4921875,
"learning_rate": 2.448979591836735e-05,
"loss": 1.662,
"step": 650
},
{
"epoch": 2.693877551020408,
"grad_norm": 78671.8359375,
"learning_rate": 2.435374149659864e-05,
"loss": 1.6588,
"step": 660
},
{
"epoch": 2.7346938775510203,
"grad_norm": 96924.7265625,
"learning_rate": 2.421768707482993e-05,
"loss": 1.7981,
"step": 670
},
{
"epoch": 2.7755102040816326,
"grad_norm": 83239.3203125,
"learning_rate": 2.4081632653061224e-05,
"loss": 1.7216,
"step": 680
},
{
"epoch": 2.816326530612245,
"grad_norm": 88283.4765625,
"learning_rate": 2.3945578231292517e-05,
"loss": 1.6632,
"step": 690
},
{
"epoch": 2.857142857142857,
"grad_norm": 57946.15234375,
"learning_rate": 2.380952380952381e-05,
"loss": 1.6614,
"step": 700
},
{
"epoch": 2.8979591836734695,
"grad_norm": 69202.7109375,
"learning_rate": 2.3673469387755103e-05,
"loss": 1.753,
"step": 710
},
{
"epoch": 2.938775510204082,
"grad_norm": 65562.8046875,
"learning_rate": 2.3537414965986396e-05,
"loss": 1.7314,
"step": 720
},
{
"epoch": 2.979591836734694,
"grad_norm": 114870.2265625,
"learning_rate": 2.340136054421769e-05,
"loss": 1.714,
"step": 730
}
],
"logging_steps": 10,
"max_steps": 2450,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 9.72903329888256e+18,
"train_batch_size": 48,
"trial_name": null,
"trial_params": null
}