|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.9921259842519685, |
|
"eval_steps": 500, |
|
"global_step": 570, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.05249343832020997, |
|
"grad_norm": 5.856240742467741, |
|
"learning_rate": 5e-06, |
|
"loss": 0.9788, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.10498687664041995, |
|
"grad_norm": 1.504758475031017, |
|
"learning_rate": 5e-06, |
|
"loss": 0.8672, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.15748031496062992, |
|
"grad_norm": 0.9699579752356402, |
|
"learning_rate": 5e-06, |
|
"loss": 0.8312, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.2099737532808399, |
|
"grad_norm": 2.3401202479822167, |
|
"learning_rate": 5e-06, |
|
"loss": 0.8176, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.26246719160104987, |
|
"grad_norm": 1.4328785905158876, |
|
"learning_rate": 5e-06, |
|
"loss": 0.8021, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.31496062992125984, |
|
"grad_norm": 1.776027549071983, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7924, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.3674540682414698, |
|
"grad_norm": 1.6731596205975654, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7862, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.4199475065616798, |
|
"grad_norm": 0.7241122381033481, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7745, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.47244094488188976, |
|
"grad_norm": 0.667641025148353, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7684, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.5249343832020997, |
|
"grad_norm": 0.5504196367515327, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7659, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.5774278215223098, |
|
"grad_norm": 0.5097698406084038, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7644, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.6299212598425197, |
|
"grad_norm": 0.7982727448245113, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7621, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.6824146981627297, |
|
"grad_norm": 0.7024212096277765, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7587, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.7349081364829396, |
|
"grad_norm": 0.6368075359040738, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7528, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.7874015748031497, |
|
"grad_norm": 0.5954968543585505, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7523, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.8398950131233596, |
|
"grad_norm": 0.670163545471037, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7534, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.8923884514435696, |
|
"grad_norm": 0.5996022754065202, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7517, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.9448818897637795, |
|
"grad_norm": 0.8415836628541776, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7459, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.9973753280839895, |
|
"grad_norm": 0.690137189983478, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7488, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.0498687664041995, |
|
"grad_norm": 0.5870306499065036, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7462, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.1023622047244095, |
|
"grad_norm": 0.6711935804641513, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7045, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.1548556430446195, |
|
"grad_norm": 0.6793309667692513, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7054, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.2073490813648293, |
|
"grad_norm": 0.4945921653765713, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7007, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.2598425196850394, |
|
"grad_norm": 0.7994708362053626, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7035, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.3123359580052494, |
|
"grad_norm": 0.6645996514564108, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7029, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.3648293963254594, |
|
"grad_norm": 0.6524179166081678, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7069, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.4173228346456692, |
|
"grad_norm": 0.8192112613994083, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7058, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.4698162729658792, |
|
"grad_norm": 0.7359384256295017, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7096, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.5223097112860893, |
|
"grad_norm": 0.8479895809019217, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7014, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.574803149606299, |
|
"grad_norm": 0.5643654766970888, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7027, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.627296587926509, |
|
"grad_norm": 0.5431958308145564, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7014, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.6797900262467191, |
|
"grad_norm": 0.520988761182218, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7004, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.7322834645669292, |
|
"grad_norm": 0.5293919501601649, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7014, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.7847769028871392, |
|
"grad_norm": 0.5558746861224813, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7051, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.8372703412073492, |
|
"grad_norm": 0.5864088226331533, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7028, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.889763779527559, |
|
"grad_norm": 0.5203657185554663, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7029, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.942257217847769, |
|
"grad_norm": 0.5249283309342107, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6985, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.9947506561679789, |
|
"grad_norm": 0.5014731402840785, |
|
"learning_rate": 5e-06, |
|
"loss": 0.698, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 2.047244094488189, |
|
"grad_norm": 0.6606881961316096, |
|
"learning_rate": 5e-06, |
|
"loss": 0.701, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 2.099737532808399, |
|
"grad_norm": 0.7956310597917616, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6552, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 2.152230971128609, |
|
"grad_norm": 0.7409823582769325, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6529, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 2.204724409448819, |
|
"grad_norm": 0.5301522848133626, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6583, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 2.257217847769029, |
|
"grad_norm": 0.5707394351699061, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6567, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 2.309711286089239, |
|
"grad_norm": 0.8305197146184463, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6591, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 2.362204724409449, |
|
"grad_norm": 0.6345171704808862, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6586, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.4146981627296586, |
|
"grad_norm": 0.540143075416848, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6577, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 2.4671916010498687, |
|
"grad_norm": 0.6047688082843984, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6583, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 2.5196850393700787, |
|
"grad_norm": 0.5089956964635003, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6597, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 2.5721784776902887, |
|
"grad_norm": 0.6189665519218218, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6593, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 2.6246719160104988, |
|
"grad_norm": 0.6080333484551841, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6586, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.677165354330709, |
|
"grad_norm": 0.6796282219494648, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6613, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 2.729658792650919, |
|
"grad_norm": 0.6734905288037869, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6593, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 2.7821522309711284, |
|
"grad_norm": 0.7251898875558835, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6579, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 2.8346456692913384, |
|
"grad_norm": 0.5903651504718952, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6613, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 2.8871391076115485, |
|
"grad_norm": 0.6998542288142676, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6589, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.9396325459317585, |
|
"grad_norm": 0.6240795370124482, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6594, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.9921259842519685, |
|
"grad_norm": 0.5469561259952066, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6579, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.9921259842519685, |
|
"step": 570, |
|
"total_flos": 954352470589440.0, |
|
"train_loss": 0.718004734474316, |
|
"train_runtime": 8541.7768, |
|
"train_samples_per_second": 34.252, |
|
"train_steps_per_second": 0.067 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 570, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 954352470589440.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|