|
{ |
|
"best_metric": 0.4873823821544647, |
|
"best_model_checkpoint": "Model-Meme/Typhoon/Typhoon1.5-Sentiment-Fold1/checkpoint-3824", |
|
"epoch": 10.0, |
|
"eval_steps": 500, |
|
"global_step": 19120, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.2615062761506276, |
|
"grad_norm": 1.630317211151123, |
|
"learning_rate": 8.710801393728223e-06, |
|
"loss": 0.888, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.5230125523012552, |
|
"grad_norm": 1.2002062797546387, |
|
"learning_rate": 9.986987223454062e-06, |
|
"loss": 0.5478, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.7845188284518828, |
|
"grad_norm": 1.279449224472046, |
|
"learning_rate": 9.93861382775312e-06, |
|
"loss": 0.551, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.49981698393821716, |
|
"eval_runtime": 106.1621, |
|
"eval_samples_per_second": 9.005, |
|
"eval_steps_per_second": 9.005, |
|
"step": 1912 |
|
}, |
|
{ |
|
"epoch": 1.0460251046025104, |
|
"grad_norm": 2.099172830581665, |
|
"learning_rate": 9.854833795480441e-06, |
|
"loss": 0.5293, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.3075313807531381, |
|
"grad_norm": 1.455888032913208, |
|
"learning_rate": 9.736519159090385e-06, |
|
"loss": 0.4982, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.5690376569037658, |
|
"grad_norm": 1.8300031423568726, |
|
"learning_rate": 9.584044234397988e-06, |
|
"loss": 0.49, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.8305439330543933, |
|
"grad_norm": 1.8547563552856445, |
|
"learning_rate": 9.39870470560879e-06, |
|
"loss": 0.4965, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.4873823821544647, |
|
"eval_runtime": 106.0089, |
|
"eval_samples_per_second": 9.018, |
|
"eval_steps_per_second": 9.018, |
|
"step": 3824 |
|
}, |
|
{ |
|
"epoch": 2.092050209205021, |
|
"grad_norm": 1.9847608804702759, |
|
"learning_rate": 9.181829336130319e-06, |
|
"loss": 0.479, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.3535564853556483, |
|
"grad_norm": 2.7613589763641357, |
|
"learning_rate": 8.934972980757115e-06, |
|
"loss": 0.43, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 2.6150627615062763, |
|
"grad_norm": 2.3412399291992188, |
|
"learning_rate": 8.660482449775261e-06, |
|
"loss": 0.4243, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 2.8765690376569037, |
|
"grad_norm": 3.2489728927612305, |
|
"learning_rate": 8.359226157693385e-06, |
|
"loss": 0.4525, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.49290716648101807, |
|
"eval_runtime": 106.1261, |
|
"eval_samples_per_second": 9.008, |
|
"eval_steps_per_second": 9.008, |
|
"step": 5736 |
|
}, |
|
{ |
|
"epoch": 3.1380753138075312, |
|
"grad_norm": 3.534010410308838, |
|
"learning_rate": 8.033886407403191e-06, |
|
"loss": 0.4122, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 3.399581589958159, |
|
"grad_norm": 4.633702754974365, |
|
"learning_rate": 7.686795672489905e-06, |
|
"loss": 0.3837, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 3.6610878661087867, |
|
"grad_norm": 4.420021057128906, |
|
"learning_rate": 7.320442366896608e-06, |
|
"loss": 0.3789, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 3.922594142259414, |
|
"grad_norm": 3.9446756839752197, |
|
"learning_rate": 6.93823377901452e-06, |
|
"loss": 0.3822, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 0.5124192237854004, |
|
"eval_runtime": 103.7839, |
|
"eval_samples_per_second": 9.211, |
|
"eval_steps_per_second": 9.211, |
|
"step": 7648 |
|
}, |
|
{ |
|
"epoch": 4.184100418410042, |
|
"grad_norm": 3.961578607559204, |
|
"learning_rate": 6.541379114428449e-06, |
|
"loss": 0.3587, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 4.445606694560669, |
|
"grad_norm": 5.545160293579102, |
|
"learning_rate": 6.13347376800588e-06, |
|
"loss": 0.3221, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 4.707112970711297, |
|
"grad_norm": 4.203147888183594, |
|
"learning_rate": 5.7174421547466815e-06, |
|
"loss": 0.3245, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 4.968619246861925, |
|
"grad_norm": 5.057765007019043, |
|
"learning_rate": 5.297112430378265e-06, |
|
"loss": 0.3273, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 0.5422254204750061, |
|
"eval_runtime": 103.666, |
|
"eval_samples_per_second": 9.222, |
|
"eval_steps_per_second": 9.222, |
|
"step": 9560 |
|
}, |
|
{ |
|
"epoch": 5.2301255230125525, |
|
"grad_norm": 4.777406215667725, |
|
"learning_rate": 4.873814405549272e-06, |
|
"loss": 0.2841, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 5.49163179916318, |
|
"grad_norm": 5.399323463439941, |
|
"learning_rate": 4.451421049042024e-06, |
|
"loss": 0.2763, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 5.7531380753138075, |
|
"grad_norm": 7.916058540344238, |
|
"learning_rate": 4.032960645429033e-06, |
|
"loss": 0.2981, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 0.595160186290741, |
|
"eval_runtime": 103.9764, |
|
"eval_samples_per_second": 9.194, |
|
"eval_steps_per_second": 9.194, |
|
"step": 11472 |
|
}, |
|
{ |
|
"epoch": 6.014644351464435, |
|
"grad_norm": 7.495786190032959, |
|
"learning_rate": 3.6222474467023006e-06, |
|
"loss": 0.2741, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 6.2761506276150625, |
|
"grad_norm": 5.999951362609863, |
|
"learning_rate": 3.220580839399443e-06, |
|
"loss": 0.2307, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 6.53765690376569, |
|
"grad_norm": 6.181816101074219, |
|
"learning_rate": 2.8316715056404507e-06, |
|
"loss": 0.242, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 6.799163179916318, |
|
"grad_norm": 4.290989398956299, |
|
"learning_rate": 2.4583076714174803e-06, |
|
"loss": 0.2416, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 0.6211022138595581, |
|
"eval_runtime": 103.368, |
|
"eval_samples_per_second": 9.249, |
|
"eval_steps_per_second": 9.249, |
|
"step": 13384 |
|
}, |
|
{ |
|
"epoch": 7.060669456066946, |
|
"grad_norm": 7.918560028076172, |
|
"learning_rate": 2.103166111641151e-06, |
|
"loss": 0.2382, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 7.322175732217573, |
|
"grad_norm": 4.114747524261475, |
|
"learning_rate": 1.7694393589342428e-06, |
|
"loss": 0.2059, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 7.583682008368201, |
|
"grad_norm": 5.888193130493164, |
|
"learning_rate": 1.4581832353425335e-06, |
|
"loss": 0.2127, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 7.845188284518828, |
|
"grad_norm": 5.456567287445068, |
|
"learning_rate": 1.1723196253920394e-06, |
|
"loss": 0.2116, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 0.6545947194099426, |
|
"eval_runtime": 107.6369, |
|
"eval_samples_per_second": 8.882, |
|
"eval_steps_per_second": 8.882, |
|
"step": 15296 |
|
}, |
|
{ |
|
"epoch": 8.106694560669457, |
|
"grad_norm": 7.760040283203125, |
|
"learning_rate": 9.138979845169372e-07, |
|
"loss": 0.1854, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 8.368200836820083, |
|
"grad_norm": 6.356107234954834, |
|
"learning_rate": 6.847710271508462e-07, |
|
"loss": 0.1877, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 8.629707112970712, |
|
"grad_norm": 8.600861549377441, |
|
"learning_rate": 4.865814439740219e-07, |
|
"loss": 0.1922, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 8.891213389121338, |
|
"grad_norm": 9.489436149597168, |
|
"learning_rate": 3.207501248937539e-07, |
|
"loss": 0.188, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 0.6717672944068909, |
|
"eval_runtime": 142.6914, |
|
"eval_samples_per_second": 6.7, |
|
"eval_steps_per_second": 6.7, |
|
"step": 17208 |
|
}, |
|
{ |
|
"epoch": 9.152719665271967, |
|
"grad_norm": 4.833051681518555, |
|
"learning_rate": 1.886963892774535e-07, |
|
"loss": 0.1797, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 9.414225941422593, |
|
"grad_norm": 4.518942356109619, |
|
"learning_rate": 9.083801976963458e-08, |
|
"loss": 0.1787, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 9.675732217573222, |
|
"grad_norm": 7.789809703826904, |
|
"learning_rate": 2.8175136316832e-08, |
|
"loss": 0.1768, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 9.93723849372385, |
|
"grad_norm": 5.010494232177734, |
|
"learning_rate": 1.1569908788744423e-09, |
|
"loss": 0.1747, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 0.6839816570281982, |
|
"eval_runtime": 142.9966, |
|
"eval_samples_per_second": 6.685, |
|
"eval_steps_per_second": 6.685, |
|
"step": 19120 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 19120, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.7134220695783014e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|