Text Generation
Transformers
PyTorch
English
llama
code
text-generation-inference
Inference Endpoints
IF-CLP-34B / trainer_state.json
Pasta009's picture
Upload folder using huggingface_hub
4cf3a8a verified
raw
history blame
10 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.0,
"eval_steps": 500,
"global_step": 784,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.03,
"learning_rate": 1.99919725148484e-05,
"loss": 0.5476,
"step": 10
},
{
"epoch": 0.05,
"learning_rate": 1.9967902947497158e-05,
"loss": 0.4891,
"step": 20
},
{
"epoch": 0.08,
"learning_rate": 1.9927829941565187e-05,
"loss": 0.4573,
"step": 30
},
{
"epoch": 0.1,
"learning_rate": 1.9871817834144506e-05,
"loss": 0.4521,
"step": 40
},
{
"epoch": 0.13,
"learning_rate": 1.9799956552507235e-05,
"loss": 0.4565,
"step": 50
},
{
"epoch": 0.15,
"learning_rate": 1.971236146972764e-05,
"loss": 0.4528,
"step": 60
},
{
"epoch": 0.18,
"learning_rate": 1.9609173219450998e-05,
"loss": 0.4456,
"step": 70
},
{
"epoch": 0.2,
"learning_rate": 1.949055747010669e-05,
"loss": 0.4407,
"step": 80
},
{
"epoch": 0.23,
"learning_rate": 1.9356704658928036e-05,
"loss": 0.4251,
"step": 90
},
{
"epoch": 0.26,
"learning_rate": 1.9207829686205882e-05,
"loss": 0.4372,
"step": 100
},
{
"epoch": 0.28,
"learning_rate": 1.904417157026683e-05,
"loss": 0.4439,
"step": 110
},
{
"epoch": 0.31,
"learning_rate": 1.8865993063730003e-05,
"loss": 0.4305,
"step": 120
},
{
"epoch": 0.33,
"learning_rate": 1.867358023165851e-05,
"loss": 0.4395,
"step": 130
},
{
"epoch": 0.36,
"learning_rate": 1.8467241992282842e-05,
"loss": 0.4402,
"step": 140
},
{
"epoch": 0.38,
"learning_rate": 1.824730962103356e-05,
"loss": 0.4382,
"step": 150
},
{
"epoch": 0.41,
"learning_rate": 1.8014136218679566e-05,
"loss": 0.4359,
"step": 160
},
{
"epoch": 0.43,
"learning_rate": 1.7768096144425903e-05,
"loss": 0.4202,
"step": 170
},
{
"epoch": 0.46,
"learning_rate": 1.7509584414881114e-05,
"loss": 0.4181,
"step": 180
},
{
"epoch": 0.48,
"learning_rate": 1.7239016069859292e-05,
"loss": 0.4263,
"step": 190
},
{
"epoch": 0.51,
"learning_rate": 1.6956825506034866e-05,
"loss": 0.4227,
"step": 200
},
{
"epoch": 0.54,
"learning_rate": 1.6663465779520042e-05,
"loss": 0.4165,
"step": 210
},
{
"epoch": 0.56,
"learning_rate": 1.635940787848455e-05,
"loss": 0.4213,
"step": 220
},
{
"epoch": 0.59,
"learning_rate": 1.6045139966985562e-05,
"loss": 0.4197,
"step": 230
},
{
"epoch": 0.61,
"learning_rate": 1.5721166601221697e-05,
"loss": 0.4257,
"step": 240
},
{
"epoch": 0.64,
"learning_rate": 1.5388007919469604e-05,
"loss": 0.4216,
"step": 250
},
{
"epoch": 0.66,
"learning_rate": 1.504619880700346e-05,
"loss": 0.4171,
"step": 260
},
{
"epoch": 0.69,
"learning_rate": 1.4696288037338258e-05,
"loss": 0.42,
"step": 270
},
{
"epoch": 0.71,
"learning_rate": 1.4338837391175582e-05,
"loss": 0.4179,
"step": 280
},
{
"epoch": 0.74,
"learning_rate": 1.397442075446633e-05,
"loss": 0.4183,
"step": 290
},
{
"epoch": 0.77,
"learning_rate": 1.3603623197038536e-05,
"loss": 0.4291,
"step": 300
},
{
"epoch": 0.79,
"learning_rate": 1.3227040033269501e-05,
"loss": 0.4142,
"step": 310
},
{
"epoch": 0.82,
"learning_rate": 1.2845275866310325e-05,
"loss": 0.4154,
"step": 320
},
{
"epoch": 0.84,
"learning_rate": 1.2458943617397346e-05,
"loss": 0.4199,
"step": 330
},
{
"epoch": 0.87,
"learning_rate": 1.206866354180891e-05,
"loss": 0.4213,
"step": 340
},
{
"epoch": 0.89,
"learning_rate": 1.1675062233047365e-05,
"loss": 0.4039,
"step": 350
},
{
"epoch": 0.92,
"learning_rate": 1.1278771616845061e-05,
"loss": 0.4065,
"step": 360
},
{
"epoch": 0.94,
"learning_rate": 1.0880427936609455e-05,
"loss": 0.4117,
"step": 370
},
{
"epoch": 0.97,
"learning_rate": 1.0480670731936209e-05,
"loss": 0.4132,
"step": 380
},
{
"epoch": 0.99,
"learning_rate": 1.0080141811830277e-05,
"loss": 0.4099,
"step": 390
},
{
"epoch": 1.02,
"learning_rate": 9.67948422428345e-06,
"loss": 0.3254,
"step": 400
},
{
"epoch": 1.05,
"learning_rate": 9.279341223862705e-06,
"loss": 0.2853,
"step": 410
},
{
"epoch": 1.07,
"learning_rate": 8.880355238966923e-06,
"loss": 0.2822,
"step": 420
},
{
"epoch": 1.1,
"learning_rate": 8.483166840409996e-06,
"loss": 0.2859,
"step": 430
},
{
"epoch": 1.12,
"learning_rate": 8.08841371298628e-06,
"loss": 0.2797,
"step": 440
},
{
"epoch": 1.15,
"learning_rate": 7.696729631669563e-06,
"loss": 0.2731,
"step": 450
},
{
"epoch": 1.17,
"learning_rate": 7.308743444089232e-06,
"loss": 0.27,
"step": 460
},
{
"epoch": 1.2,
"learning_rate": 6.925078060917245e-06,
"loss": 0.2809,
"step": 470
},
{
"epoch": 1.22,
"learning_rate": 6.546349455786926e-06,
"loss": 0.28,
"step": 480
},
{
"epoch": 1.25,
"learning_rate": 6.173165676349103e-06,
"loss": 0.2863,
"step": 490
},
{
"epoch": 1.28,
"learning_rate": 5.806125868053433e-06,
"loss": 0.2804,
"step": 500
},
{
"epoch": 1.3,
"learning_rate": 5.4458193122221405e-06,
"loss": 0.2797,
"step": 510
},
{
"epoch": 1.33,
"learning_rate": 5.092824479960625e-06,
"loss": 0.2861,
"step": 520
},
{
"epoch": 1.35,
"learning_rate": 4.747708103423791e-06,
"loss": 0.2777,
"step": 530
},
{
"epoch": 1.38,
"learning_rate": 4.411024265929283e-06,
"loss": 0.2791,
"step": 540
},
{
"epoch": 1.4,
"learning_rate": 4.083313512378368e-06,
"loss": 0.275,
"step": 550
},
{
"epoch": 1.43,
"learning_rate": 3.7651019814126656e-06,
"loss": 0.2787,
"step": 560
},
{
"epoch": 1.45,
"learning_rate": 3.456900560700158e-06,
"loss": 0.2777,
"step": 570
},
{
"epoch": 1.48,
"learning_rate": 3.1592040667065393e-06,
"loss": 0.2727,
"step": 580
},
{
"epoch": 1.51,
"learning_rate": 2.8724904502688566e-06,
"loss": 0.2748,
"step": 590
},
{
"epoch": 1.53,
"learning_rate": 2.597220029246846e-06,
"loss": 0.2809,
"step": 600
},
{
"epoch": 1.56,
"learning_rate": 2.333834749484e-06,
"loss": 0.2706,
"step": 610
},
{
"epoch": 1.58,
"learning_rate": 2.082757475264804e-06,
"loss": 0.2748,
"step": 620
},
{
"epoch": 1.61,
"learning_rate": 1.8443913104073984e-06,
"loss": 0.273,
"step": 630
},
{
"epoch": 1.63,
"learning_rate": 1.6191189510815942e-06,
"loss": 0.2788,
"step": 640
},
{
"epoch": 1.66,
"learning_rate": 1.4073020713912988e-06,
"loss": 0.2741,
"step": 650
},
{
"epoch": 1.68,
"learning_rate": 1.209280742707828e-06,
"loss": 0.2749,
"step": 660
},
{
"epoch": 1.71,
"learning_rate": 1.0253728876863256e-06,
"loss": 0.2739,
"step": 670
},
{
"epoch": 1.73,
"learning_rate": 8.558737698418762e-07,
"loss": 0.27,
"step": 680
},
{
"epoch": 1.76,
"learning_rate": 7.010555195048241e-07,
"loss": 0.2683,
"step": 690
},
{
"epoch": 1.79,
"learning_rate": 5.611666969163243e-07,
"loss": 0.2701,
"step": 700
},
{
"epoch": 1.81,
"learning_rate": 4.3643189316561864e-07,
"loss": 0.2705,
"step": 710
},
{
"epoch": 1.84,
"learning_rate": 3.2705136960970554e-07,
"loss": 0.2695,
"step": 720
},
{
"epoch": 1.86,
"learning_rate": 2.3320073635432984e-07,
"loss": 0.2747,
"step": 730
},
{
"epoch": 1.89,
"learning_rate": 1.55030670312476e-07,
"loss": 0.273,
"step": 740
},
{
"epoch": 1.91,
"learning_rate": 9.26666732930348e-08,
"loss": 0.2721,
"step": 750
},
{
"epoch": 1.94,
"learning_rate": 4.6208870508017703e-08,
"loss": 0.2798,
"step": 760
},
{
"epoch": 1.96,
"learning_rate": 1.5731849821833955e-08,
"loss": 0.2738,
"step": 770
},
{
"epoch": 1.99,
"learning_rate": 1.2845420006879494e-09,
"loss": 0.2722,
"step": 780
},
{
"epoch": 2.0,
"step": 784,
"total_flos": 1134283296079872.0,
"train_loss": 0.3541707301474347,
"train_runtime": 33786.0741,
"train_samples_per_second": 5.938,
"train_steps_per_second": 0.023
}
],
"logging_steps": 10,
"max_steps": 784,
"num_train_epochs": 2,
"save_steps": 500,
"total_flos": 1134283296079872.0,
"trial_name": null,
"trial_params": null
}