phi3nedtuned-ner-json / trainer_state.json
shujatoor's picture
End of training
10e70c5 verified
raw
history blame contribute delete
No virus
39.1 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 5.0,
"eval_steps": 500,
"global_step": 4450,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.02247191011235955,
"grad_norm": 3.921875,
"learning_rate": 4.49438202247191e-06,
"loss": 2.2098,
"step": 20
},
{
"epoch": 0.0449438202247191,
"grad_norm": 2.03125,
"learning_rate": 8.98876404494382e-06,
"loss": 2.1052,
"step": 40
},
{
"epoch": 0.06741573033707865,
"grad_norm": 1.21875,
"learning_rate": 1.348314606741573e-05,
"loss": 2.1605,
"step": 60
},
{
"epoch": 0.0898876404494382,
"grad_norm": 1.0234375,
"learning_rate": 1.797752808988764e-05,
"loss": 1.8331,
"step": 80
},
{
"epoch": 0.11235955056179775,
"grad_norm": 0.61328125,
"learning_rate": 2.2471910112359552e-05,
"loss": 1.8186,
"step": 100
},
{
"epoch": 0.1348314606741573,
"grad_norm": 0.703125,
"learning_rate": 2.696629213483146e-05,
"loss": 1.5533,
"step": 120
},
{
"epoch": 0.15730337078651685,
"grad_norm": 0.4921875,
"learning_rate": 3.1460674157303374e-05,
"loss": 1.5419,
"step": 140
},
{
"epoch": 0.1797752808988764,
"grad_norm": 0.470703125,
"learning_rate": 3.595505617977528e-05,
"loss": 1.3374,
"step": 160
},
{
"epoch": 0.20224719101123595,
"grad_norm": 0.52734375,
"learning_rate": 4.044943820224719e-05,
"loss": 1.3418,
"step": 180
},
{
"epoch": 0.2247191011235955,
"grad_norm": 0.59375,
"learning_rate": 4.4943820224719104e-05,
"loss": 1.3921,
"step": 200
},
{
"epoch": 0.24719101123595505,
"grad_norm": 0.67578125,
"learning_rate": 4.943820224719101e-05,
"loss": 1.1401,
"step": 220
},
{
"epoch": 0.2696629213483146,
"grad_norm": 0.72265625,
"learning_rate": 5.393258426966292e-05,
"loss": 1.2139,
"step": 240
},
{
"epoch": 0.29213483146067415,
"grad_norm": 0.49609375,
"learning_rate": 5.8426966292134835e-05,
"loss": 1.171,
"step": 260
},
{
"epoch": 0.3146067415730337,
"grad_norm": 0.94921875,
"learning_rate": 6.292134831460675e-05,
"loss": 1.1424,
"step": 280
},
{
"epoch": 0.33707865168539325,
"grad_norm": 0.8046875,
"learning_rate": 6.741573033707866e-05,
"loss": 1.2171,
"step": 300
},
{
"epoch": 0.3595505617977528,
"grad_norm": 1.09375,
"learning_rate": 7.191011235955056e-05,
"loss": 1.1575,
"step": 320
},
{
"epoch": 0.38202247191011235,
"grad_norm": 1.4453125,
"learning_rate": 7.640449438202247e-05,
"loss": 1.2041,
"step": 340
},
{
"epoch": 0.4044943820224719,
"grad_norm": 1.0,
"learning_rate": 8.089887640449438e-05,
"loss": 1.106,
"step": 360
},
{
"epoch": 0.42696629213483145,
"grad_norm": 1.0390625,
"learning_rate": 8.53932584269663e-05,
"loss": 1.0728,
"step": 380
},
{
"epoch": 0.449438202247191,
"grad_norm": 1.0859375,
"learning_rate": 8.988764044943821e-05,
"loss": 0.9622,
"step": 400
},
{
"epoch": 0.47191011235955055,
"grad_norm": 0.99609375,
"learning_rate": 9.438202247191012e-05,
"loss": 1.0835,
"step": 420
},
{
"epoch": 0.4943820224719101,
"grad_norm": 0.8828125,
"learning_rate": 9.887640449438202e-05,
"loss": 1.0557,
"step": 440
},
{
"epoch": 0.5168539325842697,
"grad_norm": 1.0859375,
"learning_rate": 0.00010337078651685395,
"loss": 1.1037,
"step": 460
},
{
"epoch": 0.5393258426966292,
"grad_norm": 1.3203125,
"learning_rate": 0.00010786516853932584,
"loss": 1.058,
"step": 480
},
{
"epoch": 0.5617977528089888,
"grad_norm": 1.03125,
"learning_rate": 0.00011235955056179777,
"loss": 1.1904,
"step": 500
},
{
"epoch": 0.5617977528089888,
"eval_loss": 1.0617437362670898,
"eval_runtime": 206.3616,
"eval_samples_per_second": 1.856,
"eval_steps_per_second": 1.856,
"step": 500
},
{
"epoch": 0.5842696629213483,
"grad_norm": 0.60546875,
"learning_rate": 0.00011685393258426967,
"loss": 1.1192,
"step": 520
},
{
"epoch": 0.6067415730337079,
"grad_norm": 0.96484375,
"learning_rate": 0.00012134831460674158,
"loss": 1.0356,
"step": 540
},
{
"epoch": 0.6292134831460674,
"grad_norm": 1.53125,
"learning_rate": 0.0001258426966292135,
"loss": 1.0108,
"step": 560
},
{
"epoch": 0.651685393258427,
"grad_norm": 1.453125,
"learning_rate": 0.0001303370786516854,
"loss": 0.8835,
"step": 580
},
{
"epoch": 0.6741573033707865,
"grad_norm": 0.84765625,
"learning_rate": 0.00013483146067415732,
"loss": 0.9925,
"step": 600
},
{
"epoch": 0.6966292134831461,
"grad_norm": 1.0625,
"learning_rate": 0.00013932584269662923,
"loss": 0.9548,
"step": 620
},
{
"epoch": 0.7191011235955056,
"grad_norm": 0.9375,
"learning_rate": 0.00014382022471910112,
"loss": 1.113,
"step": 640
},
{
"epoch": 0.7415730337078652,
"grad_norm": 0.8125,
"learning_rate": 0.00014831460674157306,
"loss": 0.9504,
"step": 660
},
{
"epoch": 0.7640449438202247,
"grad_norm": 0.6953125,
"learning_rate": 0.00015280898876404494,
"loss": 1.1266,
"step": 680
},
{
"epoch": 0.7865168539325843,
"grad_norm": 1.2890625,
"learning_rate": 0.00015730337078651685,
"loss": 1.0707,
"step": 700
},
{
"epoch": 0.8089887640449438,
"grad_norm": 0.88671875,
"learning_rate": 0.00016179775280898877,
"loss": 0.9222,
"step": 720
},
{
"epoch": 0.8314606741573034,
"grad_norm": 0.79296875,
"learning_rate": 0.00016629213483146068,
"loss": 1.1053,
"step": 740
},
{
"epoch": 0.8539325842696629,
"grad_norm": 0.83984375,
"learning_rate": 0.0001707865168539326,
"loss": 1.25,
"step": 760
},
{
"epoch": 0.8764044943820225,
"grad_norm": 0.46875,
"learning_rate": 0.0001752808988764045,
"loss": 0.933,
"step": 780
},
{
"epoch": 0.898876404494382,
"grad_norm": 0.66015625,
"learning_rate": 0.00017977528089887642,
"loss": 0.885,
"step": 800
},
{
"epoch": 0.9213483146067416,
"grad_norm": 1.4140625,
"learning_rate": 0.00018426966292134833,
"loss": 0.9786,
"step": 820
},
{
"epoch": 0.9438202247191011,
"grad_norm": 0.96875,
"learning_rate": 0.00018876404494382024,
"loss": 0.9101,
"step": 840
},
{
"epoch": 0.9662921348314607,
"grad_norm": 0.80859375,
"learning_rate": 0.00019325842696629215,
"loss": 0.9005,
"step": 860
},
{
"epoch": 0.9887640449438202,
"grad_norm": 0.99609375,
"learning_rate": 0.00019775280898876404,
"loss": 0.9485,
"step": 880
},
{
"epoch": 1.0112359550561798,
"grad_norm": 0.9921875,
"learning_rate": 0.00019999610626011892,
"loss": 1.0071,
"step": 900
},
{
"epoch": 1.0337078651685394,
"grad_norm": 0.84375,
"learning_rate": 0.00019996495816039186,
"loss": 0.9138,
"step": 920
},
{
"epoch": 1.0561797752808988,
"grad_norm": 0.69140625,
"learning_rate": 0.00019990267166335664,
"loss": 0.7752,
"step": 940
},
{
"epoch": 1.0786516853932584,
"grad_norm": 0.92578125,
"learning_rate": 0.00019980926617082901,
"loss": 0.9331,
"step": 960
},
{
"epoch": 1.101123595505618,
"grad_norm": 1.046875,
"learning_rate": 0.00019968477077797781,
"loss": 1.0037,
"step": 980
},
{
"epoch": 1.1235955056179776,
"grad_norm": 0.56640625,
"learning_rate": 0.00019952922426426207,
"loss": 0.765,
"step": 1000
},
{
"epoch": 1.1235955056179776,
"eval_loss": 0.944207489490509,
"eval_runtime": 206.3524,
"eval_samples_per_second": 1.856,
"eval_steps_per_second": 1.856,
"step": 1000
},
{
"epoch": 1.146067415730337,
"grad_norm": 1.09375,
"learning_rate": 0.00019934267508135164,
"loss": 0.861,
"step": 1020
},
{
"epoch": 1.1685393258426966,
"grad_norm": 1.25,
"learning_rate": 0.00019912518133803465,
"loss": 0.8251,
"step": 1040
},
{
"epoch": 1.1910112359550562,
"grad_norm": 0.609375,
"learning_rate": 0.00019887681078211707,
"loss": 0.9779,
"step": 1060
},
{
"epoch": 1.2134831460674158,
"grad_norm": 0.765625,
"learning_rate": 0.00019859764077931978,
"loss": 0.8112,
"step": 1080
},
{
"epoch": 1.2359550561797752,
"grad_norm": 0.890625,
"learning_rate": 0.00019828775828917964,
"loss": 0.9084,
"step": 1100
},
{
"epoch": 1.2584269662921348,
"grad_norm": 1.15625,
"learning_rate": 0.00019794725983796218,
"loss": 0.8429,
"step": 1120
},
{
"epoch": 1.2808988764044944,
"grad_norm": 0.7890625,
"learning_rate": 0.00019757625148859441,
"loss": 0.8029,
"step": 1140
},
{
"epoch": 1.303370786516854,
"grad_norm": 1.0078125,
"learning_rate": 0.00019717484880762685,
"loss": 0.9478,
"step": 1160
},
{
"epoch": 1.3258426966292136,
"grad_norm": 0.9765625,
"learning_rate": 0.00019674317682923532,
"loss": 0.6985,
"step": 1180
},
{
"epoch": 1.348314606741573,
"grad_norm": 0.9609375,
"learning_rate": 0.00019628137001627383,
"loss": 0.8653,
"step": 1200
},
{
"epoch": 1.3707865168539326,
"grad_norm": 0.82421875,
"learning_rate": 0.00019578957221839014,
"loss": 0.891,
"step": 1220
},
{
"epoch": 1.3932584269662922,
"grad_norm": 1.0390625,
"learning_rate": 0.00019526793662721768,
"loss": 0.861,
"step": 1240
},
{
"epoch": 1.4157303370786516,
"grad_norm": 0.62890625,
"learning_rate": 0.00019471662572865736,
"loss": 0.7591,
"step": 1260
},
{
"epoch": 1.4382022471910112,
"grad_norm": 0.8828125,
"learning_rate": 0.00019413581125226438,
"loss": 0.7109,
"step": 1280
},
{
"epoch": 1.4606741573033708,
"grad_norm": 0.6875,
"learning_rate": 0.00019352567411775565,
"loss": 0.8947,
"step": 1300
},
{
"epoch": 1.4831460674157304,
"grad_norm": 0.88671875,
"learning_rate": 0.00019288640437865445,
"loss": 0.8514,
"step": 1320
},
{
"epoch": 1.50561797752809,
"grad_norm": 0.69140625,
"learning_rate": 0.0001922182011630902,
"loss": 0.7379,
"step": 1340
},
{
"epoch": 1.5280898876404494,
"grad_norm": 0.63671875,
"learning_rate": 0.00019152127261177126,
"loss": 0.6778,
"step": 1360
},
{
"epoch": 1.550561797752809,
"grad_norm": 0.65234375,
"learning_rate": 0.00019079583581315076,
"loss": 0.6591,
"step": 1380
},
{
"epoch": 1.5730337078651684,
"grad_norm": 0.96484375,
"learning_rate": 0.0001900421167358048,
"loss": 0.8635,
"step": 1400
},
{
"epoch": 1.595505617977528,
"grad_norm": 0.8125,
"learning_rate": 0.00018926035015804488,
"loss": 0.924,
"step": 1420
},
{
"epoch": 1.6179775280898876,
"grad_norm": 1.1953125,
"learning_rate": 0.00018845077959478613,
"loss": 0.8554,
"step": 1440
},
{
"epoch": 1.6404494382022472,
"grad_norm": 0.6171875,
"learning_rate": 0.00018761365722169403,
"loss": 0.9471,
"step": 1460
},
{
"epoch": 1.6629213483146068,
"grad_norm": 0.64453125,
"learning_rate": 0.00018674924379663338,
"loss": 0.9187,
"step": 1480
},
{
"epoch": 1.6853932584269664,
"grad_norm": 0.56640625,
"learning_rate": 0.00018585780857844418,
"loss": 0.782,
"step": 1500
},
{
"epoch": 1.6853932584269664,
"eval_loss": 0.8689672350883484,
"eval_runtime": 206.6921,
"eval_samples_per_second": 1.853,
"eval_steps_per_second": 1.853,
"step": 1500
},
{
"epoch": 1.7078651685393258,
"grad_norm": 0.890625,
"learning_rate": 0.00018493962924306912,
"loss": 0.8983,
"step": 1520
},
{
"epoch": 1.7303370786516854,
"grad_norm": 0.83984375,
"learning_rate": 0.0001839949917970596,
"loss": 0.5218,
"step": 1540
},
{
"epoch": 1.7528089887640448,
"grad_norm": 0.921875,
"learning_rate": 0.00018302419048848667,
"loss": 0.6711,
"step": 1560
},
{
"epoch": 1.7752808988764044,
"grad_norm": 0.7578125,
"learning_rate": 0.0001820275277152846,
"loss": 0.7932,
"step": 1580
},
{
"epoch": 1.797752808988764,
"grad_norm": 0.97265625,
"learning_rate": 0.00018100531393105623,
"loss": 0.7181,
"step": 1600
},
{
"epoch": 1.8202247191011236,
"grad_norm": 0.95703125,
"learning_rate": 0.00017995786754836863,
"loss": 0.8525,
"step": 1620
},
{
"epoch": 1.8426966292134832,
"grad_norm": 0.6484375,
"learning_rate": 0.00017888551483956987,
"loss": 0.6968,
"step": 1640
},
{
"epoch": 1.8651685393258428,
"grad_norm": 0.78515625,
"learning_rate": 0.00017778858983515743,
"loss": 0.902,
"step": 1660
},
{
"epoch": 1.8876404494382022,
"grad_norm": 0.875,
"learning_rate": 0.00017666743421972987,
"loss": 0.954,
"step": 1680
},
{
"epoch": 1.9101123595505618,
"grad_norm": 1.0546875,
"learning_rate": 0.0001755223972255546,
"loss": 0.791,
"step": 1700
},
{
"epoch": 1.9325842696629212,
"grad_norm": 0.765625,
"learning_rate": 0.00017435383552378428,
"loss": 0.77,
"step": 1720
},
{
"epoch": 1.9550561797752808,
"grad_norm": 1.25,
"learning_rate": 0.0001731621131133564,
"loss": 0.6294,
"step": 1740
},
{
"epoch": 1.9775280898876404,
"grad_norm": 0.70703125,
"learning_rate": 0.00017194760120760986,
"loss": 0.7982,
"step": 1760
},
{
"epoch": 2.0,
"grad_norm": 0.67578125,
"learning_rate": 0.00017071067811865476,
"loss": 0.7643,
"step": 1780
},
{
"epoch": 2.0224719101123596,
"grad_norm": 1.0234375,
"learning_rate": 0.0001694517291395307,
"loss": 0.5279,
"step": 1800
},
{
"epoch": 2.044943820224719,
"grad_norm": 1.078125,
"learning_rate": 0.00016817114642419067,
"loss": 0.6667,
"step": 1820
},
{
"epoch": 2.067415730337079,
"grad_norm": 0.9375,
"learning_rate": 0.00016686932886534781,
"loss": 0.6427,
"step": 1840
},
{
"epoch": 2.0898876404494384,
"grad_norm": 0.7890625,
"learning_rate": 0.00016554668197022295,
"loss": 0.633,
"step": 1860
},
{
"epoch": 2.1123595505617976,
"grad_norm": 0.7421875,
"learning_rate": 0.00016420361773423204,
"loss": 0.5623,
"step": 1880
},
{
"epoch": 2.134831460674157,
"grad_norm": 1.078125,
"learning_rate": 0.00016284055451265246,
"loss": 0.6311,
"step": 1900
},
{
"epoch": 2.157303370786517,
"grad_norm": 1.25,
"learning_rate": 0.00016145791689030795,
"loss": 0.7469,
"step": 1920
},
{
"epoch": 2.1797752808988764,
"grad_norm": 0.68359375,
"learning_rate": 0.0001600561355493137,
"loss": 0.7196,
"step": 1940
},
{
"epoch": 2.202247191011236,
"grad_norm": 0.95703125,
"learning_rate": 0.0001586356471349215,
"loss": 0.6328,
"step": 1960
},
{
"epoch": 2.2247191011235956,
"grad_norm": 0.9765625,
"learning_rate": 0.00015719689411950808,
"loss": 0.6349,
"step": 1980
},
{
"epoch": 2.247191011235955,
"grad_norm": 1.140625,
"learning_rate": 0.00015574032466474775,
"loss": 0.5591,
"step": 2000
},
{
"epoch": 2.247191011235955,
"eval_loss": 0.8647096753120422,
"eval_runtime": 206.6878,
"eval_samples_per_second": 1.853,
"eval_steps_per_second": 1.853,
"step": 2000
},
{
"epoch": 2.2696629213483144,
"grad_norm": 1.1328125,
"learning_rate": 0.00015426639248201313,
"loss": 0.5206,
"step": 2020
},
{
"epoch": 2.292134831460674,
"grad_norm": 0.8125,
"learning_rate": 0.0001527755566910474,
"loss": 0.7186,
"step": 2040
},
{
"epoch": 2.3146067415730336,
"grad_norm": 0.97265625,
"learning_rate": 0.00015126828167695146,
"loss": 0.6533,
"step": 2060
},
{
"epoch": 2.337078651685393,
"grad_norm": 1.4609375,
"learning_rate": 0.0001497450369455312,
"loss": 0.6324,
"step": 2080
},
{
"epoch": 2.359550561797753,
"grad_norm": 0.875,
"learning_rate": 0.00014820629697704965,
"loss": 0.5276,
"step": 2100
},
{
"epoch": 2.3820224719101124,
"grad_norm": 0.74609375,
"learning_rate": 0.00014665254107842964,
"loss": 0.612,
"step": 2120
},
{
"epoch": 2.404494382022472,
"grad_norm": 0.92578125,
"learning_rate": 0.00014508425323395317,
"loss": 0.614,
"step": 2140
},
{
"epoch": 2.4269662921348316,
"grad_norm": 0.9921875,
"learning_rate": 0.0001435019219545034,
"loss": 0.4988,
"step": 2160
},
{
"epoch": 2.449438202247191,
"grad_norm": 0.8515625,
"learning_rate": 0.00014190604012539684,
"loss": 0.6777,
"step": 2180
},
{
"epoch": 2.4719101123595504,
"grad_norm": 0.96875,
"learning_rate": 0.00014029710485285324,
"loss": 0.662,
"step": 2200
},
{
"epoch": 2.49438202247191,
"grad_norm": 0.74609375,
"learning_rate": 0.00013867561730915016,
"loss": 0.6087,
"step": 2220
},
{
"epoch": 2.5168539325842696,
"grad_norm": 0.875,
"learning_rate": 0.0001370420825765114,
"loss": 0.56,
"step": 2240
},
{
"epoch": 2.539325842696629,
"grad_norm": 1.2421875,
"learning_rate": 0.00013539700948977717,
"loss": 0.572,
"step": 2260
},
{
"epoch": 2.561797752808989,
"grad_norm": 1.421875,
"learning_rate": 0.00013374091047790585,
"loss": 0.7334,
"step": 2280
},
{
"epoch": 2.5842696629213484,
"grad_norm": 1.3125,
"learning_rate": 0.00013207430140435556,
"loss": 0.5377,
"step": 2300
},
{
"epoch": 2.606741573033708,
"grad_norm": 0.86328125,
"learning_rate": 0.00013039770140639654,
"loss": 0.6306,
"step": 2320
},
{
"epoch": 2.629213483146067,
"grad_norm": 1.3046875,
"learning_rate": 0.00012871163273340307,
"loss": 0.582,
"step": 2340
},
{
"epoch": 2.6516853932584272,
"grad_norm": 0.83203125,
"learning_rate": 0.00012701662058417688,
"loss": 0.6326,
"step": 2360
},
{
"epoch": 2.6741573033707864,
"grad_norm": 1.1640625,
"learning_rate": 0.00012531319294335086,
"loss": 0.6907,
"step": 2380
},
{
"epoch": 2.696629213483146,
"grad_norm": 1.2734375,
"learning_rate": 0.00012360188041692582,
"loss": 0.656,
"step": 2400
},
{
"epoch": 2.7191011235955056,
"grad_norm": 1.234375,
"learning_rate": 0.00012188321606699016,
"loss": 0.5817,
"step": 2420
},
{
"epoch": 2.741573033707865,
"grad_norm": 0.97265625,
"learning_rate": 0.00012015773524567479,
"loss": 0.5046,
"step": 2440
},
{
"epoch": 2.764044943820225,
"grad_norm": 1.3515625,
"learning_rate": 0.00011842597542839462,
"loss": 0.6293,
"step": 2460
},
{
"epoch": 2.7865168539325844,
"grad_norm": 1.1796875,
"learning_rate": 0.00011668847604642861,
"loss": 0.6067,
"step": 2480
},
{
"epoch": 2.808988764044944,
"grad_norm": 1.1484375,
"learning_rate": 0.00011494577831889067,
"loss": 0.5669,
"step": 2500
},
{
"epoch": 2.808988764044944,
"eval_loss": 0.8295581340789795,
"eval_runtime": 206.704,
"eval_samples_per_second": 1.853,
"eval_steps_per_second": 1.853,
"step": 2500
},
{
"epoch": 2.831460674157303,
"grad_norm": 1.1484375,
"learning_rate": 0.00011319842508414365,
"loss": 0.5429,
"step": 2520
},
{
"epoch": 2.853932584269663,
"grad_norm": 1.03125,
"learning_rate": 0.00011144696063070883,
"loss": 0.5481,
"step": 2540
},
{
"epoch": 2.8764044943820224,
"grad_norm": 0.78515625,
"learning_rate": 0.00010969193052772396,
"loss": 0.5308,
"step": 2560
},
{
"epoch": 2.898876404494382,
"grad_norm": 0.703125,
"learning_rate": 0.00010793388145500198,
"loss": 0.4527,
"step": 2580
},
{
"epoch": 2.9213483146067416,
"grad_norm": 1.390625,
"learning_rate": 0.00010617336103274424,
"loss": 0.5333,
"step": 2600
},
{
"epoch": 2.943820224719101,
"grad_norm": 0.9296875,
"learning_rate": 0.00010441091765096047,
"loss": 0.5886,
"step": 2620
},
{
"epoch": 2.966292134831461,
"grad_norm": 1.296875,
"learning_rate": 0.0001026471002986491,
"loss": 0.626,
"step": 2640
},
{
"epoch": 2.98876404494382,
"grad_norm": 0.953125,
"learning_rate": 0.00010088245839279082,
"loss": 0.6703,
"step": 2660
},
{
"epoch": 3.0112359550561796,
"grad_norm": 0.64453125,
"learning_rate": 9.911754160720923e-05,
"loss": 0.4819,
"step": 2680
},
{
"epoch": 3.033707865168539,
"grad_norm": 0.71875,
"learning_rate": 9.735289970135095e-05,
"loss": 0.4379,
"step": 2700
},
{
"epoch": 3.056179775280899,
"grad_norm": 1.7734375,
"learning_rate": 9.558908234903954e-05,
"loss": 0.3811,
"step": 2720
},
{
"epoch": 3.0786516853932584,
"grad_norm": 0.6953125,
"learning_rate": 9.382663896725578e-05,
"loss": 0.3855,
"step": 2740
},
{
"epoch": 3.101123595505618,
"grad_norm": 1.4140625,
"learning_rate": 9.206611854499805e-05,
"loss": 0.4749,
"step": 2760
},
{
"epoch": 3.1235955056179776,
"grad_norm": 0.609375,
"learning_rate": 9.030806947227607e-05,
"loss": 0.501,
"step": 2780
},
{
"epoch": 3.146067415730337,
"grad_norm": 1.1015625,
"learning_rate": 8.855303936929117e-05,
"loss": 0.4239,
"step": 2800
},
{
"epoch": 3.168539325842697,
"grad_norm": 0.84375,
"learning_rate": 8.680157491585636e-05,
"loss": 0.5388,
"step": 2820
},
{
"epoch": 3.191011235955056,
"grad_norm": 1.2578125,
"learning_rate": 8.505422168110934e-05,
"loss": 0.3715,
"step": 2840
},
{
"epoch": 3.2134831460674156,
"grad_norm": 0.94921875,
"learning_rate": 8.331152395357141e-05,
"loss": 0.4274,
"step": 2860
},
{
"epoch": 3.235955056179775,
"grad_norm": 1.046875,
"learning_rate": 8.157402457160539e-05,
"loss": 0.4368,
"step": 2880
},
{
"epoch": 3.258426966292135,
"grad_norm": 1.2421875,
"learning_rate": 7.984226475432522e-05,
"loss": 0.4026,
"step": 2900
},
{
"epoch": 3.2808988764044944,
"grad_norm": 0.8359375,
"learning_rate": 7.811678393300987e-05,
"loss": 0.3971,
"step": 2920
},
{
"epoch": 3.303370786516854,
"grad_norm": 0.94921875,
"learning_rate": 7.63981195830742e-05,
"loss": 0.395,
"step": 2940
},
{
"epoch": 3.3258426966292136,
"grad_norm": 1.328125,
"learning_rate": 7.468680705664914e-05,
"loss": 0.4165,
"step": 2960
},
{
"epoch": 3.348314606741573,
"grad_norm": 0.88671875,
"learning_rate": 7.298337941582314e-05,
"loss": 0.4071,
"step": 2980
},
{
"epoch": 3.370786516853933,
"grad_norm": 1.4296875,
"learning_rate": 7.128836726659696e-05,
"loss": 0.4205,
"step": 3000
},
{
"epoch": 3.370786516853933,
"eval_loss": 0.8820343613624573,
"eval_runtime": 206.8369,
"eval_samples_per_second": 1.852,
"eval_steps_per_second": 1.852,
"step": 3000
},
{
"epoch": 3.393258426966292,
"grad_norm": 1.46875,
"learning_rate": 6.960229859360353e-05,
"loss": 0.3759,
"step": 3020
},
{
"epoch": 3.4157303370786516,
"grad_norm": 1.2265625,
"learning_rate": 6.792569859564445e-05,
"loss": 0.4457,
"step": 3040
},
{
"epoch": 3.438202247191011,
"grad_norm": 1.5390625,
"learning_rate": 6.625908952209418e-05,
"loss": 0.4088,
"step": 3060
},
{
"epoch": 3.460674157303371,
"grad_norm": 0.92578125,
"learning_rate": 6.460299051022285e-05,
"loss": 0.4221,
"step": 3080
},
{
"epoch": 3.4831460674157304,
"grad_norm": 1.34375,
"learning_rate": 6.295791742348865e-05,
"loss": 0.4304,
"step": 3100
},
{
"epoch": 3.50561797752809,
"grad_norm": 1.0,
"learning_rate": 6.132438269084985e-05,
"loss": 0.3612,
"step": 3120
},
{
"epoch": 3.5280898876404496,
"grad_norm": 1.4609375,
"learning_rate": 5.970289514714677e-05,
"loss": 0.4692,
"step": 3140
},
{
"epoch": 3.550561797752809,
"grad_norm": 1.578125,
"learning_rate": 5.8093959874603176e-05,
"loss": 0.4579,
"step": 3160
},
{
"epoch": 3.5730337078651684,
"grad_norm": 0.66796875,
"learning_rate": 5.649807804549663e-05,
"loss": 0.3754,
"step": 3180
},
{
"epoch": 3.595505617977528,
"grad_norm": 1.1328125,
"learning_rate": 5.491574676604682e-05,
"loss": 0.3685,
"step": 3200
},
{
"epoch": 3.6179775280898876,
"grad_norm": 1.09375,
"learning_rate": 5.334745892157035e-05,
"loss": 0.3809,
"step": 3220
},
{
"epoch": 3.640449438202247,
"grad_norm": 1.4453125,
"learning_rate": 5.179370302295037e-05,
"loss": 0.4809,
"step": 3240
},
{
"epoch": 3.662921348314607,
"grad_norm": 1.015625,
"learning_rate": 5.02549630544688e-05,
"loss": 0.3798,
"step": 3260
},
{
"epoch": 3.6853932584269664,
"grad_norm": 1.1796875,
"learning_rate": 4.8731718323048516e-05,
"loss": 0.4153,
"step": 3280
},
{
"epoch": 3.7078651685393256,
"grad_norm": 0.98046875,
"learning_rate": 4.722444330895256e-05,
"loss": 0.4612,
"step": 3300
},
{
"epoch": 3.7303370786516856,
"grad_norm": 1.4921875,
"learning_rate": 4.573360751798689e-05,
"loss": 0.469,
"step": 3320
},
{
"epoch": 3.752808988764045,
"grad_norm": 1.8203125,
"learning_rate": 4.425967533525229e-05,
"loss": 0.4523,
"step": 3340
},
{
"epoch": 3.7752808988764044,
"grad_norm": 1.359375,
"learning_rate": 4.2803105880491925e-05,
"loss": 0.4214,
"step": 3360
},
{
"epoch": 3.797752808988764,
"grad_norm": 1.734375,
"learning_rate": 4.136435286507849e-05,
"loss": 0.4981,
"step": 3380
},
{
"epoch": 3.8202247191011236,
"grad_norm": 1.3515625,
"learning_rate": 3.994386445068632e-05,
"loss": 0.4029,
"step": 3400
},
{
"epoch": 3.842696629213483,
"grad_norm": 1.4609375,
"learning_rate": 3.854208310969204e-05,
"loss": 0.3747,
"step": 3420
},
{
"epoch": 3.865168539325843,
"grad_norm": 1.1015625,
"learning_rate": 3.715944548734755e-05,
"loss": 0.4113,
"step": 3440
},
{
"epoch": 3.8876404494382024,
"grad_norm": 0.52734375,
"learning_rate": 3.5796382265767937e-05,
"loss": 0.3896,
"step": 3460
},
{
"epoch": 3.9101123595505616,
"grad_norm": 1.0625,
"learning_rate": 3.445331802977709e-05,
"loss": 0.4709,
"step": 3480
},
{
"epoch": 3.932584269662921,
"grad_norm": 1.4296875,
"learning_rate": 3.313067113465222e-05,
"loss": 0.3812,
"step": 3500
},
{
"epoch": 3.932584269662921,
"eval_loss": 0.8859002590179443,
"eval_runtime": 206.8365,
"eval_samples_per_second": 1.852,
"eval_steps_per_second": 1.852,
"step": 3500
},
{
"epoch": 3.955056179775281,
"grad_norm": 1.09375,
"learning_rate": 3.182885357580934e-05,
"loss": 0.3906,
"step": 3520
},
{
"epoch": 3.9775280898876404,
"grad_norm": 1.0390625,
"learning_rate": 3.054827086046931e-05,
"loss": 0.3987,
"step": 3540
},
{
"epoch": 4.0,
"grad_norm": 0.671875,
"learning_rate": 2.9289321881345254e-05,
"loss": 0.3499,
"step": 3560
},
{
"epoch": 4.022471910112359,
"grad_norm": 0.79296875,
"learning_rate": 2.8052398792390154e-05,
"loss": 0.3292,
"step": 3580
},
{
"epoch": 4.044943820224719,
"grad_norm": 0.8125,
"learning_rate": 2.6837886886643614e-05,
"loss": 0.3343,
"step": 3600
},
{
"epoch": 4.067415730337078,
"grad_norm": 0.66015625,
"learning_rate": 2.5646164476215716e-05,
"loss": 0.3236,
"step": 3620
},
{
"epoch": 4.089887640449438,
"grad_norm": 1.25,
"learning_rate": 2.447760277444543e-05,
"loss": 0.2892,
"step": 3640
},
{
"epoch": 4.112359550561798,
"grad_norm": 1.3515625,
"learning_rate": 2.3332565780270165e-05,
"loss": 0.2801,
"step": 3660
},
{
"epoch": 4.134831460674158,
"grad_norm": 1.3671875,
"learning_rate": 2.2211410164842606e-05,
"loss": 0.3082,
"step": 3680
},
{
"epoch": 4.157303370786517,
"grad_norm": 1.46875,
"learning_rate": 2.1114485160430132e-05,
"loss": 0.3128,
"step": 3700
},
{
"epoch": 4.179775280898877,
"grad_norm": 0.8515625,
"learning_rate": 2.0042132451631378e-05,
"loss": 0.3846,
"step": 3720
},
{
"epoch": 4.202247191011236,
"grad_norm": 1.1328125,
"learning_rate": 1.899468606894379e-05,
"loss": 0.2718,
"step": 3740
},
{
"epoch": 4.224719101123595,
"grad_norm": 0.80859375,
"learning_rate": 1.7972472284715415e-05,
"loss": 0.302,
"step": 3760
},
{
"epoch": 4.247191011235955,
"grad_norm": 1.6953125,
"learning_rate": 1.6975809511513353e-05,
"loss": 0.3785,
"step": 3780
},
{
"epoch": 4.269662921348314,
"grad_norm": 1.0078125,
"learning_rate": 1.600500820294041e-05,
"loss": 0.3845,
"step": 3800
},
{
"epoch": 4.292134831460674,
"grad_norm": 1.6796875,
"learning_rate": 1.5060370756930919e-05,
"loss": 0.327,
"step": 3820
},
{
"epoch": 4.314606741573034,
"grad_norm": 1.359375,
"learning_rate": 1.414219142155585e-05,
"loss": 0.3589,
"step": 3840
},
{
"epoch": 4.337078651685394,
"grad_norm": 0.75,
"learning_rate": 1.3250756203366632e-05,
"loss": 0.4057,
"step": 3860
},
{
"epoch": 4.359550561797753,
"grad_norm": 0.96484375,
"learning_rate": 1.2386342778305993e-05,
"loss": 0.3862,
"step": 3880
},
{
"epoch": 4.382022471910112,
"grad_norm": 1.4921875,
"learning_rate": 1.1549220405213878e-05,
"loss": 0.3319,
"step": 3900
},
{
"epoch": 4.404494382022472,
"grad_norm": 1.2890625,
"learning_rate": 1.0739649841955136e-05,
"loss": 0.2832,
"step": 3920
},
{
"epoch": 4.426966292134831,
"grad_norm": 1.234375,
"learning_rate": 9.957883264195223e-06,
"loss": 0.2732,
"step": 3940
},
{
"epoch": 4.449438202247191,
"grad_norm": 0.828125,
"learning_rate": 9.20416418684924e-06,
"loss": 0.2587,
"step": 3960
},
{
"epoch": 4.47191011235955,
"grad_norm": 1.5234375,
"learning_rate": 8.478727388228735e-06,
"loss": 0.3469,
"step": 3980
},
{
"epoch": 4.49438202247191,
"grad_norm": 0.90234375,
"learning_rate": 7.781798836909826e-06,
"loss": 0.3323,
"step": 4000
},
{
"epoch": 4.49438202247191,
"eval_loss": 0.9360187649726868,
"eval_runtime": 206.9644,
"eval_samples_per_second": 1.851,
"eval_steps_per_second": 1.851,
"step": 4000
},
{
"epoch": 4.51685393258427,
"grad_norm": 2.046875,
"learning_rate": 7.11359562134557e-06,
"loss": 0.3441,
"step": 4020
},
{
"epoch": 4.539325842696629,
"grad_norm": 1.296875,
"learning_rate": 6.4743258822443695e-06,
"loss": 0.3196,
"step": 4040
},
{
"epoch": 4.561797752808989,
"grad_norm": 1.515625,
"learning_rate": 5.8641887477356215e-06,
"loss": 0.3226,
"step": 4060
},
{
"epoch": 4.584269662921348,
"grad_norm": 1.6875,
"learning_rate": 5.283374271342645e-06,
"loss": 0.2859,
"step": 4080
},
{
"epoch": 4.606741573033708,
"grad_norm": 0.87109375,
"learning_rate": 4.732063372782336e-06,
"loss": 0.3164,
"step": 4100
},
{
"epoch": 4.629213483146067,
"grad_norm": 1.2734375,
"learning_rate": 4.210427781609861e-06,
"loss": 0.3275,
"step": 4120
},
{
"epoch": 4.651685393258427,
"grad_norm": 1.25,
"learning_rate": 3.718629983726185e-06,
"loss": 0.367,
"step": 4140
},
{
"epoch": 4.674157303370786,
"grad_norm": 1.6796875,
"learning_rate": 3.256823170764689e-06,
"loss": 0.3445,
"step": 4160
},
{
"epoch": 4.696629213483146,
"grad_norm": 0.9453125,
"learning_rate": 2.8251511923731655e-06,
"loss": 0.4628,
"step": 4180
},
{
"epoch": 4.719101123595506,
"grad_norm": 0.8515625,
"learning_rate": 2.423748511405577e-06,
"loss": 0.3252,
"step": 4200
},
{
"epoch": 4.741573033707866,
"grad_norm": 0.90234375,
"learning_rate": 2.052740162037814e-06,
"loss": 0.3783,
"step": 4220
},
{
"epoch": 4.764044943820225,
"grad_norm": 0.7421875,
"learning_rate": 1.7122417108203726e-06,
"loss": 0.294,
"step": 4240
},
{
"epoch": 4.786516853932584,
"grad_norm": 0.75390625,
"learning_rate": 1.4023592206802382e-06,
"loss": 0.3194,
"step": 4260
},
{
"epoch": 4.808988764044944,
"grad_norm": 1.375,
"learning_rate": 1.1231892178829472e-06,
"loss": 0.3145,
"step": 4280
},
{
"epoch": 4.831460674157303,
"grad_norm": 1.4140625,
"learning_rate": 8.74818661965382e-07,
"loss": 0.3372,
"step": 4300
},
{
"epoch": 4.853932584269663,
"grad_norm": 1.6640625,
"learning_rate": 6.573249186483721e-07,
"loss": 0.2791,
"step": 4320
},
{
"epoch": 4.876404494382022,
"grad_norm": 0.8359375,
"learning_rate": 4.707757357379383e-07,
"loss": 0.2428,
"step": 4340
},
{
"epoch": 4.898876404494382,
"grad_norm": 1.3671875,
"learning_rate": 3.152292220222064e-07,
"loss": 0.3225,
"step": 4360
},
{
"epoch": 4.921348314606742,
"grad_norm": 1.0625,
"learning_rate": 1.9073382917097483e-07,
"loss": 0.3164,
"step": 4380
},
{
"epoch": 4.943820224719101,
"grad_norm": 1.0546875,
"learning_rate": 9.732833664334307e-08,
"loss": 0.3571,
"step": 4400
},
{
"epoch": 4.966292134831461,
"grad_norm": 0.87890625,
"learning_rate": 3.5041839608151996e-08,
"loss": 0.3002,
"step": 4420
},
{
"epoch": 4.98876404494382,
"grad_norm": 0.6640625,
"learning_rate": 3.893739881088987e-09,
"loss": 0.366,
"step": 4440
},
{
"epoch": 5.0,
"step": 4450,
"total_flos": 1.024663401529344e+17,
"train_loss": 0.6764815047617708,
"train_runtime": 9686.8536,
"train_samples_per_second": 0.459,
"train_steps_per_second": 0.459
}
],
"logging_steps": 20,
"max_steps": 4450,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.024663401529344e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}