TrelisSmolLM-base / trainer_state.json
rs545837's picture
Upload folder using huggingface_hub
c6f9d28 verified
raw
history blame
No virus
162 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9999842829076621,
"eval_steps": 1590,
"global_step": 15906,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0009430255402750491,
"grad_norm": 0.390625,
"learning_rate": 0.001,
"loss": 5.5551,
"step": 15
},
{
"epoch": 0.0018860510805500982,
"grad_norm": 0.0791015625,
"learning_rate": 0.001,
"loss": 3.5038,
"step": 30
},
{
"epoch": 0.002829076620825147,
"grad_norm": 0.1298828125,
"learning_rate": 0.001,
"loss": 3.5068,
"step": 45
},
{
"epoch": 0.0037721021611001964,
"grad_norm": 0.0693359375,
"learning_rate": 0.001,
"loss": 3.4288,
"step": 60
},
{
"epoch": 0.004715127701375246,
"grad_norm": 0.12255859375,
"learning_rate": 0.001,
"loss": 3.3071,
"step": 75
},
{
"epoch": 0.005658153241650294,
"grad_norm": 0.091796875,
"learning_rate": 0.001,
"loss": 3.2653,
"step": 90
},
{
"epoch": 0.006601178781925344,
"grad_norm": 0.1318359375,
"learning_rate": 0.001,
"loss": 3.1297,
"step": 105
},
{
"epoch": 0.007544204322200393,
"grad_norm": 0.12451171875,
"learning_rate": 0.001,
"loss": 3.0482,
"step": 120
},
{
"epoch": 0.008487229862475442,
"grad_norm": 0.380859375,
"learning_rate": 0.001,
"loss": 2.9037,
"step": 135
},
{
"epoch": 0.009430255402750491,
"grad_norm": 0.1650390625,
"learning_rate": 0.001,
"loss": 2.8178,
"step": 150
},
{
"epoch": 0.01037328094302554,
"grad_norm": 0.111328125,
"learning_rate": 0.001,
"loss": 2.687,
"step": 165
},
{
"epoch": 0.011316306483300589,
"grad_norm": 0.1640625,
"learning_rate": 0.001,
"loss": 2.6247,
"step": 180
},
{
"epoch": 0.01225933202357564,
"grad_norm": 0.1298828125,
"learning_rate": 0.001,
"loss": 2.5556,
"step": 195
},
{
"epoch": 0.013202357563850688,
"grad_norm": 0.2451171875,
"learning_rate": 0.001,
"loss": 2.4524,
"step": 210
},
{
"epoch": 0.014145383104125737,
"grad_norm": 0.1083984375,
"learning_rate": 0.001,
"loss": 2.4904,
"step": 225
},
{
"epoch": 0.015088408644400786,
"grad_norm": 0.1904296875,
"learning_rate": 0.001,
"loss": 2.4211,
"step": 240
},
{
"epoch": 0.016031434184675834,
"grad_norm": 0.2412109375,
"learning_rate": 0.001,
"loss": 2.419,
"step": 255
},
{
"epoch": 0.016974459724950885,
"grad_norm": 0.130859375,
"learning_rate": 0.001,
"loss": 2.3542,
"step": 270
},
{
"epoch": 0.017917485265225932,
"grad_norm": 0.294921875,
"learning_rate": 0.001,
"loss": 2.2893,
"step": 285
},
{
"epoch": 0.018860510805500982,
"grad_norm": 0.1982421875,
"learning_rate": 0.001,
"loss": 2.2671,
"step": 300
},
{
"epoch": 0.019803536345776033,
"grad_norm": 0.2109375,
"learning_rate": 0.001,
"loss": 2.2644,
"step": 315
},
{
"epoch": 0.02074656188605108,
"grad_norm": 0.306640625,
"learning_rate": 0.001,
"loss": 2.2669,
"step": 330
},
{
"epoch": 0.02168958742632613,
"grad_norm": 0.1962890625,
"learning_rate": 0.001,
"loss": 2.2009,
"step": 345
},
{
"epoch": 0.022632612966601177,
"grad_norm": 0.197265625,
"learning_rate": 0.001,
"loss": 2.1569,
"step": 360
},
{
"epoch": 0.023575638506876228,
"grad_norm": 0.203125,
"learning_rate": 0.001,
"loss": 2.0607,
"step": 375
},
{
"epoch": 0.02451866404715128,
"grad_norm": 0.330078125,
"learning_rate": 0.001,
"loss": 2.1118,
"step": 390
},
{
"epoch": 0.025461689587426325,
"grad_norm": 0.4453125,
"learning_rate": 0.001,
"loss": 2.0465,
"step": 405
},
{
"epoch": 0.026404715127701376,
"grad_norm": 0.26953125,
"learning_rate": 0.001,
"loss": 2.0682,
"step": 420
},
{
"epoch": 0.027347740667976423,
"grad_norm": 0.158203125,
"learning_rate": 0.001,
"loss": 2.014,
"step": 435
},
{
"epoch": 0.028290766208251474,
"grad_norm": 0.2578125,
"learning_rate": 0.001,
"loss": 2.0251,
"step": 450
},
{
"epoch": 0.029233791748526524,
"grad_norm": 0.314453125,
"learning_rate": 0.001,
"loss": 1.991,
"step": 465
},
{
"epoch": 0.03017681728880157,
"grad_norm": 0.1884765625,
"learning_rate": 0.001,
"loss": 1.9579,
"step": 480
},
{
"epoch": 0.03111984282907662,
"grad_norm": 0.1630859375,
"learning_rate": 0.001,
"loss": 1.9253,
"step": 495
},
{
"epoch": 0.03206286836935167,
"grad_norm": 0.408203125,
"learning_rate": 0.001,
"loss": 1.9019,
"step": 510
},
{
"epoch": 0.033005893909626716,
"grad_norm": 0.15234375,
"learning_rate": 0.001,
"loss": 1.9208,
"step": 525
},
{
"epoch": 0.03394891944990177,
"grad_norm": 0.474609375,
"learning_rate": 0.001,
"loss": 1.9165,
"step": 540
},
{
"epoch": 0.03489194499017682,
"grad_norm": 0.2734375,
"learning_rate": 0.001,
"loss": 1.8541,
"step": 555
},
{
"epoch": 0.035834970530451864,
"grad_norm": 0.5625,
"learning_rate": 0.001,
"loss": 1.8854,
"step": 570
},
{
"epoch": 0.03677799607072692,
"grad_norm": 0.30078125,
"learning_rate": 0.001,
"loss": 1.8651,
"step": 585
},
{
"epoch": 0.037721021611001965,
"grad_norm": 0.2734375,
"learning_rate": 0.001,
"loss": 1.8392,
"step": 600
},
{
"epoch": 0.03866404715127701,
"grad_norm": 0.23828125,
"learning_rate": 0.001,
"loss": 1.843,
"step": 615
},
{
"epoch": 0.039607072691552066,
"grad_norm": 0.2578125,
"learning_rate": 0.001,
"loss": 1.7958,
"step": 630
},
{
"epoch": 0.04055009823182711,
"grad_norm": 0.197265625,
"learning_rate": 0.001,
"loss": 1.7849,
"step": 645
},
{
"epoch": 0.04149312377210216,
"grad_norm": 0.1767578125,
"learning_rate": 0.001,
"loss": 1.7397,
"step": 660
},
{
"epoch": 0.04243614931237721,
"grad_norm": 0.255859375,
"learning_rate": 0.001,
"loss": 1.7396,
"step": 675
},
{
"epoch": 0.04337917485265226,
"grad_norm": 0.291015625,
"learning_rate": 0.001,
"loss": 1.7219,
"step": 690
},
{
"epoch": 0.04432220039292731,
"grad_norm": 0.33984375,
"learning_rate": 0.001,
"loss": 1.7536,
"step": 705
},
{
"epoch": 0.045265225933202355,
"grad_norm": 0.2021484375,
"learning_rate": 0.001,
"loss": 1.697,
"step": 720
},
{
"epoch": 0.04620825147347741,
"grad_norm": 0.421875,
"learning_rate": 0.001,
"loss": 1.6725,
"step": 735
},
{
"epoch": 0.047151277013752456,
"grad_norm": 0.2080078125,
"learning_rate": 0.001,
"loss": 1.691,
"step": 750
},
{
"epoch": 0.0480943025540275,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 1.6721,
"step": 765
},
{
"epoch": 0.04903732809430256,
"grad_norm": 0.2451171875,
"learning_rate": 0.001,
"loss": 1.7221,
"step": 780
},
{
"epoch": 0.049980353634577604,
"grad_norm": 0.244140625,
"learning_rate": 0.001,
"loss": 1.6609,
"step": 795
},
{
"epoch": 0.05092337917485265,
"grad_norm": 0.494140625,
"learning_rate": 0.001,
"loss": 1.6805,
"step": 810
},
{
"epoch": 0.0518664047151277,
"grad_norm": 0.265625,
"learning_rate": 0.001,
"loss": 1.6157,
"step": 825
},
{
"epoch": 0.05280943025540275,
"grad_norm": 0.19921875,
"learning_rate": 0.001,
"loss": 1.5996,
"step": 840
},
{
"epoch": 0.0537524557956778,
"grad_norm": 0.419921875,
"learning_rate": 0.001,
"loss": 1.5686,
"step": 855
},
{
"epoch": 0.054695481335952846,
"grad_norm": 0.345703125,
"learning_rate": 0.001,
"loss": 1.6021,
"step": 870
},
{
"epoch": 0.0556385068762279,
"grad_norm": 0.447265625,
"learning_rate": 0.001,
"loss": 1.6159,
"step": 885
},
{
"epoch": 0.05658153241650295,
"grad_norm": 0.314453125,
"learning_rate": 0.001,
"loss": 1.5456,
"step": 900
},
{
"epoch": 0.057524557956777994,
"grad_norm": 0.287109375,
"learning_rate": 0.001,
"loss": 1.5764,
"step": 915
},
{
"epoch": 0.05846758349705305,
"grad_norm": 0.369140625,
"learning_rate": 0.001,
"loss": 1.5426,
"step": 930
},
{
"epoch": 0.059410609037328095,
"grad_norm": 0.380859375,
"learning_rate": 0.001,
"loss": 1.5535,
"step": 945
},
{
"epoch": 0.06035363457760314,
"grad_norm": 0.23828125,
"learning_rate": 0.001,
"loss": 1.505,
"step": 960
},
{
"epoch": 0.06129666011787819,
"grad_norm": 0.337890625,
"learning_rate": 0.001,
"loss": 1.5328,
"step": 975
},
{
"epoch": 0.06223968565815324,
"grad_norm": 0.45703125,
"learning_rate": 0.001,
"loss": 1.5274,
"step": 990
},
{
"epoch": 0.06318271119842829,
"grad_norm": 0.3515625,
"learning_rate": 0.001,
"loss": 1.5246,
"step": 1005
},
{
"epoch": 0.06412573673870334,
"grad_norm": 0.46875,
"learning_rate": 0.001,
"loss": 1.4633,
"step": 1020
},
{
"epoch": 0.06506876227897838,
"grad_norm": 0.3046875,
"learning_rate": 0.001,
"loss": 1.487,
"step": 1035
},
{
"epoch": 0.06601178781925343,
"grad_norm": 0.380859375,
"learning_rate": 0.001,
"loss": 1.4582,
"step": 1050
},
{
"epoch": 0.06695481335952849,
"grad_norm": 0.39453125,
"learning_rate": 0.001,
"loss": 1.4586,
"step": 1065
},
{
"epoch": 0.06789783889980354,
"grad_norm": 0.58984375,
"learning_rate": 0.001,
"loss": 1.4322,
"step": 1080
},
{
"epoch": 0.06884086444007859,
"grad_norm": 0.330078125,
"learning_rate": 0.001,
"loss": 1.47,
"step": 1095
},
{
"epoch": 0.06978388998035363,
"grad_norm": 0.443359375,
"learning_rate": 0.001,
"loss": 1.4215,
"step": 1110
},
{
"epoch": 0.07072691552062868,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 1.4569,
"step": 1125
},
{
"epoch": 0.07166994106090373,
"grad_norm": 0.330078125,
"learning_rate": 0.001,
"loss": 1.4428,
"step": 1140
},
{
"epoch": 0.07261296660117879,
"grad_norm": 0.3203125,
"learning_rate": 0.001,
"loss": 1.3861,
"step": 1155
},
{
"epoch": 0.07355599214145384,
"grad_norm": 0.427734375,
"learning_rate": 0.001,
"loss": 1.4478,
"step": 1170
},
{
"epoch": 0.07449901768172888,
"grad_norm": 0.27734375,
"learning_rate": 0.001,
"loss": 1.406,
"step": 1185
},
{
"epoch": 0.07544204322200393,
"grad_norm": 0.478515625,
"learning_rate": 0.001,
"loss": 1.3944,
"step": 1200
},
{
"epoch": 0.07638506876227898,
"grad_norm": 0.451171875,
"learning_rate": 0.001,
"loss": 1.3884,
"step": 1215
},
{
"epoch": 0.07732809430255402,
"grad_norm": 0.361328125,
"learning_rate": 0.001,
"loss": 1.38,
"step": 1230
},
{
"epoch": 0.07827111984282907,
"grad_norm": 0.5,
"learning_rate": 0.001,
"loss": 1.3446,
"step": 1245
},
{
"epoch": 0.07921414538310413,
"grad_norm": 0.28515625,
"learning_rate": 0.001,
"loss": 1.351,
"step": 1260
},
{
"epoch": 0.08015717092337918,
"grad_norm": 0.6953125,
"learning_rate": 0.001,
"loss": 1.352,
"step": 1275
},
{
"epoch": 0.08110019646365423,
"grad_norm": 0.296875,
"learning_rate": 0.001,
"loss": 1.3378,
"step": 1290
},
{
"epoch": 0.08204322200392927,
"grad_norm": 0.353515625,
"learning_rate": 0.001,
"loss": 1.3056,
"step": 1305
},
{
"epoch": 0.08298624754420432,
"grad_norm": 0.439453125,
"learning_rate": 0.001,
"loss": 1.3099,
"step": 1320
},
{
"epoch": 0.08392927308447937,
"grad_norm": 0.59375,
"learning_rate": 0.001,
"loss": 1.3364,
"step": 1335
},
{
"epoch": 0.08487229862475441,
"grad_norm": 0.4296875,
"learning_rate": 0.001,
"loss": 1.2865,
"step": 1350
},
{
"epoch": 0.08581532416502947,
"grad_norm": 0.337890625,
"learning_rate": 0.001,
"loss": 1.3022,
"step": 1365
},
{
"epoch": 0.08675834970530452,
"grad_norm": 0.37109375,
"learning_rate": 0.001,
"loss": 1.2641,
"step": 1380
},
{
"epoch": 0.08770137524557957,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 1.291,
"step": 1395
},
{
"epoch": 0.08864440078585462,
"grad_norm": 0.431640625,
"learning_rate": 0.001,
"loss": 1.2947,
"step": 1410
},
{
"epoch": 0.08958742632612966,
"grad_norm": 0.310546875,
"learning_rate": 0.001,
"loss": 1.2626,
"step": 1425
},
{
"epoch": 0.09053045186640471,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 1.2719,
"step": 1440
},
{
"epoch": 0.09147347740667977,
"grad_norm": 0.5,
"learning_rate": 0.001,
"loss": 1.2817,
"step": 1455
},
{
"epoch": 0.09241650294695482,
"grad_norm": 0.361328125,
"learning_rate": 0.001,
"loss": 1.2678,
"step": 1470
},
{
"epoch": 0.09335952848722986,
"grad_norm": 0.396484375,
"learning_rate": 0.001,
"loss": 1.2336,
"step": 1485
},
{
"epoch": 0.09430255402750491,
"grad_norm": 0.384765625,
"learning_rate": 0.001,
"loss": 1.2415,
"step": 1500
},
{
"epoch": 0.09524557956777996,
"grad_norm": 0.416015625,
"learning_rate": 0.001,
"loss": 1.2478,
"step": 1515
},
{
"epoch": 0.096188605108055,
"grad_norm": 0.423828125,
"learning_rate": 0.001,
"loss": 1.2475,
"step": 1530
},
{
"epoch": 0.09713163064833005,
"grad_norm": 0.400390625,
"learning_rate": 0.001,
"loss": 1.2128,
"step": 1545
},
{
"epoch": 0.09807465618860511,
"grad_norm": 0.404296875,
"learning_rate": 0.001,
"loss": 1.2292,
"step": 1560
},
{
"epoch": 0.09901768172888016,
"grad_norm": 0.294921875,
"learning_rate": 0.001,
"loss": 1.2015,
"step": 1575
},
{
"epoch": 0.09996070726915521,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 1.2088,
"step": 1590
},
{
"epoch": 0.09996070726915521,
"eval_loss": 1.5537890195846558,
"eval_runtime": 9.6819,
"eval_samples_per_second": 103.285,
"eval_steps_per_second": 1.446,
"step": 1590
},
{
"epoch": 0.10090373280943025,
"grad_norm": 0.3515625,
"learning_rate": 0.001,
"loss": 1.2156,
"step": 1605
},
{
"epoch": 0.1018467583497053,
"grad_norm": 0.443359375,
"learning_rate": 0.001,
"loss": 1.2115,
"step": 1620
},
{
"epoch": 0.10278978388998035,
"grad_norm": 0.8359375,
"learning_rate": 0.001,
"loss": 1.2202,
"step": 1635
},
{
"epoch": 0.1037328094302554,
"grad_norm": 0.314453125,
"learning_rate": 0.001,
"loss": 1.2208,
"step": 1650
},
{
"epoch": 0.10467583497053046,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 1.1911,
"step": 1665
},
{
"epoch": 0.1056188605108055,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 1.2102,
"step": 1680
},
{
"epoch": 0.10656188605108055,
"grad_norm": 0.27734375,
"learning_rate": 0.001,
"loss": 1.1984,
"step": 1695
},
{
"epoch": 0.1075049115913556,
"grad_norm": 0.90625,
"learning_rate": 0.001,
"loss": 1.2012,
"step": 1710
},
{
"epoch": 0.10844793713163065,
"grad_norm": 0.35546875,
"learning_rate": 0.001,
"loss": 1.1869,
"step": 1725
},
{
"epoch": 0.10939096267190569,
"grad_norm": 0.3828125,
"learning_rate": 0.001,
"loss": 1.1948,
"step": 1740
},
{
"epoch": 0.11033398821218075,
"grad_norm": 0.404296875,
"learning_rate": 0.001,
"loss": 1.1783,
"step": 1755
},
{
"epoch": 0.1112770137524558,
"grad_norm": 0.396484375,
"learning_rate": 0.001,
"loss": 1.1893,
"step": 1770
},
{
"epoch": 0.11222003929273085,
"grad_norm": 0.345703125,
"learning_rate": 0.001,
"loss": 1.1495,
"step": 1785
},
{
"epoch": 0.1131630648330059,
"grad_norm": 0.451171875,
"learning_rate": 0.001,
"loss": 1.175,
"step": 1800
},
{
"epoch": 0.11410609037328094,
"grad_norm": 0.61328125,
"learning_rate": 0.001,
"loss": 1.1588,
"step": 1815
},
{
"epoch": 0.11504911591355599,
"grad_norm": 0.421875,
"learning_rate": 0.001,
"loss": 1.1376,
"step": 1830
},
{
"epoch": 0.11599214145383104,
"grad_norm": 0.4453125,
"learning_rate": 0.001,
"loss": 1.1511,
"step": 1845
},
{
"epoch": 0.1169351669941061,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 1.1645,
"step": 1860
},
{
"epoch": 0.11787819253438114,
"grad_norm": 0.48046875,
"learning_rate": 0.001,
"loss": 1.1619,
"step": 1875
},
{
"epoch": 0.11882121807465619,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 1.1304,
"step": 1890
},
{
"epoch": 0.11976424361493124,
"grad_norm": 0.498046875,
"learning_rate": 0.001,
"loss": 1.1361,
"step": 1905
},
{
"epoch": 0.12070726915520628,
"grad_norm": 0.37109375,
"learning_rate": 0.001,
"loss": 1.1151,
"step": 1920
},
{
"epoch": 0.12165029469548133,
"grad_norm": 0.486328125,
"learning_rate": 0.001,
"loss": 1.1299,
"step": 1935
},
{
"epoch": 0.12259332023575638,
"grad_norm": 0.45703125,
"learning_rate": 0.001,
"loss": 1.1334,
"step": 1950
},
{
"epoch": 0.12353634577603144,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 1.112,
"step": 1965
},
{
"epoch": 0.12447937131630649,
"grad_norm": 0.7890625,
"learning_rate": 0.001,
"loss": 1.1034,
"step": 1980
},
{
"epoch": 0.12542239685658152,
"grad_norm": 0.431640625,
"learning_rate": 0.001,
"loss": 1.12,
"step": 1995
},
{
"epoch": 0.12636542239685658,
"grad_norm": 0.373046875,
"learning_rate": 0.001,
"loss": 1.0996,
"step": 2010
},
{
"epoch": 0.12730844793713164,
"grad_norm": 0.353515625,
"learning_rate": 0.001,
"loss": 1.1141,
"step": 2025
},
{
"epoch": 0.12825147347740667,
"grad_norm": 0.6640625,
"learning_rate": 0.001,
"loss": 1.1112,
"step": 2040
},
{
"epoch": 0.12919449901768174,
"grad_norm": 0.470703125,
"learning_rate": 0.001,
"loss": 1.1229,
"step": 2055
},
{
"epoch": 0.13013752455795677,
"grad_norm": 0.97265625,
"learning_rate": 0.001,
"loss": 1.074,
"step": 2070
},
{
"epoch": 0.13108055009823183,
"grad_norm": 0.3828125,
"learning_rate": 0.001,
"loss": 1.1199,
"step": 2085
},
{
"epoch": 0.13202357563850686,
"grad_norm": 0.3671875,
"learning_rate": 0.001,
"loss": 1.097,
"step": 2100
},
{
"epoch": 0.13296660117878192,
"grad_norm": 0.373046875,
"learning_rate": 0.001,
"loss": 1.0832,
"step": 2115
},
{
"epoch": 0.13390962671905698,
"grad_norm": 0.40234375,
"learning_rate": 0.001,
"loss": 1.0887,
"step": 2130
},
{
"epoch": 0.13485265225933202,
"grad_norm": 0.404296875,
"learning_rate": 0.001,
"loss": 1.066,
"step": 2145
},
{
"epoch": 0.13579567779960708,
"grad_norm": 0.4453125,
"learning_rate": 0.001,
"loss": 1.0979,
"step": 2160
},
{
"epoch": 0.1367387033398821,
"grad_norm": 0.423828125,
"learning_rate": 0.001,
"loss": 1.101,
"step": 2175
},
{
"epoch": 0.13768172888015717,
"grad_norm": 0.396484375,
"learning_rate": 0.001,
"loss": 1.0761,
"step": 2190
},
{
"epoch": 0.13862475442043223,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 1.0845,
"step": 2205
},
{
"epoch": 0.13956777996070727,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 1.0938,
"step": 2220
},
{
"epoch": 0.14051080550098233,
"grad_norm": 0.412109375,
"learning_rate": 0.001,
"loss": 1.0659,
"step": 2235
},
{
"epoch": 0.14145383104125736,
"grad_norm": 0.494140625,
"learning_rate": 0.001,
"loss": 1.0683,
"step": 2250
},
{
"epoch": 0.14239685658153242,
"grad_norm": 0.66015625,
"learning_rate": 0.001,
"loss": 1.0777,
"step": 2265
},
{
"epoch": 0.14333988212180745,
"grad_norm": 0.48828125,
"learning_rate": 0.001,
"loss": 1.0741,
"step": 2280
},
{
"epoch": 0.14428290766208252,
"grad_norm": 0.451171875,
"learning_rate": 0.001,
"loss": 1.0533,
"step": 2295
},
{
"epoch": 0.14522593320235758,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 1.0655,
"step": 2310
},
{
"epoch": 0.1461689587426326,
"grad_norm": 0.33984375,
"learning_rate": 0.001,
"loss": 1.0541,
"step": 2325
},
{
"epoch": 0.14711198428290767,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 1.0506,
"step": 2340
},
{
"epoch": 0.1480550098231827,
"grad_norm": 0.59375,
"learning_rate": 0.001,
"loss": 1.0596,
"step": 2355
},
{
"epoch": 0.14899803536345776,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 1.0586,
"step": 2370
},
{
"epoch": 0.1499410609037328,
"grad_norm": 0.40234375,
"learning_rate": 0.001,
"loss": 1.0466,
"step": 2385
},
{
"epoch": 0.15088408644400786,
"grad_norm": 0.41796875,
"learning_rate": 0.001,
"loss": 1.0485,
"step": 2400
},
{
"epoch": 0.15182711198428292,
"grad_norm": 0.482421875,
"learning_rate": 0.001,
"loss": 1.011,
"step": 2415
},
{
"epoch": 0.15277013752455795,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 1.0434,
"step": 2430
},
{
"epoch": 0.153713163064833,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 1.0353,
"step": 2445
},
{
"epoch": 0.15465618860510805,
"grad_norm": 0.45703125,
"learning_rate": 0.001,
"loss": 1.0222,
"step": 2460
},
{
"epoch": 0.1555992141453831,
"grad_norm": 0.578125,
"learning_rate": 0.001,
"loss": 1.0403,
"step": 2475
},
{
"epoch": 0.15654223968565814,
"grad_norm": 0.478515625,
"learning_rate": 0.001,
"loss": 1.0397,
"step": 2490
},
{
"epoch": 0.1574852652259332,
"grad_norm": 0.4609375,
"learning_rate": 0.001,
"loss": 1.0382,
"step": 2505
},
{
"epoch": 0.15842829076620826,
"grad_norm": 0.4609375,
"learning_rate": 0.001,
"loss": 1.0336,
"step": 2520
},
{
"epoch": 0.1593713163064833,
"grad_norm": 0.462890625,
"learning_rate": 0.001,
"loss": 1.0083,
"step": 2535
},
{
"epoch": 0.16031434184675836,
"grad_norm": 0.46875,
"learning_rate": 0.001,
"loss": 1.0236,
"step": 2550
},
{
"epoch": 0.1612573673870334,
"grad_norm": 0.45703125,
"learning_rate": 0.001,
"loss": 1.0245,
"step": 2565
},
{
"epoch": 0.16220039292730845,
"grad_norm": 0.625,
"learning_rate": 0.001,
"loss": 1.026,
"step": 2580
},
{
"epoch": 0.16314341846758348,
"grad_norm": 0.4765625,
"learning_rate": 0.001,
"loss": 1.0276,
"step": 2595
},
{
"epoch": 0.16408644400785855,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.9937,
"step": 2610
},
{
"epoch": 0.1650294695481336,
"grad_norm": 0.482421875,
"learning_rate": 0.001,
"loss": 1.0249,
"step": 2625
},
{
"epoch": 0.16597249508840864,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 1.0096,
"step": 2640
},
{
"epoch": 0.1669155206286837,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 1.0195,
"step": 2655
},
{
"epoch": 0.16785854616895873,
"grad_norm": 0.484375,
"learning_rate": 0.001,
"loss": 1.018,
"step": 2670
},
{
"epoch": 0.1688015717092338,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 1.0289,
"step": 2685
},
{
"epoch": 0.16974459724950883,
"grad_norm": 0.423828125,
"learning_rate": 0.001,
"loss": 0.9931,
"step": 2700
},
{
"epoch": 0.1706876227897839,
"grad_norm": 0.42578125,
"learning_rate": 0.001,
"loss": 1.0101,
"step": 2715
},
{
"epoch": 0.17163064833005895,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 1.0159,
"step": 2730
},
{
"epoch": 0.17257367387033398,
"grad_norm": 0.486328125,
"learning_rate": 0.001,
"loss": 1.0094,
"step": 2745
},
{
"epoch": 0.17351669941060904,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 1.0081,
"step": 2760
},
{
"epoch": 0.17445972495088408,
"grad_norm": 0.6640625,
"learning_rate": 0.001,
"loss": 0.9958,
"step": 2775
},
{
"epoch": 0.17540275049115914,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 0.9909,
"step": 2790
},
{
"epoch": 0.1763457760314342,
"grad_norm": 0.482421875,
"learning_rate": 0.001,
"loss": 0.9854,
"step": 2805
},
{
"epoch": 0.17728880157170923,
"grad_norm": 0.46875,
"learning_rate": 0.001,
"loss": 0.9858,
"step": 2820
},
{
"epoch": 0.1782318271119843,
"grad_norm": 0.49609375,
"learning_rate": 0.001,
"loss": 0.9825,
"step": 2835
},
{
"epoch": 0.17917485265225933,
"grad_norm": 0.458984375,
"learning_rate": 0.001,
"loss": 1.0153,
"step": 2850
},
{
"epoch": 0.1801178781925344,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.9984,
"step": 2865
},
{
"epoch": 0.18106090373280942,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.9832,
"step": 2880
},
{
"epoch": 0.18200392927308448,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.9843,
"step": 2895
},
{
"epoch": 0.18294695481335954,
"grad_norm": 0.41015625,
"learning_rate": 0.001,
"loss": 0.9774,
"step": 2910
},
{
"epoch": 0.18388998035363457,
"grad_norm": 0.484375,
"learning_rate": 0.001,
"loss": 0.9824,
"step": 2925
},
{
"epoch": 0.18483300589390964,
"grad_norm": 0.41796875,
"learning_rate": 0.001,
"loss": 0.9884,
"step": 2940
},
{
"epoch": 0.18577603143418467,
"grad_norm": 0.447265625,
"learning_rate": 0.001,
"loss": 0.9684,
"step": 2955
},
{
"epoch": 0.18671905697445973,
"grad_norm": 0.6640625,
"learning_rate": 0.001,
"loss": 0.9746,
"step": 2970
},
{
"epoch": 0.18766208251473476,
"grad_norm": 0.47265625,
"learning_rate": 0.001,
"loss": 0.9831,
"step": 2985
},
{
"epoch": 0.18860510805500982,
"grad_norm": 0.671875,
"learning_rate": 0.001,
"loss": 0.9868,
"step": 3000
},
{
"epoch": 0.18954813359528488,
"grad_norm": 0.482421875,
"learning_rate": 0.001,
"loss": 0.9687,
"step": 3015
},
{
"epoch": 0.19049115913555992,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 0.9759,
"step": 3030
},
{
"epoch": 0.19143418467583498,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.9755,
"step": 3045
},
{
"epoch": 0.19237721021611,
"grad_norm": 0.62109375,
"learning_rate": 0.001,
"loss": 0.9784,
"step": 3060
},
{
"epoch": 0.19332023575638507,
"grad_norm": 0.60546875,
"learning_rate": 0.001,
"loss": 0.9691,
"step": 3075
},
{
"epoch": 0.1942632612966601,
"grad_norm": 0.466796875,
"learning_rate": 0.001,
"loss": 0.9851,
"step": 3090
},
{
"epoch": 0.19520628683693517,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.9695,
"step": 3105
},
{
"epoch": 0.19614931237721023,
"grad_norm": 0.5546875,
"learning_rate": 0.001,
"loss": 0.993,
"step": 3120
},
{
"epoch": 0.19709233791748526,
"grad_norm": 0.49609375,
"learning_rate": 0.001,
"loss": 0.9625,
"step": 3135
},
{
"epoch": 0.19803536345776032,
"grad_norm": 0.3828125,
"learning_rate": 0.001,
"loss": 0.9655,
"step": 3150
},
{
"epoch": 0.19897838899803535,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.9606,
"step": 3165
},
{
"epoch": 0.19992141453831042,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.9608,
"step": 3180
},
{
"epoch": 0.19992141453831042,
"eval_loss": 1.169226050376892,
"eval_runtime": 9.7503,
"eval_samples_per_second": 102.561,
"eval_steps_per_second": 1.436,
"step": 3180
},
{
"epoch": 0.20086444007858545,
"grad_norm": 0.46484375,
"learning_rate": 0.001,
"loss": 0.9741,
"step": 3195
},
{
"epoch": 0.2018074656188605,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.9608,
"step": 3210
},
{
"epoch": 0.20275049115913557,
"grad_norm": 0.44921875,
"learning_rate": 0.001,
"loss": 0.9464,
"step": 3225
},
{
"epoch": 0.2036935166994106,
"grad_norm": 0.671875,
"learning_rate": 0.001,
"loss": 0.9683,
"step": 3240
},
{
"epoch": 0.20463654223968566,
"grad_norm": 0.43359375,
"learning_rate": 0.001,
"loss": 0.9308,
"step": 3255
},
{
"epoch": 0.2055795677799607,
"grad_norm": 0.380859375,
"learning_rate": 0.001,
"loss": 0.9541,
"step": 3270
},
{
"epoch": 0.20652259332023576,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.9452,
"step": 3285
},
{
"epoch": 0.2074656188605108,
"grad_norm": 0.62890625,
"learning_rate": 0.001,
"loss": 0.9673,
"step": 3300
},
{
"epoch": 0.20840864440078585,
"grad_norm": 0.4921875,
"learning_rate": 0.001,
"loss": 0.9508,
"step": 3315
},
{
"epoch": 0.2093516699410609,
"grad_norm": 0.490234375,
"learning_rate": 0.001,
"loss": 0.955,
"step": 3330
},
{
"epoch": 0.21029469548133595,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.9499,
"step": 3345
},
{
"epoch": 0.211237721021611,
"grad_norm": 0.64453125,
"learning_rate": 0.001,
"loss": 0.9441,
"step": 3360
},
{
"epoch": 0.21218074656188604,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.9476,
"step": 3375
},
{
"epoch": 0.2131237721021611,
"grad_norm": 0.431640625,
"learning_rate": 0.001,
"loss": 0.9506,
"step": 3390
},
{
"epoch": 0.21406679764243616,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.9546,
"step": 3405
},
{
"epoch": 0.2150098231827112,
"grad_norm": 0.4921875,
"learning_rate": 0.001,
"loss": 0.9488,
"step": 3420
},
{
"epoch": 0.21595284872298626,
"grad_norm": 0.625,
"learning_rate": 0.001,
"loss": 0.9473,
"step": 3435
},
{
"epoch": 0.2168958742632613,
"grad_norm": 0.435546875,
"learning_rate": 0.001,
"loss": 0.9491,
"step": 3450
},
{
"epoch": 0.21783889980353635,
"grad_norm": 0.41796875,
"learning_rate": 0.001,
"loss": 0.9304,
"step": 3465
},
{
"epoch": 0.21878192534381138,
"grad_norm": 0.80078125,
"learning_rate": 0.001,
"loss": 0.9482,
"step": 3480
},
{
"epoch": 0.21972495088408645,
"grad_norm": 0.466796875,
"learning_rate": 0.001,
"loss": 0.9418,
"step": 3495
},
{
"epoch": 0.2206679764243615,
"grad_norm": 0.46484375,
"learning_rate": 0.001,
"loss": 0.9226,
"step": 3510
},
{
"epoch": 0.22161100196463654,
"grad_norm": 0.427734375,
"learning_rate": 0.001,
"loss": 0.9427,
"step": 3525
},
{
"epoch": 0.2225540275049116,
"grad_norm": 0.76953125,
"learning_rate": 0.001,
"loss": 0.9261,
"step": 3540
},
{
"epoch": 0.22349705304518663,
"grad_norm": 0.69140625,
"learning_rate": 0.001,
"loss": 0.9418,
"step": 3555
},
{
"epoch": 0.2244400785854617,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.9382,
"step": 3570
},
{
"epoch": 0.22538310412573673,
"grad_norm": 0.412109375,
"learning_rate": 0.001,
"loss": 0.9353,
"step": 3585
},
{
"epoch": 0.2263261296660118,
"grad_norm": 0.447265625,
"learning_rate": 0.001,
"loss": 0.9138,
"step": 3600
},
{
"epoch": 0.22726915520628685,
"grad_norm": 0.59765625,
"learning_rate": 0.001,
"loss": 0.9033,
"step": 3615
},
{
"epoch": 0.22821218074656188,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.9337,
"step": 3630
},
{
"epoch": 0.22915520628683694,
"grad_norm": 0.4765625,
"learning_rate": 0.001,
"loss": 0.9188,
"step": 3645
},
{
"epoch": 0.23009823182711198,
"grad_norm": 0.5,
"learning_rate": 0.001,
"loss": 0.9407,
"step": 3660
},
{
"epoch": 0.23104125736738704,
"grad_norm": 0.40625,
"learning_rate": 0.001,
"loss": 0.9068,
"step": 3675
},
{
"epoch": 0.23198428290766207,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.9079,
"step": 3690
},
{
"epoch": 0.23292730844793713,
"grad_norm": 0.470703125,
"learning_rate": 0.001,
"loss": 0.9095,
"step": 3705
},
{
"epoch": 0.2338703339882122,
"grad_norm": 0.4765625,
"learning_rate": 0.001,
"loss": 0.9148,
"step": 3720
},
{
"epoch": 0.23481335952848723,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.9044,
"step": 3735
},
{
"epoch": 0.2357563850687623,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.9401,
"step": 3750
},
{
"epoch": 0.23669941060903732,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.9228,
"step": 3765
},
{
"epoch": 0.23764243614931238,
"grad_norm": 0.447265625,
"learning_rate": 0.001,
"loss": 0.9071,
"step": 3780
},
{
"epoch": 0.2385854616895874,
"grad_norm": 0.734375,
"learning_rate": 0.001,
"loss": 0.92,
"step": 3795
},
{
"epoch": 0.23952848722986247,
"grad_norm": 0.5625,
"learning_rate": 0.001,
"loss": 0.9323,
"step": 3810
},
{
"epoch": 0.24047151277013754,
"grad_norm": 0.3828125,
"learning_rate": 0.001,
"loss": 0.9013,
"step": 3825
},
{
"epoch": 0.24141453831041257,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.9045,
"step": 3840
},
{
"epoch": 0.24235756385068763,
"grad_norm": 0.43359375,
"learning_rate": 0.001,
"loss": 0.9049,
"step": 3855
},
{
"epoch": 0.24330058939096266,
"grad_norm": 0.48828125,
"learning_rate": 0.001,
"loss": 0.8902,
"step": 3870
},
{
"epoch": 0.24424361493123772,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.911,
"step": 3885
},
{
"epoch": 0.24518664047151276,
"grad_norm": 0.486328125,
"learning_rate": 0.001,
"loss": 0.9092,
"step": 3900
},
{
"epoch": 0.24612966601178782,
"grad_norm": 0.451171875,
"learning_rate": 0.001,
"loss": 0.894,
"step": 3915
},
{
"epoch": 0.24707269155206288,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.9096,
"step": 3930
},
{
"epoch": 0.2480157170923379,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 0.9147,
"step": 3945
},
{
"epoch": 0.24895874263261297,
"grad_norm": 0.8359375,
"learning_rate": 0.001,
"loss": 0.9088,
"step": 3960
},
{
"epoch": 0.249901768172888,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 0.9116,
"step": 3975
},
{
"epoch": 0.25084479371316304,
"grad_norm": 0.7421875,
"learning_rate": 0.001,
"loss": 0.901,
"step": 3990
},
{
"epoch": 0.2517878192534381,
"grad_norm": 0.58984375,
"learning_rate": 0.001,
"loss": 0.9013,
"step": 4005
},
{
"epoch": 0.25273084479371316,
"grad_norm": 0.41015625,
"learning_rate": 0.001,
"loss": 0.903,
"step": 4020
},
{
"epoch": 0.2536738703339882,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.8916,
"step": 4035
},
{
"epoch": 0.2546168958742633,
"grad_norm": 0.69140625,
"learning_rate": 0.001,
"loss": 0.897,
"step": 4050
},
{
"epoch": 0.2555599214145383,
"grad_norm": 0.462890625,
"learning_rate": 0.001,
"loss": 0.9015,
"step": 4065
},
{
"epoch": 0.25650294695481335,
"grad_norm": 0.453125,
"learning_rate": 0.001,
"loss": 0.897,
"step": 4080
},
{
"epoch": 0.2574459724950884,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.8936,
"step": 4095
},
{
"epoch": 0.25838899803536347,
"grad_norm": 0.4609375,
"learning_rate": 0.001,
"loss": 0.9048,
"step": 4110
},
{
"epoch": 0.2593320235756385,
"grad_norm": 0.4921875,
"learning_rate": 0.001,
"loss": 0.8973,
"step": 4125
},
{
"epoch": 0.26027504911591354,
"grad_norm": 0.423828125,
"learning_rate": 0.001,
"loss": 0.9053,
"step": 4140
},
{
"epoch": 0.2612180746561886,
"grad_norm": 0.75,
"learning_rate": 0.001,
"loss": 0.9121,
"step": 4155
},
{
"epoch": 0.26216110019646366,
"grad_norm": 0.3828125,
"learning_rate": 0.001,
"loss": 0.89,
"step": 4170
},
{
"epoch": 0.2631041257367387,
"grad_norm": 0.57421875,
"learning_rate": 0.001,
"loss": 0.9025,
"step": 4185
},
{
"epoch": 0.2640471512770137,
"grad_norm": 0.400390625,
"learning_rate": 0.001,
"loss": 0.899,
"step": 4200
},
{
"epoch": 0.2649901768172888,
"grad_norm": 0.484375,
"learning_rate": 0.001,
"loss": 0.8793,
"step": 4215
},
{
"epoch": 0.26593320235756385,
"grad_norm": 0.59375,
"learning_rate": 0.001,
"loss": 0.8964,
"step": 4230
},
{
"epoch": 0.2668762278978389,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.896,
"step": 4245
},
{
"epoch": 0.26781925343811397,
"grad_norm": 0.61328125,
"learning_rate": 0.001,
"loss": 0.886,
"step": 4260
},
{
"epoch": 0.268762278978389,
"grad_norm": 0.6796875,
"learning_rate": 0.001,
"loss": 0.8861,
"step": 4275
},
{
"epoch": 0.26970530451866404,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.8864,
"step": 4290
},
{
"epoch": 0.2706483300589391,
"grad_norm": 0.46484375,
"learning_rate": 0.001,
"loss": 0.8834,
"step": 4305
},
{
"epoch": 0.27159135559921416,
"grad_norm": 0.474609375,
"learning_rate": 0.001,
"loss": 0.8859,
"step": 4320
},
{
"epoch": 0.2725343811394892,
"grad_norm": 0.498046875,
"learning_rate": 0.001,
"loss": 0.8953,
"step": 4335
},
{
"epoch": 0.2734774066797642,
"grad_norm": 0.68359375,
"learning_rate": 0.001,
"loss": 0.8928,
"step": 4350
},
{
"epoch": 0.2744204322200393,
"grad_norm": 0.734375,
"learning_rate": 0.001,
"loss": 0.8821,
"step": 4365
},
{
"epoch": 0.27536345776031435,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.8872,
"step": 4380
},
{
"epoch": 0.2763064833005894,
"grad_norm": 0.490234375,
"learning_rate": 0.001,
"loss": 0.8753,
"step": 4395
},
{
"epoch": 0.27724950884086447,
"grad_norm": 0.80859375,
"learning_rate": 0.001,
"loss": 0.9047,
"step": 4410
},
{
"epoch": 0.2781925343811395,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.8876,
"step": 4425
},
{
"epoch": 0.27913555992141453,
"grad_norm": 0.46875,
"learning_rate": 0.001,
"loss": 0.864,
"step": 4440
},
{
"epoch": 0.28007858546168957,
"grad_norm": 0.443359375,
"learning_rate": 0.001,
"loss": 0.8863,
"step": 4455
},
{
"epoch": 0.28102161100196466,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.9028,
"step": 4470
},
{
"epoch": 0.2819646365422397,
"grad_norm": 0.5,
"learning_rate": 0.001,
"loss": 0.8684,
"step": 4485
},
{
"epoch": 0.2829076620825147,
"grad_norm": 0.5625,
"learning_rate": 0.001,
"loss": 0.8808,
"step": 4500
},
{
"epoch": 0.2838506876227898,
"grad_norm": 0.43359375,
"learning_rate": 0.001,
"loss": 0.8736,
"step": 4515
},
{
"epoch": 0.28479371316306484,
"grad_norm": 0.439453125,
"learning_rate": 0.001,
"loss": 0.8729,
"step": 4530
},
{
"epoch": 0.2857367387033399,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.8807,
"step": 4545
},
{
"epoch": 0.2866797642436149,
"grad_norm": 0.68359375,
"learning_rate": 0.001,
"loss": 0.8716,
"step": 4560
},
{
"epoch": 0.28762278978389,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.8754,
"step": 4575
},
{
"epoch": 0.28856581532416503,
"grad_norm": 0.46875,
"learning_rate": 0.001,
"loss": 0.866,
"step": 4590
},
{
"epoch": 0.28950884086444006,
"grad_norm": 0.50390625,
"learning_rate": 0.001,
"loss": 0.8661,
"step": 4605
},
{
"epoch": 0.29045186640471515,
"grad_norm": 0.65625,
"learning_rate": 0.001,
"loss": 0.8797,
"step": 4620
},
{
"epoch": 0.2913948919449902,
"grad_norm": 0.47265625,
"learning_rate": 0.001,
"loss": 0.8523,
"step": 4635
},
{
"epoch": 0.2923379174852652,
"grad_norm": 0.57421875,
"learning_rate": 0.001,
"loss": 0.8774,
"step": 4650
},
{
"epoch": 0.29328094302554025,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.8785,
"step": 4665
},
{
"epoch": 0.29422396856581534,
"grad_norm": 0.408203125,
"learning_rate": 0.001,
"loss": 0.8648,
"step": 4680
},
{
"epoch": 0.2951669941060904,
"grad_norm": 0.71484375,
"learning_rate": 0.001,
"loss": 0.8676,
"step": 4695
},
{
"epoch": 0.2961100196463654,
"grad_norm": 0.5,
"learning_rate": 0.001,
"loss": 0.8557,
"step": 4710
},
{
"epoch": 0.2970530451866405,
"grad_norm": 0.455078125,
"learning_rate": 0.001,
"loss": 0.8694,
"step": 4725
},
{
"epoch": 0.29799607072691553,
"grad_norm": 0.50390625,
"learning_rate": 0.001,
"loss": 0.8459,
"step": 4740
},
{
"epoch": 0.29893909626719056,
"grad_norm": 0.44921875,
"learning_rate": 0.001,
"loss": 0.8551,
"step": 4755
},
{
"epoch": 0.2998821218074656,
"grad_norm": 0.578125,
"learning_rate": 0.001,
"loss": 0.8717,
"step": 4770
},
{
"epoch": 0.2998821218074656,
"eval_loss": 1.035895824432373,
"eval_runtime": 9.7687,
"eval_samples_per_second": 102.368,
"eval_steps_per_second": 1.433,
"step": 4770
},
{
"epoch": 0.3008251473477407,
"grad_norm": 0.490234375,
"learning_rate": 0.001,
"loss": 0.8668,
"step": 4785
},
{
"epoch": 0.3017681728880157,
"grad_norm": 0.734375,
"learning_rate": 0.001,
"loss": 0.8674,
"step": 4800
},
{
"epoch": 0.30271119842829075,
"grad_norm": 0.5625,
"learning_rate": 0.001,
"loss": 0.8886,
"step": 4815
},
{
"epoch": 0.30365422396856584,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.854,
"step": 4830
},
{
"epoch": 0.3045972495088409,
"grad_norm": 0.5,
"learning_rate": 0.001,
"loss": 0.8513,
"step": 4845
},
{
"epoch": 0.3055402750491159,
"grad_norm": 0.59765625,
"learning_rate": 0.001,
"loss": 0.8574,
"step": 4860
},
{
"epoch": 0.30648330058939094,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.8437,
"step": 4875
},
{
"epoch": 0.307426326129666,
"grad_norm": 0.62109375,
"learning_rate": 0.001,
"loss": 0.8604,
"step": 4890
},
{
"epoch": 0.30836935166994106,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 0.8544,
"step": 4905
},
{
"epoch": 0.3093123772102161,
"grad_norm": 0.59375,
"learning_rate": 0.001,
"loss": 0.8607,
"step": 4920
},
{
"epoch": 0.3102554027504912,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.8454,
"step": 4935
},
{
"epoch": 0.3111984282907662,
"grad_norm": 0.458984375,
"learning_rate": 0.001,
"loss": 0.8575,
"step": 4950
},
{
"epoch": 0.31214145383104125,
"grad_norm": 0.73828125,
"learning_rate": 0.001,
"loss": 0.8401,
"step": 4965
},
{
"epoch": 0.3130844793713163,
"grad_norm": 0.470703125,
"learning_rate": 0.001,
"loss": 0.8592,
"step": 4980
},
{
"epoch": 0.31402750491159137,
"grad_norm": 0.46875,
"learning_rate": 0.001,
"loss": 0.8376,
"step": 4995
},
{
"epoch": 0.3149705304518664,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.853,
"step": 5010
},
{
"epoch": 0.31591355599214144,
"grad_norm": 0.6171875,
"learning_rate": 0.001,
"loss": 0.8659,
"step": 5025
},
{
"epoch": 0.3168565815324165,
"grad_norm": 0.62890625,
"learning_rate": 0.001,
"loss": 0.8733,
"step": 5040
},
{
"epoch": 0.31779960707269156,
"grad_norm": 0.462890625,
"learning_rate": 0.001,
"loss": 0.8541,
"step": 5055
},
{
"epoch": 0.3187426326129666,
"grad_norm": 0.66015625,
"learning_rate": 0.001,
"loss": 0.8474,
"step": 5070
},
{
"epoch": 0.3196856581532416,
"grad_norm": 0.640625,
"learning_rate": 0.001,
"loss": 0.8421,
"step": 5085
},
{
"epoch": 0.3206286836935167,
"grad_norm": 0.46875,
"learning_rate": 0.001,
"loss": 0.8501,
"step": 5100
},
{
"epoch": 0.32157170923379175,
"grad_norm": 0.44140625,
"learning_rate": 0.001,
"loss": 0.8596,
"step": 5115
},
{
"epoch": 0.3225147347740668,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.8421,
"step": 5130
},
{
"epoch": 0.32345776031434187,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.8732,
"step": 5145
},
{
"epoch": 0.3244007858546169,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.8549,
"step": 5160
},
{
"epoch": 0.32534381139489194,
"grad_norm": 0.63671875,
"learning_rate": 0.001,
"loss": 0.8468,
"step": 5175
},
{
"epoch": 0.32628683693516697,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.8419,
"step": 5190
},
{
"epoch": 0.32722986247544206,
"grad_norm": 0.431640625,
"learning_rate": 0.001,
"loss": 0.8531,
"step": 5205
},
{
"epoch": 0.3281728880157171,
"grad_norm": 0.48046875,
"learning_rate": 0.001,
"loss": 0.848,
"step": 5220
},
{
"epoch": 0.3291159135559921,
"grad_norm": 0.419921875,
"learning_rate": 0.001,
"loss": 0.8367,
"step": 5235
},
{
"epoch": 0.3300589390962672,
"grad_norm": 0.435546875,
"learning_rate": 0.001,
"loss": 0.8405,
"step": 5250
},
{
"epoch": 0.33100196463654225,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.8567,
"step": 5265
},
{
"epoch": 0.3319449901768173,
"grad_norm": 0.66796875,
"learning_rate": 0.001,
"loss": 0.8572,
"step": 5280
},
{
"epoch": 0.3328880157170923,
"grad_norm": 0.78515625,
"learning_rate": 0.001,
"loss": 0.8505,
"step": 5295
},
{
"epoch": 0.3338310412573674,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.8398,
"step": 5310
},
{
"epoch": 0.33477406679764243,
"grad_norm": 0.59375,
"learning_rate": 0.001,
"loss": 0.8475,
"step": 5325
},
{
"epoch": 0.33571709233791747,
"grad_norm": 0.6796875,
"learning_rate": 0.001,
"loss": 0.8267,
"step": 5340
},
{
"epoch": 0.33666011787819256,
"grad_norm": 0.6953125,
"learning_rate": 0.001,
"loss": 0.8442,
"step": 5355
},
{
"epoch": 0.3376031434184676,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.8605,
"step": 5370
},
{
"epoch": 0.3385461689587426,
"grad_norm": 0.48828125,
"learning_rate": 0.001,
"loss": 0.8458,
"step": 5385
},
{
"epoch": 0.33948919449901765,
"grad_norm": 0.482421875,
"learning_rate": 0.001,
"loss": 0.8474,
"step": 5400
},
{
"epoch": 0.34043222003929274,
"grad_norm": 0.486328125,
"learning_rate": 0.001,
"loss": 0.8507,
"step": 5415
},
{
"epoch": 0.3413752455795678,
"grad_norm": 0.498046875,
"learning_rate": 0.001,
"loss": 0.8449,
"step": 5430
},
{
"epoch": 0.3423182711198428,
"grad_norm": 0.62109375,
"learning_rate": 0.001,
"loss": 0.8456,
"step": 5445
},
{
"epoch": 0.3432612966601179,
"grad_norm": 0.474609375,
"learning_rate": 0.001,
"loss": 0.834,
"step": 5460
},
{
"epoch": 0.34420432220039293,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.8382,
"step": 5475
},
{
"epoch": 0.34514734774066796,
"grad_norm": 0.447265625,
"learning_rate": 0.001,
"loss": 0.8162,
"step": 5490
},
{
"epoch": 0.346090373280943,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.8331,
"step": 5505
},
{
"epoch": 0.3470333988212181,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.8461,
"step": 5520
},
{
"epoch": 0.3479764243614931,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.8277,
"step": 5535
},
{
"epoch": 0.34891944990176815,
"grad_norm": 0.490234375,
"learning_rate": 0.001,
"loss": 0.8261,
"step": 5550
},
{
"epoch": 0.34986247544204324,
"grad_norm": 0.5,
"learning_rate": 0.001,
"loss": 0.8368,
"step": 5565
},
{
"epoch": 0.3508055009823183,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.829,
"step": 5580
},
{
"epoch": 0.3517485265225933,
"grad_norm": 0.4921875,
"learning_rate": 0.001,
"loss": 0.8356,
"step": 5595
},
{
"epoch": 0.3526915520628684,
"grad_norm": 0.6484375,
"learning_rate": 0.001,
"loss": 0.8404,
"step": 5610
},
{
"epoch": 0.35363457760314343,
"grad_norm": 0.66796875,
"learning_rate": 0.001,
"loss": 0.8221,
"step": 5625
},
{
"epoch": 0.35457760314341846,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.8336,
"step": 5640
},
{
"epoch": 0.3555206286836935,
"grad_norm": 0.474609375,
"learning_rate": 0.001,
"loss": 0.8118,
"step": 5655
},
{
"epoch": 0.3564636542239686,
"grad_norm": 0.4609375,
"learning_rate": 0.001,
"loss": 0.8288,
"step": 5670
},
{
"epoch": 0.3574066797642436,
"grad_norm": 0.486328125,
"learning_rate": 0.001,
"loss": 0.8376,
"step": 5685
},
{
"epoch": 0.35834970530451865,
"grad_norm": 0.58984375,
"learning_rate": 0.001,
"loss": 0.8426,
"step": 5700
},
{
"epoch": 0.35929273084479374,
"grad_norm": 0.56640625,
"learning_rate": 0.001,
"loss": 0.8437,
"step": 5715
},
{
"epoch": 0.3602357563850688,
"grad_norm": 0.4921875,
"learning_rate": 0.001,
"loss": 0.8469,
"step": 5730
},
{
"epoch": 0.3611787819253438,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.8274,
"step": 5745
},
{
"epoch": 0.36212180746561884,
"grad_norm": 0.609375,
"learning_rate": 0.001,
"loss": 0.8306,
"step": 5760
},
{
"epoch": 0.3630648330058939,
"grad_norm": 0.6328125,
"learning_rate": 0.001,
"loss": 0.8315,
"step": 5775
},
{
"epoch": 0.36400785854616896,
"grad_norm": 0.6640625,
"learning_rate": 0.001,
"loss": 0.8379,
"step": 5790
},
{
"epoch": 0.364950884086444,
"grad_norm": 0.640625,
"learning_rate": 0.001,
"loss": 0.8342,
"step": 5805
},
{
"epoch": 0.3658939096267191,
"grad_norm": 0.625,
"learning_rate": 0.001,
"loss": 0.8374,
"step": 5820
},
{
"epoch": 0.3668369351669941,
"grad_norm": 0.640625,
"learning_rate": 0.001,
"loss": 0.8103,
"step": 5835
},
{
"epoch": 0.36777996070726915,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.8053,
"step": 5850
},
{
"epoch": 0.3687229862475442,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.8248,
"step": 5865
},
{
"epoch": 0.36966601178781927,
"grad_norm": 0.484375,
"learning_rate": 0.001,
"loss": 0.8118,
"step": 5880
},
{
"epoch": 0.3706090373280943,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.8289,
"step": 5895
},
{
"epoch": 0.37155206286836934,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.8295,
"step": 5910
},
{
"epoch": 0.3724950884086444,
"grad_norm": 0.6328125,
"learning_rate": 0.001,
"loss": 0.8158,
"step": 5925
},
{
"epoch": 0.37343811394891946,
"grad_norm": 0.478515625,
"learning_rate": 0.001,
"loss": 0.8235,
"step": 5940
},
{
"epoch": 0.3743811394891945,
"grad_norm": 0.578125,
"learning_rate": 0.001,
"loss": 0.8148,
"step": 5955
},
{
"epoch": 0.3753241650294695,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.8161,
"step": 5970
},
{
"epoch": 0.3762671905697446,
"grad_norm": 0.474609375,
"learning_rate": 0.001,
"loss": 0.812,
"step": 5985
},
{
"epoch": 0.37721021611001965,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 0.8154,
"step": 6000
},
{
"epoch": 0.3781532416502947,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.8248,
"step": 6015
},
{
"epoch": 0.37909626719056977,
"grad_norm": 0.7265625,
"learning_rate": 0.001,
"loss": 0.8104,
"step": 6030
},
{
"epoch": 0.3800392927308448,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.8228,
"step": 6045
},
{
"epoch": 0.38098231827111984,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.8392,
"step": 6060
},
{
"epoch": 0.38192534381139487,
"grad_norm": 0.50390625,
"learning_rate": 0.001,
"loss": 0.8352,
"step": 6075
},
{
"epoch": 0.38286836935166996,
"grad_norm": 0.578125,
"learning_rate": 0.001,
"loss": 0.8271,
"step": 6090
},
{
"epoch": 0.383811394891945,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.8122,
"step": 6105
},
{
"epoch": 0.38475442043222,
"grad_norm": 0.5,
"learning_rate": 0.001,
"loss": 0.8221,
"step": 6120
},
{
"epoch": 0.3856974459724951,
"grad_norm": 0.4921875,
"learning_rate": 0.001,
"loss": 0.8354,
"step": 6135
},
{
"epoch": 0.38664047151277015,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.8277,
"step": 6150
},
{
"epoch": 0.3875834970530452,
"grad_norm": 0.58984375,
"learning_rate": 0.001,
"loss": 0.8263,
"step": 6165
},
{
"epoch": 0.3885265225933202,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.8122,
"step": 6180
},
{
"epoch": 0.3894695481335953,
"grad_norm": 0.70703125,
"learning_rate": 0.001,
"loss": 0.8296,
"step": 6195
},
{
"epoch": 0.39041257367387033,
"grad_norm": 0.65234375,
"learning_rate": 0.001,
"loss": 0.8171,
"step": 6210
},
{
"epoch": 0.39135559921414537,
"grad_norm": 0.466796875,
"learning_rate": 0.001,
"loss": 0.8127,
"step": 6225
},
{
"epoch": 0.39229862475442046,
"grad_norm": 0.482421875,
"learning_rate": 0.001,
"loss": 0.806,
"step": 6240
},
{
"epoch": 0.3932416502946955,
"grad_norm": 0.462890625,
"learning_rate": 0.001,
"loss": 0.8157,
"step": 6255
},
{
"epoch": 0.3941846758349705,
"grad_norm": 0.578125,
"learning_rate": 0.001,
"loss": 0.826,
"step": 6270
},
{
"epoch": 0.39512770137524555,
"grad_norm": 0.41796875,
"learning_rate": 0.001,
"loss": 0.8208,
"step": 6285
},
{
"epoch": 0.39607072691552064,
"grad_norm": 0.62109375,
"learning_rate": 0.001,
"loss": 0.8041,
"step": 6300
},
{
"epoch": 0.3970137524557957,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.8254,
"step": 6315
},
{
"epoch": 0.3979567779960707,
"grad_norm": 0.48046875,
"learning_rate": 0.001,
"loss": 0.8332,
"step": 6330
},
{
"epoch": 0.3988998035363458,
"grad_norm": 0.48828125,
"learning_rate": 0.001,
"loss": 0.8143,
"step": 6345
},
{
"epoch": 0.39984282907662083,
"grad_norm": 0.486328125,
"learning_rate": 0.001,
"loss": 0.8087,
"step": 6360
},
{
"epoch": 0.39984282907662083,
"eval_loss": 0.9629083871841431,
"eval_runtime": 9.6716,
"eval_samples_per_second": 103.395,
"eval_steps_per_second": 1.448,
"step": 6360
},
{
"epoch": 0.40078585461689586,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.8169,
"step": 6375
},
{
"epoch": 0.4017288801571709,
"grad_norm": 0.486328125,
"learning_rate": 0.001,
"loss": 0.8229,
"step": 6390
},
{
"epoch": 0.402671905697446,
"grad_norm": 0.6015625,
"learning_rate": 0.001,
"loss": 0.8108,
"step": 6405
},
{
"epoch": 0.403614931237721,
"grad_norm": 0.5546875,
"learning_rate": 0.001,
"loss": 0.814,
"step": 6420
},
{
"epoch": 0.40455795677799605,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.8077,
"step": 6435
},
{
"epoch": 0.40550098231827114,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.8103,
"step": 6450
},
{
"epoch": 0.4064440078585462,
"grad_norm": 0.46484375,
"learning_rate": 0.001,
"loss": 0.7904,
"step": 6465
},
{
"epoch": 0.4073870333988212,
"grad_norm": 0.5625,
"learning_rate": 0.001,
"loss": 0.8006,
"step": 6480
},
{
"epoch": 0.40833005893909624,
"grad_norm": 0.5,
"learning_rate": 0.001,
"loss": 0.8112,
"step": 6495
},
{
"epoch": 0.40927308447937133,
"grad_norm": 0.470703125,
"learning_rate": 0.001,
"loss": 0.7984,
"step": 6510
},
{
"epoch": 0.41021611001964636,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.7883,
"step": 6525
},
{
"epoch": 0.4111591355599214,
"grad_norm": 1.0625,
"learning_rate": 0.001,
"loss": 0.8196,
"step": 6540
},
{
"epoch": 0.4121021611001965,
"grad_norm": 0.64453125,
"learning_rate": 0.001,
"loss": 0.8274,
"step": 6555
},
{
"epoch": 0.4130451866404715,
"grad_norm": 0.419921875,
"learning_rate": 0.001,
"loss": 0.7942,
"step": 6570
},
{
"epoch": 0.41398821218074655,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.7965,
"step": 6585
},
{
"epoch": 0.4149312377210216,
"grad_norm": 0.435546875,
"learning_rate": 0.001,
"loss": 0.7944,
"step": 6600
},
{
"epoch": 0.4158742632612967,
"grad_norm": 0.62109375,
"learning_rate": 0.001,
"loss": 0.8055,
"step": 6615
},
{
"epoch": 0.4168172888015717,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.8083,
"step": 6630
},
{
"epoch": 0.41776031434184674,
"grad_norm": 0.56640625,
"learning_rate": 0.001,
"loss": 0.8151,
"step": 6645
},
{
"epoch": 0.4187033398821218,
"grad_norm": 0.6328125,
"learning_rate": 0.001,
"loss": 0.8093,
"step": 6660
},
{
"epoch": 0.41964636542239686,
"grad_norm": 0.625,
"learning_rate": 0.001,
"loss": 0.807,
"step": 6675
},
{
"epoch": 0.4205893909626719,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.7884,
"step": 6690
},
{
"epoch": 0.4215324165029469,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 0.7958,
"step": 6705
},
{
"epoch": 0.422475442043222,
"grad_norm": 0.73046875,
"learning_rate": 0.001,
"loss": 0.8029,
"step": 6720
},
{
"epoch": 0.42341846758349705,
"grad_norm": 0.455078125,
"learning_rate": 0.001,
"loss": 0.804,
"step": 6735
},
{
"epoch": 0.4243614931237721,
"grad_norm": 0.6953125,
"learning_rate": 0.001,
"loss": 0.8235,
"step": 6750
},
{
"epoch": 0.42530451866404717,
"grad_norm": 0.6015625,
"learning_rate": 0.001,
"loss": 0.8105,
"step": 6765
},
{
"epoch": 0.4262475442043222,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.8028,
"step": 6780
},
{
"epoch": 0.42719056974459724,
"grad_norm": 0.5625,
"learning_rate": 0.001,
"loss": 0.8017,
"step": 6795
},
{
"epoch": 0.4281335952848723,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.7998,
"step": 6810
},
{
"epoch": 0.42907662082514736,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.8083,
"step": 6825
},
{
"epoch": 0.4300196463654224,
"grad_norm": 0.46484375,
"learning_rate": 0.001,
"loss": 0.7701,
"step": 6840
},
{
"epoch": 0.4309626719056974,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.7922,
"step": 6855
},
{
"epoch": 0.4319056974459725,
"grad_norm": 0.50390625,
"learning_rate": 0.001,
"loss": 0.7971,
"step": 6870
},
{
"epoch": 0.43284872298624755,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.795,
"step": 6885
},
{
"epoch": 0.4337917485265226,
"grad_norm": 0.5703125,
"learning_rate": 0.001,
"loss": 0.8004,
"step": 6900
},
{
"epoch": 0.43473477406679767,
"grad_norm": 0.59375,
"learning_rate": 0.001,
"loss": 0.7965,
"step": 6915
},
{
"epoch": 0.4356777996070727,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.7937,
"step": 6930
},
{
"epoch": 0.43662082514734774,
"grad_norm": 0.6171875,
"learning_rate": 0.001,
"loss": 0.8007,
"step": 6945
},
{
"epoch": 0.43756385068762277,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.7935,
"step": 6960
},
{
"epoch": 0.43850687622789786,
"grad_norm": 0.404296875,
"learning_rate": 0.001,
"loss": 0.8045,
"step": 6975
},
{
"epoch": 0.4394499017681729,
"grad_norm": 0.5703125,
"learning_rate": 0.001,
"loss": 0.8055,
"step": 6990
},
{
"epoch": 0.4403929273084479,
"grad_norm": 0.447265625,
"learning_rate": 0.001,
"loss": 0.8005,
"step": 7005
},
{
"epoch": 0.441335952848723,
"grad_norm": 0.72265625,
"learning_rate": 0.001,
"loss": 0.7881,
"step": 7020
},
{
"epoch": 0.44227897838899805,
"grad_norm": 0.73046875,
"learning_rate": 0.001,
"loss": 0.8212,
"step": 7035
},
{
"epoch": 0.4432220039292731,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.7984,
"step": 7050
},
{
"epoch": 0.4441650294695481,
"grad_norm": 0.5703125,
"learning_rate": 0.001,
"loss": 0.8078,
"step": 7065
},
{
"epoch": 0.4451080550098232,
"grad_norm": 0.42578125,
"learning_rate": 0.001,
"loss": 0.7773,
"step": 7080
},
{
"epoch": 0.44605108055009823,
"grad_norm": 0.486328125,
"learning_rate": 0.001,
"loss": 0.7884,
"step": 7095
},
{
"epoch": 0.44699410609037327,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.7842,
"step": 7110
},
{
"epoch": 0.44793713163064836,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.7854,
"step": 7125
},
{
"epoch": 0.4488801571709234,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.7913,
"step": 7140
},
{
"epoch": 0.4498231827111984,
"grad_norm": 0.50390625,
"learning_rate": 0.001,
"loss": 0.7944,
"step": 7155
},
{
"epoch": 0.45076620825147345,
"grad_norm": 0.578125,
"learning_rate": 0.001,
"loss": 0.7935,
"step": 7170
},
{
"epoch": 0.45170923379174854,
"grad_norm": 0.451171875,
"learning_rate": 0.001,
"loss": 0.7915,
"step": 7185
},
{
"epoch": 0.4526522593320236,
"grad_norm": 0.58984375,
"learning_rate": 0.001,
"loss": 0.7893,
"step": 7200
},
{
"epoch": 0.4535952848722986,
"grad_norm": 0.48046875,
"learning_rate": 0.001,
"loss": 0.7749,
"step": 7215
},
{
"epoch": 0.4545383104125737,
"grad_norm": 0.478515625,
"learning_rate": 0.001,
"loss": 0.7738,
"step": 7230
},
{
"epoch": 0.45548133595284873,
"grad_norm": 0.50390625,
"learning_rate": 0.001,
"loss": 0.7832,
"step": 7245
},
{
"epoch": 0.45642436149312376,
"grad_norm": 0.734375,
"learning_rate": 0.001,
"loss": 0.7935,
"step": 7260
},
{
"epoch": 0.4573673870333988,
"grad_norm": 0.6015625,
"learning_rate": 0.001,
"loss": 0.7969,
"step": 7275
},
{
"epoch": 0.4583104125736739,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 0.7891,
"step": 7290
},
{
"epoch": 0.4592534381139489,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.7854,
"step": 7305
},
{
"epoch": 0.46019646365422395,
"grad_norm": 0.62109375,
"learning_rate": 0.001,
"loss": 0.8013,
"step": 7320
},
{
"epoch": 0.46113948919449904,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.7864,
"step": 7335
},
{
"epoch": 0.4620825147347741,
"grad_norm": 0.474609375,
"learning_rate": 0.001,
"loss": 0.7932,
"step": 7350
},
{
"epoch": 0.4630255402750491,
"grad_norm": 0.59765625,
"learning_rate": 0.001,
"loss": 0.7866,
"step": 7365
},
{
"epoch": 0.46396856581532414,
"grad_norm": 0.640625,
"learning_rate": 0.001,
"loss": 0.8011,
"step": 7380
},
{
"epoch": 0.46491159135559923,
"grad_norm": 0.57421875,
"learning_rate": 0.001,
"loss": 0.7743,
"step": 7395
},
{
"epoch": 0.46585461689587426,
"grad_norm": 0.47265625,
"learning_rate": 0.001,
"loss": 0.7784,
"step": 7410
},
{
"epoch": 0.4667976424361493,
"grad_norm": 0.6328125,
"learning_rate": 0.001,
"loss": 0.7953,
"step": 7425
},
{
"epoch": 0.4677406679764244,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.7807,
"step": 7440
},
{
"epoch": 0.4686836935166994,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.7713,
"step": 7455
},
{
"epoch": 0.46962671905697445,
"grad_norm": 0.609375,
"learning_rate": 0.001,
"loss": 0.7636,
"step": 7470
},
{
"epoch": 0.4705697445972495,
"grad_norm": 0.443359375,
"learning_rate": 0.001,
"loss": 0.7773,
"step": 7485
},
{
"epoch": 0.4715127701375246,
"grad_norm": 0.65625,
"learning_rate": 0.001,
"loss": 0.8002,
"step": 7500
},
{
"epoch": 0.4724557956777996,
"grad_norm": 0.42578125,
"learning_rate": 0.001,
"loss": 0.7799,
"step": 7515
},
{
"epoch": 0.47339882121807464,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.7776,
"step": 7530
},
{
"epoch": 0.47434184675834973,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.7823,
"step": 7545
},
{
"epoch": 0.47528487229862476,
"grad_norm": 0.640625,
"learning_rate": 0.001,
"loss": 0.8059,
"step": 7560
},
{
"epoch": 0.4762278978388998,
"grad_norm": 0.5546875,
"learning_rate": 0.001,
"loss": 0.7908,
"step": 7575
},
{
"epoch": 0.4771709233791748,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.7923,
"step": 7590
},
{
"epoch": 0.4781139489194499,
"grad_norm": 0.49609375,
"learning_rate": 0.001,
"loss": 0.778,
"step": 7605
},
{
"epoch": 0.47905697445972495,
"grad_norm": 0.62890625,
"learning_rate": 0.001,
"loss": 0.8007,
"step": 7620
},
{
"epoch": 0.48,
"grad_norm": 0.64453125,
"learning_rate": 0.001,
"loss": 0.7842,
"step": 7635
},
{
"epoch": 0.48094302554027507,
"grad_norm": 0.46484375,
"learning_rate": 0.001,
"loss": 0.7968,
"step": 7650
},
{
"epoch": 0.4818860510805501,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.7812,
"step": 7665
},
{
"epoch": 0.48282907662082514,
"grad_norm": 1.5078125,
"learning_rate": 0.001,
"loss": 0.7832,
"step": 7680
},
{
"epoch": 0.48377210216110017,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.7915,
"step": 7695
},
{
"epoch": 0.48471512770137526,
"grad_norm": 0.609375,
"learning_rate": 0.001,
"loss": 0.8046,
"step": 7710
},
{
"epoch": 0.4856581532416503,
"grad_norm": 0.58984375,
"learning_rate": 0.001,
"loss": 0.7674,
"step": 7725
},
{
"epoch": 0.4866011787819253,
"grad_norm": 0.490234375,
"learning_rate": 0.001,
"loss": 0.7795,
"step": 7740
},
{
"epoch": 0.4875442043222004,
"grad_norm": 0.5703125,
"learning_rate": 0.001,
"loss": 0.7983,
"step": 7755
},
{
"epoch": 0.48848722986247545,
"grad_norm": 0.5546875,
"learning_rate": 0.001,
"loss": 0.7897,
"step": 7770
},
{
"epoch": 0.4894302554027505,
"grad_norm": 0.73828125,
"learning_rate": 0.001,
"loss": 0.772,
"step": 7785
},
{
"epoch": 0.4903732809430255,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 0.7795,
"step": 7800
},
{
"epoch": 0.4913163064833006,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.7739,
"step": 7815
},
{
"epoch": 0.49225933202357564,
"grad_norm": 0.66015625,
"learning_rate": 0.001,
"loss": 0.7891,
"step": 7830
},
{
"epoch": 0.49320235756385067,
"grad_norm": 0.57421875,
"learning_rate": 0.001,
"loss": 0.7802,
"step": 7845
},
{
"epoch": 0.49414538310412576,
"grad_norm": 0.56640625,
"learning_rate": 0.001,
"loss": 0.7843,
"step": 7860
},
{
"epoch": 0.4950884086444008,
"grad_norm": 0.5625,
"learning_rate": 0.001,
"loss": 0.7756,
"step": 7875
},
{
"epoch": 0.4960314341846758,
"grad_norm": 0.62109375,
"learning_rate": 0.001,
"loss": 0.77,
"step": 7890
},
{
"epoch": 0.49697445972495086,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.7633,
"step": 7905
},
{
"epoch": 0.49791748526522595,
"grad_norm": 0.5703125,
"learning_rate": 0.001,
"loss": 0.7842,
"step": 7920
},
{
"epoch": 0.498860510805501,
"grad_norm": 0.484375,
"learning_rate": 0.001,
"loss": 0.7742,
"step": 7935
},
{
"epoch": 0.499803536345776,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.7608,
"step": 7950
},
{
"epoch": 0.499803536345776,
"eval_loss": 0.9156466126441956,
"eval_runtime": 9.6921,
"eval_samples_per_second": 103.176,
"eval_steps_per_second": 1.444,
"step": 7950
},
{
"epoch": 0.500746561886051,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.7861,
"step": 7965
},
{
"epoch": 0.5016895874263261,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.7726,
"step": 7980
},
{
"epoch": 0.5026326129666012,
"grad_norm": 0.5,
"learning_rate": 0.001,
"loss": 0.7749,
"step": 7995
},
{
"epoch": 0.5035756385068763,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.7686,
"step": 8010
},
{
"epoch": 0.5045186640471513,
"grad_norm": 0.5703125,
"learning_rate": 0.001,
"loss": 0.7797,
"step": 8025
},
{
"epoch": 0.5054616895874263,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.7622,
"step": 8040
},
{
"epoch": 0.5064047151277014,
"grad_norm": 0.43359375,
"learning_rate": 0.001,
"loss": 0.7753,
"step": 8055
},
{
"epoch": 0.5073477406679764,
"grad_norm": 0.474609375,
"learning_rate": 0.001,
"loss": 0.7744,
"step": 8070
},
{
"epoch": 0.5082907662082514,
"grad_norm": 0.5,
"learning_rate": 0.001,
"loss": 0.7659,
"step": 8085
},
{
"epoch": 0.5092337917485266,
"grad_norm": 0.69140625,
"learning_rate": 0.001,
"loss": 0.7883,
"step": 8100
},
{
"epoch": 0.5101768172888016,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.7809,
"step": 8115
},
{
"epoch": 0.5111198428290766,
"grad_norm": 0.482421875,
"learning_rate": 0.001,
"loss": 0.7701,
"step": 8130
},
{
"epoch": 0.5120628683693517,
"grad_norm": 0.6953125,
"learning_rate": 0.001,
"loss": 0.7659,
"step": 8145
},
{
"epoch": 0.5130058939096267,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 0.7772,
"step": 8160
},
{
"epoch": 0.5139489194499017,
"grad_norm": 0.498046875,
"learning_rate": 0.001,
"loss": 0.7769,
"step": 8175
},
{
"epoch": 0.5148919449901768,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.7706,
"step": 8190
},
{
"epoch": 0.5158349705304519,
"grad_norm": 0.43359375,
"learning_rate": 0.001,
"loss": 0.7645,
"step": 8205
},
{
"epoch": 0.5167779960707269,
"grad_norm": 0.4765625,
"learning_rate": 0.001,
"loss": 0.7724,
"step": 8220
},
{
"epoch": 0.517721021611002,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.7651,
"step": 8235
},
{
"epoch": 0.518664047151277,
"grad_norm": 0.609375,
"learning_rate": 0.001,
"loss": 0.7703,
"step": 8250
},
{
"epoch": 0.519607072691552,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.7709,
"step": 8265
},
{
"epoch": 0.5205500982318271,
"grad_norm": 0.6796875,
"learning_rate": 0.001,
"loss": 0.7759,
"step": 8280
},
{
"epoch": 0.5214931237721021,
"grad_norm": 0.56640625,
"learning_rate": 0.001,
"loss": 0.7687,
"step": 8295
},
{
"epoch": 0.5224361493123773,
"grad_norm": 0.609375,
"learning_rate": 0.001,
"loss": 0.7735,
"step": 8310
},
{
"epoch": 0.5233791748526523,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.7653,
"step": 8325
},
{
"epoch": 0.5243222003929273,
"grad_norm": 0.484375,
"learning_rate": 0.001,
"loss": 0.766,
"step": 8340
},
{
"epoch": 0.5252652259332024,
"grad_norm": 0.59375,
"learning_rate": 0.001,
"loss": 0.768,
"step": 8355
},
{
"epoch": 0.5262082514734774,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.7651,
"step": 8370
},
{
"epoch": 0.5271512770137524,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.77,
"step": 8385
},
{
"epoch": 0.5280943025540275,
"grad_norm": 0.6328125,
"learning_rate": 0.001,
"loss": 0.7671,
"step": 8400
},
{
"epoch": 0.5290373280943026,
"grad_norm": 0.63671875,
"learning_rate": 0.001,
"loss": 0.7568,
"step": 8415
},
{
"epoch": 0.5299803536345776,
"grad_norm": 0.59765625,
"learning_rate": 0.001,
"loss": 0.7719,
"step": 8430
},
{
"epoch": 0.5309233791748527,
"grad_norm": 0.5703125,
"learning_rate": 0.001,
"loss": 0.7765,
"step": 8445
},
{
"epoch": 0.5318664047151277,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.7713,
"step": 8460
},
{
"epoch": 0.5328094302554027,
"grad_norm": 0.734375,
"learning_rate": 0.001,
"loss": 0.7779,
"step": 8475
},
{
"epoch": 0.5337524557956778,
"grad_norm": 0.5546875,
"learning_rate": 0.001,
"loss": 0.7675,
"step": 8490
},
{
"epoch": 0.5346954813359528,
"grad_norm": 0.431640625,
"learning_rate": 0.001,
"loss": 0.7652,
"step": 8505
},
{
"epoch": 0.5356385068762279,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.7692,
"step": 8520
},
{
"epoch": 0.536581532416503,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.7781,
"step": 8535
},
{
"epoch": 0.537524557956778,
"grad_norm": 0.59375,
"learning_rate": 0.001,
"loss": 0.765,
"step": 8550
},
{
"epoch": 0.538467583497053,
"grad_norm": 0.84765625,
"learning_rate": 0.001,
"loss": 0.7549,
"step": 8565
},
{
"epoch": 0.5394106090373281,
"grad_norm": 0.65625,
"learning_rate": 0.001,
"loss": 0.7709,
"step": 8580
},
{
"epoch": 0.5403536345776031,
"grad_norm": 0.484375,
"learning_rate": 0.001,
"loss": 0.7739,
"step": 8595
},
{
"epoch": 0.5412966601178782,
"grad_norm": 0.76171875,
"learning_rate": 0.001,
"loss": 0.769,
"step": 8610
},
{
"epoch": 0.5422396856581533,
"grad_norm": 0.6171875,
"learning_rate": 0.001,
"loss": 0.7737,
"step": 8625
},
{
"epoch": 0.5431827111984283,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.7638,
"step": 8640
},
{
"epoch": 0.5441257367387033,
"grad_norm": 0.462890625,
"learning_rate": 0.001,
"loss": 0.7392,
"step": 8655
},
{
"epoch": 0.5450687622789784,
"grad_norm": 0.71484375,
"learning_rate": 0.001,
"loss": 0.7566,
"step": 8670
},
{
"epoch": 0.5460117878192534,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.7592,
"step": 8685
},
{
"epoch": 0.5469548133595284,
"grad_norm": 0.4453125,
"learning_rate": 0.001,
"loss": 0.7485,
"step": 8700
},
{
"epoch": 0.5478978388998036,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.7678,
"step": 8715
},
{
"epoch": 0.5488408644400786,
"grad_norm": 0.59765625,
"learning_rate": 0.001,
"loss": 0.7634,
"step": 8730
},
{
"epoch": 0.5497838899803537,
"grad_norm": 0.4375,
"learning_rate": 0.001,
"loss": 0.7471,
"step": 8745
},
{
"epoch": 0.5507269155206287,
"grad_norm": 1.0234375,
"learning_rate": 0.001,
"loss": 0.7561,
"step": 8760
},
{
"epoch": 0.5516699410609037,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.7622,
"step": 8775
},
{
"epoch": 0.5526129666011788,
"grad_norm": 0.4453125,
"learning_rate": 0.001,
"loss": 0.7701,
"step": 8790
},
{
"epoch": 0.5535559921414538,
"grad_norm": 0.7109375,
"learning_rate": 0.001,
"loss": 0.7728,
"step": 8805
},
{
"epoch": 0.5544990176817289,
"grad_norm": 0.6328125,
"learning_rate": 0.001,
"loss": 0.7813,
"step": 8820
},
{
"epoch": 0.555442043222004,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.7614,
"step": 8835
},
{
"epoch": 0.556385068762279,
"grad_norm": 0.7890625,
"learning_rate": 0.001,
"loss": 0.7766,
"step": 8850
},
{
"epoch": 0.557328094302554,
"grad_norm": 0.6015625,
"learning_rate": 0.001,
"loss": 0.7735,
"step": 8865
},
{
"epoch": 0.5582711198428291,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.7641,
"step": 8880
},
{
"epoch": 0.5592141453831041,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.7798,
"step": 8895
},
{
"epoch": 0.5601571709233791,
"grad_norm": 0.484375,
"learning_rate": 0.001,
"loss": 0.7471,
"step": 8910
},
{
"epoch": 0.5611001964636543,
"grad_norm": 0.61328125,
"learning_rate": 0.001,
"loss": 0.7625,
"step": 8925
},
{
"epoch": 0.5620432220039293,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.7631,
"step": 8940
},
{
"epoch": 0.5629862475442043,
"grad_norm": 0.578125,
"learning_rate": 0.001,
"loss": 0.7679,
"step": 8955
},
{
"epoch": 0.5639292730844794,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.7647,
"step": 8970
},
{
"epoch": 0.5648722986247544,
"grad_norm": 0.466796875,
"learning_rate": 0.001,
"loss": 0.7674,
"step": 8985
},
{
"epoch": 0.5658153241650294,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.7735,
"step": 9000
},
{
"epoch": 0.5667583497053045,
"grad_norm": 0.5,
"learning_rate": 0.001,
"loss": 0.7826,
"step": 9015
},
{
"epoch": 0.5677013752455796,
"grad_norm": 0.458984375,
"learning_rate": 0.001,
"loss": 0.764,
"step": 9030
},
{
"epoch": 0.5686444007858547,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.7535,
"step": 9045
},
{
"epoch": 0.5695874263261297,
"grad_norm": 0.60546875,
"learning_rate": 0.001,
"loss": 0.7588,
"step": 9060
},
{
"epoch": 0.5705304518664047,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.7622,
"step": 9075
},
{
"epoch": 0.5714734774066798,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.7514,
"step": 9090
},
{
"epoch": 0.5724165029469548,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.7593,
"step": 9105
},
{
"epoch": 0.5733595284872298,
"grad_norm": 0.490234375,
"learning_rate": 0.001,
"loss": 0.7677,
"step": 9120
},
{
"epoch": 0.574302554027505,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.7539,
"step": 9135
},
{
"epoch": 0.57524557956778,
"grad_norm": 0.4921875,
"learning_rate": 0.001,
"loss": 0.7475,
"step": 9150
},
{
"epoch": 0.576188605108055,
"grad_norm": 0.470703125,
"learning_rate": 0.001,
"loss": 0.741,
"step": 9165
},
{
"epoch": 0.5771316306483301,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.7533,
"step": 9180
},
{
"epoch": 0.5780746561886051,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.765,
"step": 9195
},
{
"epoch": 0.5790176817288801,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 0.7741,
"step": 9210
},
{
"epoch": 0.5799607072691552,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.7598,
"step": 9225
},
{
"epoch": 0.5809037328094303,
"grad_norm": 0.453125,
"learning_rate": 0.001,
"loss": 0.7539,
"step": 9240
},
{
"epoch": 0.5818467583497053,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.7455,
"step": 9255
},
{
"epoch": 0.5827897838899804,
"grad_norm": 0.47265625,
"learning_rate": 0.001,
"loss": 0.7506,
"step": 9270
},
{
"epoch": 0.5837328094302554,
"grad_norm": 0.74609375,
"learning_rate": 0.001,
"loss": 0.7555,
"step": 9285
},
{
"epoch": 0.5846758349705304,
"grad_norm": 0.62109375,
"learning_rate": 0.001,
"loss": 0.7635,
"step": 9300
},
{
"epoch": 0.5856188605108055,
"grad_norm": 0.484375,
"learning_rate": 0.001,
"loss": 0.7351,
"step": 9315
},
{
"epoch": 0.5865618860510805,
"grad_norm": 0.59375,
"learning_rate": 0.001,
"loss": 0.7341,
"step": 9330
},
{
"epoch": 0.5875049115913556,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.7525,
"step": 9345
},
{
"epoch": 0.5884479371316307,
"grad_norm": 0.59375,
"learning_rate": 0.001,
"loss": 0.7575,
"step": 9360
},
{
"epoch": 0.5893909626719057,
"grad_norm": 0.50390625,
"learning_rate": 0.001,
"loss": 0.7608,
"step": 9375
},
{
"epoch": 0.5903339882121807,
"grad_norm": 0.59765625,
"learning_rate": 0.001,
"loss": 0.7602,
"step": 9390
},
{
"epoch": 0.5912770137524558,
"grad_norm": 0.68359375,
"learning_rate": 0.001,
"loss": 0.7615,
"step": 9405
},
{
"epoch": 0.5922200392927308,
"grad_norm": 0.4453125,
"learning_rate": 0.001,
"loss": 0.762,
"step": 9420
},
{
"epoch": 0.5931630648330058,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.7635,
"step": 9435
},
{
"epoch": 0.594106090373281,
"grad_norm": 0.455078125,
"learning_rate": 0.001,
"loss": 0.7556,
"step": 9450
},
{
"epoch": 0.595049115913556,
"grad_norm": 0.443359375,
"learning_rate": 0.001,
"loss": 0.7497,
"step": 9465
},
{
"epoch": 0.5959921414538311,
"grad_norm": 0.6015625,
"learning_rate": 0.001,
"loss": 0.7419,
"step": 9480
},
{
"epoch": 0.5969351669941061,
"grad_norm": 0.59765625,
"learning_rate": 0.001,
"loss": 0.7562,
"step": 9495
},
{
"epoch": 0.5978781925343811,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.7468,
"step": 9510
},
{
"epoch": 0.5988212180746562,
"grad_norm": 0.466796875,
"learning_rate": 0.001,
"loss": 0.7499,
"step": 9525
},
{
"epoch": 0.5997642436149312,
"grad_norm": 0.56640625,
"learning_rate": 0.001,
"loss": 0.7683,
"step": 9540
},
{
"epoch": 0.5997642436149312,
"eval_loss": 0.8865543603897095,
"eval_runtime": 9.6786,
"eval_samples_per_second": 103.32,
"eval_steps_per_second": 1.446,
"step": 9540
},
{
"epoch": 0.6007072691552063,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.7574,
"step": 9555
},
{
"epoch": 0.6016502946954814,
"grad_norm": 0.59765625,
"learning_rate": 0.001,
"loss": 0.7518,
"step": 9570
},
{
"epoch": 0.6025933202357564,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.7391,
"step": 9585
},
{
"epoch": 0.6035363457760314,
"grad_norm": 0.38671875,
"learning_rate": 0.001,
"loss": 0.7425,
"step": 9600
},
{
"epoch": 0.6044793713163065,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.7606,
"step": 9615
},
{
"epoch": 0.6054223968565815,
"grad_norm": 0.498046875,
"learning_rate": 0.001,
"loss": 0.7292,
"step": 9630
},
{
"epoch": 0.6063654223968565,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.7356,
"step": 9645
},
{
"epoch": 0.6073084479371317,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.7513,
"step": 9660
},
{
"epoch": 0.6082514734774067,
"grad_norm": 0.75,
"learning_rate": 0.001,
"loss": 0.7522,
"step": 9675
},
{
"epoch": 0.6091944990176817,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.7563,
"step": 9690
},
{
"epoch": 0.6101375245579568,
"grad_norm": 0.50390625,
"learning_rate": 0.001,
"loss": 0.7473,
"step": 9705
},
{
"epoch": 0.6110805500982318,
"grad_norm": 0.66796875,
"learning_rate": 0.001,
"loss": 0.76,
"step": 9720
},
{
"epoch": 0.6120235756385068,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.7473,
"step": 9735
},
{
"epoch": 0.6129666011787819,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.7416,
"step": 9750
},
{
"epoch": 0.613909626719057,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.7449,
"step": 9765
},
{
"epoch": 0.614852652259332,
"grad_norm": 0.6171875,
"learning_rate": 0.001,
"loss": 0.7509,
"step": 9780
},
{
"epoch": 0.6157956777996071,
"grad_norm": 0.625,
"learning_rate": 0.001,
"loss": 0.7468,
"step": 9795
},
{
"epoch": 0.6167387033398821,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.7632,
"step": 9810
},
{
"epoch": 0.6176817288801572,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.7586,
"step": 9825
},
{
"epoch": 0.6186247544204322,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.7495,
"step": 9840
},
{
"epoch": 0.6195677799607072,
"grad_norm": 0.474609375,
"learning_rate": 0.001,
"loss": 0.7548,
"step": 9855
},
{
"epoch": 0.6205108055009824,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.7484,
"step": 9870
},
{
"epoch": 0.6214538310412574,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.7683,
"step": 9885
},
{
"epoch": 0.6223968565815324,
"grad_norm": 0.47265625,
"learning_rate": 0.001,
"loss": 0.7332,
"step": 9900
},
{
"epoch": 0.6233398821218075,
"grad_norm": 0.61328125,
"learning_rate": 0.001,
"loss": 0.743,
"step": 9915
},
{
"epoch": 0.6242829076620825,
"grad_norm": 0.5625,
"learning_rate": 0.001,
"loss": 0.7527,
"step": 9930
},
{
"epoch": 0.6252259332023575,
"grad_norm": 0.5703125,
"learning_rate": 0.001,
"loss": 0.7407,
"step": 9945
},
{
"epoch": 0.6261689587426326,
"grad_norm": 0.462890625,
"learning_rate": 0.001,
"loss": 0.756,
"step": 9960
},
{
"epoch": 0.6271119842829077,
"grad_norm": 0.455078125,
"learning_rate": 0.001,
"loss": 0.7505,
"step": 9975
},
{
"epoch": 0.6280550098231827,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.7517,
"step": 9990
},
{
"epoch": 0.6289980353634578,
"grad_norm": 0.5546875,
"learning_rate": 0.001,
"loss": 0.766,
"step": 10005
},
{
"epoch": 0.6299410609037328,
"grad_norm": 0.66015625,
"learning_rate": 0.001,
"loss": 0.7385,
"step": 10020
},
{
"epoch": 0.6308840864440078,
"grad_norm": 0.7265625,
"learning_rate": 0.001,
"loss": 0.7565,
"step": 10035
},
{
"epoch": 0.6318271119842829,
"grad_norm": 0.62890625,
"learning_rate": 0.001,
"loss": 0.7508,
"step": 10050
},
{
"epoch": 0.6327701375245579,
"grad_norm": 0.60546875,
"learning_rate": 0.001,
"loss": 0.7519,
"step": 10065
},
{
"epoch": 0.633713163064833,
"grad_norm": 0.59375,
"learning_rate": 0.001,
"loss": 0.76,
"step": 10080
},
{
"epoch": 0.6346561886051081,
"grad_norm": 0.60546875,
"learning_rate": 0.001,
"loss": 0.7326,
"step": 10095
},
{
"epoch": 0.6355992141453831,
"grad_norm": 0.6328125,
"learning_rate": 0.001,
"loss": 0.7506,
"step": 10110
},
{
"epoch": 0.6365422396856582,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.7419,
"step": 10125
},
{
"epoch": 0.6374852652259332,
"grad_norm": 0.49609375,
"learning_rate": 0.001,
"loss": 0.7309,
"step": 10140
},
{
"epoch": 0.6384282907662082,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.7367,
"step": 10155
},
{
"epoch": 0.6393713163064833,
"grad_norm": 0.66796875,
"learning_rate": 0.001,
"loss": 0.7472,
"step": 10170
},
{
"epoch": 0.6403143418467584,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.7431,
"step": 10185
},
{
"epoch": 0.6412573673870334,
"grad_norm": 0.490234375,
"learning_rate": 0.001,
"loss": 0.7496,
"step": 10200
},
{
"epoch": 0.6422003929273085,
"grad_norm": 0.5625,
"learning_rate": 0.001,
"loss": 0.741,
"step": 10215
},
{
"epoch": 0.6431434184675835,
"grad_norm": 0.57421875,
"learning_rate": 0.001,
"loss": 0.7548,
"step": 10230
},
{
"epoch": 0.6440864440078585,
"grad_norm": 0.59375,
"learning_rate": 0.001,
"loss": 0.7615,
"step": 10245
},
{
"epoch": 0.6450294695481336,
"grad_norm": 0.494140625,
"learning_rate": 0.001,
"loss": 0.764,
"step": 10260
},
{
"epoch": 0.6459724950884086,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.7467,
"step": 10275
},
{
"epoch": 0.6469155206286837,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.752,
"step": 10290
},
{
"epoch": 0.6478585461689588,
"grad_norm": 0.4921875,
"learning_rate": 0.001,
"loss": 0.7238,
"step": 10305
},
{
"epoch": 0.6488015717092338,
"grad_norm": 0.48046875,
"learning_rate": 0.001,
"loss": 0.7464,
"step": 10320
},
{
"epoch": 0.6497445972495088,
"grad_norm": 0.455078125,
"learning_rate": 0.001,
"loss": 0.7376,
"step": 10335
},
{
"epoch": 0.6506876227897839,
"grad_norm": 0.65625,
"learning_rate": 0.001,
"loss": 0.7378,
"step": 10350
},
{
"epoch": 0.6516306483300589,
"grad_norm": 0.625,
"learning_rate": 0.001,
"loss": 0.7536,
"step": 10365
},
{
"epoch": 0.6525736738703339,
"grad_norm": 0.4921875,
"learning_rate": 0.001,
"loss": 0.732,
"step": 10380
},
{
"epoch": 0.6535166994106091,
"grad_norm": 0.5546875,
"learning_rate": 0.001,
"loss": 0.7554,
"step": 10395
},
{
"epoch": 0.6544597249508841,
"grad_norm": 0.4765625,
"learning_rate": 0.001,
"loss": 0.7348,
"step": 10410
},
{
"epoch": 0.6554027504911591,
"grad_norm": 0.5546875,
"learning_rate": 0.001,
"loss": 0.7446,
"step": 10425
},
{
"epoch": 0.6563457760314342,
"grad_norm": 0.62890625,
"learning_rate": 0.001,
"loss": 0.7386,
"step": 10440
},
{
"epoch": 0.6572888015717092,
"grad_norm": 0.6328125,
"learning_rate": 0.001,
"loss": 0.7456,
"step": 10455
},
{
"epoch": 0.6582318271119842,
"grad_norm": 0.75,
"learning_rate": 0.001,
"loss": 0.7447,
"step": 10470
},
{
"epoch": 0.6591748526522593,
"grad_norm": 0.6015625,
"learning_rate": 0.001,
"loss": 0.7466,
"step": 10485
},
{
"epoch": 0.6601178781925344,
"grad_norm": 0.75390625,
"learning_rate": 0.001,
"loss": 0.7638,
"step": 10500
},
{
"epoch": 0.6610609037328095,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.7454,
"step": 10515
},
{
"epoch": 0.6620039292730845,
"grad_norm": 0.48046875,
"learning_rate": 0.001,
"loss": 0.738,
"step": 10530
},
{
"epoch": 0.6629469548133595,
"grad_norm": 0.66796875,
"learning_rate": 0.001,
"loss": 0.7443,
"step": 10545
},
{
"epoch": 0.6638899803536346,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.7433,
"step": 10560
},
{
"epoch": 0.6648330058939096,
"grad_norm": 0.458984375,
"learning_rate": 0.001,
"loss": 0.7328,
"step": 10575
},
{
"epoch": 0.6657760314341846,
"grad_norm": 0.66015625,
"learning_rate": 0.001,
"loss": 0.7419,
"step": 10590
},
{
"epoch": 0.6667190569744598,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.7387,
"step": 10605
},
{
"epoch": 0.6676620825147348,
"grad_norm": 0.466796875,
"learning_rate": 0.001,
"loss": 0.7325,
"step": 10620
},
{
"epoch": 0.6686051080550098,
"grad_norm": 0.5625,
"learning_rate": 0.001,
"loss": 0.737,
"step": 10635
},
{
"epoch": 0.6695481335952849,
"grad_norm": 0.609375,
"learning_rate": 0.001,
"loss": 0.7447,
"step": 10650
},
{
"epoch": 0.6704911591355599,
"grad_norm": 0.484375,
"learning_rate": 0.001,
"loss": 0.7332,
"step": 10665
},
{
"epoch": 0.6714341846758349,
"grad_norm": 0.80078125,
"learning_rate": 0.001,
"loss": 0.7459,
"step": 10680
},
{
"epoch": 0.67237721021611,
"grad_norm": 0.5703125,
"learning_rate": 0.001,
"loss": 0.7389,
"step": 10695
},
{
"epoch": 0.6733202357563851,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.7362,
"step": 10710
},
{
"epoch": 0.6742632612966601,
"grad_norm": 0.67578125,
"learning_rate": 0.001,
"loss": 0.7297,
"step": 10725
},
{
"epoch": 0.6752062868369352,
"grad_norm": 0.64453125,
"learning_rate": 0.001,
"loss": 0.7506,
"step": 10740
},
{
"epoch": 0.6761493123772102,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.7279,
"step": 10755
},
{
"epoch": 0.6770923379174852,
"grad_norm": 0.640625,
"learning_rate": 0.001,
"loss": 0.7329,
"step": 10770
},
{
"epoch": 0.6780353634577603,
"grad_norm": 0.6953125,
"learning_rate": 0.001,
"loss": 0.736,
"step": 10785
},
{
"epoch": 0.6789783889980353,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.7168,
"step": 10800
},
{
"epoch": 0.6799214145383105,
"grad_norm": 0.57421875,
"learning_rate": 0.001,
"loss": 0.7394,
"step": 10815
},
{
"epoch": 0.6808644400785855,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.7165,
"step": 10830
},
{
"epoch": 0.6818074656188605,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.7249,
"step": 10845
},
{
"epoch": 0.6827504911591356,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.732,
"step": 10860
},
{
"epoch": 0.6836935166994106,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.747,
"step": 10875
},
{
"epoch": 0.6846365422396856,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.7268,
"step": 10890
},
{
"epoch": 0.6855795677799607,
"grad_norm": 0.59765625,
"learning_rate": 0.001,
"loss": 0.7334,
"step": 10905
},
{
"epoch": 0.6865225933202358,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.7243,
"step": 10920
},
{
"epoch": 0.6874656188605108,
"grad_norm": 0.5703125,
"learning_rate": 0.001,
"loss": 0.7402,
"step": 10935
},
{
"epoch": 0.6884086444007859,
"grad_norm": 0.5625,
"learning_rate": 0.001,
"loss": 0.738,
"step": 10950
},
{
"epoch": 0.6893516699410609,
"grad_norm": 0.5625,
"learning_rate": 0.001,
"loss": 0.7309,
"step": 10965
},
{
"epoch": 0.6902946954813359,
"grad_norm": 0.75,
"learning_rate": 0.001,
"loss": 0.7551,
"step": 10980
},
{
"epoch": 0.691237721021611,
"grad_norm": 0.59765625,
"learning_rate": 0.001,
"loss": 0.7438,
"step": 10995
},
{
"epoch": 0.692180746561886,
"grad_norm": 0.56640625,
"learning_rate": 0.001,
"loss": 0.7353,
"step": 11010
},
{
"epoch": 0.6931237721021611,
"grad_norm": 0.6015625,
"learning_rate": 0.001,
"loss": 0.728,
"step": 11025
},
{
"epoch": 0.6940667976424362,
"grad_norm": 0.6484375,
"learning_rate": 0.001,
"loss": 0.7366,
"step": 11040
},
{
"epoch": 0.6950098231827112,
"grad_norm": 0.59375,
"learning_rate": 0.001,
"loss": 0.7424,
"step": 11055
},
{
"epoch": 0.6959528487229862,
"grad_norm": 0.6328125,
"learning_rate": 0.001,
"loss": 0.7434,
"step": 11070
},
{
"epoch": 0.6968958742632613,
"grad_norm": 0.671875,
"learning_rate": 0.001,
"loss": 0.7371,
"step": 11085
},
{
"epoch": 0.6978388998035363,
"grad_norm": 0.48828125,
"learning_rate": 0.001,
"loss": 0.7326,
"step": 11100
},
{
"epoch": 0.6987819253438114,
"grad_norm": 0.5703125,
"learning_rate": 0.001,
"loss": 0.7272,
"step": 11115
},
{
"epoch": 0.6997249508840865,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.738,
"step": 11130
},
{
"epoch": 0.6997249508840865,
"eval_loss": 0.8602269291877747,
"eval_runtime": 9.6753,
"eval_samples_per_second": 103.356,
"eval_steps_per_second": 1.447,
"step": 11130
},
{
"epoch": 0.7006679764243615,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.7375,
"step": 11145
},
{
"epoch": 0.7016110019646365,
"grad_norm": 0.75,
"learning_rate": 0.001,
"loss": 0.7545,
"step": 11160
},
{
"epoch": 0.7025540275049116,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.7482,
"step": 11175
},
{
"epoch": 0.7034970530451866,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.7274,
"step": 11190
},
{
"epoch": 0.7044400785854616,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.7241,
"step": 11205
},
{
"epoch": 0.7053831041257368,
"grad_norm": 0.5625,
"learning_rate": 0.001,
"loss": 0.7303,
"step": 11220
},
{
"epoch": 0.7063261296660118,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.7267,
"step": 11235
},
{
"epoch": 0.7072691552062869,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.7267,
"step": 11250
},
{
"epoch": 0.7082121807465619,
"grad_norm": 0.48046875,
"learning_rate": 0.001,
"loss": 0.7309,
"step": 11265
},
{
"epoch": 0.7091552062868369,
"grad_norm": 0.59765625,
"learning_rate": 0.001,
"loss": 0.7377,
"step": 11280
},
{
"epoch": 0.710098231827112,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.7306,
"step": 11295
},
{
"epoch": 0.711041257367387,
"grad_norm": 0.75,
"learning_rate": 0.001,
"loss": 0.7341,
"step": 11310
},
{
"epoch": 0.7119842829076621,
"grad_norm": 0.6484375,
"learning_rate": 0.001,
"loss": 0.7349,
"step": 11325
},
{
"epoch": 0.7129273084479372,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.7407,
"step": 11340
},
{
"epoch": 0.7138703339882122,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.7358,
"step": 11355
},
{
"epoch": 0.7148133595284872,
"grad_norm": 0.474609375,
"learning_rate": 0.001,
"loss": 0.7254,
"step": 11370
},
{
"epoch": 0.7157563850687623,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.7328,
"step": 11385
},
{
"epoch": 0.7166994106090373,
"grad_norm": 0.6328125,
"learning_rate": 0.001,
"loss": 0.7304,
"step": 11400
},
{
"epoch": 0.7176424361493123,
"grad_norm": 0.484375,
"learning_rate": 0.001,
"loss": 0.7317,
"step": 11415
},
{
"epoch": 0.7185854616895875,
"grad_norm": 0.703125,
"learning_rate": 0.001,
"loss": 0.732,
"step": 11430
},
{
"epoch": 0.7195284872298625,
"grad_norm": 0.6484375,
"learning_rate": 0.001,
"loss": 0.7433,
"step": 11445
},
{
"epoch": 0.7204715127701375,
"grad_norm": 0.703125,
"learning_rate": 0.001,
"loss": 0.7415,
"step": 11460
},
{
"epoch": 0.7214145383104126,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.7285,
"step": 11475
},
{
"epoch": 0.7223575638506876,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.7299,
"step": 11490
},
{
"epoch": 0.7233005893909626,
"grad_norm": 0.72265625,
"learning_rate": 0.001,
"loss": 0.7314,
"step": 11505
},
{
"epoch": 0.7242436149312377,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.7413,
"step": 11520
},
{
"epoch": 0.7251866404715128,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.7266,
"step": 11535
},
{
"epoch": 0.7261296660117879,
"grad_norm": 0.4765625,
"learning_rate": 0.001,
"loss": 0.7104,
"step": 11550
},
{
"epoch": 0.7270726915520629,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.7252,
"step": 11565
},
{
"epoch": 0.7280157170923379,
"grad_norm": 0.578125,
"learning_rate": 0.001,
"loss": 0.726,
"step": 11580
},
{
"epoch": 0.728958742632613,
"grad_norm": 0.64453125,
"learning_rate": 0.001,
"loss": 0.7242,
"step": 11595
},
{
"epoch": 0.729901768172888,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.7313,
"step": 11610
},
{
"epoch": 0.730844793713163,
"grad_norm": 0.74609375,
"learning_rate": 0.001,
"loss": 0.7379,
"step": 11625
},
{
"epoch": 0.7317878192534382,
"grad_norm": 0.57421875,
"learning_rate": 0.001,
"loss": 0.7394,
"step": 11640
},
{
"epoch": 0.7327308447937132,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.7332,
"step": 11655
},
{
"epoch": 0.7336738703339882,
"grad_norm": 0.6484375,
"learning_rate": 0.001,
"loss": 0.7154,
"step": 11670
},
{
"epoch": 0.7346168958742633,
"grad_norm": 0.64453125,
"learning_rate": 0.001,
"loss": 0.7351,
"step": 11685
},
{
"epoch": 0.7355599214145383,
"grad_norm": 0.59765625,
"learning_rate": 0.001,
"loss": 0.7375,
"step": 11700
},
{
"epoch": 0.7365029469548133,
"grad_norm": 0.6640625,
"learning_rate": 0.001,
"loss": 0.7363,
"step": 11715
},
{
"epoch": 0.7374459724950884,
"grad_norm": 0.474609375,
"learning_rate": 0.001,
"loss": 0.7301,
"step": 11730
},
{
"epoch": 0.7383889980353635,
"grad_norm": 0.7734375,
"learning_rate": 0.001,
"loss": 0.7287,
"step": 11745
},
{
"epoch": 0.7393320235756385,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.7237,
"step": 11760
},
{
"epoch": 0.7402750491159136,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.7242,
"step": 11775
},
{
"epoch": 0.7412180746561886,
"grad_norm": 0.498046875,
"learning_rate": 0.001,
"loss": 0.7242,
"step": 11790
},
{
"epoch": 0.7421611001964636,
"grad_norm": 0.451171875,
"learning_rate": 0.001,
"loss": 0.7171,
"step": 11805
},
{
"epoch": 0.7431041257367387,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.7191,
"step": 11820
},
{
"epoch": 0.7440471512770137,
"grad_norm": 0.439453125,
"learning_rate": 0.001,
"loss": 0.7323,
"step": 11835
},
{
"epoch": 0.7449901768172889,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.7139,
"step": 11850
},
{
"epoch": 0.7459332023575639,
"grad_norm": 0.59765625,
"learning_rate": 0.001,
"loss": 0.7237,
"step": 11865
},
{
"epoch": 0.7468762278978389,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.7136,
"step": 11880
},
{
"epoch": 0.747819253438114,
"grad_norm": 0.75,
"learning_rate": 0.001,
"loss": 0.7375,
"step": 11895
},
{
"epoch": 0.748762278978389,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.7236,
"step": 11910
},
{
"epoch": 0.749705304518664,
"grad_norm": 0.62890625,
"learning_rate": 0.001,
"loss": 0.7416,
"step": 11925
},
{
"epoch": 0.750648330058939,
"grad_norm": 0.671875,
"learning_rate": 0.001,
"loss": 0.7376,
"step": 11940
},
{
"epoch": 0.7515913555992142,
"grad_norm": 0.640625,
"learning_rate": 0.001,
"loss": 0.7293,
"step": 11955
},
{
"epoch": 0.7525343811394892,
"grad_norm": 0.5,
"learning_rate": 0.001,
"loss": 0.7274,
"step": 11970
},
{
"epoch": 0.7534774066797643,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.7251,
"step": 11985
},
{
"epoch": 0.7544204322200393,
"grad_norm": 0.59765625,
"learning_rate": 0.001,
"loss": 0.7221,
"step": 12000
},
{
"epoch": 0.7553634577603143,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.7269,
"step": 12015
},
{
"epoch": 0.7563064833005894,
"grad_norm": 0.484375,
"learning_rate": 0.001,
"loss": 0.7229,
"step": 12030
},
{
"epoch": 0.7572495088408644,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.7332,
"step": 12045
},
{
"epoch": 0.7581925343811395,
"grad_norm": 0.62890625,
"learning_rate": 0.001,
"loss": 0.7425,
"step": 12060
},
{
"epoch": 0.7591355599214146,
"grad_norm": 0.50390625,
"learning_rate": 0.001,
"loss": 0.7084,
"step": 12075
},
{
"epoch": 0.7600785854616896,
"grad_norm": 0.453125,
"learning_rate": 0.001,
"loss": 0.7212,
"step": 12090
},
{
"epoch": 0.7610216110019646,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.7221,
"step": 12105
},
{
"epoch": 0.7619646365422397,
"grad_norm": 0.57421875,
"learning_rate": 0.001,
"loss": 0.7125,
"step": 12120
},
{
"epoch": 0.7629076620825147,
"grad_norm": 0.63671875,
"learning_rate": 0.001,
"loss": 0.7214,
"step": 12135
},
{
"epoch": 0.7638506876227897,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 0.7211,
"step": 12150
},
{
"epoch": 0.7647937131630649,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.7126,
"step": 12165
},
{
"epoch": 0.7657367387033399,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.726,
"step": 12180
},
{
"epoch": 0.766679764243615,
"grad_norm": 0.57421875,
"learning_rate": 0.001,
"loss": 0.7079,
"step": 12195
},
{
"epoch": 0.76762278978389,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.7282,
"step": 12210
},
{
"epoch": 0.768565815324165,
"grad_norm": 0.61328125,
"learning_rate": 0.001,
"loss": 0.7293,
"step": 12225
},
{
"epoch": 0.76950884086444,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 0.7245,
"step": 12240
},
{
"epoch": 0.7704518664047151,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.7263,
"step": 12255
},
{
"epoch": 0.7713948919449902,
"grad_norm": 0.76953125,
"learning_rate": 0.001,
"loss": 0.7483,
"step": 12270
},
{
"epoch": 0.7723379174852653,
"grad_norm": 0.6015625,
"learning_rate": 0.001,
"loss": 0.7243,
"step": 12285
},
{
"epoch": 0.7732809430255403,
"grad_norm": 0.41796875,
"learning_rate": 0.001,
"loss": 0.72,
"step": 12300
},
{
"epoch": 0.7742239685658153,
"grad_norm": 0.7421875,
"learning_rate": 0.001,
"loss": 0.7145,
"step": 12315
},
{
"epoch": 0.7751669941060904,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.7264,
"step": 12330
},
{
"epoch": 0.7761100196463654,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.7233,
"step": 12345
},
{
"epoch": 0.7770530451866404,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.7132,
"step": 12360
},
{
"epoch": 0.7779960707269156,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.7218,
"step": 12375
},
{
"epoch": 0.7789390962671906,
"grad_norm": 0.68359375,
"learning_rate": 0.001,
"loss": 0.7229,
"step": 12390
},
{
"epoch": 0.7798821218074656,
"grad_norm": 0.65625,
"learning_rate": 0.001,
"loss": 0.7244,
"step": 12405
},
{
"epoch": 0.7808251473477407,
"grad_norm": 0.50390625,
"learning_rate": 0.001,
"loss": 0.7133,
"step": 12420
},
{
"epoch": 0.7817681728880157,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.7165,
"step": 12435
},
{
"epoch": 0.7827111984282907,
"grad_norm": 0.474609375,
"learning_rate": 0.001,
"loss": 0.7125,
"step": 12450
},
{
"epoch": 0.7836542239685658,
"grad_norm": 0.49609375,
"learning_rate": 0.001,
"loss": 0.7025,
"step": 12465
},
{
"epoch": 0.7845972495088409,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.7143,
"step": 12480
},
{
"epoch": 0.7855402750491159,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.7217,
"step": 12495
},
{
"epoch": 0.786483300589391,
"grad_norm": 0.44921875,
"learning_rate": 0.001,
"loss": 0.7194,
"step": 12510
},
{
"epoch": 0.787426326129666,
"grad_norm": 0.6015625,
"learning_rate": 0.001,
"loss": 0.7117,
"step": 12525
},
{
"epoch": 0.788369351669941,
"grad_norm": 0.57421875,
"learning_rate": 0.001,
"loss": 0.7125,
"step": 12540
},
{
"epoch": 0.7893123772102161,
"grad_norm": 0.412109375,
"learning_rate": 0.001,
"loss": 0.7107,
"step": 12555
},
{
"epoch": 0.7902554027504911,
"grad_norm": 0.609375,
"learning_rate": 0.001,
"loss": 0.7392,
"step": 12570
},
{
"epoch": 0.7911984282907663,
"grad_norm": 0.67578125,
"learning_rate": 0.001,
"loss": 0.7211,
"step": 12585
},
{
"epoch": 0.7921414538310413,
"grad_norm": 0.9375,
"learning_rate": 0.001,
"loss": 0.7139,
"step": 12600
},
{
"epoch": 0.7930844793713163,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.721,
"step": 12615
},
{
"epoch": 0.7940275049115914,
"grad_norm": 0.478515625,
"learning_rate": 0.001,
"loss": 0.7258,
"step": 12630
},
{
"epoch": 0.7949705304518664,
"grad_norm": 0.5703125,
"learning_rate": 0.001,
"loss": 0.7079,
"step": 12645
},
{
"epoch": 0.7959135559921414,
"grad_norm": 0.498046875,
"learning_rate": 0.001,
"loss": 0.712,
"step": 12660
},
{
"epoch": 0.7968565815324165,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.7296,
"step": 12675
},
{
"epoch": 0.7977996070726916,
"grad_norm": 0.43359375,
"learning_rate": 0.001,
"loss": 0.7146,
"step": 12690
},
{
"epoch": 0.7987426326129666,
"grad_norm": 0.67578125,
"learning_rate": 0.001,
"loss": 0.7202,
"step": 12705
},
{
"epoch": 0.7996856581532417,
"grad_norm": 0.57421875,
"learning_rate": 0.001,
"loss": 0.7257,
"step": 12720
},
{
"epoch": 0.7996856581532417,
"eval_loss": 0.8420960307121277,
"eval_runtime": 9.6794,
"eval_samples_per_second": 103.312,
"eval_steps_per_second": 1.446,
"step": 12720
},
{
"epoch": 0.8006286836935167,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.7087,
"step": 12735
},
{
"epoch": 0.8015717092337917,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.7219,
"step": 12750
},
{
"epoch": 0.8025147347740668,
"grad_norm": 0.48046875,
"learning_rate": 0.001,
"loss": 0.7241,
"step": 12765
},
{
"epoch": 0.8034577603143418,
"grad_norm": 0.671875,
"learning_rate": 0.001,
"loss": 0.7211,
"step": 12780
},
{
"epoch": 0.8044007858546169,
"grad_norm": 0.578125,
"learning_rate": 0.001,
"loss": 0.7234,
"step": 12795
},
{
"epoch": 0.805343811394892,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.7214,
"step": 12810
},
{
"epoch": 0.806286836935167,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.7201,
"step": 12825
},
{
"epoch": 0.807229862475442,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.7261,
"step": 12840
},
{
"epoch": 0.8081728880157171,
"grad_norm": 0.56640625,
"learning_rate": 0.001,
"loss": 0.7301,
"step": 12855
},
{
"epoch": 0.8091159135559921,
"grad_norm": 0.484375,
"learning_rate": 0.001,
"loss": 0.7104,
"step": 12870
},
{
"epoch": 0.8100589390962671,
"grad_norm": 0.62890625,
"learning_rate": 0.001,
"loss": 0.7094,
"step": 12885
},
{
"epoch": 0.8110019646365423,
"grad_norm": 0.640625,
"learning_rate": 0.001,
"loss": 0.7188,
"step": 12900
},
{
"epoch": 0.8119449901768173,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.7296,
"step": 12915
},
{
"epoch": 0.8128880157170923,
"grad_norm": 0.58984375,
"learning_rate": 0.001,
"loss": 0.725,
"step": 12930
},
{
"epoch": 0.8138310412573674,
"grad_norm": 0.5546875,
"learning_rate": 0.001,
"loss": 0.716,
"step": 12945
},
{
"epoch": 0.8147740667976424,
"grad_norm": 0.60546875,
"learning_rate": 0.001,
"loss": 0.7199,
"step": 12960
},
{
"epoch": 0.8157170923379174,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.7251,
"step": 12975
},
{
"epoch": 0.8166601178781925,
"grad_norm": 0.47265625,
"learning_rate": 0.001,
"loss": 0.7153,
"step": 12990
},
{
"epoch": 0.8176031434184676,
"grad_norm": 0.6171875,
"learning_rate": 0.001,
"loss": 0.7172,
"step": 13005
},
{
"epoch": 0.8185461689587427,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.7241,
"step": 13020
},
{
"epoch": 0.8194891944990177,
"grad_norm": 0.671875,
"learning_rate": 0.001,
"loss": 0.7087,
"step": 13035
},
{
"epoch": 0.8204322200392927,
"grad_norm": 0.4375,
"learning_rate": 0.001,
"loss": 0.7146,
"step": 13050
},
{
"epoch": 0.8213752455795678,
"grad_norm": 0.42578125,
"learning_rate": 0.001,
"loss": 0.7137,
"step": 13065
},
{
"epoch": 0.8223182711198428,
"grad_norm": 0.61328125,
"learning_rate": 0.001,
"loss": 0.7309,
"step": 13080
},
{
"epoch": 0.8232612966601178,
"grad_norm": 0.74609375,
"learning_rate": 0.001,
"loss": 0.7075,
"step": 13095
},
{
"epoch": 0.824204322200393,
"grad_norm": 0.5,
"learning_rate": 0.001,
"loss": 0.7187,
"step": 13110
},
{
"epoch": 0.825147347740668,
"grad_norm": 0.56640625,
"learning_rate": 0.001,
"loss": 0.7133,
"step": 13125
},
{
"epoch": 0.826090373280943,
"grad_norm": 0.65234375,
"learning_rate": 0.001,
"loss": 0.7062,
"step": 13140
},
{
"epoch": 0.8270333988212181,
"grad_norm": 0.609375,
"learning_rate": 0.001,
"loss": 0.7139,
"step": 13155
},
{
"epoch": 0.8279764243614931,
"grad_norm": 0.43359375,
"learning_rate": 0.001,
"loss": 0.7122,
"step": 13170
},
{
"epoch": 0.8289194499017681,
"grad_norm": 0.60546875,
"learning_rate": 0.001,
"loss": 0.7089,
"step": 13185
},
{
"epoch": 0.8298624754420432,
"grad_norm": 0.6484375,
"learning_rate": 0.001,
"loss": 0.7148,
"step": 13200
},
{
"epoch": 0.8308055009823183,
"grad_norm": 0.484375,
"learning_rate": 0.001,
"loss": 0.7165,
"step": 13215
},
{
"epoch": 0.8317485265225933,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.716,
"step": 13230
},
{
"epoch": 0.8326915520628684,
"grad_norm": 0.60546875,
"learning_rate": 0.001,
"loss": 0.7324,
"step": 13245
},
{
"epoch": 0.8336345776031434,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.7028,
"step": 13260
},
{
"epoch": 0.8345776031434184,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.7162,
"step": 13275
},
{
"epoch": 0.8355206286836935,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.7115,
"step": 13290
},
{
"epoch": 0.8364636542239685,
"grad_norm": 0.60546875,
"learning_rate": 0.001,
"loss": 0.7323,
"step": 13305
},
{
"epoch": 0.8374066797642437,
"grad_norm": 0.62890625,
"learning_rate": 0.001,
"loss": 0.7154,
"step": 13320
},
{
"epoch": 0.8383497053045187,
"grad_norm": 0.640625,
"learning_rate": 0.001,
"loss": 0.7231,
"step": 13335
},
{
"epoch": 0.8392927308447937,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.7308,
"step": 13350
},
{
"epoch": 0.8402357563850688,
"grad_norm": 0.5703125,
"learning_rate": 0.001,
"loss": 0.7169,
"step": 13365
},
{
"epoch": 0.8411787819253438,
"grad_norm": 0.62890625,
"learning_rate": 0.001,
"loss": 0.7209,
"step": 13380
},
{
"epoch": 0.8421218074656188,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.706,
"step": 13395
},
{
"epoch": 0.8430648330058939,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.7079,
"step": 13410
},
{
"epoch": 0.844007858546169,
"grad_norm": 0.609375,
"learning_rate": 0.001,
"loss": 0.7062,
"step": 13425
},
{
"epoch": 0.844950884086444,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.7047,
"step": 13440
},
{
"epoch": 0.8458939096267191,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.7179,
"step": 13455
},
{
"epoch": 0.8468369351669941,
"grad_norm": 0.72265625,
"learning_rate": 0.001,
"loss": 0.7159,
"step": 13470
},
{
"epoch": 0.8477799607072691,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.7152,
"step": 13485
},
{
"epoch": 0.8487229862475442,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.709,
"step": 13500
},
{
"epoch": 0.8496660117878193,
"grad_norm": 0.498046875,
"learning_rate": 0.001,
"loss": 0.7158,
"step": 13515
},
{
"epoch": 0.8506090373280943,
"grad_norm": 0.62890625,
"learning_rate": 0.001,
"loss": 0.7026,
"step": 13530
},
{
"epoch": 0.8515520628683694,
"grad_norm": 0.640625,
"learning_rate": 0.001,
"loss": 0.7197,
"step": 13545
},
{
"epoch": 0.8524950884086444,
"grad_norm": 0.578125,
"learning_rate": 0.001,
"loss": 0.7271,
"step": 13560
},
{
"epoch": 0.8534381139489194,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.7241,
"step": 13575
},
{
"epoch": 0.8543811394891945,
"grad_norm": 0.578125,
"learning_rate": 0.001,
"loss": 0.7244,
"step": 13590
},
{
"epoch": 0.8553241650294695,
"grad_norm": 0.482421875,
"learning_rate": 0.001,
"loss": 0.7154,
"step": 13605
},
{
"epoch": 0.8562671905697447,
"grad_norm": 0.68359375,
"learning_rate": 0.001,
"loss": 0.7135,
"step": 13620
},
{
"epoch": 0.8572102161100197,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.7095,
"step": 13635
},
{
"epoch": 0.8581532416502947,
"grad_norm": 0.87109375,
"learning_rate": 0.001,
"loss": 0.7245,
"step": 13650
},
{
"epoch": 0.8590962671905698,
"grad_norm": 0.59765625,
"learning_rate": 0.001,
"loss": 0.7174,
"step": 13665
},
{
"epoch": 0.8600392927308448,
"grad_norm": 0.6875,
"learning_rate": 0.001,
"loss": 0.7131,
"step": 13680
},
{
"epoch": 0.8609823182711198,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.7227,
"step": 13695
},
{
"epoch": 0.8619253438113949,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.7067,
"step": 13710
},
{
"epoch": 0.86286836935167,
"grad_norm": 0.48828125,
"learning_rate": 0.001,
"loss": 0.7013,
"step": 13725
},
{
"epoch": 0.863811394891945,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.7046,
"step": 13740
},
{
"epoch": 0.8647544204322201,
"grad_norm": 0.578125,
"learning_rate": 0.001,
"loss": 0.7077,
"step": 13755
},
{
"epoch": 0.8656974459724951,
"grad_norm": 0.578125,
"learning_rate": 0.001,
"loss": 0.7064,
"step": 13770
},
{
"epoch": 0.8666404715127701,
"grad_norm": 0.62109375,
"learning_rate": 0.001,
"loss": 0.7177,
"step": 13785
},
{
"epoch": 0.8675834970530452,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.7128,
"step": 13800
},
{
"epoch": 0.8685265225933202,
"grad_norm": 0.68359375,
"learning_rate": 0.001,
"loss": 0.7131,
"step": 13815
},
{
"epoch": 0.8694695481335953,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.7048,
"step": 13830
},
{
"epoch": 0.8704125736738704,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.7183,
"step": 13845
},
{
"epoch": 0.8713555992141454,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.7087,
"step": 13860
},
{
"epoch": 0.8722986247544204,
"grad_norm": 0.470703125,
"learning_rate": 0.001,
"loss": 0.7117,
"step": 13875
},
{
"epoch": 0.8732416502946955,
"grad_norm": 0.4453125,
"learning_rate": 0.001,
"loss": 0.7216,
"step": 13890
},
{
"epoch": 0.8741846758349705,
"grad_norm": 0.6328125,
"learning_rate": 0.001,
"loss": 0.7159,
"step": 13905
},
{
"epoch": 0.8751277013752455,
"grad_norm": 0.75,
"learning_rate": 0.001,
"loss": 0.7096,
"step": 13920
},
{
"epoch": 0.8760707269155207,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.702,
"step": 13935
},
{
"epoch": 0.8770137524557957,
"grad_norm": 0.5625,
"learning_rate": 0.001,
"loss": 0.7101,
"step": 13950
},
{
"epoch": 0.8779567779960707,
"grad_norm": 0.59375,
"learning_rate": 0.001,
"loss": 0.7212,
"step": 13965
},
{
"epoch": 0.8788998035363458,
"grad_norm": 0.58984375,
"learning_rate": 0.001,
"loss": 0.7126,
"step": 13980
},
{
"epoch": 0.8798428290766208,
"grad_norm": 0.5078125,
"learning_rate": 0.001,
"loss": 0.7036,
"step": 13995
},
{
"epoch": 0.8807858546168958,
"grad_norm": 0.60546875,
"learning_rate": 0.001,
"loss": 0.7071,
"step": 14010
},
{
"epoch": 0.8817288801571709,
"grad_norm": 0.5546875,
"learning_rate": 0.001,
"loss": 0.7051,
"step": 14025
},
{
"epoch": 0.882671905697446,
"grad_norm": 0.46875,
"learning_rate": 0.001,
"loss": 0.7156,
"step": 14040
},
{
"epoch": 0.8836149312377211,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.709,
"step": 14055
},
{
"epoch": 0.8845579567779961,
"grad_norm": 0.6875,
"learning_rate": 0.001,
"loss": 0.7062,
"step": 14070
},
{
"epoch": 0.8855009823182711,
"grad_norm": 0.6015625,
"learning_rate": 0.001,
"loss": 0.7142,
"step": 14085
},
{
"epoch": 0.8864440078585462,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.7143,
"step": 14100
},
{
"epoch": 0.8873870333988212,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.7093,
"step": 14115
},
{
"epoch": 0.8883300589390962,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.712,
"step": 14130
},
{
"epoch": 0.8892730844793714,
"grad_norm": 0.51171875,
"learning_rate": 0.001,
"loss": 0.7085,
"step": 14145
},
{
"epoch": 0.8902161100196464,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.7197,
"step": 14160
},
{
"epoch": 0.8911591355599214,
"grad_norm": 0.64453125,
"learning_rate": 0.001,
"loss": 0.7022,
"step": 14175
},
{
"epoch": 0.8921021611001965,
"grad_norm": 0.5625,
"learning_rate": 0.001,
"loss": 0.7019,
"step": 14190
},
{
"epoch": 0.8930451866404715,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.7171,
"step": 14205
},
{
"epoch": 0.8939882121807465,
"grad_norm": 0.7890625,
"learning_rate": 0.001,
"loss": 0.7052,
"step": 14220
},
{
"epoch": 0.8949312377210216,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.7029,
"step": 14235
},
{
"epoch": 0.8958742632612967,
"grad_norm": 0.60546875,
"learning_rate": 0.001,
"loss": 0.7067,
"step": 14250
},
{
"epoch": 0.8968172888015717,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.6962,
"step": 14265
},
{
"epoch": 0.8977603143418468,
"grad_norm": 0.60546875,
"learning_rate": 0.001,
"loss": 0.702,
"step": 14280
},
{
"epoch": 0.8987033398821218,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.7066,
"step": 14295
},
{
"epoch": 0.8996463654223968,
"grad_norm": 0.62890625,
"learning_rate": 0.001,
"loss": 0.7141,
"step": 14310
},
{
"epoch": 0.8996463654223968,
"eval_loss": 0.8242524266242981,
"eval_runtime": 9.6736,
"eval_samples_per_second": 103.374,
"eval_steps_per_second": 1.447,
"step": 14310
},
{
"epoch": 0.9005893909626719,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.7051,
"step": 14325
},
{
"epoch": 0.9015324165029469,
"grad_norm": 0.6484375,
"learning_rate": 0.001,
"loss": 0.7161,
"step": 14340
},
{
"epoch": 0.902475442043222,
"grad_norm": 0.4609375,
"learning_rate": 0.001,
"loss": 0.6994,
"step": 14355
},
{
"epoch": 0.9034184675834971,
"grad_norm": 0.46875,
"learning_rate": 0.001,
"loss": 0.7121,
"step": 14370
},
{
"epoch": 0.9043614931237721,
"grad_norm": 0.62109375,
"learning_rate": 0.001,
"loss": 0.7232,
"step": 14385
},
{
"epoch": 0.9053045186640472,
"grad_norm": 0.4765625,
"learning_rate": 0.001,
"loss": 0.7122,
"step": 14400
},
{
"epoch": 0.9062475442043222,
"grad_norm": 0.52734375,
"learning_rate": 0.001,
"loss": 0.7168,
"step": 14415
},
{
"epoch": 0.9071905697445972,
"grad_norm": 0.474609375,
"learning_rate": 0.001,
"loss": 0.6997,
"step": 14430
},
{
"epoch": 0.9081335952848723,
"grad_norm": 0.56640625,
"learning_rate": 0.001,
"loss": 0.7124,
"step": 14445
},
{
"epoch": 0.9090766208251474,
"grad_norm": 0.486328125,
"learning_rate": 0.001,
"loss": 0.6995,
"step": 14460
},
{
"epoch": 0.9100196463654224,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.7087,
"step": 14475
},
{
"epoch": 0.9109626719056975,
"grad_norm": 0.58984375,
"learning_rate": 0.001,
"loss": 0.6991,
"step": 14490
},
{
"epoch": 0.9119056974459725,
"grad_norm": 0.65625,
"learning_rate": 0.001,
"loss": 0.7069,
"step": 14505
},
{
"epoch": 0.9128487229862475,
"grad_norm": 0.625,
"learning_rate": 0.001,
"loss": 0.701,
"step": 14520
},
{
"epoch": 0.9137917485265226,
"grad_norm": 0.734375,
"learning_rate": 0.001,
"loss": 0.7111,
"step": 14535
},
{
"epoch": 0.9147347740667976,
"grad_norm": 0.4765625,
"learning_rate": 0.001,
"loss": 0.6989,
"step": 14550
},
{
"epoch": 0.9156777996070727,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.7243,
"step": 14565
},
{
"epoch": 0.9166208251473478,
"grad_norm": 0.60546875,
"learning_rate": 0.001,
"loss": 0.7043,
"step": 14580
},
{
"epoch": 0.9175638506876228,
"grad_norm": 0.50390625,
"learning_rate": 0.001,
"loss": 0.6925,
"step": 14595
},
{
"epoch": 0.9185068762278978,
"grad_norm": 0.7890625,
"learning_rate": 0.001,
"loss": 0.7129,
"step": 14610
},
{
"epoch": 0.9194499017681729,
"grad_norm": 0.65625,
"learning_rate": 0.001,
"loss": 0.7064,
"step": 14625
},
{
"epoch": 0.9203929273084479,
"grad_norm": 0.451171875,
"learning_rate": 0.001,
"loss": 0.6876,
"step": 14640
},
{
"epoch": 0.9213359528487229,
"grad_norm": 0.6328125,
"learning_rate": 0.001,
"loss": 0.6978,
"step": 14655
},
{
"epoch": 0.9222789783889981,
"grad_norm": 0.57421875,
"learning_rate": 0.001,
"loss": 0.7003,
"step": 14670
},
{
"epoch": 0.9232220039292731,
"grad_norm": 0.4765625,
"learning_rate": 0.001,
"loss": 0.7009,
"step": 14685
},
{
"epoch": 0.9241650294695481,
"grad_norm": 0.609375,
"learning_rate": 0.001,
"loss": 0.7093,
"step": 14700
},
{
"epoch": 0.9251080550098232,
"grad_norm": 0.5703125,
"learning_rate": 0.001,
"loss": 0.6927,
"step": 14715
},
{
"epoch": 0.9260510805500982,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.6995,
"step": 14730
},
{
"epoch": 0.9269941060903732,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.711,
"step": 14745
},
{
"epoch": 0.9279371316306483,
"grad_norm": 0.57421875,
"learning_rate": 0.001,
"loss": 0.7156,
"step": 14760
},
{
"epoch": 0.9288801571709234,
"grad_norm": 0.72265625,
"learning_rate": 0.001,
"loss": 0.7173,
"step": 14775
},
{
"epoch": 0.9298231827111985,
"grad_norm": 0.7578125,
"learning_rate": 0.001,
"loss": 0.7132,
"step": 14790
},
{
"epoch": 0.9307662082514735,
"grad_norm": 0.47265625,
"learning_rate": 0.001,
"loss": 0.6983,
"step": 14805
},
{
"epoch": 0.9317092337917485,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.7047,
"step": 14820
},
{
"epoch": 0.9326522593320236,
"grad_norm": 0.5546875,
"learning_rate": 0.001,
"loss": 0.7115,
"step": 14835
},
{
"epoch": 0.9335952848722986,
"grad_norm": 0.5703125,
"learning_rate": 0.001,
"loss": 0.7038,
"step": 14850
},
{
"epoch": 0.9345383104125736,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.7066,
"step": 14865
},
{
"epoch": 0.9354813359528488,
"grad_norm": 0.4609375,
"learning_rate": 0.001,
"loss": 0.7062,
"step": 14880
},
{
"epoch": 0.9364243614931238,
"grad_norm": 0.4140625,
"learning_rate": 0.001,
"loss": 0.6915,
"step": 14895
},
{
"epoch": 0.9373673870333988,
"grad_norm": 0.64453125,
"learning_rate": 0.001,
"loss": 0.7031,
"step": 14910
},
{
"epoch": 0.9383104125736739,
"grad_norm": 0.6875,
"learning_rate": 0.001,
"loss": 0.7072,
"step": 14925
},
{
"epoch": 0.9392534381139489,
"grad_norm": 0.50390625,
"learning_rate": 0.001,
"loss": 0.7012,
"step": 14940
},
{
"epoch": 0.9401964636542239,
"grad_norm": 0.70703125,
"learning_rate": 0.001,
"loss": 0.7211,
"step": 14955
},
{
"epoch": 0.941139489194499,
"grad_norm": 0.4609375,
"learning_rate": 0.001,
"loss": 0.7048,
"step": 14970
},
{
"epoch": 0.9420825147347741,
"grad_norm": 0.5,
"learning_rate": 0.001,
"loss": 0.7016,
"step": 14985
},
{
"epoch": 0.9430255402750491,
"grad_norm": 0.490234375,
"learning_rate": 0.001,
"loss": 0.7095,
"step": 15000
},
{
"epoch": 0.9439685658153242,
"grad_norm": 0.458984375,
"learning_rate": 0.001,
"loss": 0.705,
"step": 15015
},
{
"epoch": 0.9449115913555992,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.6986,
"step": 15030
},
{
"epoch": 0.9458546168958742,
"grad_norm": 0.60546875,
"learning_rate": 0.001,
"loss": 0.7026,
"step": 15045
},
{
"epoch": 0.9467976424361493,
"grad_norm": 0.55859375,
"learning_rate": 0.001,
"loss": 0.709,
"step": 15060
},
{
"epoch": 0.9477406679764243,
"grad_norm": 0.59765625,
"learning_rate": 0.001,
"loss": 0.712,
"step": 15075
},
{
"epoch": 0.9486836935166995,
"grad_norm": 0.5625,
"learning_rate": 0.001,
"loss": 0.7126,
"step": 15090
},
{
"epoch": 0.9496267190569745,
"grad_norm": 0.75390625,
"learning_rate": 0.001,
"loss": 0.6879,
"step": 15105
},
{
"epoch": 0.9505697445972495,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.7031,
"step": 15120
},
{
"epoch": 0.9515127701375246,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.7146,
"step": 15135
},
{
"epoch": 0.9524557956777996,
"grad_norm": 0.48828125,
"learning_rate": 0.001,
"loss": 0.6882,
"step": 15150
},
{
"epoch": 0.9533988212180746,
"grad_norm": 0.50390625,
"learning_rate": 0.001,
"loss": 0.6981,
"step": 15165
},
{
"epoch": 0.9543418467583497,
"grad_norm": 0.578125,
"learning_rate": 0.001,
"loss": 0.7011,
"step": 15180
},
{
"epoch": 0.9552848722986248,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 0.698,
"step": 15195
},
{
"epoch": 0.9562278978388998,
"grad_norm": 0.640625,
"learning_rate": 0.001,
"loss": 0.6932,
"step": 15210
},
{
"epoch": 0.9571709233791749,
"grad_norm": 0.5546875,
"learning_rate": 0.001,
"loss": 0.69,
"step": 15225
},
{
"epoch": 0.9581139489194499,
"grad_norm": 0.609375,
"learning_rate": 0.001,
"loss": 0.695,
"step": 15240
},
{
"epoch": 0.9590569744597249,
"grad_norm": 0.5,
"learning_rate": 0.001,
"loss": 0.7002,
"step": 15255
},
{
"epoch": 0.96,
"grad_norm": 0.478515625,
"learning_rate": 0.001,
"loss": 0.6943,
"step": 15270
},
{
"epoch": 0.960943025540275,
"grad_norm": 0.58203125,
"learning_rate": 0.001,
"loss": 0.7044,
"step": 15285
},
{
"epoch": 0.9618860510805501,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.7069,
"step": 15300
},
{
"epoch": 0.9628290766208252,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.6985,
"step": 15315
},
{
"epoch": 0.9637721021611002,
"grad_norm": 0.640625,
"learning_rate": 0.001,
"loss": 0.7049,
"step": 15330
},
{
"epoch": 0.9647151277013752,
"grad_norm": 0.62890625,
"learning_rate": 0.001,
"loss": 0.7035,
"step": 15345
},
{
"epoch": 0.9656581532416503,
"grad_norm": 0.5234375,
"learning_rate": 0.001,
"loss": 0.7016,
"step": 15360
},
{
"epoch": 0.9666011787819253,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.6954,
"step": 15375
},
{
"epoch": 0.9675442043222003,
"grad_norm": 0.5859375,
"learning_rate": 0.001,
"loss": 0.7014,
"step": 15390
},
{
"epoch": 0.9684872298624755,
"grad_norm": 0.59375,
"learning_rate": 0.001,
"loss": 0.7129,
"step": 15405
},
{
"epoch": 0.9694302554027505,
"grad_norm": 0.515625,
"learning_rate": 0.001,
"loss": 0.6999,
"step": 15420
},
{
"epoch": 0.9703732809430256,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 0.7017,
"step": 15435
},
{
"epoch": 0.9713163064833006,
"grad_norm": 0.546875,
"learning_rate": 0.001,
"loss": 0.6893,
"step": 15450
},
{
"epoch": 0.9722593320235756,
"grad_norm": 0.71484375,
"learning_rate": 0.001,
"loss": 0.6993,
"step": 15465
},
{
"epoch": 0.9732023575638507,
"grad_norm": 0.6171875,
"learning_rate": 0.001,
"loss": 0.6999,
"step": 15480
},
{
"epoch": 0.9741453831041257,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.6864,
"step": 15495
},
{
"epoch": 0.9750884086444008,
"grad_norm": 0.49609375,
"learning_rate": 0.001,
"loss": 0.7057,
"step": 15510
},
{
"epoch": 0.9760314341846759,
"grad_norm": 0.58984375,
"learning_rate": 0.001,
"loss": 0.6957,
"step": 15525
},
{
"epoch": 0.9769744597249509,
"grad_norm": 0.53125,
"learning_rate": 0.001,
"loss": 0.709,
"step": 15540
},
{
"epoch": 0.9779174852652259,
"grad_norm": 0.482421875,
"learning_rate": 0.001,
"loss": 0.6965,
"step": 15555
},
{
"epoch": 0.978860510805501,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.6989,
"step": 15570
},
{
"epoch": 0.979803536345776,
"grad_norm": 0.466796875,
"learning_rate": 0.001,
"loss": 0.6995,
"step": 15585
},
{
"epoch": 0.980746561886051,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.6894,
"step": 15600
},
{
"epoch": 0.9816895874263262,
"grad_norm": 0.5390625,
"learning_rate": 0.001,
"loss": 0.7084,
"step": 15615
},
{
"epoch": 0.9826326129666012,
"grad_norm": 0.58984375,
"learning_rate": 0.001,
"loss": 0.7021,
"step": 15630
},
{
"epoch": 0.9835756385068762,
"grad_norm": 0.87109375,
"learning_rate": 0.001,
"loss": 0.6892,
"step": 15645
},
{
"epoch": 0.9845186640471513,
"grad_norm": 0.62109375,
"learning_rate": 0.001,
"loss": 0.7147,
"step": 15660
},
{
"epoch": 0.9854616895874263,
"grad_norm": 0.54296875,
"learning_rate": 0.001,
"loss": 0.7007,
"step": 15675
},
{
"epoch": 0.9864047151277013,
"grad_norm": 0.5546875,
"learning_rate": 0.001,
"loss": 0.699,
"step": 15690
},
{
"epoch": 0.9873477406679764,
"grad_norm": 0.875,
"learning_rate": 0.001,
"loss": 0.6943,
"step": 15705
},
{
"epoch": 0.9882907662082515,
"grad_norm": 0.5546875,
"learning_rate": 0.001,
"loss": 0.6943,
"step": 15720
},
{
"epoch": 0.9892337917485265,
"grad_norm": 0.466796875,
"learning_rate": 0.001,
"loss": 0.703,
"step": 15735
},
{
"epoch": 0.9901768172888016,
"grad_norm": 0.57421875,
"learning_rate": 0.001,
"loss": 0.6953,
"step": 15750
},
{
"epoch": 0.9911198428290766,
"grad_norm": 0.71875,
"learning_rate": 0.001,
"loss": 0.6884,
"step": 15765
},
{
"epoch": 0.9920628683693516,
"grad_norm": 0.68359375,
"learning_rate": 0.001,
"loss": 0.6972,
"step": 15780
},
{
"epoch": 0.9930058939096267,
"grad_norm": 0.59375,
"learning_rate": 0.001,
"loss": 0.6929,
"step": 15795
},
{
"epoch": 0.9939489194499017,
"grad_norm": 0.47265625,
"learning_rate": 0.001,
"loss": 0.6849,
"step": 15810
},
{
"epoch": 0.9948919449901769,
"grad_norm": 0.57421875,
"learning_rate": 0.001,
"loss": 0.6932,
"step": 15825
},
{
"epoch": 0.9958349705304519,
"grad_norm": 0.51953125,
"learning_rate": 0.001,
"loss": 0.7042,
"step": 15840
},
{
"epoch": 0.9967779960707269,
"grad_norm": 0.470703125,
"learning_rate": 0.001,
"loss": 0.6924,
"step": 15855
},
{
"epoch": 0.997721021611002,
"grad_norm": 0.578125,
"learning_rate": 0.001,
"loss": 0.7009,
"step": 15870
},
{
"epoch": 0.998664047151277,
"grad_norm": 0.55078125,
"learning_rate": 0.001,
"loss": 0.7059,
"step": 15885
},
{
"epoch": 0.999607072691552,
"grad_norm": 0.53515625,
"learning_rate": 0.001,
"loss": 0.691,
"step": 15900
},
{
"epoch": 0.999607072691552,
"eval_loss": 0.8118711709976196,
"eval_runtime": 9.6839,
"eval_samples_per_second": 103.264,
"eval_steps_per_second": 1.446,
"step": 15900
}
],
"logging_steps": 15,
"max_steps": 15906,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 1590,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 4.185992916964999e+18,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}