|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 40, |
|
"global_step": 198, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.005050505050505051, |
|
"grad_norm": 1.934841513633728, |
|
"learning_rate": 2e-05, |
|
"loss": 2.149, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.010101010101010102, |
|
"grad_norm": 2.4779138565063477, |
|
"learning_rate": 4e-05, |
|
"loss": 2.7234, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.015151515151515152, |
|
"grad_norm": 3.3968937397003174, |
|
"learning_rate": 6e-05, |
|
"loss": 2.9434, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.020202020202020204, |
|
"grad_norm": 2.6924426555633545, |
|
"learning_rate": 8e-05, |
|
"loss": 2.9018, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.025252525252525252, |
|
"grad_norm": 2.9207985401153564, |
|
"learning_rate": 0.0001, |
|
"loss": 2.8698, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.030303030303030304, |
|
"grad_norm": 2.5849769115448, |
|
"learning_rate": 0.00012, |
|
"loss": 2.6004, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.03535353535353535, |
|
"grad_norm": 2.265801191329956, |
|
"learning_rate": 0.00014, |
|
"loss": 1.9435, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.04040404040404041, |
|
"grad_norm": 2.9821760654449463, |
|
"learning_rate": 0.00016, |
|
"loss": 2.3524, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.045454545454545456, |
|
"grad_norm": 2.209989547729492, |
|
"learning_rate": 0.00018, |
|
"loss": 1.7031, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.050505050505050504, |
|
"grad_norm": 2.4582831859588623, |
|
"learning_rate": 0.0002, |
|
"loss": 1.858, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.05555555555555555, |
|
"grad_norm": 2.465235471725464, |
|
"learning_rate": 0.00019893617021276595, |
|
"loss": 1.9792, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.06060606060606061, |
|
"grad_norm": 2.106245756149292, |
|
"learning_rate": 0.00019787234042553193, |
|
"loss": 1.6035, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.06565656565656566, |
|
"grad_norm": 2.0670032501220703, |
|
"learning_rate": 0.00019680851063829787, |
|
"loss": 1.4089, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0707070707070707, |
|
"grad_norm": 1.869535207748413, |
|
"learning_rate": 0.00019574468085106384, |
|
"loss": 1.5282, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.07575757575757576, |
|
"grad_norm": 2.4417788982391357, |
|
"learning_rate": 0.00019468085106382982, |
|
"loss": 2.0049, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.08080808080808081, |
|
"grad_norm": 2.03873610496521, |
|
"learning_rate": 0.00019361702127659576, |
|
"loss": 1.5979, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.08585858585858586, |
|
"grad_norm": 2.6959407329559326, |
|
"learning_rate": 0.0001925531914893617, |
|
"loss": 1.7192, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.09090909090909091, |
|
"grad_norm": 2.447222948074341, |
|
"learning_rate": 0.00019148936170212768, |
|
"loss": 1.9945, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.09595959595959595, |
|
"grad_norm": 2.0988619327545166, |
|
"learning_rate": 0.00019042553191489362, |
|
"loss": 1.5905, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.10101010101010101, |
|
"grad_norm": 2.3013782501220703, |
|
"learning_rate": 0.00018936170212765957, |
|
"loss": 1.5003, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.10606060606060606, |
|
"grad_norm": 2.080165147781372, |
|
"learning_rate": 0.00018829787234042554, |
|
"loss": 1.4736, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.1111111111111111, |
|
"grad_norm": 4.3709940910339355, |
|
"learning_rate": 0.0001872340425531915, |
|
"loss": 0.9107, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.11616161616161616, |
|
"grad_norm": 5.782566070556641, |
|
"learning_rate": 0.00018617021276595746, |
|
"loss": 1.3649, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.12121212121212122, |
|
"grad_norm": 2.43636417388916, |
|
"learning_rate": 0.0001851063829787234, |
|
"loss": 1.2051, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.12626262626262627, |
|
"grad_norm": 2.063096761703491, |
|
"learning_rate": 0.00018404255319148937, |
|
"loss": 1.1806, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.13131313131313133, |
|
"grad_norm": 2.052663564682007, |
|
"learning_rate": 0.00018297872340425532, |
|
"loss": 1.1073, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.13636363636363635, |
|
"grad_norm": 2.293409585952759, |
|
"learning_rate": 0.0001819148936170213, |
|
"loss": 1.3441, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.1414141414141414, |
|
"grad_norm": 2.016857147216797, |
|
"learning_rate": 0.00018085106382978726, |
|
"loss": 1.1819, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.14646464646464646, |
|
"grad_norm": 2.0556704998016357, |
|
"learning_rate": 0.0001797872340425532, |
|
"loss": 1.2543, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.15151515151515152, |
|
"grad_norm": 2.157916784286499, |
|
"learning_rate": 0.00017872340425531915, |
|
"loss": 1.1889, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.15656565656565657, |
|
"grad_norm": 2.284308671951294, |
|
"learning_rate": 0.00017765957446808512, |
|
"loss": 1.3205, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.16161616161616163, |
|
"grad_norm": 2.55147123336792, |
|
"learning_rate": 0.00017659574468085107, |
|
"loss": 1.1276, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.16666666666666666, |
|
"grad_norm": 2.595245599746704, |
|
"learning_rate": 0.000175531914893617, |
|
"loss": 1.2703, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.1717171717171717, |
|
"grad_norm": 2.480912685394287, |
|
"learning_rate": 0.00017446808510638298, |
|
"loss": 0.9982, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.17676767676767677, |
|
"grad_norm": 2.379316806793213, |
|
"learning_rate": 0.00017340425531914896, |
|
"loss": 1.1423, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.18181818181818182, |
|
"grad_norm": 2.62664794921875, |
|
"learning_rate": 0.0001723404255319149, |
|
"loss": 1.1058, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.18686868686868688, |
|
"grad_norm": 3.0421292781829834, |
|
"learning_rate": 0.00017127659574468087, |
|
"loss": 1.3554, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.1919191919191919, |
|
"grad_norm": 2.338921308517456, |
|
"learning_rate": 0.00017021276595744682, |
|
"loss": 1.0576, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.19696969696969696, |
|
"grad_norm": 2.2380709648132324, |
|
"learning_rate": 0.00016914893617021276, |
|
"loss": 0.96, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.20202020202020202, |
|
"grad_norm": 2.5871872901916504, |
|
"learning_rate": 0.00016808510638297873, |
|
"loss": 1.1717, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.20202020202020202, |
|
"eval_loss": 1.2117999792099, |
|
"eval_runtime": 35.7116, |
|
"eval_samples_per_second": 2.8, |
|
"eval_steps_per_second": 2.8, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.20707070707070707, |
|
"grad_norm": 3.1568801403045654, |
|
"learning_rate": 0.00016702127659574468, |
|
"loss": 1.2266, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.21212121212121213, |
|
"grad_norm": 2.2659265995025635, |
|
"learning_rate": 0.00016595744680851065, |
|
"loss": 0.9661, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.21717171717171718, |
|
"grad_norm": 2.648207664489746, |
|
"learning_rate": 0.00016489361702127662, |
|
"loss": 1.4721, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.2222222222222222, |
|
"grad_norm": 2.391132116317749, |
|
"learning_rate": 0.00016382978723404257, |
|
"loss": 1.1877, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.22727272727272727, |
|
"grad_norm": 2.216590166091919, |
|
"learning_rate": 0.0001627659574468085, |
|
"loss": 1.1156, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.23232323232323232, |
|
"grad_norm": 2.7304837703704834, |
|
"learning_rate": 0.00016170212765957446, |
|
"loss": 1.0745, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.23737373737373738, |
|
"grad_norm": 2.449009418487549, |
|
"learning_rate": 0.00016063829787234043, |
|
"loss": 0.8261, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.24242424242424243, |
|
"grad_norm": 2.8799259662628174, |
|
"learning_rate": 0.00015957446808510637, |
|
"loss": 1.3139, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.2474747474747475, |
|
"grad_norm": 3.0433382987976074, |
|
"learning_rate": 0.00015851063829787235, |
|
"loss": 0.8179, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.25252525252525254, |
|
"grad_norm": 1.923201322555542, |
|
"learning_rate": 0.00015744680851063832, |
|
"loss": 1.3833, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.25757575757575757, |
|
"grad_norm": 2.040708303451538, |
|
"learning_rate": 0.00015638297872340426, |
|
"loss": 1.7501, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.26262626262626265, |
|
"grad_norm": 1.4778690338134766, |
|
"learning_rate": 0.0001553191489361702, |
|
"loss": 0.9487, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.2676767676767677, |
|
"grad_norm": 1.7248860597610474, |
|
"learning_rate": 0.00015425531914893618, |
|
"loss": 1.0579, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.2727272727272727, |
|
"grad_norm": 1.6652686595916748, |
|
"learning_rate": 0.00015319148936170213, |
|
"loss": 0.9751, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.2777777777777778, |
|
"grad_norm": 1.6558713912963867, |
|
"learning_rate": 0.0001521276595744681, |
|
"loss": 1.0705, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.2828282828282828, |
|
"grad_norm": 2.045135259628296, |
|
"learning_rate": 0.00015106382978723407, |
|
"loss": 1.4438, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.2878787878787879, |
|
"grad_norm": 2.0178723335266113, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 1.1453, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.29292929292929293, |
|
"grad_norm": 1.7096202373504639, |
|
"learning_rate": 0.00014893617021276596, |
|
"loss": 1.0412, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.29797979797979796, |
|
"grad_norm": 2.039254665374756, |
|
"learning_rate": 0.00014787234042553193, |
|
"loss": 1.3789, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.30303030303030304, |
|
"grad_norm": 2.7002921104431152, |
|
"learning_rate": 0.00014680851063829788, |
|
"loss": 1.9902, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.30808080808080807, |
|
"grad_norm": 2.1051278114318848, |
|
"learning_rate": 0.00014574468085106382, |
|
"loss": 1.2287, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.31313131313131315, |
|
"grad_norm": 2.6454594135284424, |
|
"learning_rate": 0.0001446808510638298, |
|
"loss": 1.3612, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.3181818181818182, |
|
"grad_norm": 1.9458284378051758, |
|
"learning_rate": 0.00014361702127659576, |
|
"loss": 1.1879, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.32323232323232326, |
|
"grad_norm": 2.09489369392395, |
|
"learning_rate": 0.0001425531914893617, |
|
"loss": 1.1563, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.3282828282828283, |
|
"grad_norm": 1.8626444339752197, |
|
"learning_rate": 0.00014148936170212768, |
|
"loss": 0.8783, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.3333333333333333, |
|
"grad_norm": 2.056874990463257, |
|
"learning_rate": 0.00014042553191489363, |
|
"loss": 1.1404, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.3383838383838384, |
|
"grad_norm": 2.202636241912842, |
|
"learning_rate": 0.00013936170212765957, |
|
"loss": 1.2132, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.3434343434343434, |
|
"grad_norm": 2.452498197555542, |
|
"learning_rate": 0.00013829787234042554, |
|
"loss": 1.451, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.3484848484848485, |
|
"grad_norm": 3.745300769805908, |
|
"learning_rate": 0.0001372340425531915, |
|
"loss": 1.4522, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.35353535353535354, |
|
"grad_norm": 2.425330638885498, |
|
"learning_rate": 0.00013617021276595746, |
|
"loss": 1.1973, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.35858585858585856, |
|
"grad_norm": 3.9445061683654785, |
|
"learning_rate": 0.0001351063829787234, |
|
"loss": 0.6733, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.36363636363636365, |
|
"grad_norm": 1.9044334888458252, |
|
"learning_rate": 0.00013404255319148938, |
|
"loss": 0.9999, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.3686868686868687, |
|
"grad_norm": 2.4740421772003174, |
|
"learning_rate": 0.00013297872340425532, |
|
"loss": 1.2382, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.37373737373737376, |
|
"grad_norm": 2.201266288757324, |
|
"learning_rate": 0.00013191489361702127, |
|
"loss": 0.9838, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.3787878787878788, |
|
"grad_norm": 1.951281189918518, |
|
"learning_rate": 0.00013085106382978724, |
|
"loss": 0.888, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.3838383838383838, |
|
"grad_norm": 1.6007189750671387, |
|
"learning_rate": 0.00012978723404255318, |
|
"loss": 0.7822, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.3888888888888889, |
|
"grad_norm": 2.144460916519165, |
|
"learning_rate": 0.00012872340425531915, |
|
"loss": 0.9619, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.3939393939393939, |
|
"grad_norm": 2.2907423973083496, |
|
"learning_rate": 0.00012765957446808513, |
|
"loss": 1.1213, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.398989898989899, |
|
"grad_norm": 2.3168041706085205, |
|
"learning_rate": 0.00012659574468085107, |
|
"loss": 1.0892, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.40404040404040403, |
|
"grad_norm": 2.5245144367218018, |
|
"learning_rate": 0.00012553191489361702, |
|
"loss": 1.1852, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.40404040404040403, |
|
"eval_loss": 1.081834077835083, |
|
"eval_runtime": 35.7648, |
|
"eval_samples_per_second": 2.796, |
|
"eval_steps_per_second": 2.796, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.4090909090909091, |
|
"grad_norm": 3.342719316482544, |
|
"learning_rate": 0.000124468085106383, |
|
"loss": 1.5945, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.41414141414141414, |
|
"grad_norm": 1.9120285511016846, |
|
"learning_rate": 0.00012340425531914893, |
|
"loss": 0.8123, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.41919191919191917, |
|
"grad_norm": 2.2346291542053223, |
|
"learning_rate": 0.0001223404255319149, |
|
"loss": 0.7541, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.42424242424242425, |
|
"grad_norm": 2.395176649093628, |
|
"learning_rate": 0.00012127659574468086, |
|
"loss": 1.2209, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.4292929292929293, |
|
"grad_norm": 2.275162696838379, |
|
"learning_rate": 0.00012021276595744682, |
|
"loss": 1.06, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.43434343434343436, |
|
"grad_norm": 2.27367901802063, |
|
"learning_rate": 0.00011914893617021277, |
|
"loss": 1.0375, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.4393939393939394, |
|
"grad_norm": 2.6139156818389893, |
|
"learning_rate": 0.00011808510638297874, |
|
"loss": 1.1906, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.4444444444444444, |
|
"grad_norm": 2.0727851390838623, |
|
"learning_rate": 0.00011702127659574468, |
|
"loss": 0.9182, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.4494949494949495, |
|
"grad_norm": 2.6906633377075195, |
|
"learning_rate": 0.00011595744680851064, |
|
"loss": 1.0941, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.45454545454545453, |
|
"grad_norm": 2.9968316555023193, |
|
"learning_rate": 0.00011489361702127661, |
|
"loss": 0.9502, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.4595959595959596, |
|
"grad_norm": 1.800660490989685, |
|
"learning_rate": 0.00011382978723404256, |
|
"loss": 0.7118, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.46464646464646464, |
|
"grad_norm": 1.7916479110717773, |
|
"learning_rate": 0.00011276595744680852, |
|
"loss": 0.5925, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.4696969696969697, |
|
"grad_norm": 2.894232988357544, |
|
"learning_rate": 0.00011170212765957446, |
|
"loss": 1.1137, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.47474747474747475, |
|
"grad_norm": 2.5062479972839355, |
|
"learning_rate": 0.00011063829787234043, |
|
"loss": 0.9837, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.4797979797979798, |
|
"grad_norm": 2.5845303535461426, |
|
"learning_rate": 0.00010957446808510638, |
|
"loss": 1.1791, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.48484848484848486, |
|
"grad_norm": 2.2351789474487305, |
|
"learning_rate": 0.00010851063829787234, |
|
"loss": 0.7262, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.4898989898989899, |
|
"grad_norm": 2.842940330505371, |
|
"learning_rate": 0.00010744680851063831, |
|
"loss": 1.115, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.494949494949495, |
|
"grad_norm": 2.390490770339966, |
|
"learning_rate": 0.00010638297872340425, |
|
"loss": 0.7784, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 1.5551937818527222, |
|
"learning_rate": 0.00010531914893617021, |
|
"loss": 1.323, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.5050505050505051, |
|
"grad_norm": 1.6265314817428589, |
|
"learning_rate": 0.00010425531914893618, |
|
"loss": 0.9718, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.51010101010101, |
|
"grad_norm": 1.7489873170852661, |
|
"learning_rate": 0.00010319148936170213, |
|
"loss": 1.1333, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.5151515151515151, |
|
"grad_norm": 1.857285737991333, |
|
"learning_rate": 0.00010212765957446809, |
|
"loss": 1.2764, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.5202020202020202, |
|
"grad_norm": 1.7422451972961426, |
|
"learning_rate": 0.00010106382978723406, |
|
"loss": 1.3637, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.5252525252525253, |
|
"grad_norm": 1.1181117296218872, |
|
"learning_rate": 0.0001, |
|
"loss": 0.8533, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.5303030303030303, |
|
"grad_norm": 1.7180590629577637, |
|
"learning_rate": 9.893617021276596e-05, |
|
"loss": 1.231, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.5353535353535354, |
|
"grad_norm": 1.6332601308822632, |
|
"learning_rate": 9.787234042553192e-05, |
|
"loss": 1.1621, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.5404040404040404, |
|
"grad_norm": 2.0528311729431152, |
|
"learning_rate": 9.680851063829788e-05, |
|
"loss": 1.3896, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.5454545454545454, |
|
"grad_norm": 1.4702327251434326, |
|
"learning_rate": 9.574468085106384e-05, |
|
"loss": 0.9355, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.5505050505050505, |
|
"grad_norm": 1.9625461101531982, |
|
"learning_rate": 9.468085106382978e-05, |
|
"loss": 1.5606, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.5555555555555556, |
|
"grad_norm": 1.784745693206787, |
|
"learning_rate": 9.361702127659576e-05, |
|
"loss": 1.1102, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.5606060606060606, |
|
"grad_norm": 2.0343940258026123, |
|
"learning_rate": 9.25531914893617e-05, |
|
"loss": 1.2201, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.5656565656565656, |
|
"grad_norm": 1.4346604347229004, |
|
"learning_rate": 9.148936170212766e-05, |
|
"loss": 0.7823, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.5707070707070707, |
|
"grad_norm": 1.9893382787704468, |
|
"learning_rate": 9.042553191489363e-05, |
|
"loss": 1.0692, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.5757575757575758, |
|
"grad_norm": 2.1041550636291504, |
|
"learning_rate": 8.936170212765958e-05, |
|
"loss": 1.4054, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.5808080808080808, |
|
"grad_norm": 2.3176534175872803, |
|
"learning_rate": 8.829787234042553e-05, |
|
"loss": 1.4864, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.5858585858585859, |
|
"grad_norm": 1.7559845447540283, |
|
"learning_rate": 8.723404255319149e-05, |
|
"loss": 0.8039, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.5909090909090909, |
|
"grad_norm": 1.7945481538772583, |
|
"learning_rate": 8.617021276595745e-05, |
|
"loss": 0.9355, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.5959595959595959, |
|
"grad_norm": 1.5018421411514282, |
|
"learning_rate": 8.510638297872341e-05, |
|
"loss": 0.8087, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.601010101010101, |
|
"grad_norm": 1.6970769166946411, |
|
"learning_rate": 8.404255319148937e-05, |
|
"loss": 0.8562, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.6060606060606061, |
|
"grad_norm": 2.8440420627593994, |
|
"learning_rate": 8.297872340425533e-05, |
|
"loss": 1.2596, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.6060606060606061, |
|
"eval_loss": 1.025399088859558, |
|
"eval_runtime": 35.717, |
|
"eval_samples_per_second": 2.8, |
|
"eval_steps_per_second": 2.8, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.6111111111111112, |
|
"grad_norm": 2.990412712097168, |
|
"learning_rate": 8.191489361702128e-05, |
|
"loss": 1.1597, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.6161616161616161, |
|
"grad_norm": 1.3564014434814453, |
|
"learning_rate": 8.085106382978723e-05, |
|
"loss": 0.7068, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.6212121212121212, |
|
"grad_norm": 2.5022084712982178, |
|
"learning_rate": 7.978723404255319e-05, |
|
"loss": 1.6883, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.6262626262626263, |
|
"grad_norm": 1.8736674785614014, |
|
"learning_rate": 7.872340425531916e-05, |
|
"loss": 0.6924, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.6313131313131313, |
|
"grad_norm": 2.2267813682556152, |
|
"learning_rate": 7.76595744680851e-05, |
|
"loss": 1.1193, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.6363636363636364, |
|
"grad_norm": 2.134420156478882, |
|
"learning_rate": 7.659574468085106e-05, |
|
"loss": 0.9352, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.6414141414141414, |
|
"grad_norm": 1.9499320983886719, |
|
"learning_rate": 7.553191489361703e-05, |
|
"loss": 0.8882, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.6464646464646465, |
|
"grad_norm": 2.1963298320770264, |
|
"learning_rate": 7.446808510638298e-05, |
|
"loss": 0.8756, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.6515151515151515, |
|
"grad_norm": 2.303020715713501, |
|
"learning_rate": 7.340425531914894e-05, |
|
"loss": 0.903, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.6565656565656566, |
|
"grad_norm": 2.4283013343811035, |
|
"learning_rate": 7.23404255319149e-05, |
|
"loss": 1.2209, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.6616161616161617, |
|
"grad_norm": 1.8416250944137573, |
|
"learning_rate": 7.127659574468085e-05, |
|
"loss": 0.8485, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.6666666666666666, |
|
"grad_norm": 1.958123803138733, |
|
"learning_rate": 7.021276595744681e-05, |
|
"loss": 0.955, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.6717171717171717, |
|
"grad_norm": 2.2823963165283203, |
|
"learning_rate": 6.914893617021277e-05, |
|
"loss": 0.9205, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.6767676767676768, |
|
"grad_norm": 2.2370712757110596, |
|
"learning_rate": 6.808510638297873e-05, |
|
"loss": 0.7799, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.6818181818181818, |
|
"grad_norm": 2.6467502117156982, |
|
"learning_rate": 6.702127659574469e-05, |
|
"loss": 1.0801, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.6868686868686869, |
|
"grad_norm": 2.9221208095550537, |
|
"learning_rate": 6.595744680851063e-05, |
|
"loss": 1.2855, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.6919191919191919, |
|
"grad_norm": 2.024229049682617, |
|
"learning_rate": 6.489361702127659e-05, |
|
"loss": 0.7463, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.696969696969697, |
|
"grad_norm": 2.125941276550293, |
|
"learning_rate": 6.382978723404256e-05, |
|
"loss": 0.8951, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.702020202020202, |
|
"grad_norm": 2.9005520343780518, |
|
"learning_rate": 6.276595744680851e-05, |
|
"loss": 0.7212, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.7070707070707071, |
|
"grad_norm": 2.088212251663208, |
|
"learning_rate": 6.170212765957447e-05, |
|
"loss": 0.6987, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.7121212121212122, |
|
"grad_norm": 2.779832124710083, |
|
"learning_rate": 6.063829787234043e-05, |
|
"loss": 1.2843, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.7171717171717171, |
|
"grad_norm": 1.8998260498046875, |
|
"learning_rate": 5.9574468085106384e-05, |
|
"loss": 0.7472, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.7222222222222222, |
|
"grad_norm": 2.266793727874756, |
|
"learning_rate": 5.851063829787234e-05, |
|
"loss": 0.7037, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.7272727272727273, |
|
"grad_norm": 2.3124566078186035, |
|
"learning_rate": 5.744680851063831e-05, |
|
"loss": 1.0372, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.7323232323232324, |
|
"grad_norm": 3.8273227214813232, |
|
"learning_rate": 5.638297872340426e-05, |
|
"loss": 1.1066, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.7373737373737373, |
|
"grad_norm": 1.7611407041549683, |
|
"learning_rate": 5.531914893617022e-05, |
|
"loss": 0.5528, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.7424242424242424, |
|
"grad_norm": 1.3345385789871216, |
|
"learning_rate": 5.425531914893617e-05, |
|
"loss": 0.533, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.7474747474747475, |
|
"grad_norm": 1.5304288864135742, |
|
"learning_rate": 5.319148936170213e-05, |
|
"loss": 1.1614, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.7525252525252525, |
|
"grad_norm": 2.2307026386260986, |
|
"learning_rate": 5.212765957446809e-05, |
|
"loss": 1.7027, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.7575757575757576, |
|
"grad_norm": 2.0783305168151855, |
|
"learning_rate": 5.1063829787234044e-05, |
|
"loss": 1.5548, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.7626262626262627, |
|
"grad_norm": 1.9879854917526245, |
|
"learning_rate": 5e-05, |
|
"loss": 1.3791, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.7676767676767676, |
|
"grad_norm": 1.6240109205245972, |
|
"learning_rate": 4.893617021276596e-05, |
|
"loss": 1.0744, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.7727272727272727, |
|
"grad_norm": 2.711182117462158, |
|
"learning_rate": 4.787234042553192e-05, |
|
"loss": 1.5953, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.7777777777777778, |
|
"grad_norm": 1.7542715072631836, |
|
"learning_rate": 4.680851063829788e-05, |
|
"loss": 0.8801, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.7828282828282829, |
|
"grad_norm": 2.7338082790374756, |
|
"learning_rate": 4.574468085106383e-05, |
|
"loss": 2.0285, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.7878787878787878, |
|
"grad_norm": 1.4537471532821655, |
|
"learning_rate": 4.468085106382979e-05, |
|
"loss": 0.7417, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.7929292929292929, |
|
"grad_norm": 1.9007015228271484, |
|
"learning_rate": 4.3617021276595746e-05, |
|
"loss": 1.1665, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.797979797979798, |
|
"grad_norm": 1.6936357021331787, |
|
"learning_rate": 4.2553191489361704e-05, |
|
"loss": 0.9482, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.803030303030303, |
|
"grad_norm": 1.923112392425537, |
|
"learning_rate": 4.148936170212766e-05, |
|
"loss": 1.1103, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.8080808080808081, |
|
"grad_norm": 1.7348624467849731, |
|
"learning_rate": 4.0425531914893614e-05, |
|
"loss": 0.785, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.8080808080808081, |
|
"eval_loss": 1.0044422149658203, |
|
"eval_runtime": 35.7191, |
|
"eval_samples_per_second": 2.8, |
|
"eval_steps_per_second": 2.8, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.8131313131313131, |
|
"grad_norm": 2.3853325843811035, |
|
"learning_rate": 3.936170212765958e-05, |
|
"loss": 1.2953, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.8181818181818182, |
|
"grad_norm": 1.5539531707763672, |
|
"learning_rate": 3.829787234042553e-05, |
|
"loss": 0.7532, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.8232323232323232, |
|
"grad_norm": 2.0018396377563477, |
|
"learning_rate": 3.723404255319149e-05, |
|
"loss": 1.0277, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.8282828282828283, |
|
"grad_norm": 1.8975419998168945, |
|
"learning_rate": 3.617021276595745e-05, |
|
"loss": 0.84, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.8333333333333334, |
|
"grad_norm": 2.150519847869873, |
|
"learning_rate": 3.5106382978723407e-05, |
|
"loss": 0.9874, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.8383838383838383, |
|
"grad_norm": 2.172685384750366, |
|
"learning_rate": 3.4042553191489365e-05, |
|
"loss": 1.1271, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.8434343434343434, |
|
"grad_norm": 2.1147804260253906, |
|
"learning_rate": 3.2978723404255317e-05, |
|
"loss": 0.8757, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.8484848484848485, |
|
"grad_norm": 2.024099588394165, |
|
"learning_rate": 3.191489361702128e-05, |
|
"loss": 1.2143, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.8535353535353535, |
|
"grad_norm": 2.0407485961914062, |
|
"learning_rate": 3.085106382978723e-05, |
|
"loss": 1.0318, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.8585858585858586, |
|
"grad_norm": 2.056114435195923, |
|
"learning_rate": 2.9787234042553192e-05, |
|
"loss": 1.0286, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.8636363636363636, |
|
"grad_norm": 2.441153049468994, |
|
"learning_rate": 2.8723404255319154e-05, |
|
"loss": 1.3128, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.8686868686868687, |
|
"grad_norm": 2.1208486557006836, |
|
"learning_rate": 2.765957446808511e-05, |
|
"loss": 0.7569, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.8737373737373737, |
|
"grad_norm": 2.4818294048309326, |
|
"learning_rate": 2.6595744680851064e-05, |
|
"loss": 1.5236, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.8787878787878788, |
|
"grad_norm": 1.9016083478927612, |
|
"learning_rate": 2.5531914893617022e-05, |
|
"loss": 0.7904, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.8838383838383839, |
|
"grad_norm": 2.479365348815918, |
|
"learning_rate": 2.446808510638298e-05, |
|
"loss": 1.3812, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.8888888888888888, |
|
"grad_norm": 2.3161561489105225, |
|
"learning_rate": 2.340425531914894e-05, |
|
"loss": 1.1612, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.8939393939393939, |
|
"grad_norm": 1.7003121376037598, |
|
"learning_rate": 2.2340425531914894e-05, |
|
"loss": 0.7709, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.898989898989899, |
|
"grad_norm": 2.0750014781951904, |
|
"learning_rate": 2.1276595744680852e-05, |
|
"loss": 1.0438, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.9040404040404041, |
|
"grad_norm": 1.7369530200958252, |
|
"learning_rate": 2.0212765957446807e-05, |
|
"loss": 0.711, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.9090909090909091, |
|
"grad_norm": 1.7840030193328857, |
|
"learning_rate": 1.9148936170212766e-05, |
|
"loss": 0.7699, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.9141414141414141, |
|
"grad_norm": 1.9962245225906372, |
|
"learning_rate": 1.8085106382978724e-05, |
|
"loss": 1.0592, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.9191919191919192, |
|
"grad_norm": 1.862483024597168, |
|
"learning_rate": 1.7021276595744682e-05, |
|
"loss": 0.855, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.9242424242424242, |
|
"grad_norm": 1.5625897645950317, |
|
"learning_rate": 1.595744680851064e-05, |
|
"loss": 0.6401, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.9292929292929293, |
|
"grad_norm": 2.1147549152374268, |
|
"learning_rate": 1.4893617021276596e-05, |
|
"loss": 0.9002, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.9343434343434344, |
|
"grad_norm": 1.6960692405700684, |
|
"learning_rate": 1.3829787234042554e-05, |
|
"loss": 0.7592, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.9393939393939394, |
|
"grad_norm": 2.3449957370758057, |
|
"learning_rate": 1.2765957446808511e-05, |
|
"loss": 1.1498, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.9444444444444444, |
|
"grad_norm": 1.868044376373291, |
|
"learning_rate": 1.170212765957447e-05, |
|
"loss": 0.8678, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.9494949494949495, |
|
"grad_norm": 2.05068039894104, |
|
"learning_rate": 1.0638297872340426e-05, |
|
"loss": 0.7825, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.9545454545454546, |
|
"grad_norm": 1.92442786693573, |
|
"learning_rate": 9.574468085106383e-06, |
|
"loss": 0.7249, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.9595959595959596, |
|
"grad_norm": 2.1448285579681396, |
|
"learning_rate": 8.510638297872341e-06, |
|
"loss": 0.9544, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.9646464646464646, |
|
"grad_norm": 1.7718862295150757, |
|
"learning_rate": 7.446808510638298e-06, |
|
"loss": 0.5564, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.9696969696969697, |
|
"grad_norm": 2.5203495025634766, |
|
"learning_rate": 6.3829787234042555e-06, |
|
"loss": 0.9661, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.9747474747474747, |
|
"grad_norm": 2.2129266262054443, |
|
"learning_rate": 5.319148936170213e-06, |
|
"loss": 0.9277, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.9797979797979798, |
|
"grad_norm": 2.0742154121398926, |
|
"learning_rate": 4.255319148936171e-06, |
|
"loss": 0.8203, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.9848484848484849, |
|
"grad_norm": 1.7061251401901245, |
|
"learning_rate": 3.1914893617021277e-06, |
|
"loss": 0.4595, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.98989898989899, |
|
"grad_norm": 1.7450275421142578, |
|
"learning_rate": 2.1276595744680853e-06, |
|
"loss": 0.5028, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.9949494949494949, |
|
"grad_norm": 4.231220245361328, |
|
"learning_rate": 1.0638297872340427e-06, |
|
"loss": 1.0582, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 2.37825870513916, |
|
"learning_rate": 0.0, |
|
"loss": 0.8417, |
|
"step": 198 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 198, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2341997358425088.0, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|