|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0018214936247722, |
|
"eval_steps": 500, |
|
"global_step": 550, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 0.46431864432643544, |
|
"learning_rate": 1.2121212121212122e-06, |
|
"loss": 1.4151, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 0.42969176658539837, |
|
"learning_rate": 2.4242424242424244e-06, |
|
"loss": 1.3729, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.5004307936270223, |
|
"learning_rate": 3.636363636363636e-06, |
|
"loss": 1.3989, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.43666634920041486, |
|
"learning_rate": 4.848484848484849e-06, |
|
"loss": 1.3363, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.4691114419353825, |
|
"learning_rate": 6.060606060606061e-06, |
|
"loss": 1.4293, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.4277596061377729, |
|
"learning_rate": 7.272727272727272e-06, |
|
"loss": 1.4343, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.4238339229382504, |
|
"learning_rate": 8.484848484848486e-06, |
|
"loss": 1.4462, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.40992048534183273, |
|
"learning_rate": 9.696969696969698e-06, |
|
"loss": 1.2756, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.37885700313540693, |
|
"learning_rate": 1.0909090909090909e-05, |
|
"loss": 1.3464, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.36066141170123023, |
|
"learning_rate": 1.2121212121212122e-05, |
|
"loss": 1.3419, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.35617169386863406, |
|
"learning_rate": 1.3333333333333333e-05, |
|
"loss": 1.3533, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.3040188564782602, |
|
"learning_rate": 1.4545454545454545e-05, |
|
"loss": 1.2395, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.31038319439216566, |
|
"learning_rate": 1.5757575757575756e-05, |
|
"loss": 1.3082, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.26683768372135835, |
|
"learning_rate": 1.6969696969696972e-05, |
|
"loss": 1.3063, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.3652323682563078, |
|
"learning_rate": 1.8181818181818182e-05, |
|
"loss": 1.3045, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.23559121485457843, |
|
"learning_rate": 1.9393939393939395e-05, |
|
"loss": 1.2366, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.2342299313020104, |
|
"learning_rate": 2.0606060606060608e-05, |
|
"loss": 1.2831, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.2202931700357255, |
|
"learning_rate": 2.1818181818181818e-05, |
|
"loss": 1.3064, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.2097660599292375, |
|
"learning_rate": 2.3030303030303034e-05, |
|
"loss": 1.2376, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.2356785314652122, |
|
"learning_rate": 2.4242424242424244e-05, |
|
"loss": 1.2802, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.24639302530564244, |
|
"learning_rate": 2.5454545454545454e-05, |
|
"loss": 1.3016, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.24373126133228787, |
|
"learning_rate": 2.6666666666666667e-05, |
|
"loss": 1.3407, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.24488805144123432, |
|
"learning_rate": 2.7878787878787883e-05, |
|
"loss": 1.3325, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.2653033507571198, |
|
"learning_rate": 2.909090909090909e-05, |
|
"loss": 1.2811, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.2841724819336817, |
|
"learning_rate": 3.0303030303030306e-05, |
|
"loss": 1.2837, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.2183883020111492, |
|
"learning_rate": 3.151515151515151e-05, |
|
"loss": 1.2472, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.2137995163762026, |
|
"learning_rate": 3.272727272727273e-05, |
|
"loss": 1.2854, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.19499006223503876, |
|
"learning_rate": 3.3939393939393945e-05, |
|
"loss": 1.3018, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.17367919355340256, |
|
"learning_rate": 3.515151515151515e-05, |
|
"loss": 1.2824, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.18326045693683557, |
|
"learning_rate": 3.6363636363636364e-05, |
|
"loss": 1.2192, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.17474388188066411, |
|
"learning_rate": 3.757575757575758e-05, |
|
"loss": 1.2078, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.17856970178098716, |
|
"learning_rate": 3.878787878787879e-05, |
|
"loss": 1.2683, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.18617589704298348, |
|
"learning_rate": 4e-05, |
|
"loss": 1.2265, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.17653209733215317, |
|
"learning_rate": 4.1212121212121216e-05, |
|
"loss": 1.319, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.1722921367585233, |
|
"learning_rate": 4.242424242424243e-05, |
|
"loss": 1.2117, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.176642606378719, |
|
"learning_rate": 4.3636363636363636e-05, |
|
"loss": 1.2512, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.16696442324691066, |
|
"learning_rate": 4.484848484848485e-05, |
|
"loss": 1.2637, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.17035384059517106, |
|
"learning_rate": 4.606060606060607e-05, |
|
"loss": 1.2699, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.15545801881444482, |
|
"learning_rate": 4.7272727272727275e-05, |
|
"loss": 1.2939, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.17111439344347512, |
|
"learning_rate": 4.848484848484849e-05, |
|
"loss": 1.3033, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.16994151343455458, |
|
"learning_rate": 4.9696969696969694e-05, |
|
"loss": 1.2603, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.15929214926453447, |
|
"learning_rate": 5.090909090909091e-05, |
|
"loss": 1.2626, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.16761261516238699, |
|
"learning_rate": 5.212121212121213e-05, |
|
"loss": 1.296, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.15754700542426123, |
|
"learning_rate": 5.333333333333333e-05, |
|
"loss": 1.278, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.15522526683877644, |
|
"learning_rate": 5.4545454545454546e-05, |
|
"loss": 1.2355, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.1577929926930023, |
|
"learning_rate": 5.5757575757575766e-05, |
|
"loss": 1.2879, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.31075066632858317, |
|
"learning_rate": 5.696969696969697e-05, |
|
"loss": 1.2202, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.1663780653395111, |
|
"learning_rate": 5.818181818181818e-05, |
|
"loss": 1.2319, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.16049499655883026, |
|
"learning_rate": 5.93939393939394e-05, |
|
"loss": 1.2801, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.14515773124436285, |
|
"learning_rate": 6.060606060606061e-05, |
|
"loss": 1.2588, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.14653064850325623, |
|
"learning_rate": 6.181818181818182e-05, |
|
"loss": 1.2677, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.17193239746689878, |
|
"learning_rate": 6.303030303030302e-05, |
|
"loss": 1.2742, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.1967020450342533, |
|
"learning_rate": 6.424242424242424e-05, |
|
"loss": 1.1545, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.16247531997247225, |
|
"learning_rate": 6.545454545454546e-05, |
|
"loss": 1.222, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.14990706377244528, |
|
"learning_rate": 6.666666666666667e-05, |
|
"loss": 1.2103, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.1412817445239095, |
|
"learning_rate": 6.787878787878789e-05, |
|
"loss": 1.2169, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.14575971073482757, |
|
"learning_rate": 6.90909090909091e-05, |
|
"loss": 1.2751, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.13714747569950891, |
|
"learning_rate": 7.03030303030303e-05, |
|
"loss": 1.2508, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.14334695156859903, |
|
"learning_rate": 7.151515151515152e-05, |
|
"loss": 1.2721, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.1456824177522916, |
|
"learning_rate": 7.272727272727273e-05, |
|
"loss": 1.2649, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.15030318240210044, |
|
"learning_rate": 7.393939393939395e-05, |
|
"loss": 1.2167, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.1651326066719482, |
|
"learning_rate": 7.515151515151515e-05, |
|
"loss": 1.3126, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.1408250406479118, |
|
"learning_rate": 7.636363636363637e-05, |
|
"loss": 1.2891, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.21501384376905694, |
|
"learning_rate": 7.757575757575758e-05, |
|
"loss": 1.3019, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.1365168726167339, |
|
"learning_rate": 7.878787878787879e-05, |
|
"loss": 1.2498, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.1431463689660936, |
|
"learning_rate": 8e-05, |
|
"loss": 1.2793, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.13689045214286194, |
|
"learning_rate": 8.121212121212121e-05, |
|
"loss": 1.2295, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.13483608710081227, |
|
"learning_rate": 8.242424242424243e-05, |
|
"loss": 1.2258, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.13707618564415613, |
|
"learning_rate": 8.363636363636364e-05, |
|
"loss": 1.2252, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.13780236215967515, |
|
"learning_rate": 8.484848484848486e-05, |
|
"loss": 1.2565, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.14036805493494423, |
|
"learning_rate": 8.606060606060606e-05, |
|
"loss": 1.3023, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.12776919439147982, |
|
"learning_rate": 8.727272727272727e-05, |
|
"loss": 1.2292, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.1289941815481437, |
|
"learning_rate": 8.848484848484849e-05, |
|
"loss": 1.2191, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.13943952294847306, |
|
"learning_rate": 8.96969696969697e-05, |
|
"loss": 1.2915, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.1493528502117281, |
|
"learning_rate": 9.090909090909092e-05, |
|
"loss": 1.2797, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.1252401242451818, |
|
"learning_rate": 9.212121212121214e-05, |
|
"loss": 1.2552, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.13969800467546992, |
|
"learning_rate": 9.333333333333334e-05, |
|
"loss": 1.3147, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.1277258491470434, |
|
"learning_rate": 9.454545454545455e-05, |
|
"loss": 1.2089, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.133041369314817, |
|
"learning_rate": 9.575757575757576e-05, |
|
"loss": 1.2761, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.14564572037181842, |
|
"learning_rate": 9.696969696969698e-05, |
|
"loss": 1.1901, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.13666505656492195, |
|
"learning_rate": 9.818181818181818e-05, |
|
"loss": 1.2615, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.135007805210003, |
|
"learning_rate": 9.939393939393939e-05, |
|
"loss": 1.2669, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.17287563365884975, |
|
"learning_rate": 0.00010060606060606062, |
|
"loss": 1.2669, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.12934306326048103, |
|
"learning_rate": 0.00010181818181818181, |
|
"loss": 1.1979, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.13517436169178096, |
|
"learning_rate": 0.00010303030303030303, |
|
"loss": 1.2226, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.12105351159271568, |
|
"learning_rate": 0.00010424242424242425, |
|
"loss": 1.1172, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.1281676431775383, |
|
"learning_rate": 0.00010545454545454545, |
|
"loss": 1.2046, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.11730963057933333, |
|
"learning_rate": 0.00010666666666666667, |
|
"loss": 1.1883, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.12655235108503246, |
|
"learning_rate": 0.00010787878787878789, |
|
"loss": 1.1331, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.13047560307970027, |
|
"learning_rate": 0.00010909090909090909, |
|
"loss": 1.2731, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.12193522973752649, |
|
"learning_rate": 0.00011030303030303031, |
|
"loss": 1.2161, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.12804360300116346, |
|
"learning_rate": 0.00011151515151515153, |
|
"loss": 1.3062, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.15991741754516206, |
|
"learning_rate": 0.00011272727272727272, |
|
"loss": 1.239, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.15140182244454561, |
|
"learning_rate": 0.00011393939393939394, |
|
"loss": 1.2349, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.12320241076263434, |
|
"learning_rate": 0.00011515151515151516, |
|
"loss": 1.2875, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.13235998458230466, |
|
"learning_rate": 0.00011636363636363636, |
|
"loss": 1.2218, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.11783688734798668, |
|
"learning_rate": 0.00011757575757575758, |
|
"loss": 1.1864, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.3151933420750235, |
|
"learning_rate": 0.0001187878787878788, |
|
"loss": 1.3023, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.12665632567219295, |
|
"learning_rate": 0.00012, |
|
"loss": 1.2249, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.1228886740460738, |
|
"learning_rate": 0.00012121212121212122, |
|
"loss": 1.2517, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.11892005244989344, |
|
"learning_rate": 0.00012242424242424243, |
|
"loss": 1.2586, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.1232340827222201, |
|
"learning_rate": 0.00012363636363636364, |
|
"loss": 1.3217, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.13837226869323116, |
|
"learning_rate": 0.00012484848484848487, |
|
"loss": 1.2693, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.12068217991774362, |
|
"learning_rate": 0.00012606060606060605, |
|
"loss": 1.2623, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.16779277284606545, |
|
"learning_rate": 0.00012727272727272728, |
|
"loss": 1.2415, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.13396891539963085, |
|
"learning_rate": 0.0001284848484848485, |
|
"loss": 1.2313, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.12457104490772812, |
|
"learning_rate": 0.0001296969696969697, |
|
"loss": 1.1758, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.12676816816563452, |
|
"learning_rate": 0.00013090909090909093, |
|
"loss": 1.2478, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.11973639622066906, |
|
"learning_rate": 0.00013212121212121213, |
|
"loss": 1.2335, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.1330159646034068, |
|
"learning_rate": 0.00013333333333333334, |
|
"loss": 1.26, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.1298003025338099, |
|
"learning_rate": 0.00013454545454545455, |
|
"loss": 1.1907, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.1226154813287666, |
|
"learning_rate": 0.00013575757575757578, |
|
"loss": 1.1807, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.12533753244302145, |
|
"learning_rate": 0.00013696969696969696, |
|
"loss": 1.2098, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.12673266503840944, |
|
"learning_rate": 0.0001381818181818182, |
|
"loss": 1.2265, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.1299039569361384, |
|
"learning_rate": 0.0001393939393939394, |
|
"loss": 1.2534, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.13023496663090803, |
|
"learning_rate": 0.0001406060606060606, |
|
"loss": 1.2453, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.12001793500864573, |
|
"learning_rate": 0.00014181818181818184, |
|
"loss": 1.1608, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.14561862193041028, |
|
"learning_rate": 0.00014303030303030304, |
|
"loss": 1.2233, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.12636130876430832, |
|
"learning_rate": 0.00014424242424242425, |
|
"loss": 1.2833, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.189556849271166, |
|
"learning_rate": 0.00014545454545454546, |
|
"loss": 1.3105, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.12409073764495662, |
|
"learning_rate": 0.00014666666666666666, |
|
"loss": 1.1534, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.12149212466969316, |
|
"learning_rate": 0.0001478787878787879, |
|
"loss": 1.3039, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.12147336887953522, |
|
"learning_rate": 0.0001490909090909091, |
|
"loss": 1.326, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.1176585016163167, |
|
"learning_rate": 0.0001503030303030303, |
|
"loss": 1.191, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.2066428974234372, |
|
"learning_rate": 0.00015151515151515152, |
|
"loss": 1.3054, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.29582724255710047, |
|
"learning_rate": 0.00015272727272727275, |
|
"loss": 1.2032, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.13084381204119358, |
|
"learning_rate": 0.00015393939393939393, |
|
"loss": 1.2289, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.1294157600411397, |
|
"learning_rate": 0.00015515151515151516, |
|
"loss": 1.2561, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.14039614543447027, |
|
"learning_rate": 0.00015636363636363637, |
|
"loss": 1.243, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.19939984917282128, |
|
"learning_rate": 0.00015757575757575757, |
|
"loss": 1.1286, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.14402764349968203, |
|
"learning_rate": 0.0001587878787878788, |
|
"loss": 1.1959, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.13970978861500938, |
|
"learning_rate": 0.00016, |
|
"loss": 1.1814, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.14539538472563127, |
|
"learning_rate": 0.00016121212121212122, |
|
"loss": 1.2317, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.13456425455391557, |
|
"learning_rate": 0.00016242424242424243, |
|
"loss": 1.2239, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.1314997837157779, |
|
"learning_rate": 0.00016363636363636366, |
|
"loss": 1.1986, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.14046946525591422, |
|
"learning_rate": 0.00016484848484848487, |
|
"loss": 1.2238, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.6095538041505763, |
|
"learning_rate": 0.00016606060606060607, |
|
"loss": 1.2332, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.17707289712054367, |
|
"learning_rate": 0.00016727272727272728, |
|
"loss": 1.2401, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.19335172179099247, |
|
"learning_rate": 0.00016848484848484848, |
|
"loss": 1.2361, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.13725591818701255, |
|
"learning_rate": 0.00016969696969696972, |
|
"loss": 1.193, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.15535575462507384, |
|
"learning_rate": 0.0001709090909090909, |
|
"loss": 1.2769, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.14909436560898923, |
|
"learning_rate": 0.00017212121212121213, |
|
"loss": 1.2602, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.15054368082407957, |
|
"learning_rate": 0.00017333333333333334, |
|
"loss": 1.2607, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.13386897838741724, |
|
"learning_rate": 0.00017454545454545454, |
|
"loss": 1.168, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.13567889528730145, |
|
"learning_rate": 0.00017575757575757578, |
|
"loss": 1.1984, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.13994382298003089, |
|
"learning_rate": 0.00017696969696969698, |
|
"loss": 1.2795, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.13941573210713187, |
|
"learning_rate": 0.0001781818181818182, |
|
"loss": 1.2303, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.18302605925485763, |
|
"learning_rate": 0.0001793939393939394, |
|
"loss": 1.2696, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.1547402223275396, |
|
"learning_rate": 0.00018060606060606063, |
|
"loss": 1.1276, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.19947594494850646, |
|
"learning_rate": 0.00018181818181818183, |
|
"loss": 1.271, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.1517101450465788, |
|
"learning_rate": 0.00018303030303030304, |
|
"loss": 1.2193, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.19251063116857103, |
|
"learning_rate": 0.00018424242424242427, |
|
"loss": 1.2703, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.16789099560498666, |
|
"learning_rate": 0.00018545454545454545, |
|
"loss": 1.2244, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.14907376557922342, |
|
"learning_rate": 0.0001866666666666667, |
|
"loss": 1.264, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.14276598263036905, |
|
"learning_rate": 0.0001878787878787879, |
|
"loss": 1.2545, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.14526753816999002, |
|
"learning_rate": 0.0001890909090909091, |
|
"loss": 1.2912, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.1627048894660859, |
|
"learning_rate": 0.0001903030303030303, |
|
"loss": 1.2573, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.16405036632332695, |
|
"learning_rate": 0.0001915151515151515, |
|
"loss": 1.2359, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.14533427219788658, |
|
"learning_rate": 0.00019272727272727274, |
|
"loss": 1.1718, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.13802382666732702, |
|
"learning_rate": 0.00019393939393939395, |
|
"loss": 1.2297, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.15620193618511755, |
|
"learning_rate": 0.00019515151515151516, |
|
"loss": 1.2287, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.1401696295700075, |
|
"learning_rate": 0.00019636363636363636, |
|
"loss": 1.2231, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.15816133304035035, |
|
"learning_rate": 0.0001975757575757576, |
|
"loss": 1.2804, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.14626275180535692, |
|
"learning_rate": 0.00019878787878787878, |
|
"loss": 1.2115, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.13100680398305042, |
|
"learning_rate": 0.0002, |
|
"loss": 1.2524, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.14849458896148926, |
|
"learning_rate": 0.00019999977531546566, |
|
"loss": 1.2161, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.13628125499037252, |
|
"learning_rate": 0.0001999991012628722, |
|
"loss": 1.2452, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.18617698759086793, |
|
"learning_rate": 0.00019999797784524866, |
|
"loss": 1.2197, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.14416004826313944, |
|
"learning_rate": 0.00019999640506764336, |
|
"loss": 1.2796, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.13807081386834757, |
|
"learning_rate": 0.0001999943829371238, |
|
"loss": 1.2732, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.16526927436841996, |
|
"learning_rate": 0.0001999919114627769, |
|
"loss": 1.3016, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.14479672734919855, |
|
"learning_rate": 0.0001999889906557086, |
|
"loss": 1.3106, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.13829284006072087, |
|
"learning_rate": 0.00019998562052904418, |
|
"loss": 1.3355, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.13484630104616105, |
|
"learning_rate": 0.0001999818010979279, |
|
"loss": 1.1928, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.14972770674556948, |
|
"learning_rate": 0.00019997753237952317, |
|
"loss": 1.2559, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.13378525020528342, |
|
"learning_rate": 0.00019997281439301218, |
|
"loss": 1.2673, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.13242998699125438, |
|
"learning_rate": 0.00019996764715959618, |
|
"loss": 1.2272, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.12938881004364342, |
|
"learning_rate": 0.00019996203070249516, |
|
"loss": 1.2035, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.13388032350164566, |
|
"learning_rate": 0.00019995596504694763, |
|
"loss": 1.2642, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.13893372222140873, |
|
"learning_rate": 0.00019994945022021082, |
|
"loss": 1.2235, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.14131710715500717, |
|
"learning_rate": 0.00019994248625156038, |
|
"loss": 1.1095, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.13448100369103572, |
|
"learning_rate": 0.0001999350731722902, |
|
"loss": 1.1879, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.13862444003216381, |
|
"learning_rate": 0.00019992721101571236, |
|
"loss": 1.2227, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.13506115547921224, |
|
"learning_rate": 0.00019991889981715698, |
|
"loss": 1.2833, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.13174857502600473, |
|
"learning_rate": 0.00019991013961397197, |
|
"loss": 1.2394, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.1290276308949748, |
|
"learning_rate": 0.00019990093044552304, |
|
"loss": 1.2659, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.1388159912078538, |
|
"learning_rate": 0.0001998912723531933, |
|
"loss": 1.3052, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.1256806205303357, |
|
"learning_rate": 0.00019988116538038325, |
|
"loss": 1.2031, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.13256850855084143, |
|
"learning_rate": 0.00019987060957251047, |
|
"loss": 1.211, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.13197363789890235, |
|
"learning_rate": 0.0001998596049770095, |
|
"loss": 1.2256, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.13277364593883098, |
|
"learning_rate": 0.00019984815164333163, |
|
"loss": 1.2174, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.13838072824574454, |
|
"learning_rate": 0.00019983624962294458, |
|
"loss": 1.3128, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.13524759737199996, |
|
"learning_rate": 0.0001998238989693323, |
|
"loss": 1.1806, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.12669987683723832, |
|
"learning_rate": 0.0001998110997379949, |
|
"loss": 1.2171, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.1461834612451898, |
|
"learning_rate": 0.00019979785198644806, |
|
"loss": 1.2231, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.13265793664862735, |
|
"learning_rate": 0.0001997841557742232, |
|
"loss": 1.1718, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.12842971557690963, |
|
"learning_rate": 0.00019977001116286674, |
|
"loss": 1.2758, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.12188365921206967, |
|
"learning_rate": 0.00019975541821594026, |
|
"loss": 1.2457, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.12679949330022622, |
|
"learning_rate": 0.00019974037699901993, |
|
"loss": 1.1825, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.12949746150357985, |
|
"learning_rate": 0.00019972488757969635, |
|
"loss": 1.2666, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.1363496149379173, |
|
"learning_rate": 0.00019970895002757413, |
|
"loss": 1.2031, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.14218340110669314, |
|
"learning_rate": 0.0001996925644142717, |
|
"loss": 1.3073, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.14234535389443218, |
|
"learning_rate": 0.00019967573081342103, |
|
"loss": 1.2444, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.12866113026310516, |
|
"learning_rate": 0.000199658449300667, |
|
"loss": 1.2257, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.1324053366295965, |
|
"learning_rate": 0.00019964071995366744, |
|
"loss": 1.2374, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.12906841330218152, |
|
"learning_rate": 0.00019962254285209254, |
|
"loss": 1.2334, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.13620873131846425, |
|
"learning_rate": 0.00019960391807762463, |
|
"loss": 1.242, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.14877366842835116, |
|
"learning_rate": 0.00019958484571395757, |
|
"loss": 1.1772, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.13914108740445985, |
|
"learning_rate": 0.00019956532584679675, |
|
"loss": 1.2734, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.13198394930310692, |
|
"learning_rate": 0.00019954535856385837, |
|
"loss": 1.1728, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.3807736597404611, |
|
"learning_rate": 0.0001995249439548693, |
|
"loss": 1.2089, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.1682550557564819, |
|
"learning_rate": 0.00019950408211156636, |
|
"loss": 1.2423, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.2102196862007261, |
|
"learning_rate": 0.0001994827731276963, |
|
"loss": 1.2096, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.154346739470422, |
|
"learning_rate": 0.00019946101709901514, |
|
"loss": 1.2847, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.16416668358293746, |
|
"learning_rate": 0.0001994388141232876, |
|
"loss": 1.2503, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.13134349458231093, |
|
"learning_rate": 0.0001994161643002871, |
|
"loss": 1.1231, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.15083246389185287, |
|
"learning_rate": 0.00019939306773179497, |
|
"loss": 1.1614, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.1742387260929692, |
|
"learning_rate": 0.00019936952452159995, |
|
"loss": 1.3568, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.18146911432436974, |
|
"learning_rate": 0.00019934553477549794, |
|
"loss": 1.2686, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.1393593447949332, |
|
"learning_rate": 0.00019932109860129154, |
|
"loss": 1.1141, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.14856124153987935, |
|
"learning_rate": 0.00019929621610878927, |
|
"loss": 1.234, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.14820851831477327, |
|
"learning_rate": 0.0001992708874098054, |
|
"loss": 1.2069, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.17893142790958147, |
|
"learning_rate": 0.00019924511261815926, |
|
"loss": 1.1278, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.14573658703265605, |
|
"learning_rate": 0.00019921889184967476, |
|
"loss": 1.2292, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.15282321197574994, |
|
"learning_rate": 0.00019919222522217996, |
|
"loss": 1.2482, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.16342112084119492, |
|
"learning_rate": 0.00019916511285550642, |
|
"loss": 1.2172, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.1475889153814455, |
|
"learning_rate": 0.00019913755487148876, |
|
"loss": 1.1747, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.163738064491857, |
|
"learning_rate": 0.00019910955139396396, |
|
"loss": 1.3007, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.14427856196022704, |
|
"learning_rate": 0.00019908110254877106, |
|
"loss": 1.2464, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.20204742660246344, |
|
"learning_rate": 0.00019905220846375032, |
|
"loss": 1.2515, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.15134144918251685, |
|
"learning_rate": 0.0001990228692687429, |
|
"loss": 1.1786, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.1636590177812163, |
|
"learning_rate": 0.00019899308509558998, |
|
"loss": 1.1974, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.15552319776955892, |
|
"learning_rate": 0.00019896285607813244, |
|
"loss": 1.2308, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.17104898009833774, |
|
"learning_rate": 0.00019893218235221015, |
|
"loss": 1.2828, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.16387378763964267, |
|
"learning_rate": 0.00019890106405566138, |
|
"loss": 1.2779, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.14622126798612248, |
|
"learning_rate": 0.00019886950132832207, |
|
"loss": 1.2894, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.16619841547518147, |
|
"learning_rate": 0.0001988374943120254, |
|
"loss": 1.2133, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.12664832399697545, |
|
"learning_rate": 0.00019880504315060096, |
|
"loss": 1.1807, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.2015108381613456, |
|
"learning_rate": 0.00019877214798987426, |
|
"loss": 1.1876, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.14468620723711506, |
|
"learning_rate": 0.00019873880897766598, |
|
"loss": 1.1883, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.1549018650770757, |
|
"learning_rate": 0.00019870502626379127, |
|
"loss": 1.2896, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.1492917963684983, |
|
"learning_rate": 0.0001986708000000593, |
|
"loss": 1.2102, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.178606606459489, |
|
"learning_rate": 0.00019863613034027224, |
|
"loss": 1.2292, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.206170239681528, |
|
"learning_rate": 0.00019860101744022485, |
|
"loss": 1.2666, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.13741043007948167, |
|
"learning_rate": 0.0001985654614577036, |
|
"loss": 1.2022, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.1595080658199459, |
|
"learning_rate": 0.0001985294625524861, |
|
"loss": 1.1203, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.13929705183853777, |
|
"learning_rate": 0.00019849302088634034, |
|
"loss": 1.1505, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.14045247607912964, |
|
"learning_rate": 0.00019845613662302383, |
|
"loss": 1.1897, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.15002651347444407, |
|
"learning_rate": 0.00019841880992828306, |
|
"loss": 1.2133, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.1567929487810952, |
|
"learning_rate": 0.00019838104096985267, |
|
"loss": 1.129, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.15240634543877116, |
|
"learning_rate": 0.00019834282991745464, |
|
"loss": 1.1995, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.151807679821367, |
|
"learning_rate": 0.00019830417694279766, |
|
"loss": 1.25, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.1648599156208311, |
|
"learning_rate": 0.0001982650822195762, |
|
"loss": 1.2511, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.15363401233808713, |
|
"learning_rate": 0.00019822554592346993, |
|
"loss": 1.1794, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.1569644350778875, |
|
"learning_rate": 0.00019818556823214268, |
|
"loss": 1.2033, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.15996552747294254, |
|
"learning_rate": 0.0001981451493252418, |
|
"loss": 1.2809, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.15863104885072635, |
|
"learning_rate": 0.0001981042893843974, |
|
"loss": 1.1667, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.2887466971861171, |
|
"learning_rate": 0.0001980629885932214, |
|
"loss": 1.1915, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.15233015979193984, |
|
"learning_rate": 0.00019802124713730681, |
|
"loss": 1.1734, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.18207884538436447, |
|
"learning_rate": 0.00019797906520422677, |
|
"loss": 1.2575, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.17323546756038308, |
|
"learning_rate": 0.0001979364429835339, |
|
"loss": 1.1704, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.14592153602263633, |
|
"learning_rate": 0.00019789338066675922, |
|
"loss": 1.192, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.19250697792287097, |
|
"learning_rate": 0.0001978498784474115, |
|
"loss": 1.2779, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.1429107680887097, |
|
"learning_rate": 0.0001978059365209762, |
|
"loss": 1.2529, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.48514081074992116, |
|
"learning_rate": 0.00019776155508491482, |
|
"loss": 1.1917, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.1534376167748161, |
|
"learning_rate": 0.0001977167343386638, |
|
"loss": 1.2384, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.16744875760032166, |
|
"learning_rate": 0.00019767147448363366, |
|
"loss": 1.1744, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.29195538170738244, |
|
"learning_rate": 0.00019762577572320824, |
|
"loss": 1.1418, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.1820804717651353, |
|
"learning_rate": 0.00019757963826274357, |
|
"loss": 1.2815, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.17522345110441973, |
|
"learning_rate": 0.00019753306230956718, |
|
"loss": 1.2363, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.16354388270886613, |
|
"learning_rate": 0.000197486048072977, |
|
"loss": 1.2845, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.17590082756401024, |
|
"learning_rate": 0.0001974385957642404, |
|
"loss": 1.192, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.17345720403188775, |
|
"learning_rate": 0.00019739070559659347, |
|
"loss": 1.2068, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.16070434867766506, |
|
"learning_rate": 0.00019734237778523976, |
|
"loss": 1.189, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.18983443066710415, |
|
"learning_rate": 0.0001972936125473495, |
|
"loss": 1.2223, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.15724400187981355, |
|
"learning_rate": 0.00019724441010205863, |
|
"loss": 1.2292, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.14570729442956004, |
|
"learning_rate": 0.00019719477067046766, |
|
"loss": 1.1421, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.1559242881177266, |
|
"learning_rate": 0.00019714469447564088, |
|
"loss": 1.2598, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.16621830243096108, |
|
"learning_rate": 0.0001970941817426052, |
|
"loss": 1.3038, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 4.675483994100576, |
|
"learning_rate": 0.00019704323269834927, |
|
"loss": 1.2298, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.2769699381619058, |
|
"learning_rate": 0.00019699184757182225, |
|
"loss": 1.2566, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.20189839889100783, |
|
"learning_rate": 0.00019694002659393305, |
|
"loss": 1.3181, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.19497107359413876, |
|
"learning_rate": 0.00019688776999754912, |
|
"loss": 1.1502, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.1982266815755412, |
|
"learning_rate": 0.00019683507801749545, |
|
"loss": 1.2053, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.1924340950322314, |
|
"learning_rate": 0.00019678195089055346, |
|
"loss": 1.2149, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.1725322346446431, |
|
"learning_rate": 0.00019672838885546008, |
|
"loss": 1.2553, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.2535488743520272, |
|
"learning_rate": 0.00019667439215290648, |
|
"loss": 1.2576, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.37837586860064026, |
|
"learning_rate": 0.00019661996102553718, |
|
"loss": 1.1815, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.17520419597901843, |
|
"learning_rate": 0.00019656509571794878, |
|
"loss": 1.1932, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.17056234784450633, |
|
"learning_rate": 0.00019650979647668906, |
|
"loss": 1.163, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.18272246580207432, |
|
"learning_rate": 0.00019645406355025565, |
|
"loss": 1.1887, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.17889037954429915, |
|
"learning_rate": 0.00019639789718909508, |
|
"loss": 1.2126, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.23993734971101424, |
|
"learning_rate": 0.00019634129764560168, |
|
"loss": 1.2485, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.1847578318208199, |
|
"learning_rate": 0.00019628426517411625, |
|
"loss": 1.2549, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.23185098827091005, |
|
"learning_rate": 0.00019622680003092503, |
|
"loss": 1.1599, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.220638044092583, |
|
"learning_rate": 0.00019616890247425866, |
|
"loss": 1.2281, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.2303439219825616, |
|
"learning_rate": 0.00019611057276429085, |
|
"loss": 1.2208, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.1744807302230573, |
|
"learning_rate": 0.00019605181116313724, |
|
"loss": 1.2303, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.17510946821872422, |
|
"learning_rate": 0.0001959926179348543, |
|
"loss": 1.2385, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.2218474349751746, |
|
"learning_rate": 0.00019593299334543808, |
|
"loss": 1.2153, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.1742070481516402, |
|
"learning_rate": 0.00019587293766282308, |
|
"loss": 1.1628, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.15250311715180823, |
|
"learning_rate": 0.00019581245115688094, |
|
"loss": 1.1632, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.1744397677094501, |
|
"learning_rate": 0.0001957515340994193, |
|
"loss": 1.254, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.1686772182789891, |
|
"learning_rate": 0.00019569018676418053, |
|
"loss": 1.2169, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.16404966161017623, |
|
"learning_rate": 0.00019562840942684067, |
|
"loss": 1.2221, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.16052011449463713, |
|
"learning_rate": 0.00019556620236500793, |
|
"loss": 1.2045, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.16343251390831215, |
|
"learning_rate": 0.0001955035658582216, |
|
"loss": 1.2289, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.14387162360389305, |
|
"learning_rate": 0.00019544050018795075, |
|
"loss": 1.1365, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.15304461439740238, |
|
"learning_rate": 0.00019537700563759304, |
|
"loss": 1.1931, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.17059958050065627, |
|
"learning_rate": 0.00019531308249247327, |
|
"loss": 1.2166, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.17633385530926995, |
|
"learning_rate": 0.00019524873103984235, |
|
"loss": 1.2604, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.17855814403303746, |
|
"learning_rate": 0.00019518395156887576, |
|
"loss": 1.1615, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.19823982444256988, |
|
"learning_rate": 0.00019511874437067243, |
|
"loss": 1.2153, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.1570784627362585, |
|
"learning_rate": 0.0001950531097382533, |
|
"loss": 1.2788, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.2183125402112695, |
|
"learning_rate": 0.00019498704796656018, |
|
"loss": 1.2966, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.18173933276147194, |
|
"learning_rate": 0.00019492055935245418, |
|
"loss": 1.2978, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.17483116680914407, |
|
"learning_rate": 0.00019485364419471454, |
|
"loss": 1.258, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.15490767356815494, |
|
"learning_rate": 0.0001947863027940374, |
|
"loss": 1.2088, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.14703966491934156, |
|
"learning_rate": 0.00019471853545303405, |
|
"loss": 1.2355, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.14386689086661608, |
|
"learning_rate": 0.00019465034247623003, |
|
"loss": 1.2583, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.18818904376313625, |
|
"learning_rate": 0.00019458172417006347, |
|
"loss": 1.2181, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.17393313719202513, |
|
"learning_rate": 0.00019451268084288385, |
|
"loss": 1.3453, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.14706823379985753, |
|
"learning_rate": 0.00019444321280495043, |
|
"loss": 1.2234, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.15282014755252687, |
|
"learning_rate": 0.00019437332036843118, |
|
"loss": 1.1262, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.1618727884326225, |
|
"learning_rate": 0.00019430300384740105, |
|
"loss": 1.3136, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.16090758705378874, |
|
"learning_rate": 0.00019423226355784077, |
|
"loss": 1.2055, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.15241156801091013, |
|
"learning_rate": 0.00019416109981763526, |
|
"loss": 1.2678, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.14216697909809062, |
|
"learning_rate": 0.0001940895129465724, |
|
"loss": 1.2841, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.15790232415414485, |
|
"learning_rate": 0.00019401750326634144, |
|
"loss": 1.3119, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.13322691961062616, |
|
"learning_rate": 0.0001939450711005316, |
|
"loss": 1.1293, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.14075018938835404, |
|
"learning_rate": 0.00019387221677463062, |
|
"loss": 1.2176, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.21565975459393052, |
|
"learning_rate": 0.00019379894061602335, |
|
"loss": 1.1723, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.17967631394222838, |
|
"learning_rate": 0.00019372524295399013, |
|
"loss": 1.239, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.21187969201978435, |
|
"learning_rate": 0.0001936511241197055, |
|
"loss": 1.2207, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.16967789022974608, |
|
"learning_rate": 0.00019357658444623654, |
|
"loss": 1.2478, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.14810621660374448, |
|
"learning_rate": 0.0001935016242685415, |
|
"loss": 1.1223, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.1489106421847434, |
|
"learning_rate": 0.00019342624392346824, |
|
"loss": 1.1592, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.17625176068748855, |
|
"learning_rate": 0.0001933504437497527, |
|
"loss": 1.2145, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.17250255512763446, |
|
"learning_rate": 0.00019327422408801744, |
|
"loss": 1.2504, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.16079375745566896, |
|
"learning_rate": 0.00019319758528077, |
|
"loss": 1.1795, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.15454466809245995, |
|
"learning_rate": 0.0001931205276724015, |
|
"loss": 1.2123, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.7021323604447972, |
|
"learning_rate": 0.000193043051609185, |
|
"loss": 1.2239, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.1572764339385847, |
|
"learning_rate": 0.00019296515743927399, |
|
"loss": 1.2516, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.2136637778252246, |
|
"learning_rate": 0.00019288684551270073, |
|
"loss": 1.2321, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.4546540454773654, |
|
"learning_rate": 0.00019280811618137484, |
|
"loss": 1.18, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.9809832576786297, |
|
"learning_rate": 0.00019272896979908154, |
|
"loss": 1.2081, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.5246256133291822, |
|
"learning_rate": 0.00019264940672148018, |
|
"loss": 1.2722, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.24941717134878091, |
|
"learning_rate": 0.00019256942730610268, |
|
"loss": 1.2352, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.3356068462072784, |
|
"learning_rate": 0.00019248903191235176, |
|
"loss": 1.2225, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.19535845221880543, |
|
"learning_rate": 0.00019240822090149944, |
|
"loss": 1.1669, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.22306941566416597, |
|
"learning_rate": 0.00019232699463668542, |
|
"loss": 1.2281, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.2700134013989352, |
|
"learning_rate": 0.00019224535348291542, |
|
"loss": 1.1939, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.24406908935562743, |
|
"learning_rate": 0.00019216329780705953, |
|
"loss": 1.1839, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.20465183000217488, |
|
"learning_rate": 0.00019208082797785055, |
|
"loss": 1.2277, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.21324820828129784, |
|
"learning_rate": 0.00019199794436588243, |
|
"loss": 1.2072, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.1780562512431263, |
|
"learning_rate": 0.00019191464734360844, |
|
"loss": 1.2082, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.16547971467615655, |
|
"learning_rate": 0.00019183093728533966, |
|
"loss": 1.1978, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.22904664933247196, |
|
"learning_rate": 0.00019174681456724318, |
|
"loss": 1.1562, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.1737397860007602, |
|
"learning_rate": 0.00019166227956734052, |
|
"loss": 1.2383, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.1589465455917568, |
|
"learning_rate": 0.00019157733266550575, |
|
"loss": 1.2158, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.16253126221999709, |
|
"learning_rate": 0.00019149197424346405, |
|
"loss": 1.1952, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.22436676243032663, |
|
"learning_rate": 0.00019140620468478968, |
|
"loss": 1.2315, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.19291682612950423, |
|
"learning_rate": 0.00019132002437490458, |
|
"loss": 1.2283, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.1519191258459668, |
|
"learning_rate": 0.00019123343370107637, |
|
"loss": 1.1151, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.17179909633547025, |
|
"learning_rate": 0.00019114643305241676, |
|
"loss": 1.1576, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.17992599023321432, |
|
"learning_rate": 0.00019105902281987976, |
|
"loss": 1.2592, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.17714099390314453, |
|
"learning_rate": 0.00019097120339625994, |
|
"loss": 1.2578, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.2455577642687935, |
|
"learning_rate": 0.00019088297517619055, |
|
"loss": 1.2361, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.18398518628783986, |
|
"learning_rate": 0.00019079433855614201, |
|
"loss": 1.1906, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.18944067022821645, |
|
"learning_rate": 0.00019070529393441985, |
|
"loss": 1.237, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.17639967519781063, |
|
"learning_rate": 0.00019061584171116303, |
|
"loss": 1.1841, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.15947129998283005, |
|
"learning_rate": 0.00019052598228834217, |
|
"loss": 1.1722, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.1693354353719105, |
|
"learning_rate": 0.00019043571606975777, |
|
"loss": 1.2204, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.16236190451963983, |
|
"learning_rate": 0.00019034504346103823, |
|
"loss": 1.1778, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.17702370729269964, |
|
"learning_rate": 0.00019025396486963827, |
|
"loss": 1.2065, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.19388150596154238, |
|
"learning_rate": 0.00019016248070483687, |
|
"loss": 1.2942, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.16152000400319103, |
|
"learning_rate": 0.0001900705913777356, |
|
"loss": 1.1784, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.1545267913996029, |
|
"learning_rate": 0.00018997829730125663, |
|
"loss": 1.1829, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.15421727704318197, |
|
"learning_rate": 0.000189885598890141, |
|
"loss": 1.177, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.1624966073814206, |
|
"learning_rate": 0.00018979249656094673, |
|
"loss": 1.2439, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.9490737312904575, |
|
"learning_rate": 0.00018969899073204686, |
|
"loss": 1.2085, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.8982903208613089, |
|
"learning_rate": 0.00018960508182362768, |
|
"loss": 1.2347, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.3771428474797688, |
|
"learning_rate": 0.00018951077025768678, |
|
"loss": 1.2546, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.4776152950069111, |
|
"learning_rate": 0.00018941605645803115, |
|
"loss": 1.2904, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.18786943849618057, |
|
"learning_rate": 0.00018932094085027533, |
|
"loss": 1.2122, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 1.6297025984167128, |
|
"learning_rate": 0.0001892254238618394, |
|
"loss": 1.171, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.20382660707264952, |
|
"learning_rate": 0.0001891295059219472, |
|
"loss": 1.1874, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 1.3580819775908755, |
|
"learning_rate": 0.00018903318746162429, |
|
"loss": 1.1531, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.43619056173016185, |
|
"learning_rate": 0.00018893646891369602, |
|
"loss": 1.2289, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.29385240705823723, |
|
"learning_rate": 0.0001888393507127856, |
|
"loss": 1.2073, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.3136086850525623, |
|
"learning_rate": 0.00018874183329531223, |
|
"loss": 1.1898, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.2307767217662562, |
|
"learning_rate": 0.000188643917099489, |
|
"loss": 1.207, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.18703654518135468, |
|
"learning_rate": 0.000188545602565321, |
|
"loss": 1.1688, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.4809351333934126, |
|
"learning_rate": 0.00018844689013460336, |
|
"loss": 1.2519, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.40370101428544464, |
|
"learning_rate": 0.0001883477802509192, |
|
"loss": 1.2411, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.2858848636432859, |
|
"learning_rate": 0.00018824827335963765, |
|
"loss": 1.194, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.32195602638999565, |
|
"learning_rate": 0.000188148369907912, |
|
"loss": 1.0988, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.23790306908901832, |
|
"learning_rate": 0.00018804807034467733, |
|
"loss": 1.2237, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.20126988767112128, |
|
"learning_rate": 0.0001879473751206489, |
|
"loss": 1.2731, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.3336380339194037, |
|
"learning_rate": 0.00018784628468831996, |
|
"loss": 1.2369, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.5054330893305989, |
|
"learning_rate": 0.0001877447995019596, |
|
"loss": 1.2443, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.2297866279715136, |
|
"learning_rate": 0.0001876429200176108, |
|
"loss": 1.2376, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.39350567174184636, |
|
"learning_rate": 0.00018754064669308858, |
|
"loss": 1.2126, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.2025361091435325, |
|
"learning_rate": 0.00018743797998797753, |
|
"loss": 1.2224, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.31824903419753814, |
|
"learning_rate": 0.00018733492036363005, |
|
"loss": 1.2942, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 2.4642066748643017, |
|
"learning_rate": 0.00018723146828316428, |
|
"loss": 1.2515, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.7833055646295342, |
|
"learning_rate": 0.00018712762421146183, |
|
"loss": 1.2207, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.2810249021786599, |
|
"learning_rate": 0.00018702338861516587, |
|
"loss": 1.2755, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.460995724241333, |
|
"learning_rate": 0.0001869187619626789, |
|
"loss": 1.2856, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.49139203044984286, |
|
"learning_rate": 0.00018681374472416073, |
|
"loss": 1.2392, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 2.691604613969173, |
|
"learning_rate": 0.0001867083373715264, |
|
"loss": 1.2992, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.8014112047318501, |
|
"learning_rate": 0.00018660254037844388, |
|
"loss": 1.2683, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.31614342841331383, |
|
"learning_rate": 0.00018649635422033215, |
|
"loss": 1.2356, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.2559855196513244, |
|
"learning_rate": 0.000186389779374359, |
|
"loss": 1.2053, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 1.6613999986014714, |
|
"learning_rate": 0.0001862828163194388, |
|
"loss": 1.2568, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.27190082167109786, |
|
"learning_rate": 0.0001861754655362304, |
|
"loss": 1.1288, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.43819582203066043, |
|
"learning_rate": 0.00018606772750713504, |
|
"loss": 1.1758, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.41738497400383384, |
|
"learning_rate": 0.0001859596027162941, |
|
"loss": 1.2993, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.2595142634740817, |
|
"learning_rate": 0.000185851091649587, |
|
"loss": 1.269, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.2795314201020271, |
|
"learning_rate": 0.00018574219479462878, |
|
"loss": 1.1915, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.2502992494749938, |
|
"learning_rate": 0.00018563291264076835, |
|
"loss": 1.2157, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.27422512335538374, |
|
"learning_rate": 0.00018552324567908585, |
|
"loss": 1.2541, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 1.3360989016060905, |
|
"learning_rate": 0.00018541319440239066, |
|
"loss": 1.2666, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 1.9441131913572127, |
|
"learning_rate": 0.00018530275930521924, |
|
"loss": 1.2924, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.3772373301771213, |
|
"learning_rate": 0.00018519194088383273, |
|
"loss": 1.1952, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.22091753616251295, |
|
"learning_rate": 0.0001850807396362148, |
|
"loss": 1.1858, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.21423504993321807, |
|
"learning_rate": 0.00018496915606206951, |
|
"loss": 1.2245, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.5238946238105926, |
|
"learning_rate": 0.00018485719066281892, |
|
"loss": 1.2351, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.3037858949309141, |
|
"learning_rate": 0.0001847448439416009, |
|
"loss": 1.1669, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.21553286799952254, |
|
"learning_rate": 0.00018463211640326686, |
|
"loss": 1.1454, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.202875547805464, |
|
"learning_rate": 0.0001845190085543795, |
|
"loss": 1.188, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.24385408620619278, |
|
"learning_rate": 0.00018440552090321047, |
|
"loss": 1.2307, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.23793944272430378, |
|
"learning_rate": 0.0001842916539597382, |
|
"loss": 1.2253, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.17062488448810784, |
|
"learning_rate": 0.0001841774082356455, |
|
"loss": 1.2681, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.20003742001916064, |
|
"learning_rate": 0.00018406278424431736, |
|
"loss": 1.2428, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.2696052831337752, |
|
"learning_rate": 0.0001839477825008385, |
|
"loss": 1.2945, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.23302960820538443, |
|
"learning_rate": 0.00018383240352199117, |
|
"loss": 1.1718, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.38187833239777536, |
|
"learning_rate": 0.00018371664782625287, |
|
"loss": 1.2311, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.4052561772533732, |
|
"learning_rate": 0.00018360051593379383, |
|
"loss": 1.1639, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.23379763821020377, |
|
"learning_rate": 0.0001834840083664749, |
|
"loss": 1.1809, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.2368414607613928, |
|
"learning_rate": 0.00018336712564784503, |
|
"loss": 1.2357, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.20230633988510938, |
|
"learning_rate": 0.000183249868303139, |
|
"loss": 1.1851, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.170513157244292, |
|
"learning_rate": 0.00018313223685927505, |
|
"loss": 1.205, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.18082295035256266, |
|
"learning_rate": 0.0001830142318448525, |
|
"loss": 1.2305, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.18286299264146286, |
|
"learning_rate": 0.00018289585379014942, |
|
"loss": 1.23, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.17868104103482751, |
|
"learning_rate": 0.00018277710322712012, |
|
"loss": 1.2894, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.1820411127336495, |
|
"learning_rate": 0.00018265798068939294, |
|
"loss": 1.2395, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.1738237541783663, |
|
"learning_rate": 0.0001825384867122677, |
|
"loss": 1.1576, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.15693445967795147, |
|
"learning_rate": 0.0001824186218327134, |
|
"loss": 1.0809, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.18509145652208978, |
|
"learning_rate": 0.00018229838658936564, |
|
"loss": 1.2717, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.14702488366564262, |
|
"learning_rate": 0.0001821777815225245, |
|
"loss": 1.2236, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.1828399354418095, |
|
"learning_rate": 0.00018205680717415187, |
|
"loss": 1.2565, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.17460984182013486, |
|
"learning_rate": 0.00018193546408786898, |
|
"loss": 1.2474, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.2001623109673152, |
|
"learning_rate": 0.00018181375280895416, |
|
"loss": 1.2544, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.17228631742863837, |
|
"learning_rate": 0.00018169167388434025, |
|
"loss": 1.1851, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.1644862232819482, |
|
"learning_rate": 0.00018156922786261216, |
|
"loss": 1.1817, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.19775186397477057, |
|
"learning_rate": 0.00018144641529400446, |
|
"loss": 1.257, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.1626281991220394, |
|
"learning_rate": 0.00018132323673039885, |
|
"loss": 1.2277, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.16158256707311264, |
|
"learning_rate": 0.00018119969272532166, |
|
"loss": 1.1624, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.17705809207051687, |
|
"learning_rate": 0.00018107578383394146, |
|
"loss": 1.2421, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.17639060401882287, |
|
"learning_rate": 0.00018095151061306645, |
|
"loss": 1.285, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.16918796486576196, |
|
"learning_rate": 0.00018082687362114212, |
|
"loss": 1.2606, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.15968377185965665, |
|
"learning_rate": 0.0001807018734182485, |
|
"loss": 1.194, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.17537027967397978, |
|
"learning_rate": 0.00018057651056609784, |
|
"loss": 1.1594, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.15753665403127565, |
|
"learning_rate": 0.00018045078562803203, |
|
"loss": 1.1382, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.17121200763916436, |
|
"learning_rate": 0.00018032469916902003, |
|
"loss": 1.2286, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.19120510133331003, |
|
"learning_rate": 0.00018019825175565542, |
|
"loss": 1.2835, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.1671735980123817, |
|
"learning_rate": 0.0001800714439561538, |
|
"loss": 1.2201, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.1579098534969056, |
|
"learning_rate": 0.00017994427634035015, |
|
"loss": 1.2156, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.1746075421158512, |
|
"learning_rate": 0.00017981674947969636, |
|
"loss": 1.2049, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.16878182886737042, |
|
"learning_rate": 0.00017968886394725874, |
|
"loss": 1.2204, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.16725956538286493, |
|
"learning_rate": 0.00017956062031771535, |
|
"loss": 1.2091, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.18877845951705005, |
|
"learning_rate": 0.00017943201916735335, |
|
"loss": 1.241, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.180337447476004, |
|
"learning_rate": 0.00017930306107406653, |
|
"loss": 1.2253, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.16688572366717752, |
|
"learning_rate": 0.0001791737466173527, |
|
"loss": 1.239, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.15385917621135983, |
|
"learning_rate": 0.00017904407637831099, |
|
"loss": 1.2476, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.17725645269055587, |
|
"learning_rate": 0.00017891405093963938, |
|
"loss": 1.2599, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.14758551718901028, |
|
"learning_rate": 0.00017878367088563195, |
|
"loss": 1.2249, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.15216962408661316, |
|
"learning_rate": 0.00017865293680217637, |
|
"loss": 1.2346, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.16679282848599514, |
|
"learning_rate": 0.00017852184927675112, |
|
"loss": 1.2443, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.16723562739069214, |
|
"learning_rate": 0.00017839040889842305, |
|
"loss": 1.224, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.15922276239929914, |
|
"learning_rate": 0.00017825861625784455, |
|
"loss": 1.2739, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.1510107938469514, |
|
"learning_rate": 0.00017812647194725094, |
|
"loss": 1.1764, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.16446999054333494, |
|
"learning_rate": 0.00017799397656045792, |
|
"loss": 1.2498, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.18566301651865832, |
|
"learning_rate": 0.00017786113069285874, |
|
"loss": 1.232, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.20592971655306183, |
|
"learning_rate": 0.00017772793494142167, |
|
"loss": 1.1586, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.1581947714375729, |
|
"learning_rate": 0.00017759438990468725, |
|
"loss": 1.2502, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.15466760695169174, |
|
"learning_rate": 0.00017746049618276545, |
|
"loss": 1.1605, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.16041506222444918, |
|
"learning_rate": 0.00017732625437733335, |
|
"loss": 1.2778, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.17168109661676773, |
|
"learning_rate": 0.0001771916650916321, |
|
"loss": 1.262, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.1788973186498254, |
|
"learning_rate": 0.00017705672893046425, |
|
"loss": 1.2111, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.1759644359346382, |
|
"learning_rate": 0.00017692144650019125, |
|
"loss": 1.2546, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.15710749736088767, |
|
"learning_rate": 0.0001767858184087304, |
|
"loss": 1.2487, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.1648235522911144, |
|
"learning_rate": 0.00017664984526555248, |
|
"loss": 1.2469, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.15452607969890703, |
|
"learning_rate": 0.0001765135276816787, |
|
"loss": 1.1855, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.1837695597880219, |
|
"learning_rate": 0.00017637686626967812, |
|
"loss": 1.2185, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.15861390725762364, |
|
"learning_rate": 0.00017623986164366486, |
|
"loss": 1.2056, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.1663260460966887, |
|
"learning_rate": 0.00017610251441929533, |
|
"loss": 1.1242, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.1803309720529981, |
|
"learning_rate": 0.00017596482521376546, |
|
"loss": 1.2938, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.14909085011764342, |
|
"learning_rate": 0.00017582679464580797, |
|
"loss": 1.1953, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.15779022242482527, |
|
"learning_rate": 0.00017568842333568952, |
|
"loss": 1.2792, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.1553327313967345, |
|
"learning_rate": 0.00017554971190520798, |
|
"loss": 1.2286, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.16363964666273684, |
|
"learning_rate": 0.00017541066097768963, |
|
"loss": 1.2753, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.16668099163659675, |
|
"learning_rate": 0.00017527127117798635, |
|
"loss": 1.185, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.13957568397594883, |
|
"learning_rate": 0.0001751315431324727, |
|
"loss": 1.143, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.1553111736740035, |
|
"learning_rate": 0.00017499147746904335, |
|
"loss": 1.2492, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.1691517335818193, |
|
"learning_rate": 0.00017485107481711012, |
|
"loss": 1.2619, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.15480883994395986, |
|
"learning_rate": 0.00017471033580759903, |
|
"loss": 1.2396, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.1451690143792058, |
|
"learning_rate": 0.00017456926107294765, |
|
"loss": 1.1732, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.1524398957482947, |
|
"learning_rate": 0.00017442785124710227, |
|
"loss": 1.2083, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.16790264977550012, |
|
"learning_rate": 0.0001742861069655148, |
|
"loss": 1.2201, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.1529847047636337, |
|
"learning_rate": 0.0001741440288651403, |
|
"loss": 1.243, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.1485875402374676, |
|
"learning_rate": 0.00017400161758443375, |
|
"loss": 1.2053, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.16950094279079617, |
|
"learning_rate": 0.00017385887376334742, |
|
"loss": 1.1944, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.15289337084330445, |
|
"learning_rate": 0.00017371579804332789, |
|
"loss": 1.2503, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.15337063655317973, |
|
"learning_rate": 0.00017357239106731317, |
|
"loss": 1.3092, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.1458937961897621, |
|
"learning_rate": 0.00017342865347972988, |
|
"loss": 1.2244, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.19897118610161338, |
|
"learning_rate": 0.00017328458592649027, |
|
"loss": 1.2238, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.15850805264911003, |
|
"learning_rate": 0.00017314018905498931, |
|
"loss": 1.195, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.14445183074519347, |
|
"learning_rate": 0.00017299546351410197, |
|
"loss": 1.1974, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.18180731722745677, |
|
"learning_rate": 0.00017285040995418, |
|
"loss": 1.2107, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.14943874953193587, |
|
"learning_rate": 0.00017270502902704926, |
|
"loss": 1.1843, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.15767466790910512, |
|
"learning_rate": 0.00017255932138600665, |
|
"loss": 1.1409, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.16402921378654775, |
|
"learning_rate": 0.00017241328768581726, |
|
"loss": 1.2135, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.15526246786505485, |
|
"learning_rate": 0.00017226692858271134, |
|
"loss": 1.2255, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.16608155892622348, |
|
"learning_rate": 0.00017212024473438147, |
|
"loss": 1.2691, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.14913271520144072, |
|
"learning_rate": 0.00017197323679997943, |
|
"loss": 1.1574, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.1471910610421707, |
|
"learning_rate": 0.00017182590544011347, |
|
"loss": 1.2774, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.1417464185073962, |
|
"learning_rate": 0.00017167825131684513, |
|
"loss": 1.2446, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.1610488125634495, |
|
"learning_rate": 0.0001715302750936864, |
|
"loss": 1.2862, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.20227974555123074, |
|
"learning_rate": 0.00017138197743559654, |
|
"loss": 1.207, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.1355502559749413, |
|
"learning_rate": 0.00017123335900897946, |
|
"loss": 1.1019, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.1559423167028215, |
|
"learning_rate": 0.00017108442048168038, |
|
"loss": 1.2549, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.15898973818185586, |
|
"learning_rate": 0.00017093516252298296, |
|
"loss": 1.2705, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.15169569998999652, |
|
"learning_rate": 0.00017078558580360632, |
|
"loss": 1.2454, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.15976111665597925, |
|
"learning_rate": 0.00017063569099570196, |
|
"loss": 1.2585, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.14488877221999352, |
|
"learning_rate": 0.00017048547877285077, |
|
"loss": 1.2169, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.14919533098974924, |
|
"learning_rate": 0.00017033494981006002, |
|
"loss": 1.2358, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.15251746717084805, |
|
"learning_rate": 0.00017018410478376032, |
|
"loss": 1.2241, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.1456060482002663, |
|
"learning_rate": 0.00017003294437180255, |
|
"loss": 1.2298, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.17048886778787248, |
|
"learning_rate": 0.00016988146925345484, |
|
"loss": 1.2707, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.15304381059310815, |
|
"learning_rate": 0.00016972968010939954, |
|
"loss": 1.1498, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.16590055969071696, |
|
"learning_rate": 0.0001695775776217301, |
|
"loss": 1.2481, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.14299575837437278, |
|
"learning_rate": 0.00016942516247394807, |
|
"loss": 1.2058, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.14275107775859475, |
|
"learning_rate": 0.00016927243535095997, |
|
"loss": 1.2178, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.1554250137491414, |
|
"learning_rate": 0.0001691193969390742, |
|
"loss": 1.1197, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.16958418467021688, |
|
"learning_rate": 0.0001689660479259981, |
|
"loss": 1.1768, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.1546216583314497, |
|
"learning_rate": 0.00016881238900083473, |
|
"loss": 1.1741, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.15287056494787424, |
|
"learning_rate": 0.0001686584208540797, |
|
"loss": 1.2328, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.1419329373337611, |
|
"learning_rate": 0.0001685041441776183, |
|
"loss": 1.1743, |
|
"step": 550 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 1647, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 50, |
|
"total_flos": 5103643602714624.0, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|