|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.0036429872495445, |
|
"eval_steps": 500, |
|
"global_step": 1100, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 0.46431864432643544, |
|
"learning_rate": 1.2121212121212122e-06, |
|
"loss": 1.4151, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 0.42969176658539837, |
|
"learning_rate": 2.4242424242424244e-06, |
|
"loss": 1.3729, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.5004307936270223, |
|
"learning_rate": 3.636363636363636e-06, |
|
"loss": 1.3989, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.43666634920041486, |
|
"learning_rate": 4.848484848484849e-06, |
|
"loss": 1.3363, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.4691114419353825, |
|
"learning_rate": 6.060606060606061e-06, |
|
"loss": 1.4293, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.4277596061377729, |
|
"learning_rate": 7.272727272727272e-06, |
|
"loss": 1.4343, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.4238339229382504, |
|
"learning_rate": 8.484848484848486e-06, |
|
"loss": 1.4462, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.40992048534183273, |
|
"learning_rate": 9.696969696969698e-06, |
|
"loss": 1.2756, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.37885700313540693, |
|
"learning_rate": 1.0909090909090909e-05, |
|
"loss": 1.3464, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.36066141170123023, |
|
"learning_rate": 1.2121212121212122e-05, |
|
"loss": 1.3419, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.35617169386863406, |
|
"learning_rate": 1.3333333333333333e-05, |
|
"loss": 1.3533, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.3040188564782602, |
|
"learning_rate": 1.4545454545454545e-05, |
|
"loss": 1.2395, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.31038319439216566, |
|
"learning_rate": 1.5757575757575756e-05, |
|
"loss": 1.3082, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.26683768372135835, |
|
"learning_rate": 1.6969696969696972e-05, |
|
"loss": 1.3063, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.3652323682563078, |
|
"learning_rate": 1.8181818181818182e-05, |
|
"loss": 1.3045, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.23559121485457843, |
|
"learning_rate": 1.9393939393939395e-05, |
|
"loss": 1.2366, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.2342299313020104, |
|
"learning_rate": 2.0606060606060608e-05, |
|
"loss": 1.2831, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.2202931700357255, |
|
"learning_rate": 2.1818181818181818e-05, |
|
"loss": 1.3064, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.2097660599292375, |
|
"learning_rate": 2.3030303030303034e-05, |
|
"loss": 1.2376, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.2356785314652122, |
|
"learning_rate": 2.4242424242424244e-05, |
|
"loss": 1.2802, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.24639302530564244, |
|
"learning_rate": 2.5454545454545454e-05, |
|
"loss": 1.3016, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.24373126133228787, |
|
"learning_rate": 2.6666666666666667e-05, |
|
"loss": 1.3407, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.24488805144123432, |
|
"learning_rate": 2.7878787878787883e-05, |
|
"loss": 1.3325, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.2653033507571198, |
|
"learning_rate": 2.909090909090909e-05, |
|
"loss": 1.2811, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.2841724819336817, |
|
"learning_rate": 3.0303030303030306e-05, |
|
"loss": 1.2837, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.2183883020111492, |
|
"learning_rate": 3.151515151515151e-05, |
|
"loss": 1.2472, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.2137995163762026, |
|
"learning_rate": 3.272727272727273e-05, |
|
"loss": 1.2854, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.19499006223503876, |
|
"learning_rate": 3.3939393939393945e-05, |
|
"loss": 1.3018, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.17367919355340256, |
|
"learning_rate": 3.515151515151515e-05, |
|
"loss": 1.2824, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.18326045693683557, |
|
"learning_rate": 3.6363636363636364e-05, |
|
"loss": 1.2192, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.17474388188066411, |
|
"learning_rate": 3.757575757575758e-05, |
|
"loss": 1.2078, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.17856970178098716, |
|
"learning_rate": 3.878787878787879e-05, |
|
"loss": 1.2683, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.18617589704298348, |
|
"learning_rate": 4e-05, |
|
"loss": 1.2265, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.17653209733215317, |
|
"learning_rate": 4.1212121212121216e-05, |
|
"loss": 1.319, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.1722921367585233, |
|
"learning_rate": 4.242424242424243e-05, |
|
"loss": 1.2117, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.176642606378719, |
|
"learning_rate": 4.3636363636363636e-05, |
|
"loss": 1.2512, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.16696442324691066, |
|
"learning_rate": 4.484848484848485e-05, |
|
"loss": 1.2637, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.17035384059517106, |
|
"learning_rate": 4.606060606060607e-05, |
|
"loss": 1.2699, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.15545801881444482, |
|
"learning_rate": 4.7272727272727275e-05, |
|
"loss": 1.2939, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.17111439344347512, |
|
"learning_rate": 4.848484848484849e-05, |
|
"loss": 1.3033, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.16994151343455458, |
|
"learning_rate": 4.9696969696969694e-05, |
|
"loss": 1.2603, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.15929214926453447, |
|
"learning_rate": 5.090909090909091e-05, |
|
"loss": 1.2626, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.16761261516238699, |
|
"learning_rate": 5.212121212121213e-05, |
|
"loss": 1.296, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.15754700542426123, |
|
"learning_rate": 5.333333333333333e-05, |
|
"loss": 1.278, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.15522526683877644, |
|
"learning_rate": 5.4545454545454546e-05, |
|
"loss": 1.2355, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.1577929926930023, |
|
"learning_rate": 5.5757575757575766e-05, |
|
"loss": 1.2879, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.31075066632858317, |
|
"learning_rate": 5.696969696969697e-05, |
|
"loss": 1.2202, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.1663780653395111, |
|
"learning_rate": 5.818181818181818e-05, |
|
"loss": 1.2319, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.16049499655883026, |
|
"learning_rate": 5.93939393939394e-05, |
|
"loss": 1.2801, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.14515773124436285, |
|
"learning_rate": 6.060606060606061e-05, |
|
"loss": 1.2588, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.14653064850325623, |
|
"learning_rate": 6.181818181818182e-05, |
|
"loss": 1.2677, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.17193239746689878, |
|
"learning_rate": 6.303030303030302e-05, |
|
"loss": 1.2742, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.1967020450342533, |
|
"learning_rate": 6.424242424242424e-05, |
|
"loss": 1.1545, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.16247531997247225, |
|
"learning_rate": 6.545454545454546e-05, |
|
"loss": 1.222, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.14990706377244528, |
|
"learning_rate": 6.666666666666667e-05, |
|
"loss": 1.2103, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.1412817445239095, |
|
"learning_rate": 6.787878787878789e-05, |
|
"loss": 1.2169, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.14575971073482757, |
|
"learning_rate": 6.90909090909091e-05, |
|
"loss": 1.2751, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.13714747569950891, |
|
"learning_rate": 7.03030303030303e-05, |
|
"loss": 1.2508, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.14334695156859903, |
|
"learning_rate": 7.151515151515152e-05, |
|
"loss": 1.2721, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.1456824177522916, |
|
"learning_rate": 7.272727272727273e-05, |
|
"loss": 1.2649, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.15030318240210044, |
|
"learning_rate": 7.393939393939395e-05, |
|
"loss": 1.2167, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.1651326066719482, |
|
"learning_rate": 7.515151515151515e-05, |
|
"loss": 1.3126, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.1408250406479118, |
|
"learning_rate": 7.636363636363637e-05, |
|
"loss": 1.2891, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.21501384376905694, |
|
"learning_rate": 7.757575757575758e-05, |
|
"loss": 1.3019, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.1365168726167339, |
|
"learning_rate": 7.878787878787879e-05, |
|
"loss": 1.2498, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.1431463689660936, |
|
"learning_rate": 8e-05, |
|
"loss": 1.2793, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.13689045214286194, |
|
"learning_rate": 8.121212121212121e-05, |
|
"loss": 1.2295, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.13483608710081227, |
|
"learning_rate": 8.242424242424243e-05, |
|
"loss": 1.2258, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.13707618564415613, |
|
"learning_rate": 8.363636363636364e-05, |
|
"loss": 1.2252, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.13780236215967515, |
|
"learning_rate": 8.484848484848486e-05, |
|
"loss": 1.2565, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.14036805493494423, |
|
"learning_rate": 8.606060606060606e-05, |
|
"loss": 1.3023, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.12776919439147982, |
|
"learning_rate": 8.727272727272727e-05, |
|
"loss": 1.2292, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.1289941815481437, |
|
"learning_rate": 8.848484848484849e-05, |
|
"loss": 1.2191, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.13943952294847306, |
|
"learning_rate": 8.96969696969697e-05, |
|
"loss": 1.2915, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.1493528502117281, |
|
"learning_rate": 9.090909090909092e-05, |
|
"loss": 1.2797, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.1252401242451818, |
|
"learning_rate": 9.212121212121214e-05, |
|
"loss": 1.2552, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.13969800467546992, |
|
"learning_rate": 9.333333333333334e-05, |
|
"loss": 1.3147, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.1277258491470434, |
|
"learning_rate": 9.454545454545455e-05, |
|
"loss": 1.2089, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.133041369314817, |
|
"learning_rate": 9.575757575757576e-05, |
|
"loss": 1.2761, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.14564572037181842, |
|
"learning_rate": 9.696969696969698e-05, |
|
"loss": 1.1901, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.13666505656492195, |
|
"learning_rate": 9.818181818181818e-05, |
|
"loss": 1.2615, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.135007805210003, |
|
"learning_rate": 9.939393939393939e-05, |
|
"loss": 1.2669, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.17287563365884975, |
|
"learning_rate": 0.00010060606060606062, |
|
"loss": 1.2669, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.12934306326048103, |
|
"learning_rate": 0.00010181818181818181, |
|
"loss": 1.1979, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.13517436169178096, |
|
"learning_rate": 0.00010303030303030303, |
|
"loss": 1.2226, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.12105351159271568, |
|
"learning_rate": 0.00010424242424242425, |
|
"loss": 1.1172, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.1281676431775383, |
|
"learning_rate": 0.00010545454545454545, |
|
"loss": 1.2046, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.11730963057933333, |
|
"learning_rate": 0.00010666666666666667, |
|
"loss": 1.1883, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.12655235108503246, |
|
"learning_rate": 0.00010787878787878789, |
|
"loss": 1.1331, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.13047560307970027, |
|
"learning_rate": 0.00010909090909090909, |
|
"loss": 1.2731, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.12193522973752649, |
|
"learning_rate": 0.00011030303030303031, |
|
"loss": 1.2161, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.12804360300116346, |
|
"learning_rate": 0.00011151515151515153, |
|
"loss": 1.3062, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.15991741754516206, |
|
"learning_rate": 0.00011272727272727272, |
|
"loss": 1.239, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.15140182244454561, |
|
"learning_rate": 0.00011393939393939394, |
|
"loss": 1.2349, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.12320241076263434, |
|
"learning_rate": 0.00011515151515151516, |
|
"loss": 1.2875, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.13235998458230466, |
|
"learning_rate": 0.00011636363636363636, |
|
"loss": 1.2218, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.11783688734798668, |
|
"learning_rate": 0.00011757575757575758, |
|
"loss": 1.1864, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.3151933420750235, |
|
"learning_rate": 0.0001187878787878788, |
|
"loss": 1.3023, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.12665632567219295, |
|
"learning_rate": 0.00012, |
|
"loss": 1.2249, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.1228886740460738, |
|
"learning_rate": 0.00012121212121212122, |
|
"loss": 1.2517, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.11892005244989344, |
|
"learning_rate": 0.00012242424242424243, |
|
"loss": 1.2586, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.1232340827222201, |
|
"learning_rate": 0.00012363636363636364, |
|
"loss": 1.3217, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.13837226869323116, |
|
"learning_rate": 0.00012484848484848487, |
|
"loss": 1.2693, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.12068217991774362, |
|
"learning_rate": 0.00012606060606060605, |
|
"loss": 1.2623, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.16779277284606545, |
|
"learning_rate": 0.00012727272727272728, |
|
"loss": 1.2415, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.13396891539963085, |
|
"learning_rate": 0.0001284848484848485, |
|
"loss": 1.2313, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.12457104490772812, |
|
"learning_rate": 0.0001296969696969697, |
|
"loss": 1.1758, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.12676816816563452, |
|
"learning_rate": 0.00013090909090909093, |
|
"loss": 1.2478, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.11973639622066906, |
|
"learning_rate": 0.00013212121212121213, |
|
"loss": 1.2335, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.1330159646034068, |
|
"learning_rate": 0.00013333333333333334, |
|
"loss": 1.26, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.1298003025338099, |
|
"learning_rate": 0.00013454545454545455, |
|
"loss": 1.1907, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.1226154813287666, |
|
"learning_rate": 0.00013575757575757578, |
|
"loss": 1.1807, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.12533753244302145, |
|
"learning_rate": 0.00013696969696969696, |
|
"loss": 1.2098, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.12673266503840944, |
|
"learning_rate": 0.0001381818181818182, |
|
"loss": 1.2265, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.1299039569361384, |
|
"learning_rate": 0.0001393939393939394, |
|
"loss": 1.2534, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.13023496663090803, |
|
"learning_rate": 0.0001406060606060606, |
|
"loss": 1.2453, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.12001793500864573, |
|
"learning_rate": 0.00014181818181818184, |
|
"loss": 1.1608, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.14561862193041028, |
|
"learning_rate": 0.00014303030303030304, |
|
"loss": 1.2233, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.12636130876430832, |
|
"learning_rate": 0.00014424242424242425, |
|
"loss": 1.2833, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.189556849271166, |
|
"learning_rate": 0.00014545454545454546, |
|
"loss": 1.3105, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.12409073764495662, |
|
"learning_rate": 0.00014666666666666666, |
|
"loss": 1.1534, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.12149212466969316, |
|
"learning_rate": 0.0001478787878787879, |
|
"loss": 1.3039, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.12147336887953522, |
|
"learning_rate": 0.0001490909090909091, |
|
"loss": 1.326, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.1176585016163167, |
|
"learning_rate": 0.0001503030303030303, |
|
"loss": 1.191, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.2066428974234372, |
|
"learning_rate": 0.00015151515151515152, |
|
"loss": 1.3054, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.29582724255710047, |
|
"learning_rate": 0.00015272727272727275, |
|
"loss": 1.2032, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.13084381204119358, |
|
"learning_rate": 0.00015393939393939393, |
|
"loss": 1.2289, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.1294157600411397, |
|
"learning_rate": 0.00015515151515151516, |
|
"loss": 1.2561, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.14039614543447027, |
|
"learning_rate": 0.00015636363636363637, |
|
"loss": 1.243, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.19939984917282128, |
|
"learning_rate": 0.00015757575757575757, |
|
"loss": 1.1286, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.14402764349968203, |
|
"learning_rate": 0.0001587878787878788, |
|
"loss": 1.1959, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.13970978861500938, |
|
"learning_rate": 0.00016, |
|
"loss": 1.1814, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.14539538472563127, |
|
"learning_rate": 0.00016121212121212122, |
|
"loss": 1.2317, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.13456425455391557, |
|
"learning_rate": 0.00016242424242424243, |
|
"loss": 1.2239, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.1314997837157779, |
|
"learning_rate": 0.00016363636363636366, |
|
"loss": 1.1986, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.14046946525591422, |
|
"learning_rate": 0.00016484848484848487, |
|
"loss": 1.2238, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.6095538041505763, |
|
"learning_rate": 0.00016606060606060607, |
|
"loss": 1.2332, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.17707289712054367, |
|
"learning_rate": 0.00016727272727272728, |
|
"loss": 1.2401, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.19335172179099247, |
|
"learning_rate": 0.00016848484848484848, |
|
"loss": 1.2361, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.13725591818701255, |
|
"learning_rate": 0.00016969696969696972, |
|
"loss": 1.193, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.15535575462507384, |
|
"learning_rate": 0.0001709090909090909, |
|
"loss": 1.2769, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.14909436560898923, |
|
"learning_rate": 0.00017212121212121213, |
|
"loss": 1.2602, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.15054368082407957, |
|
"learning_rate": 0.00017333333333333334, |
|
"loss": 1.2607, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.13386897838741724, |
|
"learning_rate": 0.00017454545454545454, |
|
"loss": 1.168, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.13567889528730145, |
|
"learning_rate": 0.00017575757575757578, |
|
"loss": 1.1984, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.13994382298003089, |
|
"learning_rate": 0.00017696969696969698, |
|
"loss": 1.2795, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.13941573210713187, |
|
"learning_rate": 0.0001781818181818182, |
|
"loss": 1.2303, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.18302605925485763, |
|
"learning_rate": 0.0001793939393939394, |
|
"loss": 1.2696, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.1547402223275396, |
|
"learning_rate": 0.00018060606060606063, |
|
"loss": 1.1276, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.19947594494850646, |
|
"learning_rate": 0.00018181818181818183, |
|
"loss": 1.271, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.1517101450465788, |
|
"learning_rate": 0.00018303030303030304, |
|
"loss": 1.2193, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.19251063116857103, |
|
"learning_rate": 0.00018424242424242427, |
|
"loss": 1.2703, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.16789099560498666, |
|
"learning_rate": 0.00018545454545454545, |
|
"loss": 1.2244, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.14907376557922342, |
|
"learning_rate": 0.0001866666666666667, |
|
"loss": 1.264, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.14276598263036905, |
|
"learning_rate": 0.0001878787878787879, |
|
"loss": 1.2545, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.14526753816999002, |
|
"learning_rate": 0.0001890909090909091, |
|
"loss": 1.2912, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.1627048894660859, |
|
"learning_rate": 0.0001903030303030303, |
|
"loss": 1.2573, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.16405036632332695, |
|
"learning_rate": 0.0001915151515151515, |
|
"loss": 1.2359, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.14533427219788658, |
|
"learning_rate": 0.00019272727272727274, |
|
"loss": 1.1718, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.13802382666732702, |
|
"learning_rate": 0.00019393939393939395, |
|
"loss": 1.2297, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.15620193618511755, |
|
"learning_rate": 0.00019515151515151516, |
|
"loss": 1.2287, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.1401696295700075, |
|
"learning_rate": 0.00019636363636363636, |
|
"loss": 1.2231, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.15816133304035035, |
|
"learning_rate": 0.0001975757575757576, |
|
"loss": 1.2804, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.14626275180535692, |
|
"learning_rate": 0.00019878787878787878, |
|
"loss": 1.2115, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.13100680398305042, |
|
"learning_rate": 0.0002, |
|
"loss": 1.2524, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.14849458896148926, |
|
"learning_rate": 0.00019999977531546566, |
|
"loss": 1.2161, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.13628125499037252, |
|
"learning_rate": 0.0001999991012628722, |
|
"loss": 1.2452, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.18617698759086793, |
|
"learning_rate": 0.00019999797784524866, |
|
"loss": 1.2197, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.14416004826313944, |
|
"learning_rate": 0.00019999640506764336, |
|
"loss": 1.2796, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.13807081386834757, |
|
"learning_rate": 0.0001999943829371238, |
|
"loss": 1.2732, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.16526927436841996, |
|
"learning_rate": 0.0001999919114627769, |
|
"loss": 1.3016, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.14479672734919855, |
|
"learning_rate": 0.0001999889906557086, |
|
"loss": 1.3106, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.13829284006072087, |
|
"learning_rate": 0.00019998562052904418, |
|
"loss": 1.3355, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.13484630104616105, |
|
"learning_rate": 0.0001999818010979279, |
|
"loss": 1.1928, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.14972770674556948, |
|
"learning_rate": 0.00019997753237952317, |
|
"loss": 1.2559, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.13378525020528342, |
|
"learning_rate": 0.00019997281439301218, |
|
"loss": 1.2673, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.13242998699125438, |
|
"learning_rate": 0.00019996764715959618, |
|
"loss": 1.2272, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.12938881004364342, |
|
"learning_rate": 0.00019996203070249516, |
|
"loss": 1.2035, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.13388032350164566, |
|
"learning_rate": 0.00019995596504694763, |
|
"loss": 1.2642, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.13893372222140873, |
|
"learning_rate": 0.00019994945022021082, |
|
"loss": 1.2235, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.14131710715500717, |
|
"learning_rate": 0.00019994248625156038, |
|
"loss": 1.1095, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.13448100369103572, |
|
"learning_rate": 0.0001999350731722902, |
|
"loss": 1.1879, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.13862444003216381, |
|
"learning_rate": 0.00019992721101571236, |
|
"loss": 1.2227, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.13506115547921224, |
|
"learning_rate": 0.00019991889981715698, |
|
"loss": 1.2833, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.13174857502600473, |
|
"learning_rate": 0.00019991013961397197, |
|
"loss": 1.2394, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.1290276308949748, |
|
"learning_rate": 0.00019990093044552304, |
|
"loss": 1.2659, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.1388159912078538, |
|
"learning_rate": 0.0001998912723531933, |
|
"loss": 1.3052, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.1256806205303357, |
|
"learning_rate": 0.00019988116538038325, |
|
"loss": 1.2031, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.13256850855084143, |
|
"learning_rate": 0.00019987060957251047, |
|
"loss": 1.211, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.13197363789890235, |
|
"learning_rate": 0.0001998596049770095, |
|
"loss": 1.2256, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.13277364593883098, |
|
"learning_rate": 0.00019984815164333163, |
|
"loss": 1.2174, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.13838072824574454, |
|
"learning_rate": 0.00019983624962294458, |
|
"loss": 1.3128, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.13524759737199996, |
|
"learning_rate": 0.0001998238989693323, |
|
"loss": 1.1806, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.12669987683723832, |
|
"learning_rate": 0.0001998110997379949, |
|
"loss": 1.2171, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.1461834612451898, |
|
"learning_rate": 0.00019979785198644806, |
|
"loss": 1.2231, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.13265793664862735, |
|
"learning_rate": 0.0001997841557742232, |
|
"loss": 1.1718, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.12842971557690963, |
|
"learning_rate": 0.00019977001116286674, |
|
"loss": 1.2758, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.12188365921206967, |
|
"learning_rate": 0.00019975541821594026, |
|
"loss": 1.2457, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.12679949330022622, |
|
"learning_rate": 0.00019974037699901993, |
|
"loss": 1.1825, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.12949746150357985, |
|
"learning_rate": 0.00019972488757969635, |
|
"loss": 1.2666, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.1363496149379173, |
|
"learning_rate": 0.00019970895002757413, |
|
"loss": 1.2031, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.14218340110669314, |
|
"learning_rate": 0.0001996925644142717, |
|
"loss": 1.3073, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.14234535389443218, |
|
"learning_rate": 0.00019967573081342103, |
|
"loss": 1.2444, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.12866113026310516, |
|
"learning_rate": 0.000199658449300667, |
|
"loss": 1.2257, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.1324053366295965, |
|
"learning_rate": 0.00019964071995366744, |
|
"loss": 1.2374, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.12906841330218152, |
|
"learning_rate": 0.00019962254285209254, |
|
"loss": 1.2334, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.13620873131846425, |
|
"learning_rate": 0.00019960391807762463, |
|
"loss": 1.242, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.14877366842835116, |
|
"learning_rate": 0.00019958484571395757, |
|
"loss": 1.1772, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.13914108740445985, |
|
"learning_rate": 0.00019956532584679675, |
|
"loss": 1.2734, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.13198394930310692, |
|
"learning_rate": 0.00019954535856385837, |
|
"loss": 1.1728, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.3807736597404611, |
|
"learning_rate": 0.0001995249439548693, |
|
"loss": 1.2089, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.1682550557564819, |
|
"learning_rate": 0.00019950408211156636, |
|
"loss": 1.2423, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.2102196862007261, |
|
"learning_rate": 0.0001994827731276963, |
|
"loss": 1.2096, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.154346739470422, |
|
"learning_rate": 0.00019946101709901514, |
|
"loss": 1.2847, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.16416668358293746, |
|
"learning_rate": 0.0001994388141232876, |
|
"loss": 1.2503, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.13134349458231093, |
|
"learning_rate": 0.0001994161643002871, |
|
"loss": 1.1231, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.15083246389185287, |
|
"learning_rate": 0.00019939306773179497, |
|
"loss": 1.1614, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.1742387260929692, |
|
"learning_rate": 0.00019936952452159995, |
|
"loss": 1.3568, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.18146911432436974, |
|
"learning_rate": 0.00019934553477549794, |
|
"loss": 1.2686, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.1393593447949332, |
|
"learning_rate": 0.00019932109860129154, |
|
"loss": 1.1141, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.14856124153987935, |
|
"learning_rate": 0.00019929621610878927, |
|
"loss": 1.234, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.14820851831477327, |
|
"learning_rate": 0.0001992708874098054, |
|
"loss": 1.2069, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.17893142790958147, |
|
"learning_rate": 0.00019924511261815926, |
|
"loss": 1.1278, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.14573658703265605, |
|
"learning_rate": 0.00019921889184967476, |
|
"loss": 1.2292, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.15282321197574994, |
|
"learning_rate": 0.00019919222522217996, |
|
"loss": 1.2482, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.16342112084119492, |
|
"learning_rate": 0.00019916511285550642, |
|
"loss": 1.2172, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.1475889153814455, |
|
"learning_rate": 0.00019913755487148876, |
|
"loss": 1.1747, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.163738064491857, |
|
"learning_rate": 0.00019910955139396396, |
|
"loss": 1.3007, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.14427856196022704, |
|
"learning_rate": 0.00019908110254877106, |
|
"loss": 1.2464, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.20204742660246344, |
|
"learning_rate": 0.00019905220846375032, |
|
"loss": 1.2515, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.15134144918251685, |
|
"learning_rate": 0.0001990228692687429, |
|
"loss": 1.1786, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.1636590177812163, |
|
"learning_rate": 0.00019899308509558998, |
|
"loss": 1.1974, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.15552319776955892, |
|
"learning_rate": 0.00019896285607813244, |
|
"loss": 1.2308, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.17104898009833774, |
|
"learning_rate": 0.00019893218235221015, |
|
"loss": 1.2828, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.16387378763964267, |
|
"learning_rate": 0.00019890106405566138, |
|
"loss": 1.2779, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.14622126798612248, |
|
"learning_rate": 0.00019886950132832207, |
|
"loss": 1.2894, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.16619841547518147, |
|
"learning_rate": 0.0001988374943120254, |
|
"loss": 1.2133, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.12664832399697545, |
|
"learning_rate": 0.00019880504315060096, |
|
"loss": 1.1807, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.2015108381613456, |
|
"learning_rate": 0.00019877214798987426, |
|
"loss": 1.1876, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.14468620723711506, |
|
"learning_rate": 0.00019873880897766598, |
|
"loss": 1.1883, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.1549018650770757, |
|
"learning_rate": 0.00019870502626379127, |
|
"loss": 1.2896, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.1492917963684983, |
|
"learning_rate": 0.0001986708000000593, |
|
"loss": 1.2102, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.178606606459489, |
|
"learning_rate": 0.00019863613034027224, |
|
"loss": 1.2292, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.206170239681528, |
|
"learning_rate": 0.00019860101744022485, |
|
"loss": 1.2666, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.13741043007948167, |
|
"learning_rate": 0.0001985654614577036, |
|
"loss": 1.2022, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.1595080658199459, |
|
"learning_rate": 0.0001985294625524861, |
|
"loss": 1.1203, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.13929705183853777, |
|
"learning_rate": 0.00019849302088634034, |
|
"loss": 1.1505, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.14045247607912964, |
|
"learning_rate": 0.00019845613662302383, |
|
"loss": 1.1897, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.15002651347444407, |
|
"learning_rate": 0.00019841880992828306, |
|
"loss": 1.2133, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.1567929487810952, |
|
"learning_rate": 0.00019838104096985267, |
|
"loss": 1.129, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.15240634543877116, |
|
"learning_rate": 0.00019834282991745464, |
|
"loss": 1.1995, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.151807679821367, |
|
"learning_rate": 0.00019830417694279766, |
|
"loss": 1.25, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.1648599156208311, |
|
"learning_rate": 0.0001982650822195762, |
|
"loss": 1.2511, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.15363401233808713, |
|
"learning_rate": 0.00019822554592346993, |
|
"loss": 1.1794, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.1569644350778875, |
|
"learning_rate": 0.00019818556823214268, |
|
"loss": 1.2033, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.15996552747294254, |
|
"learning_rate": 0.0001981451493252418, |
|
"loss": 1.2809, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.15863104885072635, |
|
"learning_rate": 0.0001981042893843974, |
|
"loss": 1.1667, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.2887466971861171, |
|
"learning_rate": 0.0001980629885932214, |
|
"loss": 1.1915, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.15233015979193984, |
|
"learning_rate": 0.00019802124713730681, |
|
"loss": 1.1734, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.18207884538436447, |
|
"learning_rate": 0.00019797906520422677, |
|
"loss": 1.2575, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.17323546756038308, |
|
"learning_rate": 0.0001979364429835339, |
|
"loss": 1.1704, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.14592153602263633, |
|
"learning_rate": 0.00019789338066675922, |
|
"loss": 1.192, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.19250697792287097, |
|
"learning_rate": 0.0001978498784474115, |
|
"loss": 1.2779, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.1429107680887097, |
|
"learning_rate": 0.0001978059365209762, |
|
"loss": 1.2529, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.48514081074992116, |
|
"learning_rate": 0.00019776155508491482, |
|
"loss": 1.1917, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.1534376167748161, |
|
"learning_rate": 0.0001977167343386638, |
|
"loss": 1.2384, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.16744875760032166, |
|
"learning_rate": 0.00019767147448363366, |
|
"loss": 1.1744, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.29195538170738244, |
|
"learning_rate": 0.00019762577572320824, |
|
"loss": 1.1418, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.1820804717651353, |
|
"learning_rate": 0.00019757963826274357, |
|
"loss": 1.2815, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.17522345110441973, |
|
"learning_rate": 0.00019753306230956718, |
|
"loss": 1.2363, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.16354388270886613, |
|
"learning_rate": 0.000197486048072977, |
|
"loss": 1.2845, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.17590082756401024, |
|
"learning_rate": 0.0001974385957642404, |
|
"loss": 1.192, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.17345720403188775, |
|
"learning_rate": 0.00019739070559659347, |
|
"loss": 1.2068, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.16070434867766506, |
|
"learning_rate": 0.00019734237778523976, |
|
"loss": 1.189, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.18983443066710415, |
|
"learning_rate": 0.0001972936125473495, |
|
"loss": 1.2223, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.15724400187981355, |
|
"learning_rate": 0.00019724441010205863, |
|
"loss": 1.2292, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.14570729442956004, |
|
"learning_rate": 0.00019719477067046766, |
|
"loss": 1.1421, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.1559242881177266, |
|
"learning_rate": 0.00019714469447564088, |
|
"loss": 1.2598, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.16621830243096108, |
|
"learning_rate": 0.0001970941817426052, |
|
"loss": 1.3038, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 4.675483994100576, |
|
"learning_rate": 0.00019704323269834927, |
|
"loss": 1.2298, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.2769699381619058, |
|
"learning_rate": 0.00019699184757182225, |
|
"loss": 1.2566, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.20189839889100783, |
|
"learning_rate": 0.00019694002659393305, |
|
"loss": 1.3181, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.19497107359413876, |
|
"learning_rate": 0.00019688776999754912, |
|
"loss": 1.1502, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.1982266815755412, |
|
"learning_rate": 0.00019683507801749545, |
|
"loss": 1.2053, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.1924340950322314, |
|
"learning_rate": 0.00019678195089055346, |
|
"loss": 1.2149, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.1725322346446431, |
|
"learning_rate": 0.00019672838885546008, |
|
"loss": 1.2553, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.2535488743520272, |
|
"learning_rate": 0.00019667439215290648, |
|
"loss": 1.2576, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.37837586860064026, |
|
"learning_rate": 0.00019661996102553718, |
|
"loss": 1.1815, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.17520419597901843, |
|
"learning_rate": 0.00019656509571794878, |
|
"loss": 1.1932, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.17056234784450633, |
|
"learning_rate": 0.00019650979647668906, |
|
"loss": 1.163, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.18272246580207432, |
|
"learning_rate": 0.00019645406355025565, |
|
"loss": 1.1887, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.17889037954429915, |
|
"learning_rate": 0.00019639789718909508, |
|
"loss": 1.2126, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.23993734971101424, |
|
"learning_rate": 0.00019634129764560168, |
|
"loss": 1.2485, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.1847578318208199, |
|
"learning_rate": 0.00019628426517411625, |
|
"loss": 1.2549, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.23185098827091005, |
|
"learning_rate": 0.00019622680003092503, |
|
"loss": 1.1599, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.220638044092583, |
|
"learning_rate": 0.00019616890247425866, |
|
"loss": 1.2281, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.2303439219825616, |
|
"learning_rate": 0.00019611057276429085, |
|
"loss": 1.2208, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.1744807302230573, |
|
"learning_rate": 0.00019605181116313724, |
|
"loss": 1.2303, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.17510946821872422, |
|
"learning_rate": 0.0001959926179348543, |
|
"loss": 1.2385, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.2218474349751746, |
|
"learning_rate": 0.00019593299334543808, |
|
"loss": 1.2153, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.1742070481516402, |
|
"learning_rate": 0.00019587293766282308, |
|
"loss": 1.1628, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.15250311715180823, |
|
"learning_rate": 0.00019581245115688094, |
|
"loss": 1.1632, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.1744397677094501, |
|
"learning_rate": 0.0001957515340994193, |
|
"loss": 1.254, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.1686772182789891, |
|
"learning_rate": 0.00019569018676418053, |
|
"loss": 1.2169, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.16404966161017623, |
|
"learning_rate": 0.00019562840942684067, |
|
"loss": 1.2221, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.16052011449463713, |
|
"learning_rate": 0.00019556620236500793, |
|
"loss": 1.2045, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.16343251390831215, |
|
"learning_rate": 0.0001955035658582216, |
|
"loss": 1.2289, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.14387162360389305, |
|
"learning_rate": 0.00019544050018795075, |
|
"loss": 1.1365, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.15304461439740238, |
|
"learning_rate": 0.00019537700563759304, |
|
"loss": 1.1931, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.17059958050065627, |
|
"learning_rate": 0.00019531308249247327, |
|
"loss": 1.2166, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.17633385530926995, |
|
"learning_rate": 0.00019524873103984235, |
|
"loss": 1.2604, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.17855814403303746, |
|
"learning_rate": 0.00019518395156887576, |
|
"loss": 1.1615, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.19823982444256988, |
|
"learning_rate": 0.00019511874437067243, |
|
"loss": 1.2153, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.1570784627362585, |
|
"learning_rate": 0.0001950531097382533, |
|
"loss": 1.2788, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.2183125402112695, |
|
"learning_rate": 0.00019498704796656018, |
|
"loss": 1.2966, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.18173933276147194, |
|
"learning_rate": 0.00019492055935245418, |
|
"loss": 1.2978, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.17483116680914407, |
|
"learning_rate": 0.00019485364419471454, |
|
"loss": 1.258, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.15490767356815494, |
|
"learning_rate": 0.0001947863027940374, |
|
"loss": 1.2088, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.14703966491934156, |
|
"learning_rate": 0.00019471853545303405, |
|
"loss": 1.2355, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.14386689086661608, |
|
"learning_rate": 0.00019465034247623003, |
|
"loss": 1.2583, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.18818904376313625, |
|
"learning_rate": 0.00019458172417006347, |
|
"loss": 1.2181, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.17393313719202513, |
|
"learning_rate": 0.00019451268084288385, |
|
"loss": 1.3453, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.14706823379985753, |
|
"learning_rate": 0.00019444321280495043, |
|
"loss": 1.2234, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.15282014755252687, |
|
"learning_rate": 0.00019437332036843118, |
|
"loss": 1.1262, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.1618727884326225, |
|
"learning_rate": 0.00019430300384740105, |
|
"loss": 1.3136, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.16090758705378874, |
|
"learning_rate": 0.00019423226355784077, |
|
"loss": 1.2055, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.15241156801091013, |
|
"learning_rate": 0.00019416109981763526, |
|
"loss": 1.2678, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.14216697909809062, |
|
"learning_rate": 0.0001940895129465724, |
|
"loss": 1.2841, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.15790232415414485, |
|
"learning_rate": 0.00019401750326634144, |
|
"loss": 1.3119, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.13322691961062616, |
|
"learning_rate": 0.0001939450711005316, |
|
"loss": 1.1293, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.14075018938835404, |
|
"learning_rate": 0.00019387221677463062, |
|
"loss": 1.2176, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.21565975459393052, |
|
"learning_rate": 0.00019379894061602335, |
|
"loss": 1.1723, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.17967631394222838, |
|
"learning_rate": 0.00019372524295399013, |
|
"loss": 1.239, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.21187969201978435, |
|
"learning_rate": 0.0001936511241197055, |
|
"loss": 1.2207, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.16967789022974608, |
|
"learning_rate": 0.00019357658444623654, |
|
"loss": 1.2478, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.14810621660374448, |
|
"learning_rate": 0.0001935016242685415, |
|
"loss": 1.1223, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.1489106421847434, |
|
"learning_rate": 0.00019342624392346824, |
|
"loss": 1.1592, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.17625176068748855, |
|
"learning_rate": 0.0001933504437497527, |
|
"loss": 1.2145, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.17250255512763446, |
|
"learning_rate": 0.00019327422408801744, |
|
"loss": 1.2504, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.16079375745566896, |
|
"learning_rate": 0.00019319758528077, |
|
"loss": 1.1795, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.15454466809245995, |
|
"learning_rate": 0.0001931205276724015, |
|
"loss": 1.2123, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.7021323604447972, |
|
"learning_rate": 0.000193043051609185, |
|
"loss": 1.2239, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.1572764339385847, |
|
"learning_rate": 0.00019296515743927399, |
|
"loss": 1.2516, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.2136637778252246, |
|
"learning_rate": 0.00019288684551270073, |
|
"loss": 1.2321, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.4546540454773654, |
|
"learning_rate": 0.00019280811618137484, |
|
"loss": 1.18, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.9809832576786297, |
|
"learning_rate": 0.00019272896979908154, |
|
"loss": 1.2081, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.5246256133291822, |
|
"learning_rate": 0.00019264940672148018, |
|
"loss": 1.2722, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.24941717134878091, |
|
"learning_rate": 0.00019256942730610268, |
|
"loss": 1.2352, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.3356068462072784, |
|
"learning_rate": 0.00019248903191235176, |
|
"loss": 1.2225, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.19535845221880543, |
|
"learning_rate": 0.00019240822090149944, |
|
"loss": 1.1669, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.22306941566416597, |
|
"learning_rate": 0.00019232699463668542, |
|
"loss": 1.2281, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.2700134013989352, |
|
"learning_rate": 0.00019224535348291542, |
|
"loss": 1.1939, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.24406908935562743, |
|
"learning_rate": 0.00019216329780705953, |
|
"loss": 1.1839, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.20465183000217488, |
|
"learning_rate": 0.00019208082797785055, |
|
"loss": 1.2277, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.21324820828129784, |
|
"learning_rate": 0.00019199794436588243, |
|
"loss": 1.2072, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.1780562512431263, |
|
"learning_rate": 0.00019191464734360844, |
|
"loss": 1.2082, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.16547971467615655, |
|
"learning_rate": 0.00019183093728533966, |
|
"loss": 1.1978, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.22904664933247196, |
|
"learning_rate": 0.00019174681456724318, |
|
"loss": 1.1562, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.1737397860007602, |
|
"learning_rate": 0.00019166227956734052, |
|
"loss": 1.2383, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.1589465455917568, |
|
"learning_rate": 0.00019157733266550575, |
|
"loss": 1.2158, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.16253126221999709, |
|
"learning_rate": 0.00019149197424346405, |
|
"loss": 1.1952, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.22436676243032663, |
|
"learning_rate": 0.00019140620468478968, |
|
"loss": 1.2315, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.19291682612950423, |
|
"learning_rate": 0.00019132002437490458, |
|
"loss": 1.2283, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.1519191258459668, |
|
"learning_rate": 0.00019123343370107637, |
|
"loss": 1.1151, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.17179909633547025, |
|
"learning_rate": 0.00019114643305241676, |
|
"loss": 1.1576, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.17992599023321432, |
|
"learning_rate": 0.00019105902281987976, |
|
"loss": 1.2592, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.17714099390314453, |
|
"learning_rate": 0.00019097120339625994, |
|
"loss": 1.2578, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.2455577642687935, |
|
"learning_rate": 0.00019088297517619055, |
|
"loss": 1.2361, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.18398518628783986, |
|
"learning_rate": 0.00019079433855614201, |
|
"loss": 1.1906, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.18944067022821645, |
|
"learning_rate": 0.00019070529393441985, |
|
"loss": 1.237, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.17639967519781063, |
|
"learning_rate": 0.00019061584171116303, |
|
"loss": 1.1841, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.15947129998283005, |
|
"learning_rate": 0.00019052598228834217, |
|
"loss": 1.1722, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.1693354353719105, |
|
"learning_rate": 0.00019043571606975777, |
|
"loss": 1.2204, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.16236190451963983, |
|
"learning_rate": 0.00019034504346103823, |
|
"loss": 1.1778, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.17702370729269964, |
|
"learning_rate": 0.00019025396486963827, |
|
"loss": 1.2065, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.19388150596154238, |
|
"learning_rate": 0.00019016248070483687, |
|
"loss": 1.2942, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.16152000400319103, |
|
"learning_rate": 0.0001900705913777356, |
|
"loss": 1.1784, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.1545267913996029, |
|
"learning_rate": 0.00018997829730125663, |
|
"loss": 1.1829, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.15421727704318197, |
|
"learning_rate": 0.000189885598890141, |
|
"loss": 1.177, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.1624966073814206, |
|
"learning_rate": 0.00018979249656094673, |
|
"loss": 1.2439, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.9490737312904575, |
|
"learning_rate": 0.00018969899073204686, |
|
"loss": 1.2085, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.8982903208613089, |
|
"learning_rate": 0.00018960508182362768, |
|
"loss": 1.2347, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.3771428474797688, |
|
"learning_rate": 0.00018951077025768678, |
|
"loss": 1.2546, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.4776152950069111, |
|
"learning_rate": 0.00018941605645803115, |
|
"loss": 1.2904, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.18786943849618057, |
|
"learning_rate": 0.00018932094085027533, |
|
"loss": 1.2122, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 1.6297025984167128, |
|
"learning_rate": 0.0001892254238618394, |
|
"loss": 1.171, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.20382660707264952, |
|
"learning_rate": 0.0001891295059219472, |
|
"loss": 1.1874, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 1.3580819775908755, |
|
"learning_rate": 0.00018903318746162429, |
|
"loss": 1.1531, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.43619056173016185, |
|
"learning_rate": 0.00018893646891369602, |
|
"loss": 1.2289, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.29385240705823723, |
|
"learning_rate": 0.0001888393507127856, |
|
"loss": 1.2073, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.3136086850525623, |
|
"learning_rate": 0.00018874183329531223, |
|
"loss": 1.1898, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.2307767217662562, |
|
"learning_rate": 0.000188643917099489, |
|
"loss": 1.207, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.18703654518135468, |
|
"learning_rate": 0.000188545602565321, |
|
"loss": 1.1688, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.4809351333934126, |
|
"learning_rate": 0.00018844689013460336, |
|
"loss": 1.2519, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.40370101428544464, |
|
"learning_rate": 0.0001883477802509192, |
|
"loss": 1.2411, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.2858848636432859, |
|
"learning_rate": 0.00018824827335963765, |
|
"loss": 1.194, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.32195602638999565, |
|
"learning_rate": 0.000188148369907912, |
|
"loss": 1.0988, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.23790306908901832, |
|
"learning_rate": 0.00018804807034467733, |
|
"loss": 1.2237, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.20126988767112128, |
|
"learning_rate": 0.0001879473751206489, |
|
"loss": 1.2731, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.3336380339194037, |
|
"learning_rate": 0.00018784628468831996, |
|
"loss": 1.2369, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.5054330893305989, |
|
"learning_rate": 0.0001877447995019596, |
|
"loss": 1.2443, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.2297866279715136, |
|
"learning_rate": 0.0001876429200176108, |
|
"loss": 1.2376, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.39350567174184636, |
|
"learning_rate": 0.00018754064669308858, |
|
"loss": 1.2126, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.2025361091435325, |
|
"learning_rate": 0.00018743797998797753, |
|
"loss": 1.2224, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.31824903419753814, |
|
"learning_rate": 0.00018733492036363005, |
|
"loss": 1.2942, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 2.4642066748643017, |
|
"learning_rate": 0.00018723146828316428, |
|
"loss": 1.2515, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.7833055646295342, |
|
"learning_rate": 0.00018712762421146183, |
|
"loss": 1.2207, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.2810249021786599, |
|
"learning_rate": 0.00018702338861516587, |
|
"loss": 1.2755, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.460995724241333, |
|
"learning_rate": 0.0001869187619626789, |
|
"loss": 1.2856, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.49139203044984286, |
|
"learning_rate": 0.00018681374472416073, |
|
"loss": 1.2392, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 2.691604613969173, |
|
"learning_rate": 0.0001867083373715264, |
|
"loss": 1.2992, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.8014112047318501, |
|
"learning_rate": 0.00018660254037844388, |
|
"loss": 1.2683, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.31614342841331383, |
|
"learning_rate": 0.00018649635422033215, |
|
"loss": 1.2356, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.2559855196513244, |
|
"learning_rate": 0.000186389779374359, |
|
"loss": 1.2053, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 1.6613999986014714, |
|
"learning_rate": 0.0001862828163194388, |
|
"loss": 1.2568, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.27190082167109786, |
|
"learning_rate": 0.0001861754655362304, |
|
"loss": 1.1288, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.43819582203066043, |
|
"learning_rate": 0.00018606772750713504, |
|
"loss": 1.1758, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.41738497400383384, |
|
"learning_rate": 0.0001859596027162941, |
|
"loss": 1.2993, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.2595142634740817, |
|
"learning_rate": 0.000185851091649587, |
|
"loss": 1.269, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.2795314201020271, |
|
"learning_rate": 0.00018574219479462878, |
|
"loss": 1.1915, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.2502992494749938, |
|
"learning_rate": 0.00018563291264076835, |
|
"loss": 1.2157, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.27422512335538374, |
|
"learning_rate": 0.00018552324567908585, |
|
"loss": 1.2541, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 1.3360989016060905, |
|
"learning_rate": 0.00018541319440239066, |
|
"loss": 1.2666, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 1.9441131913572127, |
|
"learning_rate": 0.00018530275930521924, |
|
"loss": 1.2924, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.3772373301771213, |
|
"learning_rate": 0.00018519194088383273, |
|
"loss": 1.1952, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.22091753616251295, |
|
"learning_rate": 0.0001850807396362148, |
|
"loss": 1.1858, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.21423504993321807, |
|
"learning_rate": 0.00018496915606206951, |
|
"loss": 1.2245, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.5238946238105926, |
|
"learning_rate": 0.00018485719066281892, |
|
"loss": 1.2351, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.3037858949309141, |
|
"learning_rate": 0.0001847448439416009, |
|
"loss": 1.1669, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.21553286799952254, |
|
"learning_rate": 0.00018463211640326686, |
|
"loss": 1.1454, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.202875547805464, |
|
"learning_rate": 0.0001845190085543795, |
|
"loss": 1.188, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.24385408620619278, |
|
"learning_rate": 0.00018440552090321047, |
|
"loss": 1.2307, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.23793944272430378, |
|
"learning_rate": 0.0001842916539597382, |
|
"loss": 1.2253, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.17062488448810784, |
|
"learning_rate": 0.0001841774082356455, |
|
"loss": 1.2681, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.20003742001916064, |
|
"learning_rate": 0.00018406278424431736, |
|
"loss": 1.2428, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.2696052831337752, |
|
"learning_rate": 0.0001839477825008385, |
|
"loss": 1.2945, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.23302960820538443, |
|
"learning_rate": 0.00018383240352199117, |
|
"loss": 1.1718, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.38187833239777536, |
|
"learning_rate": 0.00018371664782625287, |
|
"loss": 1.2311, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.4052561772533732, |
|
"learning_rate": 0.00018360051593379383, |
|
"loss": 1.1639, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.23379763821020377, |
|
"learning_rate": 0.0001834840083664749, |
|
"loss": 1.1809, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.2368414607613928, |
|
"learning_rate": 0.00018336712564784503, |
|
"loss": 1.2357, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.20230633988510938, |
|
"learning_rate": 0.000183249868303139, |
|
"loss": 1.1851, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.170513157244292, |
|
"learning_rate": 0.00018313223685927505, |
|
"loss": 1.205, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.18082295035256266, |
|
"learning_rate": 0.0001830142318448525, |
|
"loss": 1.2305, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.18286299264146286, |
|
"learning_rate": 0.00018289585379014942, |
|
"loss": 1.23, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.17868104103482751, |
|
"learning_rate": 0.00018277710322712012, |
|
"loss": 1.2894, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.1820411127336495, |
|
"learning_rate": 0.00018265798068939294, |
|
"loss": 1.2395, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.1738237541783663, |
|
"learning_rate": 0.0001825384867122677, |
|
"loss": 1.1576, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.15693445967795147, |
|
"learning_rate": 0.0001824186218327134, |
|
"loss": 1.0809, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.18509145652208978, |
|
"learning_rate": 0.00018229838658936564, |
|
"loss": 1.2717, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.14702488366564262, |
|
"learning_rate": 0.0001821777815225245, |
|
"loss": 1.2236, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.1828399354418095, |
|
"learning_rate": 0.00018205680717415187, |
|
"loss": 1.2565, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.17460984182013486, |
|
"learning_rate": 0.00018193546408786898, |
|
"loss": 1.2474, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.2001623109673152, |
|
"learning_rate": 0.00018181375280895416, |
|
"loss": 1.2544, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.17228631742863837, |
|
"learning_rate": 0.00018169167388434025, |
|
"loss": 1.1851, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.1644862232819482, |
|
"learning_rate": 0.00018156922786261216, |
|
"loss": 1.1817, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.19775186397477057, |
|
"learning_rate": 0.00018144641529400446, |
|
"loss": 1.257, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.1626281991220394, |
|
"learning_rate": 0.00018132323673039885, |
|
"loss": 1.2277, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.16158256707311264, |
|
"learning_rate": 0.00018119969272532166, |
|
"loss": 1.1624, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.17705809207051687, |
|
"learning_rate": 0.00018107578383394146, |
|
"loss": 1.2421, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.17639060401882287, |
|
"learning_rate": 0.00018095151061306645, |
|
"loss": 1.285, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.16918796486576196, |
|
"learning_rate": 0.00018082687362114212, |
|
"loss": 1.2606, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.15968377185965665, |
|
"learning_rate": 0.0001807018734182485, |
|
"loss": 1.194, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.17537027967397978, |
|
"learning_rate": 0.00018057651056609784, |
|
"loss": 1.1594, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.15753665403127565, |
|
"learning_rate": 0.00018045078562803203, |
|
"loss": 1.1382, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.17121200763916436, |
|
"learning_rate": 0.00018032469916902003, |
|
"loss": 1.2286, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.19120510133331003, |
|
"learning_rate": 0.00018019825175565542, |
|
"loss": 1.2835, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.1671735980123817, |
|
"learning_rate": 0.0001800714439561538, |
|
"loss": 1.2201, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.1579098534969056, |
|
"learning_rate": 0.00017994427634035015, |
|
"loss": 1.2156, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.1746075421158512, |
|
"learning_rate": 0.00017981674947969636, |
|
"loss": 1.2049, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.16878182886737042, |
|
"learning_rate": 0.00017968886394725874, |
|
"loss": 1.2204, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.16725956538286493, |
|
"learning_rate": 0.00017956062031771535, |
|
"loss": 1.2091, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.18877845951705005, |
|
"learning_rate": 0.00017943201916735335, |
|
"loss": 1.241, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.180337447476004, |
|
"learning_rate": 0.00017930306107406653, |
|
"loss": 1.2253, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.16688572366717752, |
|
"learning_rate": 0.0001791737466173527, |
|
"loss": 1.239, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.15385917621135983, |
|
"learning_rate": 0.00017904407637831099, |
|
"loss": 1.2476, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.17725645269055587, |
|
"learning_rate": 0.00017891405093963938, |
|
"loss": 1.2599, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.14758551718901028, |
|
"learning_rate": 0.00017878367088563195, |
|
"loss": 1.2249, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.15216962408661316, |
|
"learning_rate": 0.00017865293680217637, |
|
"loss": 1.2346, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.16679282848599514, |
|
"learning_rate": 0.00017852184927675112, |
|
"loss": 1.2443, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.16723562739069214, |
|
"learning_rate": 0.00017839040889842305, |
|
"loss": 1.224, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.15922276239929914, |
|
"learning_rate": 0.00017825861625784455, |
|
"loss": 1.2739, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.1510107938469514, |
|
"learning_rate": 0.00017812647194725094, |
|
"loss": 1.1764, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.16446999054333494, |
|
"learning_rate": 0.00017799397656045792, |
|
"loss": 1.2498, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.18566301651865832, |
|
"learning_rate": 0.00017786113069285874, |
|
"loss": 1.232, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.20592971655306183, |
|
"learning_rate": 0.00017772793494142167, |
|
"loss": 1.1586, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.1581947714375729, |
|
"learning_rate": 0.00017759438990468725, |
|
"loss": 1.2502, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.15466760695169174, |
|
"learning_rate": 0.00017746049618276545, |
|
"loss": 1.1605, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.16041506222444918, |
|
"learning_rate": 0.00017732625437733335, |
|
"loss": 1.2778, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.17168109661676773, |
|
"learning_rate": 0.0001771916650916321, |
|
"loss": 1.262, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.1788973186498254, |
|
"learning_rate": 0.00017705672893046425, |
|
"loss": 1.2111, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.1759644359346382, |
|
"learning_rate": 0.00017692144650019125, |
|
"loss": 1.2546, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.15710749736088767, |
|
"learning_rate": 0.0001767858184087304, |
|
"loss": 1.2487, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.1648235522911144, |
|
"learning_rate": 0.00017664984526555248, |
|
"loss": 1.2469, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.15452607969890703, |
|
"learning_rate": 0.0001765135276816787, |
|
"loss": 1.1855, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.1837695597880219, |
|
"learning_rate": 0.00017637686626967812, |
|
"loss": 1.2185, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.15861390725762364, |
|
"learning_rate": 0.00017623986164366486, |
|
"loss": 1.2056, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.1663260460966887, |
|
"learning_rate": 0.00017610251441929533, |
|
"loss": 1.1242, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.1803309720529981, |
|
"learning_rate": 0.00017596482521376546, |
|
"loss": 1.2938, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.14909085011764342, |
|
"learning_rate": 0.00017582679464580797, |
|
"loss": 1.1953, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.15779022242482527, |
|
"learning_rate": 0.00017568842333568952, |
|
"loss": 1.2792, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.1553327313967345, |
|
"learning_rate": 0.00017554971190520798, |
|
"loss": 1.2286, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.16363964666273684, |
|
"learning_rate": 0.00017541066097768963, |
|
"loss": 1.2753, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.16668099163659675, |
|
"learning_rate": 0.00017527127117798635, |
|
"loss": 1.185, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.13957568397594883, |
|
"learning_rate": 0.0001751315431324727, |
|
"loss": 1.143, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.1553111736740035, |
|
"learning_rate": 0.00017499147746904335, |
|
"loss": 1.2492, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.1691517335818193, |
|
"learning_rate": 0.00017485107481711012, |
|
"loss": 1.2619, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.15480883994395986, |
|
"learning_rate": 0.00017471033580759903, |
|
"loss": 1.2396, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.1451690143792058, |
|
"learning_rate": 0.00017456926107294765, |
|
"loss": 1.1732, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.1524398957482947, |
|
"learning_rate": 0.00017442785124710227, |
|
"loss": 1.2083, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.16790264977550012, |
|
"learning_rate": 0.0001742861069655148, |
|
"loss": 1.2201, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.1529847047636337, |
|
"learning_rate": 0.0001741440288651403, |
|
"loss": 1.243, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.1485875402374676, |
|
"learning_rate": 0.00017400161758443375, |
|
"loss": 1.2053, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.16950094279079617, |
|
"learning_rate": 0.00017385887376334742, |
|
"loss": 1.1944, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.15289337084330445, |
|
"learning_rate": 0.00017371579804332789, |
|
"loss": 1.2503, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.15337063655317973, |
|
"learning_rate": 0.00017357239106731317, |
|
"loss": 1.3092, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.1458937961897621, |
|
"learning_rate": 0.00017342865347972988, |
|
"loss": 1.2244, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.19897118610161338, |
|
"learning_rate": 0.00017328458592649027, |
|
"loss": 1.2238, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.15850805264911003, |
|
"learning_rate": 0.00017314018905498931, |
|
"loss": 1.195, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.14445183074519347, |
|
"learning_rate": 0.00017299546351410197, |
|
"loss": 1.1974, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.18180731722745677, |
|
"learning_rate": 0.00017285040995418, |
|
"loss": 1.2107, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.14943874953193587, |
|
"learning_rate": 0.00017270502902704926, |
|
"loss": 1.1843, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.15767466790910512, |
|
"learning_rate": 0.00017255932138600665, |
|
"loss": 1.1409, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.16402921378654775, |
|
"learning_rate": 0.00017241328768581726, |
|
"loss": 1.2135, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.15526246786505485, |
|
"learning_rate": 0.00017226692858271134, |
|
"loss": 1.2255, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.16608155892622348, |
|
"learning_rate": 0.00017212024473438147, |
|
"loss": 1.2691, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.14913271520144072, |
|
"learning_rate": 0.00017197323679997943, |
|
"loss": 1.1574, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.1471910610421707, |
|
"learning_rate": 0.00017182590544011347, |
|
"loss": 1.2774, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.1417464185073962, |
|
"learning_rate": 0.00017167825131684513, |
|
"loss": 1.2446, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.1610488125634495, |
|
"learning_rate": 0.0001715302750936864, |
|
"loss": 1.2862, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.20227974555123074, |
|
"learning_rate": 0.00017138197743559654, |
|
"loss": 1.207, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.1355502559749413, |
|
"learning_rate": 0.00017123335900897946, |
|
"loss": 1.1019, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.1559423167028215, |
|
"learning_rate": 0.00017108442048168038, |
|
"loss": 1.2549, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.15898973818185586, |
|
"learning_rate": 0.00017093516252298296, |
|
"loss": 1.2705, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.15169569998999652, |
|
"learning_rate": 0.00017078558580360632, |
|
"loss": 1.2454, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.15976111665597925, |
|
"learning_rate": 0.00017063569099570196, |
|
"loss": 1.2585, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.14488877221999352, |
|
"learning_rate": 0.00017048547877285077, |
|
"loss": 1.2169, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.14919533098974924, |
|
"learning_rate": 0.00017033494981006002, |
|
"loss": 1.2358, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.15251746717084805, |
|
"learning_rate": 0.00017018410478376032, |
|
"loss": 1.2241, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.1456060482002663, |
|
"learning_rate": 0.00017003294437180255, |
|
"loss": 1.2298, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.17048886778787248, |
|
"learning_rate": 0.00016988146925345484, |
|
"loss": 1.2707, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.15304381059310815, |
|
"learning_rate": 0.00016972968010939954, |
|
"loss": 1.1498, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.16590055969071696, |
|
"learning_rate": 0.0001695775776217301, |
|
"loss": 1.2481, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.14299575837437278, |
|
"learning_rate": 0.00016942516247394807, |
|
"loss": 1.2058, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.14275107775859475, |
|
"learning_rate": 0.00016927243535095997, |
|
"loss": 1.2178, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.1554250137491414, |
|
"learning_rate": 0.0001691193969390742, |
|
"loss": 1.1197, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.16958418467021688, |
|
"learning_rate": 0.0001689660479259981, |
|
"loss": 1.1768, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.1546216583314497, |
|
"learning_rate": 0.00016881238900083473, |
|
"loss": 1.1741, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.15287056494787424, |
|
"learning_rate": 0.0001686584208540797, |
|
"loss": 1.2328, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.1419329373337611, |
|
"learning_rate": 0.0001685041441776183, |
|
"loss": 1.1743, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.15662192296485464, |
|
"learning_rate": 0.00016834955966472213, |
|
"loss": 1.1861, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.16304778894798697, |
|
"learning_rate": 0.00016819466801004621, |
|
"loss": 1.2045, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.15690090424895087, |
|
"learning_rate": 0.00016803946990962576, |
|
"loss": 1.1553, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.15227523196027068, |
|
"learning_rate": 0.000167883966060873, |
|
"loss": 1.126, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.15442134859978873, |
|
"learning_rate": 0.00016772815716257412, |
|
"loss": 1.169, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.16235578615890994, |
|
"learning_rate": 0.00016757204391488613, |
|
"loss": 1.198, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.16631365293435893, |
|
"learning_rate": 0.00016741562701933367, |
|
"loss": 1.2191, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.15668252106453312, |
|
"learning_rate": 0.0001672589071788059, |
|
"loss": 1.1094, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.14883987789397163, |
|
"learning_rate": 0.00016710188509755329, |
|
"loss": 1.0653, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.14723412204861966, |
|
"learning_rate": 0.00016694456148118452, |
|
"loss": 1.0632, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.2694273773199372, |
|
"learning_rate": 0.00016678693703666325, |
|
"loss": 1.1577, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.3010203762394184, |
|
"learning_rate": 0.00016662901247230502, |
|
"loss": 1.1049, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.16960159731526514, |
|
"learning_rate": 0.0001664707884977739, |
|
"loss": 1.2185, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.1593257551859938, |
|
"learning_rate": 0.00016631226582407952, |
|
"loss": 1.1085, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.1695114549114674, |
|
"learning_rate": 0.00016615344516357378, |
|
"loss": 1.1863, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.17116727204552032, |
|
"learning_rate": 0.00016599432722994755, |
|
"loss": 1.0921, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.16265948588203852, |
|
"learning_rate": 0.00016583491273822765, |
|
"loss": 1.1905, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.16500105151611763, |
|
"learning_rate": 0.00016567520240477344, |
|
"loss": 1.1477, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.15445707774433456, |
|
"learning_rate": 0.00016551519694727381, |
|
"loss": 1.1247, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.17740657850033337, |
|
"learning_rate": 0.0001653548970847438, |
|
"loss": 1.0164, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.16152198390194278, |
|
"learning_rate": 0.0001651943035375214, |
|
"loss": 1.1779, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.16299638453443582, |
|
"learning_rate": 0.00016503341702726426, |
|
"loss": 1.1847, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.1486770667934901, |
|
"learning_rate": 0.00016487223827694672, |
|
"loss": 1.1002, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.15907088594322072, |
|
"learning_rate": 0.00016471076801085615, |
|
"loss": 1.1127, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.15491780431797444, |
|
"learning_rate": 0.00016454900695458998, |
|
"loss": 1.1196, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.16584586610927665, |
|
"learning_rate": 0.00016438695583505242, |
|
"loss": 1.1441, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.15197831968756778, |
|
"learning_rate": 0.00016422461538045103, |
|
"loss": 1.0532, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.15669296934233234, |
|
"learning_rate": 0.00016406198632029357, |
|
"loss": 1.1753, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.1523894998655591, |
|
"learning_rate": 0.0001638990693853848, |
|
"loss": 1.1339, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.1570057583413597, |
|
"learning_rate": 0.000163735865307823, |
|
"loss": 1.1117, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.1637341350288717, |
|
"learning_rate": 0.00016357237482099684, |
|
"loss": 1.1369, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.15703259344496112, |
|
"learning_rate": 0.0001634085986595819, |
|
"loss": 1.1448, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.1672273979359662, |
|
"learning_rate": 0.00016324453755953773, |
|
"loss": 1.1972, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.16032943590057747, |
|
"learning_rate": 0.0001630801922581041, |
|
"loss": 1.1811, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.15414179865764022, |
|
"learning_rate": 0.00016291556349379795, |
|
"loss": 1.1126, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.16133591336152153, |
|
"learning_rate": 0.00016275065200641004, |
|
"loss": 1.0664, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.17496271909044792, |
|
"learning_rate": 0.00016258545853700158, |
|
"loss": 1.2042, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.16189829224809252, |
|
"learning_rate": 0.00016241998382790095, |
|
"loss": 1.1896, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.1508460775909727, |
|
"learning_rate": 0.00016225422862270027, |
|
"loss": 1.1205, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.16848110855261456, |
|
"learning_rate": 0.00016208819366625218, |
|
"loss": 1.1132, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.15996856299765072, |
|
"learning_rate": 0.00016192187970466644, |
|
"loss": 1.1022, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.15697088450293906, |
|
"learning_rate": 0.0001617552874853065, |
|
"loss": 1.0483, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.16338270251123885, |
|
"learning_rate": 0.0001615884177567863, |
|
"loss": 1.1049, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.1605597226615145, |
|
"learning_rate": 0.0001614212712689668, |
|
"loss": 1.1319, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.1694169270615921, |
|
"learning_rate": 0.00016125384877295257, |
|
"loss": 1.2143, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.16306866904399633, |
|
"learning_rate": 0.00016108615102108855, |
|
"loss": 1.164, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.1646758516051161, |
|
"learning_rate": 0.00016091817876695655, |
|
"loss": 1.2176, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.16334394129553162, |
|
"learning_rate": 0.00016074993276537198, |
|
"loss": 1.1526, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.16103491072240317, |
|
"learning_rate": 0.00016058141377238026, |
|
"loss": 1.1022, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.16029258404100036, |
|
"learning_rate": 0.00016041262254525362, |
|
"loss": 1.1508, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.15686217802450533, |
|
"learning_rate": 0.00016024355984248768, |
|
"loss": 1.1471, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.1520812354157866, |
|
"learning_rate": 0.0001600742264237979, |
|
"loss": 1.0994, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.161954690104593, |
|
"learning_rate": 0.0001599046230501163, |
|
"loss": 1.1475, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.16470046818858164, |
|
"learning_rate": 0.00015973475048358795, |
|
"loss": 1.1425, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.17783426542152342, |
|
"learning_rate": 0.00015956460948756765, |
|
"loss": 1.1933, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.20834095672163827, |
|
"learning_rate": 0.0001593942008266164, |
|
"loss": 1.1545, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.1601922371369916, |
|
"learning_rate": 0.00015922352526649803, |
|
"loss": 1.1626, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.17590795577463478, |
|
"learning_rate": 0.00015905258357417569, |
|
"loss": 1.1428, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.16438364538914646, |
|
"learning_rate": 0.00015888137651780845, |
|
"loss": 1.1237, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.1532404311365829, |
|
"learning_rate": 0.00015870990486674792, |
|
"loss": 1.1292, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.1658092976448754, |
|
"learning_rate": 0.0001585381693915346, |
|
"loss": 1.2051, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.17009266166893572, |
|
"learning_rate": 0.00015836617086389468, |
|
"loss": 1.1574, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.17117032466097273, |
|
"learning_rate": 0.00015819391005673626, |
|
"loss": 1.1715, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.16766945870038483, |
|
"learning_rate": 0.00015802138774414622, |
|
"loss": 1.1322, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.16603674100502688, |
|
"learning_rate": 0.00015784860470138633, |
|
"loss": 1.1682, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.1538296997751752, |
|
"learning_rate": 0.00015767556170489025, |
|
"loss": 1.0979, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.16355862351540293, |
|
"learning_rate": 0.00015750225953225968, |
|
"loss": 1.0859, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.17100029950170442, |
|
"learning_rate": 0.00015732869896226094, |
|
"loss": 1.1513, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.16981502045992727, |
|
"learning_rate": 0.0001571548807748215, |
|
"loss": 1.1665, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.15981257682597005, |
|
"learning_rate": 0.00015698080575102661, |
|
"loss": 1.0862, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.17114664647857297, |
|
"learning_rate": 0.00015680647467311557, |
|
"loss": 1.1883, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.16818704328900685, |
|
"learning_rate": 0.00015663188832447833, |
|
"loss": 1.1345, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.16716958487866437, |
|
"learning_rate": 0.0001564570474896519, |
|
"loss": 1.1613, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.16425076739154273, |
|
"learning_rate": 0.00015628195295431697, |
|
"loss": 1.1153, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.1668267227341853, |
|
"learning_rate": 0.0001561066055052941, |
|
"loss": 1.1796, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.15748311738845247, |
|
"learning_rate": 0.00015593100593054064, |
|
"loss": 1.125, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.1578229320002536, |
|
"learning_rate": 0.00015575515501914668, |
|
"loss": 1.1751, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.17990453419427754, |
|
"learning_rate": 0.0001555790535613318, |
|
"loss": 1.2107, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.17170087096955436, |
|
"learning_rate": 0.0001554027023484416, |
|
"loss": 1.1704, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.1641233552033525, |
|
"learning_rate": 0.00015522610217294375, |
|
"loss": 1.1399, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.22816800313793714, |
|
"learning_rate": 0.00015504925382842487, |
|
"loss": 1.1374, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.1699514304776241, |
|
"learning_rate": 0.00015487215810958675, |
|
"loss": 1.1668, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.16580705156733896, |
|
"learning_rate": 0.00015469481581224272, |
|
"loss": 1.2467, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.1621295978537637, |
|
"learning_rate": 0.0001545172277333142, |
|
"loss": 1.095, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.15661235975760687, |
|
"learning_rate": 0.00015433939467082713, |
|
"loss": 1.1437, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.15913894864881573, |
|
"learning_rate": 0.00015416131742390827, |
|
"loss": 1.1682, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.17078546916822446, |
|
"learning_rate": 0.00015398299679278172, |
|
"loss": 1.1673, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.17593093163481005, |
|
"learning_rate": 0.00015380443357876518, |
|
"loss": 1.1588, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.16563306408223222, |
|
"learning_rate": 0.00015362562858426654, |
|
"loss": 1.1542, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.15824017459526013, |
|
"learning_rate": 0.0001534465826127801, |
|
"loss": 1.136, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.1555748015065671, |
|
"learning_rate": 0.00015326729646888314, |
|
"loss": 1.1744, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.15054650982383877, |
|
"learning_rate": 0.0001530877709582321, |
|
"loss": 1.0996, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.1732483344367302, |
|
"learning_rate": 0.00015290800688755907, |
|
"loss": 1.1375, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.17086276533942882, |
|
"learning_rate": 0.0001527280050646682, |
|
"loss": 1.1399, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.14791498583408913, |
|
"learning_rate": 0.00015254776629843205, |
|
"loss": 1.0112, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.16492159023612152, |
|
"learning_rate": 0.00015236729139878782, |
|
"loss": 1.1179, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.15550107018855622, |
|
"learning_rate": 0.0001521865811767339, |
|
"loss": 1.0872, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.17086993477527834, |
|
"learning_rate": 0.00015200563644432612, |
|
"loss": 1.1747, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.16832629109129904, |
|
"learning_rate": 0.0001518244580146742, |
|
"loss": 1.2623, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.15555098188988337, |
|
"learning_rate": 0.00015164304670193792, |
|
"loss": 1.1172, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.16504655864531245, |
|
"learning_rate": 0.00015146140332132358, |
|
"loss": 1.1615, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.15904615747067738, |
|
"learning_rate": 0.00015127952868908043, |
|
"loss": 1.1841, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.1664190253819244, |
|
"learning_rate": 0.00015109742362249672, |
|
"loss": 1.2258, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.17622098883404838, |
|
"learning_rate": 0.00015091508893989633, |
|
"loss": 1.2141, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.17253106779139057, |
|
"learning_rate": 0.00015073252546063493, |
|
"loss": 1.1289, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.15924181512457228, |
|
"learning_rate": 0.0001505497340050963, |
|
"loss": 1.1836, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.1755114666296529, |
|
"learning_rate": 0.00015036671539468878, |
|
"loss": 1.2109, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.1623211151951732, |
|
"learning_rate": 0.00015018347045184132, |
|
"loss": 1.1922, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.16895842955233992, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 1.2402, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.17126264186496687, |
|
"learning_rate": 0.00014981630486362435, |
|
"loss": 1.2327, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.19632534974860108, |
|
"learning_rate": 0.00014963238586818345, |
|
"loss": 1.1586, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.16708414744148634, |
|
"learning_rate": 0.00014944824384015236, |
|
"loss": 1.1766, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.1800283830073218, |
|
"learning_rate": 0.00014926387960700842, |
|
"loss": 1.0902, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.17570100482291343, |
|
"learning_rate": 0.0001490792939972275, |
|
"loss": 1.1875, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.15904497800793038, |
|
"learning_rate": 0.0001488944878402802, |
|
"loss": 1.083, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.1629549994998532, |
|
"learning_rate": 0.00014870946196662822, |
|
"loss": 1.1505, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.17495490666041913, |
|
"learning_rate": 0.00014852421720772062, |
|
"loss": 1.1107, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.1814722009088628, |
|
"learning_rate": 0.00014833875439599004, |
|
"loss": 1.2089, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.16312442272141373, |
|
"learning_rate": 0.00014815307436484898, |
|
"loss": 1.1518, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.18878089936272002, |
|
"learning_rate": 0.00014796717794868607, |
|
"loss": 1.1806, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.171347321491223, |
|
"learning_rate": 0.00014778106598286234, |
|
"loss": 1.3189, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.15561777538677532, |
|
"learning_rate": 0.00014759473930370736, |
|
"loss": 1.1071, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.16095434527629326, |
|
"learning_rate": 0.0001474081987485156, |
|
"loss": 1.173, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.1594519070011647, |
|
"learning_rate": 0.00014722144515554264, |
|
"loss": 1.1099, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.15853021638823037, |
|
"learning_rate": 0.00014703447936400134, |
|
"loss": 1.1563, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.21942200368949344, |
|
"learning_rate": 0.00014684730221405814, |
|
"loss": 1.1226, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.16328115550493413, |
|
"learning_rate": 0.00014665991454682924, |
|
"loss": 1.1878, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.16551970389762746, |
|
"learning_rate": 0.00014647231720437686, |
|
"loss": 1.055, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.17275962400330097, |
|
"learning_rate": 0.00014628451102970547, |
|
"loss": 1.1302, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.17269943862734122, |
|
"learning_rate": 0.00014609649686675785, |
|
"loss": 1.1635, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.18932844828433326, |
|
"learning_rate": 0.00014590827556041158, |
|
"loss": 1.0954, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.1670913427771278, |
|
"learning_rate": 0.00014571984795647494, |
|
"loss": 1.1336, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.1727053222121231, |
|
"learning_rate": 0.00014553121490168332, |
|
"loss": 1.1477, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.20281162440108022, |
|
"learning_rate": 0.00014534237724369534, |
|
"loss": 1.1364, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.17595782723751183, |
|
"learning_rate": 0.00014515333583108896, |
|
"loss": 1.1977, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.16755958800968615, |
|
"learning_rate": 0.00014496409151335785, |
|
"loss": 1.1315, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.17908885139779734, |
|
"learning_rate": 0.00014477464514090743, |
|
"loss": 1.1851, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.16047769884625865, |
|
"learning_rate": 0.00014458499756505116, |
|
"loss": 1.1137, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.18864957014640948, |
|
"learning_rate": 0.00014439514963800648, |
|
"loss": 1.1962, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.16074020271372416, |
|
"learning_rate": 0.00014420510221289137, |
|
"loss": 1.1448, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.17166433855606958, |
|
"learning_rate": 0.00014401485614372008, |
|
"loss": 1.0487, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.1773682331611409, |
|
"learning_rate": 0.0001438244122853996, |
|
"loss": 1.16, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.16596601643110168, |
|
"learning_rate": 0.00014363377149372584, |
|
"loss": 1.0843, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.16748936526973612, |
|
"learning_rate": 0.0001434429346253794, |
|
"loss": 1.1619, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.17600274748860967, |
|
"learning_rate": 0.00014325190253792222, |
|
"loss": 1.151, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.1721019830304793, |
|
"learning_rate": 0.0001430606760897934, |
|
"loss": 1.0609, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.172361054646693, |
|
"learning_rate": 0.00014286925614030542, |
|
"loss": 1.1777, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.1938123204770535, |
|
"learning_rate": 0.00014267764354964038, |
|
"loss": 1.1078, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.171152731950173, |
|
"learning_rate": 0.00014248583917884594, |
|
"loss": 1.2344, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.176915280903416, |
|
"learning_rate": 0.00014229384388983167, |
|
"loss": 1.1443, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.16620426221482948, |
|
"learning_rate": 0.00014210165854536494, |
|
"loss": 1.1635, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.17368168932295722, |
|
"learning_rate": 0.0001419092840090673, |
|
"loss": 1.2076, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.16396003568599715, |
|
"learning_rate": 0.0001417167211454104, |
|
"loss": 1.1798, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.17474861300401356, |
|
"learning_rate": 0.0001415239708197122, |
|
"loss": 1.1951, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.17687469132536782, |
|
"learning_rate": 0.00014133103389813302, |
|
"loss": 1.1863, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.1729350163866909, |
|
"learning_rate": 0.0001411379112476717, |
|
"loss": 1.1791, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.16183372577765218, |
|
"learning_rate": 0.0001409446037361617, |
|
"loss": 1.11, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.17468439327379603, |
|
"learning_rate": 0.0001407511122322672, |
|
"loss": 1.1459, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.1746393388604993, |
|
"learning_rate": 0.00014055743760547917, |
|
"loss": 1.1782, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.1709730768079813, |
|
"learning_rate": 0.00014036358072611147, |
|
"loss": 1.1729, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.16338226575160056, |
|
"learning_rate": 0.00014016954246529696, |
|
"loss": 1.1637, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.16522796461612693, |
|
"learning_rate": 0.00013997532369498355, |
|
"loss": 1.1786, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.16443017398263532, |
|
"learning_rate": 0.0001397809252879303, |
|
"loss": 1.1885, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.16678395716368374, |
|
"learning_rate": 0.0001395863481177036, |
|
"loss": 1.1576, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.15796662849679724, |
|
"learning_rate": 0.000139391593058673, |
|
"loss": 1.1719, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.1605406960709786, |
|
"learning_rate": 0.00013919666098600753, |
|
"loss": 1.1442, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.16720350808668913, |
|
"learning_rate": 0.00013900155277567157, |
|
"loss": 1.1231, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.16025549611436293, |
|
"learning_rate": 0.00013880626930442113, |
|
"loss": 1.0959, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.1682327263659866, |
|
"learning_rate": 0.00013861081144979974, |
|
"loss": 1.1816, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.16843691606816277, |
|
"learning_rate": 0.00013841518009013445, |
|
"loss": 1.1607, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.17753480625595375, |
|
"learning_rate": 0.0001382193761045322, |
|
"loss": 1.2085, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.175444570487372, |
|
"learning_rate": 0.0001380234003728754, |
|
"loss": 1.1883, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.17844909537485765, |
|
"learning_rate": 0.00013782725377581848, |
|
"loss": 1.2548, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.17827531614400435, |
|
"learning_rate": 0.00013763093719478358, |
|
"loss": 1.1753, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.16568782268956592, |
|
"learning_rate": 0.00013743445151195657, |
|
"loss": 1.1269, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.16229148230338714, |
|
"learning_rate": 0.00013723779761028347, |
|
"loss": 1.1233, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.18071555500117806, |
|
"learning_rate": 0.000137040976373466, |
|
"loss": 1.1691, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.16528568438217509, |
|
"learning_rate": 0.000136843988685958, |
|
"loss": 1.1044, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.15621961073605914, |
|
"learning_rate": 0.00013664683543296112, |
|
"loss": 1.0585, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.1821012365890871, |
|
"learning_rate": 0.00013644951750042114, |
|
"loss": 1.1061, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.18832728793079637, |
|
"learning_rate": 0.00013625203577502382, |
|
"loss": 1.2088, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.17821269011741103, |
|
"learning_rate": 0.00013605439114419094, |
|
"loss": 1.1076, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.15816928920099374, |
|
"learning_rate": 0.00013585658449607633, |
|
"loss": 1.057, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.1634582673036166, |
|
"learning_rate": 0.00013565861671956187, |
|
"loss": 1.2206, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.16402056158577089, |
|
"learning_rate": 0.00013546048870425356, |
|
"loss": 1.0809, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.18674575425072043, |
|
"learning_rate": 0.0001352622013404774, |
|
"loss": 1.0979, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.16768128529630819, |
|
"learning_rate": 0.00013506375551927547, |
|
"loss": 1.1813, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.17842009039615817, |
|
"learning_rate": 0.00013486515213240188, |
|
"loss": 1.1767, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.16460609654310757, |
|
"learning_rate": 0.0001346663920723188, |
|
"loss": 1.1711, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.16130836204900895, |
|
"learning_rate": 0.00013446747623219255, |
|
"loss": 1.1506, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.19357518099915463, |
|
"learning_rate": 0.00013426840550588933, |
|
"loss": 1.1454, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.16818483070815926, |
|
"learning_rate": 0.0001340691807879714, |
|
"loss": 1.105, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.17634247767232533, |
|
"learning_rate": 0.00013386980297369307, |
|
"loss": 1.1173, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.17275588243641285, |
|
"learning_rate": 0.0001336702729589965, |
|
"loss": 1.1614, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.18630211892037335, |
|
"learning_rate": 0.00013347059164050794, |
|
"loss": 1.0822, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.1610612417401003, |
|
"learning_rate": 0.0001332707599155334, |
|
"loss": 1.11, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.1710900937487699, |
|
"learning_rate": 0.00013307077868205487, |
|
"loss": 1.1556, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.16789088046962233, |
|
"learning_rate": 0.00013287064883872612, |
|
"loss": 1.166, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.17152792861267624, |
|
"learning_rate": 0.00013267037128486883, |
|
"loss": 1.2076, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.16987939988427406, |
|
"learning_rate": 0.00013246994692046836, |
|
"loss": 1.1892, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.17050807620104472, |
|
"learning_rate": 0.00013226937664616976, |
|
"loss": 1.1405, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.1619859173580362, |
|
"learning_rate": 0.00013206866136327388, |
|
"loss": 1.15, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.1693102604231561, |
|
"learning_rate": 0.00013186780197373306, |
|
"loss": 1.1108, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.15751817301555188, |
|
"learning_rate": 0.00013166679938014726, |
|
"loss": 1.1075, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.17344055609711165, |
|
"learning_rate": 0.00013146565448576004, |
|
"loss": 1.2239, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.16419033762795351, |
|
"learning_rate": 0.00013126436819445422, |
|
"loss": 1.2146, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.17241446484871606, |
|
"learning_rate": 0.00013106294141074825, |
|
"loss": 1.1808, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.16299230805525922, |
|
"learning_rate": 0.0001308613750397917, |
|
"loss": 1.1639, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.15993384139183944, |
|
"learning_rate": 0.00013065966998736155, |
|
"loss": 1.0664, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.17951699730636844, |
|
"learning_rate": 0.00013045782715985792, |
|
"loss": 1.1565, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.16420615622643212, |
|
"learning_rate": 0.0001302558474643, |
|
"loss": 1.1675, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.1718558627273243, |
|
"learning_rate": 0.0001300537318083221, |
|
"loss": 1.1758, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.16040004880583997, |
|
"learning_rate": 0.00012985148110016947, |
|
"loss": 1.1286, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.1644924005506398, |
|
"learning_rate": 0.0001296490962486942, |
|
"loss": 1.0901, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.17508769995003778, |
|
"learning_rate": 0.00012944657816335123, |
|
"loss": 1.158, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.16636235063611218, |
|
"learning_rate": 0.0001292439277541942, |
|
"loss": 1.1506, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.1651752039403403, |
|
"learning_rate": 0.00012904114593187136, |
|
"loss": 1.1167, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.1727449182027404, |
|
"learning_rate": 0.0001288382336076215, |
|
"loss": 1.1208, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.1678888505155555, |
|
"learning_rate": 0.00012863519169326984, |
|
"loss": 1.1164, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.1717914946350261, |
|
"learning_rate": 0.0001284320211012239, |
|
"loss": 1.174, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.15833038102003086, |
|
"learning_rate": 0.00012822872274446958, |
|
"loss": 1.0557, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.16693919668597118, |
|
"learning_rate": 0.00012802529753656668, |
|
"loss": 1.2202, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.1608034670310254, |
|
"learning_rate": 0.0001278217463916453, |
|
"loss": 1.0761, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.16522372158185394, |
|
"learning_rate": 0.0001276180702244012, |
|
"loss": 1.114, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.15699403016449, |
|
"learning_rate": 0.00012741426995009213, |
|
"loss": 1.0487, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.1724966872571711, |
|
"learning_rate": 0.00012721034648453353, |
|
"loss": 1.2198, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.16612867386952304, |
|
"learning_rate": 0.00012700630074409427, |
|
"loss": 1.1481, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.1755115998846381, |
|
"learning_rate": 0.0001268021336456929, |
|
"loss": 1.1295, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.16610612624318807, |
|
"learning_rate": 0.00012659784610679318, |
|
"loss": 1.0735, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.17210665028457767, |
|
"learning_rate": 0.0001263934390454001, |
|
"loss": 1.128, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.17115180879996908, |
|
"learning_rate": 0.00012618891338005573, |
|
"loss": 1.2114, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.1708121456807184, |
|
"learning_rate": 0.0001259842700298352, |
|
"loss": 1.2037, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.16413369503599784, |
|
"learning_rate": 0.00012577950991434248, |
|
"loss": 1.115, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.16575779512688402, |
|
"learning_rate": 0.0001255746339537061, |
|
"loss": 1.13, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.16787768159350958, |
|
"learning_rate": 0.00012536964306857526, |
|
"loss": 1.1187, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.161257935533138, |
|
"learning_rate": 0.00012516453818011566, |
|
"loss": 1.1272, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.17163817923820932, |
|
"learning_rate": 0.00012495932021000517, |
|
"loss": 1.1371, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.16451801686420228, |
|
"learning_rate": 0.0001247539900804299, |
|
"loss": 1.1234, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.4897545674498291, |
|
"learning_rate": 0.00012454854871407994, |
|
"loss": 1.1276, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.1674813107615878, |
|
"learning_rate": 0.00012434299703414524, |
|
"loss": 1.0717, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.16889265391095676, |
|
"learning_rate": 0.0001241373359643114, |
|
"loss": 1.1094, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.17807200243744, |
|
"learning_rate": 0.0001239315664287558, |
|
"loss": 1.2072, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.1821416845887793, |
|
"learning_rate": 0.00012372568935214298, |
|
"loss": 1.1574, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.16787645348154986, |
|
"learning_rate": 0.00012351970565962085, |
|
"loss": 1.0884, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.17767787321450104, |
|
"learning_rate": 0.00012331361627681645, |
|
"loss": 1.1561, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.16990542040230855, |
|
"learning_rate": 0.00012310742212983167, |
|
"loss": 1.1454, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.17040623149046621, |
|
"learning_rate": 0.00012290112414523928, |
|
"loss": 1.1143, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.17383642491755713, |
|
"learning_rate": 0.00012269472325007858, |
|
"loss": 1.1502, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.1670271756098212, |
|
"learning_rate": 0.00012248822037185138, |
|
"loss": 1.1872, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.17338597290651167, |
|
"learning_rate": 0.0001222816164385177, |
|
"loss": 1.201, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.16342647338148947, |
|
"learning_rate": 0.00012207491237849172, |
|
"loss": 1.0994, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.15847559756714308, |
|
"learning_rate": 0.0001218681091206376, |
|
"loss": 1.1272, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.17295023549456498, |
|
"learning_rate": 0.00012166120759426514, |
|
"loss": 1.1026, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.17574482838436642, |
|
"learning_rate": 0.00012145420872912585, |
|
"loss": 1.1785, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.1636907763577917, |
|
"learning_rate": 0.0001212471134554086, |
|
"loss": 1.125, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.1683476239550862, |
|
"learning_rate": 0.00012103992270373547, |
|
"loss": 1.1196, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.1634846923426851, |
|
"learning_rate": 0.00012083263740515765, |
|
"loss": 1.1666, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.20362360607016913, |
|
"learning_rate": 0.00012062525849115107, |
|
"loss": 1.2166, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.15520795046759722, |
|
"learning_rate": 0.00012041778689361254, |
|
"loss": 1.1046, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.17498352958051774, |
|
"learning_rate": 0.00012021022354485514, |
|
"loss": 1.1331, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.17105423893572264, |
|
"learning_rate": 0.00012000256937760445, |
|
"loss": 1.0799, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.16837214432904796, |
|
"learning_rate": 0.00011979482532499401, |
|
"loss": 1.1856, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.1717737899447442, |
|
"learning_rate": 0.00011958699232056134, |
|
"loss": 1.1738, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.16901679352588425, |
|
"learning_rate": 0.0001193790712982437, |
|
"loss": 1.1361, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.39457461047161113, |
|
"learning_rate": 0.00011917106319237386, |
|
"loss": 1.1095, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 0.1689082831860728, |
|
"learning_rate": 0.00011896296893767587, |
|
"loss": 1.1518, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 0.17381008421938537, |
|
"learning_rate": 0.00011875478946926093, |
|
"loss": 1.1251, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 0.1723002103848165, |
|
"learning_rate": 0.00011854652572262323, |
|
"loss": 1.1128, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 0.1799457776202214, |
|
"learning_rate": 0.00011833817863363564, |
|
"loss": 1.1702, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 0.18319688210909812, |
|
"learning_rate": 0.00011812974913854545, |
|
"loss": 1.1537, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.1716604871697964, |
|
"learning_rate": 0.0001179212381739704, |
|
"loss": 1.1567, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.17098997576829758, |
|
"learning_rate": 0.00011771264667689427, |
|
"loss": 1.1309, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.2607304915983766, |
|
"learning_rate": 0.00011750397558466273, |
|
"loss": 1.1151, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.1869030292628108, |
|
"learning_rate": 0.00011729522583497912, |
|
"loss": 1.0764, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.1969233515350324, |
|
"learning_rate": 0.00011708639836590023, |
|
"loss": 1.0986, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.17794486028770468, |
|
"learning_rate": 0.00011687749411583213, |
|
"loss": 1.1327, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 0.19408346665780069, |
|
"learning_rate": 0.00011666851402352588, |
|
"loss": 1.0768, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 0.5122187922492092, |
|
"learning_rate": 0.00011645945902807341, |
|
"loss": 1.1223, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 0.18948615786386766, |
|
"learning_rate": 0.00011625033006890315, |
|
"loss": 1.1452, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 0.16917117240636045, |
|
"learning_rate": 0.00011604112808577603, |
|
"loss": 1.131, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 0.177460873056157, |
|
"learning_rate": 0.00011583185401878101, |
|
"loss": 1.1691, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 0.2727518607538819, |
|
"learning_rate": 0.00011562250880833104, |
|
"loss": 1.185, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 0.1722336921590558, |
|
"learning_rate": 0.00011541309339515869, |
|
"loss": 1.2202, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 0.16731429410912582, |
|
"learning_rate": 0.00011520360872031209, |
|
"loss": 1.1114, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 0.1747423744661194, |
|
"learning_rate": 0.00011499405572515059, |
|
"loss": 1.0929, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 0.18705677818301408, |
|
"learning_rate": 0.00011478443535134049, |
|
"loss": 1.101, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 0.18090585152937122, |
|
"learning_rate": 0.00011457474854085096, |
|
"loss": 1.2376, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 0.1572625515956593, |
|
"learning_rate": 0.00011436499623594963, |
|
"loss": 1.1039, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 0.18868561701551964, |
|
"learning_rate": 0.00011415517937919846, |
|
"loss": 1.1091, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 0.3765201515989829, |
|
"learning_rate": 0.00011394529891344958, |
|
"loss": 1.1904, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 0.17659930390592027, |
|
"learning_rate": 0.00011373535578184082, |
|
"loss": 1.1767, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 0.16250915193456436, |
|
"learning_rate": 0.00011352535092779173, |
|
"loss": 1.0396, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 0.17679862621844297, |
|
"learning_rate": 0.00011331528529499909, |
|
"loss": 1.0899, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 0.17303502214887417, |
|
"learning_rate": 0.00011310515982743293, |
|
"loss": 1.1623, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 0.17241019588300063, |
|
"learning_rate": 0.00011289497546933212, |
|
"loss": 1.1761, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 0.19940689264510733, |
|
"learning_rate": 0.00011268473316520007, |
|
"loss": 1.186, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 0.17205111928537736, |
|
"learning_rate": 0.00011247443385980078, |
|
"loss": 1.1669, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 0.168610859884427, |
|
"learning_rate": 0.0001122640784981542, |
|
"loss": 1.1223, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 0.1571679218925349, |
|
"learning_rate": 0.0001120536680255323, |
|
"loss": 1.1017, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 0.1731635445936976, |
|
"learning_rate": 0.00011184320338745467, |
|
"loss": 1.1549, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 0.16986390394426032, |
|
"learning_rate": 0.00011163268552968423, |
|
"loss": 1.1561, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 0.17807860518694296, |
|
"learning_rate": 0.00011142211539822318, |
|
"loss": 1.1191, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 0.1848822566355218, |
|
"learning_rate": 0.0001112114939393085, |
|
"loss": 1.1265, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 0.16154933469362054, |
|
"learning_rate": 0.00011100082209940795, |
|
"loss": 1.1696, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 0.17539593545795423, |
|
"learning_rate": 0.00011079010082521557, |
|
"loss": 1.2098, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 0.16955097983529596, |
|
"learning_rate": 0.00011057933106364758, |
|
"loss": 1.134, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 0.1627286025891699, |
|
"learning_rate": 0.00011036851376183812, |
|
"loss": 1.0968, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 0.18241009216885773, |
|
"learning_rate": 0.0001101576498671349, |
|
"loss": 1.0784, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 0.1602213810951013, |
|
"learning_rate": 0.00010994674032709513, |
|
"loss": 1.0272, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 0.17710273355839293, |
|
"learning_rate": 0.00010973578608948094, |
|
"loss": 1.1807, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 0.16846043025211901, |
|
"learning_rate": 0.00010952478810225548, |
|
"loss": 1.1237, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 0.17578602240771765, |
|
"learning_rate": 0.00010931374731357841, |
|
"loss": 1.1305, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 0.1741114211076835, |
|
"learning_rate": 0.0001091026646718018, |
|
"loss": 1.0663, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 0.17419627851243186, |
|
"learning_rate": 0.0001088915411254657, |
|
"loss": 1.0864, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 0.1941495171655647, |
|
"learning_rate": 0.00010868037762329404, |
|
"loss": 1.1471, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 0.16828482820770493, |
|
"learning_rate": 0.0001084691751141903, |
|
"loss": 1.1386, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 0.1812212801001379, |
|
"learning_rate": 0.00010825793454723325, |
|
"loss": 1.193, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 0.31177412914918357, |
|
"learning_rate": 0.00010804665687167262, |
|
"loss": 1.1416, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 0.16307908317649347, |
|
"learning_rate": 0.00010783534303692493, |
|
"loss": 1.0742, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 0.17264040918528867, |
|
"learning_rate": 0.00010762399399256917, |
|
"loss": 1.1196, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 0.16662882952717148, |
|
"learning_rate": 0.00010741261068834265, |
|
"loss": 1.1095, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 0.16909860632274898, |
|
"learning_rate": 0.00010720119407413647, |
|
"loss": 1.0978, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 0.1841227749300958, |
|
"learning_rate": 0.00010698974509999158, |
|
"loss": 1.1249, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 0.163410587005958, |
|
"learning_rate": 0.00010677826471609422, |
|
"loss": 1.1027, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 0.16473194468337188, |
|
"learning_rate": 0.00010656675387277182, |
|
"loss": 1.0452, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.16465304101505895, |
|
"learning_rate": 0.00010635521352048872, |
|
"loss": 1.0975, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.1648559678877552, |
|
"learning_rate": 0.00010614364460984176, |
|
"loss": 1.1865, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.17182083260240222, |
|
"learning_rate": 0.00010593204809155628, |
|
"loss": 1.1666, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.262043699845629, |
|
"learning_rate": 0.00010572042491648149, |
|
"loss": 1.1705, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.1642146793634745, |
|
"learning_rate": 0.00010550877603558655, |
|
"loss": 1.0943, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.1740460470258003, |
|
"learning_rate": 0.00010529710239995605, |
|
"loss": 1.0492, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 0.16424504391420278, |
|
"learning_rate": 0.0001050854049607858, |
|
"loss": 1.0877, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 0.1754464752434157, |
|
"learning_rate": 0.00010487368466937866, |
|
"loss": 1.2004, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 0.17019599469065266, |
|
"learning_rate": 0.00010466194247714008, |
|
"loss": 1.1736, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 0.16453531456339282, |
|
"learning_rate": 0.00010445017933557404, |
|
"loss": 1.0682, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 0.16576841634828685, |
|
"learning_rate": 0.00010423839619627853, |
|
"loss": 1.1239, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 0.17959933112877494, |
|
"learning_rate": 0.00010402659401094152, |
|
"loss": 1.1741, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 0.1764446278189948, |
|
"learning_rate": 0.00010381477373133652, |
|
"loss": 1.1465, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 0.1704634166756464, |
|
"learning_rate": 0.0001036029363093183, |
|
"loss": 1.2092, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 0.16952721202785456, |
|
"learning_rate": 0.00010339108269681874, |
|
"loss": 1.1478, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 0.1787903145977452, |
|
"learning_rate": 0.00010317921384584244, |
|
"loss": 1.1818, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 0.1623533169508836, |
|
"learning_rate": 0.00010296733070846252, |
|
"loss": 1.1352, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 0.17651784025914782, |
|
"learning_rate": 0.00010275543423681621, |
|
"loss": 1.175, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 0.16473007072288753, |
|
"learning_rate": 0.00010254352538310075, |
|
"loss": 1.1837, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 0.16909571695790346, |
|
"learning_rate": 0.00010233160509956894, |
|
"loss": 1.1435, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 0.1904744004900901, |
|
"learning_rate": 0.000102119674338525, |
|
"loss": 1.1549, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 0.17373949711988154, |
|
"learning_rate": 0.00010190773405232024, |
|
"loss": 1.202, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 0.18192103377408994, |
|
"learning_rate": 0.00010169578519334873, |
|
"loss": 1.1851, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 0.18229237608440976, |
|
"learning_rate": 0.0001014838287140431, |
|
"loss": 1.1233, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 0.16940278596266165, |
|
"learning_rate": 0.00010127186556687019, |
|
"loss": 1.169, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 0.18262078712459354, |
|
"learning_rate": 0.00010105989670432681, |
|
"loss": 1.199, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 0.16747513400418446, |
|
"learning_rate": 0.00010084792307893552, |
|
"loss": 1.0666, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 0.1746037115339502, |
|
"learning_rate": 0.00010063594564324012, |
|
"loss": 1.1885, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 0.1732153359208303, |
|
"learning_rate": 0.00010042396534980176, |
|
"loss": 1.1155, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 0.18911672921606115, |
|
"learning_rate": 0.00010021198315119424, |
|
"loss": 1.1495, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 0.1810967486335703, |
|
"learning_rate": 0.0001, |
|
"loss": 1.2386, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 0.16208600336701465, |
|
"learning_rate": 9.978801684880578e-05, |
|
"loss": 1.1016, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 0.17108242412955507, |
|
"learning_rate": 9.957603465019826e-05, |
|
"loss": 1.1659, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 0.17777524975299183, |
|
"learning_rate": 9.93640543567599e-05, |
|
"loss": 1.1877, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 0.32271912031116656, |
|
"learning_rate": 9.91520769210645e-05, |
|
"loss": 1.1254, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 0.16598806172835462, |
|
"learning_rate": 9.894010329567323e-05, |
|
"loss": 1.187, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 0.17032736087667158, |
|
"learning_rate": 9.872813443312984e-05, |
|
"loss": 1.2073, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 0.16171590357915658, |
|
"learning_rate": 9.851617128595694e-05, |
|
"loss": 1.0839, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 0.22200097454760384, |
|
"learning_rate": 9.830421480665128e-05, |
|
"loss": 1.1299, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 0.18159512381331983, |
|
"learning_rate": 9.809226594767978e-05, |
|
"loss": 1.1991, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 0.1729262170321024, |
|
"learning_rate": 9.788032566147505e-05, |
|
"loss": 1.0962, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 0.17432638214872634, |
|
"learning_rate": 9.766839490043108e-05, |
|
"loss": 1.1669, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 0.16088749813770933, |
|
"learning_rate": 9.745647461689931e-05, |
|
"loss": 1.0548, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 0.1741820101835776, |
|
"learning_rate": 9.724456576318381e-05, |
|
"loss": 1.1986, |
|
"step": 919 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 0.1749247436598961, |
|
"learning_rate": 9.70326692915375e-05, |
|
"loss": 1.1474, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 0.1637038092120215, |
|
"learning_rate": 9.682078615415754e-05, |
|
"loss": 1.1231, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 0.18807058316399383, |
|
"learning_rate": 9.660891730318128e-05, |
|
"loss": 1.2045, |
|
"step": 922 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 0.1748352594360658, |
|
"learning_rate": 9.639706369068171e-05, |
|
"loss": 1.1689, |
|
"step": 923 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 0.1716288593635444, |
|
"learning_rate": 9.61852262686635e-05, |
|
"loss": 1.1572, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 0.18271775085545605, |
|
"learning_rate": 9.597340598905852e-05, |
|
"loss": 1.21, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 0.1717878271479738, |
|
"learning_rate": 9.576160380372149e-05, |
|
"loss": 1.0988, |
|
"step": 926 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 0.1680182822271049, |
|
"learning_rate": 9.5549820664426e-05, |
|
"loss": 1.0968, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 0.1742620341628087, |
|
"learning_rate": 9.533805752285993e-05, |
|
"loss": 1.1635, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 0.16338762068272702, |
|
"learning_rate": 9.512631533062138e-05, |
|
"loss": 0.9968, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 0.17102172627425277, |
|
"learning_rate": 9.491459503921421e-05, |
|
"loss": 1.1481, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 0.166342175891061, |
|
"learning_rate": 9.470289760004398e-05, |
|
"loss": 1.0721, |
|
"step": 931 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 0.2215993724024064, |
|
"learning_rate": 9.449122396441345e-05, |
|
"loss": 1.1877, |
|
"step": 932 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 0.1659659894828161, |
|
"learning_rate": 9.427957508351852e-05, |
|
"loss": 1.1691, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 0.1656871284817402, |
|
"learning_rate": 9.406795190844376e-05, |
|
"loss": 1.06, |
|
"step": 934 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 0.1586845482565193, |
|
"learning_rate": 9.385635539015825e-05, |
|
"loss": 1.1, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 0.16465420669616035, |
|
"learning_rate": 9.364478647951133e-05, |
|
"loss": 1.1403, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 0.15668641868741373, |
|
"learning_rate": 9.343324612722819e-05, |
|
"loss": 1.1075, |
|
"step": 937 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 0.16611169088286057, |
|
"learning_rate": 9.32217352839058e-05, |
|
"loss": 1.1256, |
|
"step": 938 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 0.17301783500328655, |
|
"learning_rate": 9.301025490000841e-05, |
|
"loss": 1.1797, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 0.17015053210523748, |
|
"learning_rate": 9.279880592586354e-05, |
|
"loss": 1.1087, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 0.17913490315482883, |
|
"learning_rate": 9.25873893116574e-05, |
|
"loss": 1.1673, |
|
"step": 941 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 0.18716341523653843, |
|
"learning_rate": 9.237600600743085e-05, |
|
"loss": 1.1448, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 0.16020067287870843, |
|
"learning_rate": 9.216465696307512e-05, |
|
"loss": 1.1152, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 0.18082661292174337, |
|
"learning_rate": 9.195334312832742e-05, |
|
"loss": 1.1923, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 0.16932315705059348, |
|
"learning_rate": 9.174206545276677e-05, |
|
"loss": 1.2025, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 0.1773723182281052, |
|
"learning_rate": 9.153082488580968e-05, |
|
"loss": 1.1583, |
|
"step": 946 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 0.1650084811263765, |
|
"learning_rate": 9.131962237670598e-05, |
|
"loss": 1.1456, |
|
"step": 947 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 0.17296637806848497, |
|
"learning_rate": 9.11084588745343e-05, |
|
"loss": 1.1188, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 0.18441757049220042, |
|
"learning_rate": 9.089733532819824e-05, |
|
"loss": 1.1057, |
|
"step": 949 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 0.16479278966433117, |
|
"learning_rate": 9.068625268642161e-05, |
|
"loss": 0.9686, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 0.1721522611977686, |
|
"learning_rate": 9.047521189774455e-05, |
|
"loss": 1.1593, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 0.17693809369137542, |
|
"learning_rate": 9.026421391051907e-05, |
|
"loss": 1.141, |
|
"step": 952 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 0.16748786371543328, |
|
"learning_rate": 9.005325967290488e-05, |
|
"loss": 1.0954, |
|
"step": 953 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 0.16738864271495277, |
|
"learning_rate": 8.984235013286511e-05, |
|
"loss": 1.1954, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 0.17608800377093817, |
|
"learning_rate": 8.963148623816191e-05, |
|
"loss": 1.2054, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 0.1572056318124515, |
|
"learning_rate": 8.942066893635246e-05, |
|
"loss": 1.062, |
|
"step": 956 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 0.17321331852374994, |
|
"learning_rate": 8.920989917478447e-05, |
|
"loss": 1.1774, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 0.24022668199221645, |
|
"learning_rate": 8.899917790059208e-05, |
|
"loss": 1.1422, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 0.17078130002322278, |
|
"learning_rate": 8.878850606069153e-05, |
|
"loss": 1.102, |
|
"step": 959 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 0.18816676453808523, |
|
"learning_rate": 8.857788460177686e-05, |
|
"loss": 1.1228, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 0.17061330738089536, |
|
"learning_rate": 8.836731447031581e-05, |
|
"loss": 1.1664, |
|
"step": 961 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 0.15744686305416977, |
|
"learning_rate": 8.815679661254537e-05, |
|
"loss": 1.0526, |
|
"step": 962 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 0.17140964720406338, |
|
"learning_rate": 8.79463319744677e-05, |
|
"loss": 1.1015, |
|
"step": 963 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 0.177839632569439, |
|
"learning_rate": 8.77359215018458e-05, |
|
"loss": 1.179, |
|
"step": 964 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 0.18976111884283367, |
|
"learning_rate": 8.752556614019923e-05, |
|
"loss": 1.1901, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 0.17757617341847579, |
|
"learning_rate": 8.731526683479992e-05, |
|
"loss": 1.1195, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 0.18025608157061349, |
|
"learning_rate": 8.710502453066792e-05, |
|
"loss": 1.1681, |
|
"step": 967 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 0.17991548690786335, |
|
"learning_rate": 8.689484017256711e-05, |
|
"loss": 1.0382, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 0.18377930665665893, |
|
"learning_rate": 8.668471470500095e-05, |
|
"loss": 1.2159, |
|
"step": 969 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 0.16785777204268718, |
|
"learning_rate": 8.647464907220832e-05, |
|
"loss": 1.1001, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 0.1718542715910871, |
|
"learning_rate": 8.626464421815919e-05, |
|
"loss": 1.1362, |
|
"step": 971 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 0.17567452985919224, |
|
"learning_rate": 8.605470108655045e-05, |
|
"loss": 1.0701, |
|
"step": 972 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 0.16645014392512258, |
|
"learning_rate": 8.584482062080154e-05, |
|
"loss": 1.1224, |
|
"step": 973 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 0.1683034297078507, |
|
"learning_rate": 8.563500376405041e-05, |
|
"loss": 1.1833, |
|
"step": 974 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 0.17027006999646288, |
|
"learning_rate": 8.542525145914905e-05, |
|
"loss": 1.2105, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 0.16661484142693264, |
|
"learning_rate": 8.521556464865954e-05, |
|
"loss": 1.1345, |
|
"step": 976 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 0.16535108955142738, |
|
"learning_rate": 8.500594427484946e-05, |
|
"loss": 1.0801, |
|
"step": 977 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 0.1619767851731314, |
|
"learning_rate": 8.479639127968792e-05, |
|
"loss": 1.1398, |
|
"step": 978 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 0.16148830590153632, |
|
"learning_rate": 8.458690660484134e-05, |
|
"loss": 1.1714, |
|
"step": 979 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 0.1718526387843433, |
|
"learning_rate": 8.4377491191669e-05, |
|
"loss": 1.1196, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 0.1589791532312397, |
|
"learning_rate": 8.4168145981219e-05, |
|
"loss": 1.0975, |
|
"step": 981 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 0.18228960673193118, |
|
"learning_rate": 8.395887191422397e-05, |
|
"loss": 1.0911, |
|
"step": 982 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 0.17051635436251017, |
|
"learning_rate": 8.374966993109687e-05, |
|
"loss": 1.1156, |
|
"step": 983 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 0.1734092864786433, |
|
"learning_rate": 8.35405409719266e-05, |
|
"loss": 1.1387, |
|
"step": 984 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 0.17960160313198717, |
|
"learning_rate": 8.333148597647414e-05, |
|
"loss": 1.1255, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.1667460148360989, |
|
"learning_rate": 8.312250588416791e-05, |
|
"loss": 1.0877, |
|
"step": 986 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.17227304205526361, |
|
"learning_rate": 8.291360163409978e-05, |
|
"loss": 1.1643, |
|
"step": 987 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.16172787345643957, |
|
"learning_rate": 8.270477416502091e-05, |
|
"loss": 1.0813, |
|
"step": 988 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.17717380607042416, |
|
"learning_rate": 8.249602441533726e-05, |
|
"loss": 1.1653, |
|
"step": 989 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.17646325594155327, |
|
"learning_rate": 8.228735332310575e-05, |
|
"loss": 1.1671, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 0.16889550666185796, |
|
"learning_rate": 8.207876182602958e-05, |
|
"loss": 1.1068, |
|
"step": 991 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 0.1668114849825541, |
|
"learning_rate": 8.187025086145458e-05, |
|
"loss": 1.1467, |
|
"step": 992 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 0.1721245962364379, |
|
"learning_rate": 8.16618213663644e-05, |
|
"loss": 1.1377, |
|
"step": 993 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 0.1929335873892253, |
|
"learning_rate": 8.145347427737678e-05, |
|
"loss": 1.119, |
|
"step": 994 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 0.16623130316399012, |
|
"learning_rate": 8.12452105307391e-05, |
|
"loss": 1.0687, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 0.18085333131496864, |
|
"learning_rate": 8.103703106232416e-05, |
|
"loss": 1.1517, |
|
"step": 996 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 0.17906886829733978, |
|
"learning_rate": 8.082893680762619e-05, |
|
"loss": 1.1325, |
|
"step": 997 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 0.16359826556853824, |
|
"learning_rate": 8.062092870175628e-05, |
|
"loss": 1.0353, |
|
"step": 998 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 0.1626009671933035, |
|
"learning_rate": 8.041300767943867e-05, |
|
"loss": 1.1598, |
|
"step": 999 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 0.16643269286545453, |
|
"learning_rate": 8.0205174675006e-05, |
|
"loss": 1.1226, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 0.1699654548245084, |
|
"learning_rate": 7.999743062239557e-05, |
|
"loss": 1.0999, |
|
"step": 1001 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 0.1827266429905637, |
|
"learning_rate": 7.978977645514487e-05, |
|
"loss": 1.2791, |
|
"step": 1002 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 0.44434388351574605, |
|
"learning_rate": 7.958221310638749e-05, |
|
"loss": 1.1693, |
|
"step": 1003 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 0.17634755268047692, |
|
"learning_rate": 7.937474150884897e-05, |
|
"loss": 1.0845, |
|
"step": 1004 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 0.177726814790075, |
|
"learning_rate": 7.916736259484239e-05, |
|
"loss": 1.1849, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 0.16479013332848777, |
|
"learning_rate": 7.896007729626457e-05, |
|
"loss": 1.1296, |
|
"step": 1006 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 0.1607507116543152, |
|
"learning_rate": 7.875288654459144e-05, |
|
"loss": 1.0829, |
|
"step": 1007 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 0.16387244898588724, |
|
"learning_rate": 7.854579127087417e-05, |
|
"loss": 1.1604, |
|
"step": 1008 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 0.163742117997135, |
|
"learning_rate": 7.833879240573487e-05, |
|
"loss": 1.0757, |
|
"step": 1009 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 0.16638334168591146, |
|
"learning_rate": 7.813189087936243e-05, |
|
"loss": 1.1637, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 0.16818864513814694, |
|
"learning_rate": 7.792508762150833e-05, |
|
"loss": 1.1564, |
|
"step": 1011 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 0.17901829622442805, |
|
"learning_rate": 7.771838356148232e-05, |
|
"loss": 1.1841, |
|
"step": 1012 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 0.16614583867770816, |
|
"learning_rate": 7.751177962814866e-05, |
|
"loss": 1.1255, |
|
"step": 1013 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 0.19013974114118293, |
|
"learning_rate": 7.730527674992143e-05, |
|
"loss": 1.1488, |
|
"step": 1014 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 0.17442438513245137, |
|
"learning_rate": 7.709887585476075e-05, |
|
"loss": 1.1066, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 0.17701832779209928, |
|
"learning_rate": 7.689257787016834e-05, |
|
"loss": 1.1755, |
|
"step": 1016 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 0.16936912919421254, |
|
"learning_rate": 7.668638372318359e-05, |
|
"loss": 1.1057, |
|
"step": 1017 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 0.19112857435989614, |
|
"learning_rate": 7.648029434037915e-05, |
|
"loss": 1.1412, |
|
"step": 1018 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 0.16103962948152212, |
|
"learning_rate": 7.627431064785706e-05, |
|
"loss": 1.0929, |
|
"step": 1019 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 0.16677548612612422, |
|
"learning_rate": 7.606843357124426e-05, |
|
"loss": 1.2046, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 0.16233738083605964, |
|
"learning_rate": 7.58626640356886e-05, |
|
"loss": 1.0678, |
|
"step": 1021 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 0.16602043351095722, |
|
"learning_rate": 7.565700296585483e-05, |
|
"loss": 1.1216, |
|
"step": 1022 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 0.16627151579102536, |
|
"learning_rate": 7.54514512859201e-05, |
|
"loss": 1.0911, |
|
"step": 1023 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 0.20281510390834787, |
|
"learning_rate": 7.524600991957012e-05, |
|
"loss": 1.1722, |
|
"step": 1024 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 0.16069933741616732, |
|
"learning_rate": 7.504067978999484e-05, |
|
"loss": 1.0839, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 0.18107514785611215, |
|
"learning_rate": 7.483546181988436e-05, |
|
"loss": 1.0717, |
|
"step": 1026 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 0.1721242495617235, |
|
"learning_rate": 7.463035693142473e-05, |
|
"loss": 1.1599, |
|
"step": 1027 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 0.17918262771658913, |
|
"learning_rate": 7.442536604629395e-05, |
|
"loss": 1.089, |
|
"step": 1028 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 0.17784491718923554, |
|
"learning_rate": 7.422049008565757e-05, |
|
"loss": 1.176, |
|
"step": 1029 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 0.15630505382249996, |
|
"learning_rate": 7.401572997016479e-05, |
|
"loss": 1.039, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 0.17070239450442767, |
|
"learning_rate": 7.381108661994429e-05, |
|
"loss": 1.1059, |
|
"step": 1031 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 0.16222624124278331, |
|
"learning_rate": 7.360656095459995e-05, |
|
"loss": 1.024, |
|
"step": 1032 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 0.17125506848213806, |
|
"learning_rate": 7.340215389320687e-05, |
|
"loss": 1.1567, |
|
"step": 1033 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 0.17959953833242476, |
|
"learning_rate": 7.31978663543071e-05, |
|
"loss": 1.141, |
|
"step": 1034 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 0.17010870239393844, |
|
"learning_rate": 7.299369925590574e-05, |
|
"loss": 1.1046, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 0.15998424148316903, |
|
"learning_rate": 7.278965351546648e-05, |
|
"loss": 1.0627, |
|
"step": 1036 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 0.1710612967988063, |
|
"learning_rate": 7.258573004990788e-05, |
|
"loss": 1.1486, |
|
"step": 1037 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 0.1607397255850716, |
|
"learning_rate": 7.238192977559884e-05, |
|
"loss": 1.0751, |
|
"step": 1038 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 0.17530360487610264, |
|
"learning_rate": 7.217825360835473e-05, |
|
"loss": 1.1281, |
|
"step": 1039 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 0.17374375355320706, |
|
"learning_rate": 7.197470246343333e-05, |
|
"loss": 1.1493, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 0.15987413285182245, |
|
"learning_rate": 7.177127725553045e-05, |
|
"loss": 1.1446, |
|
"step": 1041 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 0.1789355664165954, |
|
"learning_rate": 7.156797889877613e-05, |
|
"loss": 1.2304, |
|
"step": 1042 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 0.17282586483463735, |
|
"learning_rate": 7.136480830673019e-05, |
|
"loss": 1.1909, |
|
"step": 1043 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 0.159213004708279, |
|
"learning_rate": 7.116176639237852e-05, |
|
"loss": 1.0624, |
|
"step": 1044 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 0.16197929506166975, |
|
"learning_rate": 7.095885406812866e-05, |
|
"loss": 1.1283, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 0.16495179889101508, |
|
"learning_rate": 7.075607224580581e-05, |
|
"loss": 1.1378, |
|
"step": 1046 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 0.1662731898113741, |
|
"learning_rate": 7.05534218366488e-05, |
|
"loss": 1.0542, |
|
"step": 1047 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 0.1650210560996418, |
|
"learning_rate": 7.035090375130581e-05, |
|
"loss": 1.1326, |
|
"step": 1048 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 0.1708049460231556, |
|
"learning_rate": 7.014851889983057e-05, |
|
"loss": 1.137, |
|
"step": 1049 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 0.17558230008855352, |
|
"learning_rate": 6.994626819167789e-05, |
|
"loss": 1.1359, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 0.16645826974884173, |
|
"learning_rate": 6.974415253570003e-05, |
|
"loss": 1.1742, |
|
"step": 1051 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 0.17184188151368943, |
|
"learning_rate": 6.954217284014211e-05, |
|
"loss": 1.1953, |
|
"step": 1052 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 0.17068592296841717, |
|
"learning_rate": 6.934033001263847e-05, |
|
"loss": 1.1301, |
|
"step": 1053 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 0.16302700472895656, |
|
"learning_rate": 6.913862496020831e-05, |
|
"loss": 1.1232, |
|
"step": 1054 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 0.17057720351205383, |
|
"learning_rate": 6.893705858925178e-05, |
|
"loss": 1.1247, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 0.1582652133644645, |
|
"learning_rate": 6.873563180554583e-05, |
|
"loss": 1.1203, |
|
"step": 1056 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 0.16895339487465857, |
|
"learning_rate": 6.853434551424e-05, |
|
"loss": 1.1846, |
|
"step": 1057 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 0.1726993947504367, |
|
"learning_rate": 6.833320061985277e-05, |
|
"loss": 1.1963, |
|
"step": 1058 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 0.16055788377798388, |
|
"learning_rate": 6.813219802626698e-05, |
|
"loss": 1.1439, |
|
"step": 1059 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 0.17265920805658908, |
|
"learning_rate": 6.793133863672616e-05, |
|
"loss": 1.192, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 0.1833601588530699, |
|
"learning_rate": 6.773062335383024e-05, |
|
"loss": 1.1128, |
|
"step": 1061 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 0.1717674054389487, |
|
"learning_rate": 6.753005307953167e-05, |
|
"loss": 1.134, |
|
"step": 1062 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 0.18207071739270747, |
|
"learning_rate": 6.73296287151312e-05, |
|
"loss": 1.1723, |
|
"step": 1063 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 0.1624856029472637, |
|
"learning_rate": 6.712935116127389e-05, |
|
"loss": 1.1144, |
|
"step": 1064 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 0.19915390833072513, |
|
"learning_rate": 6.692922131794517e-05, |
|
"loss": 1.1072, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 0.1722567762020263, |
|
"learning_rate": 6.672924008446662e-05, |
|
"loss": 1.1425, |
|
"step": 1066 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 0.15963511981578907, |
|
"learning_rate": 6.652940835949208e-05, |
|
"loss": 1.1211, |
|
"step": 1067 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 0.16578458195408932, |
|
"learning_rate": 6.632972704100349e-05, |
|
"loss": 1.1244, |
|
"step": 1068 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 0.17110384656595054, |
|
"learning_rate": 6.613019702630694e-05, |
|
"loss": 1.1795, |
|
"step": 1069 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 0.16990207952122507, |
|
"learning_rate": 6.593081921202859e-05, |
|
"loss": 1.0817, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 0.16467375370533938, |
|
"learning_rate": 6.57315944941107e-05, |
|
"loss": 1.0977, |
|
"step": 1071 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 0.17152460282733728, |
|
"learning_rate": 6.553252376780748e-05, |
|
"loss": 1.1868, |
|
"step": 1072 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 0.18132610186641904, |
|
"learning_rate": 6.533360792768122e-05, |
|
"loss": 1.2005, |
|
"step": 1073 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 0.16698085910075294, |
|
"learning_rate": 6.513484786759818e-05, |
|
"loss": 1.1362, |
|
"step": 1074 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 0.16504199012773973, |
|
"learning_rate": 6.493624448072457e-05, |
|
"loss": 1.1303, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 0.1779596734054924, |
|
"learning_rate": 6.473779865952263e-05, |
|
"loss": 1.1894, |
|
"step": 1076 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 0.16364715074236819, |
|
"learning_rate": 6.453951129574644e-05, |
|
"loss": 1.1443, |
|
"step": 1077 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 0.16251550362999542, |
|
"learning_rate": 6.434138328043815e-05, |
|
"loss": 1.1238, |
|
"step": 1078 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 0.1666990603079008, |
|
"learning_rate": 6.414341550392368e-05, |
|
"loss": 1.0909, |
|
"step": 1079 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 0.15779118156939148, |
|
"learning_rate": 6.39456088558091e-05, |
|
"loss": 1.1109, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 0.1677767995314692, |
|
"learning_rate": 6.374796422497621e-05, |
|
"loss": 1.2103, |
|
"step": 1081 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 0.17165919802575422, |
|
"learning_rate": 6.355048249957886e-05, |
|
"loss": 1.1879, |
|
"step": 1082 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 0.1918026223221136, |
|
"learning_rate": 6.33531645670389e-05, |
|
"loss": 1.1733, |
|
"step": 1083 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 0.1787527260843118, |
|
"learning_rate": 6.3156011314042e-05, |
|
"loss": 1.1338, |
|
"step": 1084 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 0.1694260433762424, |
|
"learning_rate": 6.2959023626534e-05, |
|
"loss": 1.1537, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 0.18311039386845188, |
|
"learning_rate": 6.276220238971652e-05, |
|
"loss": 1.1976, |
|
"step": 1086 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 0.17113962947413622, |
|
"learning_rate": 6.256554848804343e-05, |
|
"loss": 1.0813, |
|
"step": 1087 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 0.1664054538940142, |
|
"learning_rate": 6.236906280521646e-05, |
|
"loss": 1.1514, |
|
"step": 1088 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 0.16227842164425618, |
|
"learning_rate": 6.217274622418153e-05, |
|
"loss": 1.0597, |
|
"step": 1089 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 0.17755371081391708, |
|
"learning_rate": 6.197659962712461e-05, |
|
"loss": 1.19, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 0.15844523213403552, |
|
"learning_rate": 6.178062389546784e-05, |
|
"loss": 1.0587, |
|
"step": 1091 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 0.16903107223771263, |
|
"learning_rate": 6.158481990986557e-05, |
|
"loss": 1.1339, |
|
"step": 1092 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 0.17002303254909312, |
|
"learning_rate": 6.138918855020028e-05, |
|
"loss": 1.1158, |
|
"step": 1093 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 0.16613843815618992, |
|
"learning_rate": 6.11937306955789e-05, |
|
"loss": 1.1894, |
|
"step": 1094 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 0.17111783959067672, |
|
"learning_rate": 6.099844722432843e-05, |
|
"loss": 1.1974, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.1670020953276936, |
|
"learning_rate": 6.080333901399251e-05, |
|
"loss": 1.1018, |
|
"step": 1096 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.18080514034863332, |
|
"learning_rate": 6.060840694132701e-05, |
|
"loss": 1.1845, |
|
"step": 1097 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.18525347317947505, |
|
"learning_rate": 6.0413651882296406e-05, |
|
"loss": 1.0796, |
|
"step": 1098 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.20400403778138837, |
|
"learning_rate": 6.021907471206971e-05, |
|
"loss": 1.0117, |
|
"step": 1099 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.1641784925803925, |
|
"learning_rate": 6.002467630501646e-05, |
|
"loss": 1.097, |
|
"step": 1100 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 1647, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 50, |
|
"total_flos": 1.0209608098381824e+16, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|