|
{ |
|
"best_metric": 1.8791134357452393, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-150", |
|
"epoch": 0.084942393642113, |
|
"eval_steps": 25, |
|
"global_step": 166, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.000511701166518753, |
|
"grad_norm": 0.46404364705085754, |
|
"learning_rate": 1.6666666666666667e-05, |
|
"loss": 1.9735, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.000511701166518753, |
|
"eval_loss": 4.42933464050293, |
|
"eval_runtime": 1.5342, |
|
"eval_samples_per_second": 32.589, |
|
"eval_steps_per_second": 8.473, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.001023402333037506, |
|
"grad_norm": 0.8488529324531555, |
|
"learning_rate": 3.3333333333333335e-05, |
|
"loss": 2.7654, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.001535103499556259, |
|
"grad_norm": 1.001631736755371, |
|
"learning_rate": 5e-05, |
|
"loss": 3.0228, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.002046804666075012, |
|
"grad_norm": 1.149634838104248, |
|
"learning_rate": 6.666666666666667e-05, |
|
"loss": 3.2201, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.002558505832593765, |
|
"grad_norm": 1.4123715162277222, |
|
"learning_rate": 8.333333333333334e-05, |
|
"loss": 3.3185, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.003070206999112518, |
|
"grad_norm": 1.9225883483886719, |
|
"learning_rate": 0.0001, |
|
"loss": 3.4594, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.0035819081656312713, |
|
"grad_norm": 1.8795751333236694, |
|
"learning_rate": 9.999132582169292e-05, |
|
"loss": 3.501, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.004093609332150024, |
|
"grad_norm": 2.5900070667266846, |
|
"learning_rate": 9.996530663083255e-05, |
|
"loss": 3.7938, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.004605310498668778, |
|
"grad_norm": 2.747680187225342, |
|
"learning_rate": 9.992195245831223e-05, |
|
"loss": 3.7944, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.00511701166518753, |
|
"grad_norm": 2.346953868865967, |
|
"learning_rate": 9.986128001799077e-05, |
|
"loss": 3.8185, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.005628712831706284, |
|
"grad_norm": 2.9237945079803467, |
|
"learning_rate": 9.978331270024886e-05, |
|
"loss": 3.9366, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.006140413998225036, |
|
"grad_norm": 3.622086524963379, |
|
"learning_rate": 9.96880805629717e-05, |
|
"loss": 4.2799, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.00665211516474379, |
|
"grad_norm": 2.3093104362487793, |
|
"learning_rate": 9.957562031996097e-05, |
|
"loss": 1.8889, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0071638163312625425, |
|
"grad_norm": 3.3800673484802246, |
|
"learning_rate": 9.94459753267812e-05, |
|
"loss": 2.3106, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.007675517497781296, |
|
"grad_norm": 3.375108242034912, |
|
"learning_rate": 9.929919556404513e-05, |
|
"loss": 2.3937, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.008187218664300049, |
|
"grad_norm": 2.4301416873931885, |
|
"learning_rate": 9.913533761814537e-05, |
|
"loss": 2.4152, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.008698919830818802, |
|
"grad_norm": 1.324080228805542, |
|
"learning_rate": 9.895446465943926e-05, |
|
"loss": 2.3565, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.009210620997337556, |
|
"grad_norm": 1.267471194267273, |
|
"learning_rate": 9.875664641789545e-05, |
|
"loss": 2.3992, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.009722322163856307, |
|
"grad_norm": 1.367044448852539, |
|
"learning_rate": 9.85419591562117e-05, |
|
"loss": 2.4876, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.01023402333037506, |
|
"grad_norm": 1.5106232166290283, |
|
"learning_rate": 9.831048564041413e-05, |
|
"loss": 2.6029, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.010745724496893814, |
|
"grad_norm": 1.4716838598251343, |
|
"learning_rate": 9.80623151079494e-05, |
|
"loss": 2.6521, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.011257425663412568, |
|
"grad_norm": 1.2971686124801636, |
|
"learning_rate": 9.779754323328192e-05, |
|
"loss": 2.6698, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.01176912682993132, |
|
"grad_norm": 1.4295047521591187, |
|
"learning_rate": 9.751627209100952e-05, |
|
"loss": 2.8462, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.012280827996450073, |
|
"grad_norm": 1.995835781097412, |
|
"learning_rate": 9.72186101165118e-05, |
|
"loss": 3.1823, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.012792529162968826, |
|
"grad_norm": 3.6110293865203857, |
|
"learning_rate": 9.690467206414616e-05, |
|
"loss": 3.855, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.012792529162968826, |
|
"eval_loss": 2.3177919387817383, |
|
"eval_runtime": 1.5336, |
|
"eval_samples_per_second": 32.603, |
|
"eval_steps_per_second": 8.477, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.01330423032948758, |
|
"grad_norm": 0.8223591446876526, |
|
"learning_rate": 9.657457896300791e-05, |
|
"loss": 1.6149, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.013815931496006332, |
|
"grad_norm": 1.0949565172195435, |
|
"learning_rate": 9.622845807027113e-05, |
|
"loss": 2.0488, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.014327632662525085, |
|
"grad_norm": 1.1367841958999634, |
|
"learning_rate": 9.586644282212866e-05, |
|
"loss": 2.1151, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.014839333829043838, |
|
"grad_norm": 0.9595118761062622, |
|
"learning_rate": 9.548867278234998e-05, |
|
"loss": 2.1702, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.015351034995562592, |
|
"grad_norm": 0.8941548466682434, |
|
"learning_rate": 9.509529358847655e-05, |
|
"loss": 2.2371, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.015862736162081344, |
|
"grad_norm": 0.9041470289230347, |
|
"learning_rate": 9.468645689567598e-05, |
|
"loss": 2.2612, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.016374437328600097, |
|
"grad_norm": 0.9832100868225098, |
|
"learning_rate": 9.426232031827588e-05, |
|
"loss": 2.3107, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.01688613849511885, |
|
"grad_norm": 1.1351114511489868, |
|
"learning_rate": 9.382304736900063e-05, |
|
"loss": 2.3543, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.017397839661637604, |
|
"grad_norm": 1.2356266975402832, |
|
"learning_rate": 9.336880739593416e-05, |
|
"loss": 2.4014, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.017909540828156358, |
|
"grad_norm": 1.3643877506256104, |
|
"learning_rate": 9.28997755172329e-05, |
|
"loss": 2.5325, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.01842124199467511, |
|
"grad_norm": 1.4238357543945312, |
|
"learning_rate": 9.241613255361455e-05, |
|
"loss": 2.75, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.018932943161193865, |
|
"grad_norm": 1.6282374858856201, |
|
"learning_rate": 9.191806495864812e-05, |
|
"loss": 2.8755, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.019444644327712615, |
|
"grad_norm": 0.45179080963134766, |
|
"learning_rate": 9.140576474687264e-05, |
|
"loss": 1.6157, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.019956345494231368, |
|
"grad_norm": 0.4603005051612854, |
|
"learning_rate": 9.087942941977182e-05, |
|
"loss": 1.8998, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.02046804666075012, |
|
"grad_norm": 0.5871985554695129, |
|
"learning_rate": 9.033926188963352e-05, |
|
"loss": 2.0722, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.020979747827268875, |
|
"grad_norm": 0.602536141872406, |
|
"learning_rate": 8.978547040132317e-05, |
|
"loss": 2.054, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.02149144899378763, |
|
"grad_norm": 0.72366863489151, |
|
"learning_rate": 8.921826845200139e-05, |
|
"loss": 2.1352, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.022003150160306382, |
|
"grad_norm": 0.7915085554122925, |
|
"learning_rate": 8.863787470881686e-05, |
|
"loss": 2.0386, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.022514851326825135, |
|
"grad_norm": 0.8590556979179382, |
|
"learning_rate": 8.804451292460585e-05, |
|
"loss": 2.1996, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.02302655249334389, |
|
"grad_norm": 0.9212091565132141, |
|
"learning_rate": 8.743841185163136e-05, |
|
"loss": 2.2756, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.02353825365986264, |
|
"grad_norm": 0.9804298281669617, |
|
"learning_rate": 8.681980515339464e-05, |
|
"loss": 2.295, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.024049954826381392, |
|
"grad_norm": 1.0321789979934692, |
|
"learning_rate": 8.618893131455363e-05, |
|
"loss": 2.3681, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.024561655992900146, |
|
"grad_norm": 1.3598474264144897, |
|
"learning_rate": 8.554603354898238e-05, |
|
"loss": 2.4684, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.0250733571594189, |
|
"grad_norm": 1.5326489210128784, |
|
"learning_rate": 8.489135970600768e-05, |
|
"loss": 2.6954, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.025585058325937653, |
|
"grad_norm": 2.6125943660736084, |
|
"learning_rate": 8.422516217485826e-05, |
|
"loss": 3.1982, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.025585058325937653, |
|
"eval_loss": 2.026942729949951, |
|
"eval_runtime": 1.535, |
|
"eval_samples_per_second": 32.573, |
|
"eval_steps_per_second": 8.469, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.026096759492456406, |
|
"grad_norm": 0.4928167164325714, |
|
"learning_rate": 8.354769778736406e-05, |
|
"loss": 1.5405, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.02660846065897516, |
|
"grad_norm": 0.6747775673866272, |
|
"learning_rate": 8.285922771894254e-05, |
|
"loss": 1.8902, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.027120161825493913, |
|
"grad_norm": 0.761574923992157, |
|
"learning_rate": 8.216001738791072e-05, |
|
"loss": 2.0376, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.027631862992012663, |
|
"grad_norm": 0.7025596499443054, |
|
"learning_rate": 8.14503363531613e-05, |
|
"loss": 2.0866, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.028143564158531417, |
|
"grad_norm": 0.7048724889755249, |
|
"learning_rate": 8.073045821024256e-05, |
|
"loss": 2.0865, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.02865526532505017, |
|
"grad_norm": 0.8263847231864929, |
|
"learning_rate": 8.000066048588211e-05, |
|
"loss": 2.0262, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.029166966491568923, |
|
"grad_norm": 0.8577947616577148, |
|
"learning_rate": 7.926122453099503e-05, |
|
"loss": 2.1969, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.029678667658087677, |
|
"grad_norm": 1.095415711402893, |
|
"learning_rate": 7.85124354122177e-05, |
|
"loss": 2.2191, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.03019036882460643, |
|
"grad_norm": 1.0914173126220703, |
|
"learning_rate": 7.775458180200932e-05, |
|
"loss": 2.2533, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.030702069991125184, |
|
"grad_norm": 1.104080080986023, |
|
"learning_rate": 7.698795586736298e-05, |
|
"loss": 2.3613, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.031213771157643937, |
|
"grad_norm": 1.2872527837753296, |
|
"learning_rate": 7.62128531571699e-05, |
|
"loss": 2.4725, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.03172547232416269, |
|
"grad_norm": 1.6460251808166504, |
|
"learning_rate": 7.542957248827961e-05, |
|
"loss": 2.7487, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.03223717349068144, |
|
"grad_norm": 0.5120904445648193, |
|
"learning_rate": 7.46384158303004e-05, |
|
"loss": 1.4998, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.032748874657200194, |
|
"grad_norm": 0.5734795928001404, |
|
"learning_rate": 7.383968818918426e-05, |
|
"loss": 1.8322, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.03326057582371895, |
|
"grad_norm": 0.6029512286186218, |
|
"learning_rate": 7.303369748964134e-05, |
|
"loss": 1.9155, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.0337722769902377, |
|
"grad_norm": 0.6455990076065063, |
|
"learning_rate": 7.222075445642904e-05, |
|
"loss": 2.0008, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.034283978156756455, |
|
"grad_norm": 0.6918196082115173, |
|
"learning_rate": 7.140117249456176e-05, |
|
"loss": 2.0541, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.03479567932327521, |
|
"grad_norm": 0.7618705630302429, |
|
"learning_rate": 7.057526756848719e-05, |
|
"loss": 2.062, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.03530738048979396, |
|
"grad_norm": 0.7683627605438232, |
|
"learning_rate": 6.97433580802762e-05, |
|
"loss": 2.0749, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.035819081656312715, |
|
"grad_norm": 0.8340547680854797, |
|
"learning_rate": 6.890576474687263e-05, |
|
"loss": 2.108, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.03633078282283147, |
|
"grad_norm": 0.9197708368301392, |
|
"learning_rate": 6.80628104764508e-05, |
|
"loss": 2.1471, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.03684248398935022, |
|
"grad_norm": 1.2023193836212158, |
|
"learning_rate": 6.721482024392835e-05, |
|
"loss": 2.2415, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.037354185155868976, |
|
"grad_norm": 1.1722882986068726, |
|
"learning_rate": 6.63621209656821e-05, |
|
"loss": 2.3475, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.03786588632238773, |
|
"grad_norm": 1.4040749073028564, |
|
"learning_rate": 6.550504137351576e-05, |
|
"loss": 2.6166, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.038377587488906476, |
|
"grad_norm": 2.434393882751465, |
|
"learning_rate": 6.46439118879273e-05, |
|
"loss": 2.8818, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.038377587488906476, |
|
"eval_loss": 1.9559780359268188, |
|
"eval_runtime": 1.5335, |
|
"eval_samples_per_second": 32.605, |
|
"eval_steps_per_second": 8.477, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.03888928865542523, |
|
"grad_norm": 0.4053546190261841, |
|
"learning_rate": 6.377906449072578e-05, |
|
"loss": 1.5548, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.03940098982194398, |
|
"grad_norm": 0.550913393497467, |
|
"learning_rate": 6.291083259704596e-05, |
|
"loss": 1.8655, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.039912690988462736, |
|
"grad_norm": 0.5964261889457703, |
|
"learning_rate": 6.203955092681039e-05, |
|
"loss": 1.9327, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.04042439215498149, |
|
"grad_norm": 0.7430030107498169, |
|
"learning_rate": 6.116555537568857e-05, |
|
"loss": 2.0133, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.04093609332150024, |
|
"grad_norm": 0.7743514180183411, |
|
"learning_rate": 6.0289182885602704e-05, |
|
"loss": 2.0308, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.041447794488018996, |
|
"grad_norm": 0.8544523119926453, |
|
"learning_rate": 5.941077131483025e-05, |
|
"loss": 2.0096, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.04195949565453775, |
|
"grad_norm": 0.7836202383041382, |
|
"learning_rate": 5.8530659307753036e-05, |
|
"loss": 2.0285, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.0424711968210565, |
|
"grad_norm": 0.9040354490280151, |
|
"learning_rate": 5.7649186164303506e-05, |
|
"loss": 2.0367, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.04298289798757526, |
|
"grad_norm": 0.9767117500305176, |
|
"learning_rate": 5.6766691709158096e-05, |
|
"loss": 2.1512, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.04349459915409401, |
|
"grad_norm": 1.108844518661499, |
|
"learning_rate": 5.5883516160728274e-05, |
|
"loss": 2.2661, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.044006300320612764, |
|
"grad_norm": 1.3114573955535889, |
|
"learning_rate": 5.500000000000001e-05, |
|
"loss": 2.374, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.04451800148713152, |
|
"grad_norm": 1.5609687566757202, |
|
"learning_rate": 5.4116483839271736e-05, |
|
"loss": 2.6349, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.04502970265365027, |
|
"grad_norm": 0.42394495010375977, |
|
"learning_rate": 5.3233308290841935e-05, |
|
"loss": 1.4049, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.045541403820169024, |
|
"grad_norm": 0.524941086769104, |
|
"learning_rate": 5.235081383569649e-05, |
|
"loss": 1.8021, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.04605310498668778, |
|
"grad_norm": 0.7207713723182678, |
|
"learning_rate": 5.1469340692246995e-05, |
|
"loss": 1.8878, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.046564806153206524, |
|
"grad_norm": 0.6740505695343018, |
|
"learning_rate": 5.058922868516978e-05, |
|
"loss": 1.9676, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.04707650731972528, |
|
"grad_norm": 0.71189945936203, |
|
"learning_rate": 4.9710817114397314e-05, |
|
"loss": 1.9829, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.04758820848624403, |
|
"grad_norm": 0.6990759372711182, |
|
"learning_rate": 4.883444462431145e-05, |
|
"loss": 1.9317, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.048099909652762785, |
|
"grad_norm": 0.7979338765144348, |
|
"learning_rate": 4.7960449073189606e-05, |
|
"loss": 1.9789, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.04861161081928154, |
|
"grad_norm": 0.8845726847648621, |
|
"learning_rate": 4.708916740295405e-05, |
|
"loss": 2.0438, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.04912331198580029, |
|
"grad_norm": 0.9462182521820068, |
|
"learning_rate": 4.6220935509274235e-05, |
|
"loss": 2.0737, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.049635013152319045, |
|
"grad_norm": 1.0222724676132202, |
|
"learning_rate": 4.535608811207272e-05, |
|
"loss": 2.0986, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.0501467143188378, |
|
"grad_norm": 1.254888892173767, |
|
"learning_rate": 4.4494958626484276e-05, |
|
"loss": 2.3094, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.05065841548535655, |
|
"grad_norm": 1.4137556552886963, |
|
"learning_rate": 4.36378790343179e-05, |
|
"loss": 2.4861, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.051170116651875305, |
|
"grad_norm": 2.0688607692718506, |
|
"learning_rate": 4.278517975607167e-05, |
|
"loss": 2.9678, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.051170116651875305, |
|
"eval_loss": 1.9229816198349, |
|
"eval_runtime": 1.5335, |
|
"eval_samples_per_second": 32.605, |
|
"eval_steps_per_second": 8.477, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.05168181781839406, |
|
"grad_norm": 0.40056556463241577, |
|
"learning_rate": 4.19371895235492e-05, |
|
"loss": 1.532, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.05219351898491281, |
|
"grad_norm": 0.6019428372383118, |
|
"learning_rate": 4.109423525312738e-05, |
|
"loss": 1.8149, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.052705220151431566, |
|
"grad_norm": 0.7042158842086792, |
|
"learning_rate": 4.0256641919723806e-05, |
|
"loss": 1.8967, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.05321692131795032, |
|
"grad_norm": 0.725395917892456, |
|
"learning_rate": 3.942473243151281e-05, |
|
"loss": 1.939, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.05372862248446907, |
|
"grad_norm": 0.8081313371658325, |
|
"learning_rate": 3.859882750543826e-05, |
|
"loss": 2.0155, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.054240323650987826, |
|
"grad_norm": 0.7761573195457458, |
|
"learning_rate": 3.777924554357096e-05, |
|
"loss": 2.0242, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.05475202481750658, |
|
"grad_norm": 0.7627308368682861, |
|
"learning_rate": 3.6966302510358666e-05, |
|
"loss": 2.0408, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.055263725984025326, |
|
"grad_norm": 0.954046905040741, |
|
"learning_rate": 3.616031181081575e-05, |
|
"loss": 2.0613, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.05577542715054408, |
|
"grad_norm": 1.0556938648223877, |
|
"learning_rate": 3.53615841696996e-05, |
|
"loss": 2.1036, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.05628712831706283, |
|
"grad_norm": 1.131365418434143, |
|
"learning_rate": 3.45704275117204e-05, |
|
"loss": 2.1926, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.05679882948358159, |
|
"grad_norm": 1.4441267251968384, |
|
"learning_rate": 3.378714684283011e-05, |
|
"loss": 2.3588, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.05731053065010034, |
|
"grad_norm": 1.711294412612915, |
|
"learning_rate": 3.301204413263704e-05, |
|
"loss": 2.615, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.057822231816619094, |
|
"grad_norm": 0.3940889239311218, |
|
"learning_rate": 3.224541819799071e-05, |
|
"loss": 1.4173, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.05833393298313785, |
|
"grad_norm": 0.5546037554740906, |
|
"learning_rate": 3.1487564587782306e-05, |
|
"loss": 1.6893, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.0588456341496566, |
|
"grad_norm": 0.6007658243179321, |
|
"learning_rate": 3.0738775469004985e-05, |
|
"loss": 1.8685, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.059357335316175354, |
|
"grad_norm": 0.675112247467041, |
|
"learning_rate": 2.9999339514117912e-05, |
|
"loss": 1.9225, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.05986903648269411, |
|
"grad_norm": 0.7189555168151855, |
|
"learning_rate": 2.926954178975746e-05, |
|
"loss": 1.9709, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.06038073764921286, |
|
"grad_norm": 0.7457597255706787, |
|
"learning_rate": 2.854966364683872e-05, |
|
"loss": 1.9753, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.060892438815731614, |
|
"grad_norm": 0.869260847568512, |
|
"learning_rate": 2.783998261208929e-05, |
|
"loss": 2.0098, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.06140413998225037, |
|
"grad_norm": 0.9872080087661743, |
|
"learning_rate": 2.7140772281057468e-05, |
|
"loss": 2.0447, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.06191584114876912, |
|
"grad_norm": 1.0225969552993774, |
|
"learning_rate": 2.645230221263596e-05, |
|
"loss": 2.0648, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.062427542315287875, |
|
"grad_norm": 1.0745333433151245, |
|
"learning_rate": 2.577483782514174e-05, |
|
"loss": 2.1066, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.06293924348180663, |
|
"grad_norm": 1.2674683332443237, |
|
"learning_rate": 2.5108640293992337e-05, |
|
"loss": 2.2943, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.06345094464832537, |
|
"grad_norm": 1.3520656824111938, |
|
"learning_rate": 2.445396645101762e-05, |
|
"loss": 2.3662, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.06396264581484414, |
|
"grad_norm": 1.9982293844223022, |
|
"learning_rate": 2.3811068685446395e-05, |
|
"loss": 2.8426, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.06396264581484414, |
|
"eval_loss": 1.8993180990219116, |
|
"eval_runtime": 1.5346, |
|
"eval_samples_per_second": 32.582, |
|
"eval_steps_per_second": 8.471, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.06447434698136288, |
|
"grad_norm": 0.36793649196624756, |
|
"learning_rate": 2.3180194846605367e-05, |
|
"loss": 1.4921, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.06498604814788164, |
|
"grad_norm": 0.5427601337432861, |
|
"learning_rate": 2.2561588148368657e-05, |
|
"loss": 1.8148, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.06549774931440039, |
|
"grad_norm": 0.6203567385673523, |
|
"learning_rate": 2.195548707539416e-05, |
|
"loss": 1.8883, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.06600945048091915, |
|
"grad_norm": 0.6660148501396179, |
|
"learning_rate": 2.136212529118314e-05, |
|
"loss": 1.9282, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.0665211516474379, |
|
"grad_norm": 0.7544979453086853, |
|
"learning_rate": 2.0781731547998614e-05, |
|
"loss": 1.9454, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.06703285281395666, |
|
"grad_norm": 0.7683319449424744, |
|
"learning_rate": 2.0214529598676836e-05, |
|
"loss": 1.9783, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.0675445539804754, |
|
"grad_norm": 0.8448789715766907, |
|
"learning_rate": 1.966073811036649e-05, |
|
"loss": 1.9902, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.06805625514699415, |
|
"grad_norm": 0.9101701378822327, |
|
"learning_rate": 1.9120570580228186e-05, |
|
"loss": 2.0101, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.06856795631351291, |
|
"grad_norm": 1.002808928489685, |
|
"learning_rate": 1.8594235253127375e-05, |
|
"loss": 2.1164, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.06907965748003166, |
|
"grad_norm": 1.1789017915725708, |
|
"learning_rate": 1.8081935041351887e-05, |
|
"loss": 2.2391, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.06959135864655042, |
|
"grad_norm": 1.4245861768722534, |
|
"learning_rate": 1.758386744638546e-05, |
|
"loss": 2.3897, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.07010305981306916, |
|
"grad_norm": 1.6070505380630493, |
|
"learning_rate": 1.7100224482767114e-05, |
|
"loss": 2.5421, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.07061476097958792, |
|
"grad_norm": 0.4116235673427582, |
|
"learning_rate": 1.6631192604065855e-05, |
|
"loss": 1.4141, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.07112646214610667, |
|
"grad_norm": 0.5044823288917542, |
|
"learning_rate": 1.617695263099937e-05, |
|
"loss": 1.8181, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.07163816331262543, |
|
"grad_norm": 0.5927360653877258, |
|
"learning_rate": 1.573767968172413e-05, |
|
"loss": 1.8575, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.07214986447914418, |
|
"grad_norm": 0.6108676791191101, |
|
"learning_rate": 1.531354310432403e-05, |
|
"loss": 1.9401, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.07266156564566294, |
|
"grad_norm": 0.6655164361000061, |
|
"learning_rate": 1.490470641152345e-05, |
|
"loss": 1.9377, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.07317326681218168, |
|
"grad_norm": 0.758387565612793, |
|
"learning_rate": 1.4511327217650046e-05, |
|
"loss": 1.9332, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.07368496797870044, |
|
"grad_norm": 0.8023284673690796, |
|
"learning_rate": 1.413355717787134e-05, |
|
"loss": 1.9804, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.07419666914521919, |
|
"grad_norm": 0.8641353249549866, |
|
"learning_rate": 1.3771541929728886e-05, |
|
"loss": 1.9505, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.07470837031173795, |
|
"grad_norm": 1.0415940284729004, |
|
"learning_rate": 1.3425421036992098e-05, |
|
"loss": 2.0321, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.0752200714782567, |
|
"grad_norm": 1.1581557989120483, |
|
"learning_rate": 1.309532793585384e-05, |
|
"loss": 2.0956, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.07573177264477546, |
|
"grad_norm": 1.1987535953521729, |
|
"learning_rate": 1.2781389883488218e-05, |
|
"loss": 2.2152, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.0762434738112942, |
|
"grad_norm": 1.6775120496749878, |
|
"learning_rate": 1.248372790899049e-05, |
|
"loss": 2.586, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.07675517497781295, |
|
"grad_norm": 2.6760313510894775, |
|
"learning_rate": 1.2202456766718093e-05, |
|
"loss": 3.0532, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.07675517497781295, |
|
"eval_loss": 1.8791134357452393, |
|
"eval_runtime": 1.5354, |
|
"eval_samples_per_second": 32.565, |
|
"eval_steps_per_second": 8.467, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.07726687614433171, |
|
"grad_norm": 0.33062976598739624, |
|
"learning_rate": 1.1937684892050604e-05, |
|
"loss": 1.3588, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.07777857731085046, |
|
"grad_norm": 0.5277450084686279, |
|
"learning_rate": 1.168951435958588e-05, |
|
"loss": 1.7291, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.07829027847736922, |
|
"grad_norm": 0.5764089226722717, |
|
"learning_rate": 1.1458040843788312e-05, |
|
"loss": 1.8712, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.07880197964388796, |
|
"grad_norm": 0.6534531712532043, |
|
"learning_rate": 1.1243353582104556e-05, |
|
"loss": 1.9575, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.07931368081040673, |
|
"grad_norm": 0.7147605419158936, |
|
"learning_rate": 1.1045535340560744e-05, |
|
"loss": 1.9211, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.07982538197692547, |
|
"grad_norm": 0.7351657748222351, |
|
"learning_rate": 1.0864662381854632e-05, |
|
"loss": 1.9539, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.08033708314344423, |
|
"grad_norm": 0.8152849674224854, |
|
"learning_rate": 1.070080443595488e-05, |
|
"loss": 2.0232, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.08084878430996298, |
|
"grad_norm": 0.9168938398361206, |
|
"learning_rate": 1.0554024673218807e-05, |
|
"loss": 1.9743, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.08136048547648174, |
|
"grad_norm": 1.0510733127593994, |
|
"learning_rate": 1.0424379680039025e-05, |
|
"loss": 2.0778, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.08187218664300049, |
|
"grad_norm": 1.2245820760726929, |
|
"learning_rate": 1.0311919437028318e-05, |
|
"loss": 2.2151, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.08238388780951925, |
|
"grad_norm": 1.3239753246307373, |
|
"learning_rate": 1.0216687299751144e-05, |
|
"loss": 2.2195, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.08289558897603799, |
|
"grad_norm": 1.7548948526382446, |
|
"learning_rate": 1.0138719982009242e-05, |
|
"loss": 2.5937, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.08340729014255675, |
|
"grad_norm": 0.4118451774120331, |
|
"learning_rate": 1.007804754168779e-05, |
|
"loss": 1.4343, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.0839189913090755, |
|
"grad_norm": 0.5209285616874695, |
|
"learning_rate": 1.003469336916747e-05, |
|
"loss": 1.7331, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.08443069247559425, |
|
"grad_norm": 0.5564490556716919, |
|
"learning_rate": 1.0008674178307085e-05, |
|
"loss": 1.8639, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.084942393642113, |
|
"grad_norm": 0.6351719498634338, |
|
"learning_rate": 1e-05, |
|
"loss": 1.9287, |
|
"step": 166 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 166, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 50, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 30, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.944151006450811e+18, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|