|
{ |
|
"best_metric": 0.3995036053381367, |
|
"best_model_checkpoint": "esm2_t12_35M-lora-binding-sites_2023-09-11_16-28-57/checkpoint-57940", |
|
"epoch": 2.0, |
|
"eval_steps": 500, |
|
"global_step": 57940, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0005701554508678478, |
|
"loss": 0.4954, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000570151359481901, |
|
"loss": 0.3627, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0005701445314265366, |
|
"loss": 0.325, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0005701349667673032, |
|
"loss": 0.2797, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0005701226655960202, |
|
"loss": 0.2708, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0005701076280307777, |
|
"loss": 0.2547, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0005700898542159346, |
|
"loss": 0.2395, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0005700693443221178, |
|
"loss": 0.2348, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0005700460985462198, |
|
"loss": 0.218, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0005700202538231217, |
|
"loss": 0.2284, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.000569991550655178, |
|
"loss": 0.2307, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.000569960112351964, |
|
"loss": 0.22, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.000569925939215284, |
|
"loss": 0.2117, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0005698890315731962, |
|
"loss": 0.2246, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0005698493897800102, |
|
"loss": 0.2392, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0005698070142162829, |
|
"loss": 0.2078, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0005697619052888148, |
|
"loss": 0.2098, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0005697140634306472, |
|
"loss": 0.1919, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0005696634891010563, |
|
"loss": 0.1957, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0005696101827855504, |
|
"loss": 0.2156, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0005695544319784758, |
|
"loss": 0.1983, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0005694956769058657, |
|
"loss": 0.2104, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0005694341914583195, |
|
"loss": 0.1789, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0005693699762260905, |
|
"loss": 0.1944, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0005693030318256384, |
|
"loss": 0.1895, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0005692333588996219, |
|
"loss": 0.2007, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0005691609581168934, |
|
"loss": 0.1852, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0005690858301724929, |
|
"loss": 0.1728, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0005690079757876402, |
|
"loss": 0.2076, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0005689273957097291, |
|
"loss": 0.1844, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0005688445140142318, |
|
"loss": 0.173, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0005687584985156089, |
|
"loss": 0.189, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0005686697597188824, |
|
"loss": 0.1728, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0005685782984759349, |
|
"loss": 0.1764, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0005684841156647846, |
|
"loss": 0.1817, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0005683872121895759, |
|
"loss": 0.184, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.000568287588980572, |
|
"loss": 0.1611, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0005681852469941448, |
|
"loss": 0.1782, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.000568080187212766, |
|
"loss": 0.1674, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0005679724106449985, |
|
"loss": 0.1592, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0005678619183254852, |
|
"loss": 0.188, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0005677487113149407, |
|
"loss": 0.1724, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00056763279070014, |
|
"loss": 0.1734, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0005675141575939082, |
|
"loss": 0.1802, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0005673934265999777, |
|
"loss": 0.1735, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0005672693855015051, |
|
"loss": 0.1768, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0005671426354002519, |
|
"loss": 0.1759, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0005670131775130054, |
|
"loss": 0.1646, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0005668810130825471, |
|
"loss": 0.1684, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0005667461433776413, |
|
"loss": 0.1636, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0005666085696930225, |
|
"loss": 0.1492, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0005664682933493831, |
|
"loss": 0.1636, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0005663253156933608, |
|
"loss": 0.1647, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0005661796380975255, |
|
"loss": 0.181, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0005660320105513283, |
|
"loss": 0.1708, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0005658809507792387, |
|
"loss": 0.1702, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0005657271953331912, |
|
"loss": 0.1617, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.000565570745689221, |
|
"loss": 0.1722, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.000565411603349228, |
|
"loss": 0.1702, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0005652497698409615, |
|
"loss": 0.1543, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0005650852467180052, |
|
"loss": 0.1654, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0005649180355597633, |
|
"loss": 0.1676, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.000564748137971445, |
|
"loss": 0.16, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0005645755555840486, |
|
"loss": 0.1627, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0005644002900543462, |
|
"loss": 0.1712, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.0005642232394671313, |
|
"loss": 0.1603, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.0005640426261206144, |
|
"loss": 0.1648, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.0005638593347478565, |
|
"loss": 0.1608, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.0005636733671084348, |
|
"loss": 0.162, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.0005634847249876185, |
|
"loss": 0.1516, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.0005632934101963515, |
|
"loss": 0.1638, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.000563099424571235, |
|
"loss": 0.1691, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0005629027699745098, |
|
"loss": 0.149, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.0005627034482940388, |
|
"loss": 0.1713, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0005625014614432885, |
|
"loss": 0.1615, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0005622978412332689, |
|
"loss": 0.1618, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.0005620905431860859, |
|
"loss": 0.1695, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0005618805858524459, |
|
"loss": 0.1575, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.0005616679712479168, |
|
"loss": 0.1494, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.0005614527014135753, |
|
"loss": 0.1441, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.0005612347784159886, |
|
"loss": 0.1554, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.000561014204347194, |
|
"loss": 0.1597, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.0005607909813246783, |
|
"loss": 0.1603, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.0005605651114913586, |
|
"loss": 0.1566, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 0.0005603365970155604, |
|
"loss": 0.148, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 0.0005601054400909981, |
|
"loss": 0.153, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.0005598728184863933, |
|
"loss": 0.1551, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 0.000559636396531194, |
|
"loss": 0.1808, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 0.0005593973388490791, |
|
"loss": 0.1537, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 0.0005591556477349764, |
|
"loss": 0.1488, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 0.0005589113255090943, |
|
"loss": 0.1632, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.0005586643745168998, |
|
"loss": 0.1595, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.0005584147971290952, |
|
"loss": 0.1692, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 0.0005581625957415966, |
|
"loss": 0.1454, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.0005579077727755101, |
|
"loss": 0.1608, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.0005576503306771081, |
|
"loss": 0.1506, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 0.0005573915787164281, |
|
"loss": 0.1639, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 0.000557128918857334, |
|
"loss": 0.1463, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 0.0005568636473428375, |
|
"loss": 0.159, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.0005565957667195164, |
|
"loss": 0.1792, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 0.0005563252795589956, |
|
"loss": 0.1512, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 0.0005560521884579225, |
|
"loss": 0.1604, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 0.000555776496037942, |
|
"loss": 0.1563, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 0.0005554982049456713, |
|
"loss": 0.1598, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 0.0005552173178526746, |
|
"loss": 0.1429, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 0.0005549338374554378, |
|
"loss": 0.1544, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.0005546477664753415, |
|
"loss": 0.1435, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 0.00055436055738539, |
|
"loss": 0.1562, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 0.0005540693264215625, |
|
"loss": 0.1501, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 0.0005537755131740878, |
|
"loss": 0.1511, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.0005534791204635411, |
|
"loss": 0.1456, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.0005531801511352604, |
|
"loss": 0.1507, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.0005528786080593183, |
|
"loss": 0.1477, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 0.000552574494130496, |
|
"loss": 0.1507, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 0.000552267812268254, |
|
"loss": 0.1494, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.0005519585654167048, |
|
"loss": 0.1489, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 0.0005516467565445848, |
|
"loss": 0.1495, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 0.0005513323886452253, |
|
"loss": 0.1467, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 0.0005510154647365238, |
|
"loss": 0.1465, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.0005506975915907354, |
|
"loss": 0.1725, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.0005503755775569938, |
|
"loss": 0.1604, |
|
"step": 24200 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.0005500510166991929, |
|
"loss": 0.1492, |
|
"step": 24400 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 0.0005497239121330815, |
|
"loss": 0.1446, |
|
"step": 24600 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.0005493942669988278, |
|
"loss": 0.1492, |
|
"step": 24800 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.0005490620844609896, |
|
"loss": 0.152, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 0.0005487273677084831, |
|
"loss": 0.1601, |
|
"step": 25200 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 0.0005483901199545525, |
|
"loss": 0.1432, |
|
"step": 25400 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 0.0005480503444367399, |
|
"loss": 0.153, |
|
"step": 25600 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.0005477080444168533, |
|
"loss": 0.1446, |
|
"step": 25800 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 0.0005473649535531743, |
|
"loss": 0.1543, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 0.0005470176269927267, |
|
"loss": 0.1652, |
|
"step": 26200 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 0.0005466677858441803, |
|
"loss": 0.1463, |
|
"step": 26400 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 0.0005463154334659722, |
|
"loss": 0.152, |
|
"step": 26600 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 0.0005459605732406469, |
|
"loss": 0.1572, |
|
"step": 26800 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 0.0005456032085748239, |
|
"loss": 0.1439, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 0.0005452433428991655, |
|
"loss": 0.1483, |
|
"step": 27200 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.000544880979668343, |
|
"loss": 0.1633, |
|
"step": 27400 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.0005445161223610037, |
|
"loss": 0.1559, |
|
"step": 27600 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 0.0005441487744797385, |
|
"loss": 0.1582, |
|
"step": 27800 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.000543778939551047, |
|
"loss": 0.1607, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.0005434084888892294, |
|
"loss": 0.1444, |
|
"step": 28200 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 0.0005430337029313361, |
|
"loss": 0.1411, |
|
"step": 28400 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 0.0005426564406305822, |
|
"loss": 0.145, |
|
"step": 28600 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 0.0005422767056086448, |
|
"loss": 0.1543, |
|
"step": 28800 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.9266715807717144, |
|
"eval_auc": 0.8547374969101048, |
|
"eval_f1": 0.3476562916154915, |
|
"eval_loss": 0.36462366580963135, |
|
"eval_precision": 0.22375778821030992, |
|
"eval_recall": 0.7790042059793111, |
|
"eval_runtime": 1778.9025, |
|
"eval_samples_per_second": 24.978, |
|
"eval_steps_per_second": 4.163, |
|
"step": 28970 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.000541894501510939, |
|
"loss": 0.1532, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 0.0005415098320065826, |
|
"loss": 0.146, |
|
"step": 29200 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 0.0005411227007883609, |
|
"loss": 0.142, |
|
"step": 29400 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 0.0005407331115726918, |
|
"loss": 0.1398, |
|
"step": 29600 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.0005403410680995892, |
|
"loss": 0.1609, |
|
"step": 29800 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 0.000539946574132628, |
|
"loss": 0.1443, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 0.0005395496334589074, |
|
"loss": 0.1466, |
|
"step": 30200 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 0.0005391502498890146, |
|
"loss": 0.1542, |
|
"step": 30400 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 0.0005387504424309447, |
|
"loss": 0.1624, |
|
"step": 30600 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 0.0005383461967606286, |
|
"loss": 0.1492, |
|
"step": 30800 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.000537939519747002, |
|
"loss": 0.1481, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 0.0005375304152941202, |
|
"loss": 0.1534, |
|
"step": 31200 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 0.0005371188873293412, |
|
"loss": 0.1469, |
|
"step": 31400 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.000536704939803289, |
|
"loss": 0.1443, |
|
"step": 31600 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 0.0005362885766898149, |
|
"loss": 0.148, |
|
"step": 31800 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 0.0005358698019859594, |
|
"loss": 0.1589, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 0.0005354486197119143, |
|
"loss": 0.1525, |
|
"step": 32200 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 0.0005350250339109836, |
|
"loss": 0.1514, |
|
"step": 32400 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.0005346011845377551, |
|
"loss": 0.1421, |
|
"step": 32600 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.0005341728158718659, |
|
"loss": 0.1384, |
|
"step": 32800 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 0.0005337442156678716, |
|
"loss": 0.1629, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.0005333110805040125, |
|
"loss": 0.1488, |
|
"step": 33200 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.000532875562333412, |
|
"loss": 0.1424, |
|
"step": 33400 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 0.0005324376653369972, |
|
"loss": 0.1436, |
|
"step": 33600 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 0.0005319973937185317, |
|
"loss": 0.1624, |
|
"step": 33800 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 0.0005315547517045754, |
|
"loss": 0.1381, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 0.0005311097435444433, |
|
"loss": 0.1484, |
|
"step": 34200 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 0.0005306623735101656, |
|
"loss": 0.1482, |
|
"step": 34400 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 0.0005302126458964461, |
|
"loss": 0.1439, |
|
"step": 34600 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 0.0005297605650206207, |
|
"loss": 0.1519, |
|
"step": 34800 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 0.000529306135222617, |
|
"loss": 0.1298, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 0.0005288493608649115, |
|
"loss": 0.1469, |
|
"step": 35200 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 0.0005283902463324885, |
|
"loss": 0.1431, |
|
"step": 35400 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.0005279287960327977, |
|
"loss": 0.1501, |
|
"step": 35600 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 0.0005274650143957118, |
|
"loss": 0.142, |
|
"step": 35800 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 0.0005269989058734842, |
|
"loss": 0.1433, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 0.0005265304749407059, |
|
"loss": 0.1398, |
|
"step": 36200 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 0.0005260597260942631, |
|
"loss": 0.1553, |
|
"step": 36400 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 0.0005255866638532936, |
|
"loss": 0.1545, |
|
"step": 36600 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.0005251112927591435, |
|
"loss": 0.1526, |
|
"step": 36800 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 0.0005246336173753237, |
|
"loss": 0.1503, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 0.000524153642287466, |
|
"loss": 0.139, |
|
"step": 37200 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.0005236737891556027, |
|
"loss": 0.1486, |
|
"step": 37400 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 0.000523189239945608, |
|
"loss": 0.1375, |
|
"step": 37600 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 0.0005227024048974417, |
|
"loss": 0.142, |
|
"step": 37800 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.0005222132886846675, |
|
"loss": 0.1412, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 0.000521721896002748, |
|
"loss": 0.1485, |
|
"step": 38200 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 0.0005212282315689993, |
|
"loss": 0.1416, |
|
"step": 38400 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 0.0005207323001225466, |
|
"loss": 0.1385, |
|
"step": 38600 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 0.0005202341064242777, |
|
"loss": 0.1493, |
|
"step": 38800 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 0.0005197336552567981, |
|
"loss": 0.1454, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 0.0005192309514243847, |
|
"loss": 0.1472, |
|
"step": 39200 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 0.0005187259997529396, |
|
"loss": 0.1443, |
|
"step": 39400 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 0.000518218805089944, |
|
"loss": 0.147, |
|
"step": 39600 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 0.0005177093723044116, |
|
"loss": 0.1565, |
|
"step": 39800 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 0.000517200270163978, |
|
"loss": 0.153, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 0.0005166863869556556, |
|
"loss": 0.1404, |
|
"step": 40200 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 0.000516170280335844, |
|
"loss": 0.1326, |
|
"step": 40400 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 0.0005156519552591106, |
|
"loss": 0.147, |
|
"step": 40600 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 0.0005151314167013202, |
|
"loss": 0.151, |
|
"step": 40800 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 0.000514608669659586, |
|
"loss": 0.1479, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 0.0005140837191522229, |
|
"loss": 0.1621, |
|
"step": 41200 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 0.0005135565702186988, |
|
"loss": 0.1531, |
|
"step": 41400 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 0.0005130272279195864, |
|
"loss": 0.1364, |
|
"step": 41600 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 0.0005124956973365139, |
|
"loss": 0.1419, |
|
"step": 41800 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 0.0005119619835721174, |
|
"loss": 0.1305, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 0.000511426091749991, |
|
"loss": 0.1461, |
|
"step": 42200 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 0.0005108880270146381, |
|
"loss": 0.1501, |
|
"step": 42400 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 0.0005103505010775399, |
|
"loss": 0.1587, |
|
"step": 42600 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 0.0005098081168325064, |
|
"loss": 0.1322, |
|
"step": 42800 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 0.0005092635752066296, |
|
"loss": 0.1418, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 0.0005087196202411544, |
|
"loss": 0.1564, |
|
"step": 43200 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 0.0005081707902783097, |
|
"loss": 0.1552, |
|
"step": 43400 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 0.0005076198186527788, |
|
"loss": 0.1362, |
|
"step": 43600 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 0.0005070667106538294, |
|
"loss": 0.1497, |
|
"step": 43800 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 0.0005065114715912386, |
|
"loss": 0.143, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 0.0005059541067952407, |
|
"loss": 0.1527, |
|
"step": 44200 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 0.0005053946216164774, |
|
"loss": 0.1384, |
|
"step": 44400 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 0.0005048330214259455, |
|
"loss": 0.1569, |
|
"step": 44600 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 0.0005042693116149461, |
|
"loss": 0.1349, |
|
"step": 44800 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 0.0005037034975950318, |
|
"loss": 0.1458, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 0.0005031355847979559, |
|
"loss": 0.1448, |
|
"step": 45200 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 0.0005025655786756196, |
|
"loss": 0.1386, |
|
"step": 45400 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 0.0005019934847000197, |
|
"loss": 0.146, |
|
"step": 45600 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 0.0005014193083631962, |
|
"loss": 0.1366, |
|
"step": 45800 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 0.0005008430551771795, |
|
"loss": 0.1389, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 0.0005002647306739379, |
|
"loss": 0.1378, |
|
"step": 46200 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 0.0004996843404053234, |
|
"loss": 0.1375, |
|
"step": 46400 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 0.0004991018899430197, |
|
"loss": 0.1424, |
|
"step": 46600 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 0.0004985173848784881, |
|
"loss": 0.1509, |
|
"step": 46800 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 0.0004979308308229137, |
|
"loss": 0.1438, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 0.0004973422334071515, |
|
"loss": 0.1306, |
|
"step": 47200 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 0.000496754556516732, |
|
"loss": 0.1475, |
|
"step": 47400 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 0.0004961618994976319, |
|
"loss": 0.1422, |
|
"step": 47600 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 0.0004955672160998928, |
|
"loss": 0.1411, |
|
"step": 47800 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 0.0004949705120324106, |
|
"loss": 0.1491, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 0.0004943717930234795, |
|
"loss": 0.1391, |
|
"step": 48200 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 0.0004937710648207372, |
|
"loss": 0.1339, |
|
"step": 48400 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 0.0004931683331911087, |
|
"loss": 0.1494, |
|
"step": 48600 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 0.0004925636039207527, |
|
"loss": 0.154, |
|
"step": 48800 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 0.0004919568828150041, |
|
"loss": 0.1423, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 0.0004913481756983199, |
|
"loss": 0.1407, |
|
"step": 49200 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 0.0004907405467666192, |
|
"loss": 0.1452, |
|
"step": 49400 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 0.0004901278950345495, |
|
"loss": 0.1368, |
|
"step": 49600 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 0.0004895132748496281, |
|
"loss": 0.1264, |
|
"step": 49800 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 0.0004888966921121421, |
|
"loss": 0.1478, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 0.00048827815274121847, |
|
"loss": 0.1505, |
|
"step": 50200 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 0.0004876576626747684, |
|
"loss": 0.1328, |
|
"step": 50400 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 0.0004870352278694289, |
|
"loss": 0.1511, |
|
"step": 50600 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 0.00048641085430050653, |
|
"loss": 0.1475, |
|
"step": 50800 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 0.00048578454796191977, |
|
"loss": 0.1454, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 0.00048515631486614134, |
|
"loss": 0.1472, |
|
"step": 51200 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 0.0004845261610441408, |
|
"loss": 0.1295, |
|
"step": 51400 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 0.0004838972576405494, |
|
"loss": 0.1372, |
|
"step": 51600 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 0.00048326329006063116, |
|
"loss": 0.1442, |
|
"step": 51800 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 0.0004826274199273233, |
|
"loss": 0.1379, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 0.00048198965334490996, |
|
"loss": 0.1423, |
|
"step": 52200 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 0.0004813499964358812, |
|
"loss": 0.1413, |
|
"step": 52400 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 0.0004807084553408737, |
|
"loss": 0.1423, |
|
"step": 52600 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 0.0004800650362186124, |
|
"loss": 0.1455, |
|
"step": 52800 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 0.0004794197452458509, |
|
"loss": 0.1415, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 0.00047877258861731253, |
|
"loss": 0.1452, |
|
"step": 53200 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 0.00047812357254563043, |
|
"loss": 0.1361, |
|
"step": 53400 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 0.00047747596220724844, |
|
"loss": 0.1437, |
|
"step": 53600 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 0.0004768232551777713, |
|
"loss": 0.1301, |
|
"step": 53800 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 0.0004761687074185398, |
|
"loss": 0.1401, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 0.0004755123252131414, |
|
"loss": 0.1392, |
|
"step": 54200 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 0.00047485411486277384, |
|
"loss": 0.1334, |
|
"step": 54400 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 0.00047419408268618515, |
|
"loss": 0.1486, |
|
"step": 54600 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 0.0004735322350196125, |
|
"loss": 0.146, |
|
"step": 54800 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 0.00047286857821672145, |
|
"loss": 0.1531, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 0.00047220311864854545, |
|
"loss": 0.1312, |
|
"step": 55200 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 0.0004715358627034239, |
|
"loss": 0.1376, |
|
"step": 55400 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 0.00047086681678694143, |
|
"loss": 0.138, |
|
"step": 55600 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 0.00047019934589512296, |
|
"loss": 0.1476, |
|
"step": 55800 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 0.000469526748190841, |
|
"loss": 0.1394, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 0.0004688523798024791, |
|
"loss": 0.1527, |
|
"step": 56200 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 0.00046817624720390063, |
|
"loss": 0.1422, |
|
"step": 56400 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 0.00046750175069906427, |
|
"loss": 0.1615, |
|
"step": 56600 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 0.00046682211790916814, |
|
"loss": 0.1263, |
|
"step": 56800 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 0.00046614074039934907, |
|
"loss": 0.1503, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 0.00046545762471075726, |
|
"loss": 0.1557, |
|
"step": 57200 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 0.0004647727774012289, |
|
"loss": 0.1367, |
|
"step": 57400 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 0.0004640862050452239, |
|
"loss": 0.1341, |
|
"step": 57600 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 0.00046339791423376226, |
|
"loss": 0.1449, |
|
"step": 57800 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.9439565632043402, |
|
"eval_auc": 0.8461801873862141, |
|
"eval_f1": 0.3995036053381367, |
|
"eval_loss": 0.4224541485309601, |
|
"eval_precision": 0.27316793551875035, |
|
"eval_recall": 0.7432397408207343, |
|
"eval_runtime": 1778.6249, |
|
"eval_samples_per_second": 24.982, |
|
"eval_steps_per_second": 4.164, |
|
"step": 57940 |
|
} |
|
], |
|
"logging_steps": 200, |
|
"max_steps": 202790, |
|
"num_train_epochs": 7, |
|
"save_steps": 500, |
|
"total_flos": 6.9510044394e+16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|