uztext_568Mb_Roberta_BPE / trainer_state.json
rifkat's picture
Add model weights and configuation
2f8ad03
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 9.15017728468489,
"global_step": 80000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.06,
"learning_rate": 4.97140569598536e-05,
"loss": 8.5722,
"step": 500
},
{
"epoch": 0.11,
"learning_rate": 4.94281139197072e-05,
"loss": 7.9252,
"step": 1000
},
{
"epoch": 0.17,
"learning_rate": 4.91421708795608e-05,
"loss": 7.6163,
"step": 1500
},
{
"epoch": 0.23,
"learning_rate": 4.885622783941439e-05,
"loss": 7.3317,
"step": 2000
},
{
"epoch": 0.29,
"learning_rate": 4.857028479926799e-05,
"loss": 6.9857,
"step": 2500
},
{
"epoch": 0.34,
"learning_rate": 4.8284341759121586e-05,
"loss": 6.6297,
"step": 3000
},
{
"epoch": 0.4,
"learning_rate": 4.7998398718975185e-05,
"loss": 6.3251,
"step": 3500
},
{
"epoch": 0.46,
"learning_rate": 4.7712455678828777e-05,
"loss": 6.0786,
"step": 4000
},
{
"epoch": 0.51,
"learning_rate": 4.7426512638682375e-05,
"loss": 5.8383,
"step": 4500
},
{
"epoch": 0.57,
"learning_rate": 4.7140569598535973e-05,
"loss": 5.62,
"step": 5000
},
{
"epoch": 0.63,
"learning_rate": 4.685462655838957e-05,
"loss": 5.4495,
"step": 5500
},
{
"epoch": 0.69,
"learning_rate": 4.6568683518243164e-05,
"loss": 5.2876,
"step": 6000
},
{
"epoch": 0.74,
"learning_rate": 4.628274047809676e-05,
"loss": 5.1378,
"step": 6500
},
{
"epoch": 0.8,
"learning_rate": 4.599679743795036e-05,
"loss": 5.0281,
"step": 7000
},
{
"epoch": 0.86,
"learning_rate": 4.571085439780396e-05,
"loss": 4.9177,
"step": 7500
},
{
"epoch": 0.92,
"learning_rate": 4.542491135765756e-05,
"loss": 4.8254,
"step": 8000
},
{
"epoch": 0.97,
"learning_rate": 4.5138968317511156e-05,
"loss": 4.7354,
"step": 8500
},
{
"epoch": 1.03,
"learning_rate": 4.4853025277364754e-05,
"loss": 4.6366,
"step": 9000
},
{
"epoch": 1.09,
"learning_rate": 4.4567082237218346e-05,
"loss": 4.5642,
"step": 9500
},
{
"epoch": 1.14,
"learning_rate": 4.4281139197071944e-05,
"loss": 4.5011,
"step": 10000
},
{
"epoch": 1.2,
"learning_rate": 4.399519615692554e-05,
"loss": 4.4285,
"step": 10500
},
{
"epoch": 1.26,
"learning_rate": 4.370925311677914e-05,
"loss": 4.3703,
"step": 11000
},
{
"epoch": 1.32,
"learning_rate": 4.342331007663274e-05,
"loss": 4.3203,
"step": 11500
},
{
"epoch": 1.37,
"learning_rate": 4.313736703648633e-05,
"loss": 4.2678,
"step": 12000
},
{
"epoch": 1.43,
"learning_rate": 4.285142399633993e-05,
"loss": 4.2177,
"step": 12500
},
{
"epoch": 1.49,
"learning_rate": 4.256548095619353e-05,
"loss": 4.1607,
"step": 13000
},
{
"epoch": 1.54,
"learning_rate": 4.227953791604713e-05,
"loss": 4.132,
"step": 13500
},
{
"epoch": 1.6,
"learning_rate": 4.199359487590072e-05,
"loss": 4.0738,
"step": 14000
},
{
"epoch": 1.66,
"learning_rate": 4.170765183575432e-05,
"loss": 4.0348,
"step": 14500
},
{
"epoch": 1.72,
"learning_rate": 4.1421708795607916e-05,
"loss": 3.9979,
"step": 15000
},
{
"epoch": 1.77,
"learning_rate": 4.1135765755461514e-05,
"loss": 3.9553,
"step": 15500
},
{
"epoch": 1.83,
"learning_rate": 4.084982271531511e-05,
"loss": 3.917,
"step": 16000
},
{
"epoch": 1.89,
"learning_rate": 4.056387967516871e-05,
"loss": 3.8866,
"step": 16500
},
{
"epoch": 1.94,
"learning_rate": 4.027793663502231e-05,
"loss": 3.8605,
"step": 17000
},
{
"epoch": 2.0,
"learning_rate": 3.99919935948759e-05,
"loss": 3.8163,
"step": 17500
},
{
"epoch": 2.06,
"learning_rate": 3.97060505547295e-05,
"loss": 3.7707,
"step": 18000
},
{
"epoch": 2.12,
"learning_rate": 3.94201075145831e-05,
"loss": 3.7489,
"step": 18500
},
{
"epoch": 2.17,
"learning_rate": 3.9134164474436696e-05,
"loss": 3.7237,
"step": 19000
},
{
"epoch": 2.23,
"learning_rate": 3.8848221434290295e-05,
"loss": 3.6926,
"step": 19500
},
{
"epoch": 2.29,
"learning_rate": 3.8562278394143887e-05,
"loss": 3.6643,
"step": 20000
},
{
"epoch": 2.34,
"learning_rate": 3.8276335353997485e-05,
"loss": 3.6353,
"step": 20500
},
{
"epoch": 2.4,
"learning_rate": 3.7990392313851084e-05,
"loss": 3.6286,
"step": 21000
},
{
"epoch": 2.46,
"learning_rate": 3.7704449273704675e-05,
"loss": 3.6052,
"step": 21500
},
{
"epoch": 2.52,
"learning_rate": 3.7418506233558274e-05,
"loss": 3.5797,
"step": 22000
},
{
"epoch": 2.57,
"learning_rate": 3.713256319341187e-05,
"loss": 3.5624,
"step": 22500
},
{
"epoch": 2.63,
"learning_rate": 3.684662015326547e-05,
"loss": 3.5298,
"step": 23000
},
{
"epoch": 2.69,
"learning_rate": 3.656067711311907e-05,
"loss": 3.5125,
"step": 23500
},
{
"epoch": 2.75,
"learning_rate": 3.627473407297267e-05,
"loss": 3.4832,
"step": 24000
},
{
"epoch": 2.8,
"learning_rate": 3.5988791032826266e-05,
"loss": 3.4755,
"step": 24500
},
{
"epoch": 2.86,
"learning_rate": 3.570284799267986e-05,
"loss": 3.4514,
"step": 25000
},
{
"epoch": 2.92,
"learning_rate": 3.5416904952533456e-05,
"loss": 3.441,
"step": 25500
},
{
"epoch": 2.97,
"learning_rate": 3.5130961912387055e-05,
"loss": 3.4208,
"step": 26000
},
{
"epoch": 3.03,
"learning_rate": 3.484501887224065e-05,
"loss": 3.3992,
"step": 26500
},
{
"epoch": 3.09,
"learning_rate": 3.455907583209425e-05,
"loss": 3.3641,
"step": 27000
},
{
"epoch": 3.15,
"learning_rate": 3.427313279194784e-05,
"loss": 3.3579,
"step": 27500
},
{
"epoch": 3.2,
"learning_rate": 3.398718975180144e-05,
"loss": 3.3453,
"step": 28000
},
{
"epoch": 3.26,
"learning_rate": 3.370124671165504e-05,
"loss": 3.3162,
"step": 28500
},
{
"epoch": 3.32,
"learning_rate": 3.341530367150864e-05,
"loss": 3.2976,
"step": 29000
},
{
"epoch": 3.37,
"learning_rate": 3.312936063136223e-05,
"loss": 3.2955,
"step": 29500
},
{
"epoch": 3.43,
"learning_rate": 3.284341759121583e-05,
"loss": 3.2894,
"step": 30000
},
{
"epoch": 3.49,
"learning_rate": 3.255747455106943e-05,
"loss": 3.2687,
"step": 30500
},
{
"epoch": 3.55,
"learning_rate": 3.2271531510923026e-05,
"loss": 3.2548,
"step": 31000
},
{
"epoch": 3.6,
"learning_rate": 3.1985588470776624e-05,
"loss": 3.2446,
"step": 31500
},
{
"epoch": 3.66,
"learning_rate": 3.169964543063022e-05,
"loss": 3.2264,
"step": 32000
},
{
"epoch": 3.72,
"learning_rate": 3.141370239048382e-05,
"loss": 3.2166,
"step": 32500
},
{
"epoch": 3.77,
"learning_rate": 3.112775935033741e-05,
"loss": 3.2097,
"step": 33000
},
{
"epoch": 3.83,
"learning_rate": 3.084181631019101e-05,
"loss": 3.1955,
"step": 33500
},
{
"epoch": 3.89,
"learning_rate": 3.055587327004461e-05,
"loss": 3.181,
"step": 34000
},
{
"epoch": 3.95,
"learning_rate": 3.0269930229898208e-05,
"loss": 3.1758,
"step": 34500
},
{
"epoch": 4.0,
"learning_rate": 2.9983987189751807e-05,
"loss": 3.1688,
"step": 35000
},
{
"epoch": 4.06,
"learning_rate": 2.9698044149605398e-05,
"loss": 3.1284,
"step": 35500
},
{
"epoch": 4.12,
"learning_rate": 2.9412101109458993e-05,
"loss": 3.1256,
"step": 36000
},
{
"epoch": 4.17,
"learning_rate": 2.9126158069312592e-05,
"loss": 3.1062,
"step": 36500
},
{
"epoch": 4.23,
"learning_rate": 2.884021502916619e-05,
"loss": 3.1051,
"step": 37000
},
{
"epoch": 4.29,
"learning_rate": 2.855427198901979e-05,
"loss": 3.108,
"step": 37500
},
{
"epoch": 4.35,
"learning_rate": 2.8268328948873384e-05,
"loss": 3.0932,
"step": 38000
},
{
"epoch": 4.4,
"learning_rate": 2.7982385908726982e-05,
"loss": 3.0766,
"step": 38500
},
{
"epoch": 4.46,
"learning_rate": 2.769644286858058e-05,
"loss": 3.0501,
"step": 39000
},
{
"epoch": 4.52,
"learning_rate": 2.7410499828434176e-05,
"loss": 3.0634,
"step": 39500
},
{
"epoch": 4.58,
"learning_rate": 2.7124556788287774e-05,
"loss": 3.0524,
"step": 40000
},
{
"epoch": 4.63,
"learning_rate": 2.6838613748141373e-05,
"loss": 3.0556,
"step": 40500
},
{
"epoch": 4.69,
"learning_rate": 2.655267070799497e-05,
"loss": 3.0276,
"step": 41000
},
{
"epoch": 4.75,
"learning_rate": 2.6266727667848566e-05,
"loss": 3.0364,
"step": 41500
},
{
"epoch": 4.8,
"learning_rate": 2.5980784627702165e-05,
"loss": 3.0186,
"step": 42000
},
{
"epoch": 4.86,
"learning_rate": 2.5694841587555763e-05,
"loss": 3.0071,
"step": 42500
},
{
"epoch": 4.92,
"learning_rate": 2.540889854740936e-05,
"loss": 3.0144,
"step": 43000
},
{
"epoch": 4.98,
"learning_rate": 2.5122955507262953e-05,
"loss": 3.0003,
"step": 43500
},
{
"epoch": 5.03,
"learning_rate": 2.4837012467116552e-05,
"loss": 2.9854,
"step": 44000
},
{
"epoch": 5.09,
"learning_rate": 2.455106942697015e-05,
"loss": 2.9696,
"step": 44500
},
{
"epoch": 5.15,
"learning_rate": 2.426512638682375e-05,
"loss": 2.9651,
"step": 45000
},
{
"epoch": 5.2,
"learning_rate": 2.397918334667734e-05,
"loss": 2.9463,
"step": 45500
},
{
"epoch": 5.26,
"learning_rate": 2.369324030653094e-05,
"loss": 2.9379,
"step": 46000
},
{
"epoch": 5.32,
"learning_rate": 2.3407297266384537e-05,
"loss": 2.9409,
"step": 46500
},
{
"epoch": 5.38,
"learning_rate": 2.3121354226238136e-05,
"loss": 2.9315,
"step": 47000
},
{
"epoch": 5.43,
"learning_rate": 2.283541118609173e-05,
"loss": 2.9407,
"step": 47500
},
{
"epoch": 5.49,
"learning_rate": 2.254946814594533e-05,
"loss": 2.9327,
"step": 48000
},
{
"epoch": 5.55,
"learning_rate": 2.2263525105798928e-05,
"loss": 2.9282,
"step": 48500
},
{
"epoch": 5.6,
"learning_rate": 2.1977582065652523e-05,
"loss": 2.9177,
"step": 49000
},
{
"epoch": 5.66,
"learning_rate": 2.1691639025506118e-05,
"loss": 2.9109,
"step": 49500
},
{
"epoch": 5.72,
"learning_rate": 2.1405695985359716e-05,
"loss": 2.9009,
"step": 50000
},
{
"epoch": 5.78,
"learning_rate": 2.1119752945213315e-05,
"loss": 2.8929,
"step": 50500
},
{
"epoch": 5.83,
"learning_rate": 2.0833809905066913e-05,
"loss": 2.8863,
"step": 51000
},
{
"epoch": 5.89,
"learning_rate": 2.0547866864920508e-05,
"loss": 2.8912,
"step": 51500
},
{
"epoch": 5.95,
"learning_rate": 2.0261923824774107e-05,
"loss": 2.8757,
"step": 52000
},
{
"epoch": 6.0,
"learning_rate": 1.9975980784627705e-05,
"loss": 2.874,
"step": 52500
},
{
"epoch": 6.06,
"learning_rate": 1.96900377444813e-05,
"loss": 2.8468,
"step": 53000
},
{
"epoch": 6.12,
"learning_rate": 1.9404094704334895e-05,
"loss": 2.866,
"step": 53500
},
{
"epoch": 6.18,
"learning_rate": 1.9118151664188494e-05,
"loss": 2.8428,
"step": 54000
},
{
"epoch": 6.23,
"learning_rate": 1.8832208624042092e-05,
"loss": 2.8474,
"step": 54500
},
{
"epoch": 6.29,
"learning_rate": 1.8546265583895687e-05,
"loss": 2.8368,
"step": 55000
},
{
"epoch": 6.35,
"learning_rate": 1.8260322543749286e-05,
"loss": 2.8377,
"step": 55500
},
{
"epoch": 6.41,
"learning_rate": 1.7974379503602884e-05,
"loss": 2.838,
"step": 56000
},
{
"epoch": 6.46,
"learning_rate": 1.7688436463456483e-05,
"loss": 2.83,
"step": 56500
},
{
"epoch": 6.52,
"learning_rate": 1.7402493423310078e-05,
"loss": 2.8275,
"step": 57000
},
{
"epoch": 6.58,
"learning_rate": 1.7116550383163673e-05,
"loss": 2.8106,
"step": 57500
},
{
"epoch": 6.63,
"learning_rate": 1.683060734301727e-05,
"loss": 2.8158,
"step": 58000
},
{
"epoch": 6.69,
"learning_rate": 1.654466430287087e-05,
"loss": 2.814,
"step": 58500
},
{
"epoch": 6.75,
"learning_rate": 1.6258721262724465e-05,
"loss": 2.8101,
"step": 59000
},
{
"epoch": 6.81,
"learning_rate": 1.5972778222578063e-05,
"loss": 2.8015,
"step": 59500
},
{
"epoch": 6.86,
"learning_rate": 1.5686835182431662e-05,
"loss": 2.7999,
"step": 60000
},
{
"epoch": 6.92,
"learning_rate": 1.540089214228526e-05,
"loss": 2.7919,
"step": 60500
},
{
"epoch": 6.98,
"learning_rate": 1.5114949102138854e-05,
"loss": 2.7953,
"step": 61000
},
{
"epoch": 7.03,
"learning_rate": 1.482900606199245e-05,
"loss": 2.7795,
"step": 61500
},
{
"epoch": 7.09,
"learning_rate": 1.4543063021846049e-05,
"loss": 2.7648,
"step": 62000
},
{
"epoch": 7.15,
"learning_rate": 1.4257119981699646e-05,
"loss": 2.773,
"step": 62500
},
{
"epoch": 7.21,
"learning_rate": 1.3971176941553244e-05,
"loss": 2.7663,
"step": 63000
},
{
"epoch": 7.26,
"learning_rate": 1.368523390140684e-05,
"loss": 2.7592,
"step": 63500
},
{
"epoch": 7.32,
"learning_rate": 1.3399290861260438e-05,
"loss": 2.7543,
"step": 64000
},
{
"epoch": 7.38,
"learning_rate": 1.3113347821114036e-05,
"loss": 2.7626,
"step": 64500
},
{
"epoch": 7.43,
"learning_rate": 1.2827404780967631e-05,
"loss": 2.755,
"step": 65000
},
{
"epoch": 7.49,
"learning_rate": 1.2541461740821228e-05,
"loss": 2.7538,
"step": 65500
},
{
"epoch": 7.55,
"learning_rate": 1.2255518700674826e-05,
"loss": 2.7381,
"step": 66000
},
{
"epoch": 7.61,
"learning_rate": 1.1969575660528423e-05,
"loss": 2.7491,
"step": 66500
},
{
"epoch": 7.66,
"learning_rate": 1.168363262038202e-05,
"loss": 2.7441,
"step": 67000
},
{
"epoch": 7.72,
"learning_rate": 1.1397689580235618e-05,
"loss": 2.7369,
"step": 67500
},
{
"epoch": 7.78,
"learning_rate": 1.1111746540089215e-05,
"loss": 2.7466,
"step": 68000
},
{
"epoch": 7.83,
"learning_rate": 1.0825803499942812e-05,
"loss": 2.7363,
"step": 68500
},
{
"epoch": 7.89,
"learning_rate": 1.0539860459796409e-05,
"loss": 2.7294,
"step": 69000
},
{
"epoch": 7.95,
"learning_rate": 1.0253917419650007e-05,
"loss": 2.7304,
"step": 69500
},
{
"epoch": 8.01,
"learning_rate": 9.967974379503602e-06,
"loss": 2.7278,
"step": 70000
},
{
"epoch": 8.06,
"learning_rate": 9.6820313393572e-06,
"loss": 2.7231,
"step": 70500
},
{
"epoch": 8.12,
"learning_rate": 9.396088299210797e-06,
"loss": 2.7267,
"step": 71000
},
{
"epoch": 8.18,
"learning_rate": 9.110145259064396e-06,
"loss": 2.7133,
"step": 71500
},
{
"epoch": 8.24,
"learning_rate": 8.824202218917991e-06,
"loss": 2.7088,
"step": 72000
},
{
"epoch": 8.29,
"learning_rate": 8.53825917877159e-06,
"loss": 2.7115,
"step": 72500
},
{
"epoch": 8.35,
"learning_rate": 8.252316138625186e-06,
"loss": 2.7049,
"step": 73000
},
{
"epoch": 8.41,
"learning_rate": 7.966373098478785e-06,
"loss": 2.7083,
"step": 73500
},
{
"epoch": 8.46,
"learning_rate": 7.68043005833238e-06,
"loss": 2.6946,
"step": 74000
},
{
"epoch": 8.52,
"learning_rate": 7.394487018185977e-06,
"loss": 2.691,
"step": 74500
},
{
"epoch": 8.58,
"learning_rate": 7.108543978039575e-06,
"loss": 2.7066,
"step": 75000
},
{
"epoch": 8.64,
"learning_rate": 6.8226009378931726e-06,
"loss": 2.6896,
"step": 75500
},
{
"epoch": 8.69,
"learning_rate": 6.5366578977467685e-06,
"loss": 2.6976,
"step": 76000
},
{
"epoch": 8.75,
"learning_rate": 6.250714857600366e-06,
"loss": 2.6989,
"step": 76500
},
{
"epoch": 8.81,
"learning_rate": 5.964771817453964e-06,
"loss": 2.6869,
"step": 77000
},
{
"epoch": 8.86,
"learning_rate": 5.6788287773075605e-06,
"loss": 2.6898,
"step": 77500
},
{
"epoch": 8.92,
"learning_rate": 5.392885737161157e-06,
"loss": 2.6851,
"step": 78000
},
{
"epoch": 8.98,
"learning_rate": 5.106942697014755e-06,
"loss": 2.6871,
"step": 78500
},
{
"epoch": 9.04,
"learning_rate": 4.820999656868352e-06,
"loss": 2.6719,
"step": 79000
},
{
"epoch": 9.09,
"learning_rate": 4.535056616721949e-06,
"loss": 2.6704,
"step": 79500
},
{
"epoch": 9.15,
"learning_rate": 4.249113576575546e-06,
"loss": 2.6776,
"step": 80000
}
],
"max_steps": 87430,
"num_train_epochs": 10,
"total_flos": 9.648556042711978e+17,
"trial_name": null,
"trial_params": null
}