|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 10686, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02807411566535654, |
|
"grad_norm": 8.254485130310059, |
|
"learning_rate": 4.9550814149354296e-05, |
|
"loss": 5.1202, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.05614823133071308, |
|
"grad_norm": 14.410130500793457, |
|
"learning_rate": 4.908291222159835e-05, |
|
"loss": 4.827, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.08422234699606962, |
|
"grad_norm": 12.091243743896484, |
|
"learning_rate": 4.861501029384241e-05, |
|
"loss": 4.8012, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.11229646266142616, |
|
"grad_norm": 9.615291595458984, |
|
"learning_rate": 4.814710836608647e-05, |
|
"loss": 4.7478, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.1403705783267827, |
|
"grad_norm": 7.660737991333008, |
|
"learning_rate": 4.767920643833052e-05, |
|
"loss": 4.7, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.16844469399213924, |
|
"grad_norm": 8.466590881347656, |
|
"learning_rate": 4.7211304510574585e-05, |
|
"loss": 4.7237, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.1965188096574958, |
|
"grad_norm": 9.133238792419434, |
|
"learning_rate": 4.674340258281865e-05, |
|
"loss": 4.6992, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.22459292532285233, |
|
"grad_norm": 11.180377006530762, |
|
"learning_rate": 4.62755006550627e-05, |
|
"loss": 4.6911, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.25266704098820886, |
|
"grad_norm": 9.834083557128906, |
|
"learning_rate": 4.5807598727306764e-05, |
|
"loss": 4.6565, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.2807411566535654, |
|
"grad_norm": 8.723132133483887, |
|
"learning_rate": 4.533969679955082e-05, |
|
"loss": 4.6333, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.308815272318922, |
|
"grad_norm": 12.12197494506836, |
|
"learning_rate": 4.4871794871794874e-05, |
|
"loss": 4.5906, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.3368893879842785, |
|
"grad_norm": 10.753365516662598, |
|
"learning_rate": 4.440389294403893e-05, |
|
"loss": 4.5926, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.36496350364963503, |
|
"grad_norm": 8.904462814331055, |
|
"learning_rate": 4.393599101628299e-05, |
|
"loss": 4.5735, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.3930376193149916, |
|
"grad_norm": 7.0933098793029785, |
|
"learning_rate": 4.346808908852705e-05, |
|
"loss": 4.6032, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.4211117349803481, |
|
"grad_norm": 6.25530481338501, |
|
"learning_rate": 4.30001871607711e-05, |
|
"loss": 4.6286, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.44918585064570465, |
|
"grad_norm": 9.821544647216797, |
|
"learning_rate": 4.2532285233015164e-05, |
|
"loss": 4.5785, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.4772599663110612, |
|
"grad_norm": 8.794018745422363, |
|
"learning_rate": 4.206438330525922e-05, |
|
"loss": 4.5527, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.5053340819764177, |
|
"grad_norm": 9.12629222869873, |
|
"learning_rate": 4.1596481377503274e-05, |
|
"loss": 4.5306, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.5334081976417743, |
|
"grad_norm": 8.344796180725098, |
|
"learning_rate": 4.1128579449747336e-05, |
|
"loss": 4.4843, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.5614823133071308, |
|
"grad_norm": 11.365053176879883, |
|
"learning_rate": 4.066067752199139e-05, |
|
"loss": 4.4841, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.5895564289724874, |
|
"grad_norm": 8.378218650817871, |
|
"learning_rate": 4.0192775594235447e-05, |
|
"loss": 4.5196, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.617630544637844, |
|
"grad_norm": 11.260799407958984, |
|
"learning_rate": 3.972487366647951e-05, |
|
"loss": 4.4967, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.6457046603032004, |
|
"grad_norm": 11.992345809936523, |
|
"learning_rate": 3.9256971738723564e-05, |
|
"loss": 4.509, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.673778775968557, |
|
"grad_norm": 15.889375686645508, |
|
"learning_rate": 3.878906981096762e-05, |
|
"loss": 4.5183, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.7018528916339135, |
|
"grad_norm": 6.526124954223633, |
|
"learning_rate": 3.832116788321168e-05, |
|
"loss": 4.4316, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.7299270072992701, |
|
"grad_norm": 8.866271018981934, |
|
"learning_rate": 3.7853265955455736e-05, |
|
"loss": 4.4426, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.7580011229646266, |
|
"grad_norm": 8.329658508300781, |
|
"learning_rate": 3.73853640276998e-05, |
|
"loss": 4.445, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.7860752386299832, |
|
"grad_norm": 8.942349433898926, |
|
"learning_rate": 3.691746209994385e-05, |
|
"loss": 4.4743, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.8141493542953397, |
|
"grad_norm": 7.263224124908447, |
|
"learning_rate": 3.6449560172187915e-05, |
|
"loss": 4.4459, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.8422234699606962, |
|
"grad_norm": 8.767597198486328, |
|
"learning_rate": 3.598165824443197e-05, |
|
"loss": 4.3944, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.8702975856260527, |
|
"grad_norm": 7.760046482086182, |
|
"learning_rate": 3.5513756316676025e-05, |
|
"loss": 4.4683, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.8983717012914093, |
|
"grad_norm": 13.036920547485352, |
|
"learning_rate": 3.504585438892009e-05, |
|
"loss": 4.3891, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.9264458169567659, |
|
"grad_norm": 10.132473945617676, |
|
"learning_rate": 3.457795246116414e-05, |
|
"loss": 4.3914, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.9545199326221224, |
|
"grad_norm": 10.985709190368652, |
|
"learning_rate": 3.41100505334082e-05, |
|
"loss": 4.3861, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.982594048287479, |
|
"grad_norm": 12.080757141113281, |
|
"learning_rate": 3.364214860565226e-05, |
|
"loss": 4.4272, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.0106681639528354, |
|
"grad_norm": 10.2216157913208, |
|
"learning_rate": 3.3174246677896315e-05, |
|
"loss": 4.1802, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.038742279618192, |
|
"grad_norm": 7.825625896453857, |
|
"learning_rate": 3.270634475014037e-05, |
|
"loss": 3.7209, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 1.0668163952835485, |
|
"grad_norm": 9.543134689331055, |
|
"learning_rate": 3.223844282238443e-05, |
|
"loss": 3.7451, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.094890510948905, |
|
"grad_norm": 6.093784809112549, |
|
"learning_rate": 3.177054089462849e-05, |
|
"loss": 3.7287, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.1229646266142617, |
|
"grad_norm": 9.133221626281738, |
|
"learning_rate": 3.130263896687254e-05, |
|
"loss": 3.7896, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.1510387422796182, |
|
"grad_norm": 8.328789710998535, |
|
"learning_rate": 3.0834737039116604e-05, |
|
"loss": 3.7729, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 1.1791128579449748, |
|
"grad_norm": 12.665905952453613, |
|
"learning_rate": 3.036683511136066e-05, |
|
"loss": 3.768, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.2071869736103313, |
|
"grad_norm": 9.492897987365723, |
|
"learning_rate": 2.9898933183604715e-05, |
|
"loss": 3.7355, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.235261089275688, |
|
"grad_norm": 10.17910385131836, |
|
"learning_rate": 2.9431031255848773e-05, |
|
"loss": 3.8411, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.2633352049410442, |
|
"grad_norm": 7.939426898956299, |
|
"learning_rate": 2.896312932809283e-05, |
|
"loss": 3.7935, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.2914093206064008, |
|
"grad_norm": 8.300692558288574, |
|
"learning_rate": 2.8495227400336887e-05, |
|
"loss": 3.7495, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.3194834362717573, |
|
"grad_norm": 11.419922828674316, |
|
"learning_rate": 2.8027325472580952e-05, |
|
"loss": 3.8478, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.347557551937114, |
|
"grad_norm": 12.745059967041016, |
|
"learning_rate": 2.7559423544825007e-05, |
|
"loss": 3.8148, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.3756316676024705, |
|
"grad_norm": 12.494312286376953, |
|
"learning_rate": 2.7091521617069066e-05, |
|
"loss": 3.7696, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 1.403705783267827, |
|
"grad_norm": 7.439785480499268, |
|
"learning_rate": 2.6623619689313124e-05, |
|
"loss": 3.7559, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.4317798989331836, |
|
"grad_norm": 12.009989738464355, |
|
"learning_rate": 2.615571776155718e-05, |
|
"loss": 3.7759, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 1.4598540145985401, |
|
"grad_norm": 10.613336563110352, |
|
"learning_rate": 2.5687815833801238e-05, |
|
"loss": 3.7577, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 1.4879281302638967, |
|
"grad_norm": 10.656264305114746, |
|
"learning_rate": 2.5219913906045293e-05, |
|
"loss": 3.7515, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 1.5160022459292533, |
|
"grad_norm": 9.189164161682129, |
|
"learning_rate": 2.4752011978289352e-05, |
|
"loss": 3.8009, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 1.5440763615946098, |
|
"grad_norm": 9.947083473205566, |
|
"learning_rate": 2.428411005053341e-05, |
|
"loss": 3.8175, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 1.5721504772599664, |
|
"grad_norm": 14.514716148376465, |
|
"learning_rate": 2.3816208122777466e-05, |
|
"loss": 3.7141, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 1.600224592925323, |
|
"grad_norm": 8.071717262268066, |
|
"learning_rate": 2.3348306195021524e-05, |
|
"loss": 3.7252, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 1.6282987085906795, |
|
"grad_norm": 11.72280216217041, |
|
"learning_rate": 2.2880404267265583e-05, |
|
"loss": 3.7949, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 1.656372824256036, |
|
"grad_norm": 9.227740287780762, |
|
"learning_rate": 2.2412502339509638e-05, |
|
"loss": 3.733, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 1.6844469399213926, |
|
"grad_norm": 13.090754508972168, |
|
"learning_rate": 2.1944600411753697e-05, |
|
"loss": 3.8362, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 1.7125210555867492, |
|
"grad_norm": 12.168136596679688, |
|
"learning_rate": 2.1476698483997755e-05, |
|
"loss": 3.7261, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 1.7405951712521057, |
|
"grad_norm": 8.332971572875977, |
|
"learning_rate": 2.1008796556241814e-05, |
|
"loss": 3.8215, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 1.768669286917462, |
|
"grad_norm": 8.5094633102417, |
|
"learning_rate": 2.0540894628485872e-05, |
|
"loss": 3.804, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 1.7967434025828186, |
|
"grad_norm": 11.575583457946777, |
|
"learning_rate": 2.0072992700729927e-05, |
|
"loss": 3.7369, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 1.8248175182481752, |
|
"grad_norm": 12.597991943359375, |
|
"learning_rate": 1.9605090772973986e-05, |
|
"loss": 3.7377, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 1.8528916339135317, |
|
"grad_norm": 9.923151016235352, |
|
"learning_rate": 1.9137188845218044e-05, |
|
"loss": 3.7808, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 1.8809657495788883, |
|
"grad_norm": 10.516582489013672, |
|
"learning_rate": 1.86692869174621e-05, |
|
"loss": 3.7584, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 1.9090398652442448, |
|
"grad_norm": 8.432146072387695, |
|
"learning_rate": 1.8201384989706158e-05, |
|
"loss": 3.7837, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 1.9371139809096012, |
|
"grad_norm": 12.533796310424805, |
|
"learning_rate": 1.7733483061950217e-05, |
|
"loss": 3.6908, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 1.9651880965749577, |
|
"grad_norm": 11.924694061279297, |
|
"learning_rate": 1.7265581134194272e-05, |
|
"loss": 3.7846, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 1.9932622122403143, |
|
"grad_norm": 10.164834976196289, |
|
"learning_rate": 1.6797679206438334e-05, |
|
"loss": 3.766, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 2.021336327905671, |
|
"grad_norm": 9.073919296264648, |
|
"learning_rate": 1.633445629795995e-05, |
|
"loss": 3.3417, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 2.0494104435710274, |
|
"grad_norm": 10.306299209594727, |
|
"learning_rate": 1.5866554370204005e-05, |
|
"loss": 3.2644, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 2.077484559236384, |
|
"grad_norm": 11.006250381469727, |
|
"learning_rate": 1.5398652442448064e-05, |
|
"loss": 3.243, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 2.1055586749017405, |
|
"grad_norm": 9.138731956481934, |
|
"learning_rate": 1.4930750514692121e-05, |
|
"loss": 3.2938, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 2.133632790567097, |
|
"grad_norm": 9.330265998840332, |
|
"learning_rate": 1.4462848586936178e-05, |
|
"loss": 3.247, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 2.1617069062324537, |
|
"grad_norm": 9.435338973999023, |
|
"learning_rate": 1.3994946659180236e-05, |
|
"loss": 3.2847, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 2.18978102189781, |
|
"grad_norm": 9.446837425231934, |
|
"learning_rate": 1.3527044731424293e-05, |
|
"loss": 3.2522, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 2.2178551375631668, |
|
"grad_norm": 11.070699691772461, |
|
"learning_rate": 1.305914280366835e-05, |
|
"loss": 3.2073, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 2.2459292532285233, |
|
"grad_norm": 9.919466018676758, |
|
"learning_rate": 1.259124087591241e-05, |
|
"loss": 3.2816, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 2.27400336889388, |
|
"grad_norm": 11.001142501831055, |
|
"learning_rate": 1.2123338948156467e-05, |
|
"loss": 3.2669, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 2.3020774845592364, |
|
"grad_norm": 10.917780876159668, |
|
"learning_rate": 1.1655437020400524e-05, |
|
"loss": 3.2488, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 2.330151600224593, |
|
"grad_norm": 10.114619255065918, |
|
"learning_rate": 1.1187535092644583e-05, |
|
"loss": 3.2317, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 2.3582257158899496, |
|
"grad_norm": 8.730778694152832, |
|
"learning_rate": 1.071963316488864e-05, |
|
"loss": 3.274, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 2.386299831555306, |
|
"grad_norm": 7.530703544616699, |
|
"learning_rate": 1.0251731237132698e-05, |
|
"loss": 3.2506, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 2.4143739472206627, |
|
"grad_norm": 13.161232948303223, |
|
"learning_rate": 9.783829309376755e-06, |
|
"loss": 3.3056, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 2.4424480628860192, |
|
"grad_norm": 7.7642598152160645, |
|
"learning_rate": 9.315927381620813e-06, |
|
"loss": 3.2745, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 2.470522178551376, |
|
"grad_norm": 8.774062156677246, |
|
"learning_rate": 8.84802545386487e-06, |
|
"loss": 3.263, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 2.4985962942167323, |
|
"grad_norm": 13.077786445617676, |
|
"learning_rate": 8.380123526108929e-06, |
|
"loss": 3.2418, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 2.5266704098820885, |
|
"grad_norm": 20.591571807861328, |
|
"learning_rate": 7.912221598352986e-06, |
|
"loss": 3.3027, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 2.554744525547445, |
|
"grad_norm": 10.915307998657227, |
|
"learning_rate": 7.444319670597043e-06, |
|
"loss": 3.2292, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 2.5828186412128016, |
|
"grad_norm": 10.937311172485352, |
|
"learning_rate": 6.976417742841101e-06, |
|
"loss": 3.3281, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 2.610892756878158, |
|
"grad_norm": 8.748698234558105, |
|
"learning_rate": 6.508515815085159e-06, |
|
"loss": 3.2823, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 2.6389668725435147, |
|
"grad_norm": 8.032021522521973, |
|
"learning_rate": 6.040613887329216e-06, |
|
"loss": 3.2265, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 2.6670409882088713, |
|
"grad_norm": 14.668137550354004, |
|
"learning_rate": 5.5727119595732735e-06, |
|
"loss": 3.2673, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 2.695115103874228, |
|
"grad_norm": 9.834213256835938, |
|
"learning_rate": 5.104810031817331e-06, |
|
"loss": 3.2213, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 2.7231892195395844, |
|
"grad_norm": 14.294054985046387, |
|
"learning_rate": 4.636908104061389e-06, |
|
"loss": 3.2487, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 2.751263335204941, |
|
"grad_norm": 8.172629356384277, |
|
"learning_rate": 4.169006176305447e-06, |
|
"loss": 3.2598, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 2.7793374508702975, |
|
"grad_norm": 18.083383560180664, |
|
"learning_rate": 3.7011042485495043e-06, |
|
"loss": 3.2655, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 2.807411566535654, |
|
"grad_norm": 13.841841697692871, |
|
"learning_rate": 3.2332023207935616e-06, |
|
"loss": 3.2676, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 2.8354856822010106, |
|
"grad_norm": 16.98509407043457, |
|
"learning_rate": 2.7653003930376193e-06, |
|
"loss": 3.2499, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 2.863559797866367, |
|
"grad_norm": 10.365632057189941, |
|
"learning_rate": 2.297398465281677e-06, |
|
"loss": 3.2457, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 2.8916339135317237, |
|
"grad_norm": 10.886728286743164, |
|
"learning_rate": 1.8294965375257348e-06, |
|
"loss": 3.2445, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 2.9197080291970803, |
|
"grad_norm": 10.890334129333496, |
|
"learning_rate": 1.3615946097697923e-06, |
|
"loss": 3.211, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 2.947782144862437, |
|
"grad_norm": 11.197315216064453, |
|
"learning_rate": 8.936926820138499e-07, |
|
"loss": 3.2483, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 2.9758562605277934, |
|
"grad_norm": 9.767260551452637, |
|
"learning_rate": 4.2579075425790754e-07, |
|
"loss": 3.2594, |
|
"step": 10600 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 10686, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 5583353462784000.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|