gemmy-9b-ckpt-252 / trainer_state.json
lucyknada's picture
Upload folder using huggingface_hub
fa3018e verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.7558488302339532,
"eval_steps": 500,
"global_step": 252,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.002999400119976005,
"grad_norm": 12.213529318427376,
"learning_rate": 1.0000000000000002e-06,
"loss": 1.6991,
"step": 1
},
{
"epoch": 0.00599880023995201,
"grad_norm": 11.149836011452198,
"learning_rate": 2.0000000000000003e-06,
"loss": 1.5997,
"step": 2
},
{
"epoch": 0.008998200359928014,
"grad_norm": 5.296165512046405,
"learning_rate": 3e-06,
"loss": 1.3037,
"step": 3
},
{
"epoch": 0.01199760047990402,
"grad_norm": 3.7277911972297235,
"learning_rate": 4.000000000000001e-06,
"loss": 1.3472,
"step": 4
},
{
"epoch": 0.014997000599880024,
"grad_norm": 6.970291148182751,
"learning_rate": 5e-06,
"loss": 1.5234,
"step": 5
},
{
"epoch": 0.017996400719856028,
"grad_norm": 5.825870871741097,
"learning_rate": 6e-06,
"loss": 1.2331,
"step": 6
},
{
"epoch": 0.020995800839832032,
"grad_norm": 4.789226733983147,
"learning_rate": 7e-06,
"loss": 1.2847,
"step": 7
},
{
"epoch": 0.02399520095980804,
"grad_norm": 5.735809923380043,
"learning_rate": 8.000000000000001e-06,
"loss": 1.3148,
"step": 8
},
{
"epoch": 0.026994601079784044,
"grad_norm": 4.146594143559111,
"learning_rate": 9e-06,
"loss": 1.3241,
"step": 9
},
{
"epoch": 0.029994001199760048,
"grad_norm": 4.3548572301609365,
"learning_rate": 1e-05,
"loss": 1.2048,
"step": 10
},
{
"epoch": 0.032993401319736056,
"grad_norm": 3.186860649490824,
"learning_rate": 9.999942663491213e-06,
"loss": 1.271,
"step": 11
},
{
"epoch": 0.035992801439712056,
"grad_norm": 5.683449148461711,
"learning_rate": 9.999770655279843e-06,
"loss": 1.4619,
"step": 12
},
{
"epoch": 0.038992201559688064,
"grad_norm": 2.918639590322084,
"learning_rate": 9.99948397931083e-06,
"loss": 1.2698,
"step": 13
},
{
"epoch": 0.041991601679664065,
"grad_norm": 3.1049642688975228,
"learning_rate": 9.999082642158972e-06,
"loss": 1.4018,
"step": 14
},
{
"epoch": 0.04499100179964007,
"grad_norm": 2.396816425806027,
"learning_rate": 9.99856665302878e-06,
"loss": 1.1897,
"step": 15
},
{
"epoch": 0.04799040191961608,
"grad_norm": 2.6334601975355283,
"learning_rate": 9.997936023754258e-06,
"loss": 1.3976,
"step": 16
},
{
"epoch": 0.05098980203959208,
"grad_norm": 2.619689679258694,
"learning_rate": 9.997190768798639e-06,
"loss": 1.286,
"step": 17
},
{
"epoch": 0.05398920215956809,
"grad_norm": 2.7084119507983377,
"learning_rate": 9.99633090525405e-06,
"loss": 1.3328,
"step": 18
},
{
"epoch": 0.05698860227954409,
"grad_norm": 2.3481903809919404,
"learning_rate": 9.995356452841122e-06,
"loss": 1.1662,
"step": 19
},
{
"epoch": 0.059988002399520096,
"grad_norm": 4.651416427067284,
"learning_rate": 9.994267433908533e-06,
"loss": 1.5033,
"step": 20
},
{
"epoch": 0.0629874025194961,
"grad_norm": 2.625598226472764,
"learning_rate": 9.9930638734325e-06,
"loss": 1.2022,
"step": 21
},
{
"epoch": 0.06598680263947211,
"grad_norm": 2.4104347021265973,
"learning_rate": 9.991745799016206e-06,
"loss": 1.1979,
"step": 22
},
{
"epoch": 0.06898620275944811,
"grad_norm": 3.1305838351835478,
"learning_rate": 9.990313240889167e-06,
"loss": 1.3105,
"step": 23
},
{
"epoch": 0.07198560287942411,
"grad_norm": 3.18309366674778,
"learning_rate": 9.988766231906532e-06,
"loss": 1.3079,
"step": 24
},
{
"epoch": 0.07498500299940011,
"grad_norm": 3.5781057246328563,
"learning_rate": 9.987104807548341e-06,
"loss": 1.3255,
"step": 25
},
{
"epoch": 0.07798440311937613,
"grad_norm": 2.259604288692876,
"learning_rate": 9.985329005918702e-06,
"loss": 1.242,
"step": 26
},
{
"epoch": 0.08098380323935213,
"grad_norm": 5.16511455231151,
"learning_rate": 9.983438867744923e-06,
"loss": 1.5295,
"step": 27
},
{
"epoch": 0.08398320335932813,
"grad_norm": 2.5507458276752955,
"learning_rate": 9.981434436376572e-06,
"loss": 1.1835,
"step": 28
},
{
"epoch": 0.08698260347930414,
"grad_norm": 2.374301337319529,
"learning_rate": 9.97931575778449e-06,
"loss": 1.2602,
"step": 29
},
{
"epoch": 0.08998200359928014,
"grad_norm": 2.869182508074545,
"learning_rate": 9.977082880559725e-06,
"loss": 1.2714,
"step": 30
},
{
"epoch": 0.09298140371925614,
"grad_norm": 3.352349678924823,
"learning_rate": 9.974735855912436e-06,
"loss": 1.2895,
"step": 31
},
{
"epoch": 0.09598080383923216,
"grad_norm": 2.402498887438619,
"learning_rate": 9.972274737670702e-06,
"loss": 1.2169,
"step": 32
},
{
"epoch": 0.09898020395920816,
"grad_norm": 2.3026043458108356,
"learning_rate": 9.969699582279294e-06,
"loss": 1.2486,
"step": 33
},
{
"epoch": 0.10197960407918416,
"grad_norm": 2.7234482697971796,
"learning_rate": 9.967010448798376e-06,
"loss": 1.3611,
"step": 34
},
{
"epoch": 0.10497900419916016,
"grad_norm": 2.3328312481588096,
"learning_rate": 9.964207398902163e-06,
"loss": 1.2112,
"step": 35
},
{
"epoch": 0.10797840431913618,
"grad_norm": 2.3738667556158304,
"learning_rate": 9.961290496877492e-06,
"loss": 1.1489,
"step": 36
},
{
"epoch": 0.11097780443911218,
"grad_norm": 2.381064080119671,
"learning_rate": 9.958259809622353e-06,
"loss": 1.3426,
"step": 37
},
{
"epoch": 0.11397720455908818,
"grad_norm": 2.789399877084244,
"learning_rate": 9.955115406644357e-06,
"loss": 1.2355,
"step": 38
},
{
"epoch": 0.11697660467906419,
"grad_norm": 2.2360091447066885,
"learning_rate": 9.951857360059141e-06,
"loss": 1.298,
"step": 39
},
{
"epoch": 0.11997600479904019,
"grad_norm": 2.3420608415672644,
"learning_rate": 9.948485744588709e-06,
"loss": 1.2597,
"step": 40
},
{
"epoch": 0.12297540491901619,
"grad_norm": 2.0514216630297786,
"learning_rate": 9.945000637559728e-06,
"loss": 1.2464,
"step": 41
},
{
"epoch": 0.1259748050389922,
"grad_norm": 2.1083906964286765,
"learning_rate": 9.941402118901743e-06,
"loss": 1.2223,
"step": 42
},
{
"epoch": 0.1289742051589682,
"grad_norm": 2.584585224378793,
"learning_rate": 9.937690271145356e-06,
"loss": 1.2575,
"step": 43
},
{
"epoch": 0.13197360527894422,
"grad_norm": 2.3779090684986652,
"learning_rate": 9.93386517942032e-06,
"loss": 1.1451,
"step": 44
},
{
"epoch": 0.1349730053989202,
"grad_norm": 2.108339055225494,
"learning_rate": 9.9299269314536e-06,
"loss": 1.2248,
"step": 45
},
{
"epoch": 0.13797240551889622,
"grad_norm": 2.121719885374642,
"learning_rate": 9.92587561756735e-06,
"loss": 1.207,
"step": 46
},
{
"epoch": 0.14097180563887224,
"grad_norm": 2.205771486341789,
"learning_rate": 9.921711330676848e-06,
"loss": 1.2321,
"step": 47
},
{
"epoch": 0.14397120575884823,
"grad_norm": 2.288650968193418,
"learning_rate": 9.917434166288364e-06,
"loss": 1.244,
"step": 48
},
{
"epoch": 0.14697060587882424,
"grad_norm": 2.3703349873199917,
"learning_rate": 9.913044222496966e-06,
"loss": 1.2758,
"step": 49
},
{
"epoch": 0.14997000599880023,
"grad_norm": 2.342816783402187,
"learning_rate": 9.908541599984276e-06,
"loss": 1.2799,
"step": 50
},
{
"epoch": 0.15296940611877624,
"grad_norm": 2.290011867632185,
"learning_rate": 9.903926402016153e-06,
"loss": 1.137,
"step": 51
},
{
"epoch": 0.15596880623875226,
"grad_norm": 2.2455094517388825,
"learning_rate": 9.899198734440335e-06,
"loss": 1.2133,
"step": 52
},
{
"epoch": 0.15896820635872824,
"grad_norm": 2.263487895355519,
"learning_rate": 9.894358705684002e-06,
"loss": 1.1898,
"step": 53
},
{
"epoch": 0.16196760647870426,
"grad_norm": 2.094119128560065,
"learning_rate": 9.889406426751296e-06,
"loss": 1.1847,
"step": 54
},
{
"epoch": 0.16496700659868027,
"grad_norm": 2.45640150031141,
"learning_rate": 9.88434201122077e-06,
"loss": 1.2662,
"step": 55
},
{
"epoch": 0.16796640671865626,
"grad_norm": 3.7462847341409606,
"learning_rate": 9.879165575242788e-06,
"loss": 1.1912,
"step": 56
},
{
"epoch": 0.17096580683863227,
"grad_norm": 2.382097038596032,
"learning_rate": 9.873877237536854e-06,
"loss": 1.2652,
"step": 57
},
{
"epoch": 0.1739652069586083,
"grad_norm": 2.320080713494513,
"learning_rate": 9.868477119388897e-06,
"loss": 1.2682,
"step": 58
},
{
"epoch": 0.17696460707858427,
"grad_norm": 2.1974546622057938,
"learning_rate": 9.862965344648485e-06,
"loss": 1.1956,
"step": 59
},
{
"epoch": 0.1799640071985603,
"grad_norm": 2.40241845961337,
"learning_rate": 9.85734203972599e-06,
"loss": 1.2948,
"step": 60
},
{
"epoch": 0.1829634073185363,
"grad_norm": 2.1855343450625053,
"learning_rate": 9.851607333589677e-06,
"loss": 1.1967,
"step": 61
},
{
"epoch": 0.1859628074385123,
"grad_norm": 2.772912562685191,
"learning_rate": 9.84576135776276e-06,
"loss": 1.1864,
"step": 62
},
{
"epoch": 0.1889622075584883,
"grad_norm": 2.420057020746961,
"learning_rate": 9.839804246320374e-06,
"loss": 1.2534,
"step": 63
},
{
"epoch": 0.19196160767846432,
"grad_norm": 2.265555099135192,
"learning_rate": 9.833736135886513e-06,
"loss": 1.2563,
"step": 64
},
{
"epoch": 0.1949610077984403,
"grad_norm": 2.292076182842636,
"learning_rate": 9.827557165630879e-06,
"loss": 1.193,
"step": 65
},
{
"epoch": 0.19796040791841632,
"grad_norm": 2.032874490713336,
"learning_rate": 9.821267477265705e-06,
"loss": 1.231,
"step": 66
},
{
"epoch": 0.20095980803839233,
"grad_norm": 2.1341683347971543,
"learning_rate": 9.814867215042503e-06,
"loss": 1.2317,
"step": 67
},
{
"epoch": 0.20395920815836832,
"grad_norm": 2.2463443930345077,
"learning_rate": 9.808356525748748e-06,
"loss": 1.2911,
"step": 68
},
{
"epoch": 0.20695860827834434,
"grad_norm": 2.0325804061530848,
"learning_rate": 9.801735558704516e-06,
"loss": 1.254,
"step": 69
},
{
"epoch": 0.20995800839832032,
"grad_norm": 2.0568500164008086,
"learning_rate": 9.795004465759067e-06,
"loss": 1.206,
"step": 70
},
{
"epoch": 0.21295740851829634,
"grad_norm": 2.082577302356669,
"learning_rate": 9.78816340128734e-06,
"loss": 1.2027,
"step": 71
},
{
"epoch": 0.21595680863827235,
"grad_norm": 2.048998263808703,
"learning_rate": 9.781212522186442e-06,
"loss": 1.2288,
"step": 72
},
{
"epoch": 0.21895620875824834,
"grad_norm": 2.49581568276726,
"learning_rate": 9.774151987872029e-06,
"loss": 1.2493,
"step": 73
},
{
"epoch": 0.22195560887822435,
"grad_norm": 1.8323511888260988,
"learning_rate": 9.766981960274653e-06,
"loss": 1.2133,
"step": 74
},
{
"epoch": 0.22495500899820037,
"grad_norm": 1.9244897973012691,
"learning_rate": 9.759702603836059e-06,
"loss": 1.1556,
"step": 75
},
{
"epoch": 0.22795440911817635,
"grad_norm": 2.2391757657702476,
"learning_rate": 9.752314085505396e-06,
"loss": 1.139,
"step": 76
},
{
"epoch": 0.23095380923815237,
"grad_norm": 2.1062513435343955,
"learning_rate": 9.744816574735405e-06,
"loss": 1.1568,
"step": 77
},
{
"epoch": 0.23395320935812838,
"grad_norm": 2.244216493001031,
"learning_rate": 9.737210243478522e-06,
"loss": 1.1672,
"step": 78
},
{
"epoch": 0.23695260947810437,
"grad_norm": 2.2301876958310585,
"learning_rate": 9.72949526618294e-06,
"loss": 1.1805,
"step": 79
},
{
"epoch": 0.23995200959808038,
"grad_norm": 2.087760281351449,
"learning_rate": 9.721671819788603e-06,
"loss": 1.2595,
"step": 80
},
{
"epoch": 0.2429514097180564,
"grad_norm": 2.040923837775886,
"learning_rate": 9.713740083723152e-06,
"loss": 1.1853,
"step": 81
},
{
"epoch": 0.24595080983803239,
"grad_norm": 1.9835044903865013,
"learning_rate": 9.705700239897809e-06,
"loss": 1.1165,
"step": 82
},
{
"epoch": 0.2489502099580084,
"grad_norm": 2.4048211037733433,
"learning_rate": 9.697552472703205e-06,
"loss": 1.2442,
"step": 83
},
{
"epoch": 0.2519496100779844,
"grad_norm": 1.9274434259225983,
"learning_rate": 9.689296969005151e-06,
"loss": 1.2233,
"step": 84
},
{
"epoch": 0.2549490101979604,
"grad_norm": 2.0930221605237813,
"learning_rate": 9.680933918140348e-06,
"loss": 1.2042,
"step": 85
},
{
"epoch": 0.2579484103179364,
"grad_norm": 2.050947291358715,
"learning_rate": 9.672463511912056e-06,
"loss": 1.1526,
"step": 86
},
{
"epoch": 0.26094781043791243,
"grad_norm": 2.1905662044973218,
"learning_rate": 9.66388594458568e-06,
"loss": 1.2428,
"step": 87
},
{
"epoch": 0.26394721055788845,
"grad_norm": 2.194939852549527,
"learning_rate": 9.655201412884328e-06,
"loss": 1.1842,
"step": 88
},
{
"epoch": 0.2669466106778644,
"grad_norm": 1.9191574847419615,
"learning_rate": 9.64641011598429e-06,
"loss": 1.2005,
"step": 89
},
{
"epoch": 0.2699460107978404,
"grad_norm": 2.0406272846772597,
"learning_rate": 9.637512255510475e-06,
"loss": 1.1275,
"step": 90
},
{
"epoch": 0.27294541091781643,
"grad_norm": 1.9546291627174466,
"learning_rate": 9.628508035531785e-06,
"loss": 1.2274,
"step": 91
},
{
"epoch": 0.27594481103779245,
"grad_norm": 1.9691367121837948,
"learning_rate": 9.619397662556434e-06,
"loss": 1.207,
"step": 92
},
{
"epoch": 0.27894421115776846,
"grad_norm": 2.3613406410056323,
"learning_rate": 9.610181345527217e-06,
"loss": 1.2273,
"step": 93
},
{
"epoch": 0.2819436112777445,
"grad_norm": 1.869890124546451,
"learning_rate": 9.600859295816708e-06,
"loss": 1.2272,
"step": 94
},
{
"epoch": 0.28494301139772044,
"grad_norm": 2.1493924596390834,
"learning_rate": 9.591431727222425e-06,
"loss": 1.1787,
"step": 95
},
{
"epoch": 0.28794241151769645,
"grad_norm": 2.055643425325773,
"learning_rate": 9.581898855961911e-06,
"loss": 1.1993,
"step": 96
},
{
"epoch": 0.29094181163767247,
"grad_norm": 1.84642386683885,
"learning_rate": 9.572260900667794e-06,
"loss": 1.1839,
"step": 97
},
{
"epoch": 0.2939412117576485,
"grad_norm": 1.890951805519698,
"learning_rate": 9.562518082382751e-06,
"loss": 1.207,
"step": 98
},
{
"epoch": 0.2969406118776245,
"grad_norm": 2.37835708136971,
"learning_rate": 9.55267062455446e-06,
"loss": 1.1894,
"step": 99
},
{
"epoch": 0.29994001199760045,
"grad_norm": 1.9298763701114428,
"learning_rate": 9.542718753030463e-06,
"loss": 1.17,
"step": 100
},
{
"epoch": 0.30293941211757647,
"grad_norm": 1.9738559838328023,
"learning_rate": 9.532662696052986e-06,
"loss": 1.1911,
"step": 101
},
{
"epoch": 0.3059388122375525,
"grad_norm": 2.097481526316711,
"learning_rate": 9.522502684253709e-06,
"loss": 1.2047,
"step": 102
},
{
"epoch": 0.3089382123575285,
"grad_norm": 2.2773066391553405,
"learning_rate": 9.512238950648474e-06,
"loss": 1.1683,
"step": 103
},
{
"epoch": 0.3119376124775045,
"grad_norm": 2.0545409984803435,
"learning_rate": 9.501871730631944e-06,
"loss": 1.1777,
"step": 104
},
{
"epoch": 0.3149370125974805,
"grad_norm": 2.0765643204952435,
"learning_rate": 9.491401261972194e-06,
"loss": 1.2306,
"step": 105
},
{
"epoch": 0.3179364127174565,
"grad_norm": 1.9891174989072355,
"learning_rate": 9.480827784805278e-06,
"loss": 1.2351,
"step": 106
},
{
"epoch": 0.3209358128374325,
"grad_norm": 1.889574475483626,
"learning_rate": 9.4701515416297e-06,
"loss": 1.1742,
"step": 107
},
{
"epoch": 0.3239352129574085,
"grad_norm": 2.0131114525216955,
"learning_rate": 9.459372777300863e-06,
"loss": 1.127,
"step": 108
},
{
"epoch": 0.32693461307738453,
"grad_norm": 2.0094971545889506,
"learning_rate": 9.448491739025455e-06,
"loss": 1.1203,
"step": 109
},
{
"epoch": 0.32993401319736054,
"grad_norm": 2.007024349964355,
"learning_rate": 9.437508676355774e-06,
"loss": 1.2005,
"step": 110
},
{
"epoch": 0.33293341331733656,
"grad_norm": 1.834949376512209,
"learning_rate": 9.426423841184007e-06,
"loss": 1.1451,
"step": 111
},
{
"epoch": 0.3359328134373125,
"grad_norm": 2.0659739662077863,
"learning_rate": 9.415237487736452e-06,
"loss": 1.1978,
"step": 112
},
{
"epoch": 0.33893221355728853,
"grad_norm": 2.0021385422334714,
"learning_rate": 9.403949872567696e-06,
"loss": 1.1969,
"step": 113
},
{
"epoch": 0.34193161367726455,
"grad_norm": 2.131713368907865,
"learning_rate": 9.392561254554712e-06,
"loss": 1.0958,
"step": 114
},
{
"epoch": 0.34493101379724056,
"grad_norm": 2.009265355301516,
"learning_rate": 9.381071894890942e-06,
"loss": 1.2154,
"step": 115
},
{
"epoch": 0.3479304139172166,
"grad_norm": 1.9134985847126271,
"learning_rate": 9.369482057080293e-06,
"loss": 1.2242,
"step": 116
},
{
"epoch": 0.3509298140371926,
"grad_norm": 1.9567206999616176,
"learning_rate": 9.3577920069311e-06,
"loss": 1.172,
"step": 117
},
{
"epoch": 0.35392921415716855,
"grad_norm": 1.9433526958741427,
"learning_rate": 9.346002012550027e-06,
"loss": 1.1785,
"step": 118
},
{
"epoch": 0.35692861427714456,
"grad_norm": 3.7758127026886776,
"learning_rate": 9.334112344335924e-06,
"loss": 1.2509,
"step": 119
},
{
"epoch": 0.3599280143971206,
"grad_norm": 2.2288097190307754,
"learning_rate": 9.322123274973613e-06,
"loss": 1.2034,
"step": 120
},
{
"epoch": 0.3629274145170966,
"grad_norm": 2.210018938006893,
"learning_rate": 9.310035079427651e-06,
"loss": 1.2089,
"step": 121
},
{
"epoch": 0.3659268146370726,
"grad_norm": 1.8381206653711355,
"learning_rate": 9.297848034936007e-06,
"loss": 1.186,
"step": 122
},
{
"epoch": 0.36892621475704857,
"grad_norm": 2.1078020542808913,
"learning_rate": 9.285562421003716e-06,
"loss": 1.2042,
"step": 123
},
{
"epoch": 0.3719256148770246,
"grad_norm": 2.2181514311801496,
"learning_rate": 9.273178519396459e-06,
"loss": 1.1561,
"step": 124
},
{
"epoch": 0.3749250149970006,
"grad_norm": 2.127427295141141,
"learning_rate": 9.260696614134115e-06,
"loss": 1.1935,
"step": 125
},
{
"epoch": 0.3779244151169766,
"grad_norm": 2.179542019285163,
"learning_rate": 9.24811699148423e-06,
"loss": 1.2001,
"step": 126
},
{
"epoch": 0.3809238152369526,
"grad_norm": 2.95737675688332,
"learning_rate": 9.235439939955458e-06,
"loss": 1.1333,
"step": 127
},
{
"epoch": 0.38392321535692864,
"grad_norm": 2.408346714267433,
"learning_rate": 9.222665750290953e-06,
"loss": 1.2035,
"step": 128
},
{
"epoch": 0.3869226154769046,
"grad_norm": 1.9803537052769442,
"learning_rate": 9.209794715461691e-06,
"loss": 1.2133,
"step": 129
},
{
"epoch": 0.3899220155968806,
"grad_norm": 2.0753391828539844,
"learning_rate": 9.196827130659752e-06,
"loss": 1.2389,
"step": 130
},
{
"epoch": 0.3929214157168566,
"grad_norm": 2.057624573882372,
"learning_rate": 9.18376329329155e-06,
"loss": 1.1612,
"step": 131
},
{
"epoch": 0.39592081583683264,
"grad_norm": 2.095421490379302,
"learning_rate": 9.170603502971017e-06,
"loss": 1.2319,
"step": 132
},
{
"epoch": 0.39892021595680865,
"grad_norm": 1.9629007266889844,
"learning_rate": 9.157348061512728e-06,
"loss": 1.2061,
"step": 133
},
{
"epoch": 0.40191961607678467,
"grad_norm": 1.9702164374790287,
"learning_rate": 9.143997272924974e-06,
"loss": 1.197,
"step": 134
},
{
"epoch": 0.40491901619676063,
"grad_norm": 2.094524351597277,
"learning_rate": 9.1305514434028e-06,
"loss": 1.1508,
"step": 135
},
{
"epoch": 0.40791841631673664,
"grad_norm": 2.078336809149187,
"learning_rate": 9.117010881320973e-06,
"loss": 1.2006,
"step": 136
},
{
"epoch": 0.41091781643671266,
"grad_norm": 2.1047922013296385,
"learning_rate": 9.103375897226919e-06,
"loss": 1.2225,
"step": 137
},
{
"epoch": 0.41391721655668867,
"grad_norm": 1.9800892010976672,
"learning_rate": 9.089646803833589e-06,
"loss": 1.2117,
"step": 138
},
{
"epoch": 0.4169166166766647,
"grad_norm": 1.948925596250933,
"learning_rate": 9.075823916012298e-06,
"loss": 1.1917,
"step": 139
},
{
"epoch": 0.41991601679664065,
"grad_norm": 1.8720535653109143,
"learning_rate": 9.061907550785498e-06,
"loss": 1.2048,
"step": 140
},
{
"epoch": 0.42291541691661666,
"grad_norm": 2.1914812005829316,
"learning_rate": 9.047898027319508e-06,
"loss": 1.2393,
"step": 141
},
{
"epoch": 0.4259148170365927,
"grad_norm": 3.541065061637063,
"learning_rate": 9.033795666917191e-06,
"loss": 1.1728,
"step": 142
},
{
"epoch": 0.4289142171565687,
"grad_norm": 1.958943800312053,
"learning_rate": 9.019600793010596e-06,
"loss": 1.1204,
"step": 143
},
{
"epoch": 0.4319136172765447,
"grad_norm": 1.8982949525316917,
"learning_rate": 9.005313731153525e-06,
"loss": 1.136,
"step": 144
},
{
"epoch": 0.4349130173965207,
"grad_norm": 2.1962092793560775,
"learning_rate": 8.990934809014079e-06,
"loss": 1.2025,
"step": 145
},
{
"epoch": 0.4379124175164967,
"grad_norm": 2.148888648427292,
"learning_rate": 8.976464356367133e-06,
"loss": 1.1971,
"step": 146
},
{
"epoch": 0.4409118176364727,
"grad_norm": 2.1656007482087243,
"learning_rate": 8.961902705086785e-06,
"loss": 1.1531,
"step": 147
},
{
"epoch": 0.4439112177564487,
"grad_norm": 2.245423572348511,
"learning_rate": 8.947250189138732e-06,
"loss": 1.205,
"step": 148
},
{
"epoch": 0.4469106178764247,
"grad_norm": 2.2015971794373037,
"learning_rate": 8.932507144572616e-06,
"loss": 1.1663,
"step": 149
},
{
"epoch": 0.44991001799640074,
"grad_norm": 2.040881991969036,
"learning_rate": 8.917673909514321e-06,
"loss": 1.2318,
"step": 150
},
{
"epoch": 0.45290941811637675,
"grad_norm": 2.0365534777961884,
"learning_rate": 8.902750824158213e-06,
"loss": 1.1634,
"step": 151
},
{
"epoch": 0.4559088182363527,
"grad_norm": 2.0931596144968903,
"learning_rate": 8.887738230759334e-06,
"loss": 1.1547,
"step": 152
},
{
"epoch": 0.4589082183563287,
"grad_norm": 2.040316251155918,
"learning_rate": 8.872636473625564e-06,
"loss": 1.1948,
"step": 153
},
{
"epoch": 0.46190761847630474,
"grad_norm": 2.0411032824079514,
"learning_rate": 8.857445899109716e-06,
"loss": 1.1292,
"step": 154
},
{
"epoch": 0.46490701859628075,
"grad_norm": 2.9762584994335923,
"learning_rate": 8.84216685560159e-06,
"loss": 1.1308,
"step": 155
},
{
"epoch": 0.46790641871625677,
"grad_norm": 2.097668580242586,
"learning_rate": 8.826799693519996e-06,
"loss": 1.1631,
"step": 156
},
{
"epoch": 0.4709058188362327,
"grad_norm": 2.0737253370280384,
"learning_rate": 8.811344765304698e-06,
"loss": 1.1484,
"step": 157
},
{
"epoch": 0.47390521895620874,
"grad_norm": 2.1733316818144033,
"learning_rate": 8.795802425408352e-06,
"loss": 1.1992,
"step": 158
},
{
"epoch": 0.47690461907618475,
"grad_norm": 3.0100700234706927,
"learning_rate": 8.780173030288359e-06,
"loss": 1.2115,
"step": 159
},
{
"epoch": 0.47990401919616077,
"grad_norm": 2.0471682767814374,
"learning_rate": 8.7644569383987e-06,
"loss": 1.2007,
"step": 160
},
{
"epoch": 0.4829034193161368,
"grad_norm": 2.069677080263759,
"learning_rate": 8.748654510181709e-06,
"loss": 1.1719,
"step": 161
},
{
"epoch": 0.4859028194361128,
"grad_norm": 2.136133148960194,
"learning_rate": 8.732766108059814e-06,
"loss": 1.1097,
"step": 162
},
{
"epoch": 0.48890221955608876,
"grad_norm": 2.2078781415683704,
"learning_rate": 8.716792096427217e-06,
"loss": 1.2822,
"step": 163
},
{
"epoch": 0.49190161967606477,
"grad_norm": 1.8108893344256605,
"learning_rate": 8.700732841641542e-06,
"loss": 1.1984,
"step": 164
},
{
"epoch": 0.4949010197960408,
"grad_norm": 1.9202600268802827,
"learning_rate": 8.68458871201543e-06,
"loss": 1.1822,
"step": 165
},
{
"epoch": 0.4979004199160168,
"grad_norm": 2.043987291062948,
"learning_rate": 8.668360077808093e-06,
"loss": 1.1588,
"step": 166
},
{
"epoch": 0.5008998200359928,
"grad_norm": 1.8608294617929213,
"learning_rate": 8.652047311216823e-06,
"loss": 1.1739,
"step": 167
},
{
"epoch": 0.5038992201559688,
"grad_norm": 2.0614264560161812,
"learning_rate": 8.635650786368452e-06,
"loss": 1.2053,
"step": 168
},
{
"epoch": 0.5068986202759448,
"grad_norm": 1.8415134777925706,
"learning_rate": 8.61917087931078e-06,
"loss": 1.1629,
"step": 169
},
{
"epoch": 0.5098980203959208,
"grad_norm": 2.067056790142203,
"learning_rate": 8.602607968003935e-06,
"loss": 1.1656,
"step": 170
},
{
"epoch": 0.5128974205158968,
"grad_norm": 1.9885496712381465,
"learning_rate": 8.585962432311728e-06,
"loss": 1.1061,
"step": 171
},
{
"epoch": 0.5158968206358728,
"grad_norm": 1.8542146492076237,
"learning_rate": 8.569234653992916e-06,
"loss": 1.2416,
"step": 172
},
{
"epoch": 0.5188962207558488,
"grad_norm": 2.2130247278463453,
"learning_rate": 8.552425016692464e-06,
"loss": 1.2651,
"step": 173
},
{
"epoch": 0.5218956208758249,
"grad_norm": 1.9214372548459986,
"learning_rate": 8.535533905932739e-06,
"loss": 1.2038,
"step": 174
},
{
"epoch": 0.5248950209958009,
"grad_norm": 1.8202998106553783,
"learning_rate": 8.518561709104667e-06,
"loss": 1.1806,
"step": 175
},
{
"epoch": 0.5278944211157769,
"grad_norm": 2.033480383542316,
"learning_rate": 8.501508815458856e-06,
"loss": 1.1138,
"step": 176
},
{
"epoch": 0.5308938212357528,
"grad_norm": 1.8173739641856819,
"learning_rate": 8.484375616096658e-06,
"loss": 1.0463,
"step": 177
},
{
"epoch": 0.5338932213557288,
"grad_norm": 1.9159018093250713,
"learning_rate": 8.467162503961209e-06,
"loss": 1.1438,
"step": 178
},
{
"epoch": 0.5368926214757048,
"grad_norm": 2.0146669190076416,
"learning_rate": 8.449869873828411e-06,
"loss": 1.2241,
"step": 179
},
{
"epoch": 0.5398920215956808,
"grad_norm": 2.07880204207024,
"learning_rate": 8.432498122297879e-06,
"loss": 1.2059,
"step": 180
},
{
"epoch": 0.5428914217156569,
"grad_norm": 2.1530579237214593,
"learning_rate": 8.415047647783847e-06,
"loss": 1.1945,
"step": 181
},
{
"epoch": 0.5458908218356329,
"grad_norm": 1.9176382821131852,
"learning_rate": 8.39751885050603e-06,
"loss": 1.2025,
"step": 182
},
{
"epoch": 0.5488902219556089,
"grad_norm": 2.032474526366641,
"learning_rate": 8.379912132480441e-06,
"loss": 1.1962,
"step": 183
},
{
"epoch": 0.5518896220755849,
"grad_norm": 2.454514968342902,
"learning_rate": 8.36222789751018e-06,
"loss": 1.2075,
"step": 184
},
{
"epoch": 0.5548890221955609,
"grad_norm": 2.033627242661126,
"learning_rate": 8.344466551176163e-06,
"loss": 1.2134,
"step": 185
},
{
"epoch": 0.5578884223155369,
"grad_norm": 2.0441091460883585,
"learning_rate": 8.326628500827826e-06,
"loss": 1.2038,
"step": 186
},
{
"epoch": 0.5608878224355129,
"grad_norm": 1.9807070431779146,
"learning_rate": 8.308714155573785e-06,
"loss": 1.12,
"step": 187
},
{
"epoch": 0.563887222555489,
"grad_norm": 1.9349184602121337,
"learning_rate": 8.290723926272439e-06,
"loss": 1.1802,
"step": 188
},
{
"epoch": 0.5668866226754649,
"grad_norm": 1.8384106015909107,
"learning_rate": 8.27265822552257e-06,
"loss": 1.1826,
"step": 189
},
{
"epoch": 0.5698860227954409,
"grad_norm": 1.9156655051619718,
"learning_rate": 8.254517467653858e-06,
"loss": 1.1257,
"step": 190
},
{
"epoch": 0.5728854229154169,
"grad_norm": 2.1724786856863973,
"learning_rate": 8.236302068717393e-06,
"loss": 1.1839,
"step": 191
},
{
"epoch": 0.5758848230353929,
"grad_norm": 1.8396024645135292,
"learning_rate": 8.218012446476128e-06,
"loss": 1.16,
"step": 192
},
{
"epoch": 0.5788842231553689,
"grad_norm": 1.9408487099445306,
"learning_rate": 8.199649020395299e-06,
"loss": 1.2241,
"step": 193
},
{
"epoch": 0.5818836232753449,
"grad_norm": 1.9108055995475,
"learning_rate": 8.1812122116328e-06,
"loss": 1.164,
"step": 194
},
{
"epoch": 0.584883023395321,
"grad_norm": 2.098400399811293,
"learning_rate": 8.16270244302953e-06,
"loss": 1.1549,
"step": 195
},
{
"epoch": 0.587882423515297,
"grad_norm": 2.1600698972560015,
"learning_rate": 8.144120139099697e-06,
"loss": 1.2257,
"step": 196
},
{
"epoch": 0.590881823635273,
"grad_norm": 2.4157151117755826,
"learning_rate": 8.125465726021068e-06,
"loss": 1.2377,
"step": 197
},
{
"epoch": 0.593881223755249,
"grad_norm": 2.0852805021246246,
"learning_rate": 8.106739631625216e-06,
"loss": 1.2469,
"step": 198
},
{
"epoch": 0.596880623875225,
"grad_norm": 1.9748280355919685,
"learning_rate": 8.08794228538769e-06,
"loss": 1.1286,
"step": 199
},
{
"epoch": 0.5998800239952009,
"grad_norm": 1.9281839006734618,
"learning_rate": 8.06907411841817e-06,
"loss": 1.1319,
"step": 200
},
{
"epoch": 0.6028794241151769,
"grad_norm": 1.8690571721258256,
"learning_rate": 8.050135563450587e-06,
"loss": 1.1856,
"step": 201
},
{
"epoch": 0.6058788242351529,
"grad_norm": 1.9748712035222622,
"learning_rate": 8.031127054833192e-06,
"loss": 1.1948,
"step": 202
},
{
"epoch": 0.608878224355129,
"grad_norm": 1.991151818988732,
"learning_rate": 8.01204902851859e-06,
"loss": 1.1659,
"step": 203
},
{
"epoch": 0.611877624475105,
"grad_norm": 2.958946009978542,
"learning_rate": 7.992901922053751e-06,
"loss": 1.1856,
"step": 204
},
{
"epoch": 0.614877024595081,
"grad_norm": 2.065918241043226,
"learning_rate": 7.973686174569971e-06,
"loss": 1.0692,
"step": 205
},
{
"epoch": 0.617876424715057,
"grad_norm": 2.123952469982477,
"learning_rate": 7.954402226772804e-06,
"loss": 1.1462,
"step": 206
},
{
"epoch": 0.620875824835033,
"grad_norm": 2.384554268129999,
"learning_rate": 7.93505052093194e-06,
"loss": 1.2331,
"step": 207
},
{
"epoch": 0.623875224955009,
"grad_norm": 2.213816272262928,
"learning_rate": 7.915631500871084e-06,
"loss": 1.2179,
"step": 208
},
{
"epoch": 0.626874625074985,
"grad_norm": 2.041537464539173,
"learning_rate": 7.896145611957759e-06,
"loss": 1.1869,
"step": 209
},
{
"epoch": 0.629874025194961,
"grad_norm": 1.9032099687526678,
"learning_rate": 7.876593301093104e-06,
"loss": 1.1041,
"step": 210
},
{
"epoch": 0.6328734253149371,
"grad_norm": 2.0485786710751452,
"learning_rate": 7.856975016701616e-06,
"loss": 1.2018,
"step": 211
},
{
"epoch": 0.635872825434913,
"grad_norm": 1.985579363766555,
"learning_rate": 7.837291208720867e-06,
"loss": 1.2284,
"step": 212
},
{
"epoch": 0.638872225554889,
"grad_norm": 1.8372693078207498,
"learning_rate": 7.81754232859119e-06,
"loss": 1.2194,
"step": 213
},
{
"epoch": 0.641871625674865,
"grad_norm": 2.034191799766837,
"learning_rate": 7.797728829245321e-06,
"loss": 1.2119,
"step": 214
},
{
"epoch": 0.644871025794841,
"grad_norm": 2.172768388345461,
"learning_rate": 7.777851165098012e-06,
"loss": 1.2264,
"step": 215
},
{
"epoch": 0.647870425914817,
"grad_norm": 2.036813209788241,
"learning_rate": 7.757909792035608e-06,
"loss": 1.2233,
"step": 216
},
{
"epoch": 0.650869826034793,
"grad_norm": 1.7687192297667345,
"learning_rate": 7.737905167405596e-06,
"loss": 1.2008,
"step": 217
},
{
"epoch": 0.6538692261547691,
"grad_norm": 1.9034513819160714,
"learning_rate": 7.717837750006106e-06,
"loss": 1.2116,
"step": 218
},
{
"epoch": 0.6568686262747451,
"grad_norm": 1.896055349909721,
"learning_rate": 7.697708000075404e-06,
"loss": 1.0943,
"step": 219
},
{
"epoch": 0.6598680263947211,
"grad_norm": 2.01228606334524,
"learning_rate": 7.67751637928132e-06,
"loss": 1.1236,
"step": 220
},
{
"epoch": 0.6628674265146971,
"grad_norm": 2.156992216970816,
"learning_rate": 7.657263350710676e-06,
"loss": 1.1291,
"step": 221
},
{
"epoch": 0.6658668266346731,
"grad_norm": 1.9699773673672667,
"learning_rate": 7.636949378858647e-06,
"loss": 1.1458,
"step": 222
},
{
"epoch": 0.668866226754649,
"grad_norm": 2.1989186993052754,
"learning_rate": 7.616574929618126e-06,
"loss": 1.1121,
"step": 223
},
{
"epoch": 0.671865626874625,
"grad_norm": 1.993731506914194,
"learning_rate": 7.596140470269029e-06,
"loss": 1.0735,
"step": 224
},
{
"epoch": 0.674865026994601,
"grad_norm": 1.8252968956264437,
"learning_rate": 7.575646469467576e-06,
"loss": 1.149,
"step": 225
},
{
"epoch": 0.6778644271145771,
"grad_norm": 2.055158428727496,
"learning_rate": 7.555093397235553e-06,
"loss": 1.154,
"step": 226
},
{
"epoch": 0.6808638272345531,
"grad_norm": 1.9915560854905294,
"learning_rate": 7.5344817249495195e-06,
"loss": 1.0942,
"step": 227
},
{
"epoch": 0.6838632273545291,
"grad_norm": 2.06764585971135,
"learning_rate": 7.51381192533001e-06,
"loss": 1.1884,
"step": 228
},
{
"epoch": 0.6868626274745051,
"grad_norm": 1.9341247237628625,
"learning_rate": 7.493084472430683e-06,
"loss": 1.1791,
"step": 229
},
{
"epoch": 0.6898620275944811,
"grad_norm": 1.9931838011874805,
"learning_rate": 7.472299841627452e-06,
"loss": 1.1299,
"step": 230
},
{
"epoch": 0.6928614277144571,
"grad_norm": 1.9296797551564173,
"learning_rate": 7.451458509607583e-06,
"loss": 1.1513,
"step": 231
},
{
"epoch": 0.6958608278344331,
"grad_norm": 2.0306426071035544,
"learning_rate": 7.430560954358764e-06,
"loss": 1.1837,
"step": 232
},
{
"epoch": 0.6988602279544092,
"grad_norm": 1.8564265209276956,
"learning_rate": 7.409607655158139e-06,
"loss": 1.2164,
"step": 233
},
{
"epoch": 0.7018596280743852,
"grad_norm": 2.2341388452646727,
"learning_rate": 7.388599092561315e-06,
"loss": 1.1252,
"step": 234
},
{
"epoch": 0.7048590281943611,
"grad_norm": 2.1161776966895496,
"learning_rate": 7.367535748391349e-06,
"loss": 1.213,
"step": 235
},
{
"epoch": 0.7078584283143371,
"grad_norm": 1.9243533852197758,
"learning_rate": 7.3464181057276864e-06,
"loss": 1.1046,
"step": 236
},
{
"epoch": 0.7108578284343131,
"grad_norm": 1.9237683937108236,
"learning_rate": 7.325246648895089e-06,
"loss": 1.1577,
"step": 237
},
{
"epoch": 0.7138572285542891,
"grad_norm": 1.7984916602185241,
"learning_rate": 7.304021863452525e-06,
"loss": 1.2052,
"step": 238
},
{
"epoch": 0.7168566286742651,
"grad_norm": 1.8721994512639362,
"learning_rate": 7.282744236182033e-06,
"loss": 1.1676,
"step": 239
},
{
"epoch": 0.7198560287942412,
"grad_norm": 1.8334072031704922,
"learning_rate": 7.261414255077561e-06,
"loss": 1.1118,
"step": 240
},
{
"epoch": 0.7228554289142172,
"grad_norm": 2.2489377670615474,
"learning_rate": 7.240032409333765e-06,
"loss": 1.1682,
"step": 241
},
{
"epoch": 0.7258548290341932,
"grad_norm": 1.8400363638367707,
"learning_rate": 7.218599189334799e-06,
"loss": 1.1638,
"step": 242
},
{
"epoch": 0.7288542291541692,
"grad_norm": 1.8851878635905157,
"learning_rate": 7.197115086643069e-06,
"loss": 1.1176,
"step": 243
},
{
"epoch": 0.7318536292741452,
"grad_norm": 1.8023550249649072,
"learning_rate": 7.175580593987952e-06,
"loss": 1.1164,
"step": 244
},
{
"epoch": 0.7348530293941212,
"grad_norm": 2.694191659057711,
"learning_rate": 7.153996205254495e-06,
"loss": 1.1573,
"step": 245
},
{
"epoch": 0.7378524295140971,
"grad_norm": 1.964255695607597,
"learning_rate": 7.132362415472099e-06,
"loss": 1.0549,
"step": 246
},
{
"epoch": 0.7408518296340731,
"grad_norm": 1.8017763616156734,
"learning_rate": 7.1106797208031554e-06,
"loss": 1.1446,
"step": 247
},
{
"epoch": 0.7438512297540492,
"grad_norm": 1.9428097126313124,
"learning_rate": 7.088948618531668e-06,
"loss": 1.1081,
"step": 248
},
{
"epoch": 0.7468506298740252,
"grad_norm": 2.002394147288212,
"learning_rate": 7.067169607051851e-06,
"loss": 1.1359,
"step": 249
},
{
"epoch": 0.7498500299940012,
"grad_norm": 1.8198961995762413,
"learning_rate": 7.045343185856701e-06,
"loss": 1.1201,
"step": 250
},
{
"epoch": 0.7528494301139772,
"grad_norm": 1.9151331196894974,
"learning_rate": 7.02346985552653e-06,
"loss": 1.0402,
"step": 251
},
{
"epoch": 0.7558488302339532,
"grad_norm": 1.9384378002461773,
"learning_rate": 7.001550117717499e-06,
"loss": 1.175,
"step": 252
}
],
"logging_steps": 1,
"max_steps": 666,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 84,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 139507934232576.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}