Gwanwoo's picture
Upload folder using huggingface_hub
591221e verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.974025974025974,
"eval_steps": 39,
"global_step": 308,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.006493506493506494,
"grad_norm": 0.39616659283638,
"learning_rate": 1.0000000000000002e-06,
"loss": 1.6277,
"step": 1
},
{
"epoch": 0.006493506493506494,
"eval_loss": 1.6379814147949219,
"eval_runtime": 25.2293,
"eval_samples_per_second": 11.693,
"eval_steps_per_second": 1.467,
"step": 1
},
{
"epoch": 0.012987012987012988,
"grad_norm": 0.42422759532928467,
"learning_rate": 2.0000000000000003e-06,
"loss": 1.6549,
"step": 2
},
{
"epoch": 0.01948051948051948,
"grad_norm": 0.40566202998161316,
"learning_rate": 3e-06,
"loss": 1.6115,
"step": 3
},
{
"epoch": 0.025974025974025976,
"grad_norm": 0.3940100371837616,
"learning_rate": 4.000000000000001e-06,
"loss": 1.617,
"step": 4
},
{
"epoch": 0.032467532467532464,
"grad_norm": 0.3876812756061554,
"learning_rate": 5e-06,
"loss": 1.6561,
"step": 5
},
{
"epoch": 0.03896103896103896,
"grad_norm": 0.41130709648132324,
"learning_rate": 6e-06,
"loss": 1.606,
"step": 6
},
{
"epoch": 0.045454545454545456,
"grad_norm": 0.3654179573059082,
"learning_rate": 7e-06,
"loss": 1.605,
"step": 7
},
{
"epoch": 0.05194805194805195,
"grad_norm": 0.35875341296195984,
"learning_rate": 8.000000000000001e-06,
"loss": 1.5651,
"step": 8
},
{
"epoch": 0.05844155844155844,
"grad_norm": 0.3530851900577545,
"learning_rate": 9e-06,
"loss": 1.5201,
"step": 9
},
{
"epoch": 0.06493506493506493,
"grad_norm": 0.38295701146125793,
"learning_rate": 1e-05,
"loss": 1.6376,
"step": 10
},
{
"epoch": 0.07142857142857142,
"grad_norm": 0.39765068888664246,
"learning_rate": 9.999722154604716e-06,
"loss": 1.5167,
"step": 11
},
{
"epoch": 0.07792207792207792,
"grad_norm": 0.44107645750045776,
"learning_rate": 9.99888864929809e-06,
"loss": 1.6259,
"step": 12
},
{
"epoch": 0.08441558441558442,
"grad_norm": 0.4228634238243103,
"learning_rate": 9.997499576714369e-06,
"loss": 1.6767,
"step": 13
},
{
"epoch": 0.09090909090909091,
"grad_norm": 0.4122574031352997,
"learning_rate": 9.995555091232516e-06,
"loss": 1.6082,
"step": 14
},
{
"epoch": 0.09740259740259741,
"grad_norm": 0.43088310956954956,
"learning_rate": 9.99305540895907e-06,
"loss": 1.6651,
"step": 15
},
{
"epoch": 0.1038961038961039,
"grad_norm": 0.4325619637966156,
"learning_rate": 9.990000807704114e-06,
"loss": 1.6473,
"step": 16
},
{
"epoch": 0.11038961038961038,
"grad_norm": 0.4201916456222534,
"learning_rate": 9.986391626950405e-06,
"loss": 1.5726,
"step": 17
},
{
"epoch": 0.11688311688311688,
"grad_norm": 0.4616028070449829,
"learning_rate": 9.982228267815644e-06,
"loss": 1.5994,
"step": 18
},
{
"epoch": 0.12337662337662338,
"grad_norm": 0.4239194989204407,
"learning_rate": 9.977511193007896e-06,
"loss": 1.6018,
"step": 19
},
{
"epoch": 0.12987012987012986,
"grad_norm": 0.41312310099601746,
"learning_rate": 9.972240926774167e-06,
"loss": 1.6497,
"step": 20
},
{
"epoch": 0.13636363636363635,
"grad_norm": 0.45023512840270996,
"learning_rate": 9.966418054842143e-06,
"loss": 1.557,
"step": 21
},
{
"epoch": 0.14285714285714285,
"grad_norm": 0.4155130982398987,
"learning_rate": 9.960043224355081e-06,
"loss": 1.5782,
"step": 22
},
{
"epoch": 0.14935064935064934,
"grad_norm": 0.39791709184646606,
"learning_rate": 9.9531171437999e-06,
"loss": 1.548,
"step": 23
},
{
"epoch": 0.15584415584415584,
"grad_norm": 0.40294748544692993,
"learning_rate": 9.945640582928438e-06,
"loss": 1.5745,
"step": 24
},
{
"epoch": 0.16233766233766234,
"grad_norm": 0.3911401927471161,
"learning_rate": 9.937614372671896e-06,
"loss": 1.5653,
"step": 25
},
{
"epoch": 0.16883116883116883,
"grad_norm": 0.407604843378067,
"learning_rate": 9.929039405048502e-06,
"loss": 1.5627,
"step": 26
},
{
"epoch": 0.17532467532467533,
"grad_norm": 0.4176672697067261,
"learning_rate": 9.919916633064363e-06,
"loss": 1.5413,
"step": 27
},
{
"epoch": 0.18181818181818182,
"grad_norm": 0.40910884737968445,
"learning_rate": 9.91024707060755e-06,
"loss": 1.5688,
"step": 28
},
{
"epoch": 0.18831168831168832,
"grad_norm": 0.393100768327713,
"learning_rate": 9.900031792335432e-06,
"loss": 1.5259,
"step": 29
},
{
"epoch": 0.19480519480519481,
"grad_norm": 0.39668065309524536,
"learning_rate": 9.889271933555214e-06,
"loss": 1.5471,
"step": 30
},
{
"epoch": 0.2012987012987013,
"grad_norm": 0.37264811992645264,
"learning_rate": 9.877968690097785e-06,
"loss": 1.488,
"step": 31
},
{
"epoch": 0.2077922077922078,
"grad_norm": 0.39069730043411255,
"learning_rate": 9.866123318184803e-06,
"loss": 1.5863,
"step": 32
},
{
"epoch": 0.21428571428571427,
"grad_norm": 0.37075284123420715,
"learning_rate": 9.853737134289086e-06,
"loss": 1.4967,
"step": 33
},
{
"epoch": 0.22077922077922077,
"grad_norm": 0.37381497025489807,
"learning_rate": 9.840811514988294e-06,
"loss": 1.4219,
"step": 34
},
{
"epoch": 0.22727272727272727,
"grad_norm": 0.37528884410858154,
"learning_rate": 9.827347896811954e-06,
"loss": 1.4754,
"step": 35
},
{
"epoch": 0.23376623376623376,
"grad_norm": 0.3899977505207062,
"learning_rate": 9.81334777608179e-06,
"loss": 1.5201,
"step": 36
},
{
"epoch": 0.24025974025974026,
"grad_norm": 0.37340793013572693,
"learning_rate": 9.798812708745431e-06,
"loss": 1.4467,
"step": 37
},
{
"epoch": 0.24675324675324675,
"grad_norm": 0.3682302236557007,
"learning_rate": 9.783744310203492e-06,
"loss": 1.3911,
"step": 38
},
{
"epoch": 0.2532467532467532,
"grad_norm": 0.37759163975715637,
"learning_rate": 9.76814425513003e-06,
"loss": 1.4242,
"step": 39
},
{
"epoch": 0.2532467532467532,
"eval_loss": 1.451162338256836,
"eval_runtime": 25.2272,
"eval_samples_per_second": 11.694,
"eval_steps_per_second": 1.467,
"step": 39
},
{
"epoch": 0.2597402597402597,
"grad_norm": 0.3705059885978699,
"learning_rate": 9.752014277286433e-06,
"loss": 1.4542,
"step": 40
},
{
"epoch": 0.2662337662337662,
"grad_norm": 0.34180477261543274,
"learning_rate": 9.73535616932873e-06,
"loss": 1.4108,
"step": 41
},
{
"epoch": 0.2727272727272727,
"grad_norm": 0.34464818239212036,
"learning_rate": 9.718171782608355e-06,
"loss": 1.4691,
"step": 42
},
{
"epoch": 0.2792207792207792,
"grad_norm": 0.37641096115112305,
"learning_rate": 9.7004630269664e-06,
"loss": 1.3784,
"step": 43
},
{
"epoch": 0.2857142857142857,
"grad_norm": 0.335043340921402,
"learning_rate": 9.682231870521347e-06,
"loss": 1.4899,
"step": 44
},
{
"epoch": 0.2922077922077922,
"grad_norm": 0.3437371551990509,
"learning_rate": 9.663480339450344e-06,
"loss": 1.4003,
"step": 45
},
{
"epoch": 0.2987012987012987,
"grad_norm": 0.3729651868343353,
"learning_rate": 9.644210517764014e-06,
"loss": 1.3884,
"step": 46
},
{
"epoch": 0.3051948051948052,
"grad_norm": 0.33896854519844055,
"learning_rate": 9.624424547074851e-06,
"loss": 1.4258,
"step": 47
},
{
"epoch": 0.3116883116883117,
"grad_norm": 0.36020371317863464,
"learning_rate": 9.60412462635919e-06,
"loss": 1.3703,
"step": 48
},
{
"epoch": 0.3181818181818182,
"grad_norm": 0.3405916392803192,
"learning_rate": 9.583313011712832e-06,
"loss": 1.3407,
"step": 49
},
{
"epoch": 0.3246753246753247,
"grad_norm": 0.3447131812572479,
"learning_rate": 9.561992016100293e-06,
"loss": 1.3456,
"step": 50
},
{
"epoch": 0.33116883116883117,
"grad_norm": 0.35361772775650024,
"learning_rate": 9.540164009097756e-06,
"loss": 1.3384,
"step": 51
},
{
"epoch": 0.33766233766233766,
"grad_norm": 0.3525960147380829,
"learning_rate": 9.517831416629717e-06,
"loss": 1.3726,
"step": 52
},
{
"epoch": 0.34415584415584416,
"grad_norm": 0.3269696831703186,
"learning_rate": 9.494996720699363e-06,
"loss": 1.3664,
"step": 53
},
{
"epoch": 0.35064935064935066,
"grad_norm": 0.3153081238269806,
"learning_rate": 9.471662459112747e-06,
"loss": 1.3448,
"step": 54
},
{
"epoch": 0.35714285714285715,
"grad_norm": 0.3017883002758026,
"learning_rate": 9.44783122519672e-06,
"loss": 1.3228,
"step": 55
},
{
"epoch": 0.36363636363636365,
"grad_norm": 0.314627081155777,
"learning_rate": 9.423505667510724e-06,
"loss": 1.3565,
"step": 56
},
{
"epoch": 0.37012987012987014,
"grad_norm": 0.3048722743988037,
"learning_rate": 9.398688489552437e-06,
"loss": 1.2669,
"step": 57
},
{
"epoch": 0.37662337662337664,
"grad_norm": 0.3107397258281708,
"learning_rate": 9.373382449457305e-06,
"loss": 1.2871,
"step": 58
},
{
"epoch": 0.38311688311688313,
"grad_norm": 0.3061436116695404,
"learning_rate": 9.347590359692015e-06,
"loss": 1.3015,
"step": 59
},
{
"epoch": 0.38961038961038963,
"grad_norm": 0.31199324131011963,
"learning_rate": 9.321315086741916e-06,
"loss": 1.3196,
"step": 60
},
{
"epoch": 0.3961038961038961,
"grad_norm": 0.3002881109714508,
"learning_rate": 9.294559550792451e-06,
"loss": 1.3404,
"step": 61
},
{
"epoch": 0.4025974025974026,
"grad_norm": 0.2965368330478668,
"learning_rate": 9.2673267254046e-06,
"loss": 1.2931,
"step": 62
},
{
"epoch": 0.4090909090909091,
"grad_norm": 0.2912820279598236,
"learning_rate": 9.23961963718442e-06,
"loss": 1.2941,
"step": 63
},
{
"epoch": 0.4155844155844156,
"grad_norm": 0.29220762848854065,
"learning_rate": 9.211441365446661e-06,
"loss": 1.306,
"step": 64
},
{
"epoch": 0.42207792207792205,
"grad_norm": 0.2802380323410034,
"learning_rate": 9.182795041872543e-06,
"loss": 1.2641,
"step": 65
},
{
"epoch": 0.42857142857142855,
"grad_norm": 0.28170621395111084,
"learning_rate": 9.153683850161706e-06,
"loss": 1.2459,
"step": 66
},
{
"epoch": 0.43506493506493504,
"grad_norm": 0.2919502854347229,
"learning_rate": 9.124111025678378e-06,
"loss": 1.2736,
"step": 67
},
{
"epoch": 0.44155844155844154,
"grad_norm": 0.2793984115123749,
"learning_rate": 9.094079855091797e-06,
"loss": 1.265,
"step": 68
},
{
"epoch": 0.44805194805194803,
"grad_norm": 0.2946363091468811,
"learning_rate": 9.063593676010954e-06,
"loss": 1.2669,
"step": 69
},
{
"epoch": 0.45454545454545453,
"grad_norm": 0.2771857678890228,
"learning_rate": 9.032655876613636e-06,
"loss": 1.2479,
"step": 70
},
{
"epoch": 0.461038961038961,
"grad_norm": 0.26628848910331726,
"learning_rate": 9.001269895269886e-06,
"loss": 1.2764,
"step": 71
},
{
"epoch": 0.4675324675324675,
"grad_norm": 0.3006434440612793,
"learning_rate": 8.969439220159861e-06,
"loss": 1.2286,
"step": 72
},
{
"epoch": 0.474025974025974,
"grad_norm": 0.27274397015571594,
"learning_rate": 8.937167388886163e-06,
"loss": 1.3059,
"step": 73
},
{
"epoch": 0.4805194805194805,
"grad_norm": 0.2763414978981018,
"learning_rate": 8.904457988080682e-06,
"loss": 1.2095,
"step": 74
},
{
"epoch": 0.487012987012987,
"grad_norm": 0.28402575850486755,
"learning_rate": 8.871314653005972e-06,
"loss": 1.2601,
"step": 75
},
{
"epoch": 0.4935064935064935,
"grad_norm": 0.2892557680606842,
"learning_rate": 8.837741067151251e-06,
"loss": 1.2342,
"step": 76
},
{
"epoch": 0.5,
"grad_norm": 0.2644696533679962,
"learning_rate": 8.80374096182301e-06,
"loss": 1.2744,
"step": 77
},
{
"epoch": 0.5064935064935064,
"grad_norm": 0.25198492407798767,
"learning_rate": 8.76931811573033e-06,
"loss": 1.2885,
"step": 78
},
{
"epoch": 0.5064935064935064,
"eval_loss": 1.2617864608764648,
"eval_runtime": 25.3815,
"eval_samples_per_second": 11.623,
"eval_steps_per_second": 1.458,
"step": 78
},
{
"epoch": 0.512987012987013,
"grad_norm": 0.2804642915725708,
"learning_rate": 8.734476354564924e-06,
"loss": 1.2583,
"step": 79
},
{
"epoch": 0.5194805194805194,
"grad_norm": 0.2729627788066864,
"learning_rate": 8.699219550575954e-06,
"loss": 1.2246,
"step": 80
},
{
"epoch": 0.525974025974026,
"grad_norm": 0.2520177662372589,
"learning_rate": 8.663551622139674e-06,
"loss": 1.2599,
"step": 81
},
{
"epoch": 0.5324675324675324,
"grad_norm": 0.2614675760269165,
"learning_rate": 8.627476533323957e-06,
"loss": 1.2165,
"step": 82
},
{
"epoch": 0.538961038961039,
"grad_norm": 0.3191888928413391,
"learning_rate": 8.590998293447728e-06,
"loss": 1.2558,
"step": 83
},
{
"epoch": 0.5454545454545454,
"grad_norm": 0.27159151434898376,
"learning_rate": 8.554120956635375e-06,
"loss": 1.2197,
"step": 84
},
{
"epoch": 0.551948051948052,
"grad_norm": 0.291990727186203,
"learning_rate": 8.516848621366188e-06,
"loss": 1.219,
"step": 85
},
{
"epoch": 0.5584415584415584,
"grad_norm": 0.23849813640117645,
"learning_rate": 8.47918543001886e-06,
"loss": 1.2199,
"step": 86
},
{
"epoch": 0.564935064935065,
"grad_norm": 0.26883506774902344,
"learning_rate": 8.441135568411102e-06,
"loss": 1.1959,
"step": 87
},
{
"epoch": 0.5714285714285714,
"grad_norm": 0.2667544484138489,
"learning_rate": 8.402703265334455e-06,
"loss": 1.2682,
"step": 88
},
{
"epoch": 0.577922077922078,
"grad_norm": 0.24013420939445496,
"learning_rate": 8.363892792084291e-06,
"loss": 1.1649,
"step": 89
},
{
"epoch": 0.5844155844155844,
"grad_norm": 0.2544495463371277,
"learning_rate": 8.324708461985124e-06,
"loss": 1.1929,
"step": 90
},
{
"epoch": 0.5909090909090909,
"grad_norm": 0.2654297351837158,
"learning_rate": 8.285154629911227e-06,
"loss": 1.18,
"step": 91
},
{
"epoch": 0.5974025974025974,
"grad_norm": 0.26719850301742554,
"learning_rate": 8.245235691802644e-06,
"loss": 1.2933,
"step": 92
},
{
"epoch": 0.6038961038961039,
"grad_norm": 0.2760712802410126,
"learning_rate": 8.20495608417663e-06,
"loss": 1.1612,
"step": 93
},
{
"epoch": 0.6103896103896104,
"grad_norm": 0.2652733623981476,
"learning_rate": 8.164320283634585e-06,
"loss": 1.2125,
"step": 94
},
{
"epoch": 0.6168831168831169,
"grad_norm": 0.24604123830795288,
"learning_rate": 8.123332806364537e-06,
"loss": 1.1801,
"step": 95
},
{
"epoch": 0.6233766233766234,
"grad_norm": 0.23077791929244995,
"learning_rate": 8.081998207639212e-06,
"loss": 1.2016,
"step": 96
},
{
"epoch": 0.6298701298701299,
"grad_norm": 0.25489139556884766,
"learning_rate": 8.040321081309783e-06,
"loss": 1.2049,
"step": 97
},
{
"epoch": 0.6363636363636364,
"grad_norm": 0.2564036250114441,
"learning_rate": 7.998306059295302e-06,
"loss": 1.2377,
"step": 98
},
{
"epoch": 0.6428571428571429,
"grad_norm": 0.2734230160713196,
"learning_rate": 7.955957811067932e-06,
"loss": 1.1107,
"step": 99
},
{
"epoch": 0.6493506493506493,
"grad_norm": 0.2672719359397888,
"learning_rate": 7.913281043133978e-06,
"loss": 1.1863,
"step": 100
},
{
"epoch": 0.6558441558441559,
"grad_norm": 0.263724684715271,
"learning_rate": 7.870280498510824e-06,
"loss": 1.2678,
"step": 101
},
{
"epoch": 0.6623376623376623,
"grad_norm": 0.27098724246025085,
"learning_rate": 7.826960956199796e-06,
"loss": 1.1656,
"step": 102
},
{
"epoch": 0.6688311688311688,
"grad_norm": 0.29257479310035706,
"learning_rate": 7.783327230655036e-06,
"loss": 1.1749,
"step": 103
},
{
"epoch": 0.6753246753246753,
"grad_norm": 0.26874226331710815,
"learning_rate": 7.739384171248436e-06,
"loss": 1.2013,
"step": 104
},
{
"epoch": 0.6818181818181818,
"grad_norm": 0.26934632658958435,
"learning_rate": 7.695136661730677e-06,
"loss": 1.1507,
"step": 105
},
{
"epoch": 0.6883116883116883,
"grad_norm": 0.27807483077049255,
"learning_rate": 7.650589619688468e-06,
"loss": 1.1729,
"step": 106
},
{
"epoch": 0.6948051948051948,
"grad_norm": 0.2936646342277527,
"learning_rate": 7.6057479959980145e-06,
"loss": 1.1646,
"step": 107
},
{
"epoch": 0.7012987012987013,
"grad_norm": 0.28149378299713135,
"learning_rate": 7.560616774274775e-06,
"loss": 1.1939,
"step": 108
},
{
"epoch": 0.7077922077922078,
"grad_norm": 0.25706660747528076,
"learning_rate": 7.5152009703196105e-06,
"loss": 1.1708,
"step": 109
},
{
"epoch": 0.7142857142857143,
"grad_norm": 0.29964110255241394,
"learning_rate": 7.469505631561318e-06,
"loss": 1.2161,
"step": 110
},
{
"epoch": 0.7207792207792207,
"grad_norm": 0.24933487176895142,
"learning_rate": 7.423535836495683e-06,
"loss": 1.1641,
"step": 111
},
{
"epoch": 0.7272727272727273,
"grad_norm": 0.27039459347724915,
"learning_rate": 7.3772966941210585e-06,
"loss": 1.1563,
"step": 112
},
{
"epoch": 0.7337662337662337,
"grad_norm": 0.2490512579679489,
"learning_rate": 7.33079334337056e-06,
"loss": 1.1887,
"step": 113
},
{
"epoch": 0.7402597402597403,
"grad_norm": 0.28315550088882446,
"learning_rate": 7.284030952540937e-06,
"loss": 1.1189,
"step": 114
},
{
"epoch": 0.7467532467532467,
"grad_norm": 0.2557179033756256,
"learning_rate": 7.2370147187181736e-06,
"loss": 1.1812,
"step": 115
},
{
"epoch": 0.7532467532467533,
"grad_norm": 0.2875461280345917,
"learning_rate": 7.189749867199899e-06,
"loss": 1.1534,
"step": 116
},
{
"epoch": 0.7597402597402597,
"grad_norm": 0.26117077469825745,
"learning_rate": 7.142241650914654e-06,
"loss": 1.1618,
"step": 117
},
{
"epoch": 0.7597402597402597,
"eval_loss": 1.2008626461029053,
"eval_runtime": 25.2533,
"eval_samples_per_second": 11.682,
"eval_steps_per_second": 1.465,
"step": 117
},
{
"epoch": 0.7662337662337663,
"grad_norm": 0.29663676023483276,
"learning_rate": 7.094495349838093e-06,
"loss": 1.1064,
"step": 118
},
{
"epoch": 0.7727272727272727,
"grad_norm": 0.23107394576072693,
"learning_rate": 7.046516270406174e-06,
"loss": 1.1464,
"step": 119
},
{
"epoch": 0.7792207792207793,
"grad_norm": 0.2502164840698242,
"learning_rate": 6.998309744925411e-06,
"loss": 1.1998,
"step": 120
},
{
"epoch": 0.7857142857142857,
"grad_norm": 0.25331148505210876,
"learning_rate": 6.9498811309802595e-06,
"loss": 1.1784,
"step": 121
},
{
"epoch": 0.7922077922077922,
"grad_norm": 0.2596096694469452,
"learning_rate": 6.901235810837668e-06,
"loss": 1.1034,
"step": 122
},
{
"epoch": 0.7987012987012987,
"grad_norm": 0.26797452569007874,
"learning_rate": 6.852379190848923e-06,
"loss": 1.1264,
"step": 123
},
{
"epoch": 0.8051948051948052,
"grad_norm": 0.31169766187667847,
"learning_rate": 6.8033167008487784e-06,
"loss": 1.1386,
"step": 124
},
{
"epoch": 0.8116883116883117,
"grad_norm": 0.26767072081565857,
"learning_rate": 6.754053793552005e-06,
"loss": 1.2137,
"step": 125
},
{
"epoch": 0.8181818181818182,
"grad_norm": 0.2596385180950165,
"learning_rate": 6.704595943947385e-06,
"loss": 1.1466,
"step": 126
},
{
"epoch": 0.8246753246753247,
"grad_norm": 0.27891236543655396,
"learning_rate": 6.654948648689228e-06,
"loss": 1.1371,
"step": 127
},
{
"epoch": 0.8311688311688312,
"grad_norm": 0.28439176082611084,
"learning_rate": 6.605117425486483e-06,
"loss": 1.1948,
"step": 128
},
{
"epoch": 0.8376623376623377,
"grad_norm": 0.2944129705429077,
"learning_rate": 6.555107812489513e-06,
"loss": 1.1169,
"step": 129
},
{
"epoch": 0.8441558441558441,
"grad_norm": 0.2609187960624695,
"learning_rate": 6.504925367674595e-06,
"loss": 1.1503,
"step": 130
},
{
"epoch": 0.8506493506493507,
"grad_norm": 0.27614086866378784,
"learning_rate": 6.454575668226215e-06,
"loss": 1.1835,
"step": 131
},
{
"epoch": 0.8571428571428571,
"grad_norm": 0.2971368134021759,
"learning_rate": 6.40406430991723e-06,
"loss": 1.1816,
"step": 132
},
{
"epoch": 0.8636363636363636,
"grad_norm": 0.28436651825904846,
"learning_rate": 6.353396906486971e-06,
"loss": 1.1947,
"step": 133
},
{
"epoch": 0.8701298701298701,
"grad_norm": 0.2339404821395874,
"learning_rate": 6.302579089017328e-06,
"loss": 1.1027,
"step": 134
},
{
"epoch": 0.8766233766233766,
"grad_norm": 0.27757248282432556,
"learning_rate": 6.251616505306933e-06,
"loss": 1.1294,
"step": 135
},
{
"epoch": 0.8831168831168831,
"grad_norm": 0.27656033635139465,
"learning_rate": 6.200514819243476e-06,
"loss": 1.1313,
"step": 136
},
{
"epoch": 0.8896103896103896,
"grad_norm": 0.26819008588790894,
"learning_rate": 6.149279710174219e-06,
"loss": 1.2036,
"step": 137
},
{
"epoch": 0.8961038961038961,
"grad_norm": 0.3008396029472351,
"learning_rate": 6.097916872274815e-06,
"loss": 1.1512,
"step": 138
},
{
"epoch": 0.9025974025974026,
"grad_norm": 0.29651182889938354,
"learning_rate": 6.046432013916467e-06,
"loss": 1.1412,
"step": 139
},
{
"epoch": 0.9090909090909091,
"grad_norm": 0.275259405374527,
"learning_rate": 5.9948308570315e-06,
"loss": 1.1726,
"step": 140
},
{
"epoch": 0.9155844155844156,
"grad_norm": 0.26858457922935486,
"learning_rate": 5.943119136477449e-06,
"loss": 1.1701,
"step": 141
},
{
"epoch": 0.922077922077922,
"grad_norm": 0.273671954870224,
"learning_rate": 5.891302599399686e-06,
"loss": 1.165,
"step": 142
},
{
"epoch": 0.9285714285714286,
"grad_norm": 0.26044774055480957,
"learning_rate": 5.839387004592705e-06,
"loss": 1.1119,
"step": 143
},
{
"epoch": 0.935064935064935,
"grad_norm": 0.24865947663784027,
"learning_rate": 5.78737812186009e-06,
"loss": 1.1598,
"step": 144
},
{
"epoch": 0.9415584415584416,
"grad_norm": 0.2713409960269928,
"learning_rate": 5.735281731373271e-06,
"loss": 1.1543,
"step": 145
},
{
"epoch": 0.948051948051948,
"grad_norm": 0.2865453362464905,
"learning_rate": 5.6831036230291345e-06,
"loss": 1.1379,
"step": 146
},
{
"epoch": 0.9545454545454546,
"grad_norm": 0.26891422271728516,
"learning_rate": 5.630849595806534e-06,
"loss": 1.1382,
"step": 147
},
{
"epoch": 0.961038961038961,
"grad_norm": 0.3001209795475006,
"learning_rate": 5.578525457121807e-06,
"loss": 1.1674,
"step": 148
},
{
"epoch": 0.9675324675324676,
"grad_norm": 0.2672886848449707,
"learning_rate": 5.526137022183356e-06,
"loss": 1.1209,
"step": 149
},
{
"epoch": 0.974025974025974,
"grad_norm": 0.27608615159988403,
"learning_rate": 5.473690113345343e-06,
"loss": 1.1855,
"step": 150
},
{
"epoch": 0.9805194805194806,
"grad_norm": 0.2823050618171692,
"learning_rate": 5.4211905594606165e-06,
"loss": 1.1433,
"step": 151
},
{
"epoch": 0.987012987012987,
"grad_norm": 0.2888166308403015,
"learning_rate": 5.368644195232896e-06,
"loss": 1.1413,
"step": 152
},
{
"epoch": 0.9935064935064936,
"grad_norm": 0.2720174193382263,
"learning_rate": 5.316056860568318e-06,
"loss": 1.1657,
"step": 153
},
{
"epoch": 1.0,
"grad_norm": 0.28704148530960083,
"learning_rate": 5.2634343999263985e-06,
"loss": 1.1606,
"step": 154
},
{
"epoch": 1.0064935064935066,
"grad_norm": 0.2424069494009018,
"learning_rate": 5.210782661670486e-06,
"loss": 1.1506,
"step": 155
},
{
"epoch": 1.0129870129870129,
"grad_norm": 0.2754787802696228,
"learning_rate": 5.158107497417795e-06,
"loss": 1.1186,
"step": 156
},
{
"epoch": 1.0129870129870129,
"eval_loss": 1.1744325160980225,
"eval_runtime": 25.2167,
"eval_samples_per_second": 11.699,
"eval_steps_per_second": 1.467,
"step": 156
},
{
"epoch": 1.0194805194805194,
"grad_norm": 0.2637002170085907,
"learning_rate": 5.105414761389056e-06,
"loss": 1.178,
"step": 157
},
{
"epoch": 1.025974025974026,
"grad_norm": 0.26754647493362427,
"learning_rate": 5.052710309757899e-06,
"loss": 1.1329,
"step": 158
},
{
"epoch": 1.0064935064935066,
"grad_norm": 0.2765344977378845,
"learning_rate": 5e-06,
"loss": 1.0933,
"step": 159
},
{
"epoch": 1.0129870129870129,
"grad_norm": 0.2661243677139282,
"learning_rate": 4.947289690242103e-06,
"loss": 1.0931,
"step": 160
},
{
"epoch": 1.0194805194805194,
"grad_norm": 0.2913898527622223,
"learning_rate": 4.894585238610946e-06,
"loss": 1.0963,
"step": 161
},
{
"epoch": 1.025974025974026,
"grad_norm": 0.2579714357852936,
"learning_rate": 4.841892502582206e-06,
"loss": 1.1984,
"step": 162
},
{
"epoch": 1.0324675324675325,
"grad_norm": 0.2802336513996124,
"learning_rate": 4.789217338329515e-06,
"loss": 1.1592,
"step": 163
},
{
"epoch": 1.0389610389610389,
"grad_norm": 0.27885061502456665,
"learning_rate": 4.736565600073602e-06,
"loss": 1.1184,
"step": 164
},
{
"epoch": 1.0454545454545454,
"grad_norm": 0.26521897315979004,
"learning_rate": 4.683943139431683e-06,
"loss": 1.1685,
"step": 165
},
{
"epoch": 1.051948051948052,
"grad_norm": 0.30475732684135437,
"learning_rate": 4.631355804767106e-06,
"loss": 1.105,
"step": 166
},
{
"epoch": 1.0584415584415585,
"grad_norm": 0.2804529070854187,
"learning_rate": 4.578809440539386e-06,
"loss": 1.0735,
"step": 167
},
{
"epoch": 1.0649350649350648,
"grad_norm": 0.2721943259239197,
"learning_rate": 4.526309886654659e-06,
"loss": 1.187,
"step": 168
},
{
"epoch": 1.0714285714285714,
"grad_norm": 0.2899441421031952,
"learning_rate": 4.473862977816647e-06,
"loss": 1.1375,
"step": 169
},
{
"epoch": 1.077922077922078,
"grad_norm": 0.27229487895965576,
"learning_rate": 4.4214745428781946e-06,
"loss": 1.1079,
"step": 170
},
{
"epoch": 1.0844155844155845,
"grad_norm": 0.28253939747810364,
"learning_rate": 4.369150404193467e-06,
"loss": 1.1283,
"step": 171
},
{
"epoch": 1.0909090909090908,
"grad_norm": 0.31305843591690063,
"learning_rate": 4.316896376970866e-06,
"loss": 1.1576,
"step": 172
},
{
"epoch": 1.0974025974025974,
"grad_norm": 0.26900458335876465,
"learning_rate": 4.264718268626729e-06,
"loss": 1.1694,
"step": 173
},
{
"epoch": 1.103896103896104,
"grad_norm": 0.27618759870529175,
"learning_rate": 4.212621878139912e-06,
"loss": 1.0882,
"step": 174
},
{
"epoch": 1.1103896103896105,
"grad_norm": 0.3342863619327545,
"learning_rate": 4.160612995407296e-06,
"loss": 1.1103,
"step": 175
},
{
"epoch": 1.1168831168831168,
"grad_norm": 0.3191912770271301,
"learning_rate": 4.108697400600316e-06,
"loss": 1.1528,
"step": 176
},
{
"epoch": 1.1233766233766234,
"grad_norm": 0.28136956691741943,
"learning_rate": 4.056880863522553e-06,
"loss": 1.1239,
"step": 177
},
{
"epoch": 1.12987012987013,
"grad_norm": 0.3044598400592804,
"learning_rate": 4.005169142968503e-06,
"loss": 1.1396,
"step": 178
},
{
"epoch": 1.1363636363636362,
"grad_norm": 0.26966431736946106,
"learning_rate": 3.953567986083535e-06,
"loss": 1.1166,
"step": 179
},
{
"epoch": 1.1428571428571428,
"grad_norm": 0.2642230689525604,
"learning_rate": 3.902083127725186e-06,
"loss": 1.1156,
"step": 180
},
{
"epoch": 1.1493506493506493,
"grad_norm": 0.27723589539527893,
"learning_rate": 3.850720289825783e-06,
"loss": 1.106,
"step": 181
},
{
"epoch": 1.155844155844156,
"grad_norm": 0.25195732712745667,
"learning_rate": 3.799485180756526e-06,
"loss": 1.1473,
"step": 182
},
{
"epoch": 1.1623376623376624,
"grad_norm": 0.2634832561016083,
"learning_rate": 3.7483834946930682e-06,
"loss": 1.12,
"step": 183
},
{
"epoch": 1.1688311688311688,
"grad_norm": 0.2719174027442932,
"learning_rate": 3.6974209109826724e-06,
"loss": 1.1379,
"step": 184
},
{
"epoch": 1.1753246753246753,
"grad_norm": 0.2815384566783905,
"learning_rate": 3.64660309351303e-06,
"loss": 1.1548,
"step": 185
},
{
"epoch": 1.1818181818181819,
"grad_norm": 0.2685678005218506,
"learning_rate": 3.595935690082769e-06,
"loss": 1.1321,
"step": 186
},
{
"epoch": 1.1883116883116882,
"grad_norm": 0.282886803150177,
"learning_rate": 3.545424331773787e-06,
"loss": 1.159,
"step": 187
},
{
"epoch": 1.1948051948051948,
"grad_norm": 0.2961716055870056,
"learning_rate": 3.495074632325407e-06,
"loss": 1.0941,
"step": 188
},
{
"epoch": 1.2012987012987013,
"grad_norm": 0.26541033387184143,
"learning_rate": 3.4448921875104898e-06,
"loss": 1.1487,
"step": 189
},
{
"epoch": 1.2077922077922079,
"grad_norm": 0.3059718608856201,
"learning_rate": 3.3948825745135196e-06,
"loss": 1.1529,
"step": 190
},
{
"epoch": 1.2142857142857142,
"grad_norm": 0.2965088486671448,
"learning_rate": 3.345051351310774e-06,
"loss": 1.0652,
"step": 191
},
{
"epoch": 1.2207792207792207,
"grad_norm": 0.2668808102607727,
"learning_rate": 3.295404056052616e-06,
"loss": 1.1231,
"step": 192
},
{
"epoch": 1.2272727272727273,
"grad_norm": 0.2890554666519165,
"learning_rate": 3.2459462064479972e-06,
"loss": 1.15,
"step": 193
},
{
"epoch": 1.2337662337662338,
"grad_norm": 0.25789421796798706,
"learning_rate": 3.1966832991512232e-06,
"loss": 1.1434,
"step": 194
},
{
"epoch": 1.2402597402597402,
"grad_norm": 0.28866487741470337,
"learning_rate": 3.147620809151078e-06,
"loss": 1.1237,
"step": 195
},
{
"epoch": 1.2402597402597402,
"eval_loss": 1.1614325046539307,
"eval_runtime": 25.3147,
"eval_samples_per_second": 11.653,
"eval_steps_per_second": 1.462,
"step": 195
},
{
"epoch": 1.2467532467532467,
"grad_norm": 0.26898378133773804,
"learning_rate": 3.098764189162332e-06,
"loss": 1.1265,
"step": 196
},
{
"epoch": 1.2532467532467533,
"grad_norm": 0.2718038260936737,
"learning_rate": 3.0501188690197418e-06,
"loss": 1.1012,
"step": 197
},
{
"epoch": 1.2597402597402598,
"grad_norm": 0.27878862619400024,
"learning_rate": 3.0016902550745896e-06,
"loss": 1.1454,
"step": 198
},
{
"epoch": 1.2662337662337662,
"grad_norm": 0.28224632143974304,
"learning_rate": 2.9534837295938268e-06,
"loss": 1.1561,
"step": 199
},
{
"epoch": 1.2727272727272727,
"grad_norm": 0.2997869849205017,
"learning_rate": 2.9055046501619088e-06,
"loss": 1.1408,
"step": 200
},
{
"epoch": 1.2792207792207793,
"grad_norm": 0.29859447479248047,
"learning_rate": 2.857758349085348e-06,
"loss": 1.1105,
"step": 201
},
{
"epoch": 1.2857142857142856,
"grad_norm": 0.2930034101009369,
"learning_rate": 2.810250132800103e-06,
"loss": 1.1113,
"step": 202
},
{
"epoch": 1.2922077922077921,
"grad_norm": 0.29845744371414185,
"learning_rate": 2.762985281281828e-06,
"loss": 1.1702,
"step": 203
},
{
"epoch": 1.2987012987012987,
"grad_norm": 0.27331170439720154,
"learning_rate": 2.715969047459066e-06,
"loss": 1.1197,
"step": 204
},
{
"epoch": 1.3051948051948052,
"grad_norm": 0.31652095913887024,
"learning_rate": 2.6692066566294393e-06,
"loss": 1.1422,
"step": 205
},
{
"epoch": 1.3116883116883118,
"grad_norm": 0.3164924383163452,
"learning_rate": 2.622703305878941e-06,
"loss": 1.1179,
"step": 206
},
{
"epoch": 1.3181818181818181,
"grad_norm": 0.25467604398727417,
"learning_rate": 2.5764641635043174e-06,
"loss": 1.0839,
"step": 207
},
{
"epoch": 1.3246753246753247,
"grad_norm": 0.2731601297855377,
"learning_rate": 2.530494368438683e-06,
"loss": 1.1033,
"step": 208
},
{
"epoch": 1.3311688311688312,
"grad_norm": 0.2869837284088135,
"learning_rate": 2.4847990296803907e-06,
"loss": 1.1179,
"step": 209
},
{
"epoch": 1.3376623376623376,
"grad_norm": 0.2819492816925049,
"learning_rate": 2.4393832257252253e-06,
"loss": 1.012,
"step": 210
},
{
"epoch": 1.344155844155844,
"grad_norm": 0.267045259475708,
"learning_rate": 2.394252004001989e-06,
"loss": 1.0931,
"step": 211
},
{
"epoch": 1.3506493506493507,
"grad_norm": 0.32312628626823425,
"learning_rate": 2.349410380311532e-06,
"loss": 1.1414,
"step": 212
},
{
"epoch": 1.3571428571428572,
"grad_norm": 0.25884246826171875,
"learning_rate": 2.304863338269326e-06,
"loss": 1.126,
"step": 213
},
{
"epoch": 1.3636363636363638,
"grad_norm": 0.3011741638183594,
"learning_rate": 2.2606158287515662e-06,
"loss": 1.1247,
"step": 214
},
{
"epoch": 1.37012987012987,
"grad_norm": 0.2756052017211914,
"learning_rate": 2.216672769344965e-06,
"loss": 1.1196,
"step": 215
},
{
"epoch": 1.3766233766233766,
"grad_norm": 0.2793034613132477,
"learning_rate": 2.1730390438002056e-06,
"loss": 1.1372,
"step": 216
},
{
"epoch": 1.3831168831168832,
"grad_norm": 0.2940700948238373,
"learning_rate": 2.129719501489177e-06,
"loss": 1.0846,
"step": 217
},
{
"epoch": 1.3896103896103895,
"grad_norm": 0.30984777212142944,
"learning_rate": 2.086718956866024e-06,
"loss": 1.1404,
"step": 218
},
{
"epoch": 1.396103896103896,
"grad_norm": 0.25352832674980164,
"learning_rate": 2.044042188932068e-06,
"loss": 1.1223,
"step": 219
},
{
"epoch": 1.4025974025974026,
"grad_norm": 0.31949445605278015,
"learning_rate": 2.0016939407046987e-06,
"loss": 1.0899,
"step": 220
},
{
"epoch": 1.4090909090909092,
"grad_norm": 0.2721538245677948,
"learning_rate": 1.9596789186902184e-06,
"loss": 1.0571,
"step": 221
},
{
"epoch": 1.4155844155844157,
"grad_norm": 0.2586475610733032,
"learning_rate": 1.9180017923607884e-06,
"loss": 1.0164,
"step": 222
},
{
"epoch": 1.422077922077922,
"grad_norm": 0.29858991503715515,
"learning_rate": 1.8766671936354647e-06,
"loss": 1.155,
"step": 223
},
{
"epoch": 1.4285714285714286,
"grad_norm": 0.31114456057548523,
"learning_rate": 1.8356797163654172e-06,
"loss": 1.1546,
"step": 224
},
{
"epoch": 1.435064935064935,
"grad_norm": 0.27457138895988464,
"learning_rate": 1.795043915823373e-06,
"loss": 1.1159,
"step": 225
},
{
"epoch": 1.4415584415584415,
"grad_norm": 0.31147733330726624,
"learning_rate": 1.754764308197358e-06,
"loss": 1.1316,
"step": 226
},
{
"epoch": 1.448051948051948,
"grad_norm": 0.30539748072624207,
"learning_rate": 1.7148453700887747e-06,
"loss": 1.1135,
"step": 227
},
{
"epoch": 1.4545454545454546,
"grad_norm": 0.2620432674884796,
"learning_rate": 1.6752915380148772e-06,
"loss": 1.136,
"step": 228
},
{
"epoch": 1.4610389610389611,
"grad_norm": 0.3202950060367584,
"learning_rate": 1.6361072079157092e-06,
"loss": 1.1635,
"step": 229
},
{
"epoch": 1.4675324675324675,
"grad_norm": 0.3076479732990265,
"learning_rate": 1.5972967346655449e-06,
"loss": 1.1305,
"step": 230
},
{
"epoch": 1.474025974025974,
"grad_norm": 0.2635696828365326,
"learning_rate": 1.5588644315888978e-06,
"loss": 1.0887,
"step": 231
},
{
"epoch": 1.4805194805194806,
"grad_norm": 0.30404505133628845,
"learning_rate": 1.5208145699811417e-06,
"loss": 1.1176,
"step": 232
},
{
"epoch": 1.487012987012987,
"grad_norm": 0.29464152455329895,
"learning_rate": 1.4831513786338126e-06,
"loss": 1.1724,
"step": 233
},
{
"epoch": 1.4935064935064934,
"grad_norm": 0.3045382499694824,
"learning_rate": 1.4458790433646264e-06,
"loss": 1.1151,
"step": 234
},
{
"epoch": 1.4935064935064934,
"eval_loss": 1.1553453207015991,
"eval_runtime": 25.2983,
"eval_samples_per_second": 11.661,
"eval_steps_per_second": 1.463,
"step": 234
},
{
"epoch": 1.5,
"grad_norm": 0.2871938943862915,
"learning_rate": 1.4090017065522731e-06,
"loss": 1.1687,
"step": 235
},
{
"epoch": 1.5064935064935066,
"grad_norm": 0.2857874035835266,
"learning_rate": 1.3725234666760428e-06,
"loss": 1.1089,
"step": 236
},
{
"epoch": 1.512987012987013,
"grad_norm": 0.3428025245666504,
"learning_rate": 1.3364483778603272e-06,
"loss": 1.1335,
"step": 237
},
{
"epoch": 1.5194805194805194,
"grad_norm": 0.3187818229198456,
"learning_rate": 1.3007804494240478e-06,
"loss": 1.163,
"step": 238
},
{
"epoch": 1.525974025974026,
"grad_norm": 0.28887903690338135,
"learning_rate": 1.2655236454350777e-06,
"loss": 1.119,
"step": 239
},
{
"epoch": 1.5324675324675323,
"grad_norm": 0.28905734419822693,
"learning_rate": 1.2306818842696716e-06,
"loss": 1.1448,
"step": 240
},
{
"epoch": 1.5389610389610389,
"grad_norm": 0.25675976276397705,
"learning_rate": 1.1962590381769923e-06,
"loss": 1.1513,
"step": 241
},
{
"epoch": 1.5454545454545454,
"grad_norm": 0.2782272696495056,
"learning_rate": 1.1622589328487505e-06,
"loss": 1.0953,
"step": 242
},
{
"epoch": 1.551948051948052,
"grad_norm": 0.2892334461212158,
"learning_rate": 1.128685346994028e-06,
"loss": 1.1721,
"step": 243
},
{
"epoch": 1.5584415584415585,
"grad_norm": 0.31845828890800476,
"learning_rate": 1.09554201191932e-06,
"loss": 1.1076,
"step": 244
},
{
"epoch": 1.564935064935065,
"grad_norm": 0.27880439162254333,
"learning_rate": 1.0628326111138377e-06,
"loss": 1.1691,
"step": 245
},
{
"epoch": 1.5714285714285714,
"grad_norm": 0.2526175081729889,
"learning_rate": 1.03056077984014e-06,
"loss": 1.1443,
"step": 246
},
{
"epoch": 1.577922077922078,
"grad_norm": 0.2798580527305603,
"learning_rate": 9.98730104730115e-07,
"loss": 1.1156,
"step": 247
},
{
"epoch": 1.5844155844155843,
"grad_norm": 0.3181169927120209,
"learning_rate": 9.673441233863661e-07,
"loss": 1.0997,
"step": 248
},
{
"epoch": 1.5909090909090908,
"grad_norm": 0.2808782756328583,
"learning_rate": 9.364063239890476e-07,
"loss": 1.1658,
"step": 249
},
{
"epoch": 1.5974025974025974,
"grad_norm": 0.2726938724517822,
"learning_rate": 9.059201449082045e-07,
"loss": 1.0976,
"step": 250
},
{
"epoch": 1.603896103896104,
"grad_norm": 0.29625144600868225,
"learning_rate": 8.758889743216247e-07,
"loss": 1.1165,
"step": 251
},
{
"epoch": 1.6103896103896105,
"grad_norm": 0.29083919525146484,
"learning_rate": 8.463161498382949e-07,
"loss": 1.158,
"step": 252
},
{
"epoch": 1.616883116883117,
"grad_norm": 0.27802902460098267,
"learning_rate": 8.172049581274571e-07,
"loss": 1.1136,
"step": 253
},
{
"epoch": 1.6233766233766234,
"grad_norm": 0.2722671627998352,
"learning_rate": 7.885586345533397e-07,
"loss": 1.1543,
"step": 254
},
{
"epoch": 1.62987012987013,
"grad_norm": 0.26996558904647827,
"learning_rate": 7.603803628155821e-07,
"loss": 1.1196,
"step": 255
},
{
"epoch": 1.6363636363636362,
"grad_norm": 0.2988632023334503,
"learning_rate": 7.326732745954001e-07,
"loss": 1.1126,
"step": 256
},
{
"epoch": 1.6428571428571428,
"grad_norm": 0.2721429467201233,
"learning_rate": 7.054404492075512e-07,
"loss": 1.1334,
"step": 257
},
{
"epoch": 1.6493506493506493,
"grad_norm": 0.2960205078125,
"learning_rate": 6.786849132580841e-07,
"loss": 1.1289,
"step": 258
},
{
"epoch": 1.655844155844156,
"grad_norm": 0.2805614173412323,
"learning_rate": 6.524096403079861e-07,
"loss": 1.1566,
"step": 259
},
{
"epoch": 1.6623376623376624,
"grad_norm": 0.26631271839141846,
"learning_rate": 6.266175505426958e-07,
"loss": 1.1385,
"step": 260
},
{
"epoch": 1.6688311688311688,
"grad_norm": 0.2959340810775757,
"learning_rate": 6.013115104475653e-07,
"loss": 1.061,
"step": 261
},
{
"epoch": 1.6753246753246753,
"grad_norm": 0.3008975088596344,
"learning_rate": 5.76494332489278e-07,
"loss": 1.1383,
"step": 262
},
{
"epoch": 1.6818181818181817,
"grad_norm": 0.28491881489753723,
"learning_rate": 5.521687748032805e-07,
"loss": 1.1827,
"step": 263
},
{
"epoch": 1.6883116883116882,
"grad_norm": 0.2625648081302643,
"learning_rate": 5.283375408872538e-07,
"loss": 1.1496,
"step": 264
},
{
"epoch": 1.6948051948051948,
"grad_norm": 0.29310908913612366,
"learning_rate": 5.05003279300637e-07,
"loss": 1.1256,
"step": 265
},
{
"epoch": 1.7012987012987013,
"grad_norm": 0.28792184591293335,
"learning_rate": 4.82168583370285e-07,
"loss": 1.1269,
"step": 266
},
{
"epoch": 1.7077922077922079,
"grad_norm": 0.3059506118297577,
"learning_rate": 4.598359909022443e-07,
"loss": 1.1275,
"step": 267
},
{
"epoch": 1.7142857142857144,
"grad_norm": 0.2944253087043762,
"learning_rate": 4.380079838997087e-07,
"loss": 1.0612,
"step": 268
},
{
"epoch": 1.7207792207792207,
"grad_norm": 0.3492392599582672,
"learning_rate": 4.1668698828716994e-07,
"loss": 1.1408,
"step": 269
},
{
"epoch": 1.7272727272727273,
"grad_norm": 0.2833608090877533,
"learning_rate": 3.958753736408105e-07,
"loss": 1.1345,
"step": 270
},
{
"epoch": 1.7337662337662336,
"grad_norm": 0.30626702308654785,
"learning_rate": 3.7557545292514987e-07,
"loss": 1.1533,
"step": 271
},
{
"epoch": 1.7402597402597402,
"grad_norm": 0.2807121276855469,
"learning_rate": 3.557894822359864e-07,
"loss": 1.1166,
"step": 272
},
{
"epoch": 1.7467532467532467,
"grad_norm": 0.31310421228408813,
"learning_rate": 3.3651966054965734e-07,
"loss": 1.1238,
"step": 273
},
{
"epoch": 1.7467532467532467,
"eval_loss": 1.153025507926941,
"eval_runtime": 25.365,
"eval_samples_per_second": 11.63,
"eval_steps_per_second": 1.459,
"step": 273
},
{
"epoch": 1.7532467532467533,
"grad_norm": 0.29958927631378174,
"learning_rate": 3.177681294786539e-07,
"loss": 1.1579,
"step": 274
},
{
"epoch": 1.7597402597402598,
"grad_norm": 0.3180530071258545,
"learning_rate": 2.995369730336012e-07,
"loss": 1.0656,
"step": 275
},
{
"epoch": 1.7662337662337664,
"grad_norm": 0.30499908328056335,
"learning_rate": 2.8182821739164534e-07,
"loss": 1.0744,
"step": 276
},
{
"epoch": 1.7727272727272727,
"grad_norm": 0.2803168296813965,
"learning_rate": 2.6464383067127175e-07,
"loss": 1.0652,
"step": 277
},
{
"epoch": 1.7792207792207793,
"grad_norm": 0.2866427004337311,
"learning_rate": 2.479857227135685e-07,
"loss": 1.1175,
"step": 278
},
{
"epoch": 1.7857142857142856,
"grad_norm": 0.26658734679222107,
"learning_rate": 2.3185574486997264e-07,
"loss": 1.1494,
"step": 279
},
{
"epoch": 1.7922077922077921,
"grad_norm": 0.3132326602935791,
"learning_rate": 2.1625568979651012e-07,
"loss": 1.0939,
"step": 280
},
{
"epoch": 1.7987012987012987,
"grad_norm": 0.29077062010765076,
"learning_rate": 2.0118729125457036e-07,
"loss": 1.1686,
"step": 281
},
{
"epoch": 1.8051948051948052,
"grad_norm": 0.2711000144481659,
"learning_rate": 1.866522239182117e-07,
"loss": 1.1246,
"step": 282
},
{
"epoch": 1.8116883116883118,
"grad_norm": 0.28556856513023376,
"learning_rate": 1.7265210318804683e-07,
"loss": 1.0459,
"step": 283
},
{
"epoch": 1.8181818181818183,
"grad_norm": 0.30168288946151733,
"learning_rate": 1.5918848501170647e-07,
"loss": 1.1087,
"step": 284
},
{
"epoch": 1.8246753246753247,
"grad_norm": 0.30205830931663513,
"learning_rate": 1.4626286571091664e-07,
"loss": 1.1333,
"step": 285
},
{
"epoch": 1.8311688311688312,
"grad_norm": 0.2765739858150482,
"learning_rate": 1.338766818151982e-07,
"loss": 1.0988,
"step": 286
},
{
"epoch": 1.8376623376623376,
"grad_norm": 0.3049326241016388,
"learning_rate": 1.2203130990221635e-07,
"loss": 1.0979,
"step": 287
},
{
"epoch": 1.844155844155844,
"grad_norm": 0.2744823694229126,
"learning_rate": 1.107280664447874e-07,
"loss": 1.1496,
"step": 288
},
{
"epoch": 1.8506493506493507,
"grad_norm": 0.2835923731327057,
"learning_rate": 9.996820766456916e-08,
"loss": 1.0894,
"step": 289
},
{
"epoch": 1.8571428571428572,
"grad_norm": 0.31562480330467224,
"learning_rate": 8.975292939244928e-08,
"loss": 1.1204,
"step": 290
},
{
"epoch": 1.8636363636363638,
"grad_norm": 0.26750344038009644,
"learning_rate": 8.008336693563823e-08,
"loss": 1.1243,
"step": 291
},
{
"epoch": 1.87012987012987,
"grad_norm": 0.26439955830574036,
"learning_rate": 7.096059495149855e-08,
"loss": 1.1027,
"step": 292
},
{
"epoch": 1.8766233766233766,
"grad_norm": 0.2618618905544281,
"learning_rate": 6.238562732810427e-08,
"loss": 1.0328,
"step": 293
},
{
"epoch": 1.883116883116883,
"grad_norm": 0.3233661949634552,
"learning_rate": 5.435941707156389e-08,
"loss": 1.1159,
"step": 294
},
{
"epoch": 1.8896103896103895,
"grad_norm": 0.28882384300231934,
"learning_rate": 4.6882856200101135e-08,
"loss": 1.1039,
"step": 295
},
{
"epoch": 1.896103896103896,
"grad_norm": 0.3103022277355194,
"learning_rate": 3.99567756449204e-08,
"loss": 1.1315,
"step": 296
},
{
"epoch": 1.9025974025974026,
"grad_norm": 0.2719733715057373,
"learning_rate": 3.358194515785784e-08,
"loss": 1.1128,
"step": 297
},
{
"epoch": 1.9090909090909092,
"grad_norm": 0.266874223947525,
"learning_rate": 2.77590732258326e-08,
"loss": 1.1333,
"step": 298
},
{
"epoch": 1.9155844155844157,
"grad_norm": 0.2862412631511688,
"learning_rate": 2.2488806992105317e-08,
"loss": 1.1387,
"step": 299
},
{
"epoch": 1.922077922077922,
"grad_norm": 0.3009204864501953,
"learning_rate": 1.7771732184357905e-08,
"loss": 1.1328,
"step": 300
},
{
"epoch": 1.9285714285714286,
"grad_norm": 0.28494757413864136,
"learning_rate": 1.3608373049596724e-08,
"loss": 1.1189,
"step": 301
},
{
"epoch": 1.935064935064935,
"grad_norm": 0.26805561780929565,
"learning_rate": 9.999192295886973e-09,
"loss": 1.0664,
"step": 302
},
{
"epoch": 1.9415584415584415,
"grad_norm": 0.3299916088581085,
"learning_rate": 6.944591040930481e-09,
"loss": 1.1117,
"step": 303
},
{
"epoch": 1.948051948051948,
"grad_norm": 0.2890622913837433,
"learning_rate": 4.444908767484712e-09,
"loss": 1.1396,
"step": 304
},
{
"epoch": 1.9545454545454546,
"grad_norm": 0.28758740425109863,
"learning_rate": 2.500423285632381e-09,
"loss": 1.1277,
"step": 305
},
{
"epoch": 1.9610389610389611,
"grad_norm": 0.2983172535896301,
"learning_rate": 1.111350701909486e-09,
"loss": 1.1452,
"step": 306
},
{
"epoch": 1.9675324675324677,
"grad_norm": 0.30814212560653687,
"learning_rate": 2.7784539528397104e-10,
"loss": 1.0888,
"step": 307
},
{
"epoch": 1.974025974025974,
"grad_norm": 0.2840464115142822,
"learning_rate": 0.0,
"loss": 1.1205,
"step": 308
}
],
"logging_steps": 1,
"max_steps": 308,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 77,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.703755799815258e+18,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}