{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 780, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.01, "grad_norm": 3.158679656326677, "learning_rate": 8.974358974358974e-08, "loss": 0.5861, "step": 1 }, { "epoch": 0.01, "grad_norm": 3.0322708666943594, "learning_rate": 1.7948717948717948e-07, "loss": 0.6075, "step": 2 }, { "epoch": 0.02, "grad_norm": 2.822065509661876, "learning_rate": 2.692307692307692e-07, "loss": 0.5896, "step": 3 }, { "epoch": 0.03, "grad_norm": 3.01990383007436, "learning_rate": 3.5897435897435896e-07, "loss": 0.6534, "step": 4 }, { "epoch": 0.03, "grad_norm": 3.1514015311739887, "learning_rate": 4.4871794871794865e-07, "loss": 0.6318, "step": 5 }, { "epoch": 0.04, "grad_norm": 3.770082837335288, "learning_rate": 5.384615384615384e-07, "loss": 0.5567, "step": 6 }, { "epoch": 0.04, "grad_norm": 3.1473432471953484, "learning_rate": 6.282051282051282e-07, "loss": 0.556, "step": 7 }, { "epoch": 0.05, "grad_norm": 2.867277473433583, "learning_rate": 7.179487179487179e-07, "loss": 0.626, "step": 8 }, { "epoch": 0.06, "grad_norm": 3.116538344911058, "learning_rate": 8.076923076923077e-07, "loss": 0.5956, "step": 9 }, { "epoch": 0.06, "grad_norm": 2.883927442084639, "learning_rate": 8.974358974358973e-07, "loss": 0.6348, "step": 10 }, { "epoch": 0.07, "grad_norm": 2.7296086958649717, "learning_rate": 9.871794871794872e-07, "loss": 0.6012, "step": 11 }, { "epoch": 0.08, "grad_norm": 2.888291481225885, "learning_rate": 1.0769230769230769e-06, "loss": 0.6158, "step": 12 }, { "epoch": 0.08, "grad_norm": 2.6798480814733354, "learning_rate": 1.1666666666666666e-06, "loss": 0.5465, "step": 13 }, { "epoch": 0.09, "grad_norm": 2.5286944008685213, "learning_rate": 1.2564102564102565e-06, "loss": 0.538, "step": 14 }, { "epoch": 0.1, "grad_norm": 2.8054433121795896, "learning_rate": 1.3461538461538462e-06, "loss": 0.5605, "step": 15 }, { "epoch": 0.1, "grad_norm": 2.301783853298887, "learning_rate": 1.4358974358974359e-06, "loss": 0.5839, "step": 16 }, { "epoch": 0.11, "grad_norm": 2.0055412267413795, "learning_rate": 1.5256410256410255e-06, "loss": 0.563, "step": 17 }, { "epoch": 0.12, "grad_norm": 2.337406004341634, "learning_rate": 1.6153846153846154e-06, "loss": 0.5564, "step": 18 }, { "epoch": 0.12, "grad_norm": 2.120974098777836, "learning_rate": 1.7051282051282051e-06, "loss": 0.5074, "step": 19 }, { "epoch": 0.13, "grad_norm": 2.2235871141473913, "learning_rate": 1.7948717948717946e-06, "loss": 0.5682, "step": 20 }, { "epoch": 0.13, "grad_norm": 2.325580776574494, "learning_rate": 1.8846153846153845e-06, "loss": 0.7112, "step": 21 }, { "epoch": 0.14, "grad_norm": 2.459952422821355, "learning_rate": 1.9743589743589744e-06, "loss": 0.6661, "step": 22 }, { "epoch": 0.15, "grad_norm": 1.95839275930406, "learning_rate": 2.064102564102564e-06, "loss": 0.4998, "step": 23 }, { "epoch": 0.15, "grad_norm": 1.974144020858437, "learning_rate": 2.1538461538461538e-06, "loss": 0.5123, "step": 24 }, { "epoch": 0.16, "grad_norm": 1.7091650511198182, "learning_rate": 2.243589743589744e-06, "loss": 0.4679, "step": 25 }, { "epoch": 0.17, "grad_norm": 1.8800999304301238, "learning_rate": 2.333333333333333e-06, "loss": 0.4695, "step": 26 }, { "epoch": 0.17, "grad_norm": 2.11180387597458, "learning_rate": 2.423076923076923e-06, "loss": 0.5301, "step": 27 }, { "epoch": 0.18, "grad_norm": 1.9925416173236044, "learning_rate": 2.512820512820513e-06, "loss": 0.5425, "step": 28 }, { "epoch": 0.19, "grad_norm": 1.9501180051573095, "learning_rate": 2.6025641025641026e-06, "loss": 0.5811, "step": 29 }, { "epoch": 0.19, "grad_norm": 1.8590839448588985, "learning_rate": 2.6923076923076923e-06, "loss": 0.5084, "step": 30 }, { "epoch": 0.2, "grad_norm": 1.9713208570956429, "learning_rate": 2.782051282051282e-06, "loss": 0.5148, "step": 31 }, { "epoch": 0.21, "grad_norm": 1.6902323535320234, "learning_rate": 2.8717948717948717e-06, "loss": 0.4844, "step": 32 }, { "epoch": 0.21, "grad_norm": 1.8356045743488203, "learning_rate": 2.9615384615384614e-06, "loss": 0.4298, "step": 33 }, { "epoch": 0.22, "grad_norm": 1.7798007609043585, "learning_rate": 3.051282051282051e-06, "loss": 0.4968, "step": 34 }, { "epoch": 0.22, "grad_norm": 2.102982471446415, "learning_rate": 3.141025641025641e-06, "loss": 0.5953, "step": 35 }, { "epoch": 0.23, "grad_norm": 2.1598493191848167, "learning_rate": 3.230769230769231e-06, "loss": 0.5369, "step": 36 }, { "epoch": 0.24, "grad_norm": 1.681878732897283, "learning_rate": 3.32051282051282e-06, "loss": 0.4529, "step": 37 }, { "epoch": 0.24, "grad_norm": 1.7703525285846058, "learning_rate": 3.4102564102564103e-06, "loss": 0.4955, "step": 38 }, { "epoch": 0.25, "grad_norm": 1.7683469960023375, "learning_rate": 3.5e-06, "loss": 0.5052, "step": 39 }, { "epoch": 0.26, "grad_norm": 1.7419535992669049, "learning_rate": 3.5897435897435892e-06, "loss": 0.5162, "step": 40 }, { "epoch": 0.26, "grad_norm": 1.7449819975072198, "learning_rate": 3.6794871794871797e-06, "loss": 0.5277, "step": 41 }, { "epoch": 0.27, "grad_norm": 1.7648482981202087, "learning_rate": 3.769230769230769e-06, "loss": 0.4876, "step": 42 }, { "epoch": 0.28, "grad_norm": 1.6946980645629792, "learning_rate": 3.858974358974359e-06, "loss": 0.441, "step": 43 }, { "epoch": 0.28, "grad_norm": 1.8314518184385051, "learning_rate": 3.948717948717949e-06, "loss": 0.4556, "step": 44 }, { "epoch": 0.29, "grad_norm": 1.9141285726499875, "learning_rate": 4.038461538461538e-06, "loss": 0.4898, "step": 45 }, { "epoch": 0.29, "grad_norm": 1.7938139036996628, "learning_rate": 4.128205128205128e-06, "loss": 0.4583, "step": 46 }, { "epoch": 0.3, "grad_norm": 1.6929510536316692, "learning_rate": 4.217948717948718e-06, "loss": 0.4323, "step": 47 }, { "epoch": 0.31, "grad_norm": 1.8092163175147657, "learning_rate": 4.3076923076923076e-06, "loss": 0.4285, "step": 48 }, { "epoch": 0.31, "grad_norm": 1.766659033823101, "learning_rate": 4.397435897435897e-06, "loss": 0.4975, "step": 49 }, { "epoch": 0.32, "grad_norm": 1.6616034395207262, "learning_rate": 4.487179487179488e-06, "loss": 0.3902, "step": 50 }, { "epoch": 0.33, "grad_norm": 1.717963883722435, "learning_rate": 4.576923076923077e-06, "loss": 0.4552, "step": 51 }, { "epoch": 0.33, "grad_norm": 2.9526480455423094, "learning_rate": 4.666666666666666e-06, "loss": 0.6078, "step": 52 }, { "epoch": 0.34, "grad_norm": 1.7249377723052146, "learning_rate": 4.756410256410257e-06, "loss": 0.4963, "step": 53 }, { "epoch": 0.35, "grad_norm": 1.6946332682174237, "learning_rate": 4.846153846153846e-06, "loss": 0.523, "step": 54 }, { "epoch": 0.35, "grad_norm": 1.7252362147444964, "learning_rate": 4.935897435897436e-06, "loss": 0.4866, "step": 55 }, { "epoch": 0.36, "grad_norm": 1.6674378519984692, "learning_rate": 5.025641025641026e-06, "loss": 0.4364, "step": 56 }, { "epoch": 0.37, "grad_norm": 1.9277904713149823, "learning_rate": 5.115384615384615e-06, "loss": 0.4977, "step": 57 }, { "epoch": 0.37, "grad_norm": 1.8110319448592889, "learning_rate": 5.205128205128205e-06, "loss": 0.4954, "step": 58 }, { "epoch": 0.38, "grad_norm": 1.7639887815910555, "learning_rate": 5.294871794871795e-06, "loss": 0.4816, "step": 59 }, { "epoch": 0.38, "grad_norm": 1.5942284653278298, "learning_rate": 5.384615384615385e-06, "loss": 0.4693, "step": 60 }, { "epoch": 0.39, "grad_norm": 1.5325212943981656, "learning_rate": 5.474358974358974e-06, "loss": 0.3896, "step": 61 }, { "epoch": 0.4, "grad_norm": 1.5026157299230636, "learning_rate": 5.564102564102564e-06, "loss": 0.4682, "step": 62 }, { "epoch": 0.4, "grad_norm": 1.4154344705468702, "learning_rate": 5.653846153846154e-06, "loss": 0.3658, "step": 63 }, { "epoch": 0.41, "grad_norm": 1.7736758802266976, "learning_rate": 5.743589743589743e-06, "loss": 0.5029, "step": 64 }, { "epoch": 0.42, "grad_norm": 1.7765092780230023, "learning_rate": 5.833333333333333e-06, "loss": 0.4608, "step": 65 }, { "epoch": 0.42, "grad_norm": 1.9404511561666338, "learning_rate": 5.923076923076923e-06, "loss": 0.4652, "step": 66 }, { "epoch": 0.43, "grad_norm": 1.753353616574486, "learning_rate": 6.0128205128205125e-06, "loss": 0.5038, "step": 67 }, { "epoch": 0.44, "grad_norm": 1.8599416931364385, "learning_rate": 6.102564102564102e-06, "loss": 0.5354, "step": 68 }, { "epoch": 0.44, "grad_norm": 1.5941444460994902, "learning_rate": 6.192307692307692e-06, "loss": 0.4343, "step": 69 }, { "epoch": 0.45, "grad_norm": 2.017933692057373, "learning_rate": 6.282051282051282e-06, "loss": 0.5443, "step": 70 }, { "epoch": 0.46, "grad_norm": 1.7072140285820805, "learning_rate": 6.371794871794871e-06, "loss": 0.4895, "step": 71 }, { "epoch": 0.46, "grad_norm": 1.5695015424217655, "learning_rate": 6.461538461538462e-06, "loss": 0.4482, "step": 72 }, { "epoch": 0.47, "grad_norm": 1.7562460888971851, "learning_rate": 6.5512820512820515e-06, "loss": 0.4835, "step": 73 }, { "epoch": 0.47, "grad_norm": 1.7316693884254104, "learning_rate": 6.64102564102564e-06, "loss": 0.4877, "step": 74 }, { "epoch": 0.48, "grad_norm": 1.6249360298584317, "learning_rate": 6.730769230769231e-06, "loss": 0.4522, "step": 75 }, { "epoch": 0.49, "grad_norm": 1.6936430758296457, "learning_rate": 6.8205128205128205e-06, "loss": 0.4546, "step": 76 }, { "epoch": 0.49, "grad_norm": 1.611309437065777, "learning_rate": 6.91025641025641e-06, "loss": 0.4106, "step": 77 }, { "epoch": 0.5, "grad_norm": 1.7957088623001527, "learning_rate": 7e-06, "loss": 0.4938, "step": 78 }, { "epoch": 0.51, "grad_norm": 1.6681604508744847, "learning_rate": 6.9999649520318915e-06, "loss": 0.4654, "step": 79 }, { "epoch": 0.51, "grad_norm": 1.6377954005820725, "learning_rate": 6.999859808829483e-06, "loss": 0.3843, "step": 80 }, { "epoch": 0.52, "grad_norm": 1.701805354673672, "learning_rate": 6.999684572498523e-06, "loss": 0.523, "step": 81 }, { "epoch": 0.53, "grad_norm": 1.8137342603270725, "learning_rate": 6.999439246548541e-06, "loss": 0.4227, "step": 82 }, { "epoch": 0.53, "grad_norm": 1.610322923292699, "learning_rate": 6.999123835892781e-06, "loss": 0.3851, "step": 83 }, { "epoch": 0.54, "grad_norm": 1.7424356339897609, "learning_rate": 6.998738346848099e-06, "loss": 0.5343, "step": 84 }, { "epoch": 0.54, "grad_norm": 1.5684238404352413, "learning_rate": 6.998282787134845e-06, "loss": 0.4003, "step": 85 }, { "epoch": 0.55, "grad_norm": 1.6444093877109798, "learning_rate": 6.997757165876698e-06, "loss": 0.5267, "step": 86 }, { "epoch": 0.56, "grad_norm": 1.658881843691408, "learning_rate": 6.9971614936004935e-06, "loss": 0.4371, "step": 87 }, { "epoch": 0.56, "grad_norm": 1.752404442163129, "learning_rate": 6.996495782236003e-06, "loss": 0.4437, "step": 88 }, { "epoch": 0.57, "grad_norm": 1.7726886879652404, "learning_rate": 6.9957600451157e-06, "loss": 0.4805, "step": 89 }, { "epoch": 0.58, "grad_norm": 1.7130722569388794, "learning_rate": 6.9949542969744955e-06, "loss": 0.4753, "step": 90 }, { "epoch": 0.58, "grad_norm": 1.66475783464997, "learning_rate": 6.9940785539494385e-06, "loss": 0.4647, "step": 91 }, { "epoch": 0.59, "grad_norm": 1.6802006865988512, "learning_rate": 6.9931328335793926e-06, "loss": 0.4353, "step": 92 }, { "epoch": 0.6, "grad_norm": 1.7344675321937646, "learning_rate": 6.992117154804688e-06, "loss": 0.4741, "step": 93 }, { "epoch": 0.6, "grad_norm": 1.683261676978249, "learning_rate": 6.991031537966741e-06, "loss": 0.4209, "step": 94 }, { "epoch": 0.61, "grad_norm": 1.8314669990022636, "learning_rate": 6.989876004807644e-06, "loss": 0.4895, "step": 95 }, { "epoch": 0.62, "grad_norm": 1.6286010700850748, "learning_rate": 6.9886505784697354e-06, "loss": 0.4316, "step": 96 }, { "epoch": 0.62, "grad_norm": 1.79287017890556, "learning_rate": 6.98735528349513e-06, "loss": 0.5164, "step": 97 }, { "epoch": 0.63, "grad_norm": 1.5957068076822012, "learning_rate": 6.985990145825233e-06, "loss": 0.4151, "step": 98 }, { "epoch": 0.63, "grad_norm": 1.7661104640217256, "learning_rate": 6.984555192800216e-06, "loss": 0.5416, "step": 99 }, { "epoch": 0.64, "grad_norm": 1.5981307636582884, "learning_rate": 6.983050453158471e-06, "loss": 0.4675, "step": 100 }, { "epoch": 0.65, "grad_norm": 1.6430144323111846, "learning_rate": 6.981475957036039e-06, "loss": 0.4333, "step": 101 }, { "epoch": 0.65, "grad_norm": 1.7167975371990496, "learning_rate": 6.979831735965997e-06, "loss": 0.3996, "step": 102 }, { "epoch": 0.66, "grad_norm": 1.8627029109499222, "learning_rate": 6.9781178228778385e-06, "loss": 0.4818, "step": 103 }, { "epoch": 0.67, "grad_norm": 1.7395258209832791, "learning_rate": 6.9763342520968e-06, "loss": 0.5318, "step": 104 }, { "epoch": 0.67, "grad_norm": 1.7354412977551585, "learning_rate": 6.974481059343188e-06, "loss": 0.4553, "step": 105 }, { "epoch": 0.68, "grad_norm": 1.7554448227091846, "learning_rate": 6.972558281731655e-06, "loss": 0.473, "step": 106 }, { "epoch": 0.69, "grad_norm": 1.8267052663331127, "learning_rate": 6.970565957770456e-06, "loss": 0.4606, "step": 107 }, { "epoch": 0.69, "grad_norm": 1.4809121128025398, "learning_rate": 6.96850412736068e-06, "loss": 0.4078, "step": 108 }, { "epoch": 0.7, "grad_norm": 1.7118223458061004, "learning_rate": 6.9663728317954505e-06, "loss": 0.4936, "step": 109 }, { "epoch": 0.71, "grad_norm": 1.7727261373643637, "learning_rate": 6.9641721137591e-06, "loss": 0.5248, "step": 110 }, { "epoch": 0.71, "grad_norm": 1.6530960350112813, "learning_rate": 6.961902017326311e-06, "loss": 0.4673, "step": 111 }, { "epoch": 0.72, "grad_norm": 1.621743811161392, "learning_rate": 6.959562587961235e-06, "loss": 0.4532, "step": 112 }, { "epoch": 0.72, "grad_norm": 1.7097624964244162, "learning_rate": 6.9571538725165855e-06, "loss": 0.4599, "step": 113 }, { "epoch": 0.73, "grad_norm": 1.6251091941091769, "learning_rate": 6.9546759192326944e-06, "loss": 0.4622, "step": 114 }, { "epoch": 0.74, "grad_norm": 1.600988114121893, "learning_rate": 6.95212877773655e-06, "loss": 0.3924, "step": 115 }, { "epoch": 0.74, "grad_norm": 1.7438877798521037, "learning_rate": 6.949512499040799e-06, "loss": 0.4436, "step": 116 }, { "epoch": 0.75, "grad_norm": 1.7329586181529388, "learning_rate": 6.946827135542729e-06, "loss": 0.4049, "step": 117 }, { "epoch": 0.76, "grad_norm": 1.4988532720543135, "learning_rate": 6.944072741023215e-06, "loss": 0.3812, "step": 118 }, { "epoch": 0.76, "grad_norm": 1.767535618217951, "learning_rate": 6.941249370645649e-06, "loss": 0.4399, "step": 119 }, { "epoch": 0.77, "grad_norm": 1.6724445288790142, "learning_rate": 6.938357080954826e-06, "loss": 0.4523, "step": 120 }, { "epoch": 0.78, "grad_norm": 1.666547625101208, "learning_rate": 6.935395929875821e-06, "loss": 0.4782, "step": 121 }, { "epoch": 0.78, "grad_norm": 1.85593796476523, "learning_rate": 6.93236597671282e-06, "loss": 0.5285, "step": 122 }, { "epoch": 0.79, "grad_norm": 1.5583837042326505, "learning_rate": 6.929267282147936e-06, "loss": 0.4111, "step": 123 }, { "epoch": 0.79, "grad_norm": 1.5489104990791143, "learning_rate": 6.9260999082400014e-06, "loss": 0.4232, "step": 124 }, { "epoch": 0.8, "grad_norm": 1.7270840700085746, "learning_rate": 6.922863918423311e-06, "loss": 0.4385, "step": 125 }, { "epoch": 0.81, "grad_norm": 1.7851688586632903, "learning_rate": 6.91955937750636e-06, "loss": 0.503, "step": 126 }, { "epoch": 0.81, "grad_norm": 1.6825892167200216, "learning_rate": 6.916186351670546e-06, "loss": 0.4429, "step": 127 }, { "epoch": 0.82, "grad_norm": 1.6005946819762804, "learning_rate": 6.912744908468841e-06, "loss": 0.428, "step": 128 }, { "epoch": 0.83, "grad_norm": 1.5740880543799294, "learning_rate": 6.909235116824441e-06, "loss": 0.4859, "step": 129 }, { "epoch": 0.83, "grad_norm": 1.5701248349607777, "learning_rate": 6.905657047029383e-06, "loss": 0.4114, "step": 130 }, { "epoch": 0.84, "grad_norm": 1.8117183546435858, "learning_rate": 6.90201077074314e-06, "loss": 0.5386, "step": 131 }, { "epoch": 0.85, "grad_norm": 1.7176967426159182, "learning_rate": 6.898296360991182e-06, "loss": 0.4988, "step": 132 }, { "epoch": 0.85, "grad_norm": 1.6119557103460176, "learning_rate": 6.894513892163519e-06, "loss": 0.4353, "step": 133 }, { "epoch": 0.86, "grad_norm": 1.8163909694175877, "learning_rate": 6.890663440013204e-06, "loss": 0.4624, "step": 134 }, { "epoch": 0.87, "grad_norm": 1.5906374611348832, "learning_rate": 6.886745081654823e-06, "loss": 0.4413, "step": 135 }, { "epoch": 0.87, "grad_norm": 1.7083476998308251, "learning_rate": 6.882758895562948e-06, "loss": 0.4798, "step": 136 }, { "epoch": 0.88, "grad_norm": 1.716310498038562, "learning_rate": 6.8787049615705635e-06, "loss": 0.4478, "step": 137 }, { "epoch": 0.88, "grad_norm": 1.7408667095266788, "learning_rate": 6.8745833608674685e-06, "loss": 0.5123, "step": 138 }, { "epoch": 0.89, "grad_norm": 1.4736521035752084, "learning_rate": 6.870394175998651e-06, "loss": 0.4126, "step": 139 }, { "epoch": 0.9, "grad_norm": 1.6728053416467819, "learning_rate": 6.866137490862636e-06, "loss": 0.479, "step": 140 }, { "epoch": 0.9, "grad_norm": 1.6162869139636344, "learning_rate": 6.861813390709803e-06, "loss": 0.4, "step": 141 }, { "epoch": 0.91, "grad_norm": 1.7100413135119708, "learning_rate": 6.857421962140681e-06, "loss": 0.4366, "step": 142 }, { "epoch": 0.92, "grad_norm": 1.5997870652975197, "learning_rate": 6.852963293104211e-06, "loss": 0.4237, "step": 143 }, { "epoch": 0.92, "grad_norm": 1.661897923842565, "learning_rate": 6.848437472895989e-06, "loss": 0.3599, "step": 144 }, { "epoch": 0.93, "grad_norm": 1.5437806414367865, "learning_rate": 6.84384459215647e-06, "loss": 0.382, "step": 145 }, { "epoch": 0.94, "grad_norm": 1.8494864927398265, "learning_rate": 6.839184742869166e-06, "loss": 0.4811, "step": 146 }, { "epoch": 0.94, "grad_norm": 1.7136464365889568, "learning_rate": 6.8344580183587866e-06, "loss": 0.4596, "step": 147 }, { "epoch": 0.95, "grad_norm": 1.5679341769016597, "learning_rate": 6.829664513289387e-06, "loss": 0.4479, "step": 148 }, { "epoch": 0.96, "grad_norm": 1.5364949608026977, "learning_rate": 6.824804323662456e-06, "loss": 0.4242, "step": 149 }, { "epoch": 0.96, "grad_norm": 1.7792768326461645, "learning_rate": 6.8198775468150085e-06, "loss": 0.5057, "step": 150 }, { "epoch": 0.97, "grad_norm": 1.64880328624309, "learning_rate": 6.814884281417627e-06, "loss": 0.4682, "step": 151 }, { "epoch": 0.97, "grad_norm": 1.6611210694948921, "learning_rate": 6.8098246274724835e-06, "loss": 0.4179, "step": 152 }, { "epoch": 0.98, "grad_norm": 1.5595830417210474, "learning_rate": 6.8046986863113455e-06, "loss": 0.3934, "step": 153 }, { "epoch": 0.99, "grad_norm": 1.5319462691982524, "learning_rate": 6.7995065605935405e-06, "loss": 0.433, "step": 154 }, { "epoch": 0.99, "grad_norm": 1.5929306277662778, "learning_rate": 6.7942483543039e-06, "loss": 0.4027, "step": 155 }, { "epoch": 1.0, "grad_norm": 1.6960544276623803, "learning_rate": 6.788924172750679e-06, "loss": 0.4456, "step": 156 }, { "epoch": 1.01, "grad_norm": 1.583925185372818, "learning_rate": 6.783534122563447e-06, "loss": 0.3919, "step": 157 }, { "epoch": 1.01, "grad_norm": 1.5898722296830252, "learning_rate": 6.7780783116909495e-06, "loss": 0.4287, "step": 158 }, { "epoch": 1.02, "grad_norm": 1.4205967061693483, "learning_rate": 6.772556849398952e-06, "loss": 0.414, "step": 159 }, { "epoch": 1.03, "grad_norm": 1.5791464079003765, "learning_rate": 6.7669698462680434e-06, "loss": 0.4554, "step": 160 }, { "epoch": 1.03, "grad_norm": 1.655295470225546, "learning_rate": 6.761317414191428e-06, "loss": 0.4268, "step": 161 }, { "epoch": 1.04, "grad_norm": 1.5529231674584578, "learning_rate": 6.755599666372685e-06, "loss": 0.3546, "step": 162 }, { "epoch": 1.04, "grad_norm": 1.6006425620300517, "learning_rate": 6.749816717323493e-06, "loss": 0.3596, "step": 163 }, { "epoch": 1.05, "grad_norm": 1.5607773236397822, "learning_rate": 6.743968682861346e-06, "loss": 0.4296, "step": 164 }, { "epoch": 1.06, "grad_norm": 1.5655082531240243, "learning_rate": 6.738055680107233e-06, "loss": 0.3895, "step": 165 }, { "epoch": 1.06, "grad_norm": 1.7066313168254195, "learning_rate": 6.7320778274832836e-06, "loss": 0.4153, "step": 166 }, { "epoch": 1.07, "grad_norm": 1.659306747460643, "learning_rate": 6.726035244710406e-06, "loss": 0.4053, "step": 167 }, { "epoch": 1.08, "grad_norm": 1.6891499067659266, "learning_rate": 6.7199280528058844e-06, "loss": 0.3971, "step": 168 }, { "epoch": 1.08, "grad_norm": 1.5705522381169368, "learning_rate": 6.713756374080959e-06, "loss": 0.3442, "step": 169 }, { "epoch": 1.09, "grad_norm": 1.560896872295485, "learning_rate": 6.70752033213837e-06, "loss": 0.3426, "step": 170 }, { "epoch": 1.1, "grad_norm": 1.698691051754592, "learning_rate": 6.7012200518698904e-06, "loss": 0.3358, "step": 171 }, { "epoch": 1.1, "grad_norm": 1.583062054049607, "learning_rate": 6.6948556594538185e-06, "loss": 0.3902, "step": 172 }, { "epoch": 1.11, "grad_norm": 1.5865690802506587, "learning_rate": 6.688427282352449e-06, "loss": 0.3747, "step": 173 }, { "epoch": 1.12, "grad_norm": 1.6540144652784194, "learning_rate": 6.681935049309533e-06, "loss": 0.358, "step": 174 }, { "epoch": 1.12, "grad_norm": 1.63701712650459, "learning_rate": 6.6753790903476814e-06, "loss": 0.3225, "step": 175 }, { "epoch": 1.13, "grad_norm": 1.7170307214130864, "learning_rate": 6.668759536765778e-06, "loss": 0.3457, "step": 176 }, { "epoch": 1.13, "grad_norm": 1.862784857769338, "learning_rate": 6.6620765211363376e-06, "loss": 0.4721, "step": 177 }, { "epoch": 1.14, "grad_norm": 2.0372982900949275, "learning_rate": 6.655330177302857e-06, "loss": 0.4379, "step": 178 }, { "epoch": 1.15, "grad_norm": 1.6108247505727127, "learning_rate": 6.64852064037713e-06, "loss": 0.3245, "step": 179 }, { "epoch": 1.15, "grad_norm": 1.5709194699574685, "learning_rate": 6.6416480467365494e-06, "loss": 0.3274, "step": 180 }, { "epoch": 1.16, "grad_norm": 1.4120453761226166, "learning_rate": 6.634712534021367e-06, "loss": 0.3123, "step": 181 }, { "epoch": 1.17, "grad_norm": 1.5271738412739237, "learning_rate": 6.627714241131943e-06, "loss": 0.2981, "step": 182 }, { "epoch": 1.17, "grad_norm": 1.657027378920605, "learning_rate": 6.62065330822596e-06, "loss": 0.3302, "step": 183 }, { "epoch": 1.18, "grad_norm": 1.6455506118501428, "learning_rate": 6.613529876715619e-06, "loss": 0.3508, "step": 184 }, { "epoch": 1.19, "grad_norm": 1.728702644768987, "learning_rate": 6.606344089264805e-06, "loss": 0.3717, "step": 185 }, { "epoch": 1.19, "grad_norm": 1.66605902530422, "learning_rate": 6.599096089786234e-06, "loss": 0.3268, "step": 186 }, { "epoch": 1.2, "grad_norm": 1.7301577213147818, "learning_rate": 6.591786023438565e-06, "loss": 0.3184, "step": 187 }, { "epoch": 1.21, "grad_norm": 1.5156585831764764, "learning_rate": 6.5844140366234956e-06, "loss": 0.3003, "step": 188 }, { "epoch": 1.21, "grad_norm": 1.5266516234302516, "learning_rate": 6.576980276982832e-06, "loss": 0.2672, "step": 189 }, { "epoch": 1.22, "grad_norm": 1.551348526011579, "learning_rate": 6.569484893395527e-06, "loss": 0.3095, "step": 190 }, { "epoch": 1.22, "grad_norm": 1.8181092851281593, "learning_rate": 6.5619280359747045e-06, "loss": 0.3643, "step": 191 }, { "epoch": 1.23, "grad_norm": 1.7302846932168594, "learning_rate": 6.55430985606465e-06, "loss": 0.3177, "step": 192 }, { "epoch": 1.24, "grad_norm": 1.587436238285827, "learning_rate": 6.546630506237778e-06, "loss": 0.273, "step": 193 }, { "epoch": 1.24, "grad_norm": 1.5878702336844899, "learning_rate": 6.538890140291578e-06, "loss": 0.2945, "step": 194 }, { "epoch": 1.25, "grad_norm": 1.6055049355849913, "learning_rate": 6.531088913245536e-06, "loss": 0.2917, "step": 195 }, { "epoch": 1.26, "grad_norm": 1.6136330635697707, "learning_rate": 6.5232269813380254e-06, "loss": 0.3031, "step": 196 }, { "epoch": 1.26, "grad_norm": 1.6772420948466469, "learning_rate": 6.5153045020231855e-06, "loss": 0.3063, "step": 197 }, { "epoch": 1.27, "grad_norm": 1.6946657606437059, "learning_rate": 6.507321633967758e-06, "loss": 0.2792, "step": 198 }, { "epoch": 1.28, "grad_norm": 1.6400336393596016, "learning_rate": 6.499278537047919e-06, "loss": 0.2533, "step": 199 }, { "epoch": 1.28, "grad_norm": 1.663092604458537, "learning_rate": 6.49117537234607e-06, "loss": 0.2457, "step": 200 }, { "epoch": 1.29, "grad_norm": 1.6920287652097301, "learning_rate": 6.483012302147617e-06, "loss": 0.2654, "step": 201 }, { "epoch": 1.29, "grad_norm": 1.5534173050131097, "learning_rate": 6.474789489937715e-06, "loss": 0.2534, "step": 202 }, { "epoch": 1.3, "grad_norm": 1.5689308391296257, "learning_rate": 6.4665071003979985e-06, "loss": 0.2247, "step": 203 }, { "epoch": 1.31, "grad_norm": 1.6586248558217869, "learning_rate": 6.4581652994032816e-06, "loss": 0.223, "step": 204 }, { "epoch": 1.31, "grad_norm": 1.5984104105693349, "learning_rate": 6.449764254018236e-06, "loss": 0.2672, "step": 205 }, { "epoch": 1.32, "grad_norm": 1.5608644731307773, "learning_rate": 6.441304132494045e-06, "loss": 0.2067, "step": 206 }, { "epoch": 1.33, "grad_norm": 1.6111918566862335, "learning_rate": 6.432785104265034e-06, "loss": 0.2327, "step": 207 }, { "epoch": 1.33, "grad_norm": 1.820458743704155, "learning_rate": 6.424207339945278e-06, "loss": 0.3055, "step": 208 }, { "epoch": 1.34, "grad_norm": 1.6194035399467694, "learning_rate": 6.415571011325181e-06, "loss": 0.2638, "step": 209 }, { "epoch": 1.35, "grad_norm": 1.8508729271576532, "learning_rate": 6.406876291368041e-06, "loss": 0.2854, "step": 210 }, { "epoch": 1.35, "grad_norm": 1.621939945199255, "learning_rate": 6.3981233542065824e-06, "loss": 0.2559, "step": 211 }, { "epoch": 1.36, "grad_norm": 1.6253283893715365, "learning_rate": 6.3893123751394695e-06, "loss": 0.2087, "step": 212 }, { "epoch": 1.37, "grad_norm": 1.6425706873157386, "learning_rate": 6.380443530627797e-06, "loss": 0.2439, "step": 213 }, { "epoch": 1.37, "grad_norm": 1.6238804132316198, "learning_rate": 6.371516998291552e-06, "loss": 0.245, "step": 214 }, { "epoch": 1.38, "grad_norm": 1.6717735651678982, "learning_rate": 6.3625329569060595e-06, "loss": 0.2439, "step": 215 }, { "epoch": 1.38, "grad_norm": 1.542780390527117, "learning_rate": 6.3534915863984045e-06, "loss": 0.2473, "step": 216 }, { "epoch": 1.39, "grad_norm": 1.4913673007510586, "learning_rate": 6.344393067843825e-06, "loss": 0.192, "step": 217 }, { "epoch": 1.4, "grad_norm": 1.5417375030490306, "learning_rate": 6.335237583462083e-06, "loss": 0.2459, "step": 218 }, { "epoch": 1.4, "grad_norm": 1.435171944665851, "learning_rate": 6.326025316613824e-06, "loss": 0.1888, "step": 219 }, { "epoch": 1.41, "grad_norm": 1.6884981117835218, "learning_rate": 6.3167564517968944e-06, "loss": 0.2381, "step": 220 }, { "epoch": 1.42, "grad_norm": 1.5331309342695914, "learning_rate": 6.307431174642653e-06, "loss": 0.2019, "step": 221 }, { "epoch": 1.42, "grad_norm": 1.6107084406380612, "learning_rate": 6.2980496719122544e-06, "loss": 0.2217, "step": 222 }, { "epoch": 1.43, "grad_norm": 1.5873058124403931, "learning_rate": 6.288612131492901e-06, "loss": 0.2413, "step": 223 }, { "epoch": 1.44, "grad_norm": 1.6440866109246133, "learning_rate": 6.279118742394089e-06, "loss": 0.2265, "step": 224 }, { "epoch": 1.44, "grad_norm": 1.4675158233520256, "learning_rate": 6.2695696947438165e-06, "loss": 0.1992, "step": 225 }, { "epoch": 1.45, "grad_norm": 1.8369618646827899, "learning_rate": 6.25996517978478e-06, "loss": 0.2351, "step": 226 }, { "epoch": 1.46, "grad_norm": 1.6374154520403525, "learning_rate": 6.2503053898705416e-06, "loss": 0.2224, "step": 227 }, { "epoch": 1.46, "grad_norm": 1.5927118853075677, "learning_rate": 6.2405905184616776e-06, "loss": 0.2155, "step": 228 }, { "epoch": 1.47, "grad_norm": 1.8012434043943897, "learning_rate": 6.230820760121904e-06, "loss": 0.2047, "step": 229 }, { "epoch": 1.47, "grad_norm": 1.7040064897533143, "learning_rate": 6.220996310514181e-06, "loss": 0.2268, "step": 230 }, { "epoch": 1.48, "grad_norm": 1.5756797325887457, "learning_rate": 6.21111736639679e-06, "loss": 0.2067, "step": 231 }, { "epoch": 1.49, "grad_norm": 1.5807736828548324, "learning_rate": 6.201184125619403e-06, "loss": 0.1972, "step": 232 }, { "epoch": 1.49, "grad_norm": 1.5794194589821984, "learning_rate": 6.191196787119104e-06, "loss": 0.1881, "step": 233 }, { "epoch": 1.5, "grad_norm": 1.5803383370350332, "learning_rate": 6.181155550916423e-06, "loss": 0.2186, "step": 234 }, { "epoch": 1.51, "grad_norm": 1.5109465274494256, "learning_rate": 6.171060618111317e-06, "loss": 0.203, "step": 235 }, { "epoch": 1.51, "grad_norm": 1.4013551973113128, "learning_rate": 6.160912190879146e-06, "loss": 0.1516, "step": 236 }, { "epoch": 1.52, "grad_norm": 1.6401705227130554, "learning_rate": 6.15071047246663e-06, "loss": 0.2405, "step": 237 }, { "epoch": 1.53, "grad_norm": 1.532406183260015, "learning_rate": 6.140455667187765e-06, "loss": 0.1584, "step": 238 }, { "epoch": 1.53, "grad_norm": 1.4242864943638422, "learning_rate": 6.13014798041975e-06, "loss": 0.1605, "step": 239 }, { "epoch": 1.54, "grad_norm": 1.658783600906005, "learning_rate": 6.119787618598854e-06, "loss": 0.251, "step": 240 }, { "epoch": 1.54, "grad_norm": 1.5412515372331743, "learning_rate": 6.109374789216296e-06, "loss": 0.1843, "step": 241 }, { "epoch": 1.55, "grad_norm": 1.5415956887491478, "learning_rate": 6.098909700814082e-06, "loss": 0.2413, "step": 242 }, { "epoch": 1.56, "grad_norm": 1.642788987391343, "learning_rate": 6.08839256298083e-06, "loss": 0.2066, "step": 243 }, { "epoch": 1.56, "grad_norm": 1.5497708571437974, "learning_rate": 6.077823586347579e-06, "loss": 0.1812, "step": 244 }, { "epoch": 1.57, "grad_norm": 1.5028500577138046, "learning_rate": 6.06720298258356e-06, "loss": 0.1936, "step": 245 }, { "epoch": 1.58, "grad_norm": 1.6048175825890045, "learning_rate": 6.056530964391961e-06, "loss": 0.2167, "step": 246 }, { "epoch": 1.58, "grad_norm": 1.5024506404579518, "learning_rate": 6.0458077455056704e-06, "loss": 0.2027, "step": 247 }, { "epoch": 1.59, "grad_norm": 1.563040859063659, "learning_rate": 6.035033540682993e-06, "loss": 0.1914, "step": 248 }, { "epoch": 1.6, "grad_norm": 1.5792427326316212, "learning_rate": 6.024208565703351e-06, "loss": 0.2111, "step": 249 }, { "epoch": 1.6, "grad_norm": 1.5265351502978457, "learning_rate": 6.013333037362959e-06, "loss": 0.1771, "step": 250 }, { "epoch": 1.61, "grad_norm": 1.6864551041061244, "learning_rate": 6.002407173470486e-06, "loss": 0.2253, "step": 251 }, { "epoch": 1.62, "grad_norm": 1.648662118810667, "learning_rate": 5.991431192842692e-06, "loss": 0.1919, "step": 252 }, { "epoch": 1.62, "grad_norm": 1.650471958543779, "learning_rate": 5.980405315300045e-06, "loss": 0.235, "step": 253 }, { "epoch": 1.63, "grad_norm": 1.6364021616200621, "learning_rate": 5.969329761662319e-06, "loss": 0.1889, "step": 254 }, { "epoch": 1.63, "grad_norm": 1.669326431103446, "learning_rate": 5.9582047537441716e-06, "loss": 0.2336, "step": 255 }, { "epoch": 1.64, "grad_norm": 1.5745390091314653, "learning_rate": 5.9470305143507e-06, "loss": 0.2231, "step": 256 }, { "epoch": 1.65, "grad_norm": 1.5373757070684195, "learning_rate": 5.9358072672729845e-06, "loss": 0.2089, "step": 257 }, { "epoch": 1.65, "grad_norm": 1.538852347237115, "learning_rate": 5.924535237283598e-06, "loss": 0.1658, "step": 258 }, { "epoch": 1.66, "grad_norm": 1.6284277440436594, "learning_rate": 5.913214650132112e-06, "loss": 0.1896, "step": 259 }, { "epoch": 1.67, "grad_norm": 1.6636487154066815, "learning_rate": 5.901845732540568e-06, "loss": 0.2591, "step": 260 }, { "epoch": 1.67, "grad_norm": 1.5490784239861026, "learning_rate": 5.8904287121989455e-06, "loss": 0.1826, "step": 261 }, { "epoch": 1.68, "grad_norm": 1.6410593482616802, "learning_rate": 5.878963817760597e-06, "loss": 0.2031, "step": 262 }, { "epoch": 1.69, "grad_norm": 1.6496652774449363, "learning_rate": 5.867451278837666e-06, "loss": 0.179, "step": 263 }, { "epoch": 1.69, "grad_norm": 1.4140543483810792, "learning_rate": 5.855891325996495e-06, "loss": 0.1951, "step": 264 }, { "epoch": 1.7, "grad_norm": 1.6277068884971182, "learning_rate": 5.8442841907530035e-06, "loss": 0.2323, "step": 265 }, { "epoch": 1.71, "grad_norm": 1.9911845501519134, "learning_rate": 5.83263010556805e-06, "loss": 0.2472, "step": 266 }, { "epoch": 1.71, "grad_norm": 1.681859260330041, "learning_rate": 5.820929303842783e-06, "loss": 0.2256, "step": 267 }, { "epoch": 1.72, "grad_norm": 1.6655852843392644, "learning_rate": 5.809182019913959e-06, "loss": 0.2081, "step": 268 }, { "epoch": 1.72, "grad_norm": 1.6025126221307116, "learning_rate": 5.797388489049253e-06, "loss": 0.2025, "step": 269 }, { "epoch": 1.73, "grad_norm": 1.554404375701696, "learning_rate": 5.785548947442547e-06, "loss": 0.2131, "step": 270 }, { "epoch": 1.74, "grad_norm": 1.4734916428150344, "learning_rate": 5.7736636322092016e-06, "loss": 0.1751, "step": 271 }, { "epoch": 1.74, "grad_norm": 1.6900522728772658, "learning_rate": 5.7617327813813e-06, "loss": 0.1833, "step": 272 }, { "epoch": 1.75, "grad_norm": 1.5488301168933416, "learning_rate": 5.749756633902887e-06, "loss": 0.1581, "step": 273 }, { "epoch": 1.76, "grad_norm": 1.537366544574601, "learning_rate": 5.7377354296251855e-06, "loss": 0.1873, "step": 274 }, { "epoch": 1.76, "grad_norm": 1.639976903485692, "learning_rate": 5.725669409301782e-06, "loss": 0.1645, "step": 275 }, { "epoch": 1.77, "grad_norm": 1.6628953838068126, "learning_rate": 5.71355881458382e-06, "loss": 0.2042, "step": 276 }, { "epoch": 1.78, "grad_norm": 1.6468632976145132, "learning_rate": 5.701403888015149e-06, "loss": 0.2166, "step": 277 }, { "epoch": 1.78, "grad_norm": 1.7952592417746278, "learning_rate": 5.689204873027471e-06, "loss": 0.2341, "step": 278 }, { "epoch": 1.79, "grad_norm": 1.5636852158640757, "learning_rate": 5.676962013935464e-06, "loss": 0.1987, "step": 279 }, { "epoch": 1.79, "grad_norm": 1.6527989174526452, "learning_rate": 5.664675555931892e-06, "loss": 0.1985, "step": 280 }, { "epoch": 1.8, "grad_norm": 1.7788761906790338, "learning_rate": 5.652345745082691e-06, "loss": 0.1919, "step": 281 }, { "epoch": 1.81, "grad_norm": 1.8065167705575866, "learning_rate": 5.639972828322043e-06, "loss": 0.2279, "step": 282 }, { "epoch": 1.81, "grad_norm": 1.5716803825130072, "learning_rate": 5.627557053447427e-06, "loss": 0.1928, "step": 283 }, { "epoch": 1.82, "grad_norm": 1.5390117314749268, "learning_rate": 5.615098669114664e-06, "loss": 0.1967, "step": 284 }, { "epoch": 1.83, "grad_norm": 1.5711837517647724, "learning_rate": 5.6025979248329265e-06, "loss": 0.2509, "step": 285 }, { "epoch": 1.83, "grad_norm": 1.4610065215464945, "learning_rate": 5.590055070959752e-06, "loss": 0.1823, "step": 286 }, { "epoch": 1.84, "grad_norm": 1.6474450019294298, "learning_rate": 5.577470358696021e-06, "loss": 0.2568, "step": 287 }, { "epoch": 1.85, "grad_norm": 1.593759652911112, "learning_rate": 5.564844040080931e-06, "loss": 0.226, "step": 288 }, { "epoch": 1.85, "grad_norm": 1.4905383734267907, "learning_rate": 5.5521763679869445e-06, "loss": 0.2009, "step": 289 }, { "epoch": 1.86, "grad_norm": 1.857438854432097, "learning_rate": 5.53946759611473e-06, "loss": 0.1911, "step": 290 }, { "epoch": 1.87, "grad_norm": 1.4583746703530132, "learning_rate": 5.526717978988076e-06, "loss": 0.2056, "step": 291 }, { "epoch": 1.87, "grad_norm": 1.5753247070423766, "learning_rate": 5.513927771948798e-06, "loss": 0.2117, "step": 292 }, { "epoch": 1.88, "grad_norm": 1.5953163221997346, "learning_rate": 5.5010972311516184e-06, "loss": 0.1946, "step": 293 }, { "epoch": 1.88, "grad_norm": 1.6835676204377235, "learning_rate": 5.488226613559045e-06, "loss": 0.239, "step": 294 }, { "epoch": 1.89, "grad_norm": 1.46235220992734, "learning_rate": 5.475316176936217e-06, "loss": 0.2106, "step": 295 }, { "epoch": 1.9, "grad_norm": 1.5871461769010167, "learning_rate": 5.462366179845746e-06, "loss": 0.2103, "step": 296 }, { "epoch": 1.9, "grad_norm": 1.6639931565192387, "learning_rate": 5.449376881642537e-06, "loss": 0.182, "step": 297 }, { "epoch": 1.91, "grad_norm": 1.57603760113916, "learning_rate": 5.436348542468598e-06, "loss": 0.186, "step": 298 }, { "epoch": 1.92, "grad_norm": 1.5868679006153603, "learning_rate": 5.423281423247821e-06, "loss": 0.1849, "step": 299 }, { "epoch": 1.92, "grad_norm": 1.5360967555172775, "learning_rate": 5.4101757856807655e-06, "loss": 0.1389, "step": 300 }, { "epoch": 1.93, "grad_norm": 1.4790770040597105, "learning_rate": 5.397031892239414e-06, "loss": 0.1666, "step": 301 }, { "epoch": 1.94, "grad_norm": 1.6761503047677448, "learning_rate": 5.383850006161913e-06, "loss": 0.1932, "step": 302 }, { "epoch": 1.94, "grad_norm": 1.6271280533728048, "learning_rate": 5.370630391447303e-06, "loss": 0.1959, "step": 303 }, { "epoch": 1.95, "grad_norm": 1.73188421274081, "learning_rate": 5.357373312850236e-06, "loss": 0.2206, "step": 304 }, { "epoch": 1.96, "grad_norm": 1.532708439171332, "learning_rate": 5.3440790358756615e-06, "loss": 0.2077, "step": 305 }, { "epoch": 1.96, "grad_norm": 1.7867777204429343, "learning_rate": 5.330747826773522e-06, "loss": 0.2253, "step": 306 }, { "epoch": 1.97, "grad_norm": 1.6343254051389262, "learning_rate": 5.317379952533411e-06, "loss": 0.2139, "step": 307 }, { "epoch": 1.97, "grad_norm": 1.6108378458429666, "learning_rate": 5.303975680879232e-06, "loss": 0.189, "step": 308 }, { "epoch": 1.98, "grad_norm": 1.5104269336113685, "learning_rate": 5.290535280263835e-06, "loss": 0.179, "step": 309 }, { "epoch": 1.99, "grad_norm": 1.5594047944760525, "learning_rate": 5.277059019863637e-06, "loss": 0.1933, "step": 310 }, { "epoch": 1.99, "grad_norm": 1.480689209423561, "learning_rate": 5.263547169573235e-06, "loss": 0.1817, "step": 311 }, { "epoch": 2.0, "grad_norm": 1.5361954565514202, "learning_rate": 5.25e-06, "loss": 0.1898, "step": 312 }, { "epoch": 2.01, "grad_norm": 1.524359102459848, "learning_rate": 5.236417782458656e-06, "loss": 0.1665, "step": 313 }, { "epoch": 2.01, "grad_norm": 1.5142652247408335, "learning_rate": 5.222800788965847e-06, "loss": 0.1968, "step": 314 }, { "epoch": 2.02, "grad_norm": 1.4018217314302317, "learning_rate": 5.2091492922346894e-06, "loss": 0.201, "step": 315 }, { "epoch": 2.03, "grad_norm": 1.4467159018792048, "learning_rate": 5.195463565669309e-06, "loss": 0.2011, "step": 316 }, { "epoch": 2.03, "grad_norm": 1.4366910804420947, "learning_rate": 5.18174388335937e-06, "loss": 0.1713, "step": 317 }, { "epoch": 2.04, "grad_norm": 1.4023140768213624, "learning_rate": 5.167990520074577e-06, "loss": 0.1413, "step": 318 }, { "epoch": 2.04, "grad_norm": 1.4524819834529294, "learning_rate": 5.154203751259183e-06, "loss": 0.1481, "step": 319 }, { "epoch": 2.05, "grad_norm": 1.8373309015708064, "learning_rate": 5.140383853026463e-06, "loss": 0.2036, "step": 320 }, { "epoch": 2.06, "grad_norm": 1.4579603843340663, "learning_rate": 5.12653110215319e-06, "loss": 0.1637, "step": 321 }, { "epoch": 2.06, "grad_norm": 1.4630253394454602, "learning_rate": 5.11264577607409e-06, "loss": 0.1679, "step": 322 }, { "epoch": 2.07, "grad_norm": 1.5010582913270236, "learning_rate": 5.098728152876287e-06, "loss": 0.1776, "step": 323 }, { "epoch": 2.08, "grad_norm": 1.6137474108058125, "learning_rate": 5.084778511293731e-06, "loss": 0.1446, "step": 324 }, { "epoch": 2.08, "grad_norm": 1.3840161878955395, "learning_rate": 5.070797130701618e-06, "loss": 0.1277, "step": 325 }, { "epoch": 2.09, "grad_norm": 1.5250613932316681, "learning_rate": 5.056784291110794e-06, "loss": 0.127, "step": 326 }, { "epoch": 2.1, "grad_norm": 1.4249393893437852, "learning_rate": 5.04274027316215e-06, "loss": 0.1077, "step": 327 }, { "epoch": 2.1, "grad_norm": 1.6057528976372746, "learning_rate": 5.028665358120995e-06, "loss": 0.1525, "step": 328 }, { "epoch": 2.11, "grad_norm": 1.5619085988004227, "learning_rate": 5.014559827871426e-06, "loss": 0.1328, "step": 329 }, { "epoch": 2.12, "grad_norm": 1.5710261957830611, "learning_rate": 5.00042396491069e-06, "loss": 0.1256, "step": 330 }, { "epoch": 2.12, "grad_norm": 1.559080830559306, "learning_rate": 4.9862580523435116e-06, "loss": 0.0978, "step": 331 }, { "epoch": 2.13, "grad_norm": 1.55462651273927, "learning_rate": 4.972062373876435e-06, "loss": 0.0938, "step": 332 }, { "epoch": 2.13, "grad_norm": 1.8403497241572058, "learning_rate": 4.95783721381214e-06, "loss": 0.157, "step": 333 }, { "epoch": 2.14, "grad_norm": 1.871192614146256, "learning_rate": 4.943582857043742e-06, "loss": 0.1324, "step": 334 }, { "epoch": 2.15, "grad_norm": 1.6434827968562997, "learning_rate": 4.9292995890490945e-06, "loss": 0.1022, "step": 335 }, { "epoch": 2.15, "grad_norm": 1.5886538225877374, "learning_rate": 4.914987695885067e-06, "loss": 0.1105, "step": 336 }, { "epoch": 2.16, "grad_norm": 1.50849954963733, "learning_rate": 4.900647464181817e-06, "loss": 0.1248, "step": 337 }, { "epoch": 2.17, "grad_norm": 1.4878345169368261, "learning_rate": 4.886279181137049e-06, "loss": 0.0967, "step": 338 }, { "epoch": 2.17, "grad_norm": 1.5621938707921055, "learning_rate": 4.871883134510263e-06, "loss": 0.1021, "step": 339 }, { "epoch": 2.18, "grad_norm": 1.6173058805664053, "learning_rate": 4.8574596126169925e-06, "loss": 0.126, "step": 340 }, { "epoch": 2.19, "grad_norm": 1.5633693529236932, "learning_rate": 4.843008904323029e-06, "loss": 0.1243, "step": 341 }, { "epoch": 2.19, "grad_norm": 1.4883031561898439, "learning_rate": 4.828531299038638e-06, "loss": 0.1101, "step": 342 }, { "epoch": 2.2, "grad_norm": 1.6093361309083989, "learning_rate": 4.81402708671276e-06, "loss": 0.0982, "step": 343 }, { "epoch": 2.21, "grad_norm": 1.5365242697494994, "learning_rate": 4.799496557827208e-06, "loss": 0.1082, "step": 344 }, { "epoch": 2.21, "grad_norm": 1.3677800601638932, "learning_rate": 4.7849400033908465e-06, "loss": 0.0984, "step": 345 }, { "epoch": 2.22, "grad_norm": 1.4697749904794721, "learning_rate": 4.770357714933765e-06, "loss": 0.1106, "step": 346 }, { "epoch": 2.22, "grad_norm": 1.515659406646426, "learning_rate": 4.755749984501437e-06, "loss": 0.1083, "step": 347 }, { "epoch": 2.23, "grad_norm": 1.5033210621791604, "learning_rate": 4.741117104648874e-06, "loss": 0.1004, "step": 348 }, { "epoch": 2.24, "grad_norm": 1.29511877140105, "learning_rate": 4.726459368434768e-06, "loss": 0.0826, "step": 349 }, { "epoch": 2.24, "grad_norm": 1.3129120468041393, "learning_rate": 4.711777069415615e-06, "loss": 0.0948, "step": 350 }, { "epoch": 2.25, "grad_norm": 1.4547928582961123, "learning_rate": 4.697070501639841e-06, "loss": 0.0891, "step": 351 }, { "epoch": 2.26, "grad_norm": 1.444163293133138, "learning_rate": 4.682339959641915e-06, "loss": 0.0902, "step": 352 }, { "epoch": 2.26, "grad_norm": 1.5803225389429953, "learning_rate": 4.667585738436448e-06, "loss": 0.0966, "step": 353 }, { "epoch": 2.27, "grad_norm": 1.6247464395630296, "learning_rate": 4.652808133512279e-06, "loss": 0.0854, "step": 354 }, { "epoch": 2.28, "grad_norm": 1.684325097074753, "learning_rate": 4.638007440826568e-06, "loss": 0.0821, "step": 355 }, { "epoch": 2.28, "grad_norm": 1.5373768221128852, "learning_rate": 4.62318395679886e-06, "loss": 0.0715, "step": 356 }, { "epoch": 2.29, "grad_norm": 1.689616257480933, "learning_rate": 4.6083379783051545e-06, "loss": 0.0877, "step": 357 }, { "epoch": 2.29, "grad_norm": 1.4679120367508793, "learning_rate": 4.593469802671951e-06, "loss": 0.0775, "step": 358 }, { "epoch": 2.3, "grad_norm": 1.3313559666874426, "learning_rate": 4.5785797276703075e-06, "loss": 0.0589, "step": 359 }, { "epoch": 2.31, "grad_norm": 1.5399713230048182, "learning_rate": 4.563668051509864e-06, "loss": 0.058, "step": 360 }, { "epoch": 2.31, "grad_norm": 1.6540628385359142, "learning_rate": 4.548735072832879e-06, "loss": 0.0824, "step": 361 }, { "epoch": 2.32, "grad_norm": 1.4382670193400522, "learning_rate": 4.533781090708244e-06, "loss": 0.0651, "step": 362 }, { "epoch": 2.33, "grad_norm": 1.3614359116885117, "learning_rate": 4.518806404625495e-06, "loss": 0.0702, "step": 363 }, { "epoch": 2.33, "grad_norm": 1.6935996818522705, "learning_rate": 4.503811314488816e-06, "loss": 0.0981, "step": 364 }, { "epoch": 2.34, "grad_norm": 1.5512615145482254, "learning_rate": 4.48879612061103e-06, "loss": 0.0805, "step": 365 }, { "epoch": 2.35, "grad_norm": 1.50599871994829, "learning_rate": 4.473761123707584e-06, "loss": 0.0937, "step": 366 }, { "epoch": 2.35, "grad_norm": 1.4324990501767214, "learning_rate": 4.458706624890534e-06, "loss": 0.0785, "step": 367 }, { "epoch": 2.36, "grad_norm": 1.4491214242293755, "learning_rate": 4.443632925662504e-06, "loss": 0.0634, "step": 368 }, { "epoch": 2.37, "grad_norm": 1.4290846682285696, "learning_rate": 4.428540327910652e-06, "loss": 0.0715, "step": 369 }, { "epoch": 2.37, "grad_norm": 1.5712380885133443, "learning_rate": 4.41342913390063e-06, "loss": 0.0809, "step": 370 }, { "epoch": 2.38, "grad_norm": 1.673544546540396, "learning_rate": 4.398299646270518e-06, "loss": 0.0767, "step": 371 }, { "epoch": 2.38, "grad_norm": 1.4244151960543419, "learning_rate": 4.3831521680247765e-06, "loss": 0.0843, "step": 372 }, { "epoch": 2.39, "grad_norm": 1.4326265228498654, "learning_rate": 4.3679870025281644e-06, "loss": 0.0607, "step": 373 }, { "epoch": 2.4, "grad_norm": 1.4882885049005354, "learning_rate": 4.352804453499677e-06, "loss": 0.0776, "step": 374 }, { "epoch": 2.4, "grad_norm": 1.4736757455586167, "learning_rate": 4.3376048250064525e-06, "loss": 0.0657, "step": 375 }, { "epoch": 2.41, "grad_norm": 1.574256990342458, "learning_rate": 4.322388421457687e-06, "loss": 0.082, "step": 376 }, { "epoch": 2.42, "grad_norm": 1.611911060390727, "learning_rate": 4.30715554759854e-06, "loss": 0.0695, "step": 377 }, { "epoch": 2.42, "grad_norm": 1.5323569330355356, "learning_rate": 4.2919065085040285e-06, "loss": 0.0736, "step": 378 }, { "epoch": 2.43, "grad_norm": 1.4158459967105763, "learning_rate": 4.276641609572911e-06, "loss": 0.0774, "step": 379 }, { "epoch": 2.44, "grad_norm": 1.411415928220124, "learning_rate": 4.261361156521586e-06, "loss": 0.0609, "step": 380 }, { "epoch": 2.44, "grad_norm": 1.3129410201180565, "learning_rate": 4.246065455377956e-06, "loss": 0.066, "step": 381 }, { "epoch": 2.45, "grad_norm": 1.7101784619118496, "learning_rate": 4.230754812475306e-06, "loss": 0.0773, "step": 382 }, { "epoch": 2.46, "grad_norm": 1.386968561190813, "learning_rate": 4.215429534446161e-06, "loss": 0.0663, "step": 383 }, { "epoch": 2.46, "grad_norm": 1.402321547197101, "learning_rate": 4.200089928216156e-06, "loss": 0.0703, "step": 384 }, { "epoch": 2.47, "grad_norm": 1.5024096069853687, "learning_rate": 4.1847363009978776e-06, "loss": 0.0666, "step": 385 }, { "epoch": 2.47, "grad_norm": 1.5771691557161291, "learning_rate": 4.169368960284718e-06, "loss": 0.0752, "step": 386 }, { "epoch": 2.48, "grad_norm": 1.5004612326192561, "learning_rate": 4.153988213844717e-06, "loss": 0.0659, "step": 387 }, { "epoch": 2.49, "grad_norm": 1.5934299696891325, "learning_rate": 4.138594369714394e-06, "loss": 0.0649, "step": 388 }, { "epoch": 2.49, "grad_norm": 1.531764702409795, "learning_rate": 4.123187736192583e-06, "loss": 0.0598, "step": 389 }, { "epoch": 2.5, "grad_norm": 1.5143115245379815, "learning_rate": 4.107768621834257e-06, "loss": 0.0645, "step": 390 }, { "epoch": 2.51, "grad_norm": 1.4197919164126784, "learning_rate": 4.092337335444343e-06, "loss": 0.0579, "step": 391 }, { "epoch": 2.51, "grad_norm": 1.2419987125261145, "learning_rate": 4.076894186071548e-06, "loss": 0.0485, "step": 392 }, { "epoch": 2.52, "grad_norm": 1.6270989771595925, "learning_rate": 4.061439483002161e-06, "loss": 0.0752, "step": 393 }, { "epoch": 2.53, "grad_norm": 1.6668781562334298, "learning_rate": 4.045973535753863e-06, "loss": 0.0506, "step": 394 }, { "epoch": 2.53, "grad_norm": 1.4111176343177239, "learning_rate": 4.030496654069524e-06, "loss": 0.0531, "step": 395 }, { "epoch": 2.54, "grad_norm": 1.551403757100443, "learning_rate": 4.015009147911007e-06, "loss": 0.0829, "step": 396 }, { "epoch": 2.54, "grad_norm": 1.5266748331664532, "learning_rate": 3.9995113274529506e-06, "loss": 0.0579, "step": 397 }, { "epoch": 2.55, "grad_norm": 1.2952363610150155, "learning_rate": 3.984003503076566e-06, "loss": 0.075, "step": 398 }, { "epoch": 2.56, "grad_norm": 1.4737121699978741, "learning_rate": 3.968485985363416e-06, "loss": 0.0667, "step": 399 }, { "epoch": 2.56, "grad_norm": 1.3510918382261339, "learning_rate": 3.952959085089193e-06, "loss": 0.0506, "step": 400 }, { "epoch": 2.57, "grad_norm": 1.2593798377497472, "learning_rate": 3.937423113217505e-06, "loss": 0.0586, "step": 401 }, { "epoch": 2.58, "grad_norm": 1.3649343014206317, "learning_rate": 3.92187838089363e-06, "loss": 0.0722, "step": 402 }, { "epoch": 2.58, "grad_norm": 1.3123387897850538, "learning_rate": 3.9063251994383055e-06, "loss": 0.0681, "step": 403 }, { "epoch": 2.59, "grad_norm": 1.3958865510693308, "learning_rate": 3.8907638803414774e-06, "loss": 0.0641, "step": 404 }, { "epoch": 2.6, "grad_norm": 1.3955726113701625, "learning_rate": 3.875194735256067e-06, "loss": 0.0692, "step": 405 }, { "epoch": 2.6, "grad_norm": 1.4482062071185025, "learning_rate": 3.859618075991735e-06, "loss": 0.0612, "step": 406 }, { "epoch": 2.61, "grad_norm": 1.599652437336555, "learning_rate": 3.844034214508625e-06, "loss": 0.0819, "step": 407 }, { "epoch": 2.62, "grad_norm": 1.4310645914637838, "learning_rate": 3.828443462911128e-06, "loss": 0.063, "step": 408 }, { "epoch": 2.62, "grad_norm": 1.4893479185732388, "learning_rate": 3.8128461334416223e-06, "loss": 0.0762, "step": 409 }, { "epoch": 2.63, "grad_norm": 1.3002764419943207, "learning_rate": 3.7972425384742264e-06, "loss": 0.0595, "step": 410 }, { "epoch": 2.63, "grad_norm": 1.5160157250102182, "learning_rate": 3.781632990508541e-06, "loss": 0.0762, "step": 411 }, { "epoch": 2.64, "grad_norm": 1.3825833601713984, "learning_rate": 3.766017802163386e-06, "loss": 0.0672, "step": 412 }, { "epoch": 2.65, "grad_norm": 1.3919309630976415, "learning_rate": 3.7503972861705478e-06, "loss": 0.0711, "step": 413 }, { "epoch": 2.65, "grad_norm": 1.3171909579170673, "learning_rate": 3.7347717553685084e-06, "loss": 0.0492, "step": 414 }, { "epoch": 2.66, "grad_norm": 1.4882892958051253, "learning_rate": 3.7191415226961867e-06, "loss": 0.0578, "step": 415 }, { "epoch": 2.67, "grad_norm": 1.5208283595273544, "learning_rate": 3.703506901186665e-06, "loss": 0.0858, "step": 416 }, { "epoch": 2.67, "grad_norm": 1.255493106442096, "learning_rate": 3.6878682039609253e-06, "loss": 0.0467, "step": 417 }, { "epoch": 2.68, "grad_norm": 1.3608611259131593, "learning_rate": 3.6722257442215736e-06, "loss": 0.0601, "step": 418 }, { "epoch": 2.69, "grad_norm": 1.2991974178801189, "learning_rate": 3.6565798352465697e-06, "loss": 0.0471, "step": 419 }, { "epoch": 2.69, "grad_norm": 1.2001791330016007, "learning_rate": 3.640930790382953e-06, "loss": 0.0601, "step": 420 }, { "epoch": 2.7, "grad_norm": 1.4083800701711358, "learning_rate": 3.625278923040567e-06, "loss": 0.0764, "step": 421 }, { "epoch": 2.71, "grad_norm": 1.4598166249307498, "learning_rate": 3.6096245466857808e-06, "loss": 0.0825, "step": 422 }, { "epoch": 2.71, "grad_norm": 1.5075568056733875, "learning_rate": 3.5939679748352146e-06, "loss": 0.0773, "step": 423 }, { "epoch": 2.72, "grad_norm": 1.6813722670975189, "learning_rate": 3.578309521049456e-06, "loss": 0.0678, "step": 424 }, { "epoch": 2.72, "grad_norm": 1.4325596151868998, "learning_rate": 3.562649498926785e-06, "loss": 0.0628, "step": 425 }, { "epoch": 2.73, "grad_norm": 1.485710743290764, "learning_rate": 3.546988222096891e-06, "loss": 0.0682, "step": 426 }, { "epoch": 2.74, "grad_norm": 1.4464307460259824, "learning_rate": 3.531326004214592e-06, "loss": 0.0546, "step": 427 }, { "epoch": 2.74, "grad_norm": 1.607327721588889, "learning_rate": 3.515663158953552e-06, "loss": 0.0631, "step": 428 }, { "epoch": 2.75, "grad_norm": 1.5697673191081656, "learning_rate": 3.5e-06, "loss": 0.0522, "step": 429 }, { "epoch": 2.76, "grad_norm": 1.3868779096239099, "learning_rate": 3.484336841046448e-06, "loss": 0.0651, "step": 430 }, { "epoch": 2.76, "grad_norm": 1.3231395905165282, "learning_rate": 3.468673995785409e-06, "loss": 0.0481, "step": 431 }, { "epoch": 2.77, "grad_norm": 1.3593013701575807, "learning_rate": 3.4530117779031096e-06, "loss": 0.0619, "step": 432 }, { "epoch": 2.78, "grad_norm": 1.3228492850140474, "learning_rate": 3.4373505010732152e-06, "loss": 0.0675, "step": 433 }, { "epoch": 2.78, "grad_norm": 1.5367657186057377, "learning_rate": 3.4216904789505444e-06, "loss": 0.0648, "step": 434 }, { "epoch": 2.79, "grad_norm": 1.3753764939376347, "learning_rate": 3.4060320251647866e-06, "loss": 0.0623, "step": 435 }, { "epoch": 2.79, "grad_norm": 1.2778004930662803, "learning_rate": 3.3903754533142195e-06, "loss": 0.0641, "step": 436 }, { "epoch": 2.8, "grad_norm": 1.4635871262219515, "learning_rate": 3.374721076959433e-06, "loss": 0.0563, "step": 437 }, { "epoch": 2.81, "grad_norm": 1.437055332759444, "learning_rate": 3.359069209617048e-06, "loss": 0.0745, "step": 438 }, { "epoch": 2.81, "grad_norm": 1.339007375027239, "learning_rate": 3.3434201647534306e-06, "loss": 0.055, "step": 439 }, { "epoch": 2.82, "grad_norm": 1.301365126421519, "learning_rate": 3.3277742557784263e-06, "loss": 0.0598, "step": 440 }, { "epoch": 2.83, "grad_norm": 1.4760103555439992, "learning_rate": 3.312131796039074e-06, "loss": 0.0892, "step": 441 }, { "epoch": 2.83, "grad_norm": 1.2237525507670013, "learning_rate": 3.296493098813335e-06, "loss": 0.0542, "step": 442 }, { "epoch": 2.84, "grad_norm": 1.5222822001085714, "learning_rate": 3.280858477303813e-06, "loss": 0.0868, "step": 443 }, { "epoch": 2.85, "grad_norm": 1.39320313785298, "learning_rate": 3.265228244631491e-06, "loss": 0.0736, "step": 444 }, { "epoch": 2.85, "grad_norm": 1.3573613063742243, "learning_rate": 3.2496027138294534e-06, "loss": 0.0628, "step": 445 }, { "epoch": 2.86, "grad_norm": 1.4654821857892797, "learning_rate": 3.2339821978366144e-06, "loss": 0.0563, "step": 446 }, { "epoch": 2.87, "grad_norm": 1.4095736477135694, "learning_rate": 3.2183670094914596e-06, "loss": 0.079, "step": 447 }, { "epoch": 2.87, "grad_norm": 1.2720604821604127, "learning_rate": 3.2027574615257726e-06, "loss": 0.0593, "step": 448 }, { "epoch": 2.88, "grad_norm": 1.4181511906965734, "learning_rate": 3.1871538665583784e-06, "loss": 0.0657, "step": 449 }, { "epoch": 2.88, "grad_norm": 1.4715832293023445, "learning_rate": 3.171556537088873e-06, "loss": 0.0841, "step": 450 }, { "epoch": 2.89, "grad_norm": 1.3651848095300287, "learning_rate": 3.155965785491375e-06, "loss": 0.0763, "step": 451 }, { "epoch": 2.9, "grad_norm": 1.2733406993933158, "learning_rate": 3.140381924008266e-06, "loss": 0.0642, "step": 452 }, { "epoch": 2.9, "grad_norm": 1.4649815014884033, "learning_rate": 3.1248052647439327e-06, "loss": 0.0571, "step": 453 }, { "epoch": 2.91, "grad_norm": 1.2811384885805435, "learning_rate": 3.109236119658523e-06, "loss": 0.0546, "step": 454 }, { "epoch": 2.92, "grad_norm": 1.2820286884512915, "learning_rate": 3.0936748005616936e-06, "loss": 0.0572, "step": 455 }, { "epoch": 2.92, "grad_norm": 1.2499223607002325, "learning_rate": 3.0781216191063695e-06, "loss": 0.0413, "step": 456 }, { "epoch": 2.93, "grad_norm": 1.2767469475357223, "learning_rate": 3.0625768867824957e-06, "loss": 0.0509, "step": 457 }, { "epoch": 2.94, "grad_norm": 1.4259647549315762, "learning_rate": 3.047040914910806e-06, "loss": 0.0586, "step": 458 }, { "epoch": 2.94, "grad_norm": 1.3134194512796191, "learning_rate": 3.0315140146365854e-06, "loss": 0.0538, "step": 459 }, { "epoch": 2.95, "grad_norm": 1.8805869988982469, "learning_rate": 3.015996496923435e-06, "loss": 0.1031, "step": 460 }, { "epoch": 2.96, "grad_norm": 1.3905086869403662, "learning_rate": 3.00048867254705e-06, "loss": 0.072, "step": 461 }, { "epoch": 2.96, "grad_norm": 1.5495009008005376, "learning_rate": 2.9849908520889936e-06, "loss": 0.0724, "step": 462 }, { "epoch": 2.97, "grad_norm": 1.4887450497609702, "learning_rate": 2.9695033459304766e-06, "loss": 0.072, "step": 463 }, { "epoch": 2.97, "grad_norm": 1.4706761990324224, "learning_rate": 2.954026464246138e-06, "loss": 0.0586, "step": 464 }, { "epoch": 2.98, "grad_norm": 1.3362836710115724, "learning_rate": 2.9385605169978387e-06, "loss": 0.0537, "step": 465 }, { "epoch": 2.99, "grad_norm": 1.3505739518905349, "learning_rate": 2.923105813928453e-06, "loss": 0.0538, "step": 466 }, { "epoch": 2.99, "grad_norm": 1.2122588192580181, "learning_rate": 2.907662664555658e-06, "loss": 0.0487, "step": 467 }, { "epoch": 3.0, "grad_norm": 1.4238620533550892, "learning_rate": 2.8922313781657437e-06, "loss": 0.0552, "step": 468 }, { "epoch": 3.01, "grad_norm": 1.1892063414140128, "learning_rate": 2.876812263807417e-06, "loss": 0.0444, "step": 469 }, { "epoch": 3.01, "grad_norm": 1.476004528679644, "learning_rate": 2.861405630285606e-06, "loss": 0.0661, "step": 470 }, { "epoch": 3.02, "grad_norm": 1.2283441956799137, "learning_rate": 2.8460117861552833e-06, "loss": 0.0685, "step": 471 }, { "epoch": 3.03, "grad_norm": 1.334586988917645, "learning_rate": 2.8306310397152817e-06, "loss": 0.067, "step": 472 }, { "epoch": 3.03, "grad_norm": 1.2195524245988179, "learning_rate": 2.815263699002124e-06, "loss": 0.0474, "step": 473 }, { "epoch": 3.04, "grad_norm": 1.2153155028002154, "learning_rate": 2.799910071783845e-06, "loss": 0.0411, "step": 474 }, { "epoch": 3.04, "grad_norm": 1.2126762221994973, "learning_rate": 2.7845704655538383e-06, "loss": 0.0449, "step": 475 }, { "epoch": 3.05, "grad_norm": 1.9487314192826695, "learning_rate": 2.7692451875246956e-06, "loss": 0.0699, "step": 476 }, { "epoch": 3.06, "grad_norm": 1.2517341101809802, "learning_rate": 2.7539345446220444e-06, "loss": 0.0501, "step": 477 }, { "epoch": 3.06, "grad_norm": 1.1687876365657441, "learning_rate": 2.7386388434784143e-06, "loss": 0.0542, "step": 478 }, { "epoch": 3.07, "grad_norm": 1.2867517360302896, "learning_rate": 2.723358390427089e-06, "loss": 0.0594, "step": 479 }, { "epoch": 3.08, "grad_norm": 1.2340318913389399, "learning_rate": 2.708093491495973e-06, "loss": 0.0448, "step": 480 }, { "epoch": 3.08, "grad_norm": 1.027195991965926, "learning_rate": 2.6928444524014595e-06, "loss": 0.0369, "step": 481 }, { "epoch": 3.09, "grad_norm": 1.1529153587540384, "learning_rate": 2.6776115785423123e-06, "loss": 0.0366, "step": 482 }, { "epoch": 3.1, "grad_norm": 1.0936389016778483, "learning_rate": 2.6623951749935487e-06, "loss": 0.0318, "step": 483 }, { "epoch": 3.1, "grad_norm": 1.2450006020964501, "learning_rate": 2.6471955465003237e-06, "loss": 0.0495, "step": 484 }, { "epoch": 3.11, "grad_norm": 1.2226035789766154, "learning_rate": 2.6320129974718355e-06, "loss": 0.0445, "step": 485 }, { "epoch": 3.12, "grad_norm": 1.25075066205537, "learning_rate": 2.616847831975224e-06, "loss": 0.0384, "step": 486 }, { "epoch": 3.12, "grad_norm": 1.2463934852518022, "learning_rate": 2.601700353729481e-06, "loss": 0.03, "step": 487 }, { "epoch": 3.13, "grad_norm": 1.2325826640793016, "learning_rate": 2.58657086609937e-06, "loss": 0.0301, "step": 488 }, { "epoch": 3.13, "grad_norm": 1.3592627844900937, "learning_rate": 2.5714596720893473e-06, "loss": 0.0556, "step": 489 }, { "epoch": 3.14, "grad_norm": 1.2447561831293539, "learning_rate": 2.5563670743374973e-06, "loss": 0.0458, "step": 490 }, { "epoch": 3.15, "grad_norm": 1.2857400436948982, "learning_rate": 2.5412933751094662e-06, "loss": 0.0334, "step": 491 }, { "epoch": 3.15, "grad_norm": 1.2532170542444407, "learning_rate": 2.5262388762924157e-06, "loss": 0.0344, "step": 492 }, { "epoch": 3.16, "grad_norm": 1.203530717357317, "learning_rate": 2.5112038793889706e-06, "loss": 0.0421, "step": 493 }, { "epoch": 3.17, "grad_norm": 1.3294042133808428, "learning_rate": 2.496188685511185e-06, "loss": 0.0321, "step": 494 }, { "epoch": 3.17, "grad_norm": 1.2370090074937525, "learning_rate": 2.481193595374505e-06, "loss": 0.0303, "step": 495 }, { "epoch": 3.18, "grad_norm": 1.2854698199833365, "learning_rate": 2.4662189092917563e-06, "loss": 0.0437, "step": 496 }, { "epoch": 3.19, "grad_norm": 1.258966876492369, "learning_rate": 2.4512649271671214e-06, "loss": 0.0348, "step": 497 }, { "epoch": 3.19, "grad_norm": 1.3543020541420905, "learning_rate": 2.436331948490136e-06, "loss": 0.0381, "step": 498 }, { "epoch": 3.2, "grad_norm": 1.524191201432137, "learning_rate": 2.4214202723296924e-06, "loss": 0.0349, "step": 499 }, { "epoch": 3.21, "grad_norm": 1.2112359861077326, "learning_rate": 2.4065301973280486e-06, "loss": 0.0383, "step": 500 }, { "epoch": 3.21, "grad_norm": 1.1519148621388768, "learning_rate": 2.391662021694847e-06, "loss": 0.0321, "step": 501 }, { "epoch": 3.22, "grad_norm": 1.3479943611058076, "learning_rate": 2.3768160432011395e-06, "loss": 0.0437, "step": 502 }, { "epoch": 3.22, "grad_norm": 1.5161099076368707, "learning_rate": 2.3619925591734323e-06, "loss": 0.0319, "step": 503 }, { "epoch": 3.23, "grad_norm": 1.2636677740958409, "learning_rate": 2.3471918664877217e-06, "loss": 0.0332, "step": 504 }, { "epoch": 3.24, "grad_norm": 1.144672254581016, "learning_rate": 2.332414261563553e-06, "loss": 0.0321, "step": 505 }, { "epoch": 3.24, "grad_norm": 0.9299371323757989, "learning_rate": 2.317660040358085e-06, "loss": 0.032, "step": 506 }, { "epoch": 3.25, "grad_norm": 1.1313131663449125, "learning_rate": 2.3029294983601598e-06, "loss": 0.0313, "step": 507 }, { "epoch": 3.26, "grad_norm": 1.0138929678425856, "learning_rate": 2.2882229305843866e-06, "loss": 0.0307, "step": 508 }, { "epoch": 3.26, "grad_norm": 1.1199838222245246, "learning_rate": 2.2735406315652323e-06, "loss": 0.0334, "step": 509 }, { "epoch": 3.27, "grad_norm": 1.1253274246781595, "learning_rate": 2.258882895351125e-06, "loss": 0.0291, "step": 510 }, { "epoch": 3.28, "grad_norm": 1.0864211988064494, "learning_rate": 2.2442500154985643e-06, "loss": 0.0283, "step": 511 }, { "epoch": 3.28, "grad_norm": 1.154223930393599, "learning_rate": 2.229642285066236e-06, "loss": 0.0286, "step": 512 }, { "epoch": 3.29, "grad_norm": 1.1685223712205286, "learning_rate": 2.215059996609154e-06, "loss": 0.0391, "step": 513 }, { "epoch": 3.29, "grad_norm": 1.075832341378613, "learning_rate": 2.200503442172792e-06, "loss": 0.0304, "step": 514 }, { "epoch": 3.3, "grad_norm": 0.9059326924484261, "learning_rate": 2.185972913287241e-06, "loss": 0.0189, "step": 515 }, { "epoch": 3.31, "grad_norm": 1.0918674140346547, "learning_rate": 2.1714687009613628e-06, "loss": 0.0191, "step": 516 }, { "epoch": 3.31, "grad_norm": 1.1821702256912072, "learning_rate": 2.156991095676971e-06, "loss": 0.0362, "step": 517 }, { "epoch": 3.32, "grad_norm": 0.97509557211284, "learning_rate": 2.1425403873830083e-06, "loss": 0.0225, "step": 518 }, { "epoch": 3.33, "grad_norm": 0.9490638193518203, "learning_rate": 2.1281168654897376e-06, "loss": 0.0246, "step": 519 }, { "epoch": 3.33, "grad_norm": 1.2487827485548177, "learning_rate": 2.113720818862951e-06, "loss": 0.0355, "step": 520 }, { "epoch": 3.34, "grad_norm": 1.1256405423581581, "learning_rate": 2.099352535818182e-06, "loss": 0.0285, "step": 521 }, { "epoch": 3.35, "grad_norm": 0.9095020425799777, "learning_rate": 2.085012304114933e-06, "loss": 0.029, "step": 522 }, { "epoch": 3.35, "grad_norm": 1.1211507222049146, "learning_rate": 2.070700410950906e-06, "loss": 0.0295, "step": 523 }, { "epoch": 3.36, "grad_norm": 1.3066574168869387, "learning_rate": 2.0564171429562587e-06, "loss": 0.0253, "step": 524 }, { "epoch": 3.37, "grad_norm": 0.9756346110567761, "learning_rate": 2.042162786187862e-06, "loss": 0.0247, "step": 525 }, { "epoch": 3.37, "grad_norm": 1.1674473433223178, "learning_rate": 2.027937626123565e-06, "loss": 0.0309, "step": 526 }, { "epoch": 3.38, "grad_norm": 1.125722578814156, "learning_rate": 2.0137419476564896e-06, "loss": 0.0257, "step": 527 }, { "epoch": 3.38, "grad_norm": 0.9997980698465747, "learning_rate": 1.9995760350893098e-06, "loss": 0.0279, "step": 528 }, { "epoch": 3.39, "grad_norm": 0.9080372864182613, "learning_rate": 1.985440172128573e-06, "loss": 0.0197, "step": 529 }, { "epoch": 3.4, "grad_norm": 1.0431027439667329, "learning_rate": 1.9713346418790058e-06, "loss": 0.0277, "step": 530 }, { "epoch": 3.4, "grad_norm": 1.0520465883120045, "learning_rate": 1.957259726837849e-06, "loss": 0.0221, "step": 531 }, { "epoch": 3.41, "grad_norm": 1.1131126853324889, "learning_rate": 1.9432157088892064e-06, "loss": 0.0263, "step": 532 }, { "epoch": 3.42, "grad_norm": 1.0287024420083004, "learning_rate": 1.9292028692983824e-06, "loss": 0.025, "step": 533 }, { "epoch": 3.42, "grad_norm": 1.0138678830635144, "learning_rate": 1.91522148870627e-06, "loss": 0.0222, "step": 534 }, { "epoch": 3.43, "grad_norm": 0.9566591334023489, "learning_rate": 1.9012718471237144e-06, "loss": 0.0283, "step": 535 }, { "epoch": 3.44, "grad_norm": 1.0842644448648473, "learning_rate": 1.887354223925911e-06, "loss": 0.0214, "step": 536 }, { "epoch": 3.44, "grad_norm": 0.975574990683092, "learning_rate": 1.87346889784681e-06, "loss": 0.0246, "step": 537 }, { "epoch": 3.45, "grad_norm": 1.1478718825084286, "learning_rate": 1.8596161469735374e-06, "loss": 0.028, "step": 538 }, { "epoch": 3.46, "grad_norm": 1.0607801924078737, "learning_rate": 1.8457962487408175e-06, "loss": 0.0225, "step": 539 }, { "epoch": 3.46, "grad_norm": 0.9032740986095086, "learning_rate": 1.8320094799254222e-06, "loss": 0.0241, "step": 540 }, { "epoch": 3.47, "grad_norm": 0.9873082880576691, "learning_rate": 1.8182561166406308e-06, "loss": 0.0233, "step": 541 }, { "epoch": 3.47, "grad_norm": 1.036096577241665, "learning_rate": 1.8045364343306915e-06, "loss": 0.0225, "step": 542 }, { "epoch": 3.48, "grad_norm": 1.011020516763835, "learning_rate": 1.7908507077653124e-06, "loss": 0.0231, "step": 543 }, { "epoch": 3.49, "grad_norm": 0.9671254615577292, "learning_rate": 1.7771992110341533e-06, "loss": 0.0224, "step": 544 }, { "epoch": 3.49, "grad_norm": 0.9590953302318, "learning_rate": 1.7635822175413446e-06, "loss": 0.0202, "step": 545 }, { "epoch": 3.5, "grad_norm": 0.9693202672140285, "learning_rate": 1.7500000000000008e-06, "loss": 0.0207, "step": 546 }, { "epoch": 3.51, "grad_norm": 1.1439987999618473, "learning_rate": 1.7364528304267646e-06, "loss": 0.0255, "step": 547 }, { "epoch": 3.51, "grad_norm": 0.9251345782373769, "learning_rate": 1.7229409801363635e-06, "loss": 0.0179, "step": 548 }, { "epoch": 3.52, "grad_norm": 0.9771154935646014, "learning_rate": 1.7094647197361656e-06, "loss": 0.0255, "step": 549 }, { "epoch": 3.53, "grad_norm": 1.2798180991649053, "learning_rate": 1.6960243191207686e-06, "loss": 0.0232, "step": 550 }, { "epoch": 3.53, "grad_norm": 1.0317246192313414, "learning_rate": 1.6826200474665891e-06, "loss": 0.0221, "step": 551 }, { "epoch": 3.54, "grad_norm": 1.088333248492394, "learning_rate": 1.669252173226479e-06, "loss": 0.0282, "step": 552 }, { "epoch": 3.54, "grad_norm": 1.0599165259760654, "learning_rate": 1.6559209641243388e-06, "loss": 0.0213, "step": 553 }, { "epoch": 3.55, "grad_norm": 0.8840390529269968, "learning_rate": 1.642626687149765e-06, "loss": 0.0266, "step": 554 }, { "epoch": 3.56, "grad_norm": 1.022733278705135, "learning_rate": 1.629369608552696e-06, "loss": 0.0189, "step": 555 }, { "epoch": 3.56, "grad_norm": 0.9370338460268343, "learning_rate": 1.6161499938380873e-06, "loss": 0.0165, "step": 556 }, { "epoch": 3.57, "grad_norm": 0.8403429260662558, "learning_rate": 1.6029681077605864e-06, "loss": 0.0194, "step": 557 }, { "epoch": 3.58, "grad_norm": 1.0755284457395846, "learning_rate": 1.5898242143192336e-06, "loss": 0.0237, "step": 558 }, { "epoch": 3.58, "grad_norm": 0.9907763333997255, "learning_rate": 1.576718576752179e-06, "loss": 0.0211, "step": 559 }, { "epoch": 3.59, "grad_norm": 0.888869735944081, "learning_rate": 1.5636514575314024e-06, "loss": 0.0179, "step": 560 }, { "epoch": 3.6, "grad_norm": 0.885718782686732, "learning_rate": 1.550623118357463e-06, "loss": 0.0201, "step": 561 }, { "epoch": 3.6, "grad_norm": 1.0264516969794926, "learning_rate": 1.5376338201542538e-06, "loss": 0.0194, "step": 562 }, { "epoch": 3.61, "grad_norm": 0.9766407203376387, "learning_rate": 1.5246838230637831e-06, "loss": 0.0221, "step": 563 }, { "epoch": 3.62, "grad_norm": 0.725003276513177, "learning_rate": 1.511773386440955e-06, "loss": 0.0149, "step": 564 }, { "epoch": 3.62, "grad_norm": 0.8499489911772146, "learning_rate": 1.4989027688483808e-06, "loss": 0.0224, "step": 565 }, { "epoch": 3.63, "grad_norm": 0.8758345450457246, "learning_rate": 1.4860722280512022e-06, "loss": 0.0186, "step": 566 }, { "epoch": 3.63, "grad_norm": 1.0827335434130656, "learning_rate": 1.473282021011924e-06, "loss": 0.0259, "step": 567 }, { "epoch": 3.64, "grad_norm": 0.997864768606809, "learning_rate": 1.4605324038852707e-06, "loss": 0.0196, "step": 568 }, { "epoch": 3.65, "grad_norm": 0.906378937725106, "learning_rate": 1.4478236320130554e-06, "loss": 0.0209, "step": 569 }, { "epoch": 3.65, "grad_norm": 0.7169173279534029, "learning_rate": 1.4351559599190708e-06, "loss": 0.0116, "step": 570 }, { "epoch": 3.66, "grad_norm": 0.8890755643373717, "learning_rate": 1.4225296413039794e-06, "loss": 0.0154, "step": 571 }, { "epoch": 3.67, "grad_norm": 1.0180803274467207, "learning_rate": 1.4099449290402492e-06, "loss": 0.0239, "step": 572 }, { "epoch": 3.67, "grad_norm": 0.8050132864758199, "learning_rate": 1.3974020751670734e-06, "loss": 0.0142, "step": 573 }, { "epoch": 3.68, "grad_norm": 0.9413472638967014, "learning_rate": 1.3849013308853369e-06, "loss": 0.0244, "step": 574 }, { "epoch": 3.69, "grad_norm": 0.7466941890622708, "learning_rate": 1.3724429465525733e-06, "loss": 0.012, "step": 575 }, { "epoch": 3.69, "grad_norm": 0.7962567303062643, "learning_rate": 1.360027171677957e-06, "loss": 0.0171, "step": 576 }, { "epoch": 3.7, "grad_norm": 0.8736628823165199, "learning_rate": 1.3476542549173097e-06, "loss": 0.0228, "step": 577 }, { "epoch": 3.71, "grad_norm": 0.8479619989881332, "learning_rate": 1.335324444068108e-06, "loss": 0.0237, "step": 578 }, { "epoch": 3.71, "grad_norm": 1.016855000511962, "learning_rate": 1.3230379860645363e-06, "loss": 0.0239, "step": 579 }, { "epoch": 3.72, "grad_norm": 0.9313677958108982, "learning_rate": 1.3107951269725286e-06, "loss": 0.0183, "step": 580 }, { "epoch": 3.72, "grad_norm": 0.769512727084196, "learning_rate": 1.2985961119848508e-06, "loss": 0.0151, "step": 581 }, { "epoch": 3.73, "grad_norm": 0.8107437247743629, "learning_rate": 1.28644118541618e-06, "loss": 0.0214, "step": 582 }, { "epoch": 3.74, "grad_norm": 0.7390308078541298, "learning_rate": 1.2743305906982184e-06, "loss": 0.014, "step": 583 }, { "epoch": 3.74, "grad_norm": 0.8815024064424067, "learning_rate": 1.2622645703748163e-06, "loss": 0.0176, "step": 584 }, { "epoch": 3.75, "grad_norm": 1.0125087727077078, "learning_rate": 1.2502433660971122e-06, "loss": 0.0161, "step": 585 }, { "epoch": 3.76, "grad_norm": 0.7991134263859656, "learning_rate": 1.2382672186187003e-06, "loss": 0.0181, "step": 586 }, { "epoch": 3.76, "grad_norm": 0.7695977276266822, "learning_rate": 1.2263363677907975e-06, "loss": 0.0144, "step": 587 }, { "epoch": 3.77, "grad_norm": 0.7707016914429152, "learning_rate": 1.214451052557453e-06, "loss": 0.0152, "step": 588 }, { "epoch": 3.78, "grad_norm": 0.7456989741080493, "learning_rate": 1.202611510950747e-06, "loss": 0.0171, "step": 589 }, { "epoch": 3.78, "grad_norm": 0.9961563272393459, "learning_rate": 1.1908179800860415e-06, "loss": 0.0185, "step": 590 }, { "epoch": 3.79, "grad_norm": 0.8118922886096522, "learning_rate": 1.1790706961572176e-06, "loss": 0.0148, "step": 591 }, { "epoch": 3.79, "grad_norm": 0.8151702390898807, "learning_rate": 1.167369894431949e-06, "loss": 0.021, "step": 592 }, { "epoch": 3.8, "grad_norm": 0.9901756508859597, "learning_rate": 1.1557158092469968e-06, "loss": 0.0163, "step": 593 }, { "epoch": 3.81, "grad_norm": 0.768365224491104, "learning_rate": 1.1441086740035036e-06, "loss": 0.0166, "step": 594 }, { "epoch": 3.81, "grad_norm": 0.8630559070642096, "learning_rate": 1.1325487211623343e-06, "loss": 0.0159, "step": 595 }, { "epoch": 3.82, "grad_norm": 0.6603172130754364, "learning_rate": 1.121036182239403e-06, "loss": 0.0131, "step": 596 }, { "epoch": 3.83, "grad_norm": 0.9272515645659455, "learning_rate": 1.1095712878010542e-06, "loss": 0.0265, "step": 597 }, { "epoch": 3.83, "grad_norm": 0.7321005503092809, "learning_rate": 1.0981542674594327e-06, "loss": 0.0171, "step": 598 }, { "epoch": 3.84, "grad_norm": 1.044787276822999, "learning_rate": 1.08678534986789e-06, "loss": 0.0236, "step": 599 }, { "epoch": 3.85, "grad_norm": 0.8453699406585213, "learning_rate": 1.0754647627164022e-06, "loss": 0.0214, "step": 600 }, { "epoch": 3.85, "grad_norm": 0.7651768008769598, "learning_rate": 1.064192732727016e-06, "loss": 0.0152, "step": 601 }, { "epoch": 3.86, "grad_norm": 0.8101285929402818, "learning_rate": 1.0529694856493002e-06, "loss": 0.0156, "step": 602 }, { "epoch": 3.87, "grad_norm": 0.8055953151019704, "learning_rate": 1.0417952462558286e-06, "loss": 0.02, "step": 603 }, { "epoch": 3.87, "grad_norm": 0.7870823941792655, "learning_rate": 1.0306702383376813e-06, "loss": 0.0171, "step": 604 }, { "epoch": 3.88, "grad_norm": 0.7771789691810701, "learning_rate": 1.0195946846999551e-06, "loss": 0.0173, "step": 605 }, { "epoch": 3.88, "grad_norm": 0.8068664036381833, "learning_rate": 1.0085688071573086e-06, "loss": 0.0233, "step": 606 }, { "epoch": 3.89, "grad_norm": 0.8738920300362064, "learning_rate": 9.97592826529514e-07, "loss": 0.0237, "step": 607 }, { "epoch": 3.9, "grad_norm": 0.841806738949766, "learning_rate": 9.866669626370412e-07, "loss": 0.0192, "step": 608 }, { "epoch": 3.9, "grad_norm": 0.8875635219003588, "learning_rate": 9.757914342966495e-07, "loss": 0.0154, "step": 609 }, { "epoch": 3.91, "grad_norm": 0.757847160378628, "learning_rate": 9.649664593170062e-07, "loss": 0.0136, "step": 610 }, { "epoch": 3.92, "grad_norm": 0.8909717857074549, "learning_rate": 9.541922544943295e-07, "loss": 0.0192, "step": 611 }, { "epoch": 3.92, "grad_norm": 0.8943576006561003, "learning_rate": 9.434690356080394e-07, "loss": 0.0139, "step": 612 }, { "epoch": 3.93, "grad_norm": 0.6645490438288841, "learning_rate": 9.327970174164409e-07, "loss": 0.0142, "step": 613 }, { "epoch": 3.94, "grad_norm": 0.67608989131137, "learning_rate": 9.221764136524202e-07, "loss": 0.0157, "step": 614 }, { "epoch": 3.94, "grad_norm": 0.7343067591040218, "learning_rate": 9.116074370191705e-07, "loss": 0.0142, "step": 615 }, { "epoch": 3.95, "grad_norm": 0.9063118920210492, "learning_rate": 9.010902991859196e-07, "loss": 0.0352, "step": 616 }, { "epoch": 3.96, "grad_norm": 0.7302861209580851, "learning_rate": 8.906252107837054e-07, "loss": 0.021, "step": 617 }, { "epoch": 3.96, "grad_norm": 0.951381903299599, "learning_rate": 8.802123814011458e-07, "loss": 0.0219, "step": 618 }, { "epoch": 3.97, "grad_norm": 0.8401460617775475, "learning_rate": 8.698520195802499e-07, "loss": 0.0181, "step": 619 }, { "epoch": 3.97, "grad_norm": 0.9542163084486763, "learning_rate": 8.595443328122345e-07, "loss": 0.0149, "step": 620 }, { "epoch": 3.98, "grad_norm": 0.7339711918314168, "learning_rate": 8.492895275333705e-07, "loss": 0.0135, "step": 621 }, { "epoch": 3.99, "grad_norm": 1.0384660655095288, "learning_rate": 8.390878091208544e-07, "loss": 0.0126, "step": 622 }, { "epoch": 3.99, "grad_norm": 0.882866454798963, "learning_rate": 8.289393818886837e-07, "loss": 0.0145, "step": 623 }, { "epoch": 4.0, "grad_norm": 1.0887387218063342, "learning_rate": 8.188444490835774e-07, "loss": 0.0163, "step": 624 }, { "epoch": 4.01, "grad_norm": 0.8850634568075848, "learning_rate": 8.088032128808952e-07, "loss": 0.0129, "step": 625 }, { "epoch": 4.01, "grad_norm": 0.8517304003704518, "learning_rate": 7.988158743805973e-07, "loss": 0.0199, "step": 626 }, { "epoch": 4.02, "grad_norm": 0.7431241826485437, "learning_rate": 7.888826336032093e-07, "loss": 0.0201, "step": 627 }, { "epoch": 4.03, "grad_norm": 0.6695280139807532, "learning_rate": 7.790036894858198e-07, "loss": 0.0201, "step": 628 }, { "epoch": 4.03, "grad_norm": 0.6069369320073023, "learning_rate": 7.691792398780962e-07, "loss": 0.0118, "step": 629 }, { "epoch": 4.04, "grad_norm": 0.7630660606526203, "learning_rate": 7.594094815383223e-07, "loss": 0.0109, "step": 630 }, { "epoch": 4.04, "grad_norm": 0.7085481933951672, "learning_rate": 7.496946101294585e-07, "loss": 0.0123, "step": 631 }, { "epoch": 4.05, "grad_norm": 0.7837964413939096, "learning_rate": 7.400348202152192e-07, "loss": 0.0196, "step": 632 }, { "epoch": 4.06, "grad_norm": 0.6565704481331924, "learning_rate": 7.304303052561841e-07, "loss": 0.0122, "step": 633 }, { "epoch": 4.06, "grad_norm": 0.6955396545607133, "learning_rate": 7.208812576059113e-07, "loss": 0.0156, "step": 634 }, { "epoch": 4.07, "grad_norm": 0.5956060246770807, "learning_rate": 7.113878685070994e-07, "loss": 0.015, "step": 635 }, { "epoch": 4.08, "grad_norm": 0.5935447259636408, "learning_rate": 7.019503280877466e-07, "loss": 0.0105, "step": 636 }, { "epoch": 4.08, "grad_norm": 0.676191247655706, "learning_rate": 6.925688253573465e-07, "loss": 0.0128, "step": 637 }, { "epoch": 4.09, "grad_norm": 0.5351296056518017, "learning_rate": 6.832435482031064e-07, "loss": 0.0094, "step": 638 }, { "epoch": 4.1, "grad_norm": 0.5478275824393115, "learning_rate": 6.73974683386176e-07, "loss": 0.0103, "step": 639 }, { "epoch": 4.1, "grad_norm": 0.5637592397358228, "learning_rate": 6.647624165379173e-07, "loss": 0.0131, "step": 640 }, { "epoch": 4.11, "grad_norm": 0.5844732480577218, "learning_rate": 6.55606932156175e-07, "loss": 0.012, "step": 641 }, { "epoch": 4.12, "grad_norm": 0.5560435069211428, "learning_rate": 6.465084136015951e-07, "loss": 0.0084, "step": 642 }, { "epoch": 4.12, "grad_norm": 0.4885438511561934, "learning_rate": 6.374670430939404e-07, "loss": 0.0072, "step": 643 }, { "epoch": 4.13, "grad_norm": 0.5653559793263204, "learning_rate": 6.284830017084488e-07, "loss": 0.0068, "step": 644 }, { "epoch": 4.13, "grad_norm": 0.6135164676196407, "learning_rate": 6.195564693722027e-07, "loss": 0.0142, "step": 645 }, { "epoch": 4.14, "grad_norm": 0.4518891844084337, "learning_rate": 6.106876248605299e-07, "loss": 0.0102, "step": 646 }, { "epoch": 4.15, "grad_norm": 0.4974300377898031, "learning_rate": 6.018766457934177e-07, "loss": 0.0054, "step": 647 }, { "epoch": 4.15, "grad_norm": 0.5411109097379481, "learning_rate": 5.931237086319592e-07, "loss": 0.009, "step": 648 }, { "epoch": 4.16, "grad_norm": 0.5359394089968081, "learning_rate": 5.844289886748196e-07, "loss": 0.0134, "step": 649 }, { "epoch": 4.17, "grad_norm": 0.80432664191503, "learning_rate": 5.757926600547231e-07, "loss": 0.0075, "step": 650 }, { "epoch": 4.17, "grad_norm": 0.7830393837005951, "learning_rate": 5.672148957349661e-07, "loss": 0.0106, "step": 651 }, { "epoch": 4.18, "grad_norm": 0.6788613866964288, "learning_rate": 5.586958675059548e-07, "loss": 0.013, "step": 652 }, { "epoch": 4.19, "grad_norm": 0.47264907469733186, "learning_rate": 5.502357459817639e-07, "loss": 0.0095, "step": 653 }, { "epoch": 4.19, "grad_norm": 0.48175069610687854, "learning_rate": 5.418347005967189e-07, "loss": 0.0071, "step": 654 }, { "epoch": 4.2, "grad_norm": 0.6790920380461669, "learning_rate": 5.334928996020013e-07, "loss": 0.0115, "step": 655 }, { "epoch": 4.21, "grad_norm": 0.4090145392612429, "learning_rate": 5.252105100622848e-07, "loss": 0.0073, "step": 656 }, { "epoch": 4.21, "grad_norm": 0.49819569886726583, "learning_rate": 5.169876978523828e-07, "loss": 0.0091, "step": 657 }, { "epoch": 4.22, "grad_norm": 0.6045179104834907, "learning_rate": 5.088246276539292e-07, "loss": 0.0129, "step": 658 }, { "epoch": 4.22, "grad_norm": 0.40815948732981716, "learning_rate": 5.0072146295208e-07, "loss": 0.0067, "step": 659 }, { "epoch": 4.23, "grad_norm": 0.5355302694809556, "learning_rate": 4.926783660322411e-07, "loss": 0.0091, "step": 660 }, { "epoch": 4.24, "grad_norm": 0.4672977933370541, "learning_rate": 4.846954979768149e-07, "loss": 0.0081, "step": 661 }, { "epoch": 4.24, "grad_norm": 0.6156756390110313, "learning_rate": 4.7677301866197455e-07, "loss": 0.0103, "step": 662 }, { "epoch": 4.25, "grad_norm": 0.4919976018539734, "learning_rate": 4.6891108675446453e-07, "loss": 0.0075, "step": 663 }, { "epoch": 4.26, "grad_norm": 0.3845755495693945, "learning_rate": 4.611098597084226e-07, "loss": 0.0075, "step": 664 }, { "epoch": 4.26, "grad_norm": 0.5879587628160349, "learning_rate": 4.533694937622227e-07, "loss": 0.0096, "step": 665 }, { "epoch": 4.27, "grad_norm": 0.4812494262734211, "learning_rate": 4.456901439353499e-07, "loss": 0.0066, "step": 666 }, { "epoch": 4.28, "grad_norm": 0.5332547275950345, "learning_rate": 4.3807196402529535e-07, "loss": 0.0077, "step": 667 }, { "epoch": 4.28, "grad_norm": 0.6890481785285228, "learning_rate": 4.3051510660447336e-07, "loss": 0.0057, "step": 668 }, { "epoch": 4.29, "grad_norm": 0.47168726099727676, "learning_rate": 4.2301972301716934e-07, "loss": 0.0086, "step": 669 }, { "epoch": 4.29, "grad_norm": 0.5049838939089353, "learning_rate": 4.155859633765044e-07, "loss": 0.0078, "step": 670 }, { "epoch": 4.3, "grad_norm": 0.34008519374005064, "learning_rate": 4.0821397656143503e-07, "loss": 0.0048, "step": 671 }, { "epoch": 4.31, "grad_norm": 0.607517447320986, "learning_rate": 4.009039102137657e-07, "loss": 0.0051, "step": 672 }, { "epoch": 4.31, "grad_norm": 0.5920287455206938, "learning_rate": 3.9365591073519387e-07, "loss": 0.0089, "step": 673 }, { "epoch": 4.32, "grad_norm": 0.5406111609382083, "learning_rate": 3.8647012328438085e-07, "loss": 0.0063, "step": 674 }, { "epoch": 4.33, "grad_norm": 0.35958261003969805, "learning_rate": 3.793466917740402e-07, "loss": 0.0061, "step": 675 }, { "epoch": 4.33, "grad_norm": 0.5049810947725443, "learning_rate": 3.7228575886805744e-07, "loss": 0.0091, "step": 676 }, { "epoch": 4.34, "grad_norm": 0.641243030116075, "learning_rate": 3.6528746597863283e-07, "loss": 0.0091, "step": 677 }, { "epoch": 4.35, "grad_norm": 0.5551219906137611, "learning_rate": 3.583519532634516e-07, "loss": 0.0093, "step": 678 }, { "epoch": 4.35, "grad_norm": 0.5639034541755932, "learning_rate": 3.514793596228702e-07, "loss": 0.0078, "step": 679 }, { "epoch": 4.36, "grad_norm": 0.6129837509479856, "learning_rate": 3.44669822697144e-07, "loss": 0.0072, "step": 680 }, { "epoch": 4.37, "grad_norm": 0.4367986472936705, "learning_rate": 3.3792347886366265e-07, "loss": 0.0088, "step": 681 }, { "epoch": 4.37, "grad_norm": 0.46298536553833297, "learning_rate": 3.31240463234221e-07, "loss": 0.0077, "step": 682 }, { "epoch": 4.38, "grad_norm": 0.47004028200875053, "learning_rate": 3.2462090965231767e-07, "loss": 0.0057, "step": 683 }, { "epoch": 4.38, "grad_norm": 0.4778435373643932, "learning_rate": 3.180649506904667e-07, "loss": 0.0084, "step": 684 }, { "epoch": 4.39, "grad_norm": 0.3789331021336801, "learning_rate": 3.1157271764755085e-07, "loss": 0.0059, "step": 685 }, { "epoch": 4.4, "grad_norm": 0.5304877562885468, "learning_rate": 3.0514434054618216e-07, "loss": 0.0071, "step": 686 }, { "epoch": 4.4, "grad_norm": 0.453404842137191, "learning_rate": 2.987799481301091e-07, "loss": 0.005, "step": 687 }, { "epoch": 4.41, "grad_norm": 0.4946215142129279, "learning_rate": 2.924796678616297e-07, "loss": 0.0064, "step": 688 }, { "epoch": 4.42, "grad_norm": 0.43531588103388913, "learning_rate": 2.862436259190414e-07, "loss": 0.0074, "step": 689 }, { "epoch": 4.42, "grad_norm": 0.42256615584272145, "learning_rate": 2.800719471941152e-07, "loss": 0.0052, "step": 690 }, { "epoch": 4.43, "grad_norm": 0.5292189519846633, "learning_rate": 2.739647552895949e-07, "loss": 0.0098, "step": 691 }, { "epoch": 4.44, "grad_norm": 0.2916439347063657, "learning_rate": 2.6792217251671744e-07, "loss": 0.0042, "step": 692 }, { "epoch": 4.44, "grad_norm": 0.42877425290392307, "learning_rate": 2.619443198927677e-07, "loss": 0.0075, "step": 693 }, { "epoch": 4.45, "grad_norm": 0.35484337293926443, "learning_rate": 2.5603131713865374e-07, "loss": 0.0078, "step": 694 }, { "epoch": 4.46, "grad_norm": 0.4543586667955343, "learning_rate": 2.50183282676508e-07, "loss": 0.0059, "step": 695 }, { "epoch": 4.46, "grad_norm": 0.48615348530445246, "learning_rate": 2.444003336273163e-07, "loss": 0.0075, "step": 696 }, { "epoch": 4.47, "grad_norm": 0.3946759106084988, "learning_rate": 2.3868258580857164e-07, "loss": 0.0067, "step": 697 }, { "epoch": 4.47, "grad_norm": 0.3682990107394935, "learning_rate": 2.3303015373195713e-07, "loss": 0.0065, "step": 698 }, { "epoch": 4.48, "grad_norm": 0.3469846297857803, "learning_rate": 2.2744315060104846e-07, "loss": 0.0057, "step": 699 }, { "epoch": 4.49, "grad_norm": 0.3055870761212841, "learning_rate": 2.2192168830904963e-07, "loss": 0.0046, "step": 700 }, { "epoch": 4.49, "grad_norm": 0.2684039161842392, "learning_rate": 2.1646587743655287e-07, "loss": 0.0044, "step": 701 }, { "epoch": 4.5, "grad_norm": 0.551851617769783, "learning_rate": 2.1107582724932088e-07, "loss": 0.0076, "step": 702 }, { "epoch": 4.51, "grad_norm": 0.3773077489097412, "learning_rate": 2.0575164569610016e-07, "loss": 0.0058, "step": 703 }, { "epoch": 4.51, "grad_norm": 0.30157672847907924, "learning_rate": 2.0049343940645935e-07, "loss": 0.005, "step": 704 }, { "epoch": 4.52, "grad_norm": 0.37102460832798273, "learning_rate": 1.953013136886541e-07, "loss": 0.0072, "step": 705 }, { "epoch": 4.53, "grad_norm": 0.5134825511016379, "learning_rate": 1.901753725275166e-07, "loss": 0.0054, "step": 706 }, { "epoch": 4.53, "grad_norm": 0.3418486411468999, "learning_rate": 1.8511571858237357e-07, "loss": 0.0049, "step": 707 }, { "epoch": 4.54, "grad_norm": 0.37310826466217933, "learning_rate": 1.801224531849908e-07, "loss": 0.0068, "step": 708 }, { "epoch": 4.54, "grad_norm": 0.5358213340322415, "learning_rate": 1.7519567633754352e-07, "loss": 0.0055, "step": 709 }, { "epoch": 4.55, "grad_norm": 0.43174093605907204, "learning_rate": 1.70335486710614e-07, "loss": 0.0104, "step": 710 }, { "epoch": 4.56, "grad_norm": 0.3429914758954129, "learning_rate": 1.6554198164121265e-07, "loss": 0.0052, "step": 711 }, { "epoch": 4.56, "grad_norm": 0.24916354059181808, "learning_rate": 1.6081525713083428e-07, "loss": 0.0038, "step": 712 }, { "epoch": 4.57, "grad_norm": 0.4382399839925513, "learning_rate": 1.561554078435296e-07, "loss": 0.0057, "step": 713 }, { "epoch": 4.58, "grad_norm": 0.4974839410031333, "learning_rate": 1.5156252710401207e-07, "loss": 0.0091, "step": 714 }, { "epoch": 4.58, "grad_norm": 0.3839389633196022, "learning_rate": 1.4703670689578884e-07, "loss": 0.0064, "step": 715 }, { "epoch": 4.59, "grad_norm": 0.3399047282423716, "learning_rate": 1.4257803785931926e-07, "loss": 0.006, "step": 716 }, { "epoch": 4.6, "grad_norm": 0.26059500112902745, "learning_rate": 1.3818660929019717e-07, "loss": 0.0054, "step": 717 }, { "epoch": 4.6, "grad_norm": 0.324557199088587, "learning_rate": 1.3386250913736408e-07, "loss": 0.0053, "step": 718 }, { "epoch": 4.61, "grad_norm": 0.31523371331282024, "learning_rate": 1.296058240013491e-07, "loss": 0.0069, "step": 719 }, { "epoch": 4.62, "grad_norm": 0.2736291729957554, "learning_rate": 1.2541663913253191e-07, "loss": 0.0055, "step": 720 }, { "epoch": 4.62, "grad_norm": 0.3431360054012468, "learning_rate": 1.2129503842943645e-07, "loss": 0.0072, "step": 721 }, { "epoch": 4.63, "grad_norm": 0.30255537482868156, "learning_rate": 1.1724110443705115e-07, "loss": 0.0055, "step": 722 }, { "epoch": 4.63, "grad_norm": 0.35751371687669753, "learning_rate": 1.1325491834517676e-07, "loss": 0.007, "step": 723 }, { "epoch": 4.64, "grad_norm": 0.37341244248321687, "learning_rate": 1.0933655998679653e-07, "loss": 0.0049, "step": 724 }, { "epoch": 4.65, "grad_norm": 0.36294551103333006, "learning_rate": 1.0548610783648199e-07, "loss": 0.0075, "step": 725 }, { "epoch": 4.65, "grad_norm": 0.23289924927760813, "learning_rate": 1.0170363900881795e-07, "loss": 0.003, "step": 726 }, { "epoch": 4.66, "grad_norm": 0.23590969846667792, "learning_rate": 9.798922925685994e-08, "loss": 0.0037, "step": 727 }, { "epoch": 4.67, "grad_norm": 0.371740523030214, "learning_rate": 9.434295297061668e-08, "loss": 0.0081, "step": 728 }, { "epoch": 4.67, "grad_norm": 0.35348995399607075, "learning_rate": 9.076488317555886e-08, "loss": 0.0053, "step": 729 }, { "epoch": 4.68, "grad_norm": 0.5059722433604245, "learning_rate": 8.725509153115918e-08, "loss": 0.0092, "step": 730 }, { "epoch": 4.69, "grad_norm": 0.29220233975264737, "learning_rate": 8.38136483294546e-08, "loss": 0.0033, "step": 731 }, { "epoch": 4.69, "grad_norm": 0.28309868206988914, "learning_rate": 8.044062249364048e-08, "loss": 0.0052, "step": 732 }, { "epoch": 4.7, "grad_norm": 0.4083770685160588, "learning_rate": 7.713608157668921e-08, "loss": 0.0082, "step": 733 }, { "epoch": 4.71, "grad_norm": 0.3242480770397174, "learning_rate": 7.390009175999835e-08, "loss": 0.0087, "step": 734 }, { "epoch": 4.71, "grad_norm": 0.3427949444072894, "learning_rate": 7.073271785206314e-08, "loss": 0.0072, "step": 735 }, { "epoch": 4.72, "grad_norm": 0.4125538286496571, "learning_rate": 6.763402328718116e-08, "loss": 0.0065, "step": 736 }, { "epoch": 4.72, "grad_norm": 0.266510337541702, "learning_rate": 6.460407012417918e-08, "loss": 0.0042, "step": 737 }, { "epoch": 4.73, "grad_norm": 0.3107043699376071, "learning_rate": 6.164291904517333e-08, "loss": 0.0083, "step": 738 }, { "epoch": 4.74, "grad_norm": 0.2613236810681545, "learning_rate": 5.875062935435121e-08, "loss": 0.0051, "step": 739 }, { "epoch": 4.74, "grad_norm": 0.2851078758987879, "learning_rate": 5.592725897678446e-08, "loss": 0.0057, "step": 740 }, { "epoch": 4.75, "grad_norm": 0.21054304193906503, "learning_rate": 5.3172864457271926e-08, "loss": 0.0031, "step": 741 }, { "epoch": 4.76, "grad_norm": 0.3265715406591838, "learning_rate": 5.048750095920151e-08, "loss": 0.0067, "step": 742 }, { "epoch": 4.76, "grad_norm": 0.45665464799625405, "learning_rate": 4.787122226345014e-08, "loss": 0.0055, "step": 743 }, { "epoch": 4.77, "grad_norm": 0.2843677568177273, "learning_rate": 4.532408076730504e-08, "loss": 0.0057, "step": 744 }, { "epoch": 4.78, "grad_norm": 0.28779532309845957, "learning_rate": 4.2846127483414206e-08, "loss": 0.0059, "step": 745 }, { "epoch": 4.78, "grad_norm": 0.3003525952411292, "learning_rate": 4.043741203876483e-08, "loss": 0.0051, "step": 746 }, { "epoch": 4.79, "grad_norm": 0.26104706018588686, "learning_rate": 3.80979826736893e-08, "loss": 0.0049, "step": 747 }, { "epoch": 4.79, "grad_norm": 0.39096885913150037, "learning_rate": 3.58278862409e-08, "loss": 0.0074, "step": 748 }, { "epoch": 4.8, "grad_norm": 0.3088992283183089, "learning_rate": 3.3627168204549306e-08, "loss": 0.0041, "step": 749 }, { "epoch": 4.81, "grad_norm": 0.31604556262271916, "learning_rate": 3.1495872639320357e-08, "loss": 0.0062, "step": 750 }, { "epoch": 4.81, "grad_norm": 0.3121260022084396, "learning_rate": 2.9434042229544543e-08, "loss": 0.0064, "step": 751 }, { "epoch": 4.82, "grad_norm": 0.2529095358959997, "learning_rate": 2.7441718268344737e-08, "loss": 0.0054, "step": 752 }, { "epoch": 4.83, "grad_norm": 0.3758371548507174, "learning_rate": 2.5518940656811095e-08, "loss": 0.0097, "step": 753 }, { "epoch": 4.83, "grad_norm": 0.3434048070755047, "learning_rate": 2.3665747903199418e-08, "loss": 0.0075, "step": 754 }, { "epoch": 4.84, "grad_norm": 0.3932959717890068, "learning_rate": 2.1882177122162173e-08, "loss": 0.0087, "step": 755 }, { "epoch": 4.85, "grad_norm": 0.3941078588894348, "learning_rate": 2.0168264034002404e-08, "loss": 0.0105, "step": 756 }, { "epoch": 4.85, "grad_norm": 0.2780367397788764, "learning_rate": 1.8524042963961095e-08, "loss": 0.0052, "step": 757 }, { "epoch": 4.86, "grad_norm": 0.3463950374581184, "learning_rate": 1.6949546841528607e-08, "loss": 0.0057, "step": 758 }, { "epoch": 4.87, "grad_norm": 0.36448273851642904, "learning_rate": 1.544480719978447e-08, "loss": 0.0085, "step": 759 }, { "epoch": 4.87, "grad_norm": 0.2802121622475311, "learning_rate": 1.4009854174767521e-08, "loss": 0.008, "step": 760 }, { "epoch": 4.88, "grad_norm": 0.313518910941428, "learning_rate": 1.2644716504870091e-08, "loss": 0.007, "step": 761 }, { "epoch": 4.88, "grad_norm": 0.3661778282505006, "learning_rate": 1.1349421530265246e-08, "loss": 0.0093, "step": 762 }, { "epoch": 4.89, "grad_norm": 0.39469666067358067, "learning_rate": 1.0123995192356183e-08, "loss": 0.0102, "step": 763 }, { "epoch": 4.9, "grad_norm": 0.35114903900718974, "learning_rate": 8.968462033259405e-09, "loss": 0.0084, "step": 764 }, { "epoch": 4.9, "grad_norm": 0.3674950171762429, "learning_rate": 7.882845195312016e-09, "loss": 0.0054, "step": 765 }, { "epoch": 4.91, "grad_norm": 0.2775079654458955, "learning_rate": 6.8671664206073625e-09, "loss": 0.0049, "step": 766 }, { "epoch": 4.92, "grad_norm": 0.3258792327477565, "learning_rate": 5.921446050561386e-09, "loss": 0.0085, "step": 767 }, { "epoch": 4.92, "grad_norm": 0.3545502145110547, "learning_rate": 5.0457030255038334e-09, "loss": 0.0044, "step": 768 }, { "epoch": 4.93, "grad_norm": 0.3203502791392638, "learning_rate": 4.239954884299401e-09, "loss": 0.0066, "step": 769 }, { "epoch": 4.94, "grad_norm": 0.322043236377632, "learning_rate": 3.5042177639972304e-09, "loss": 0.0081, "step": 770 }, { "epoch": 4.94, "grad_norm": 0.3262082117453807, "learning_rate": 2.838506399506446e-09, "loss": 0.0068, "step": 771 }, { "epoch": 4.95, "grad_norm": 0.46623582650195866, "learning_rate": 2.2428341233012294e-09, "loss": 0.0197, "step": 772 }, { "epoch": 4.96, "grad_norm": 0.41216660386641557, "learning_rate": 1.7172128651554152e-09, "loss": 0.0102, "step": 773 }, { "epoch": 4.96, "grad_norm": 0.41233357005851606, "learning_rate": 1.2616531519011874e-09, "loss": 0.0081, "step": 774 }, { "epoch": 4.97, "grad_norm": 0.2876638035649477, "learning_rate": 8.761641072196346e-10, "loss": 0.0071, "step": 775 }, { "epoch": 4.97, "grad_norm": 0.3046501833564902, "learning_rate": 5.607534514585066e-10, "loss": 0.0053, "step": 776 }, { "epoch": 4.98, "grad_norm": 0.2748887054069138, "learning_rate": 3.1542750147639517e-10, "loss": 0.0055, "step": 777 }, { "epoch": 4.99, "grad_norm": 0.324688595353322, "learning_rate": 1.401911705168346e-10, "loss": 0.0057, "step": 778 }, { "epoch": 4.99, "grad_norm": 0.2944466244233151, "learning_rate": 3.5047968109214176e-11, "loss": 0.0059, "step": 779 }, { "epoch": 5.0, "grad_norm": 0.31412002348727536, "learning_rate": 0.0, "loss": 0.0066, "step": 780 }, { "epoch": 5.0, "step": 780, "total_flos": 0.0, "train_loss": 0.17025724985541252, "train_runtime": 1577.4481, "train_samples_per_second": 15.848, "train_steps_per_second": 0.494 } ], "logging_steps": 1.0, "max_steps": 780, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }