diff --git "a/VPE2/trainer_state.json" "b/VPE2/trainer_state.json" new file mode 100644--- /dev/null +++ "b/VPE2/trainer_state.json" @@ -0,0 +1,8625 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.03918367346938775, + "eval_steps": 50, + "global_step": 1200, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 3.265306122448979e-05, + "grad_norm": 12.553070068359375, + "learning_rate": 2.0000000000000002e-07, + "loss": 2.3591, + "step": 1 + }, + { + "epoch": 6.530612244897959e-05, + "grad_norm": 12.028986930847168, + "learning_rate": 4.0000000000000003e-07, + "loss": 2.3144, + "step": 2 + }, + { + "epoch": 9.79591836734694e-05, + "grad_norm": 12.335774421691895, + "learning_rate": 6.000000000000001e-07, + "loss": 2.3323, + "step": 3 + }, + { + "epoch": 0.00013061224489795917, + "grad_norm": 12.695511817932129, + "learning_rate": 8.000000000000001e-07, + "loss": 2.3651, + "step": 4 + }, + { + "epoch": 0.00016326530612244898, + "grad_norm": 12.702936172485352, + "learning_rate": 1.0000000000000002e-06, + "loss": 2.3563, + "step": 5 + }, + { + "epoch": 0.0001959183673469388, + "grad_norm": 11.689526557922363, + "learning_rate": 1.2000000000000002e-06, + "loss": 2.3263, + "step": 6 + }, + { + "epoch": 0.00022857142857142857, + "grad_norm": 11.332144737243652, + "learning_rate": 1.4000000000000001e-06, + "loss": 2.2861, + "step": 7 + }, + { + "epoch": 0.00026122448979591835, + "grad_norm": 11.530237197875977, + "learning_rate": 1.6000000000000001e-06, + "loss": 2.2867, + "step": 8 + }, + { + "epoch": 0.0002938775510204082, + "grad_norm": 10.193593978881836, + "learning_rate": 1.8000000000000001e-06, + "loss": 2.2521, + "step": 9 + }, + { + "epoch": 0.00032653061224489796, + "grad_norm": 9.98508071899414, + "learning_rate": 2.0000000000000003e-06, + "loss": 2.249, + "step": 10 + }, + { + "epoch": 0.00035918367346938774, + "grad_norm": 9.584476470947266, + "learning_rate": 2.2e-06, + "loss": 2.2116, + "step": 11 + }, + { + "epoch": 0.0003918367346938776, + "grad_norm": 9.574864387512207, + "learning_rate": 2.4000000000000003e-06, + "loss": 2.2343, + "step": 12 + }, + { + "epoch": 0.00042448979591836735, + "grad_norm": 8.929769515991211, + "learning_rate": 2.6e-06, + "loss": 2.1599, + "step": 13 + }, + { + "epoch": 0.00045714285714285713, + "grad_norm": 8.262470245361328, + "learning_rate": 2.8000000000000003e-06, + "loss": 2.1152, + "step": 14 + }, + { + "epoch": 0.0004897959183673469, + "grad_norm": 8.28957748413086, + "learning_rate": 3e-06, + "loss": 2.1048, + "step": 15 + }, + { + "epoch": 0.0005224489795918367, + "grad_norm": 8.160731315612793, + "learning_rate": 3.2000000000000003e-06, + "loss": 2.054, + "step": 16 + }, + { + "epoch": 0.0005551020408163266, + "grad_norm": 7.5022077560424805, + "learning_rate": 3.4000000000000005e-06, + "loss": 2.0559, + "step": 17 + }, + { + "epoch": 0.0005877551020408164, + "grad_norm": 7.718921661376953, + "learning_rate": 3.6000000000000003e-06, + "loss": 2.003, + "step": 18 + }, + { + "epoch": 0.0006204081632653061, + "grad_norm": 7.6420392990112305, + "learning_rate": 3.8000000000000005e-06, + "loss": 1.9646, + "step": 19 + }, + { + "epoch": 0.0006530612244897959, + "grad_norm": 7.363818645477295, + "learning_rate": 4.000000000000001e-06, + "loss": 1.918, + "step": 20 + }, + { + "epoch": 0.0006857142857142857, + "grad_norm": 7.367091178894043, + "learning_rate": 4.2000000000000004e-06, + "loss": 1.9065, + "step": 21 + }, + { + "epoch": 0.0007183673469387755, + "grad_norm": 7.201758861541748, + "learning_rate": 4.4e-06, + "loss": 1.8989, + "step": 22 + }, + { + "epoch": 0.0007510204081632653, + "grad_norm": 7.093530654907227, + "learning_rate": 4.600000000000001e-06, + "loss": 1.8381, + "step": 23 + }, + { + "epoch": 0.0007836734693877551, + "grad_norm": 7.22179651260376, + "learning_rate": 4.800000000000001e-06, + "loss": 1.7959, + "step": 24 + }, + { + "epoch": 0.0008163265306122449, + "grad_norm": 7.264500141143799, + "learning_rate": 5e-06, + "loss": 1.7654, + "step": 25 + }, + { + "epoch": 0.0008489795918367347, + "grad_norm": 6.931406021118164, + "learning_rate": 5.2e-06, + "loss": 1.6971, + "step": 26 + }, + { + "epoch": 0.0008816326530612245, + "grad_norm": 6.8040547370910645, + "learning_rate": 5.400000000000001e-06, + "loss": 1.6738, + "step": 27 + }, + { + "epoch": 0.0009142857142857143, + "grad_norm": 7.414416313171387, + "learning_rate": 5.600000000000001e-06, + "loss": 1.6131, + "step": 28 + }, + { + "epoch": 0.000946938775510204, + "grad_norm": 8.00181770324707, + "learning_rate": 5.8e-06, + "loss": 1.5743, + "step": 29 + }, + { + "epoch": 0.0009795918367346938, + "grad_norm": 8.280529022216797, + "learning_rate": 6e-06, + "loss": 1.5139, + "step": 30 + }, + { + "epoch": 0.0010122448979591836, + "grad_norm": 24.16852378845215, + "learning_rate": 6.200000000000001e-06, + "loss": 1.4998, + "step": 31 + }, + { + "epoch": 0.0010448979591836734, + "grad_norm": 25.735715866088867, + "learning_rate": 6.4000000000000006e-06, + "loss": 1.4774, + "step": 32 + }, + { + "epoch": 0.0010775510204081632, + "grad_norm": 18.589033126831055, + "learning_rate": 6.600000000000001e-06, + "loss": 1.3795, + "step": 33 + }, + { + "epoch": 0.0011102040816326532, + "grad_norm": 7.6697096824646, + "learning_rate": 6.800000000000001e-06, + "loss": 1.3513, + "step": 34 + }, + { + "epoch": 0.001142857142857143, + "grad_norm": 12.835890769958496, + "learning_rate": 7e-06, + "loss": 1.2908, + "step": 35 + }, + { + "epoch": 0.0011755102040816327, + "grad_norm": 12.289154052734375, + "learning_rate": 7.2000000000000005e-06, + "loss": 1.2354, + "step": 36 + }, + { + "epoch": 0.0012081632653061225, + "grad_norm": 5.951991558074951, + "learning_rate": 7.4e-06, + "loss": 1.2242, + "step": 37 + }, + { + "epoch": 0.0012408163265306123, + "grad_norm": 5.6563825607299805, + "learning_rate": 7.600000000000001e-06, + "loss": 1.2448, + "step": 38 + }, + { + "epoch": 0.001273469387755102, + "grad_norm": 5.884244441986084, + "learning_rate": 7.800000000000002e-06, + "loss": 1.2011, + "step": 39 + }, + { + "epoch": 0.0013061224489795918, + "grad_norm": 4.279225826263428, + "learning_rate": 8.000000000000001e-06, + "loss": 1.1616, + "step": 40 + }, + { + "epoch": 0.0013387755102040816, + "grad_norm": 4.273413181304932, + "learning_rate": 8.2e-06, + "loss": 1.1361, + "step": 41 + }, + { + "epoch": 0.0013714285714285714, + "grad_norm": 4.514435768127441, + "learning_rate": 8.400000000000001e-06, + "loss": 1.1194, + "step": 42 + }, + { + "epoch": 0.0014040816326530612, + "grad_norm": 3.1108927726745605, + "learning_rate": 8.6e-06, + "loss": 1.1359, + "step": 43 + }, + { + "epoch": 0.001436734693877551, + "grad_norm": 3.187659740447998, + "learning_rate": 8.8e-06, + "loss": 1.1088, + "step": 44 + }, + { + "epoch": 0.0014693877551020407, + "grad_norm": 3.0588572025299072, + "learning_rate": 9e-06, + "loss": 1.0367, + "step": 45 + }, + { + "epoch": 0.0015020408163265305, + "grad_norm": 3.0041353702545166, + "learning_rate": 9.200000000000002e-06, + "loss": 1.065, + "step": 46 + }, + { + "epoch": 0.0015346938775510203, + "grad_norm": 3.0067152976989746, + "learning_rate": 9.4e-06, + "loss": 1.106, + "step": 47 + }, + { + "epoch": 0.0015673469387755103, + "grad_norm": 2.8196375370025635, + "learning_rate": 9.600000000000001e-06, + "loss": 1.0463, + "step": 48 + }, + { + "epoch": 0.0016, + "grad_norm": 2.7588930130004883, + "learning_rate": 9.800000000000001e-06, + "loss": 1.0707, + "step": 49 + }, + { + "epoch": 0.0016326530612244899, + "grad_norm": 2.776578903198242, + "learning_rate": 1e-05, + "loss": 1.057, + "step": 50 + }, + { + "epoch": 0.0016326530612244899, + "eval_loss": 1.0555493831634521, + "eval_runtime": 80.9919, + "eval_samples_per_second": 1.235, + "eval_steps_per_second": 1.235, + "step": 50 + }, + { + "epoch": 0.0016653061224489796, + "grad_norm": 3.0007307529449463, + "learning_rate": 9.999967341606794e-06, + "loss": 1.0359, + "step": 51 + }, + { + "epoch": 0.0016979591836734694, + "grad_norm": 2.820908308029175, + "learning_rate": 9.999934683213587e-06, + "loss": 1.0659, + "step": 52 + }, + { + "epoch": 0.0017306122448979592, + "grad_norm": 3.070078134536743, + "learning_rate": 9.999902024820379e-06, + "loss": 1.0676, + "step": 53 + }, + { + "epoch": 0.001763265306122449, + "grad_norm": 2.826664686203003, + "learning_rate": 9.999869366427172e-06, + "loss": 1.0396, + "step": 54 + }, + { + "epoch": 0.0017959183673469388, + "grad_norm": 2.8928892612457275, + "learning_rate": 9.999836708033965e-06, + "loss": 1.0043, + "step": 55 + }, + { + "epoch": 0.0018285714285714285, + "grad_norm": 2.9964358806610107, + "learning_rate": 9.999804049640759e-06, + "loss": 1.0335, + "step": 56 + }, + { + "epoch": 0.0018612244897959183, + "grad_norm": 2.796006202697754, + "learning_rate": 9.999771391247552e-06, + "loss": 0.9906, + "step": 57 + }, + { + "epoch": 0.001893877551020408, + "grad_norm": 2.638707399368286, + "learning_rate": 9.999738732854345e-06, + "loss": 1.0347, + "step": 58 + }, + { + "epoch": 0.0019265306122448979, + "grad_norm": 2.6315219402313232, + "learning_rate": 9.999706074461137e-06, + "loss": 1.0213, + "step": 59 + }, + { + "epoch": 0.0019591836734693877, + "grad_norm": 2.963063955307007, + "learning_rate": 9.99967341606793e-06, + "loss": 0.9669, + "step": 60 + }, + { + "epoch": 0.0019918367346938777, + "grad_norm": 2.913827657699585, + "learning_rate": 9.999640757674723e-06, + "loss": 1.013, + "step": 61 + }, + { + "epoch": 0.002024489795918367, + "grad_norm": 2.830906391143799, + "learning_rate": 9.999608099281516e-06, + "loss": 1.0036, + "step": 62 + }, + { + "epoch": 0.002057142857142857, + "grad_norm": 2.856300115585327, + "learning_rate": 9.99957544088831e-06, + "loss": 0.9806, + "step": 63 + }, + { + "epoch": 0.0020897959183673468, + "grad_norm": 2.8459157943725586, + "learning_rate": 9.999542782495103e-06, + "loss": 1.0253, + "step": 64 + }, + { + "epoch": 0.0021224489795918368, + "grad_norm": 2.748532772064209, + "learning_rate": 9.999510124101896e-06, + "loss": 1.0017, + "step": 65 + }, + { + "epoch": 0.0021551020408163263, + "grad_norm": 2.910276412963867, + "learning_rate": 9.999477465708688e-06, + "loss": 0.978, + "step": 66 + }, + { + "epoch": 0.0021877551020408163, + "grad_norm": 2.8665144443511963, + "learning_rate": 9.99944480731548e-06, + "loss": 0.9914, + "step": 67 + }, + { + "epoch": 0.0022204081632653063, + "grad_norm": 2.6549315452575684, + "learning_rate": 9.999412148922274e-06, + "loss": 0.9735, + "step": 68 + }, + { + "epoch": 0.002253061224489796, + "grad_norm": 2.7277936935424805, + "learning_rate": 9.999379490529067e-06, + "loss": 0.9498, + "step": 69 + }, + { + "epoch": 0.002285714285714286, + "grad_norm": 2.849747657775879, + "learning_rate": 9.99934683213586e-06, + "loss": 0.9976, + "step": 70 + }, + { + "epoch": 0.0023183673469387754, + "grad_norm": 2.8592312335968018, + "learning_rate": 9.999314173742652e-06, + "loss": 0.9523, + "step": 71 + }, + { + "epoch": 0.0023510204081632654, + "grad_norm": 2.7041122913360596, + "learning_rate": 9.999281515349445e-06, + "loss": 0.9692, + "step": 72 + }, + { + "epoch": 0.002383673469387755, + "grad_norm": 2.717122793197632, + "learning_rate": 9.999248856956238e-06, + "loss": 0.9491, + "step": 73 + }, + { + "epoch": 0.002416326530612245, + "grad_norm": 2.516847610473633, + "learning_rate": 9.999216198563032e-06, + "loss": 0.9632, + "step": 74 + }, + { + "epoch": 0.0024489795918367346, + "grad_norm": 2.766266107559204, + "learning_rate": 9.999183540169825e-06, + "loss": 0.932, + "step": 75 + }, + { + "epoch": 0.0024816326530612246, + "grad_norm": 2.8596270084381104, + "learning_rate": 9.999150881776616e-06, + "loss": 0.9479, + "step": 76 + }, + { + "epoch": 0.002514285714285714, + "grad_norm": 2.6034138202667236, + "learning_rate": 9.99911822338341e-06, + "loss": 0.9712, + "step": 77 + }, + { + "epoch": 0.002546938775510204, + "grad_norm": 2.662513256072998, + "learning_rate": 9.999085564990203e-06, + "loss": 0.9882, + "step": 78 + }, + { + "epoch": 0.0025795918367346937, + "grad_norm": 2.900061845779419, + "learning_rate": 9.999052906596996e-06, + "loss": 0.9232, + "step": 79 + }, + { + "epoch": 0.0026122448979591837, + "grad_norm": 2.7503371238708496, + "learning_rate": 9.99902024820379e-06, + "loss": 0.9483, + "step": 80 + }, + { + "epoch": 0.0026448979591836737, + "grad_norm": 2.61838960647583, + "learning_rate": 9.998987589810583e-06, + "loss": 0.9104, + "step": 81 + }, + { + "epoch": 0.0026775510204081632, + "grad_norm": 2.775689125061035, + "learning_rate": 9.998954931417376e-06, + "loss": 0.9145, + "step": 82 + }, + { + "epoch": 0.0027102040816326532, + "grad_norm": 2.695878744125366, + "learning_rate": 9.998922273024167e-06, + "loss": 0.9075, + "step": 83 + }, + { + "epoch": 0.002742857142857143, + "grad_norm": 2.748655080795288, + "learning_rate": 9.99888961463096e-06, + "loss": 0.966, + "step": 84 + }, + { + "epoch": 0.002775510204081633, + "grad_norm": 2.5372986793518066, + "learning_rate": 9.998856956237754e-06, + "loss": 0.9541, + "step": 85 + }, + { + "epoch": 0.0028081632653061224, + "grad_norm": 2.6666107177734375, + "learning_rate": 9.998824297844547e-06, + "loss": 0.9678, + "step": 86 + }, + { + "epoch": 0.0028408163265306124, + "grad_norm": 2.450654983520508, + "learning_rate": 9.99879163945134e-06, + "loss": 0.9404, + "step": 87 + }, + { + "epoch": 0.002873469387755102, + "grad_norm": 2.6407573223114014, + "learning_rate": 9.998758981058134e-06, + "loss": 0.9661, + "step": 88 + }, + { + "epoch": 0.002906122448979592, + "grad_norm": 2.589488983154297, + "learning_rate": 9.998726322664925e-06, + "loss": 0.9528, + "step": 89 + }, + { + "epoch": 0.0029387755102040815, + "grad_norm": 3.260467767715454, + "learning_rate": 9.998693664271718e-06, + "loss": 0.9625, + "step": 90 + }, + { + "epoch": 0.0029714285714285715, + "grad_norm": 2.8341681957244873, + "learning_rate": 9.998661005878512e-06, + "loss": 0.9088, + "step": 91 + }, + { + "epoch": 0.003004081632653061, + "grad_norm": 3.00934100151062, + "learning_rate": 9.998628347485305e-06, + "loss": 0.9148, + "step": 92 + }, + { + "epoch": 0.003036734693877551, + "grad_norm": 2.8187415599823, + "learning_rate": 9.998595689092098e-06, + "loss": 0.9028, + "step": 93 + }, + { + "epoch": 0.0030693877551020406, + "grad_norm": 2.617946147918701, + "learning_rate": 9.99856303069889e-06, + "loss": 0.8886, + "step": 94 + }, + { + "epoch": 0.0031020408163265306, + "grad_norm": 2.7720468044281006, + "learning_rate": 9.998530372305683e-06, + "loss": 0.9417, + "step": 95 + }, + { + "epoch": 0.0031346938775510206, + "grad_norm": 2.8178319931030273, + "learning_rate": 9.998497713912476e-06, + "loss": 0.9012, + "step": 96 + }, + { + "epoch": 0.00316734693877551, + "grad_norm": 2.9068961143493652, + "learning_rate": 9.99846505551927e-06, + "loss": 0.9318, + "step": 97 + }, + { + "epoch": 0.0032, + "grad_norm": 3.227482557296753, + "learning_rate": 9.998432397126063e-06, + "loss": 0.8956, + "step": 98 + }, + { + "epoch": 0.0032326530612244897, + "grad_norm": 2.8666388988494873, + "learning_rate": 9.998399738732854e-06, + "loss": 0.9288, + "step": 99 + }, + { + "epoch": 0.0032653061224489797, + "grad_norm": 2.861967086791992, + "learning_rate": 9.998367080339647e-06, + "loss": 0.8923, + "step": 100 + }, + { + "epoch": 0.0032653061224489797, + "eval_loss": 0.9108777046203613, + "eval_runtime": 73.7729, + "eval_samples_per_second": 1.356, + "eval_steps_per_second": 1.356, + "step": 100 + }, + { + "epoch": 0.0032979591836734693, + "grad_norm": 3.026766777038574, + "learning_rate": 9.99833442194644e-06, + "loss": 0.9085, + "step": 101 + }, + { + "epoch": 0.0033306122448979593, + "grad_norm": 3.0358951091766357, + "learning_rate": 9.998301763553234e-06, + "loss": 0.9019, + "step": 102 + }, + { + "epoch": 0.003363265306122449, + "grad_norm": 2.7646968364715576, + "learning_rate": 9.998269105160027e-06, + "loss": 0.8954, + "step": 103 + }, + { + "epoch": 0.003395918367346939, + "grad_norm": 2.880887746810913, + "learning_rate": 9.99823644676682e-06, + "loss": 0.9257, + "step": 104 + }, + { + "epoch": 0.0034285714285714284, + "grad_norm": 3.1517140865325928, + "learning_rate": 9.998203788373614e-06, + "loss": 0.8951, + "step": 105 + }, + { + "epoch": 0.0034612244897959184, + "grad_norm": 2.7021565437316895, + "learning_rate": 9.998171129980407e-06, + "loss": 0.9521, + "step": 106 + }, + { + "epoch": 0.003493877551020408, + "grad_norm": 2.860952854156494, + "learning_rate": 9.998138471587198e-06, + "loss": 0.902, + "step": 107 + }, + { + "epoch": 0.003526530612244898, + "grad_norm": 3.0253970623016357, + "learning_rate": 9.998105813193992e-06, + "loss": 0.8935, + "step": 108 + }, + { + "epoch": 0.003559183673469388, + "grad_norm": 2.776489734649658, + "learning_rate": 9.998073154800785e-06, + "loss": 0.9323, + "step": 109 + }, + { + "epoch": 0.0035918367346938775, + "grad_norm": 2.3988196849823, + "learning_rate": 9.998040496407578e-06, + "loss": 0.8951, + "step": 110 + }, + { + "epoch": 0.0036244897959183675, + "grad_norm": 2.6600584983825684, + "learning_rate": 9.998007838014371e-06, + "loss": 0.8913, + "step": 111 + }, + { + "epoch": 0.003657142857142857, + "grad_norm": 2.4915781021118164, + "learning_rate": 9.997975179621163e-06, + "loss": 0.8814, + "step": 112 + }, + { + "epoch": 0.003689795918367347, + "grad_norm": 2.7426276206970215, + "learning_rate": 9.997942521227956e-06, + "loss": 0.9027, + "step": 113 + }, + { + "epoch": 0.0037224489795918366, + "grad_norm": 2.80908465385437, + "learning_rate": 9.99790986283475e-06, + "loss": 0.9033, + "step": 114 + }, + { + "epoch": 0.0037551020408163266, + "grad_norm": 2.5127768516540527, + "learning_rate": 9.997877204441543e-06, + "loss": 0.8886, + "step": 115 + }, + { + "epoch": 0.003787755102040816, + "grad_norm": 2.7205052375793457, + "learning_rate": 9.997844546048336e-06, + "loss": 0.9199, + "step": 116 + }, + { + "epoch": 0.003820408163265306, + "grad_norm": 2.6644845008850098, + "learning_rate": 9.997811887655127e-06, + "loss": 0.8722, + "step": 117 + }, + { + "epoch": 0.0038530612244897957, + "grad_norm": 3.05825138092041, + "learning_rate": 9.99777922926192e-06, + "loss": 0.8926, + "step": 118 + }, + { + "epoch": 0.0038857142857142857, + "grad_norm": 2.8610692024230957, + "learning_rate": 9.997746570868714e-06, + "loss": 0.886, + "step": 119 + }, + { + "epoch": 0.003918367346938775, + "grad_norm": 2.8283677101135254, + "learning_rate": 9.997713912475507e-06, + "loss": 0.8838, + "step": 120 + }, + { + "epoch": 0.003951020408163266, + "grad_norm": 3.36906361579895, + "learning_rate": 9.9976812540823e-06, + "loss": 0.8614, + "step": 121 + }, + { + "epoch": 0.003983673469387755, + "grad_norm": 2.949343681335449, + "learning_rate": 9.997648595689093e-06, + "loss": 0.8915, + "step": 122 + }, + { + "epoch": 0.004016326530612245, + "grad_norm": 2.986492156982422, + "learning_rate": 9.997615937295885e-06, + "loss": 0.9053, + "step": 123 + }, + { + "epoch": 0.004048979591836734, + "grad_norm": 2.7720727920532227, + "learning_rate": 9.997583278902678e-06, + "loss": 0.8915, + "step": 124 + }, + { + "epoch": 0.004081632653061225, + "grad_norm": 2.5769472122192383, + "learning_rate": 9.997550620509471e-06, + "loss": 0.8845, + "step": 125 + }, + { + "epoch": 0.004114285714285714, + "grad_norm": 2.9634106159210205, + "learning_rate": 9.997517962116265e-06, + "loss": 0.9057, + "step": 126 + }, + { + "epoch": 0.004146938775510204, + "grad_norm": 2.6103193759918213, + "learning_rate": 9.997485303723058e-06, + "loss": 0.8843, + "step": 127 + }, + { + "epoch": 0.0041795918367346935, + "grad_norm": 2.892089366912842, + "learning_rate": 9.997452645329851e-06, + "loss": 0.8875, + "step": 128 + }, + { + "epoch": 0.004212244897959184, + "grad_norm": 3.0502076148986816, + "learning_rate": 9.997419986936644e-06, + "loss": 0.8654, + "step": 129 + }, + { + "epoch": 0.0042448979591836735, + "grad_norm": 2.968538522720337, + "learning_rate": 9.997387328543436e-06, + "loss": 0.9211, + "step": 130 + }, + { + "epoch": 0.004277551020408163, + "grad_norm": 2.7077767848968506, + "learning_rate": 9.99735467015023e-06, + "loss": 0.849, + "step": 131 + }, + { + "epoch": 0.004310204081632653, + "grad_norm": 2.8962769508361816, + "learning_rate": 9.997322011757022e-06, + "loss": 0.8844, + "step": 132 + }, + { + "epoch": 0.004342857142857143, + "grad_norm": 2.5692780017852783, + "learning_rate": 9.997289353363816e-06, + "loss": 0.8494, + "step": 133 + }, + { + "epoch": 0.004375510204081633, + "grad_norm": 2.603320837020874, + "learning_rate": 9.997256694970609e-06, + "loss": 0.8994, + "step": 134 + }, + { + "epoch": 0.004408163265306122, + "grad_norm": 2.762920618057251, + "learning_rate": 9.9972240365774e-06, + "loss": 0.8894, + "step": 135 + }, + { + "epoch": 0.004440816326530613, + "grad_norm": 2.7908272743225098, + "learning_rate": 9.997191378184194e-06, + "loss": 0.888, + "step": 136 + }, + { + "epoch": 0.004473469387755102, + "grad_norm": 2.405191421508789, + "learning_rate": 9.997158719790987e-06, + "loss": 0.8953, + "step": 137 + }, + { + "epoch": 0.004506122448979592, + "grad_norm": 3.086392879486084, + "learning_rate": 9.99712606139778e-06, + "loss": 0.9024, + "step": 138 + }, + { + "epoch": 0.004538775510204081, + "grad_norm": 3.10508394241333, + "learning_rate": 9.997093403004573e-06, + "loss": 0.8636, + "step": 139 + }, + { + "epoch": 0.004571428571428572, + "grad_norm": 2.7642314434051514, + "learning_rate": 9.997060744611365e-06, + "loss": 0.8589, + "step": 140 + }, + { + "epoch": 0.004604081632653061, + "grad_norm": 2.9966800212860107, + "learning_rate": 9.997028086218158e-06, + "loss": 0.8564, + "step": 141 + }, + { + "epoch": 0.004636734693877551, + "grad_norm": 2.998603582382202, + "learning_rate": 9.996995427824951e-06, + "loss": 0.8596, + "step": 142 + }, + { + "epoch": 0.0046693877551020405, + "grad_norm": 2.815833330154419, + "learning_rate": 9.996962769431745e-06, + "loss": 0.9006, + "step": 143 + }, + { + "epoch": 0.004702040816326531, + "grad_norm": 2.654766798019409, + "learning_rate": 9.996930111038538e-06, + "loss": 0.8683, + "step": 144 + }, + { + "epoch": 0.0047346938775510205, + "grad_norm": 2.9405479431152344, + "learning_rate": 9.996897452645331e-06, + "loss": 0.8756, + "step": 145 + }, + { + "epoch": 0.00476734693877551, + "grad_norm": 3.789085626602173, + "learning_rate": 9.996864794252124e-06, + "loss": 0.8991, + "step": 146 + }, + { + "epoch": 0.0048, + "grad_norm": 3.062678813934326, + "learning_rate": 9.996832135858918e-06, + "loss": 0.8461, + "step": 147 + }, + { + "epoch": 0.00483265306122449, + "grad_norm": 2.656879425048828, + "learning_rate": 9.996799477465709e-06, + "loss": 0.8755, + "step": 148 + }, + { + "epoch": 0.00486530612244898, + "grad_norm": 2.66681170463562, + "learning_rate": 9.996766819072502e-06, + "loss": 0.8882, + "step": 149 + }, + { + "epoch": 0.004897959183673469, + "grad_norm": 2.69744873046875, + "learning_rate": 9.996734160679296e-06, + "loss": 0.8904, + "step": 150 + }, + { + "epoch": 0.004897959183673469, + "eval_loss": 0.8796091675758362, + "eval_runtime": 74.0605, + "eval_samples_per_second": 1.35, + "eval_steps_per_second": 1.35, + "step": 150 + }, + { + "epoch": 0.0049306122448979596, + "grad_norm": 2.9235658645629883, + "learning_rate": 9.996701502286089e-06, + "loss": 0.819, + "step": 151 + }, + { + "epoch": 0.004963265306122449, + "grad_norm": 3.2201597690582275, + "learning_rate": 9.996668843892882e-06, + "loss": 0.8964, + "step": 152 + }, + { + "epoch": 0.004995918367346939, + "grad_norm": 2.7854557037353516, + "learning_rate": 9.996636185499674e-06, + "loss": 0.8886, + "step": 153 + }, + { + "epoch": 0.005028571428571428, + "grad_norm": 2.4900546073913574, + "learning_rate": 9.996603527106467e-06, + "loss": 0.8496, + "step": 154 + }, + { + "epoch": 0.005061224489795919, + "grad_norm": 2.7506489753723145, + "learning_rate": 9.99657086871326e-06, + "loss": 0.9044, + "step": 155 + }, + { + "epoch": 0.005093877551020408, + "grad_norm": 2.8616607189178467, + "learning_rate": 9.996538210320053e-06, + "loss": 0.8685, + "step": 156 + }, + { + "epoch": 0.005126530612244898, + "grad_norm": 2.814704656600952, + "learning_rate": 9.996505551926847e-06, + "loss": 0.8771, + "step": 157 + }, + { + "epoch": 0.005159183673469387, + "grad_norm": 3.006065845489502, + "learning_rate": 9.996472893533638e-06, + "loss": 0.8571, + "step": 158 + }, + { + "epoch": 0.005191836734693878, + "grad_norm": 2.821923017501831, + "learning_rate": 9.996440235140431e-06, + "loss": 0.8629, + "step": 159 + }, + { + "epoch": 0.005224489795918367, + "grad_norm": 3.278881072998047, + "learning_rate": 9.996407576747225e-06, + "loss": 0.893, + "step": 160 + }, + { + "epoch": 0.005257142857142857, + "grad_norm": 2.744616985321045, + "learning_rate": 9.996374918354018e-06, + "loss": 0.8682, + "step": 161 + }, + { + "epoch": 0.005289795918367347, + "grad_norm": 2.697544574737549, + "learning_rate": 9.996342259960811e-06, + "loss": 0.8974, + "step": 162 + }, + { + "epoch": 0.005322448979591837, + "grad_norm": 2.9768247604370117, + "learning_rate": 9.996309601567604e-06, + "loss": 0.8931, + "step": 163 + }, + { + "epoch": 0.0053551020408163265, + "grad_norm": 3.027183771133423, + "learning_rate": 9.996276943174396e-06, + "loss": 0.8897, + "step": 164 + }, + { + "epoch": 0.005387755102040816, + "grad_norm": 2.7222235202789307, + "learning_rate": 9.996244284781189e-06, + "loss": 0.8787, + "step": 165 + }, + { + "epoch": 0.0054204081632653065, + "grad_norm": 2.7184181213378906, + "learning_rate": 9.996211626387982e-06, + "loss": 0.8832, + "step": 166 + }, + { + "epoch": 0.005453061224489796, + "grad_norm": 2.8517744541168213, + "learning_rate": 9.996178967994776e-06, + "loss": 0.8463, + "step": 167 + }, + { + "epoch": 0.005485714285714286, + "grad_norm": 3.104189395904541, + "learning_rate": 9.996146309601569e-06, + "loss": 0.8724, + "step": 168 + }, + { + "epoch": 0.005518367346938775, + "grad_norm": 2.9883522987365723, + "learning_rate": 9.996113651208362e-06, + "loss": 0.8654, + "step": 169 + }, + { + "epoch": 0.005551020408163266, + "grad_norm": 3.018421173095703, + "learning_rate": 9.996080992815155e-06, + "loss": 0.8836, + "step": 170 + }, + { + "epoch": 0.005583673469387755, + "grad_norm": 2.795041561126709, + "learning_rate": 9.996048334421947e-06, + "loss": 0.8843, + "step": 171 + }, + { + "epoch": 0.005616326530612245, + "grad_norm": 2.594553232192993, + "learning_rate": 9.99601567602874e-06, + "loss": 0.8874, + "step": 172 + }, + { + "epoch": 0.005648979591836734, + "grad_norm": 2.946117877960205, + "learning_rate": 9.995983017635533e-06, + "loss": 0.8265, + "step": 173 + }, + { + "epoch": 0.005681632653061225, + "grad_norm": 3.060215950012207, + "learning_rate": 9.995950359242327e-06, + "loss": 0.8872, + "step": 174 + }, + { + "epoch": 0.005714285714285714, + "grad_norm": 2.5649561882019043, + "learning_rate": 9.99591770084912e-06, + "loss": 0.898, + "step": 175 + }, + { + "epoch": 0.005746938775510204, + "grad_norm": 2.7028987407684326, + "learning_rate": 9.995885042455911e-06, + "loss": 0.8168, + "step": 176 + }, + { + "epoch": 0.005779591836734694, + "grad_norm": 2.918105125427246, + "learning_rate": 9.995852384062705e-06, + "loss": 0.9006, + "step": 177 + }, + { + "epoch": 0.005812244897959184, + "grad_norm": 3.0559136867523193, + "learning_rate": 9.995819725669498e-06, + "loss": 0.9106, + "step": 178 + }, + { + "epoch": 0.005844897959183673, + "grad_norm": 2.7586793899536133, + "learning_rate": 9.995787067276291e-06, + "loss": 0.8643, + "step": 179 + }, + { + "epoch": 0.005877551020408163, + "grad_norm": 2.6476991176605225, + "learning_rate": 9.995754408883084e-06, + "loss": 0.8952, + "step": 180 + }, + { + "epoch": 0.005910204081632653, + "grad_norm": 2.624241352081299, + "learning_rate": 9.995721750489876e-06, + "loss": 0.8836, + "step": 181 + }, + { + "epoch": 0.005942857142857143, + "grad_norm": 3.0197315216064453, + "learning_rate": 9.995689092096669e-06, + "loss": 0.87, + "step": 182 + }, + { + "epoch": 0.0059755102040816325, + "grad_norm": 2.9282002449035645, + "learning_rate": 9.995656433703462e-06, + "loss": 0.8904, + "step": 183 + }, + { + "epoch": 0.006008163265306122, + "grad_norm": 2.6132211685180664, + "learning_rate": 9.995623775310255e-06, + "loss": 0.8565, + "step": 184 + }, + { + "epoch": 0.0060408163265306125, + "grad_norm": 2.6105284690856934, + "learning_rate": 9.995591116917049e-06, + "loss": 0.8842, + "step": 185 + }, + { + "epoch": 0.006073469387755102, + "grad_norm": 2.868211030960083, + "learning_rate": 9.995558458523842e-06, + "loss": 0.8575, + "step": 186 + }, + { + "epoch": 0.006106122448979592, + "grad_norm": 2.7017822265625, + "learning_rate": 9.995525800130633e-06, + "loss": 0.8429, + "step": 187 + }, + { + "epoch": 0.006138775510204081, + "grad_norm": 2.7130422592163086, + "learning_rate": 9.995493141737427e-06, + "loss": 0.8451, + "step": 188 + }, + { + "epoch": 0.006171428571428572, + "grad_norm": 2.681940793991089, + "learning_rate": 9.99546048334422e-06, + "loss": 0.866, + "step": 189 + }, + { + "epoch": 0.006204081632653061, + "grad_norm": 2.6012067794799805, + "learning_rate": 9.995427824951013e-06, + "loss": 0.8575, + "step": 190 + }, + { + "epoch": 0.006236734693877551, + "grad_norm": 2.7982876300811768, + "learning_rate": 9.995395166557806e-06, + "loss": 0.8818, + "step": 191 + }, + { + "epoch": 0.006269387755102041, + "grad_norm": 2.7027719020843506, + "learning_rate": 9.9953625081646e-06, + "loss": 0.8638, + "step": 192 + }, + { + "epoch": 0.006302040816326531, + "grad_norm": 2.6796092987060547, + "learning_rate": 9.995329849771393e-06, + "loss": 0.8473, + "step": 193 + }, + { + "epoch": 0.00633469387755102, + "grad_norm": 2.882390022277832, + "learning_rate": 9.995297191378184e-06, + "loss": 0.8854, + "step": 194 + }, + { + "epoch": 0.00636734693877551, + "grad_norm": 2.970914125442505, + "learning_rate": 9.995264532984978e-06, + "loss": 0.8499, + "step": 195 + }, + { + "epoch": 0.0064, + "grad_norm": 2.7777442932128906, + "learning_rate": 9.995231874591771e-06, + "loss": 0.8761, + "step": 196 + }, + { + "epoch": 0.00643265306122449, + "grad_norm": 2.6736974716186523, + "learning_rate": 9.995199216198564e-06, + "loss": 0.8645, + "step": 197 + }, + { + "epoch": 0.006465306122448979, + "grad_norm": 2.9803805351257324, + "learning_rate": 9.995166557805357e-06, + "loss": 0.8257, + "step": 198 + }, + { + "epoch": 0.006497959183673469, + "grad_norm": 2.7522194385528564, + "learning_rate": 9.995133899412149e-06, + "loss": 0.8467, + "step": 199 + }, + { + "epoch": 0.006530612244897959, + "grad_norm": 2.6975879669189453, + "learning_rate": 9.995101241018942e-06, + "loss": 0.872, + "step": 200 + }, + { + "epoch": 0.006530612244897959, + "eval_loss": 0.8632190823554993, + "eval_runtime": 74.7709, + "eval_samples_per_second": 1.337, + "eval_steps_per_second": 1.337, + "step": 200 + }, + { + "epoch": 0.006563265306122449, + "grad_norm": 2.8385777473449707, + "learning_rate": 9.995068582625735e-06, + "loss": 0.8592, + "step": 201 + }, + { + "epoch": 0.0065959183673469386, + "grad_norm": 2.8501343727111816, + "learning_rate": 9.995035924232529e-06, + "loss": 0.8516, + "step": 202 + }, + { + "epoch": 0.006628571428571429, + "grad_norm": 2.841095447540283, + "learning_rate": 9.995003265839322e-06, + "loss": 0.8589, + "step": 203 + }, + { + "epoch": 0.0066612244897959185, + "grad_norm": 2.7659661769866943, + "learning_rate": 9.994970607446113e-06, + "loss": 0.8434, + "step": 204 + }, + { + "epoch": 0.006693877551020408, + "grad_norm": 2.892019748687744, + "learning_rate": 9.994937949052907e-06, + "loss": 0.8896, + "step": 205 + }, + { + "epoch": 0.006726530612244898, + "grad_norm": 2.7645156383514404, + "learning_rate": 9.9949052906597e-06, + "loss": 0.8654, + "step": 206 + }, + { + "epoch": 0.006759183673469388, + "grad_norm": 2.7840843200683594, + "learning_rate": 9.994872632266493e-06, + "loss": 0.8558, + "step": 207 + }, + { + "epoch": 0.006791836734693878, + "grad_norm": 2.843766212463379, + "learning_rate": 9.994839973873286e-06, + "loss": 0.8482, + "step": 208 + }, + { + "epoch": 0.006824489795918367, + "grad_norm": 2.7724156379699707, + "learning_rate": 9.99480731548008e-06, + "loss": 0.8664, + "step": 209 + }, + { + "epoch": 0.006857142857142857, + "grad_norm": 2.8699095249176025, + "learning_rate": 9.994774657086873e-06, + "loss": 0.8506, + "step": 210 + }, + { + "epoch": 0.006889795918367347, + "grad_norm": 2.988041877746582, + "learning_rate": 9.994741998693666e-06, + "loss": 0.8781, + "step": 211 + }, + { + "epoch": 0.006922448979591837, + "grad_norm": 2.597806930541992, + "learning_rate": 9.994709340300458e-06, + "loss": 0.8586, + "step": 212 + }, + { + "epoch": 0.006955102040816326, + "grad_norm": 2.7888684272766113, + "learning_rate": 9.994676681907251e-06, + "loss": 0.8431, + "step": 213 + }, + { + "epoch": 0.006987755102040816, + "grad_norm": 2.80710768699646, + "learning_rate": 9.994644023514044e-06, + "loss": 0.8911, + "step": 214 + }, + { + "epoch": 0.007020408163265306, + "grad_norm": 2.6957030296325684, + "learning_rate": 9.994611365120837e-06, + "loss": 0.8626, + "step": 215 + }, + { + "epoch": 0.007053061224489796, + "grad_norm": 2.8306329250335693, + "learning_rate": 9.99457870672763e-06, + "loss": 0.8619, + "step": 216 + }, + { + "epoch": 0.0070857142857142855, + "grad_norm": 2.735477924346924, + "learning_rate": 9.994546048334422e-06, + "loss": 0.8661, + "step": 217 + }, + { + "epoch": 0.007118367346938776, + "grad_norm": 2.737856388092041, + "learning_rate": 9.994513389941215e-06, + "loss": 0.8718, + "step": 218 + }, + { + "epoch": 0.0071510204081632655, + "grad_norm": 2.743589401245117, + "learning_rate": 9.994480731548009e-06, + "loss": 0.8781, + "step": 219 + }, + { + "epoch": 0.007183673469387755, + "grad_norm": 2.8569047451019287, + "learning_rate": 9.994448073154802e-06, + "loss": 0.8512, + "step": 220 + }, + { + "epoch": 0.007216326530612245, + "grad_norm": 3.129603385925293, + "learning_rate": 9.994415414761595e-06, + "loss": 0.8839, + "step": 221 + }, + { + "epoch": 0.007248979591836735, + "grad_norm": 2.8460073471069336, + "learning_rate": 9.994382756368387e-06, + "loss": 0.8568, + "step": 222 + }, + { + "epoch": 0.007281632653061225, + "grad_norm": 3.024534225463867, + "learning_rate": 9.99435009797518e-06, + "loss": 0.8672, + "step": 223 + }, + { + "epoch": 0.007314285714285714, + "grad_norm": 3.0231759548187256, + "learning_rate": 9.994317439581973e-06, + "loss": 0.8899, + "step": 224 + }, + { + "epoch": 0.007346938775510204, + "grad_norm": 2.8645284175872803, + "learning_rate": 9.994284781188766e-06, + "loss": 0.8319, + "step": 225 + }, + { + "epoch": 0.007379591836734694, + "grad_norm": 2.741297483444214, + "learning_rate": 9.99425212279556e-06, + "loss": 0.8726, + "step": 226 + }, + { + "epoch": 0.007412244897959184, + "grad_norm": 2.8565313816070557, + "learning_rate": 9.994219464402353e-06, + "loss": 0.8807, + "step": 227 + }, + { + "epoch": 0.007444897959183673, + "grad_norm": 2.9202141761779785, + "learning_rate": 9.994186806009144e-06, + "loss": 0.8098, + "step": 228 + }, + { + "epoch": 0.007477551020408164, + "grad_norm": 2.955820083618164, + "learning_rate": 9.994154147615938e-06, + "loss": 0.8638, + "step": 229 + }, + { + "epoch": 0.007510204081632653, + "grad_norm": 2.7195394039154053, + "learning_rate": 9.99412148922273e-06, + "loss": 0.8591, + "step": 230 + }, + { + "epoch": 0.007542857142857143, + "grad_norm": 2.91768479347229, + "learning_rate": 9.994088830829524e-06, + "loss": 0.8774, + "step": 231 + }, + { + "epoch": 0.007575510204081632, + "grad_norm": 2.7152891159057617, + "learning_rate": 9.994056172436317e-06, + "loss": 0.8488, + "step": 232 + }, + { + "epoch": 0.007608163265306123, + "grad_norm": 2.5756800174713135, + "learning_rate": 9.99402351404311e-06, + "loss": 0.8598, + "step": 233 + }, + { + "epoch": 0.007640816326530612, + "grad_norm": 2.603003978729248, + "learning_rate": 9.993990855649904e-06, + "loss": 0.8473, + "step": 234 + }, + { + "epoch": 0.007673469387755102, + "grad_norm": 2.758294105529785, + "learning_rate": 9.993958197256695e-06, + "loss": 0.8624, + "step": 235 + }, + { + "epoch": 0.0077061224489795915, + "grad_norm": 2.8660902976989746, + "learning_rate": 9.993925538863489e-06, + "loss": 0.8564, + "step": 236 + }, + { + "epoch": 0.007738775510204082, + "grad_norm": 2.8767125606536865, + "learning_rate": 9.993892880470282e-06, + "loss": 0.8578, + "step": 237 + }, + { + "epoch": 0.0077714285714285715, + "grad_norm": 2.906562089920044, + "learning_rate": 9.993860222077075e-06, + "loss": 0.8383, + "step": 238 + }, + { + "epoch": 0.007804081632653061, + "grad_norm": 2.74647855758667, + "learning_rate": 9.993827563683868e-06, + "loss": 0.8656, + "step": 239 + }, + { + "epoch": 0.00783673469387755, + "grad_norm": 3.078082323074341, + "learning_rate": 9.99379490529066e-06, + "loss": 0.8244, + "step": 240 + }, + { + "epoch": 0.00786938775510204, + "grad_norm": 2.7712790966033936, + "learning_rate": 9.993762246897453e-06, + "loss": 0.8536, + "step": 241 + }, + { + "epoch": 0.007902040816326531, + "grad_norm": 2.852389335632324, + "learning_rate": 9.993729588504246e-06, + "loss": 0.8496, + "step": 242 + }, + { + "epoch": 0.007934693877551021, + "grad_norm": 2.7907660007476807, + "learning_rate": 9.99369693011104e-06, + "loss": 0.8993, + "step": 243 + }, + { + "epoch": 0.00796734693877551, + "grad_norm": 2.8219473361968994, + "learning_rate": 9.993664271717833e-06, + "loss": 0.8652, + "step": 244 + }, + { + "epoch": 0.008, + "grad_norm": 2.6678013801574707, + "learning_rate": 9.993631613324624e-06, + "loss": 0.8636, + "step": 245 + }, + { + "epoch": 0.00803265306122449, + "grad_norm": 2.724881887435913, + "learning_rate": 9.993598954931417e-06, + "loss": 0.8559, + "step": 246 + }, + { + "epoch": 0.00806530612244898, + "grad_norm": 2.8275516033172607, + "learning_rate": 9.99356629653821e-06, + "loss": 0.8646, + "step": 247 + }, + { + "epoch": 0.008097959183673469, + "grad_norm": 2.8110673427581787, + "learning_rate": 9.993533638145004e-06, + "loss": 0.8713, + "step": 248 + }, + { + "epoch": 0.008130612244897958, + "grad_norm": 2.652221202850342, + "learning_rate": 9.993500979751797e-06, + "loss": 0.8373, + "step": 249 + }, + { + "epoch": 0.00816326530612245, + "grad_norm": 2.6357979774475098, + "learning_rate": 9.99346832135859e-06, + "loss": 0.8537, + "step": 250 + }, + { + "epoch": 0.00816326530612245, + "eval_loss": 0.8545005917549133, + "eval_runtime": 74.0963, + "eval_samples_per_second": 1.35, + "eval_steps_per_second": 1.35, + "step": 250 + }, + { + "epoch": 0.00819591836734694, + "grad_norm": 2.6132781505584717, + "learning_rate": 9.993435662965384e-06, + "loss": 0.8592, + "step": 251 + }, + { + "epoch": 0.008228571428571429, + "grad_norm": 2.942228078842163, + "learning_rate": 9.993403004572177e-06, + "loss": 0.7732, + "step": 252 + }, + { + "epoch": 0.008261224489795918, + "grad_norm": 2.5452983379364014, + "learning_rate": 9.993370346178968e-06, + "loss": 0.8666, + "step": 253 + }, + { + "epoch": 0.008293877551020408, + "grad_norm": 2.863976240158081, + "learning_rate": 9.993337687785762e-06, + "loss": 0.8232, + "step": 254 + }, + { + "epoch": 0.008326530612244898, + "grad_norm": 2.640972137451172, + "learning_rate": 9.993305029392555e-06, + "loss": 0.8703, + "step": 255 + }, + { + "epoch": 0.008359183673469387, + "grad_norm": 2.98362135887146, + "learning_rate": 9.993272370999348e-06, + "loss": 0.8165, + "step": 256 + }, + { + "epoch": 0.008391836734693878, + "grad_norm": 2.848294258117676, + "learning_rate": 9.993239712606141e-06, + "loss": 0.9104, + "step": 257 + }, + { + "epoch": 0.008424489795918368, + "grad_norm": 2.7992308139801025, + "learning_rate": 9.993207054212933e-06, + "loss": 0.8146, + "step": 258 + }, + { + "epoch": 0.008457142857142858, + "grad_norm": 2.905052900314331, + "learning_rate": 9.993174395819726e-06, + "loss": 0.8606, + "step": 259 + }, + { + "epoch": 0.008489795918367347, + "grad_norm": 2.97420334815979, + "learning_rate": 9.99314173742652e-06, + "loss": 0.7909, + "step": 260 + }, + { + "epoch": 0.008522448979591837, + "grad_norm": 2.7139410972595215, + "learning_rate": 9.993109079033313e-06, + "loss": 0.8593, + "step": 261 + }, + { + "epoch": 0.008555102040816326, + "grad_norm": 2.7178266048431396, + "learning_rate": 9.993076420640106e-06, + "loss": 0.8785, + "step": 262 + }, + { + "epoch": 0.008587755102040816, + "grad_norm": 2.6858327388763428, + "learning_rate": 9.993043762246897e-06, + "loss": 0.846, + "step": 263 + }, + { + "epoch": 0.008620408163265305, + "grad_norm": 2.813204288482666, + "learning_rate": 9.99301110385369e-06, + "loss": 0.8386, + "step": 264 + }, + { + "epoch": 0.008653061224489797, + "grad_norm": 2.795591115951538, + "learning_rate": 9.992978445460484e-06, + "loss": 0.8417, + "step": 265 + }, + { + "epoch": 0.008685714285714286, + "grad_norm": 2.8094005584716797, + "learning_rate": 9.992945787067277e-06, + "loss": 0.8396, + "step": 266 + }, + { + "epoch": 0.008718367346938776, + "grad_norm": 2.723392963409424, + "learning_rate": 9.99291312867407e-06, + "loss": 0.8357, + "step": 267 + }, + { + "epoch": 0.008751020408163265, + "grad_norm": 2.9623281955718994, + "learning_rate": 9.992880470280864e-06, + "loss": 0.8874, + "step": 268 + }, + { + "epoch": 0.008783673469387755, + "grad_norm": 2.8696374893188477, + "learning_rate": 9.992847811887655e-06, + "loss": 0.8595, + "step": 269 + }, + { + "epoch": 0.008816326530612244, + "grad_norm": 2.906726837158203, + "learning_rate": 9.992815153494448e-06, + "loss": 0.8736, + "step": 270 + }, + { + "epoch": 0.008848979591836734, + "grad_norm": 2.814854383468628, + "learning_rate": 9.992782495101242e-06, + "loss": 0.8603, + "step": 271 + }, + { + "epoch": 0.008881632653061225, + "grad_norm": 2.799457550048828, + "learning_rate": 9.992749836708035e-06, + "loss": 0.8469, + "step": 272 + }, + { + "epoch": 0.008914285714285715, + "grad_norm": 2.755631923675537, + "learning_rate": 9.992717178314828e-06, + "loss": 0.8702, + "step": 273 + }, + { + "epoch": 0.008946938775510204, + "grad_norm": 2.9200055599212646, + "learning_rate": 9.992684519921621e-06, + "loss": 0.8407, + "step": 274 + }, + { + "epoch": 0.008979591836734694, + "grad_norm": 3.2251176834106445, + "learning_rate": 9.992651861528415e-06, + "loss": 0.8572, + "step": 275 + }, + { + "epoch": 0.009012244897959184, + "grad_norm": 3.8850901126861572, + "learning_rate": 9.992619203135206e-06, + "loss": 0.8458, + "step": 276 + }, + { + "epoch": 0.009044897959183673, + "grad_norm": 2.8082637786865234, + "learning_rate": 9.992586544742e-06, + "loss": 0.8415, + "step": 277 + }, + { + "epoch": 0.009077551020408163, + "grad_norm": 2.8923559188842773, + "learning_rate": 9.992553886348793e-06, + "loss": 0.8116, + "step": 278 + }, + { + "epoch": 0.009110204081632652, + "grad_norm": 2.849003791809082, + "learning_rate": 9.992521227955586e-06, + "loss": 0.834, + "step": 279 + }, + { + "epoch": 0.009142857142857144, + "grad_norm": 2.7201125621795654, + "learning_rate": 9.992488569562379e-06, + "loss": 0.8518, + "step": 280 + }, + { + "epoch": 0.009175510204081633, + "grad_norm": 2.9798922538757324, + "learning_rate": 9.99245591116917e-06, + "loss": 0.8691, + "step": 281 + }, + { + "epoch": 0.009208163265306123, + "grad_norm": 2.805250406265259, + "learning_rate": 9.992423252775964e-06, + "loss": 0.8537, + "step": 282 + }, + { + "epoch": 0.009240816326530612, + "grad_norm": 2.6364924907684326, + "learning_rate": 9.992390594382757e-06, + "loss": 0.8368, + "step": 283 + }, + { + "epoch": 0.009273469387755102, + "grad_norm": 2.692296028137207, + "learning_rate": 9.99235793598955e-06, + "loss": 0.8373, + "step": 284 + }, + { + "epoch": 0.009306122448979591, + "grad_norm": 2.876549005508423, + "learning_rate": 9.992325277596344e-06, + "loss": 0.8348, + "step": 285 + }, + { + "epoch": 0.009338775510204081, + "grad_norm": 2.77553129196167, + "learning_rate": 9.992292619203135e-06, + "loss": 0.8377, + "step": 286 + }, + { + "epoch": 0.009371428571428572, + "grad_norm": 2.900949478149414, + "learning_rate": 9.992259960809928e-06, + "loss": 0.8469, + "step": 287 + }, + { + "epoch": 0.009404081632653062, + "grad_norm": 2.99700927734375, + "learning_rate": 9.992227302416722e-06, + "loss": 0.8926, + "step": 288 + }, + { + "epoch": 0.009436734693877551, + "grad_norm": 2.738802194595337, + "learning_rate": 9.992194644023515e-06, + "loss": 0.8511, + "step": 289 + }, + { + "epoch": 0.009469387755102041, + "grad_norm": 2.7471158504486084, + "learning_rate": 9.992161985630308e-06, + "loss": 0.8449, + "step": 290 + }, + { + "epoch": 0.00950204081632653, + "grad_norm": 2.7274839878082275, + "learning_rate": 9.992129327237101e-06, + "loss": 0.8308, + "step": 291 + }, + { + "epoch": 0.00953469387755102, + "grad_norm": 2.554326057434082, + "learning_rate": 9.992096668843893e-06, + "loss": 0.868, + "step": 292 + }, + { + "epoch": 0.00956734693877551, + "grad_norm": 2.578237295150757, + "learning_rate": 9.992064010450688e-06, + "loss": 0.8557, + "step": 293 + }, + { + "epoch": 0.0096, + "grad_norm": 3.0283281803131104, + "learning_rate": 9.99203135205748e-06, + "loss": 0.8289, + "step": 294 + }, + { + "epoch": 0.00963265306122449, + "grad_norm": 2.785170316696167, + "learning_rate": 9.991998693664272e-06, + "loss": 0.878, + "step": 295 + }, + { + "epoch": 0.00966530612244898, + "grad_norm": 2.8442277908325195, + "learning_rate": 9.991966035271066e-06, + "loss": 0.8282, + "step": 296 + }, + { + "epoch": 0.00969795918367347, + "grad_norm": 2.847158908843994, + "learning_rate": 9.991933376877859e-06, + "loss": 0.844, + "step": 297 + }, + { + "epoch": 0.00973061224489796, + "grad_norm": 2.6960904598236084, + "learning_rate": 9.991900718484652e-06, + "loss": 0.863, + "step": 298 + }, + { + "epoch": 0.009763265306122449, + "grad_norm": 2.97739839553833, + "learning_rate": 9.991868060091444e-06, + "loss": 0.8448, + "step": 299 + }, + { + "epoch": 0.009795918367346938, + "grad_norm": 2.9316720962524414, + "learning_rate": 9.991835401698237e-06, + "loss": 0.8512, + "step": 300 + }, + { + "epoch": 0.009795918367346938, + "eval_loss": 0.8451775312423706, + "eval_runtime": 73.1659, + "eval_samples_per_second": 1.367, + "eval_steps_per_second": 1.367, + "step": 300 + }, + { + "epoch": 0.009828571428571428, + "grad_norm": 3.2117464542388916, + "learning_rate": 9.99180274330503e-06, + "loss": 0.852, + "step": 301 + }, + { + "epoch": 0.009861224489795919, + "grad_norm": 2.7478814125061035, + "learning_rate": 9.991770084911823e-06, + "loss": 0.8454, + "step": 302 + }, + { + "epoch": 0.009893877551020409, + "grad_norm": 2.7774980068206787, + "learning_rate": 9.991737426518617e-06, + "loss": 0.8581, + "step": 303 + }, + { + "epoch": 0.009926530612244898, + "grad_norm": 2.691710948944092, + "learning_rate": 9.991704768125408e-06, + "loss": 0.8483, + "step": 304 + }, + { + "epoch": 0.009959183673469388, + "grad_norm": 2.803921937942505, + "learning_rate": 9.991672109732201e-06, + "loss": 0.8394, + "step": 305 + }, + { + "epoch": 0.009991836734693877, + "grad_norm": 2.9755606651306152, + "learning_rate": 9.991639451338995e-06, + "loss": 0.8245, + "step": 306 + }, + { + "epoch": 0.010024489795918367, + "grad_norm": 3.09531569480896, + "learning_rate": 9.991606792945788e-06, + "loss": 0.8701, + "step": 307 + }, + { + "epoch": 0.010057142857142857, + "grad_norm": 2.8512625694274902, + "learning_rate": 9.991574134552581e-06, + "loss": 0.8597, + "step": 308 + }, + { + "epoch": 0.010089795918367346, + "grad_norm": 2.7423555850982666, + "learning_rate": 9.991541476159374e-06, + "loss": 0.8116, + "step": 309 + }, + { + "epoch": 0.010122448979591837, + "grad_norm": 2.8539059162139893, + "learning_rate": 9.991508817766166e-06, + "loss": 0.841, + "step": 310 + }, + { + "epoch": 0.010155102040816327, + "grad_norm": 2.843325138092041, + "learning_rate": 9.99147615937296e-06, + "loss": 0.8491, + "step": 311 + }, + { + "epoch": 0.010187755102040816, + "grad_norm": 2.6367223262786865, + "learning_rate": 9.991443500979752e-06, + "loss": 0.8452, + "step": 312 + }, + { + "epoch": 0.010220408163265306, + "grad_norm": 2.6435811519622803, + "learning_rate": 9.991410842586546e-06, + "loss": 0.8415, + "step": 313 + }, + { + "epoch": 0.010253061224489796, + "grad_norm": 2.727060317993164, + "learning_rate": 9.991378184193339e-06, + "loss": 0.7985, + "step": 314 + }, + { + "epoch": 0.010285714285714285, + "grad_norm": 2.9816014766693115, + "learning_rate": 9.991345525800132e-06, + "loss": 0.8129, + "step": 315 + }, + { + "epoch": 0.010318367346938775, + "grad_norm": 3.3140499591827393, + "learning_rate": 9.991312867406925e-06, + "loss": 0.8822, + "step": 316 + }, + { + "epoch": 0.010351020408163266, + "grad_norm": 2.8612375259399414, + "learning_rate": 9.991280209013717e-06, + "loss": 0.7966, + "step": 317 + }, + { + "epoch": 0.010383673469387756, + "grad_norm": 2.692972183227539, + "learning_rate": 9.99124755062051e-06, + "loss": 0.8288, + "step": 318 + }, + { + "epoch": 0.010416326530612245, + "grad_norm": 2.980607271194458, + "learning_rate": 9.991214892227303e-06, + "loss": 0.8238, + "step": 319 + }, + { + "epoch": 0.010448979591836735, + "grad_norm": 3.0936498641967773, + "learning_rate": 9.991182233834097e-06, + "loss": 0.8338, + "step": 320 + }, + { + "epoch": 0.010481632653061224, + "grad_norm": 2.7482964992523193, + "learning_rate": 9.99114957544089e-06, + "loss": 0.8511, + "step": 321 + }, + { + "epoch": 0.010514285714285714, + "grad_norm": 2.6563820838928223, + "learning_rate": 9.991116917047681e-06, + "loss": 0.8568, + "step": 322 + }, + { + "epoch": 0.010546938775510203, + "grad_norm": 2.7284178733825684, + "learning_rate": 9.991084258654475e-06, + "loss": 0.8615, + "step": 323 + }, + { + "epoch": 0.010579591836734695, + "grad_norm": 2.7720155715942383, + "learning_rate": 9.991051600261268e-06, + "loss": 0.8459, + "step": 324 + }, + { + "epoch": 0.010612244897959184, + "grad_norm": 2.6759071350097656, + "learning_rate": 9.991018941868061e-06, + "loss": 0.8489, + "step": 325 + }, + { + "epoch": 0.010644897959183674, + "grad_norm": 2.8001279830932617, + "learning_rate": 9.990986283474854e-06, + "loss": 0.83, + "step": 326 + }, + { + "epoch": 0.010677551020408163, + "grad_norm": 2.7164299488067627, + "learning_rate": 9.990953625081646e-06, + "loss": 0.8059, + "step": 327 + }, + { + "epoch": 0.010710204081632653, + "grad_norm": 2.833035707473755, + "learning_rate": 9.990920966688439e-06, + "loss": 0.8394, + "step": 328 + }, + { + "epoch": 0.010742857142857143, + "grad_norm": 2.6928770542144775, + "learning_rate": 9.990888308295232e-06, + "loss": 0.8297, + "step": 329 + }, + { + "epoch": 0.010775510204081632, + "grad_norm": 2.7575488090515137, + "learning_rate": 9.990855649902026e-06, + "loss": 0.8149, + "step": 330 + }, + { + "epoch": 0.010808163265306122, + "grad_norm": 2.7353413105010986, + "learning_rate": 9.990822991508819e-06, + "loss": 0.839, + "step": 331 + }, + { + "epoch": 0.010840816326530613, + "grad_norm": 2.7132623195648193, + "learning_rate": 9.990790333115612e-06, + "loss": 0.8452, + "step": 332 + }, + { + "epoch": 0.010873469387755103, + "grad_norm": 2.8830623626708984, + "learning_rate": 9.990757674722404e-06, + "loss": 0.8156, + "step": 333 + }, + { + "epoch": 0.010906122448979592, + "grad_norm": 2.7021992206573486, + "learning_rate": 9.990725016329197e-06, + "loss": 0.8368, + "step": 334 + }, + { + "epoch": 0.010938775510204082, + "grad_norm": 2.6805949211120605, + "learning_rate": 9.99069235793599e-06, + "loss": 0.8402, + "step": 335 + }, + { + "epoch": 0.010971428571428571, + "grad_norm": 3.0374648571014404, + "learning_rate": 9.990659699542783e-06, + "loss": 0.8332, + "step": 336 + }, + { + "epoch": 0.01100408163265306, + "grad_norm": 2.9531807899475098, + "learning_rate": 9.990627041149577e-06, + "loss": 0.8451, + "step": 337 + }, + { + "epoch": 0.01103673469387755, + "grad_norm": 2.8179969787597656, + "learning_rate": 9.99059438275637e-06, + "loss": 0.8336, + "step": 338 + }, + { + "epoch": 0.011069387755102042, + "grad_norm": 3.139448642730713, + "learning_rate": 9.990561724363163e-06, + "loss": 0.865, + "step": 339 + }, + { + "epoch": 0.011102040816326531, + "grad_norm": 3.040363073348999, + "learning_rate": 9.990529065969955e-06, + "loss": 0.8387, + "step": 340 + }, + { + "epoch": 0.01113469387755102, + "grad_norm": 2.700171709060669, + "learning_rate": 9.990496407576748e-06, + "loss": 0.8219, + "step": 341 + }, + { + "epoch": 0.01116734693877551, + "grad_norm": 2.7326748371124268, + "learning_rate": 9.990463749183541e-06, + "loss": 0.7956, + "step": 342 + }, + { + "epoch": 0.0112, + "grad_norm": 2.7744908332824707, + "learning_rate": 9.990431090790334e-06, + "loss": 0.8133, + "step": 343 + }, + { + "epoch": 0.01123265306122449, + "grad_norm": 2.8565139770507812, + "learning_rate": 9.990398432397128e-06, + "loss": 0.825, + "step": 344 + }, + { + "epoch": 0.011265306122448979, + "grad_norm": 2.9909799098968506, + "learning_rate": 9.990365774003919e-06, + "loss": 0.8683, + "step": 345 + }, + { + "epoch": 0.011297959183673469, + "grad_norm": 2.642660140991211, + "learning_rate": 9.990333115610712e-06, + "loss": 0.8598, + "step": 346 + }, + { + "epoch": 0.01133061224489796, + "grad_norm": 2.7017788887023926, + "learning_rate": 9.990300457217506e-06, + "loss": 0.8164, + "step": 347 + }, + { + "epoch": 0.01136326530612245, + "grad_norm": 2.7391109466552734, + "learning_rate": 9.990267798824299e-06, + "loss": 0.8498, + "step": 348 + }, + { + "epoch": 0.011395918367346939, + "grad_norm": 2.7465527057647705, + "learning_rate": 9.990235140431092e-06, + "loss": 0.8225, + "step": 349 + }, + { + "epoch": 0.011428571428571429, + "grad_norm": 2.6583974361419678, + "learning_rate": 9.990202482037884e-06, + "loss": 0.8505, + "step": 350 + }, + { + "epoch": 0.011428571428571429, + "eval_loss": 0.8395382165908813, + "eval_runtime": 73.8611, + "eval_samples_per_second": 1.354, + "eval_steps_per_second": 1.354, + "step": 350 + }, + { + "epoch": 0.011461224489795918, + "grad_norm": 2.7981696128845215, + "learning_rate": 9.990169823644677e-06, + "loss": 0.8312, + "step": 351 + }, + { + "epoch": 0.011493877551020408, + "grad_norm": 2.924536943435669, + "learning_rate": 9.99013716525147e-06, + "loss": 0.8498, + "step": 352 + }, + { + "epoch": 0.011526530612244897, + "grad_norm": 2.819368362426758, + "learning_rate": 9.990104506858263e-06, + "loss": 0.8022, + "step": 353 + }, + { + "epoch": 0.011559183673469389, + "grad_norm": 2.9508090019226074, + "learning_rate": 9.990071848465056e-06, + "loss": 0.8363, + "step": 354 + }, + { + "epoch": 0.011591836734693878, + "grad_norm": 2.643820285797119, + "learning_rate": 9.99003919007185e-06, + "loss": 0.8172, + "step": 355 + }, + { + "epoch": 0.011624489795918368, + "grad_norm": 2.5882935523986816, + "learning_rate": 9.990006531678641e-06, + "loss": 0.8095, + "step": 356 + }, + { + "epoch": 0.011657142857142857, + "grad_norm": 2.968686103820801, + "learning_rate": 9.989973873285436e-06, + "loss": 0.8673, + "step": 357 + }, + { + "epoch": 0.011689795918367347, + "grad_norm": 2.8162810802459717, + "learning_rate": 9.989941214892228e-06, + "loss": 0.8693, + "step": 358 + }, + { + "epoch": 0.011722448979591836, + "grad_norm": 2.8215367794036865, + "learning_rate": 9.989908556499021e-06, + "loss": 0.8053, + "step": 359 + }, + { + "epoch": 0.011755102040816326, + "grad_norm": 2.7967352867126465, + "learning_rate": 9.989875898105814e-06, + "loss": 0.8094, + "step": 360 + }, + { + "epoch": 0.011787755102040815, + "grad_norm": 2.873476505279541, + "learning_rate": 9.989843239712607e-06, + "loss": 0.8174, + "step": 361 + }, + { + "epoch": 0.011820408163265307, + "grad_norm": 2.8655476570129395, + "learning_rate": 9.9898105813194e-06, + "loss": 0.7988, + "step": 362 + }, + { + "epoch": 0.011853061224489796, + "grad_norm": 2.794279098510742, + "learning_rate": 9.989777922926192e-06, + "loss": 0.8404, + "step": 363 + }, + { + "epoch": 0.011885714285714286, + "grad_norm": 2.8534135818481445, + "learning_rate": 9.989745264532985e-06, + "loss": 0.8474, + "step": 364 + }, + { + "epoch": 0.011918367346938775, + "grad_norm": 2.716606855392456, + "learning_rate": 9.989712606139779e-06, + "loss": 0.8256, + "step": 365 + }, + { + "epoch": 0.011951020408163265, + "grad_norm": 2.8670265674591064, + "learning_rate": 9.989679947746572e-06, + "loss": 0.834, + "step": 366 + }, + { + "epoch": 0.011983673469387755, + "grad_norm": 2.9551408290863037, + "learning_rate": 9.989647289353365e-06, + "loss": 0.8295, + "step": 367 + }, + { + "epoch": 0.012016326530612244, + "grad_norm": 3.209449529647827, + "learning_rate": 9.989614630960157e-06, + "loss": 0.8167, + "step": 368 + }, + { + "epoch": 0.012048979591836735, + "grad_norm": 2.8796558380126953, + "learning_rate": 9.98958197256695e-06, + "loss": 0.8245, + "step": 369 + }, + { + "epoch": 0.012081632653061225, + "grad_norm": 2.9785306453704834, + "learning_rate": 9.989549314173743e-06, + "loss": 0.7946, + "step": 370 + }, + { + "epoch": 0.012114285714285715, + "grad_norm": 2.8382294178009033, + "learning_rate": 9.989516655780536e-06, + "loss": 0.8392, + "step": 371 + }, + { + "epoch": 0.012146938775510204, + "grad_norm": 3.2182369232177734, + "learning_rate": 9.98948399738733e-06, + "loss": 0.8343, + "step": 372 + }, + { + "epoch": 0.012179591836734694, + "grad_norm": 2.9380509853363037, + "learning_rate": 9.989451338994123e-06, + "loss": 0.8417, + "step": 373 + }, + { + "epoch": 0.012212244897959183, + "grad_norm": 2.6060733795166016, + "learning_rate": 9.989418680600914e-06, + "loss": 0.8471, + "step": 374 + }, + { + "epoch": 0.012244897959183673, + "grad_norm": 3.0596020221710205, + "learning_rate": 9.989386022207708e-06, + "loss": 0.8408, + "step": 375 + }, + { + "epoch": 0.012277551020408162, + "grad_norm": 2.91416335105896, + "learning_rate": 9.989353363814501e-06, + "loss": 0.8648, + "step": 376 + }, + { + "epoch": 0.012310204081632654, + "grad_norm": 2.8798558712005615, + "learning_rate": 9.989320705421294e-06, + "loss": 0.7871, + "step": 377 + }, + { + "epoch": 0.012342857142857143, + "grad_norm": 2.9666454792022705, + "learning_rate": 9.989288047028087e-06, + "loss": 0.8461, + "step": 378 + }, + { + "epoch": 0.012375510204081633, + "grad_norm": 2.9145092964172363, + "learning_rate": 9.98925538863488e-06, + "loss": 0.8288, + "step": 379 + }, + { + "epoch": 0.012408163265306122, + "grad_norm": 2.844569444656372, + "learning_rate": 9.989222730241674e-06, + "loss": 0.8685, + "step": 380 + }, + { + "epoch": 0.012440816326530612, + "grad_norm": 2.790707588195801, + "learning_rate": 9.989190071848465e-06, + "loss": 0.8467, + "step": 381 + }, + { + "epoch": 0.012473469387755102, + "grad_norm": 2.7228894233703613, + "learning_rate": 9.989157413455259e-06, + "loss": 0.8149, + "step": 382 + }, + { + "epoch": 0.012506122448979591, + "grad_norm": 2.8930270671844482, + "learning_rate": 9.989124755062052e-06, + "loss": 0.8521, + "step": 383 + }, + { + "epoch": 0.012538775510204082, + "grad_norm": 2.7201409339904785, + "learning_rate": 9.989092096668845e-06, + "loss": 0.8303, + "step": 384 + }, + { + "epoch": 0.012571428571428572, + "grad_norm": 2.558375597000122, + "learning_rate": 9.989059438275638e-06, + "loss": 0.8165, + "step": 385 + }, + { + "epoch": 0.012604081632653062, + "grad_norm": 2.893265724182129, + "learning_rate": 9.98902677988243e-06, + "loss": 0.833, + "step": 386 + }, + { + "epoch": 0.012636734693877551, + "grad_norm": 2.6960701942443848, + "learning_rate": 9.988994121489223e-06, + "loss": 0.7972, + "step": 387 + }, + { + "epoch": 0.01266938775510204, + "grad_norm": 2.5734570026397705, + "learning_rate": 9.988961463096016e-06, + "loss": 0.8748, + "step": 388 + }, + { + "epoch": 0.01270204081632653, + "grad_norm": 2.993227243423462, + "learning_rate": 9.98892880470281e-06, + "loss": 0.8405, + "step": 389 + }, + { + "epoch": 0.01273469387755102, + "grad_norm": 3.2019479274749756, + "learning_rate": 9.988896146309603e-06, + "loss": 0.8423, + "step": 390 + }, + { + "epoch": 0.012767346938775511, + "grad_norm": 3.2083091735839844, + "learning_rate": 9.988863487916394e-06, + "loss": 0.8088, + "step": 391 + }, + { + "epoch": 0.0128, + "grad_norm": 3.1379964351654053, + "learning_rate": 9.988830829523188e-06, + "loss": 0.8681, + "step": 392 + }, + { + "epoch": 0.01283265306122449, + "grad_norm": 2.7037453651428223, + "learning_rate": 9.98879817112998e-06, + "loss": 0.8379, + "step": 393 + }, + { + "epoch": 0.01286530612244898, + "grad_norm": 2.828984022140503, + "learning_rate": 9.988765512736774e-06, + "loss": 0.8148, + "step": 394 + }, + { + "epoch": 0.01289795918367347, + "grad_norm": 2.9447624683380127, + "learning_rate": 9.988732854343567e-06, + "loss": 0.8263, + "step": 395 + }, + { + "epoch": 0.012930612244897959, + "grad_norm": 2.7229952812194824, + "learning_rate": 9.98870019595036e-06, + "loss": 0.8613, + "step": 396 + }, + { + "epoch": 0.012963265306122448, + "grad_norm": 2.820530652999878, + "learning_rate": 9.988667537557152e-06, + "loss": 0.8174, + "step": 397 + }, + { + "epoch": 0.012995918367346938, + "grad_norm": 2.897592067718506, + "learning_rate": 9.988634879163947e-06, + "loss": 0.8416, + "step": 398 + }, + { + "epoch": 0.01302857142857143, + "grad_norm": 3.032390832901001, + "learning_rate": 9.988602220770739e-06, + "loss": 0.842, + "step": 399 + }, + { + "epoch": 0.013061224489795919, + "grad_norm": 2.949514389038086, + "learning_rate": 9.988569562377532e-06, + "loss": 0.8283, + "step": 400 + }, + { + "epoch": 0.013061224489795919, + "eval_loss": 0.8356085419654846, + "eval_runtime": 76.0573, + "eval_samples_per_second": 1.315, + "eval_steps_per_second": 1.315, + "step": 400 + }, + { + "epoch": 0.013093877551020408, + "grad_norm": 2.811276912689209, + "learning_rate": 9.988536903984325e-06, + "loss": 0.7906, + "step": 401 + }, + { + "epoch": 0.013126530612244898, + "grad_norm": 2.9961533546447754, + "learning_rate": 9.988504245591118e-06, + "loss": 0.839, + "step": 402 + }, + { + "epoch": 0.013159183673469388, + "grad_norm": 3.0735530853271484, + "learning_rate": 9.988471587197911e-06, + "loss": 0.8053, + "step": 403 + }, + { + "epoch": 0.013191836734693877, + "grad_norm": 2.793745279312134, + "learning_rate": 9.988438928804703e-06, + "loss": 0.8231, + "step": 404 + }, + { + "epoch": 0.013224489795918367, + "grad_norm": 3.1552298069000244, + "learning_rate": 9.988406270411496e-06, + "loss": 0.8328, + "step": 405 + }, + { + "epoch": 0.013257142857142858, + "grad_norm": 2.984865427017212, + "learning_rate": 9.98837361201829e-06, + "loss": 0.811, + "step": 406 + }, + { + "epoch": 0.013289795918367348, + "grad_norm": 2.8718035221099854, + "learning_rate": 9.988340953625083e-06, + "loss": 0.853, + "step": 407 + }, + { + "epoch": 0.013322448979591837, + "grad_norm": 3.0865938663482666, + "learning_rate": 9.988308295231876e-06, + "loss": 0.8003, + "step": 408 + }, + { + "epoch": 0.013355102040816327, + "grad_norm": 2.7381820678710938, + "learning_rate": 9.988275636838668e-06, + "loss": 0.8184, + "step": 409 + }, + { + "epoch": 0.013387755102040816, + "grad_norm": 2.8170037269592285, + "learning_rate": 9.98824297844546e-06, + "loss": 0.8311, + "step": 410 + }, + { + "epoch": 0.013420408163265306, + "grad_norm": 2.9140095710754395, + "learning_rate": 9.988210320052254e-06, + "loss": 0.834, + "step": 411 + }, + { + "epoch": 0.013453061224489795, + "grad_norm": 2.933525800704956, + "learning_rate": 9.988177661659047e-06, + "loss": 0.8294, + "step": 412 + }, + { + "epoch": 0.013485714285714285, + "grad_norm": 2.8974623680114746, + "learning_rate": 9.98814500326584e-06, + "loss": 0.8704, + "step": 413 + }, + { + "epoch": 0.013518367346938776, + "grad_norm": 2.794966697692871, + "learning_rate": 9.988112344872634e-06, + "loss": 0.8579, + "step": 414 + }, + { + "epoch": 0.013551020408163266, + "grad_norm": 2.856208324432373, + "learning_rate": 9.988079686479425e-06, + "loss": 0.8239, + "step": 415 + }, + { + "epoch": 0.013583673469387755, + "grad_norm": 2.999901533126831, + "learning_rate": 9.988047028086218e-06, + "loss": 0.8166, + "step": 416 + }, + { + "epoch": 0.013616326530612245, + "grad_norm": 2.8142268657684326, + "learning_rate": 9.988014369693012e-06, + "loss": 0.8343, + "step": 417 + }, + { + "epoch": 0.013648979591836734, + "grad_norm": 2.808567523956299, + "learning_rate": 9.987981711299805e-06, + "loss": 0.8414, + "step": 418 + }, + { + "epoch": 0.013681632653061224, + "grad_norm": 2.6575918197631836, + "learning_rate": 9.987949052906598e-06, + "loss": 0.8256, + "step": 419 + }, + { + "epoch": 0.013714285714285714, + "grad_norm": 2.9415652751922607, + "learning_rate": 9.987916394513391e-06, + "loss": 0.811, + "step": 420 + }, + { + "epoch": 0.013746938775510205, + "grad_norm": 2.7710697650909424, + "learning_rate": 9.987883736120185e-06, + "loss": 0.8132, + "step": 421 + }, + { + "epoch": 0.013779591836734694, + "grad_norm": 2.6907401084899902, + "learning_rate": 9.987851077726976e-06, + "loss": 0.8022, + "step": 422 + }, + { + "epoch": 0.013812244897959184, + "grad_norm": 3.010432481765747, + "learning_rate": 9.98781841933377e-06, + "loss": 0.8196, + "step": 423 + }, + { + "epoch": 0.013844897959183674, + "grad_norm": 2.868598461151123, + "learning_rate": 9.987785760940563e-06, + "loss": 0.8234, + "step": 424 + }, + { + "epoch": 0.013877551020408163, + "grad_norm": 2.9146206378936768, + "learning_rate": 9.987753102547356e-06, + "loss": 0.8357, + "step": 425 + }, + { + "epoch": 0.013910204081632653, + "grad_norm": 3.148287534713745, + "learning_rate": 9.987720444154149e-06, + "loss": 0.8168, + "step": 426 + }, + { + "epoch": 0.013942857142857142, + "grad_norm": 3.0107545852661133, + "learning_rate": 9.98768778576094e-06, + "loss": 0.8313, + "step": 427 + }, + { + "epoch": 0.013975510204081632, + "grad_norm": 2.7397043704986572, + "learning_rate": 9.987655127367734e-06, + "loss": 0.825, + "step": 428 + }, + { + "epoch": 0.014008163265306123, + "grad_norm": 2.8118340969085693, + "learning_rate": 9.987622468974527e-06, + "loss": 0.8101, + "step": 429 + }, + { + "epoch": 0.014040816326530613, + "grad_norm": 2.862879514694214, + "learning_rate": 9.98758981058132e-06, + "loss": 0.8504, + "step": 430 + }, + { + "epoch": 0.014073469387755102, + "grad_norm": 3.0519521236419678, + "learning_rate": 9.987557152188114e-06, + "loss": 0.8337, + "step": 431 + }, + { + "epoch": 0.014106122448979592, + "grad_norm": 2.8947372436523438, + "learning_rate": 9.987524493794905e-06, + "loss": 0.8041, + "step": 432 + }, + { + "epoch": 0.014138775510204081, + "grad_norm": 2.7874197959899902, + "learning_rate": 9.987491835401698e-06, + "loss": 0.8143, + "step": 433 + }, + { + "epoch": 0.014171428571428571, + "grad_norm": 3.011812686920166, + "learning_rate": 9.987459177008492e-06, + "loss": 0.7936, + "step": 434 + }, + { + "epoch": 0.01420408163265306, + "grad_norm": 2.8212246894836426, + "learning_rate": 9.987426518615285e-06, + "loss": 0.8028, + "step": 435 + }, + { + "epoch": 0.014236734693877552, + "grad_norm": 2.7385244369506836, + "learning_rate": 9.987393860222078e-06, + "loss": 0.8292, + "step": 436 + }, + { + "epoch": 0.014269387755102041, + "grad_norm": 2.7048816680908203, + "learning_rate": 9.987361201828871e-06, + "loss": 0.7567, + "step": 437 + }, + { + "epoch": 0.014302040816326531, + "grad_norm": 3.001534938812256, + "learning_rate": 9.987328543435663e-06, + "loss": 0.7897, + "step": 438 + }, + { + "epoch": 0.01433469387755102, + "grad_norm": 3.1388514041900635, + "learning_rate": 9.987295885042456e-06, + "loss": 0.8373, + "step": 439 + }, + { + "epoch": 0.01436734693877551, + "grad_norm": 2.902858018875122, + "learning_rate": 9.98726322664925e-06, + "loss": 0.8342, + "step": 440 + }, + { + "epoch": 0.0144, + "grad_norm": 2.871511459350586, + "learning_rate": 9.987230568256043e-06, + "loss": 0.8107, + "step": 441 + }, + { + "epoch": 0.01443265306122449, + "grad_norm": 3.0739245414733887, + "learning_rate": 9.987197909862836e-06, + "loss": 0.8171, + "step": 442 + }, + { + "epoch": 0.014465306122448979, + "grad_norm": 2.9680142402648926, + "learning_rate": 9.987165251469629e-06, + "loss": 0.8492, + "step": 443 + }, + { + "epoch": 0.01449795918367347, + "grad_norm": 2.763535737991333, + "learning_rate": 9.987132593076422e-06, + "loss": 0.8201, + "step": 444 + }, + { + "epoch": 0.01453061224489796, + "grad_norm": 2.8160109519958496, + "learning_rate": 9.987099934683214e-06, + "loss": 0.8175, + "step": 445 + }, + { + "epoch": 0.01456326530612245, + "grad_norm": 2.746919870376587, + "learning_rate": 9.987067276290007e-06, + "loss": 0.8121, + "step": 446 + }, + { + "epoch": 0.014595918367346939, + "grad_norm": 2.9417343139648438, + "learning_rate": 9.9870346178968e-06, + "loss": 0.8211, + "step": 447 + }, + { + "epoch": 0.014628571428571428, + "grad_norm": 2.886810302734375, + "learning_rate": 9.987001959503594e-06, + "loss": 0.7889, + "step": 448 + }, + { + "epoch": 0.014661224489795918, + "grad_norm": 2.89780592918396, + "learning_rate": 9.986969301110387e-06, + "loss": 0.8386, + "step": 449 + }, + { + "epoch": 0.014693877551020407, + "grad_norm": 2.8015475273132324, + "learning_rate": 9.986936642717178e-06, + "loss": 0.8023, + "step": 450 + }, + { + "epoch": 0.014693877551020407, + "eval_loss": 0.83209228515625, + "eval_runtime": 76.371, + "eval_samples_per_second": 1.309, + "eval_steps_per_second": 1.309, + "step": 450 + }, + { + "epoch": 0.014726530612244899, + "grad_norm": 3.0286951065063477, + "learning_rate": 9.986903984323972e-06, + "loss": 0.795, + "step": 451 + }, + { + "epoch": 0.014759183673469388, + "grad_norm": 2.798675060272217, + "learning_rate": 9.986871325930765e-06, + "loss": 0.778, + "step": 452 + }, + { + "epoch": 0.014791836734693878, + "grad_norm": 2.9744062423706055, + "learning_rate": 9.986838667537558e-06, + "loss": 0.8034, + "step": 453 + }, + { + "epoch": 0.014824489795918367, + "grad_norm": 3.131807804107666, + "learning_rate": 9.986806009144351e-06, + "loss": 0.8147, + "step": 454 + }, + { + "epoch": 0.014857142857142857, + "grad_norm": 2.759378671646118, + "learning_rate": 9.986773350751145e-06, + "loss": 0.7798, + "step": 455 + }, + { + "epoch": 0.014889795918367347, + "grad_norm": 2.79740047454834, + "learning_rate": 9.986740692357936e-06, + "loss": 0.793, + "step": 456 + }, + { + "epoch": 0.014922448979591836, + "grad_norm": 3.367506742477417, + "learning_rate": 9.98670803396473e-06, + "loss": 0.8295, + "step": 457 + }, + { + "epoch": 0.014955102040816327, + "grad_norm": 2.965843677520752, + "learning_rate": 9.986675375571523e-06, + "loss": 0.7908, + "step": 458 + }, + { + "epoch": 0.014987755102040817, + "grad_norm": 2.815945625305176, + "learning_rate": 9.986642717178316e-06, + "loss": 0.821, + "step": 459 + }, + { + "epoch": 0.015020408163265307, + "grad_norm": 2.8080756664276123, + "learning_rate": 9.986610058785109e-06, + "loss": 0.7952, + "step": 460 + }, + { + "epoch": 0.015053061224489796, + "grad_norm": 3.2081828117370605, + "learning_rate": 9.9865774003919e-06, + "loss": 0.8306, + "step": 461 + }, + { + "epoch": 0.015085714285714286, + "grad_norm": 2.866701602935791, + "learning_rate": 9.986544741998695e-06, + "loss": 0.8438, + "step": 462 + }, + { + "epoch": 0.015118367346938775, + "grad_norm": 2.806741952896118, + "learning_rate": 9.986512083605487e-06, + "loss": 0.8082, + "step": 463 + }, + { + "epoch": 0.015151020408163265, + "grad_norm": 2.8352231979370117, + "learning_rate": 9.98647942521228e-06, + "loss": 0.7805, + "step": 464 + }, + { + "epoch": 0.015183673469387754, + "grad_norm": 2.838469982147217, + "learning_rate": 9.986446766819073e-06, + "loss": 0.844, + "step": 465 + }, + { + "epoch": 0.015216326530612246, + "grad_norm": 2.7868754863739014, + "learning_rate": 9.986414108425867e-06, + "loss": 0.8078, + "step": 466 + }, + { + "epoch": 0.015248979591836735, + "grad_norm": 2.809109687805176, + "learning_rate": 9.98638145003266e-06, + "loss": 0.8425, + "step": 467 + }, + { + "epoch": 0.015281632653061225, + "grad_norm": 2.7377090454101562, + "learning_rate": 9.986348791639451e-06, + "loss": 0.8332, + "step": 468 + }, + { + "epoch": 0.015314285714285714, + "grad_norm": 2.863290786743164, + "learning_rate": 9.986316133246245e-06, + "loss": 0.8352, + "step": 469 + }, + { + "epoch": 0.015346938775510204, + "grad_norm": 2.720506429672241, + "learning_rate": 9.986283474853038e-06, + "loss": 0.8114, + "step": 470 + }, + { + "epoch": 0.015379591836734693, + "grad_norm": 2.7052462100982666, + "learning_rate": 9.986250816459831e-06, + "loss": 0.8093, + "step": 471 + }, + { + "epoch": 0.015412244897959183, + "grad_norm": 3.0364644527435303, + "learning_rate": 9.986218158066624e-06, + "loss": 0.8156, + "step": 472 + }, + { + "epoch": 0.015444897959183674, + "grad_norm": 2.9919137954711914, + "learning_rate": 9.986185499673416e-06, + "loss": 0.7906, + "step": 473 + }, + { + "epoch": 0.015477551020408164, + "grad_norm": 3.102023124694824, + "learning_rate": 9.98615284128021e-06, + "loss": 0.8247, + "step": 474 + }, + { + "epoch": 0.015510204081632653, + "grad_norm": 2.8093349933624268, + "learning_rate": 9.986120182887002e-06, + "loss": 0.8339, + "step": 475 + }, + { + "epoch": 0.015542857142857143, + "grad_norm": 2.893361806869507, + "learning_rate": 9.986087524493796e-06, + "loss": 0.7698, + "step": 476 + }, + { + "epoch": 0.015575510204081633, + "grad_norm": 2.813911199569702, + "learning_rate": 9.986054866100589e-06, + "loss": 0.8204, + "step": 477 + }, + { + "epoch": 0.015608163265306122, + "grad_norm": 2.7959954738616943, + "learning_rate": 9.986022207707382e-06, + "loss": 0.8099, + "step": 478 + }, + { + "epoch": 0.015640816326530613, + "grad_norm": 2.637242317199707, + "learning_rate": 9.985989549314174e-06, + "loss": 0.8218, + "step": 479 + }, + { + "epoch": 0.0156734693877551, + "grad_norm": 2.694908380508423, + "learning_rate": 9.985956890920967e-06, + "loss": 0.843, + "step": 480 + }, + { + "epoch": 0.015706122448979593, + "grad_norm": 2.65104603767395, + "learning_rate": 9.98592423252776e-06, + "loss": 0.8142, + "step": 481 + }, + { + "epoch": 0.01573877551020408, + "grad_norm": 2.7440900802612305, + "learning_rate": 9.985891574134553e-06, + "loss": 0.8079, + "step": 482 + }, + { + "epoch": 0.01577142857142857, + "grad_norm": 2.7440459728240967, + "learning_rate": 9.985858915741347e-06, + "loss": 0.8106, + "step": 483 + }, + { + "epoch": 0.015804081632653063, + "grad_norm": 2.8990628719329834, + "learning_rate": 9.98582625734814e-06, + "loss": 0.7954, + "step": 484 + }, + { + "epoch": 0.01583673469387755, + "grad_norm": 2.8682186603546143, + "learning_rate": 9.985793598954933e-06, + "loss": 0.7892, + "step": 485 + }, + { + "epoch": 0.015869387755102042, + "grad_norm": 3.1332759857177734, + "learning_rate": 9.985760940561725e-06, + "loss": 0.8182, + "step": 486 + }, + { + "epoch": 0.01590204081632653, + "grad_norm": 2.9516282081604004, + "learning_rate": 9.985728282168518e-06, + "loss": 0.7985, + "step": 487 + }, + { + "epoch": 0.01593469387755102, + "grad_norm": 3.3294899463653564, + "learning_rate": 9.985695623775311e-06, + "loss": 0.8368, + "step": 488 + }, + { + "epoch": 0.01596734693877551, + "grad_norm": 2.95438814163208, + "learning_rate": 9.985662965382104e-06, + "loss": 0.8151, + "step": 489 + }, + { + "epoch": 0.016, + "grad_norm": 2.7308006286621094, + "learning_rate": 9.985630306988898e-06, + "loss": 0.827, + "step": 490 + }, + { + "epoch": 0.016032653061224488, + "grad_norm": 2.695218801498413, + "learning_rate": 9.985597648595689e-06, + "loss": 0.7846, + "step": 491 + }, + { + "epoch": 0.01606530612244898, + "grad_norm": 2.842834949493408, + "learning_rate": 9.985564990202482e-06, + "loss": 0.783, + "step": 492 + }, + { + "epoch": 0.01609795918367347, + "grad_norm": 2.944671869277954, + "learning_rate": 9.985532331809276e-06, + "loss": 0.8273, + "step": 493 + }, + { + "epoch": 0.01613061224489796, + "grad_norm": 2.8893423080444336, + "learning_rate": 9.985499673416069e-06, + "loss": 0.8131, + "step": 494 + }, + { + "epoch": 0.01616326530612245, + "grad_norm": 2.917097568511963, + "learning_rate": 9.985467015022862e-06, + "loss": 0.8491, + "step": 495 + }, + { + "epoch": 0.016195918367346938, + "grad_norm": 2.678619384765625, + "learning_rate": 9.985434356629654e-06, + "loss": 0.7789, + "step": 496 + }, + { + "epoch": 0.01622857142857143, + "grad_norm": 2.86249041557312, + "learning_rate": 9.985401698236447e-06, + "loss": 0.8461, + "step": 497 + }, + { + "epoch": 0.016261224489795917, + "grad_norm": 2.9272475242614746, + "learning_rate": 9.98536903984324e-06, + "loss": 0.8047, + "step": 498 + }, + { + "epoch": 0.016293877551020408, + "grad_norm": 2.894817590713501, + "learning_rate": 9.985336381450033e-06, + "loss": 0.8157, + "step": 499 + }, + { + "epoch": 0.0163265306122449, + "grad_norm": 2.9469430446624756, + "learning_rate": 9.985303723056827e-06, + "loss": 0.8409, + "step": 500 + }, + { + "epoch": 0.0163265306122449, + "eval_loss": 0.8293061852455139, + "eval_runtime": 76.9824, + "eval_samples_per_second": 1.299, + "eval_steps_per_second": 1.299, + "step": 500 + }, + { + "epoch": 0.016359183673469387, + "grad_norm": 2.7225022315979004, + "learning_rate": 9.98527106466362e-06, + "loss": 0.7978, + "step": 501 + }, + { + "epoch": 0.01639183673469388, + "grad_norm": 2.6931450366973877, + "learning_rate": 9.985238406270411e-06, + "loss": 0.8089, + "step": 502 + }, + { + "epoch": 0.016424489795918366, + "grad_norm": 2.697438955307007, + "learning_rate": 9.985205747877206e-06, + "loss": 0.8231, + "step": 503 + }, + { + "epoch": 0.016457142857142858, + "grad_norm": 2.997002124786377, + "learning_rate": 9.985173089483998e-06, + "loss": 0.8211, + "step": 504 + }, + { + "epoch": 0.016489795918367346, + "grad_norm": 2.952876567840576, + "learning_rate": 9.985140431090791e-06, + "loss": 0.786, + "step": 505 + }, + { + "epoch": 0.016522448979591837, + "grad_norm": 3.085167407989502, + "learning_rate": 9.985107772697584e-06, + "loss": 0.8261, + "step": 506 + }, + { + "epoch": 0.016555102040816328, + "grad_norm": 2.8937699794769287, + "learning_rate": 9.985075114304378e-06, + "loss": 0.8009, + "step": 507 + }, + { + "epoch": 0.016587755102040816, + "grad_norm": 2.934492588043213, + "learning_rate": 9.98504245591117e-06, + "loss": 0.8059, + "step": 508 + }, + { + "epoch": 0.016620408163265307, + "grad_norm": 2.9281036853790283, + "learning_rate": 9.985009797517962e-06, + "loss": 0.8488, + "step": 509 + }, + { + "epoch": 0.016653061224489795, + "grad_norm": 2.8904941082000732, + "learning_rate": 9.984977139124756e-06, + "loss": 0.8431, + "step": 510 + }, + { + "epoch": 0.016685714285714286, + "grad_norm": 2.7994799613952637, + "learning_rate": 9.984944480731549e-06, + "loss": 0.8339, + "step": 511 + }, + { + "epoch": 0.016718367346938774, + "grad_norm": 3.0456082820892334, + "learning_rate": 9.984911822338342e-06, + "loss": 0.8077, + "step": 512 + }, + { + "epoch": 0.016751020408163265, + "grad_norm": 2.969820261001587, + "learning_rate": 9.984879163945135e-06, + "loss": 0.8254, + "step": 513 + }, + { + "epoch": 0.016783673469387757, + "grad_norm": 2.691143035888672, + "learning_rate": 9.984846505551927e-06, + "loss": 0.8208, + "step": 514 + }, + { + "epoch": 0.016816326530612245, + "grad_norm": 3.0714621543884277, + "learning_rate": 9.98481384715872e-06, + "loss": 0.8094, + "step": 515 + }, + { + "epoch": 0.016848979591836736, + "grad_norm": 3.146872043609619, + "learning_rate": 9.984781188765513e-06, + "loss": 0.8295, + "step": 516 + }, + { + "epoch": 0.016881632653061224, + "grad_norm": 3.0401713848114014, + "learning_rate": 9.984748530372307e-06, + "loss": 0.8438, + "step": 517 + }, + { + "epoch": 0.016914285714285715, + "grad_norm": 3.1565845012664795, + "learning_rate": 9.9847158719791e-06, + "loss": 0.8238, + "step": 518 + }, + { + "epoch": 0.016946938775510203, + "grad_norm": 3.035167932510376, + "learning_rate": 9.984683213585893e-06, + "loss": 0.7907, + "step": 519 + }, + { + "epoch": 0.016979591836734694, + "grad_norm": 2.9173972606658936, + "learning_rate": 9.984650555192685e-06, + "loss": 0.8298, + "step": 520 + }, + { + "epoch": 0.017012244897959182, + "grad_norm": 3.0475914478302, + "learning_rate": 9.984617896799478e-06, + "loss": 0.7896, + "step": 521 + }, + { + "epoch": 0.017044897959183673, + "grad_norm": 2.8846919536590576, + "learning_rate": 9.984585238406271e-06, + "loss": 0.8268, + "step": 522 + }, + { + "epoch": 0.017077551020408165, + "grad_norm": 2.9851925373077393, + "learning_rate": 9.984552580013064e-06, + "loss": 0.8166, + "step": 523 + }, + { + "epoch": 0.017110204081632652, + "grad_norm": 2.960608720779419, + "learning_rate": 9.984519921619857e-06, + "loss": 0.838, + "step": 524 + }, + { + "epoch": 0.017142857142857144, + "grad_norm": 2.8663570880889893, + "learning_rate": 9.984487263226649e-06, + "loss": 0.8134, + "step": 525 + }, + { + "epoch": 0.01717551020408163, + "grad_norm": 3.2871882915496826, + "learning_rate": 9.984454604833444e-06, + "loss": 0.842, + "step": 526 + }, + { + "epoch": 0.017208163265306123, + "grad_norm": 2.9747695922851562, + "learning_rate": 9.984421946440235e-06, + "loss": 0.8371, + "step": 527 + }, + { + "epoch": 0.01724081632653061, + "grad_norm": 3.065366744995117, + "learning_rate": 9.984389288047029e-06, + "loss": 0.799, + "step": 528 + }, + { + "epoch": 0.017273469387755102, + "grad_norm": 3.3719398975372314, + "learning_rate": 9.984356629653822e-06, + "loss": 0.7875, + "step": 529 + }, + { + "epoch": 0.017306122448979593, + "grad_norm": 3.0407257080078125, + "learning_rate": 9.984323971260615e-06, + "loss": 0.833, + "step": 530 + }, + { + "epoch": 0.01733877551020408, + "grad_norm": 2.992464780807495, + "learning_rate": 9.984291312867408e-06, + "loss": 0.83, + "step": 531 + }, + { + "epoch": 0.017371428571428572, + "grad_norm": 2.655125379562378, + "learning_rate": 9.9842586544742e-06, + "loss": 0.8243, + "step": 532 + }, + { + "epoch": 0.01740408163265306, + "grad_norm": 2.7782280445098877, + "learning_rate": 9.984225996080993e-06, + "loss": 0.8279, + "step": 533 + }, + { + "epoch": 0.01743673469387755, + "grad_norm": 2.6449782848358154, + "learning_rate": 9.984193337687786e-06, + "loss": 0.8517, + "step": 534 + }, + { + "epoch": 0.01746938775510204, + "grad_norm": 3.011317491531372, + "learning_rate": 9.98416067929458e-06, + "loss": 0.7763, + "step": 535 + }, + { + "epoch": 0.01750204081632653, + "grad_norm": 2.915745496749878, + "learning_rate": 9.984128020901373e-06, + "loss": 0.8283, + "step": 536 + }, + { + "epoch": 0.017534693877551022, + "grad_norm": 2.7176997661590576, + "learning_rate": 9.984095362508164e-06, + "loss": 0.7846, + "step": 537 + }, + { + "epoch": 0.01756734693877551, + "grad_norm": 2.9343252182006836, + "learning_rate": 9.984062704114958e-06, + "loss": 0.7733, + "step": 538 + }, + { + "epoch": 0.0176, + "grad_norm": 3.041198492050171, + "learning_rate": 9.984030045721751e-06, + "loss": 0.8024, + "step": 539 + }, + { + "epoch": 0.01763265306122449, + "grad_norm": 2.8848764896392822, + "learning_rate": 9.983997387328544e-06, + "loss": 0.8064, + "step": 540 + }, + { + "epoch": 0.01766530612244898, + "grad_norm": 2.847885847091675, + "learning_rate": 9.983964728935337e-06, + "loss": 0.7638, + "step": 541 + }, + { + "epoch": 0.017697959183673468, + "grad_norm": 2.7816402912139893, + "learning_rate": 9.98393207054213e-06, + "loss": 0.7839, + "step": 542 + }, + { + "epoch": 0.01773061224489796, + "grad_norm": 2.8443193435668945, + "learning_rate": 9.983899412148922e-06, + "loss": 0.8005, + "step": 543 + }, + { + "epoch": 0.01776326530612245, + "grad_norm": 2.923306941986084, + "learning_rate": 9.983866753755717e-06, + "loss": 0.8048, + "step": 544 + }, + { + "epoch": 0.01779591836734694, + "grad_norm": 3.0685269832611084, + "learning_rate": 9.983834095362509e-06, + "loss": 0.7676, + "step": 545 + }, + { + "epoch": 0.01782857142857143, + "grad_norm": 2.841459274291992, + "learning_rate": 9.983801436969302e-06, + "loss": 0.7828, + "step": 546 + }, + { + "epoch": 0.017861224489795918, + "grad_norm": 3.1108107566833496, + "learning_rate": 9.983768778576095e-06, + "loss": 0.8287, + "step": 547 + }, + { + "epoch": 0.01789387755102041, + "grad_norm": 2.8307478427886963, + "learning_rate": 9.983736120182888e-06, + "loss": 0.7956, + "step": 548 + }, + { + "epoch": 0.017926530612244897, + "grad_norm": 2.83941650390625, + "learning_rate": 9.983703461789682e-06, + "loss": 0.7963, + "step": 549 + }, + { + "epoch": 0.017959183673469388, + "grad_norm": 2.710853338241577, + "learning_rate": 9.983670803396473e-06, + "loss": 0.8023, + "step": 550 + }, + { + "epoch": 0.017959183673469388, + "eval_loss": 0.8259984850883484, + "eval_runtime": 98.251, + "eval_samples_per_second": 1.018, + "eval_steps_per_second": 1.018, + "step": 550 + }, + { + "epoch": 0.01799183673469388, + "grad_norm": 2.7674920558929443, + "learning_rate": 9.983638145003266e-06, + "loss": 0.8121, + "step": 551 + }, + { + "epoch": 0.018024489795918367, + "grad_norm": 2.756206512451172, + "learning_rate": 9.98360548661006e-06, + "loss": 0.7886, + "step": 552 + }, + { + "epoch": 0.01805714285714286, + "grad_norm": 3.000251054763794, + "learning_rate": 9.983572828216853e-06, + "loss": 0.8181, + "step": 553 + }, + { + "epoch": 0.018089795918367346, + "grad_norm": 2.7608935832977295, + "learning_rate": 9.983540169823646e-06, + "loss": 0.8047, + "step": 554 + }, + { + "epoch": 0.018122448979591838, + "grad_norm": 2.7886579036712646, + "learning_rate": 9.983507511430438e-06, + "loss": 0.8023, + "step": 555 + }, + { + "epoch": 0.018155102040816325, + "grad_norm": 2.7562673091888428, + "learning_rate": 9.983474853037231e-06, + "loss": 0.7935, + "step": 556 + }, + { + "epoch": 0.018187755102040817, + "grad_norm": 2.8150694370269775, + "learning_rate": 9.983442194644024e-06, + "loss": 0.8252, + "step": 557 + }, + { + "epoch": 0.018220408163265304, + "grad_norm": 3.069070339202881, + "learning_rate": 9.983409536250817e-06, + "loss": 0.8352, + "step": 558 + }, + { + "epoch": 0.018253061224489796, + "grad_norm": 3.065143585205078, + "learning_rate": 9.98337687785761e-06, + "loss": 0.8322, + "step": 559 + }, + { + "epoch": 0.018285714285714287, + "grad_norm": 2.808361530303955, + "learning_rate": 9.983344219464404e-06, + "loss": 0.8316, + "step": 560 + }, + { + "epoch": 0.018318367346938775, + "grad_norm": 3.086597204208374, + "learning_rate": 9.983311561071195e-06, + "loss": 0.8063, + "step": 561 + }, + { + "epoch": 0.018351020408163266, + "grad_norm": 2.8510525226593018, + "learning_rate": 9.983278902677989e-06, + "loss": 0.7735, + "step": 562 + }, + { + "epoch": 0.018383673469387754, + "grad_norm": 2.8289899826049805, + "learning_rate": 9.983246244284782e-06, + "loss": 0.831, + "step": 563 + }, + { + "epoch": 0.018416326530612245, + "grad_norm": 2.6663060188293457, + "learning_rate": 9.983213585891575e-06, + "loss": 0.8109, + "step": 564 + }, + { + "epoch": 0.018448979591836733, + "grad_norm": 2.8458051681518555, + "learning_rate": 9.983180927498368e-06, + "loss": 0.8069, + "step": 565 + }, + { + "epoch": 0.018481632653061224, + "grad_norm": 3.2364461421966553, + "learning_rate": 9.98314826910516e-06, + "loss": 0.8214, + "step": 566 + }, + { + "epoch": 0.018514285714285716, + "grad_norm": 3.0581226348876953, + "learning_rate": 9.983115610711955e-06, + "loss": 0.8115, + "step": 567 + }, + { + "epoch": 0.018546938775510204, + "grad_norm": 2.7239277362823486, + "learning_rate": 9.983082952318746e-06, + "loss": 0.8248, + "step": 568 + }, + { + "epoch": 0.018579591836734695, + "grad_norm": 2.827193260192871, + "learning_rate": 9.98305029392554e-06, + "loss": 0.8282, + "step": 569 + }, + { + "epoch": 0.018612244897959183, + "grad_norm": 3.018669605255127, + "learning_rate": 9.983017635532333e-06, + "loss": 0.8196, + "step": 570 + }, + { + "epoch": 0.018644897959183674, + "grad_norm": 2.9022018909454346, + "learning_rate": 9.982984977139126e-06, + "loss": 0.7748, + "step": 571 + }, + { + "epoch": 0.018677551020408162, + "grad_norm": 2.7094552516937256, + "learning_rate": 9.98295231874592e-06, + "loss": 0.8381, + "step": 572 + }, + { + "epoch": 0.018710204081632653, + "grad_norm": 3.020749568939209, + "learning_rate": 9.98291966035271e-06, + "loss": 0.7954, + "step": 573 + }, + { + "epoch": 0.018742857142857144, + "grad_norm": 2.9567782878875732, + "learning_rate": 9.982887001959504e-06, + "loss": 0.8197, + "step": 574 + }, + { + "epoch": 0.018775510204081632, + "grad_norm": 2.939793348312378, + "learning_rate": 9.982854343566297e-06, + "loss": 0.7891, + "step": 575 + }, + { + "epoch": 0.018808163265306124, + "grad_norm": 2.851130485534668, + "learning_rate": 9.98282168517309e-06, + "loss": 0.8669, + "step": 576 + }, + { + "epoch": 0.01884081632653061, + "grad_norm": 2.745495080947876, + "learning_rate": 9.982789026779884e-06, + "loss": 0.8028, + "step": 577 + }, + { + "epoch": 0.018873469387755103, + "grad_norm": 2.83685302734375, + "learning_rate": 9.982756368386675e-06, + "loss": 0.7981, + "step": 578 + }, + { + "epoch": 0.01890612244897959, + "grad_norm": 2.845992088317871, + "learning_rate": 9.982723709993469e-06, + "loss": 0.7752, + "step": 579 + }, + { + "epoch": 0.018938775510204082, + "grad_norm": 3.1280617713928223, + "learning_rate": 9.982691051600262e-06, + "loss": 0.824, + "step": 580 + }, + { + "epoch": 0.018971428571428573, + "grad_norm": 2.8821194171905518, + "learning_rate": 9.982658393207055e-06, + "loss": 0.7925, + "step": 581 + }, + { + "epoch": 0.01900408163265306, + "grad_norm": 2.827406167984009, + "learning_rate": 9.982625734813848e-06, + "loss": 0.8311, + "step": 582 + }, + { + "epoch": 0.019036734693877552, + "grad_norm": 2.7628026008605957, + "learning_rate": 9.982593076420641e-06, + "loss": 0.8228, + "step": 583 + }, + { + "epoch": 0.01906938775510204, + "grad_norm": 2.8712172508239746, + "learning_rate": 9.982560418027433e-06, + "loss": 0.8014, + "step": 584 + }, + { + "epoch": 0.01910204081632653, + "grad_norm": 2.85799241065979, + "learning_rate": 9.982527759634226e-06, + "loss": 0.8281, + "step": 585 + }, + { + "epoch": 0.01913469387755102, + "grad_norm": 2.917358160018921, + "learning_rate": 9.98249510124102e-06, + "loss": 0.8049, + "step": 586 + }, + { + "epoch": 0.01916734693877551, + "grad_norm": 2.8612101078033447, + "learning_rate": 9.982462442847813e-06, + "loss": 0.7645, + "step": 587 + }, + { + "epoch": 0.0192, + "grad_norm": 2.8899152278900146, + "learning_rate": 9.982429784454606e-06, + "loss": 0.7655, + "step": 588 + }, + { + "epoch": 0.01923265306122449, + "grad_norm": 2.761504888534546, + "learning_rate": 9.9823971260614e-06, + "loss": 0.784, + "step": 589 + }, + { + "epoch": 0.01926530612244898, + "grad_norm": 3.1624319553375244, + "learning_rate": 9.982364467668192e-06, + "loss": 0.8538, + "step": 590 + }, + { + "epoch": 0.01929795918367347, + "grad_norm": 2.840982675552368, + "learning_rate": 9.982331809274984e-06, + "loss": 0.7879, + "step": 591 + }, + { + "epoch": 0.01933061224489796, + "grad_norm": 2.8063557147979736, + "learning_rate": 9.982299150881777e-06, + "loss": 0.7858, + "step": 592 + }, + { + "epoch": 0.019363265306122448, + "grad_norm": 2.761817455291748, + "learning_rate": 9.98226649248857e-06, + "loss": 0.7932, + "step": 593 + }, + { + "epoch": 0.01939591836734694, + "grad_norm": 2.845803737640381, + "learning_rate": 9.982233834095364e-06, + "loss": 0.8315, + "step": 594 + }, + { + "epoch": 0.019428571428571427, + "grad_norm": 2.820237398147583, + "learning_rate": 9.982201175702157e-06, + "loss": 0.8115, + "step": 595 + }, + { + "epoch": 0.01946122448979592, + "grad_norm": 2.8963451385498047, + "learning_rate": 9.982168517308948e-06, + "loss": 0.7965, + "step": 596 + }, + { + "epoch": 0.01949387755102041, + "grad_norm": 2.8076157569885254, + "learning_rate": 9.982135858915742e-06, + "loss": 0.8205, + "step": 597 + }, + { + "epoch": 0.019526530612244897, + "grad_norm": 2.9265244007110596, + "learning_rate": 9.982103200522535e-06, + "loss": 0.7975, + "step": 598 + }, + { + "epoch": 0.01955918367346939, + "grad_norm": 3.044168472290039, + "learning_rate": 9.982070542129328e-06, + "loss": 0.7738, + "step": 599 + }, + { + "epoch": 0.019591836734693877, + "grad_norm": 2.991462469100952, + "learning_rate": 9.982037883736121e-06, + "loss": 0.808, + "step": 600 + }, + { + "epoch": 0.019591836734693877, + "eval_loss": 0.8216409087181091, + "eval_runtime": 96.6305, + "eval_samples_per_second": 1.035, + "eval_steps_per_second": 1.035, + "step": 600 + }, + { + "epoch": 0.019624489795918368, + "grad_norm": 2.856177568435669, + "learning_rate": 9.982005225342915e-06, + "loss": 0.8002, + "step": 601 + }, + { + "epoch": 0.019657142857142856, + "grad_norm": 2.881652355194092, + "learning_rate": 9.981972566949706e-06, + "loss": 0.8012, + "step": 602 + }, + { + "epoch": 0.019689795918367347, + "grad_norm": 2.9178857803344727, + "learning_rate": 9.9819399085565e-06, + "loss": 0.7724, + "step": 603 + }, + { + "epoch": 0.019722448979591838, + "grad_norm": 2.7211968898773193, + "learning_rate": 9.981907250163293e-06, + "loss": 0.798, + "step": 604 + }, + { + "epoch": 0.019755102040816326, + "grad_norm": 3.050656795501709, + "learning_rate": 9.981874591770086e-06, + "loss": 0.8564, + "step": 605 + }, + { + "epoch": 0.019787755102040817, + "grad_norm": 2.9224750995635986, + "learning_rate": 9.981841933376879e-06, + "loss": 0.8058, + "step": 606 + }, + { + "epoch": 0.019820408163265305, + "grad_norm": 2.9750068187713623, + "learning_rate": 9.98180927498367e-06, + "loss": 0.7774, + "step": 607 + }, + { + "epoch": 0.019853061224489797, + "grad_norm": 3.0843162536621094, + "learning_rate": 9.981776616590466e-06, + "loss": 0.8017, + "step": 608 + }, + { + "epoch": 0.019885714285714284, + "grad_norm": 3.043372631072998, + "learning_rate": 9.981743958197257e-06, + "loss": 0.7763, + "step": 609 + }, + { + "epoch": 0.019918367346938776, + "grad_norm": 2.8992819786071777, + "learning_rate": 9.98171129980405e-06, + "loss": 0.8483, + "step": 610 + }, + { + "epoch": 0.019951020408163267, + "grad_norm": 2.5121912956237793, + "learning_rate": 9.981678641410844e-06, + "loss": 0.8014, + "step": 611 + }, + { + "epoch": 0.019983673469387755, + "grad_norm": 2.8126091957092285, + "learning_rate": 9.981645983017637e-06, + "loss": 0.8, + "step": 612 + }, + { + "epoch": 0.020016326530612246, + "grad_norm": 2.760281801223755, + "learning_rate": 9.98161332462443e-06, + "loss": 0.8089, + "step": 613 + }, + { + "epoch": 0.020048979591836734, + "grad_norm": 2.75252103805542, + "learning_rate": 9.981580666231222e-06, + "loss": 0.7991, + "step": 614 + }, + { + "epoch": 0.020081632653061225, + "grad_norm": 2.7450368404388428, + "learning_rate": 9.981548007838015e-06, + "loss": 0.7995, + "step": 615 + }, + { + "epoch": 0.020114285714285713, + "grad_norm": 2.752061367034912, + "learning_rate": 9.981515349444808e-06, + "loss": 0.8167, + "step": 616 + }, + { + "epoch": 0.020146938775510204, + "grad_norm": 2.882789134979248, + "learning_rate": 9.981482691051601e-06, + "loss": 0.8438, + "step": 617 + }, + { + "epoch": 0.020179591836734692, + "grad_norm": 2.8994760513305664, + "learning_rate": 9.981450032658395e-06, + "loss": 0.7794, + "step": 618 + }, + { + "epoch": 0.020212244897959183, + "grad_norm": 2.925473928451538, + "learning_rate": 9.981417374265186e-06, + "loss": 0.798, + "step": 619 + }, + { + "epoch": 0.020244897959183675, + "grad_norm": 2.909290075302124, + "learning_rate": 9.98138471587198e-06, + "loss": 0.8487, + "step": 620 + }, + { + "epoch": 0.020277551020408163, + "grad_norm": 3.004021644592285, + "learning_rate": 9.981352057478773e-06, + "loss": 0.8214, + "step": 621 + }, + { + "epoch": 0.020310204081632654, + "grad_norm": 2.850106954574585, + "learning_rate": 9.981319399085566e-06, + "loss": 0.8158, + "step": 622 + }, + { + "epoch": 0.02034285714285714, + "grad_norm": 2.903092622756958, + "learning_rate": 9.981286740692359e-06, + "loss": 0.8078, + "step": 623 + }, + { + "epoch": 0.020375510204081633, + "grad_norm": 2.9145193099975586, + "learning_rate": 9.981254082299152e-06, + "loss": 0.8224, + "step": 624 + }, + { + "epoch": 0.02040816326530612, + "grad_norm": 3.0473721027374268, + "learning_rate": 9.981221423905944e-06, + "loss": 0.7681, + "step": 625 + }, + { + "epoch": 0.020440816326530612, + "grad_norm": 2.8615925312042236, + "learning_rate": 9.981188765512737e-06, + "loss": 0.815, + "step": 626 + }, + { + "epoch": 0.020473469387755103, + "grad_norm": 2.8572824001312256, + "learning_rate": 9.98115610711953e-06, + "loss": 0.7672, + "step": 627 + }, + { + "epoch": 0.02050612244897959, + "grad_norm": 2.9113521575927734, + "learning_rate": 9.981123448726324e-06, + "loss": 0.7952, + "step": 628 + }, + { + "epoch": 0.020538775510204083, + "grad_norm": 3.0004873275756836, + "learning_rate": 9.981090790333117e-06, + "loss": 0.8181, + "step": 629 + }, + { + "epoch": 0.02057142857142857, + "grad_norm": 2.9250986576080322, + "learning_rate": 9.981058131939908e-06, + "loss": 0.7681, + "step": 630 + }, + { + "epoch": 0.02060408163265306, + "grad_norm": 3.1994357109069824, + "learning_rate": 9.981025473546703e-06, + "loss": 0.8292, + "step": 631 + }, + { + "epoch": 0.02063673469387755, + "grad_norm": 3.061445713043213, + "learning_rate": 9.980992815153495e-06, + "loss": 0.8241, + "step": 632 + }, + { + "epoch": 0.02066938775510204, + "grad_norm": 3.1536760330200195, + "learning_rate": 9.980960156760288e-06, + "loss": 0.7889, + "step": 633 + }, + { + "epoch": 0.020702040816326532, + "grad_norm": 2.7358641624450684, + "learning_rate": 9.980927498367081e-06, + "loss": 0.8167, + "step": 634 + }, + { + "epoch": 0.02073469387755102, + "grad_norm": 2.952697992324829, + "learning_rate": 9.980894839973874e-06, + "loss": 0.7741, + "step": 635 + }, + { + "epoch": 0.02076734693877551, + "grad_norm": 3.003535747528076, + "learning_rate": 9.980862181580668e-06, + "loss": 0.8101, + "step": 636 + }, + { + "epoch": 0.0208, + "grad_norm": 2.895209789276123, + "learning_rate": 9.98082952318746e-06, + "loss": 0.8025, + "step": 637 + }, + { + "epoch": 0.02083265306122449, + "grad_norm": 2.974769115447998, + "learning_rate": 9.980796864794252e-06, + "loss": 0.8236, + "step": 638 + }, + { + "epoch": 0.020865306122448978, + "grad_norm": 2.8859550952911377, + "learning_rate": 9.980764206401046e-06, + "loss": 0.8214, + "step": 639 + }, + { + "epoch": 0.02089795918367347, + "grad_norm": 3.1916568279266357, + "learning_rate": 9.980731548007839e-06, + "loss": 0.7618, + "step": 640 + }, + { + "epoch": 0.02093061224489796, + "grad_norm": 3.1790311336517334, + "learning_rate": 9.980698889614632e-06, + "loss": 0.779, + "step": 641 + }, + { + "epoch": 0.02096326530612245, + "grad_norm": 9.082115173339844, + "learning_rate": 9.980666231221424e-06, + "loss": 0.7972, + "step": 642 + }, + { + "epoch": 0.02099591836734694, + "grad_norm": 3.0865111351013184, + "learning_rate": 9.980633572828217e-06, + "loss": 0.7972, + "step": 643 + }, + { + "epoch": 0.021028571428571428, + "grad_norm": 3.3228487968444824, + "learning_rate": 9.98060091443501e-06, + "loss": 0.8245, + "step": 644 + }, + { + "epoch": 0.02106122448979592, + "grad_norm": 2.7116079330444336, + "learning_rate": 9.980568256041803e-06, + "loss": 0.8129, + "step": 645 + }, + { + "epoch": 0.021093877551020407, + "grad_norm": 2.9123730659484863, + "learning_rate": 9.980535597648597e-06, + "loss": 0.832, + "step": 646 + }, + { + "epoch": 0.021126530612244898, + "grad_norm": 3.1061103343963623, + "learning_rate": 9.98050293925539e-06, + "loss": 0.8067, + "step": 647 + }, + { + "epoch": 0.02115918367346939, + "grad_norm": 3.2617485523223877, + "learning_rate": 9.980470280862181e-06, + "loss": 0.7897, + "step": 648 + }, + { + "epoch": 0.021191836734693877, + "grad_norm": 3.0000083446502686, + "learning_rate": 9.980437622468976e-06, + "loss": 0.7663, + "step": 649 + }, + { + "epoch": 0.02122448979591837, + "grad_norm": 3.029799461364746, + "learning_rate": 9.980404964075768e-06, + "loss": 0.7688, + "step": 650 + }, + { + "epoch": 0.02122448979591837, + "eval_loss": 0.8194996118545532, + "eval_runtime": 74.2652, + "eval_samples_per_second": 1.347, + "eval_steps_per_second": 1.347, + "step": 650 + }, + { + "epoch": 0.021257142857142856, + "grad_norm": 2.762091875076294, + "learning_rate": 9.980372305682561e-06, + "loss": 0.8051, + "step": 651 + }, + { + "epoch": 0.021289795918367348, + "grad_norm": 3.030320167541504, + "learning_rate": 9.980339647289354e-06, + "loss": 0.8201, + "step": 652 + }, + { + "epoch": 0.021322448979591836, + "grad_norm": 2.9875736236572266, + "learning_rate": 9.980306988896148e-06, + "loss": 0.7841, + "step": 653 + }, + { + "epoch": 0.021355102040816327, + "grad_norm": 2.847587823867798, + "learning_rate": 9.980274330502941e-06, + "loss": 0.7807, + "step": 654 + }, + { + "epoch": 0.021387755102040815, + "grad_norm": 3.0852084159851074, + "learning_rate": 9.980241672109732e-06, + "loss": 0.7606, + "step": 655 + }, + { + "epoch": 0.021420408163265306, + "grad_norm": 3.159208297729492, + "learning_rate": 9.980209013716526e-06, + "loss": 0.8011, + "step": 656 + }, + { + "epoch": 0.021453061224489797, + "grad_norm": 3.063185214996338, + "learning_rate": 9.980176355323319e-06, + "loss": 0.7839, + "step": 657 + }, + { + "epoch": 0.021485714285714285, + "grad_norm": 2.9843311309814453, + "learning_rate": 9.980143696930112e-06, + "loss": 0.7961, + "step": 658 + }, + { + "epoch": 0.021518367346938776, + "grad_norm": 3.0662293434143066, + "learning_rate": 9.980111038536905e-06, + "loss": 0.8065, + "step": 659 + }, + { + "epoch": 0.021551020408163264, + "grad_norm": 2.9915835857391357, + "learning_rate": 9.980078380143697e-06, + "loss": 0.8022, + "step": 660 + }, + { + "epoch": 0.021583673469387755, + "grad_norm": 2.7856531143188477, + "learning_rate": 9.98004572175049e-06, + "loss": 0.7853, + "step": 661 + }, + { + "epoch": 0.021616326530612243, + "grad_norm": 2.8300893306732178, + "learning_rate": 9.980013063357283e-06, + "loss": 0.8097, + "step": 662 + }, + { + "epoch": 0.021648979591836735, + "grad_norm": 2.7422966957092285, + "learning_rate": 9.979980404964077e-06, + "loss": 0.8307, + "step": 663 + }, + { + "epoch": 0.021681632653061226, + "grad_norm": 2.80092453956604, + "learning_rate": 9.97994774657087e-06, + "loss": 0.7932, + "step": 664 + }, + { + "epoch": 0.021714285714285714, + "grad_norm": 2.8894505500793457, + "learning_rate": 9.979915088177663e-06, + "loss": 0.8042, + "step": 665 + }, + { + "epoch": 0.021746938775510205, + "grad_norm": 2.7492778301239014, + "learning_rate": 9.979882429784455e-06, + "loss": 0.7817, + "step": 666 + }, + { + "epoch": 0.021779591836734693, + "grad_norm": 2.734226703643799, + "learning_rate": 9.979849771391248e-06, + "loss": 0.7984, + "step": 667 + }, + { + "epoch": 0.021812244897959184, + "grad_norm": 2.7127978801727295, + "learning_rate": 9.979817112998041e-06, + "loss": 0.7955, + "step": 668 + }, + { + "epoch": 0.021844897959183672, + "grad_norm": 2.881192207336426, + "learning_rate": 9.979784454604834e-06, + "loss": 0.8257, + "step": 669 + }, + { + "epoch": 0.021877551020408163, + "grad_norm": 3.02278995513916, + "learning_rate": 9.979751796211628e-06, + "loss": 0.7892, + "step": 670 + }, + { + "epoch": 0.021910204081632655, + "grad_norm": 3.157317876815796, + "learning_rate": 9.979719137818419e-06, + "loss": 0.8227, + "step": 671 + }, + { + "epoch": 0.021942857142857142, + "grad_norm": 2.876741409301758, + "learning_rate": 9.979686479425214e-06, + "loss": 0.8116, + "step": 672 + }, + { + "epoch": 0.021975510204081634, + "grad_norm": 3.1728057861328125, + "learning_rate": 9.979653821032006e-06, + "loss": 0.8207, + "step": 673 + }, + { + "epoch": 0.02200816326530612, + "grad_norm": 3.2653119564056396, + "learning_rate": 9.979621162638799e-06, + "loss": 0.8087, + "step": 674 + }, + { + "epoch": 0.022040816326530613, + "grad_norm": 3.0209734439849854, + "learning_rate": 9.979588504245592e-06, + "loss": 0.7958, + "step": 675 + }, + { + "epoch": 0.0220734693877551, + "grad_norm": 2.9444637298583984, + "learning_rate": 9.979555845852385e-06, + "loss": 0.8031, + "step": 676 + }, + { + "epoch": 0.022106122448979592, + "grad_norm": 2.8222014904022217, + "learning_rate": 9.979523187459179e-06, + "loss": 0.7842, + "step": 677 + }, + { + "epoch": 0.022138775510204083, + "grad_norm": 3.034346342086792, + "learning_rate": 9.97949052906597e-06, + "loss": 0.7979, + "step": 678 + }, + { + "epoch": 0.02217142857142857, + "grad_norm": 2.9454312324523926, + "learning_rate": 9.979457870672763e-06, + "loss": 0.8209, + "step": 679 + }, + { + "epoch": 0.022204081632653062, + "grad_norm": 2.9699203968048096, + "learning_rate": 9.979425212279557e-06, + "loss": 0.7783, + "step": 680 + }, + { + "epoch": 0.02223673469387755, + "grad_norm": 3.0586094856262207, + "learning_rate": 9.97939255388635e-06, + "loss": 0.8257, + "step": 681 + }, + { + "epoch": 0.02226938775510204, + "grad_norm": 3.1194658279418945, + "learning_rate": 9.979359895493143e-06, + "loss": 0.7824, + "step": 682 + }, + { + "epoch": 0.02230204081632653, + "grad_norm": 2.950078248977661, + "learning_rate": 9.979327237099935e-06, + "loss": 0.7866, + "step": 683 + }, + { + "epoch": 0.02233469387755102, + "grad_norm": 3.135443925857544, + "learning_rate": 9.979294578706728e-06, + "loss": 0.786, + "step": 684 + }, + { + "epoch": 0.02236734693877551, + "grad_norm": 3.1549785137176514, + "learning_rate": 9.979261920313521e-06, + "loss": 0.8063, + "step": 685 + }, + { + "epoch": 0.0224, + "grad_norm": 3.064702033996582, + "learning_rate": 9.979229261920314e-06, + "loss": 0.7789, + "step": 686 + }, + { + "epoch": 0.02243265306122449, + "grad_norm": 2.9797706604003906, + "learning_rate": 9.979196603527108e-06, + "loss": 0.8022, + "step": 687 + }, + { + "epoch": 0.02246530612244898, + "grad_norm": 3.0248265266418457, + "learning_rate": 9.9791639451339e-06, + "loss": 0.7935, + "step": 688 + }, + { + "epoch": 0.02249795918367347, + "grad_norm": 2.865000009536743, + "learning_rate": 9.979131286740692e-06, + "loss": 0.7539, + "step": 689 + }, + { + "epoch": 0.022530612244897958, + "grad_norm": 2.7899985313415527, + "learning_rate": 9.979098628347487e-06, + "loss": 0.766, + "step": 690 + }, + { + "epoch": 0.02256326530612245, + "grad_norm": 3.016523838043213, + "learning_rate": 9.979065969954279e-06, + "loss": 0.7904, + "step": 691 + }, + { + "epoch": 0.022595918367346937, + "grad_norm": 2.954990863800049, + "learning_rate": 9.979033311561072e-06, + "loss": 0.8276, + "step": 692 + }, + { + "epoch": 0.02262857142857143, + "grad_norm": 2.932527780532837, + "learning_rate": 9.979000653167865e-06, + "loss": 0.7614, + "step": 693 + }, + { + "epoch": 0.02266122448979592, + "grad_norm": 2.892082691192627, + "learning_rate": 9.978967994774657e-06, + "loss": 0.8026, + "step": 694 + }, + { + "epoch": 0.022693877551020408, + "grad_norm": 2.8288159370422363, + "learning_rate": 9.978935336381452e-06, + "loss": 0.7956, + "step": 695 + }, + { + "epoch": 0.0227265306122449, + "grad_norm": 2.8160433769226074, + "learning_rate": 9.978902677988243e-06, + "loss": 0.7765, + "step": 696 + }, + { + "epoch": 0.022759183673469387, + "grad_norm": 2.8383493423461914, + "learning_rate": 9.978870019595036e-06, + "loss": 0.7926, + "step": 697 + }, + { + "epoch": 0.022791836734693878, + "grad_norm": 3.058093786239624, + "learning_rate": 9.97883736120183e-06, + "loss": 0.7878, + "step": 698 + }, + { + "epoch": 0.022824489795918366, + "grad_norm": 3.0352556705474854, + "learning_rate": 9.978804702808623e-06, + "loss": 0.7885, + "step": 699 + }, + { + "epoch": 0.022857142857142857, + "grad_norm": 2.9252421855926514, + "learning_rate": 9.978772044415416e-06, + "loss": 0.7759, + "step": 700 + }, + { + "epoch": 0.022857142857142857, + "eval_loss": 0.8187767267227173, + "eval_runtime": 74.1789, + "eval_samples_per_second": 1.348, + "eval_steps_per_second": 1.348, + "step": 700 + }, + { + "epoch": 0.02288979591836735, + "grad_norm": 2.860393762588501, + "learning_rate": 9.978739386022208e-06, + "loss": 0.8177, + "step": 701 + }, + { + "epoch": 0.022922448979591836, + "grad_norm": 2.960848331451416, + "learning_rate": 9.978706727629001e-06, + "loss": 0.8298, + "step": 702 + }, + { + "epoch": 0.022955102040816328, + "grad_norm": 2.8897740840911865, + "learning_rate": 9.978674069235794e-06, + "loss": 0.7625, + "step": 703 + }, + { + "epoch": 0.022987755102040815, + "grad_norm": 3.0318431854248047, + "learning_rate": 9.978641410842587e-06, + "loss": 0.7759, + "step": 704 + }, + { + "epoch": 0.023020408163265307, + "grad_norm": 2.959118366241455, + "learning_rate": 9.97860875244938e-06, + "loss": 0.7921, + "step": 705 + }, + { + "epoch": 0.023053061224489795, + "grad_norm": 2.871290922164917, + "learning_rate": 9.978576094056174e-06, + "loss": 0.7943, + "step": 706 + }, + { + "epoch": 0.023085714285714286, + "grad_norm": 2.699939489364624, + "learning_rate": 9.978543435662965e-06, + "loss": 0.7708, + "step": 707 + }, + { + "epoch": 0.023118367346938777, + "grad_norm": 2.8150787353515625, + "learning_rate": 9.978510777269759e-06, + "loss": 0.7888, + "step": 708 + }, + { + "epoch": 0.023151020408163265, + "grad_norm": 2.9636101722717285, + "learning_rate": 9.978478118876552e-06, + "loss": 0.7861, + "step": 709 + }, + { + "epoch": 0.023183673469387756, + "grad_norm": 3.122624397277832, + "learning_rate": 9.978445460483345e-06, + "loss": 0.7967, + "step": 710 + }, + { + "epoch": 0.023216326530612244, + "grad_norm": 3.070082902908325, + "learning_rate": 9.978412802090138e-06, + "loss": 0.793, + "step": 711 + }, + { + "epoch": 0.023248979591836735, + "grad_norm": 3.063530206680298, + "learning_rate": 9.97838014369693e-06, + "loss": 0.8115, + "step": 712 + }, + { + "epoch": 0.023281632653061223, + "grad_norm": 2.8320116996765137, + "learning_rate": 9.978347485303725e-06, + "loss": 0.831, + "step": 713 + }, + { + "epoch": 0.023314285714285714, + "grad_norm": 2.9792158603668213, + "learning_rate": 9.978314826910516e-06, + "loss": 0.745, + "step": 714 + }, + { + "epoch": 0.023346938775510206, + "grad_norm": 2.9203131198883057, + "learning_rate": 9.97828216851731e-06, + "loss": 0.7912, + "step": 715 + }, + { + "epoch": 0.023379591836734694, + "grad_norm": 3.151167631149292, + "learning_rate": 9.978249510124103e-06, + "loss": 0.7687, + "step": 716 + }, + { + "epoch": 0.023412244897959185, + "grad_norm": 2.8722281455993652, + "learning_rate": 9.978216851730896e-06, + "loss": 0.7975, + "step": 717 + }, + { + "epoch": 0.023444897959183673, + "grad_norm": 2.989642858505249, + "learning_rate": 9.97818419333769e-06, + "loss": 0.8136, + "step": 718 + }, + { + "epoch": 0.023477551020408164, + "grad_norm": 3.2293035984039307, + "learning_rate": 9.978151534944481e-06, + "loss": 0.8469, + "step": 719 + }, + { + "epoch": 0.023510204081632652, + "grad_norm": 3.016899585723877, + "learning_rate": 9.978118876551274e-06, + "loss": 0.7901, + "step": 720 + }, + { + "epoch": 0.023542857142857143, + "grad_norm": 3.194211006164551, + "learning_rate": 9.978086218158067e-06, + "loss": 0.8112, + "step": 721 + }, + { + "epoch": 0.02357551020408163, + "grad_norm": 2.810654640197754, + "learning_rate": 9.97805355976486e-06, + "loss": 0.7878, + "step": 722 + }, + { + "epoch": 0.023608163265306122, + "grad_norm": 2.788594961166382, + "learning_rate": 9.978020901371654e-06, + "loss": 0.8245, + "step": 723 + }, + { + "epoch": 0.023640816326530614, + "grad_norm": 2.976698875427246, + "learning_rate": 9.977988242978445e-06, + "loss": 0.7869, + "step": 724 + }, + { + "epoch": 0.0236734693877551, + "grad_norm": 2.7467331886291504, + "learning_rate": 9.977955584585239e-06, + "loss": 0.7837, + "step": 725 + }, + { + "epoch": 0.023706122448979593, + "grad_norm": 2.8149194717407227, + "learning_rate": 9.977922926192032e-06, + "loss": 0.7713, + "step": 726 + }, + { + "epoch": 0.02373877551020408, + "grad_norm": 3.07743501663208, + "learning_rate": 9.977890267798825e-06, + "loss": 0.7519, + "step": 727 + }, + { + "epoch": 0.023771428571428572, + "grad_norm": 2.779167413711548, + "learning_rate": 9.977857609405618e-06, + "loss": 0.8015, + "step": 728 + }, + { + "epoch": 0.02380408163265306, + "grad_norm": 2.861452579498291, + "learning_rate": 9.977824951012412e-06, + "loss": 0.8097, + "step": 729 + }, + { + "epoch": 0.02383673469387755, + "grad_norm": 2.9489521980285645, + "learning_rate": 9.977792292619203e-06, + "loss": 0.7988, + "step": 730 + }, + { + "epoch": 0.023869387755102042, + "grad_norm": 2.7786002159118652, + "learning_rate": 9.977759634225996e-06, + "loss": 0.7828, + "step": 731 + }, + { + "epoch": 0.02390204081632653, + "grad_norm": 2.9412615299224854, + "learning_rate": 9.97772697583279e-06, + "loss": 0.7684, + "step": 732 + }, + { + "epoch": 0.02393469387755102, + "grad_norm": 2.9088659286499023, + "learning_rate": 9.977694317439583e-06, + "loss": 0.7899, + "step": 733 + }, + { + "epoch": 0.02396734693877551, + "grad_norm": 2.811553955078125, + "learning_rate": 9.977661659046376e-06, + "loss": 0.7705, + "step": 734 + }, + { + "epoch": 0.024, + "grad_norm": 2.7078611850738525, + "learning_rate": 9.977629000653168e-06, + "loss": 0.7803, + "step": 735 + }, + { + "epoch": 0.02403265306122449, + "grad_norm": 2.9198622703552246, + "learning_rate": 9.977596342259963e-06, + "loss": 0.794, + "step": 736 + }, + { + "epoch": 0.02406530612244898, + "grad_norm": 3.1856372356414795, + "learning_rate": 9.977563683866754e-06, + "loss": 0.8146, + "step": 737 + }, + { + "epoch": 0.02409795918367347, + "grad_norm": 2.914483070373535, + "learning_rate": 9.977531025473547e-06, + "loss": 0.813, + "step": 738 + }, + { + "epoch": 0.02413061224489796, + "grad_norm": 2.837502956390381, + "learning_rate": 9.97749836708034e-06, + "loss": 0.7807, + "step": 739 + }, + { + "epoch": 0.02416326530612245, + "grad_norm": 2.7452003955841064, + "learning_rate": 9.977465708687134e-06, + "loss": 0.8272, + "step": 740 + }, + { + "epoch": 0.024195918367346938, + "grad_norm": 2.922658681869507, + "learning_rate": 9.977433050293927e-06, + "loss": 0.8096, + "step": 741 + }, + { + "epoch": 0.02422857142857143, + "grad_norm": 2.9923369884490967, + "learning_rate": 9.977400391900719e-06, + "loss": 0.8232, + "step": 742 + }, + { + "epoch": 0.024261224489795917, + "grad_norm": 2.879516363143921, + "learning_rate": 9.977367733507512e-06, + "loss": 0.8128, + "step": 743 + }, + { + "epoch": 0.02429387755102041, + "grad_norm": 2.7809741497039795, + "learning_rate": 9.977335075114305e-06, + "loss": 0.7824, + "step": 744 + }, + { + "epoch": 0.0243265306122449, + "grad_norm": 2.8293521404266357, + "learning_rate": 9.977302416721098e-06, + "loss": 0.7637, + "step": 745 + }, + { + "epoch": 0.024359183673469387, + "grad_norm": 2.900975227355957, + "learning_rate": 9.977269758327891e-06, + "loss": 0.8065, + "step": 746 + }, + { + "epoch": 0.02439183673469388, + "grad_norm": 2.8670005798339844, + "learning_rate": 9.977237099934683e-06, + "loss": 0.7747, + "step": 747 + }, + { + "epoch": 0.024424489795918367, + "grad_norm": 2.987907886505127, + "learning_rate": 9.977204441541476e-06, + "loss": 0.7907, + "step": 748 + }, + { + "epoch": 0.024457142857142858, + "grad_norm": 2.8523452281951904, + "learning_rate": 9.97717178314827e-06, + "loss": 0.781, + "step": 749 + }, + { + "epoch": 0.024489795918367346, + "grad_norm": 2.9037506580352783, + "learning_rate": 9.977139124755063e-06, + "loss": 0.7505, + "step": 750 + }, + { + "epoch": 0.024489795918367346, + "eval_loss": 0.8159348964691162, + "eval_runtime": 74.1956, + "eval_samples_per_second": 1.348, + "eval_steps_per_second": 1.348, + "step": 750 + }, + { + "epoch": 0.024522448979591837, + "grad_norm": 2.8492870330810547, + "learning_rate": 9.977106466361856e-06, + "loss": 0.8186, + "step": 751 + }, + { + "epoch": 0.024555102040816325, + "grad_norm": 3.0403406620025635, + "learning_rate": 9.97707380796865e-06, + "loss": 0.8152, + "step": 752 + }, + { + "epoch": 0.024587755102040816, + "grad_norm": 2.8419723510742188, + "learning_rate": 9.97704114957544e-06, + "loss": 0.7908, + "step": 753 + }, + { + "epoch": 0.024620408163265307, + "grad_norm": 3.0176303386688232, + "learning_rate": 9.977008491182236e-06, + "loss": 0.7921, + "step": 754 + }, + { + "epoch": 0.024653061224489795, + "grad_norm": 2.928659439086914, + "learning_rate": 9.976975832789027e-06, + "loss": 0.7784, + "step": 755 + }, + { + "epoch": 0.024685714285714287, + "grad_norm": 2.800809621810913, + "learning_rate": 9.97694317439582e-06, + "loss": 0.7362, + "step": 756 + }, + { + "epoch": 0.024718367346938774, + "grad_norm": 2.698071002960205, + "learning_rate": 9.976910516002614e-06, + "loss": 0.8046, + "step": 757 + }, + { + "epoch": 0.024751020408163266, + "grad_norm": 2.8206839561462402, + "learning_rate": 9.976877857609407e-06, + "loss": 0.8169, + "step": 758 + }, + { + "epoch": 0.024783673469387753, + "grad_norm": 2.927217960357666, + "learning_rate": 9.9768451992162e-06, + "loss": 0.7816, + "step": 759 + }, + { + "epoch": 0.024816326530612245, + "grad_norm": 2.9845449924468994, + "learning_rate": 9.976812540822992e-06, + "loss": 0.801, + "step": 760 + }, + { + "epoch": 0.024848979591836736, + "grad_norm": 2.881765604019165, + "learning_rate": 9.976779882429785e-06, + "loss": 0.8052, + "step": 761 + }, + { + "epoch": 0.024881632653061224, + "grad_norm": 2.9570679664611816, + "learning_rate": 9.976747224036578e-06, + "loss": 0.8044, + "step": 762 + }, + { + "epoch": 0.024914285714285715, + "grad_norm": 3.069812774658203, + "learning_rate": 9.976714565643371e-06, + "loss": 0.8166, + "step": 763 + }, + { + "epoch": 0.024946938775510203, + "grad_norm": 3.0275254249572754, + "learning_rate": 9.976681907250165e-06, + "loss": 0.7835, + "step": 764 + }, + { + "epoch": 0.024979591836734694, + "grad_norm": 3.007643461227417, + "learning_rate": 9.976649248856956e-06, + "loss": 0.8088, + "step": 765 + }, + { + "epoch": 0.025012244897959182, + "grad_norm": 2.788733720779419, + "learning_rate": 9.97661659046375e-06, + "loss": 0.7852, + "step": 766 + }, + { + "epoch": 0.025044897959183673, + "grad_norm": 3.187893867492676, + "learning_rate": 9.976583932070543e-06, + "loss": 0.7887, + "step": 767 + }, + { + "epoch": 0.025077551020408165, + "grad_norm": 3.0475502014160156, + "learning_rate": 9.976551273677336e-06, + "loss": 0.7837, + "step": 768 + }, + { + "epoch": 0.025110204081632653, + "grad_norm": 2.9664242267608643, + "learning_rate": 9.976518615284129e-06, + "loss": 0.8021, + "step": 769 + }, + { + "epoch": 0.025142857142857144, + "grad_norm": 3.1118979454040527, + "learning_rate": 9.976485956890922e-06, + "loss": 0.7992, + "step": 770 + }, + { + "epoch": 0.02517551020408163, + "grad_norm": 2.909301280975342, + "learning_rate": 9.976453298497714e-06, + "loss": 0.743, + "step": 771 + }, + { + "epoch": 0.025208163265306123, + "grad_norm": 2.9557557106018066, + "learning_rate": 9.976420640104507e-06, + "loss": 0.8007, + "step": 772 + }, + { + "epoch": 0.02524081632653061, + "grad_norm": 3.108922243118286, + "learning_rate": 9.9763879817113e-06, + "loss": 0.8127, + "step": 773 + }, + { + "epoch": 0.025273469387755102, + "grad_norm": 2.927797555923462, + "learning_rate": 9.976355323318094e-06, + "loss": 0.7577, + "step": 774 + }, + { + "epoch": 0.025306122448979593, + "grad_norm": 2.929468870162964, + "learning_rate": 9.976322664924887e-06, + "loss": 0.7662, + "step": 775 + }, + { + "epoch": 0.02533877551020408, + "grad_norm": 3.0343034267425537, + "learning_rate": 9.976290006531678e-06, + "loss": 0.8283, + "step": 776 + }, + { + "epoch": 0.025371428571428573, + "grad_norm": 3.136800765991211, + "learning_rate": 9.976257348138473e-06, + "loss": 0.7935, + "step": 777 + }, + { + "epoch": 0.02540408163265306, + "grad_norm": 2.916213035583496, + "learning_rate": 9.976224689745265e-06, + "loss": 0.7651, + "step": 778 + }, + { + "epoch": 0.02543673469387755, + "grad_norm": 2.929622173309326, + "learning_rate": 9.976192031352058e-06, + "loss": 0.7677, + "step": 779 + }, + { + "epoch": 0.02546938775510204, + "grad_norm": 3.2419509887695312, + "learning_rate": 9.976159372958851e-06, + "loss": 0.7789, + "step": 780 + }, + { + "epoch": 0.02550204081632653, + "grad_norm": 2.8888614177703857, + "learning_rate": 9.976126714565645e-06, + "loss": 0.7735, + "step": 781 + }, + { + "epoch": 0.025534693877551022, + "grad_norm": 3.1505699157714844, + "learning_rate": 9.976094056172438e-06, + "loss": 0.8128, + "step": 782 + }, + { + "epoch": 0.02556734693877551, + "grad_norm": 3.030273675918579, + "learning_rate": 9.97606139777923e-06, + "loss": 0.7678, + "step": 783 + }, + { + "epoch": 0.0256, + "grad_norm": 3.0003929138183594, + "learning_rate": 9.976028739386023e-06, + "loss": 0.8076, + "step": 784 + }, + { + "epoch": 0.02563265306122449, + "grad_norm": 2.957667350769043, + "learning_rate": 9.975996080992816e-06, + "loss": 0.7886, + "step": 785 + }, + { + "epoch": 0.02566530612244898, + "grad_norm": 3.0816612243652344, + "learning_rate": 9.975963422599609e-06, + "loss": 0.7871, + "step": 786 + }, + { + "epoch": 0.025697959183673468, + "grad_norm": 2.947930097579956, + "learning_rate": 9.975930764206402e-06, + "loss": 0.8288, + "step": 787 + }, + { + "epoch": 0.02573061224489796, + "grad_norm": 3.121185302734375, + "learning_rate": 9.975898105813194e-06, + "loss": 0.793, + "step": 788 + }, + { + "epoch": 0.025763265306122447, + "grad_norm": 3.0452706813812256, + "learning_rate": 9.975865447419987e-06, + "loss": 0.7819, + "step": 789 + }, + { + "epoch": 0.02579591836734694, + "grad_norm": 2.903444528579712, + "learning_rate": 9.97583278902678e-06, + "loss": 0.798, + "step": 790 + }, + { + "epoch": 0.02582857142857143, + "grad_norm": 3.088996410369873, + "learning_rate": 9.975800130633574e-06, + "loss": 0.8171, + "step": 791 + }, + { + "epoch": 0.025861224489795918, + "grad_norm": 2.957293748855591, + "learning_rate": 9.975767472240367e-06, + "loss": 0.7862, + "step": 792 + }, + { + "epoch": 0.02589387755102041, + "grad_norm": 2.7569499015808105, + "learning_rate": 9.97573481384716e-06, + "loss": 0.7899, + "step": 793 + }, + { + "epoch": 0.025926530612244897, + "grad_norm": 2.931257724761963, + "learning_rate": 9.975702155453952e-06, + "loss": 0.8051, + "step": 794 + }, + { + "epoch": 0.025959183673469388, + "grad_norm": 2.9962334632873535, + "learning_rate": 9.975669497060746e-06, + "loss": 0.7914, + "step": 795 + }, + { + "epoch": 0.025991836734693876, + "grad_norm": 3.016993761062622, + "learning_rate": 9.975636838667538e-06, + "loss": 0.8011, + "step": 796 + }, + { + "epoch": 0.026024489795918367, + "grad_norm": 3.066710948944092, + "learning_rate": 9.975604180274331e-06, + "loss": 0.7885, + "step": 797 + }, + { + "epoch": 0.02605714285714286, + "grad_norm": 2.9981632232666016, + "learning_rate": 9.975571521881125e-06, + "loss": 0.7968, + "step": 798 + }, + { + "epoch": 0.026089795918367346, + "grad_norm": 2.971733331680298, + "learning_rate": 9.975538863487916e-06, + "loss": 0.7997, + "step": 799 + }, + { + "epoch": 0.026122448979591838, + "grad_norm": 3.0647032260894775, + "learning_rate": 9.975506205094711e-06, + "loss": 0.8122, + "step": 800 + }, + { + "epoch": 0.026122448979591838, + "eval_loss": 0.8141899108886719, + "eval_runtime": 73.6209, + "eval_samples_per_second": 1.358, + "eval_steps_per_second": 1.358, + "step": 800 + }, + { + "epoch": 0.026155102040816326, + "grad_norm": 3.015854597091675, + "learning_rate": 9.975473546701503e-06, + "loss": 0.8062, + "step": 801 + }, + { + "epoch": 0.026187755102040817, + "grad_norm": 3.065284013748169, + "learning_rate": 9.975440888308296e-06, + "loss": 0.7973, + "step": 802 + }, + { + "epoch": 0.026220408163265305, + "grad_norm": 2.8390414714813232, + "learning_rate": 9.975408229915089e-06, + "loss": 0.7943, + "step": 803 + }, + { + "epoch": 0.026253061224489796, + "grad_norm": 3.254967451095581, + "learning_rate": 9.975375571521882e-06, + "loss": 0.8059, + "step": 804 + }, + { + "epoch": 0.026285714285714287, + "grad_norm": 3.0705697536468506, + "learning_rate": 9.975342913128675e-06, + "loss": 0.7871, + "step": 805 + }, + { + "epoch": 0.026318367346938775, + "grad_norm": 2.900188446044922, + "learning_rate": 9.975310254735467e-06, + "loss": 0.7672, + "step": 806 + }, + { + "epoch": 0.026351020408163266, + "grad_norm": 2.8207056522369385, + "learning_rate": 9.97527759634226e-06, + "loss": 0.7921, + "step": 807 + }, + { + "epoch": 0.026383673469387754, + "grad_norm": 2.891328811645508, + "learning_rate": 9.975244937949053e-06, + "loss": 0.8177, + "step": 808 + }, + { + "epoch": 0.026416326530612246, + "grad_norm": 2.974545955657959, + "learning_rate": 9.975212279555847e-06, + "loss": 0.7773, + "step": 809 + }, + { + "epoch": 0.026448979591836733, + "grad_norm": 2.8445703983306885, + "learning_rate": 9.97517962116264e-06, + "loss": 0.8131, + "step": 810 + }, + { + "epoch": 0.026481632653061225, + "grad_norm": 2.93383526802063, + "learning_rate": 9.975146962769433e-06, + "loss": 0.7693, + "step": 811 + }, + { + "epoch": 0.026514285714285716, + "grad_norm": 3.0080511569976807, + "learning_rate": 9.975114304376225e-06, + "loss": 0.7593, + "step": 812 + }, + { + "epoch": 0.026546938775510204, + "grad_norm": 2.967305898666382, + "learning_rate": 9.975081645983018e-06, + "loss": 0.7773, + "step": 813 + }, + { + "epoch": 0.026579591836734695, + "grad_norm": 2.9201626777648926, + "learning_rate": 9.975048987589811e-06, + "loss": 0.7908, + "step": 814 + }, + { + "epoch": 0.026612244897959183, + "grad_norm": 3.0869758129119873, + "learning_rate": 9.975016329196604e-06, + "loss": 0.8015, + "step": 815 + }, + { + "epoch": 0.026644897959183674, + "grad_norm": 3.120110511779785, + "learning_rate": 9.974983670803398e-06, + "loss": 0.7993, + "step": 816 + }, + { + "epoch": 0.026677551020408162, + "grad_norm": 3.1260945796966553, + "learning_rate": 9.97495101241019e-06, + "loss": 0.8407, + "step": 817 + }, + { + "epoch": 0.026710204081632653, + "grad_norm": 2.858689785003662, + "learning_rate": 9.974918354016984e-06, + "loss": 0.8001, + "step": 818 + }, + { + "epoch": 0.02674285714285714, + "grad_norm": 2.854126214981079, + "learning_rate": 9.974885695623776e-06, + "loss": 0.754, + "step": 819 + }, + { + "epoch": 0.026775510204081632, + "grad_norm": 2.932650566101074, + "learning_rate": 9.974853037230569e-06, + "loss": 0.8051, + "step": 820 + }, + { + "epoch": 0.026808163265306124, + "grad_norm": 3.063242197036743, + "learning_rate": 9.974820378837362e-06, + "loss": 0.8114, + "step": 821 + }, + { + "epoch": 0.02684081632653061, + "grad_norm": 2.8537795543670654, + "learning_rate": 9.974787720444155e-06, + "loss": 0.7842, + "step": 822 + }, + { + "epoch": 0.026873469387755103, + "grad_norm": 2.986083745956421, + "learning_rate": 9.974755062050949e-06, + "loss": 0.7949, + "step": 823 + }, + { + "epoch": 0.02690612244897959, + "grad_norm": 3.0178754329681396, + "learning_rate": 9.97472240365774e-06, + "loss": 0.8005, + "step": 824 + }, + { + "epoch": 0.026938775510204082, + "grad_norm": 3.1746835708618164, + "learning_rate": 9.974689745264533e-06, + "loss": 0.7994, + "step": 825 + }, + { + "epoch": 0.02697142857142857, + "grad_norm": 2.8827731609344482, + "learning_rate": 9.974657086871327e-06, + "loss": 0.7764, + "step": 826 + }, + { + "epoch": 0.02700408163265306, + "grad_norm": 2.7529618740081787, + "learning_rate": 9.97462442847812e-06, + "loss": 0.8092, + "step": 827 + }, + { + "epoch": 0.027036734693877552, + "grad_norm": 2.8854501247406006, + "learning_rate": 9.974591770084913e-06, + "loss": 0.7713, + "step": 828 + }, + { + "epoch": 0.02706938775510204, + "grad_norm": 2.8492770195007324, + "learning_rate": 9.974559111691705e-06, + "loss": 0.75, + "step": 829 + }, + { + "epoch": 0.02710204081632653, + "grad_norm": 2.9942445755004883, + "learning_rate": 9.974526453298498e-06, + "loss": 0.7838, + "step": 830 + }, + { + "epoch": 0.02713469387755102, + "grad_norm": 2.9380862712860107, + "learning_rate": 9.974493794905291e-06, + "loss": 0.8111, + "step": 831 + }, + { + "epoch": 0.02716734693877551, + "grad_norm": 2.8200275897979736, + "learning_rate": 9.974461136512084e-06, + "loss": 0.7818, + "step": 832 + }, + { + "epoch": 0.0272, + "grad_norm": 3.0098073482513428, + "learning_rate": 9.974428478118878e-06, + "loss": 0.7786, + "step": 833 + }, + { + "epoch": 0.02723265306122449, + "grad_norm": 2.96885347366333, + "learning_rate": 9.97439581972567e-06, + "loss": 0.762, + "step": 834 + }, + { + "epoch": 0.02726530612244898, + "grad_norm": 2.9910387992858887, + "learning_rate": 9.974363161332462e-06, + "loss": 0.8142, + "step": 835 + }, + { + "epoch": 0.02729795918367347, + "grad_norm": 3.1752495765686035, + "learning_rate": 9.974330502939257e-06, + "loss": 0.8031, + "step": 836 + }, + { + "epoch": 0.02733061224489796, + "grad_norm": 2.9318926334381104, + "learning_rate": 9.974297844546049e-06, + "loss": 0.7873, + "step": 837 + }, + { + "epoch": 0.027363265306122448, + "grad_norm": 2.8536527156829834, + "learning_rate": 9.974265186152842e-06, + "loss": 0.7878, + "step": 838 + }, + { + "epoch": 0.02739591836734694, + "grad_norm": 2.7921385765075684, + "learning_rate": 9.974232527759635e-06, + "loss": 0.8042, + "step": 839 + }, + { + "epoch": 0.027428571428571427, + "grad_norm": 2.97587251663208, + "learning_rate": 9.974199869366427e-06, + "loss": 0.8044, + "step": 840 + }, + { + "epoch": 0.02746122448979592, + "grad_norm": 2.879415273666382, + "learning_rate": 9.974167210973222e-06, + "loss": 0.7902, + "step": 841 + }, + { + "epoch": 0.02749387755102041, + "grad_norm": 2.930534839630127, + "learning_rate": 9.974134552580013e-06, + "loss": 0.7809, + "step": 842 + }, + { + "epoch": 0.027526530612244898, + "grad_norm": 2.8264150619506836, + "learning_rate": 9.974101894186807e-06, + "loss": 0.7565, + "step": 843 + }, + { + "epoch": 0.02755918367346939, + "grad_norm": 2.879453182220459, + "learning_rate": 9.9740692357936e-06, + "loss": 0.8139, + "step": 844 + }, + { + "epoch": 0.027591836734693877, + "grad_norm": 2.805103063583374, + "learning_rate": 9.974036577400393e-06, + "loss": 0.798, + "step": 845 + }, + { + "epoch": 0.027624489795918368, + "grad_norm": 2.8710720539093018, + "learning_rate": 9.974003919007186e-06, + "loss": 0.7412, + "step": 846 + }, + { + "epoch": 0.027657142857142856, + "grad_norm": 2.9521493911743164, + "learning_rate": 9.973971260613978e-06, + "loss": 0.7943, + "step": 847 + }, + { + "epoch": 0.027689795918367347, + "grad_norm": 3.062432050704956, + "learning_rate": 9.973938602220771e-06, + "loss": 0.7884, + "step": 848 + }, + { + "epoch": 0.02772244897959184, + "grad_norm": 2.98431658744812, + "learning_rate": 9.973905943827564e-06, + "loss": 0.7643, + "step": 849 + }, + { + "epoch": 0.027755102040816326, + "grad_norm": 2.852004289627075, + "learning_rate": 9.973873285434358e-06, + "loss": 0.8126, + "step": 850 + }, + { + "epoch": 0.027755102040816326, + "eval_loss": 0.8143028020858765, + "eval_runtime": 85.3794, + "eval_samples_per_second": 1.171, + "eval_steps_per_second": 1.171, + "step": 850 + }, + { + "epoch": 0.027787755102040818, + "grad_norm": 3.2032933235168457, + "learning_rate": 9.97384062704115e-06, + "loss": 0.7847, + "step": 851 + }, + { + "epoch": 0.027820408163265305, + "grad_norm": 3.0217716693878174, + "learning_rate": 9.973807968647944e-06, + "loss": 0.7812, + "step": 852 + }, + { + "epoch": 0.027853061224489797, + "grad_norm": 2.961899995803833, + "learning_rate": 9.973775310254736e-06, + "loss": 0.7917, + "step": 853 + }, + { + "epoch": 0.027885714285714285, + "grad_norm": 2.9481585025787354, + "learning_rate": 9.973742651861529e-06, + "loss": 0.7866, + "step": 854 + }, + { + "epoch": 0.027918367346938776, + "grad_norm": 2.8054749965667725, + "learning_rate": 9.973709993468322e-06, + "loss": 0.7808, + "step": 855 + }, + { + "epoch": 0.027951020408163264, + "grad_norm": 3.106367349624634, + "learning_rate": 9.973677335075115e-06, + "loss": 0.8281, + "step": 856 + }, + { + "epoch": 0.027983673469387755, + "grad_norm": 3.7496819496154785, + "learning_rate": 9.973644676681908e-06, + "loss": 0.7851, + "step": 857 + }, + { + "epoch": 0.028016326530612246, + "grad_norm": 2.7985293865203857, + "learning_rate": 9.9736120182887e-06, + "loss": 0.7843, + "step": 858 + }, + { + "epoch": 0.028048979591836734, + "grad_norm": 2.8734278678894043, + "learning_rate": 9.973579359895495e-06, + "loss": 0.7482, + "step": 859 + }, + { + "epoch": 0.028081632653061225, + "grad_norm": 3.0287320613861084, + "learning_rate": 9.973546701502287e-06, + "loss": 0.7901, + "step": 860 + }, + { + "epoch": 0.028114285714285713, + "grad_norm": 3.2069811820983887, + "learning_rate": 9.97351404310908e-06, + "loss": 0.7762, + "step": 861 + }, + { + "epoch": 0.028146938775510204, + "grad_norm": 3.193441152572632, + "learning_rate": 9.973481384715873e-06, + "loss": 0.7867, + "step": 862 + }, + { + "epoch": 0.028179591836734692, + "grad_norm": 2.932454824447632, + "learning_rate": 9.973448726322665e-06, + "loss": 0.801, + "step": 863 + }, + { + "epoch": 0.028212244897959184, + "grad_norm": 3.009593963623047, + "learning_rate": 9.97341606792946e-06, + "loss": 0.7612, + "step": 864 + }, + { + "epoch": 0.028244897959183675, + "grad_norm": 3.2580459117889404, + "learning_rate": 9.973383409536251e-06, + "loss": 0.7295, + "step": 865 + }, + { + "epoch": 0.028277551020408163, + "grad_norm": 2.934058904647827, + "learning_rate": 9.973350751143044e-06, + "loss": 0.7766, + "step": 866 + }, + { + "epoch": 0.028310204081632654, + "grad_norm": 3.0753285884857178, + "learning_rate": 9.973318092749837e-06, + "loss": 0.7647, + "step": 867 + }, + { + "epoch": 0.028342857142857142, + "grad_norm": 3.1178665161132812, + "learning_rate": 9.97328543435663e-06, + "loss": 0.8071, + "step": 868 + }, + { + "epoch": 0.028375510204081633, + "grad_norm": 3.235008955001831, + "learning_rate": 9.973252775963424e-06, + "loss": 0.7899, + "step": 869 + }, + { + "epoch": 0.02840816326530612, + "grad_norm": 3.198587417602539, + "learning_rate": 9.973220117570215e-06, + "loss": 0.7918, + "step": 870 + }, + { + "epoch": 0.028440816326530612, + "grad_norm": 3.1364388465881348, + "learning_rate": 9.973187459177009e-06, + "loss": 0.8178, + "step": 871 + }, + { + "epoch": 0.028473469387755104, + "grad_norm": 3.1181514263153076, + "learning_rate": 9.973154800783802e-06, + "loss": 0.7922, + "step": 872 + }, + { + "epoch": 0.02850612244897959, + "grad_norm": 3.0627710819244385, + "learning_rate": 9.973122142390595e-06, + "loss": 0.8063, + "step": 873 + }, + { + "epoch": 0.028538775510204083, + "grad_norm": 3.042802572250366, + "learning_rate": 9.973089483997388e-06, + "loss": 0.8304, + "step": 874 + }, + { + "epoch": 0.02857142857142857, + "grad_norm": 2.935323715209961, + "learning_rate": 9.973056825604182e-06, + "loss": 0.7983, + "step": 875 + }, + { + "epoch": 0.028604081632653062, + "grad_norm": 2.9549400806427, + "learning_rate": 9.973024167210973e-06, + "loss": 0.8063, + "step": 876 + }, + { + "epoch": 0.02863673469387755, + "grad_norm": 2.7875542640686035, + "learning_rate": 9.972991508817766e-06, + "loss": 0.7981, + "step": 877 + }, + { + "epoch": 0.02866938775510204, + "grad_norm": 2.9394519329071045, + "learning_rate": 9.97295885042456e-06, + "loss": 0.792, + "step": 878 + }, + { + "epoch": 0.028702040816326532, + "grad_norm": 3.0494203567504883, + "learning_rate": 9.972926192031353e-06, + "loss": 0.8128, + "step": 879 + }, + { + "epoch": 0.02873469387755102, + "grad_norm": 2.9686169624328613, + "learning_rate": 9.972893533638146e-06, + "loss": 0.7797, + "step": 880 + }, + { + "epoch": 0.02876734693877551, + "grad_norm": 2.964944839477539, + "learning_rate": 9.972860875244938e-06, + "loss": 0.7494, + "step": 881 + }, + { + "epoch": 0.0288, + "grad_norm": 2.9027297496795654, + "learning_rate": 9.972828216851733e-06, + "loss": 0.7526, + "step": 882 + }, + { + "epoch": 0.02883265306122449, + "grad_norm": 3.0997474193573, + "learning_rate": 9.972795558458524e-06, + "loss": 0.7878, + "step": 883 + }, + { + "epoch": 0.02886530612244898, + "grad_norm": 2.999011516571045, + "learning_rate": 9.972762900065317e-06, + "loss": 0.785, + "step": 884 + }, + { + "epoch": 0.02889795918367347, + "grad_norm": 3.1621158123016357, + "learning_rate": 9.97273024167211e-06, + "loss": 0.8338, + "step": 885 + }, + { + "epoch": 0.028930612244897957, + "grad_norm": 2.943216323852539, + "learning_rate": 9.972697583278904e-06, + "loss": 0.7844, + "step": 886 + }, + { + "epoch": 0.02896326530612245, + "grad_norm": 3.0958337783813477, + "learning_rate": 9.972664924885697e-06, + "loss": 0.7895, + "step": 887 + }, + { + "epoch": 0.02899591836734694, + "grad_norm": 2.982701539993286, + "learning_rate": 9.972632266492489e-06, + "loss": 0.7783, + "step": 888 + }, + { + "epoch": 0.029028571428571428, + "grad_norm": 2.9064059257507324, + "learning_rate": 9.972599608099282e-06, + "loss": 0.7915, + "step": 889 + }, + { + "epoch": 0.02906122448979592, + "grad_norm": 2.8960936069488525, + "learning_rate": 9.972566949706075e-06, + "loss": 0.7914, + "step": 890 + }, + { + "epoch": 0.029093877551020407, + "grad_norm": 2.881122350692749, + "learning_rate": 9.972534291312868e-06, + "loss": 0.7673, + "step": 891 + }, + { + "epoch": 0.0291265306122449, + "grad_norm": 2.9828569889068604, + "learning_rate": 9.972501632919662e-06, + "loss": 0.8089, + "step": 892 + }, + { + "epoch": 0.029159183673469386, + "grad_norm": 3.0593910217285156, + "learning_rate": 9.972468974526453e-06, + "loss": 0.8079, + "step": 893 + }, + { + "epoch": 0.029191836734693877, + "grad_norm": 2.993431568145752, + "learning_rate": 9.972436316133246e-06, + "loss": 0.7855, + "step": 894 + }, + { + "epoch": 0.02922448979591837, + "grad_norm": 3.132528066635132, + "learning_rate": 9.97240365774004e-06, + "loss": 0.7948, + "step": 895 + }, + { + "epoch": 0.029257142857142857, + "grad_norm": 3.0954887866973877, + "learning_rate": 9.972370999346833e-06, + "loss": 0.765, + "step": 896 + }, + { + "epoch": 0.029289795918367348, + "grad_norm": 3.0066490173339844, + "learning_rate": 9.972338340953626e-06, + "loss": 0.7616, + "step": 897 + }, + { + "epoch": 0.029322448979591836, + "grad_norm": 3.05830717086792, + "learning_rate": 9.97230568256042e-06, + "loss": 0.7585, + "step": 898 + }, + { + "epoch": 0.029355102040816327, + "grad_norm": 3.2171273231506348, + "learning_rate": 9.972273024167211e-06, + "loss": 0.8138, + "step": 899 + }, + { + "epoch": 0.029387755102040815, + "grad_norm": 2.9297590255737305, + "learning_rate": 9.972240365774006e-06, + "loss": 0.762, + "step": 900 + }, + { + "epoch": 0.029387755102040815, + "eval_loss": 0.8132917284965515, + "eval_runtime": 76.0316, + "eval_samples_per_second": 1.315, + "eval_steps_per_second": 1.315, + "step": 900 + }, + { + "epoch": 0.029420408163265306, + "grad_norm": 2.8318891525268555, + "learning_rate": 9.972207707380797e-06, + "loss": 0.7571, + "step": 901 + }, + { + "epoch": 0.029453061224489797, + "grad_norm": 3.222458600997925, + "learning_rate": 9.97217504898759e-06, + "loss": 0.7869, + "step": 902 + }, + { + "epoch": 0.029485714285714285, + "grad_norm": 3.2787888050079346, + "learning_rate": 9.972142390594384e-06, + "loss": 0.7922, + "step": 903 + }, + { + "epoch": 0.029518367346938777, + "grad_norm": 2.9158124923706055, + "learning_rate": 9.972109732201175e-06, + "loss": 0.8008, + "step": 904 + }, + { + "epoch": 0.029551020408163264, + "grad_norm": 2.931711435317993, + "learning_rate": 9.97207707380797e-06, + "loss": 0.7958, + "step": 905 + }, + { + "epoch": 0.029583673469387756, + "grad_norm": 3.1747889518737793, + "learning_rate": 9.972044415414762e-06, + "loss": 0.7953, + "step": 906 + }, + { + "epoch": 0.029616326530612243, + "grad_norm": 3.23126482963562, + "learning_rate": 9.972011757021555e-06, + "loss": 0.8133, + "step": 907 + }, + { + "epoch": 0.029648979591836735, + "grad_norm": 2.8572208881378174, + "learning_rate": 9.971979098628348e-06, + "loss": 0.7832, + "step": 908 + }, + { + "epoch": 0.029681632653061226, + "grad_norm": 3.004584312438965, + "learning_rate": 9.971946440235142e-06, + "loss": 0.8025, + "step": 909 + }, + { + "epoch": 0.029714285714285714, + "grad_norm": 2.7122766971588135, + "learning_rate": 9.971913781841935e-06, + "loss": 0.7819, + "step": 910 + }, + { + "epoch": 0.029746938775510205, + "grad_norm": 2.9016056060791016, + "learning_rate": 9.971881123448726e-06, + "loss": 0.7852, + "step": 911 + }, + { + "epoch": 0.029779591836734693, + "grad_norm": 2.985293388366699, + "learning_rate": 9.97184846505552e-06, + "loss": 0.8294, + "step": 912 + }, + { + "epoch": 0.029812244897959184, + "grad_norm": 2.9722225666046143, + "learning_rate": 9.971815806662313e-06, + "loss": 0.7924, + "step": 913 + }, + { + "epoch": 0.029844897959183672, + "grad_norm": 2.8796117305755615, + "learning_rate": 9.971783148269106e-06, + "loss": 0.7549, + "step": 914 + }, + { + "epoch": 0.029877551020408163, + "grad_norm": 3.0465636253356934, + "learning_rate": 9.9717504898759e-06, + "loss": 0.7875, + "step": 915 + }, + { + "epoch": 0.029910204081632655, + "grad_norm": 3.179034948348999, + "learning_rate": 9.971717831482692e-06, + "loss": 0.774, + "step": 916 + }, + { + "epoch": 0.029942857142857143, + "grad_norm": 3.2824554443359375, + "learning_rate": 9.971685173089484e-06, + "loss": 0.788, + "step": 917 + }, + { + "epoch": 0.029975510204081634, + "grad_norm": 3.2578866481781006, + "learning_rate": 9.971652514696277e-06, + "loss": 0.7672, + "step": 918 + }, + { + "epoch": 0.03000816326530612, + "grad_norm": 3.1106038093566895, + "learning_rate": 9.97161985630307e-06, + "loss": 0.8133, + "step": 919 + }, + { + "epoch": 0.030040816326530613, + "grad_norm": 2.948033094406128, + "learning_rate": 9.971587197909864e-06, + "loss": 0.7819, + "step": 920 + }, + { + "epoch": 0.0300734693877551, + "grad_norm": 3.094900131225586, + "learning_rate": 9.971554539516657e-06, + "loss": 0.7833, + "step": 921 + }, + { + "epoch": 0.030106122448979592, + "grad_norm": 2.9197640419006348, + "learning_rate": 9.971521881123449e-06, + "loss": 0.7804, + "step": 922 + }, + { + "epoch": 0.03013877551020408, + "grad_norm": 3.011596441268921, + "learning_rate": 9.971489222730243e-06, + "loss": 0.7706, + "step": 923 + }, + { + "epoch": 0.03017142857142857, + "grad_norm": 2.9059317111968994, + "learning_rate": 9.971456564337035e-06, + "loss": 0.7594, + "step": 924 + }, + { + "epoch": 0.030204081632653063, + "grad_norm": 3.198932409286499, + "learning_rate": 9.971423905943828e-06, + "loss": 0.7876, + "step": 925 + }, + { + "epoch": 0.03023673469387755, + "grad_norm": 3.1033425331115723, + "learning_rate": 9.971391247550621e-06, + "loss": 0.7628, + "step": 926 + }, + { + "epoch": 0.03026938775510204, + "grad_norm": 3.210116147994995, + "learning_rate": 9.971358589157413e-06, + "loss": 0.748, + "step": 927 + }, + { + "epoch": 0.03030204081632653, + "grad_norm": 2.9255874156951904, + "learning_rate": 9.971325930764208e-06, + "loss": 0.7654, + "step": 928 + }, + { + "epoch": 0.03033469387755102, + "grad_norm": 2.949495553970337, + "learning_rate": 9.971293272371e-06, + "loss": 0.7875, + "step": 929 + }, + { + "epoch": 0.03036734693877551, + "grad_norm": 2.8590776920318604, + "learning_rate": 9.971260613977793e-06, + "loss": 0.7365, + "step": 930 + }, + { + "epoch": 0.0304, + "grad_norm": 2.9772932529449463, + "learning_rate": 9.971227955584586e-06, + "loss": 0.7683, + "step": 931 + }, + { + "epoch": 0.03043265306122449, + "grad_norm": 3.1256070137023926, + "learning_rate": 9.97119529719138e-06, + "loss": 0.8071, + "step": 932 + }, + { + "epoch": 0.03046530612244898, + "grad_norm": 3.2088918685913086, + "learning_rate": 9.971162638798172e-06, + "loss": 0.7676, + "step": 933 + }, + { + "epoch": 0.03049795918367347, + "grad_norm": 2.9030396938323975, + "learning_rate": 9.971129980404964e-06, + "loss": 0.8101, + "step": 934 + }, + { + "epoch": 0.030530612244897958, + "grad_norm": 2.968740940093994, + "learning_rate": 9.971097322011757e-06, + "loss": 0.7716, + "step": 935 + }, + { + "epoch": 0.03056326530612245, + "grad_norm": 2.748076915740967, + "learning_rate": 9.97106466361855e-06, + "loss": 0.7889, + "step": 936 + }, + { + "epoch": 0.030595918367346937, + "grad_norm": 2.941471815109253, + "learning_rate": 9.971032005225344e-06, + "loss": 0.7453, + "step": 937 + }, + { + "epoch": 0.03062857142857143, + "grad_norm": 2.9281535148620605, + "learning_rate": 9.970999346832137e-06, + "loss": 0.7891, + "step": 938 + }, + { + "epoch": 0.03066122448979592, + "grad_norm": 2.85978627204895, + "learning_rate": 9.97096668843893e-06, + "loss": 0.7772, + "step": 939 + }, + { + "epoch": 0.030693877551020408, + "grad_norm": 3.195918560028076, + "learning_rate": 9.970934030045722e-06, + "loss": 0.8068, + "step": 940 + }, + { + "epoch": 0.0307265306122449, + "grad_norm": 3.1167869567871094, + "learning_rate": 9.970901371652517e-06, + "loss": 0.7549, + "step": 941 + }, + { + "epoch": 0.030759183673469387, + "grad_norm": 2.9762868881225586, + "learning_rate": 9.970868713259308e-06, + "loss": 0.7923, + "step": 942 + }, + { + "epoch": 0.030791836734693878, + "grad_norm": 3.1077170372009277, + "learning_rate": 9.970836054866101e-06, + "loss": 0.7636, + "step": 943 + }, + { + "epoch": 0.030824489795918366, + "grad_norm": 3.062537431716919, + "learning_rate": 9.970803396472895e-06, + "loss": 0.7796, + "step": 944 + }, + { + "epoch": 0.030857142857142857, + "grad_norm": 3.2072839736938477, + "learning_rate": 9.970770738079686e-06, + "loss": 0.793, + "step": 945 + }, + { + "epoch": 0.03088979591836735, + "grad_norm": 3.26112699508667, + "learning_rate": 9.970738079686481e-06, + "loss": 0.8107, + "step": 946 + }, + { + "epoch": 0.030922448979591836, + "grad_norm": 2.9786770343780518, + "learning_rate": 9.970705421293273e-06, + "loss": 0.8025, + "step": 947 + }, + { + "epoch": 0.030955102040816328, + "grad_norm": 3.0765345096588135, + "learning_rate": 9.970672762900066e-06, + "loss": 0.8047, + "step": 948 + }, + { + "epoch": 0.030987755102040816, + "grad_norm": 2.958984375, + "learning_rate": 9.970640104506859e-06, + "loss": 0.762, + "step": 949 + }, + { + "epoch": 0.031020408163265307, + "grad_norm": 3.08040452003479, + "learning_rate": 9.970607446113652e-06, + "loss": 0.7928, + "step": 950 + }, + { + "epoch": 0.031020408163265307, + "eval_loss": 0.8111075758934021, + "eval_runtime": 74.5279, + "eval_samples_per_second": 1.342, + "eval_steps_per_second": 1.342, + "step": 950 + }, + { + "epoch": 0.031053061224489795, + "grad_norm": 2.965144395828247, + "learning_rate": 9.970574787720446e-06, + "loss": 0.8011, + "step": 951 + }, + { + "epoch": 0.031085714285714286, + "grad_norm": 2.9179527759552, + "learning_rate": 9.970542129327237e-06, + "loss": 0.7703, + "step": 952 + }, + { + "epoch": 0.031118367346938774, + "grad_norm": 2.9587607383728027, + "learning_rate": 9.97050947093403e-06, + "loss": 0.7536, + "step": 953 + }, + { + "epoch": 0.031151020408163265, + "grad_norm": 2.9393138885498047, + "learning_rate": 9.970476812540824e-06, + "loss": 0.7922, + "step": 954 + }, + { + "epoch": 0.031183673469387756, + "grad_norm": 3.013162612915039, + "learning_rate": 9.970444154147617e-06, + "loss": 0.7668, + "step": 955 + }, + { + "epoch": 0.031216326530612244, + "grad_norm": 2.8993349075317383, + "learning_rate": 9.97041149575441e-06, + "loss": 0.7559, + "step": 956 + }, + { + "epoch": 0.031248979591836736, + "grad_norm": 3.0861432552337646, + "learning_rate": 9.970378837361203e-06, + "loss": 0.7632, + "step": 957 + }, + { + "epoch": 0.03128163265306123, + "grad_norm": 2.9803247451782227, + "learning_rate": 9.970346178967995e-06, + "loss": 0.7933, + "step": 958 + }, + { + "epoch": 0.03131428571428571, + "grad_norm": 3.0155863761901855, + "learning_rate": 9.970313520574788e-06, + "loss": 0.8215, + "step": 959 + }, + { + "epoch": 0.0313469387755102, + "grad_norm": 3.0787782669067383, + "learning_rate": 9.970280862181581e-06, + "loss": 0.7936, + "step": 960 + }, + { + "epoch": 0.031379591836734694, + "grad_norm": 3.0619027614593506, + "learning_rate": 9.970248203788375e-06, + "loss": 0.7958, + "step": 961 + }, + { + "epoch": 0.031412244897959185, + "grad_norm": 2.7557358741760254, + "learning_rate": 9.970215545395168e-06, + "loss": 0.761, + "step": 962 + }, + { + "epoch": 0.031444897959183676, + "grad_norm": 2.8766896724700928, + "learning_rate": 9.97018288700196e-06, + "loss": 0.7563, + "step": 963 + }, + { + "epoch": 0.03147755102040816, + "grad_norm": 2.976086139678955, + "learning_rate": 9.970150228608754e-06, + "loss": 0.7839, + "step": 964 + }, + { + "epoch": 0.03151020408163265, + "grad_norm": 2.788193941116333, + "learning_rate": 9.970117570215546e-06, + "loss": 0.789, + "step": 965 + }, + { + "epoch": 0.03154285714285714, + "grad_norm": 3.0085878372192383, + "learning_rate": 9.970084911822339e-06, + "loss": 0.771, + "step": 966 + }, + { + "epoch": 0.031575510204081635, + "grad_norm": 3.1194570064544678, + "learning_rate": 9.970052253429132e-06, + "loss": 0.7889, + "step": 967 + }, + { + "epoch": 0.031608163265306126, + "grad_norm": 2.8725526332855225, + "learning_rate": 9.970019595035924e-06, + "loss": 0.7882, + "step": 968 + }, + { + "epoch": 0.03164081632653061, + "grad_norm": 2.982837677001953, + "learning_rate": 9.969986936642719e-06, + "loss": 0.7674, + "step": 969 + }, + { + "epoch": 0.0316734693877551, + "grad_norm": 2.9641151428222656, + "learning_rate": 9.96995427824951e-06, + "loss": 0.7871, + "step": 970 + }, + { + "epoch": 0.03170612244897959, + "grad_norm": 2.973365068435669, + "learning_rate": 9.969921619856304e-06, + "loss": 0.7939, + "step": 971 + }, + { + "epoch": 0.031738775510204084, + "grad_norm": 3.0161428451538086, + "learning_rate": 9.969888961463097e-06, + "loss": 0.7865, + "step": 972 + }, + { + "epoch": 0.03177142857142857, + "grad_norm": 3.369993209838867, + "learning_rate": 9.96985630306989e-06, + "loss": 0.8231, + "step": 973 + }, + { + "epoch": 0.03180408163265306, + "grad_norm": 3.0183799266815186, + "learning_rate": 9.969823644676683e-06, + "loss": 0.7853, + "step": 974 + }, + { + "epoch": 0.03183673469387755, + "grad_norm": 3.244966983795166, + "learning_rate": 9.969790986283475e-06, + "loss": 0.7607, + "step": 975 + }, + { + "epoch": 0.03186938775510204, + "grad_norm": 3.0273585319519043, + "learning_rate": 9.969758327890268e-06, + "loss": 0.77, + "step": 976 + }, + { + "epoch": 0.031902040816326534, + "grad_norm": 3.168429136276245, + "learning_rate": 9.969725669497061e-06, + "loss": 0.7537, + "step": 977 + }, + { + "epoch": 0.03193469387755102, + "grad_norm": 3.006373882293701, + "learning_rate": 9.969693011103854e-06, + "loss": 0.7927, + "step": 978 + }, + { + "epoch": 0.03196734693877551, + "grad_norm": 3.1100571155548096, + "learning_rate": 9.969660352710648e-06, + "loss": 0.7846, + "step": 979 + }, + { + "epoch": 0.032, + "grad_norm": 2.941803216934204, + "learning_rate": 9.969627694317441e-06, + "loss": 0.7821, + "step": 980 + }, + { + "epoch": 0.03203265306122449, + "grad_norm": 3.077153444290161, + "learning_rate": 9.969595035924232e-06, + "loss": 0.8004, + "step": 981 + }, + { + "epoch": 0.032065306122448976, + "grad_norm": 2.9901161193847656, + "learning_rate": 9.969562377531026e-06, + "loss": 0.7848, + "step": 982 + }, + { + "epoch": 0.03209795918367347, + "grad_norm": 3.0473811626434326, + "learning_rate": 9.969529719137819e-06, + "loss": 0.7823, + "step": 983 + }, + { + "epoch": 0.03213061224489796, + "grad_norm": 2.894256353378296, + "learning_rate": 9.969497060744612e-06, + "loss": 0.7757, + "step": 984 + }, + { + "epoch": 0.03216326530612245, + "grad_norm": 3.0119597911834717, + "learning_rate": 9.969464402351405e-06, + "loss": 0.7799, + "step": 985 + }, + { + "epoch": 0.03219591836734694, + "grad_norm": 3.020019292831421, + "learning_rate": 9.969431743958197e-06, + "loss": 0.7817, + "step": 986 + }, + { + "epoch": 0.032228571428571426, + "grad_norm": 3.0470404624938965, + "learning_rate": 9.969399085564992e-06, + "loss": 0.7873, + "step": 987 + }, + { + "epoch": 0.03226122448979592, + "grad_norm": 2.9759464263916016, + "learning_rate": 9.969366427171783e-06, + "loss": 0.7857, + "step": 988 + }, + { + "epoch": 0.03229387755102041, + "grad_norm": 3.1986935138702393, + "learning_rate": 9.969333768778577e-06, + "loss": 0.786, + "step": 989 + }, + { + "epoch": 0.0323265306122449, + "grad_norm": 3.0360336303710938, + "learning_rate": 9.96930111038537e-06, + "loss": 0.8174, + "step": 990 + }, + { + "epoch": 0.03235918367346939, + "grad_norm": 3.093979597091675, + "learning_rate": 9.969268451992163e-06, + "loss": 0.7619, + "step": 991 + }, + { + "epoch": 0.032391836734693875, + "grad_norm": 2.935920238494873, + "learning_rate": 9.969235793598956e-06, + "loss": 0.7661, + "step": 992 + }, + { + "epoch": 0.03242448979591837, + "grad_norm": 3.1429708003997803, + "learning_rate": 9.969203135205748e-06, + "loss": 0.7981, + "step": 993 + }, + { + "epoch": 0.03245714285714286, + "grad_norm": 2.8831772804260254, + "learning_rate": 9.969170476812541e-06, + "loss": 0.7969, + "step": 994 + }, + { + "epoch": 0.03248979591836735, + "grad_norm": 2.8683125972747803, + "learning_rate": 9.969137818419334e-06, + "loss": 0.7688, + "step": 995 + }, + { + "epoch": 0.032522448979591834, + "grad_norm": 3.026094436645508, + "learning_rate": 9.969105160026128e-06, + "loss": 0.782, + "step": 996 + }, + { + "epoch": 0.032555102040816325, + "grad_norm": 3.065110445022583, + "learning_rate": 9.969072501632921e-06, + "loss": 0.782, + "step": 997 + }, + { + "epoch": 0.032587755102040816, + "grad_norm": 2.9125099182128906, + "learning_rate": 9.969039843239714e-06, + "loss": 0.8056, + "step": 998 + }, + { + "epoch": 0.03262040816326531, + "grad_norm": 2.978609800338745, + "learning_rate": 9.969007184846506e-06, + "loss": 0.7839, + "step": 999 + }, + { + "epoch": 0.0326530612244898, + "grad_norm": 3.037384510040283, + "learning_rate": 9.968974526453299e-06, + "loss": 0.7877, + "step": 1000 + }, + { + "epoch": 0.0326530612244898, + "eval_loss": 0.8111925721168518, + "eval_runtime": 72.9629, + "eval_samples_per_second": 1.371, + "eval_steps_per_second": 1.371, + "step": 1000 + }, + { + "epoch": 0.03268571428571428, + "grad_norm": 2.9627959728240967, + "learning_rate": 9.968941868060092e-06, + "loss": 0.7855, + "step": 1001 + }, + { + "epoch": 0.032718367346938775, + "grad_norm": 3.2459042072296143, + "learning_rate": 9.968909209666885e-06, + "loss": 0.7897, + "step": 1002 + }, + { + "epoch": 0.032751020408163266, + "grad_norm": 3.0746281147003174, + "learning_rate": 9.968876551273679e-06, + "loss": 0.7415, + "step": 1003 + }, + { + "epoch": 0.03278367346938776, + "grad_norm": 3.005556583404541, + "learning_rate": 9.96884389288047e-06, + "loss": 0.8075, + "step": 1004 + }, + { + "epoch": 0.03281632653061224, + "grad_norm": 2.9875948429107666, + "learning_rate": 9.968811234487265e-06, + "loss": 0.7982, + "step": 1005 + }, + { + "epoch": 0.03284897959183673, + "grad_norm": 2.8231279850006104, + "learning_rate": 9.968778576094057e-06, + "loss": 0.758, + "step": 1006 + }, + { + "epoch": 0.032881632653061224, + "grad_norm": 2.8995821475982666, + "learning_rate": 9.96874591770085e-06, + "loss": 0.7653, + "step": 1007 + }, + { + "epoch": 0.032914285714285715, + "grad_norm": 3.0024149417877197, + "learning_rate": 9.968713259307643e-06, + "loss": 0.7777, + "step": 1008 + }, + { + "epoch": 0.03294693877551021, + "grad_norm": 3.019351005554199, + "learning_rate": 9.968680600914435e-06, + "loss": 0.7921, + "step": 1009 + }, + { + "epoch": 0.03297959183673469, + "grad_norm": 2.916240930557251, + "learning_rate": 9.96864794252123e-06, + "loss": 0.75, + "step": 1010 + }, + { + "epoch": 0.03301224489795918, + "grad_norm": 3.005613088607788, + "learning_rate": 9.968615284128021e-06, + "loss": 0.7943, + "step": 1011 + }, + { + "epoch": 0.033044897959183674, + "grad_norm": 2.9690959453582764, + "learning_rate": 9.968582625734814e-06, + "loss": 0.7855, + "step": 1012 + }, + { + "epoch": 0.033077551020408165, + "grad_norm": 3.0370657444000244, + "learning_rate": 9.968549967341608e-06, + "loss": 0.7567, + "step": 1013 + }, + { + "epoch": 0.033110204081632656, + "grad_norm": 3.010715961456299, + "learning_rate": 9.9685173089484e-06, + "loss": 0.7545, + "step": 1014 + }, + { + "epoch": 0.03314285714285714, + "grad_norm": 3.1019158363342285, + "learning_rate": 9.968484650555194e-06, + "loss": 0.7881, + "step": 1015 + }, + { + "epoch": 0.03317551020408163, + "grad_norm": 2.9990556240081787, + "learning_rate": 9.968451992161986e-06, + "loss": 0.7892, + "step": 1016 + }, + { + "epoch": 0.03320816326530612, + "grad_norm": 2.9631762504577637, + "learning_rate": 9.968419333768779e-06, + "loss": 0.7459, + "step": 1017 + }, + { + "epoch": 0.033240816326530614, + "grad_norm": 3.1492321491241455, + "learning_rate": 9.968386675375572e-06, + "loss": 0.796, + "step": 1018 + }, + { + "epoch": 0.0332734693877551, + "grad_norm": 3.0348122119903564, + "learning_rate": 9.968354016982365e-06, + "loss": 0.795, + "step": 1019 + }, + { + "epoch": 0.03330612244897959, + "grad_norm": 3.2612578868865967, + "learning_rate": 9.968321358589159e-06, + "loss": 0.7781, + "step": 1020 + }, + { + "epoch": 0.03333877551020408, + "grad_norm": 3.0605742931365967, + "learning_rate": 9.968288700195952e-06, + "loss": 0.7329, + "step": 1021 + }, + { + "epoch": 0.03337142857142857, + "grad_norm": 3.0548489093780518, + "learning_rate": 9.968256041802743e-06, + "loss": 0.7601, + "step": 1022 + }, + { + "epoch": 0.033404081632653064, + "grad_norm": 2.946451187133789, + "learning_rate": 9.968223383409537e-06, + "loss": 0.7756, + "step": 1023 + }, + { + "epoch": 0.03343673469387755, + "grad_norm": 3.0444273948669434, + "learning_rate": 9.96819072501633e-06, + "loss": 0.762, + "step": 1024 + }, + { + "epoch": 0.03346938775510204, + "grad_norm": 3.207759380340576, + "learning_rate": 9.968158066623123e-06, + "loss": 0.7755, + "step": 1025 + }, + { + "epoch": 0.03350204081632653, + "grad_norm": 3.0307605266571045, + "learning_rate": 9.968125408229916e-06, + "loss": 0.7604, + "step": 1026 + }, + { + "epoch": 0.03353469387755102, + "grad_norm": 2.975473165512085, + "learning_rate": 9.968092749836708e-06, + "loss": 0.7929, + "step": 1027 + }, + { + "epoch": 0.033567346938775514, + "grad_norm": 3.00290846824646, + "learning_rate": 9.968060091443503e-06, + "loss": 0.8294, + "step": 1028 + }, + { + "epoch": 0.0336, + "grad_norm": 3.235747814178467, + "learning_rate": 9.968027433050294e-06, + "loss": 0.7807, + "step": 1029 + }, + { + "epoch": 0.03363265306122449, + "grad_norm": 3.031163215637207, + "learning_rate": 9.967994774657088e-06, + "loss": 0.7549, + "step": 1030 + }, + { + "epoch": 0.03366530612244898, + "grad_norm": 2.9528584480285645, + "learning_rate": 9.96796211626388e-06, + "loss": 0.7542, + "step": 1031 + }, + { + "epoch": 0.03369795918367347, + "grad_norm": 3.0693178176879883, + "learning_rate": 9.967929457870672e-06, + "loss": 0.7773, + "step": 1032 + }, + { + "epoch": 0.033730612244897956, + "grad_norm": 3.159449338912964, + "learning_rate": 9.967896799477467e-06, + "loss": 0.7998, + "step": 1033 + }, + { + "epoch": 0.03376326530612245, + "grad_norm": 3.0627989768981934, + "learning_rate": 9.967864141084259e-06, + "loss": 0.7624, + "step": 1034 + }, + { + "epoch": 0.03379591836734694, + "grad_norm": 3.088207721710205, + "learning_rate": 9.967831482691052e-06, + "loss": 0.7757, + "step": 1035 + }, + { + "epoch": 0.03382857142857143, + "grad_norm": 3.1921637058258057, + "learning_rate": 9.967798824297845e-06, + "loss": 0.7923, + "step": 1036 + }, + { + "epoch": 0.03386122448979592, + "grad_norm": 3.039994955062866, + "learning_rate": 9.967766165904638e-06, + "loss": 0.7628, + "step": 1037 + }, + { + "epoch": 0.033893877551020406, + "grad_norm": 3.221714973449707, + "learning_rate": 9.967733507511432e-06, + "loss": 0.7684, + "step": 1038 + }, + { + "epoch": 0.0339265306122449, + "grad_norm": 2.9829065799713135, + "learning_rate": 9.967700849118223e-06, + "loss": 0.7956, + "step": 1039 + }, + { + "epoch": 0.03395918367346939, + "grad_norm": 2.9304864406585693, + "learning_rate": 9.967668190725016e-06, + "loss": 0.7865, + "step": 1040 + }, + { + "epoch": 0.03399183673469388, + "grad_norm": 3.0041542053222656, + "learning_rate": 9.96763553233181e-06, + "loss": 0.8113, + "step": 1041 + }, + { + "epoch": 0.034024489795918364, + "grad_norm": 3.1308107376098633, + "learning_rate": 9.967602873938603e-06, + "loss": 0.7622, + "step": 1042 + }, + { + "epoch": 0.034057142857142855, + "grad_norm": 3.0761616230010986, + "learning_rate": 9.967570215545396e-06, + "loss": 0.7822, + "step": 1043 + }, + { + "epoch": 0.03408979591836735, + "grad_norm": 3.073542594909668, + "learning_rate": 9.96753755715219e-06, + "loss": 0.782, + "step": 1044 + }, + { + "epoch": 0.03412244897959184, + "grad_norm": 2.954620361328125, + "learning_rate": 9.967504898758981e-06, + "loss": 0.7686, + "step": 1045 + }, + { + "epoch": 0.03415510204081633, + "grad_norm": 3.1652603149414062, + "learning_rate": 9.967472240365776e-06, + "loss": 0.7743, + "step": 1046 + }, + { + "epoch": 0.034187755102040814, + "grad_norm": 2.8833162784576416, + "learning_rate": 9.967439581972567e-06, + "loss": 0.7741, + "step": 1047 + }, + { + "epoch": 0.034220408163265305, + "grad_norm": 2.910767078399658, + "learning_rate": 9.96740692357936e-06, + "loss": 0.7768, + "step": 1048 + }, + { + "epoch": 0.034253061224489796, + "grad_norm": 3.0124878883361816, + "learning_rate": 9.967374265186154e-06, + "loss": 0.7656, + "step": 1049 + }, + { + "epoch": 0.03428571428571429, + "grad_norm": 3.200578451156616, + "learning_rate": 9.967341606792945e-06, + "loss": 0.7979, + "step": 1050 + }, + { + "epoch": 0.03428571428571429, + "eval_loss": 0.809950053691864, + "eval_runtime": 78.3268, + "eval_samples_per_second": 1.277, + "eval_steps_per_second": 1.277, + "step": 1050 + }, + { + "epoch": 0.03431836734693878, + "grad_norm": 3.1058106422424316, + "learning_rate": 9.96730894839974e-06, + "loss": 0.7929, + "step": 1051 + }, + { + "epoch": 0.03435102040816326, + "grad_norm": 2.981287956237793, + "learning_rate": 9.967276290006532e-06, + "loss": 0.7922, + "step": 1052 + }, + { + "epoch": 0.034383673469387754, + "grad_norm": 3.0732994079589844, + "learning_rate": 9.967243631613325e-06, + "loss": 0.7598, + "step": 1053 + }, + { + "epoch": 0.034416326530612246, + "grad_norm": 2.965428352355957, + "learning_rate": 9.967210973220118e-06, + "loss": 0.7796, + "step": 1054 + }, + { + "epoch": 0.03444897959183674, + "grad_norm": 3.047433853149414, + "learning_rate": 9.967178314826912e-06, + "loss": 0.7806, + "step": 1055 + }, + { + "epoch": 0.03448163265306122, + "grad_norm": 2.9787747859954834, + "learning_rate": 9.967145656433705e-06, + "loss": 0.7548, + "step": 1056 + }, + { + "epoch": 0.03451428571428571, + "grad_norm": 3.3675661087036133, + "learning_rate": 9.967112998040496e-06, + "loss": 0.7764, + "step": 1057 + }, + { + "epoch": 0.034546938775510204, + "grad_norm": 2.90116024017334, + "learning_rate": 9.96708033964729e-06, + "loss": 0.7819, + "step": 1058 + }, + { + "epoch": 0.034579591836734695, + "grad_norm": 2.965744972229004, + "learning_rate": 9.967047681254083e-06, + "loss": 0.7603, + "step": 1059 + }, + { + "epoch": 0.034612244897959187, + "grad_norm": 3.116170644760132, + "learning_rate": 9.967015022860876e-06, + "loss": 0.7885, + "step": 1060 + }, + { + "epoch": 0.03464489795918367, + "grad_norm": 3.1526293754577637, + "learning_rate": 9.96698236446767e-06, + "loss": 0.7752, + "step": 1061 + }, + { + "epoch": 0.03467755102040816, + "grad_norm": 3.0937249660491943, + "learning_rate": 9.966949706074463e-06, + "loss": 0.7845, + "step": 1062 + }, + { + "epoch": 0.034710204081632653, + "grad_norm": 3.069850444793701, + "learning_rate": 9.966917047681254e-06, + "loss": 0.7645, + "step": 1063 + }, + { + "epoch": 0.034742857142857145, + "grad_norm": 2.979881763458252, + "learning_rate": 9.966884389288047e-06, + "loss": 0.7687, + "step": 1064 + }, + { + "epoch": 0.034775510204081636, + "grad_norm": 3.1711268424987793, + "learning_rate": 9.96685173089484e-06, + "loss": 0.8136, + "step": 1065 + }, + { + "epoch": 0.03480816326530612, + "grad_norm": 3.197852373123169, + "learning_rate": 9.966819072501634e-06, + "loss": 0.8152, + "step": 1066 + }, + { + "epoch": 0.03484081632653061, + "grad_norm": 3.098540782928467, + "learning_rate": 9.966786414108427e-06, + "loss": 0.7645, + "step": 1067 + }, + { + "epoch": 0.0348734693877551, + "grad_norm": 3.1754651069641113, + "learning_rate": 9.966753755715219e-06, + "loss": 0.7663, + "step": 1068 + }, + { + "epoch": 0.034906122448979594, + "grad_norm": 3.0712950229644775, + "learning_rate": 9.966721097322014e-06, + "loss": 0.7851, + "step": 1069 + }, + { + "epoch": 0.03493877551020408, + "grad_norm": 3.285234212875366, + "learning_rate": 9.966688438928805e-06, + "loss": 0.774, + "step": 1070 + }, + { + "epoch": 0.03497142857142857, + "grad_norm": 3.2550666332244873, + "learning_rate": 9.966655780535598e-06, + "loss": 0.751, + "step": 1071 + }, + { + "epoch": 0.03500408163265306, + "grad_norm": 3.238384246826172, + "learning_rate": 9.966623122142392e-06, + "loss": 0.7666, + "step": 1072 + }, + { + "epoch": 0.03503673469387755, + "grad_norm": 3.0862374305725098, + "learning_rate": 9.966590463749183e-06, + "loss": 0.7981, + "step": 1073 + }, + { + "epoch": 0.035069387755102044, + "grad_norm": 3.172941207885742, + "learning_rate": 9.966557805355978e-06, + "loss": 0.7549, + "step": 1074 + }, + { + "epoch": 0.03510204081632653, + "grad_norm": 3.060302495956421, + "learning_rate": 9.96652514696277e-06, + "loss": 0.7835, + "step": 1075 + }, + { + "epoch": 0.03513469387755102, + "grad_norm": 3.122617244720459, + "learning_rate": 9.966492488569563e-06, + "loss": 0.7952, + "step": 1076 + }, + { + "epoch": 0.03516734693877551, + "grad_norm": 2.904637336730957, + "learning_rate": 9.966459830176356e-06, + "loss": 0.7595, + "step": 1077 + }, + { + "epoch": 0.0352, + "grad_norm": 3.1672565937042236, + "learning_rate": 9.96642717178315e-06, + "loss": 0.7926, + "step": 1078 + }, + { + "epoch": 0.035232653061224486, + "grad_norm": 3.0276424884796143, + "learning_rate": 9.966394513389943e-06, + "loss": 0.7801, + "step": 1079 + }, + { + "epoch": 0.03526530612244898, + "grad_norm": 3.0199339389801025, + "learning_rate": 9.966361854996734e-06, + "loss": 0.7977, + "step": 1080 + }, + { + "epoch": 0.03529795918367347, + "grad_norm": 3.1580443382263184, + "learning_rate": 9.966329196603527e-06, + "loss": 0.7638, + "step": 1081 + }, + { + "epoch": 0.03533061224489796, + "grad_norm": 2.9399259090423584, + "learning_rate": 9.96629653821032e-06, + "loss": 0.7594, + "step": 1082 + }, + { + "epoch": 0.03536326530612245, + "grad_norm": 3.1555962562561035, + "learning_rate": 9.966263879817114e-06, + "loss": 0.7688, + "step": 1083 + }, + { + "epoch": 0.035395918367346936, + "grad_norm": 3.03648042678833, + "learning_rate": 9.966231221423907e-06, + "loss": 0.8, + "step": 1084 + }, + { + "epoch": 0.03542857142857143, + "grad_norm": 3.142136573791504, + "learning_rate": 9.9661985630307e-06, + "loss": 0.7502, + "step": 1085 + }, + { + "epoch": 0.03546122448979592, + "grad_norm": 3.103031635284424, + "learning_rate": 9.966165904637492e-06, + "loss": 0.7979, + "step": 1086 + }, + { + "epoch": 0.03549387755102041, + "grad_norm": 3.0145888328552246, + "learning_rate": 9.966133246244287e-06, + "loss": 0.7669, + "step": 1087 + }, + { + "epoch": 0.0355265306122449, + "grad_norm": 3.1782283782958984, + "learning_rate": 9.966100587851078e-06, + "loss": 0.7677, + "step": 1088 + }, + { + "epoch": 0.035559183673469386, + "grad_norm": 3.102893829345703, + "learning_rate": 9.966067929457871e-06, + "loss": 0.7651, + "step": 1089 + }, + { + "epoch": 0.03559183673469388, + "grad_norm": 3.0069074630737305, + "learning_rate": 9.966035271064665e-06, + "loss": 0.7729, + "step": 1090 + }, + { + "epoch": 0.03562448979591837, + "grad_norm": 3.022388458251953, + "learning_rate": 9.966002612671456e-06, + "loss": 0.7537, + "step": 1091 + }, + { + "epoch": 0.03565714285714286, + "grad_norm": 3.0949289798736572, + "learning_rate": 9.965969954278251e-06, + "loss": 0.77, + "step": 1092 + }, + { + "epoch": 0.035689795918367344, + "grad_norm": 3.48496413230896, + "learning_rate": 9.965937295885043e-06, + "loss": 0.7671, + "step": 1093 + }, + { + "epoch": 0.035722448979591835, + "grad_norm": 3.1143581867218018, + "learning_rate": 9.965904637491836e-06, + "loss": 0.7451, + "step": 1094 + }, + { + "epoch": 0.035755102040816326, + "grad_norm": 2.93046498298645, + "learning_rate": 9.96587197909863e-06, + "loss": 0.7878, + "step": 1095 + }, + { + "epoch": 0.03578775510204082, + "grad_norm": 3.0415232181549072, + "learning_rate": 9.96583932070542e-06, + "loss": 0.7465, + "step": 1096 + }, + { + "epoch": 0.03582040816326531, + "grad_norm": 3.1707770824432373, + "learning_rate": 9.965806662312216e-06, + "loss": 0.7548, + "step": 1097 + }, + { + "epoch": 0.03585306122448979, + "grad_norm": 3.0290682315826416, + "learning_rate": 9.965774003919007e-06, + "loss": 0.7703, + "step": 1098 + }, + { + "epoch": 0.035885714285714285, + "grad_norm": 2.925924777984619, + "learning_rate": 9.9657413455258e-06, + "loss": 0.7653, + "step": 1099 + }, + { + "epoch": 0.035918367346938776, + "grad_norm": 3.2360856533050537, + "learning_rate": 9.965708687132594e-06, + "loss": 0.7705, + "step": 1100 + }, + { + "epoch": 0.035918367346938776, + "eval_loss": 0.8095739483833313, + "eval_runtime": 95.0564, + "eval_samples_per_second": 1.052, + "eval_steps_per_second": 1.052, + "step": 1100 + }, + { + "epoch": 0.03595102040816327, + "grad_norm": 3.276266574859619, + "learning_rate": 9.965676028739387e-06, + "loss": 0.7926, + "step": 1101 + }, + { + "epoch": 0.03598367346938776, + "grad_norm": 3.069849967956543, + "learning_rate": 9.96564337034618e-06, + "loss": 0.7966, + "step": 1102 + }, + { + "epoch": 0.03601632653061224, + "grad_norm": 3.484065532684326, + "learning_rate": 9.965610711952973e-06, + "loss": 0.7744, + "step": 1103 + }, + { + "epoch": 0.036048979591836734, + "grad_norm": 3.468071699142456, + "learning_rate": 9.965578053559765e-06, + "loss": 0.7575, + "step": 1104 + }, + { + "epoch": 0.036081632653061226, + "grad_norm": 3.8356101512908936, + "learning_rate": 9.965545395166558e-06, + "loss": 0.7817, + "step": 1105 + }, + { + "epoch": 0.03611428571428572, + "grad_norm": 3.2512879371643066, + "learning_rate": 9.965512736773351e-06, + "loss": 0.7356, + "step": 1106 + }, + { + "epoch": 0.0361469387755102, + "grad_norm": 3.004352569580078, + "learning_rate": 9.965480078380145e-06, + "loss": 0.7286, + "step": 1107 + }, + { + "epoch": 0.03617959183673469, + "grad_norm": 3.194286823272705, + "learning_rate": 9.965447419986938e-06, + "loss": 0.7474, + "step": 1108 + }, + { + "epoch": 0.036212244897959184, + "grad_norm": 3.509319543838501, + "learning_rate": 9.96541476159373e-06, + "loss": 0.7888, + "step": 1109 + }, + { + "epoch": 0.036244897959183675, + "grad_norm": 3.528698205947876, + "learning_rate": 9.965382103200524e-06, + "loss": 0.794, + "step": 1110 + }, + { + "epoch": 0.036277551020408166, + "grad_norm": 3.2880067825317383, + "learning_rate": 9.965349444807316e-06, + "loss": 0.7408, + "step": 1111 + }, + { + "epoch": 0.03631020408163265, + "grad_norm": 3.031257152557373, + "learning_rate": 9.965316786414109e-06, + "loss": 0.7847, + "step": 1112 + }, + { + "epoch": 0.03634285714285714, + "grad_norm": 3.2594382762908936, + "learning_rate": 9.965284128020902e-06, + "loss": 0.7911, + "step": 1113 + }, + { + "epoch": 0.03637551020408163, + "grad_norm": 3.2341842651367188, + "learning_rate": 9.965251469627694e-06, + "loss": 0.7412, + "step": 1114 + }, + { + "epoch": 0.036408163265306125, + "grad_norm": 3.251246213912964, + "learning_rate": 9.965218811234489e-06, + "loss": 0.7549, + "step": 1115 + }, + { + "epoch": 0.03644081632653061, + "grad_norm": 3.0365469455718994, + "learning_rate": 9.96518615284128e-06, + "loss": 0.78, + "step": 1116 + }, + { + "epoch": 0.0364734693877551, + "grad_norm": 3.059936285018921, + "learning_rate": 9.965153494448074e-06, + "loss": 0.805, + "step": 1117 + }, + { + "epoch": 0.03650612244897959, + "grad_norm": 3.0952272415161133, + "learning_rate": 9.965120836054867e-06, + "loss": 0.7563, + "step": 1118 + }, + { + "epoch": 0.03653877551020408, + "grad_norm": 3.136528491973877, + "learning_rate": 9.96508817766166e-06, + "loss": 0.7767, + "step": 1119 + }, + { + "epoch": 0.036571428571428574, + "grad_norm": 2.979304313659668, + "learning_rate": 9.965055519268453e-06, + "loss": 0.8002, + "step": 1120 + }, + { + "epoch": 0.03660408163265306, + "grad_norm": 2.8801610469818115, + "learning_rate": 9.965022860875245e-06, + "loss": 0.7793, + "step": 1121 + }, + { + "epoch": 0.03663673469387755, + "grad_norm": 2.9416167736053467, + "learning_rate": 9.964990202482038e-06, + "loss": 0.7703, + "step": 1122 + }, + { + "epoch": 0.03666938775510204, + "grad_norm": 3.1201093196868896, + "learning_rate": 9.964957544088831e-06, + "loss": 0.7699, + "step": 1123 + }, + { + "epoch": 0.03670204081632653, + "grad_norm": 3.1061949729919434, + "learning_rate": 9.964924885695625e-06, + "loss": 0.7974, + "step": 1124 + }, + { + "epoch": 0.036734693877551024, + "grad_norm": 3.0530447959899902, + "learning_rate": 9.964892227302418e-06, + "loss": 0.7576, + "step": 1125 + }, + { + "epoch": 0.03676734693877551, + "grad_norm": 3.002209186553955, + "learning_rate": 9.964859568909211e-06, + "loss": 0.7699, + "step": 1126 + }, + { + "epoch": 0.0368, + "grad_norm": 3.1953022480010986, + "learning_rate": 9.964826910516003e-06, + "loss": 0.7153, + "step": 1127 + }, + { + "epoch": 0.03683265306122449, + "grad_norm": 3.091996192932129, + "learning_rate": 9.964794252122796e-06, + "loss": 0.7368, + "step": 1128 + }, + { + "epoch": 0.03686530612244898, + "grad_norm": 3.173081398010254, + "learning_rate": 9.964761593729589e-06, + "loss": 0.7839, + "step": 1129 + }, + { + "epoch": 0.036897959183673466, + "grad_norm": 3.0133304595947266, + "learning_rate": 9.964728935336382e-06, + "loss": 0.7827, + "step": 1130 + }, + { + "epoch": 0.03693061224489796, + "grad_norm": 3.1596643924713135, + "learning_rate": 9.964696276943176e-06, + "loss": 0.7649, + "step": 1131 + }, + { + "epoch": 0.03696326530612245, + "grad_norm": 3.2854256629943848, + "learning_rate": 9.964663618549967e-06, + "loss": 0.751, + "step": 1132 + }, + { + "epoch": 0.03699591836734694, + "grad_norm": 2.8992016315460205, + "learning_rate": 9.964630960156762e-06, + "loss": 0.7535, + "step": 1133 + }, + { + "epoch": 0.03702857142857143, + "grad_norm": 2.9541709423065186, + "learning_rate": 9.964598301763554e-06, + "loss": 0.7433, + "step": 1134 + }, + { + "epoch": 0.037061224489795916, + "grad_norm": 2.9385082721710205, + "learning_rate": 9.964565643370347e-06, + "loss": 0.8005, + "step": 1135 + }, + { + "epoch": 0.03709387755102041, + "grad_norm": 3.042072057723999, + "learning_rate": 9.96453298497714e-06, + "loss": 0.7819, + "step": 1136 + }, + { + "epoch": 0.0371265306122449, + "grad_norm": 2.9819841384887695, + "learning_rate": 9.964500326583932e-06, + "loss": 0.7911, + "step": 1137 + }, + { + "epoch": 0.03715918367346939, + "grad_norm": 3.242607831954956, + "learning_rate": 9.964467668190726e-06, + "loss": 0.7926, + "step": 1138 + }, + { + "epoch": 0.037191836734693874, + "grad_norm": 2.9996449947357178, + "learning_rate": 9.964435009797518e-06, + "loss": 0.7904, + "step": 1139 + }, + { + "epoch": 0.037224489795918365, + "grad_norm": 3.1378376483917236, + "learning_rate": 9.964402351404311e-06, + "loss": 0.8, + "step": 1140 + }, + { + "epoch": 0.03725714285714286, + "grad_norm": 2.9573757648468018, + "learning_rate": 9.964369693011105e-06, + "loss": 0.7589, + "step": 1141 + }, + { + "epoch": 0.03728979591836735, + "grad_norm": 3.122229814529419, + "learning_rate": 9.964337034617898e-06, + "loss": 0.7852, + "step": 1142 + }, + { + "epoch": 0.03732244897959184, + "grad_norm": 3.3379294872283936, + "learning_rate": 9.964304376224691e-06, + "loss": 0.7868, + "step": 1143 + }, + { + "epoch": 0.037355102040816324, + "grad_norm": 3.2145421504974365, + "learning_rate": 9.964271717831484e-06, + "loss": 0.7516, + "step": 1144 + }, + { + "epoch": 0.037387755102040815, + "grad_norm": 3.068854808807373, + "learning_rate": 9.964239059438276e-06, + "loss": 0.7539, + "step": 1145 + }, + { + "epoch": 0.037420408163265306, + "grad_norm": 3.1995043754577637, + "learning_rate": 9.964206401045069e-06, + "loss": 0.7927, + "step": 1146 + }, + { + "epoch": 0.0374530612244898, + "grad_norm": 3.0854032039642334, + "learning_rate": 9.964173742651862e-06, + "loss": 0.7608, + "step": 1147 + }, + { + "epoch": 0.03748571428571429, + "grad_norm": 3.3977510929107666, + "learning_rate": 9.964141084258655e-06, + "loss": 0.779, + "step": 1148 + }, + { + "epoch": 0.03751836734693877, + "grad_norm": 3.0493321418762207, + "learning_rate": 9.964108425865449e-06, + "loss": 0.7329, + "step": 1149 + }, + { + "epoch": 0.037551020408163265, + "grad_norm": 2.8519468307495117, + "learning_rate": 9.96407576747224e-06, + "loss": 0.7954, + "step": 1150 + }, + { + "epoch": 0.037551020408163265, + "eval_loss": 0.8091681003570557, + "eval_runtime": 85.0018, + "eval_samples_per_second": 1.176, + "eval_steps_per_second": 1.176, + "step": 1150 + }, + { + "epoch": 0.037583673469387756, + "grad_norm": 2.9106438159942627, + "learning_rate": 9.964043109079035e-06, + "loss": 0.7796, + "step": 1151 + }, + { + "epoch": 0.03761632653061225, + "grad_norm": 3.1851396560668945, + "learning_rate": 9.964010450685827e-06, + "loss": 0.7718, + "step": 1152 + }, + { + "epoch": 0.03764897959183673, + "grad_norm": 3.101126194000244, + "learning_rate": 9.96397779229262e-06, + "loss": 0.8017, + "step": 1153 + }, + { + "epoch": 0.03768163265306122, + "grad_norm": 3.1867284774780273, + "learning_rate": 9.963945133899413e-06, + "loss": 0.7973, + "step": 1154 + }, + { + "epoch": 0.037714285714285714, + "grad_norm": 2.98366379737854, + "learning_rate": 9.963912475506205e-06, + "loss": 0.8105, + "step": 1155 + }, + { + "epoch": 0.037746938775510205, + "grad_norm": 2.998530387878418, + "learning_rate": 9.963879817113e-06, + "loss": 0.7738, + "step": 1156 + }, + { + "epoch": 0.0377795918367347, + "grad_norm": 3.0798680782318115, + "learning_rate": 9.963847158719791e-06, + "loss": 0.7693, + "step": 1157 + }, + { + "epoch": 0.03781224489795918, + "grad_norm": 3.2727184295654297, + "learning_rate": 9.963814500326584e-06, + "loss": 0.7271, + "step": 1158 + }, + { + "epoch": 0.03784489795918367, + "grad_norm": 3.05658221244812, + "learning_rate": 9.963781841933378e-06, + "loss": 0.7759, + "step": 1159 + }, + { + "epoch": 0.037877551020408164, + "grad_norm": 3.0974197387695312, + "learning_rate": 9.963749183540171e-06, + "loss": 0.7844, + "step": 1160 + }, + { + "epoch": 0.037910204081632655, + "grad_norm": 3.1191930770874023, + "learning_rate": 9.963716525146964e-06, + "loss": 0.8066, + "step": 1161 + }, + { + "epoch": 0.037942857142857146, + "grad_norm": 3.0258705615997314, + "learning_rate": 9.963683866753756e-06, + "loss": 0.7373, + "step": 1162 + }, + { + "epoch": 0.03797551020408163, + "grad_norm": 3.1126925945281982, + "learning_rate": 9.963651208360549e-06, + "loss": 0.7577, + "step": 1163 + }, + { + "epoch": 0.03800816326530612, + "grad_norm": 3.1630795001983643, + "learning_rate": 9.963618549967342e-06, + "loss": 0.7333, + "step": 1164 + }, + { + "epoch": 0.03804081632653061, + "grad_norm": 3.100018262863159, + "learning_rate": 9.963585891574135e-06, + "loss": 0.7697, + "step": 1165 + }, + { + "epoch": 0.038073469387755104, + "grad_norm": 3.369335889816284, + "learning_rate": 9.963553233180929e-06, + "loss": 0.7788, + "step": 1166 + }, + { + "epoch": 0.03810612244897959, + "grad_norm": 3.357944965362549, + "learning_rate": 9.963520574787722e-06, + "loss": 0.7744, + "step": 1167 + }, + { + "epoch": 0.03813877551020408, + "grad_norm": 3.253232717514038, + "learning_rate": 9.963487916394513e-06, + "loss": 0.752, + "step": 1168 + }, + { + "epoch": 0.03817142857142857, + "grad_norm": 2.9491517543792725, + "learning_rate": 9.963455258001307e-06, + "loss": 0.7657, + "step": 1169 + }, + { + "epoch": 0.03820408163265306, + "grad_norm": 3.1153948307037354, + "learning_rate": 9.9634225996081e-06, + "loss": 0.7545, + "step": 1170 + }, + { + "epoch": 0.038236734693877554, + "grad_norm": 3.0334460735321045, + "learning_rate": 9.963389941214893e-06, + "loss": 0.7999, + "step": 1171 + }, + { + "epoch": 0.03826938775510204, + "grad_norm": 3.0638813972473145, + "learning_rate": 9.963357282821686e-06, + "loss": 0.7922, + "step": 1172 + }, + { + "epoch": 0.03830204081632653, + "grad_norm": 3.0759074687957764, + "learning_rate": 9.963324624428478e-06, + "loss": 0.7521, + "step": 1173 + }, + { + "epoch": 0.03833469387755102, + "grad_norm": 3.0248312950134277, + "learning_rate": 9.963291966035273e-06, + "loss": 0.7695, + "step": 1174 + }, + { + "epoch": 0.03836734693877551, + "grad_norm": 3.101550817489624, + "learning_rate": 9.963259307642064e-06, + "loss": 0.7446, + "step": 1175 + }, + { + "epoch": 0.0384, + "grad_norm": 3.0817489624023438, + "learning_rate": 9.963226649248858e-06, + "loss": 0.7653, + "step": 1176 + }, + { + "epoch": 0.03843265306122449, + "grad_norm": 3.051537036895752, + "learning_rate": 9.96319399085565e-06, + "loss": 0.7723, + "step": 1177 + }, + { + "epoch": 0.03846530612244898, + "grad_norm": 3.030776023864746, + "learning_rate": 9.963161332462442e-06, + "loss": 0.7762, + "step": 1178 + }, + { + "epoch": 0.03849795918367347, + "grad_norm": 3.1081271171569824, + "learning_rate": 9.963128674069237e-06, + "loss": 0.8198, + "step": 1179 + }, + { + "epoch": 0.03853061224489796, + "grad_norm": 3.1396594047546387, + "learning_rate": 9.963096015676029e-06, + "loss": 0.7389, + "step": 1180 + }, + { + "epoch": 0.038563265306122446, + "grad_norm": 2.977698564529419, + "learning_rate": 9.963063357282822e-06, + "loss": 0.7961, + "step": 1181 + }, + { + "epoch": 0.03859591836734694, + "grad_norm": 3.048220157623291, + "learning_rate": 9.963030698889615e-06, + "loss": 0.8025, + "step": 1182 + }, + { + "epoch": 0.03862857142857143, + "grad_norm": 3.087770700454712, + "learning_rate": 9.962998040496409e-06, + "loss": 0.6993, + "step": 1183 + }, + { + "epoch": 0.03866122448979592, + "grad_norm": 3.0148611068725586, + "learning_rate": 9.962965382103202e-06, + "loss": 0.7555, + "step": 1184 + }, + { + "epoch": 0.03869387755102041, + "grad_norm": 3.4241836071014404, + "learning_rate": 9.962932723709993e-06, + "loss": 0.7876, + "step": 1185 + }, + { + "epoch": 0.038726530612244896, + "grad_norm": 3.200383424758911, + "learning_rate": 9.962900065316787e-06, + "loss": 0.7739, + "step": 1186 + }, + { + "epoch": 0.03875918367346939, + "grad_norm": 3.0834004878997803, + "learning_rate": 9.96286740692358e-06, + "loss": 0.7524, + "step": 1187 + }, + { + "epoch": 0.03879183673469388, + "grad_norm": 2.9796390533447266, + "learning_rate": 9.962834748530373e-06, + "loss": 0.7495, + "step": 1188 + }, + { + "epoch": 0.03882448979591837, + "grad_norm": 3.1456053256988525, + "learning_rate": 9.962802090137166e-06, + "loss": 0.7939, + "step": 1189 + }, + { + "epoch": 0.038857142857142854, + "grad_norm": 3.227130651473999, + "learning_rate": 9.96276943174396e-06, + "loss": 0.7283, + "step": 1190 + }, + { + "epoch": 0.038889795918367345, + "grad_norm": 3.2170636653900146, + "learning_rate": 9.962736773350751e-06, + "loss": 0.7481, + "step": 1191 + }, + { + "epoch": 0.03892244897959184, + "grad_norm": 6.70438814163208, + "learning_rate": 9.962704114957546e-06, + "loss": 0.7565, + "step": 1192 + }, + { + "epoch": 0.03895510204081633, + "grad_norm": 3.2916693687438965, + "learning_rate": 9.962671456564338e-06, + "loss": 0.7732, + "step": 1193 + }, + { + "epoch": 0.03898775510204082, + "grad_norm": 3.4526243209838867, + "learning_rate": 9.96263879817113e-06, + "loss": 0.7563, + "step": 1194 + }, + { + "epoch": 0.039020408163265304, + "grad_norm": 3.188997507095337, + "learning_rate": 9.962606139777924e-06, + "loss": 0.771, + "step": 1195 + }, + { + "epoch": 0.039053061224489795, + "grad_norm": 3.0594732761383057, + "learning_rate": 9.962573481384716e-06, + "loss": 0.7433, + "step": 1196 + }, + { + "epoch": 0.039085714285714286, + "grad_norm": 3.239617109298706, + "learning_rate": 9.96254082299151e-06, + "loss": 0.7823, + "step": 1197 + }, + { + "epoch": 0.03911836734693878, + "grad_norm": 2.983693838119507, + "learning_rate": 9.962508164598302e-06, + "loss": 0.7557, + "step": 1198 + }, + { + "epoch": 0.03915102040816327, + "grad_norm": 2.9180068969726562, + "learning_rate": 9.962475506205095e-06, + "loss": 0.7915, + "step": 1199 + }, + { + "epoch": 0.03918367346938775, + "grad_norm": 3.102025270462036, + "learning_rate": 9.962442847811888e-06, + "loss": 0.764, + "step": 1200 + }, + { + "epoch": 0.03918367346938775, + "eval_loss": 0.807881772518158, + "eval_runtime": 85.0203, + "eval_samples_per_second": 1.176, + "eval_steps_per_second": 1.176, + "step": 1200 + } + ], + "logging_steps": 1, + "max_steps": 306250, + "num_input_tokens_seen": 0, + "num_train_epochs": 10, + "save_steps": 50, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.305701335141549e+18, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}