{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9827003787491043, "eval_steps": 200, "global_step": 1200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.008189169822909202, "grad_norm": 4.80268669128418, "learning_rate": 0.0001, "loss": 1.6951, "step": 10 }, { "epoch": 0.016378339645818404, "grad_norm": 2.047800302505493, "learning_rate": 0.0001, "loss": 0.983, "step": 20 }, { "epoch": 0.024567509468727607, "grad_norm": 1.5396971702575684, "learning_rate": 0.0001, "loss": 0.9092, "step": 30 }, { "epoch": 0.03275667929163681, "grad_norm": 4.307969093322754, "learning_rate": 0.0001, "loss": 0.8544, "step": 40 }, { "epoch": 0.04094584911454601, "grad_norm": 2.1945383548736572, "learning_rate": 0.0001, "loss": 0.8015, "step": 50 }, { "epoch": 0.04913501893745521, "grad_norm": 1.5352895259857178, "learning_rate": 0.0001, "loss": 0.7917, "step": 60 }, { "epoch": 0.057324188760364415, "grad_norm": 1.9893757104873657, "learning_rate": 0.0001, "loss": 0.7561, "step": 70 }, { "epoch": 0.06551335858327362, "grad_norm": 1.067208170890808, "learning_rate": 0.0001, "loss": 0.7457, "step": 80 }, { "epoch": 0.07370252840618283, "grad_norm": 1.8355977535247803, "learning_rate": 0.0001, "loss": 0.7197, "step": 90 }, { "epoch": 0.08189169822909202, "grad_norm": 0.9938833713531494, "learning_rate": 0.0001, "loss": 0.6959, "step": 100 }, { "epoch": 0.09008086805200123, "grad_norm": 1.5569779872894287, "learning_rate": 0.0001, "loss": 0.6875, "step": 110 }, { "epoch": 0.09827003787491043, "grad_norm": 1.362318515777588, "learning_rate": 0.0001, "loss": 0.7031, "step": 120 }, { "epoch": 0.10645920769781964, "grad_norm": 0.9789265990257263, "learning_rate": 0.0001, "loss": 0.6744, "step": 130 }, { "epoch": 0.11464837752072883, "grad_norm": 0.9227413535118103, "learning_rate": 0.0001, "loss": 0.6496, "step": 140 }, { "epoch": 0.12283754734363804, "grad_norm": 1.0309714078903198, "learning_rate": 0.0001, "loss": 0.6365, "step": 150 }, { "epoch": 0.13102671716654724, "grad_norm": 1.2037794589996338, "learning_rate": 0.0001, "loss": 0.6629, "step": 160 }, { "epoch": 0.13921588698945644, "grad_norm": 0.8785035014152527, "learning_rate": 0.0001, "loss": 0.6462, "step": 170 }, { "epoch": 0.14740505681236565, "grad_norm": 1.5502973794937134, "learning_rate": 0.0001, "loss": 0.6146, "step": 180 }, { "epoch": 0.15559422663527486, "grad_norm": 0.9982532858848572, "learning_rate": 0.0001, "loss": 0.6205, "step": 190 }, { "epoch": 0.16378339645818404, "grad_norm": 1.0146996974945068, "learning_rate": 0.0001, "loss": 0.6107, "step": 200 }, { "epoch": 0.16378339645818404, "eval_loss": 0.5019610524177551, "eval_runtime": 993.1256, "eval_samples_per_second": 36.249, "eval_steps_per_second": 18.125, "step": 200 }, { "epoch": 0.17197256628109325, "grad_norm": 1.0556907653808594, "learning_rate": 0.0001, "loss": 0.6103, "step": 210 }, { "epoch": 0.18016173610400246, "grad_norm": 1.069365382194519, "learning_rate": 0.0001, "loss": 0.6309, "step": 220 }, { "epoch": 0.18835090592691167, "grad_norm": 1.2085084915161133, "learning_rate": 0.0001, "loss": 0.598, "step": 230 }, { "epoch": 0.19654007574982085, "grad_norm": 0.8124738931655884, "learning_rate": 0.0001, "loss": 0.603, "step": 240 }, { "epoch": 0.20472924557273006, "grad_norm": 1.032139778137207, "learning_rate": 0.0001, "loss": 0.611, "step": 250 }, { "epoch": 0.21291841539563927, "grad_norm": 0.8203685283660889, "learning_rate": 0.0001, "loss": 0.5709, "step": 260 }, { "epoch": 0.22110758521854848, "grad_norm": 0.7408131957054138, "learning_rate": 0.0001, "loss": 0.5753, "step": 270 }, { "epoch": 0.22929675504145766, "grad_norm": 0.858842134475708, "learning_rate": 0.0001, "loss": 0.5649, "step": 280 }, { "epoch": 0.23748592486436687, "grad_norm": 0.9077485799789429, "learning_rate": 0.0001, "loss": 0.5769, "step": 290 }, { "epoch": 0.24567509468727608, "grad_norm": 1.1604504585266113, "learning_rate": 0.0001, "loss": 0.5558, "step": 300 }, { "epoch": 0.2538642645101853, "grad_norm": 0.7758939266204834, "learning_rate": 0.0001, "loss": 0.5711, "step": 310 }, { "epoch": 0.26205343433309447, "grad_norm": 1.2495347261428833, "learning_rate": 0.0001, "loss": 0.5845, "step": 320 }, { "epoch": 0.2702426041560037, "grad_norm": 0.8986139297485352, "learning_rate": 0.0001, "loss": 0.5483, "step": 330 }, { "epoch": 0.2784317739789129, "grad_norm": 0.7849992513656616, "learning_rate": 0.0001, "loss": 0.5682, "step": 340 }, { "epoch": 0.28662094380182207, "grad_norm": 0.8479835987091064, "learning_rate": 0.0001, "loss": 0.5381, "step": 350 }, { "epoch": 0.2948101136247313, "grad_norm": 0.9730711579322815, "learning_rate": 0.0001, "loss": 0.5572, "step": 360 }, { "epoch": 0.3029992834476405, "grad_norm": 0.8130871653556824, "learning_rate": 0.0001, "loss": 0.5285, "step": 370 }, { "epoch": 0.3111884532705497, "grad_norm": 0.7159671783447266, "learning_rate": 0.0001, "loss": 0.529, "step": 380 }, { "epoch": 0.3193776230934589, "grad_norm": 0.7021393775939941, "learning_rate": 0.0001, "loss": 0.5371, "step": 390 }, { "epoch": 0.3275667929163681, "grad_norm": 0.9046493172645569, "learning_rate": 0.0001, "loss": 0.5412, "step": 400 }, { "epoch": 0.3275667929163681, "eval_loss": 0.43820247054100037, "eval_runtime": 1014.548, "eval_samples_per_second": 35.484, "eval_steps_per_second": 17.742, "step": 400 }, { "epoch": 0.3357559627392773, "grad_norm": 0.9188250303268433, "learning_rate": 0.0001, "loss": 0.5567, "step": 410 }, { "epoch": 0.3439451325621865, "grad_norm": 0.6707028150558472, "learning_rate": 0.0001, "loss": 0.5303, "step": 420 }, { "epoch": 0.3521343023850957, "grad_norm": 1.1012392044067383, "learning_rate": 0.0001, "loss": 0.5356, "step": 430 }, { "epoch": 0.3603234722080049, "grad_norm": 0.6614859104156494, "learning_rate": 0.0001, "loss": 0.5443, "step": 440 }, { "epoch": 0.3685126420309141, "grad_norm": 0.8753280639648438, "learning_rate": 0.0001, "loss": 0.5278, "step": 450 }, { "epoch": 0.37670181185382334, "grad_norm": 0.713115394115448, "learning_rate": 0.0001, "loss": 0.5344, "step": 460 }, { "epoch": 0.3848909816767325, "grad_norm": 1.034822940826416, "learning_rate": 0.0001, "loss": 0.511, "step": 470 }, { "epoch": 0.3930801514996417, "grad_norm": 0.7497478723526001, "learning_rate": 0.0001, "loss": 0.5143, "step": 480 }, { "epoch": 0.40126932132255094, "grad_norm": 0.9680531024932861, "learning_rate": 0.0001, "loss": 0.5372, "step": 490 }, { "epoch": 0.4094584911454601, "grad_norm": 0.6382943391799927, "learning_rate": 0.0001, "loss": 0.5242, "step": 500 }, { "epoch": 0.4176476609683693, "grad_norm": 0.7657376527786255, "learning_rate": 0.0001, "loss": 0.5125, "step": 510 }, { "epoch": 0.42583683079127854, "grad_norm": 0.9952341914176941, "learning_rate": 0.0001, "loss": 0.507, "step": 520 }, { "epoch": 0.4340260006141877, "grad_norm": 0.8375737071037292, "learning_rate": 0.0001, "loss": 0.5142, "step": 530 }, { "epoch": 0.44221517043709696, "grad_norm": 0.7849007844924927, "learning_rate": 0.0001, "loss": 0.5049, "step": 540 }, { "epoch": 0.45040434026000614, "grad_norm": 0.7818809151649475, "learning_rate": 0.0001, "loss": 0.4921, "step": 550 }, { "epoch": 0.4585935100829153, "grad_norm": 0.6086965203285217, "learning_rate": 0.0001, "loss": 0.5254, "step": 560 }, { "epoch": 0.46678267990582456, "grad_norm": 0.6038303971290588, "learning_rate": 0.0001, "loss": 0.5044, "step": 570 }, { "epoch": 0.47497184972873374, "grad_norm": 0.7107850313186646, "learning_rate": 0.0001, "loss": 0.5189, "step": 580 }, { "epoch": 0.483161019551643, "grad_norm": 0.7399956583976746, "learning_rate": 0.0001, "loss": 0.4814, "step": 590 }, { "epoch": 0.49135018937455216, "grad_norm": 0.7447919249534607, "learning_rate": 0.0001, "loss": 0.5273, "step": 600 }, { "epoch": 0.49135018937455216, "eval_loss": 0.41115716099739075, "eval_runtime": 974.1489, "eval_samples_per_second": 36.955, "eval_steps_per_second": 18.478, "step": 600 }, { "epoch": 0.49953935919746134, "grad_norm": 0.7186319231987, "learning_rate": 0.0001, "loss": 0.4964, "step": 610 }, { "epoch": 0.5077285290203706, "grad_norm": 0.6659027338027954, "learning_rate": 0.0001, "loss": 0.4813, "step": 620 }, { "epoch": 0.5159176988432798, "grad_norm": 0.8509200811386108, "learning_rate": 0.0001, "loss": 0.5007, "step": 630 }, { "epoch": 0.5241068686661889, "grad_norm": 0.7163340449333191, "learning_rate": 0.0001, "loss": 0.4918, "step": 640 }, { "epoch": 0.5322960384890981, "grad_norm": 0.6762372851371765, "learning_rate": 0.0001, "loss": 0.4721, "step": 650 }, { "epoch": 0.5404852083120074, "grad_norm": 0.741247832775116, "learning_rate": 0.0001, "loss": 0.5035, "step": 660 }, { "epoch": 0.5486743781349166, "grad_norm": 0.6757375001907349, "learning_rate": 0.0001, "loss": 0.5044, "step": 670 }, { "epoch": 0.5568635479578258, "grad_norm": 0.6605280637741089, "learning_rate": 0.0001, "loss": 0.5036, "step": 680 }, { "epoch": 0.565052717780735, "grad_norm": 0.7928522825241089, "learning_rate": 0.0001, "loss": 0.4884, "step": 690 }, { "epoch": 0.5732418876036441, "grad_norm": 0.7482662796974182, "learning_rate": 0.0001, "loss": 0.4831, "step": 700 }, { "epoch": 0.5814310574265534, "grad_norm": 0.7094814777374268, "learning_rate": 0.0001, "loss": 0.493, "step": 710 }, { "epoch": 0.5896202272494626, "grad_norm": 0.708988606929779, "learning_rate": 0.0001, "loss": 0.4878, "step": 720 }, { "epoch": 0.5978093970723718, "grad_norm": 0.7677808403968811, "learning_rate": 0.0001, "loss": 0.4819, "step": 730 }, { "epoch": 0.605998566895281, "grad_norm": 0.6716774702072144, "learning_rate": 0.0001, "loss": 0.505, "step": 740 }, { "epoch": 0.6141877367181902, "grad_norm": 0.6802059412002563, "learning_rate": 0.0001, "loss": 0.4612, "step": 750 }, { "epoch": 0.6223769065410995, "grad_norm": 0.7577608227729797, "learning_rate": 0.0001, "loss": 0.4814, "step": 760 }, { "epoch": 0.6305660763640086, "grad_norm": 0.6441067457199097, "learning_rate": 0.0001, "loss": 0.4838, "step": 770 }, { "epoch": 0.6387552461869178, "grad_norm": 0.666208028793335, "learning_rate": 0.0001, "loss": 0.4908, "step": 780 }, { "epoch": 0.646944416009827, "grad_norm": 0.7665196657180786, "learning_rate": 0.0001, "loss": 0.4716, "step": 790 }, { "epoch": 0.6551335858327362, "grad_norm": 0.8056005239486694, "learning_rate": 0.0001, "loss": 0.4717, "step": 800 }, { "epoch": 0.6551335858327362, "eval_loss": 0.41053226590156555, "eval_runtime": 952.7571, "eval_samples_per_second": 37.785, "eval_steps_per_second": 18.893, "step": 800 }, { "epoch": 0.6633227556556454, "grad_norm": 1.5036696195602417, "learning_rate": 0.0001, "loss": 0.4711, "step": 810 }, { "epoch": 0.6715119254785546, "grad_norm": 0.7547106146812439, "learning_rate": 0.0001, "loss": 0.476, "step": 820 }, { "epoch": 0.6797010953014638, "grad_norm": 0.7936639785766602, "learning_rate": 0.0001, "loss": 0.4762, "step": 830 }, { "epoch": 0.687890265124373, "grad_norm": 0.7227616310119629, "learning_rate": 0.0001, "loss": 0.4798, "step": 840 }, { "epoch": 0.6960794349472822, "grad_norm": 0.7741641402244568, "learning_rate": 0.0001, "loss": 0.4992, "step": 850 }, { "epoch": 0.7042686047701914, "grad_norm": 0.6723213791847229, "learning_rate": 0.0001, "loss": 0.4818, "step": 860 }, { "epoch": 0.7124577745931007, "grad_norm": 0.5631268620491028, "learning_rate": 0.0001, "loss": 0.481, "step": 870 }, { "epoch": 0.7206469444160098, "grad_norm": 0.7071460485458374, "learning_rate": 0.0001, "loss": 0.4744, "step": 880 }, { "epoch": 0.728836114238919, "grad_norm": 0.6503288745880127, "learning_rate": 0.0001, "loss": 0.4466, "step": 890 }, { "epoch": 0.7370252840618282, "grad_norm": 0.6489100456237793, "learning_rate": 0.0001, "loss": 0.4708, "step": 900 }, { "epoch": 0.7452144538847374, "grad_norm": 0.7342277765274048, "learning_rate": 0.0001, "loss": 0.4687, "step": 910 }, { "epoch": 0.7534036237076467, "grad_norm": 0.6502553224563599, "learning_rate": 0.0001, "loss": 0.4649, "step": 920 }, { "epoch": 0.7615927935305559, "grad_norm": 0.6611005663871765, "learning_rate": 0.0001, "loss": 0.4771, "step": 930 }, { "epoch": 0.769781963353465, "grad_norm": 0.5838377475738525, "learning_rate": 0.0001, "loss": 0.4671, "step": 940 }, { "epoch": 0.7779711331763742, "grad_norm": 0.6084617972373962, "learning_rate": 0.0001, "loss": 0.4701, "step": 950 }, { "epoch": 0.7861603029992834, "grad_norm": 0.6489084959030151, "learning_rate": 0.0001, "loss": 0.4586, "step": 960 }, { "epoch": 0.7943494728221927, "grad_norm": 0.6316090226173401, "learning_rate": 0.0001, "loss": 0.451, "step": 970 }, { "epoch": 0.8025386426451019, "grad_norm": 0.6724498271942139, "learning_rate": 0.0001, "loss": 0.4776, "step": 980 }, { "epoch": 0.8107278124680111, "grad_norm": 0.6327974796295166, "learning_rate": 0.0001, "loss": 0.4708, "step": 990 }, { "epoch": 0.8189169822909202, "grad_norm": 0.8041887879371643, "learning_rate": 0.0001, "loss": 0.4835, "step": 1000 }, { "epoch": 0.8189169822909202, "eval_loss": 0.39249786734580994, "eval_runtime": 953.0333, "eval_samples_per_second": 37.774, "eval_steps_per_second": 18.887, "step": 1000 }, { "epoch": 0.8271061521138294, "grad_norm": 0.7333071827888489, "learning_rate": 0.0001, "loss": 0.4662, "step": 1010 }, { "epoch": 0.8352953219367386, "grad_norm": 0.6036032438278198, "learning_rate": 0.0001, "loss": 0.449, "step": 1020 }, { "epoch": 0.8434844917596479, "grad_norm": 0.6087955236434937, "learning_rate": 0.0001, "loss": 0.4651, "step": 1030 }, { "epoch": 0.8516736615825571, "grad_norm": 0.8231128454208374, "learning_rate": 0.0001, "loss": 0.4672, "step": 1040 }, { "epoch": 0.8598628314054663, "grad_norm": 0.6176358461380005, "learning_rate": 0.0001, "loss": 0.4616, "step": 1050 }, { "epoch": 0.8680520012283754, "grad_norm": 0.5654678344726562, "learning_rate": 0.0001, "loss": 0.437, "step": 1060 }, { "epoch": 0.8762411710512846, "grad_norm": 0.6934413909912109, "learning_rate": 0.0001, "loss": 0.4302, "step": 1070 }, { "epoch": 0.8844303408741939, "grad_norm": 0.6048303842544556, "learning_rate": 0.0001, "loss": 0.4584, "step": 1080 }, { "epoch": 0.8926195106971031, "grad_norm": 0.6487530469894409, "learning_rate": 0.0001, "loss": 0.4647, "step": 1090 }, { "epoch": 0.9008086805200123, "grad_norm": 0.6674748659133911, "learning_rate": 0.0001, "loss": 0.467, "step": 1100 }, { "epoch": 0.9089978503429215, "grad_norm": 0.6509521007537842, "learning_rate": 0.0001, "loss": 0.4525, "step": 1110 }, { "epoch": 0.9171870201658306, "grad_norm": 0.6949372291564941, "learning_rate": 0.0001, "loss": 0.4663, "step": 1120 }, { "epoch": 0.9253761899887399, "grad_norm": 0.7340756058692932, "learning_rate": 0.0001, "loss": 0.4297, "step": 1130 }, { "epoch": 0.9335653598116491, "grad_norm": 0.5543187856674194, "learning_rate": 0.0001, "loss": 0.4606, "step": 1140 }, { "epoch": 0.9417545296345583, "grad_norm": 0.6061782836914062, "learning_rate": 0.0001, "loss": 0.4589, "step": 1150 }, { "epoch": 0.9499436994574675, "grad_norm": 0.6638914942741394, "learning_rate": 0.0001, "loss": 0.4252, "step": 1160 }, { "epoch": 0.9581328692803767, "grad_norm": 0.5913025736808777, "learning_rate": 0.0001, "loss": 0.44, "step": 1170 }, { "epoch": 0.966322039103286, "grad_norm": 0.5672889947891235, "learning_rate": 0.0001, "loss": 0.4659, "step": 1180 }, { "epoch": 0.9745112089261951, "grad_norm": 0.5614264607429504, "learning_rate": 0.0001, "loss": 0.4471, "step": 1190 }, { "epoch": 0.9827003787491043, "grad_norm": 0.5809562802314758, "learning_rate": 0.0001, "loss": 0.4288, "step": 1200 }, { "epoch": 0.9827003787491043, "eval_loss": 0.37333595752716064, "eval_runtime": 961.0219, "eval_samples_per_second": 37.46, "eval_steps_per_second": 18.73, "step": 1200 } ], "logging_steps": 10, "max_steps": 1221, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 400, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 7.304264713469952e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }