diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,14433 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.8806026624471123, + "eval_steps": 200, + "global_step": 3200, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0005503766640294451, + "grad_norm": 2.2015435695648193, + "learning_rate": 1.4775011317868612e-06, + "loss": 0.7802, + "step": 2 + }, + { + "epoch": 0.0011007533280588903, + "grad_norm": 2.0623114109039307, + "learning_rate": 2.9550022635737224e-06, + "loss": 0.6659, + "step": 4 + }, + { + "epoch": 0.0016511299920883354, + "grad_norm": 0.8444932699203491, + "learning_rate": 3.819285020442103e-06, + "loss": 0.6275, + "step": 6 + }, + { + "epoch": 0.0022015066561177805, + "grad_norm": 0.7291238307952881, + "learning_rate": 4.432503395360583e-06, + "loss": 0.5955, + "step": 8 + }, + { + "epoch": 0.0027518833201472257, + "grad_norm": 0.591098427772522, + "learning_rate": 4.90815251991065e-06, + "loss": 0.5906, + "step": 10 + }, + { + "epoch": 0.003302259984176671, + "grad_norm": 0.6075527667999268, + "learning_rate": 5.2967861522289644e-06, + "loss": 0.5451, + "step": 12 + }, + { + "epoch": 0.003852636648206116, + "grad_norm": 0.5598031878471375, + "learning_rate": 5.625371206454386e-06, + "loss": 0.5539, + "step": 14 + }, + { + "epoch": 0.004403013312235561, + "grad_norm": 0.5352339148521423, + "learning_rate": 5.910004527147445e-06, + "loss": 0.5452, + "step": 16 + }, + { + "epoch": 0.004953389976265006, + "grad_norm": 0.524741530418396, + "learning_rate": 6.161068909097345e-06, + "loss": 0.5536, + "step": 18 + }, + { + "epoch": 0.005503766640294451, + "grad_norm": 0.4852159321308136, + "learning_rate": 6.38565365169751e-06, + "loss": 0.5439, + "step": 20 + }, + { + "epoch": 0.0060541433043238965, + "grad_norm": 0.4764852225780487, + "learning_rate": 6.5888152636627215e-06, + "loss": 0.5468, + "step": 22 + }, + { + "epoch": 0.006604519968353342, + "grad_norm": 0.463278591632843, + "learning_rate": 6.774287284015826e-06, + "loss": 0.541, + "step": 24 + }, + { + "epoch": 0.007154896632382787, + "grad_norm": 0.4566305875778198, + "learning_rate": 6.944905003449378e-06, + "loss": 0.5258, + "step": 26 + }, + { + "epoch": 0.007705273296412232, + "grad_norm": 0.4572094678878784, + "learning_rate": 7.102872338241248e-06, + "loss": 0.5385, + "step": 28 + }, + { + "epoch": 0.008255649960441678, + "grad_norm": 0.4581094980239868, + "learning_rate": 7.2499364085658915e-06, + "loss": 0.5258, + "step": 30 + }, + { + "epoch": 0.008806026624471122, + "grad_norm": 0.4602491557598114, + "learning_rate": 7.387505658934305e-06, + "loss": 0.5239, + "step": 32 + }, + { + "epoch": 0.009356403288500568, + "grad_norm": 0.4633028507232666, + "learning_rate": 7.516732105870977e-06, + "loss": 0.5237, + "step": 34 + }, + { + "epoch": 0.009906779952530012, + "grad_norm": 0.4267115592956543, + "learning_rate": 7.638570040884206e-06, + "loss": 0.5467, + "step": 36 + }, + { + "epoch": 0.010457156616559458, + "grad_norm": 0.4587521255016327, + "learning_rate": 7.753818840648305e-06, + "loss": 0.5282, + "step": 38 + }, + { + "epoch": 0.011007533280588903, + "grad_norm": 0.44529175758361816, + "learning_rate": 7.863154783484372e-06, + "loss": 0.536, + "step": 40 + }, + { + "epoch": 0.011557909944618349, + "grad_norm": 0.436199814081192, + "learning_rate": 7.967155095109629e-06, + "loss": 0.5259, + "step": 42 + }, + { + "epoch": 0.012108286608647793, + "grad_norm": 0.43157511949539185, + "learning_rate": 8.066316395449581e-06, + "loss": 0.5173, + "step": 44 + }, + { + "epoch": 0.012658663272677239, + "grad_norm": 0.4393196105957031, + "learning_rate": 8.161069041569085e-06, + "loss": 0.5037, + "step": 46 + }, + { + "epoch": 0.013209039936706683, + "grad_norm": 0.5085024237632751, + "learning_rate": 8.251788415802687e-06, + "loss": 0.5015, + "step": 48 + }, + { + "epoch": 0.01375941660073613, + "grad_norm": 0.40056705474853516, + "learning_rate": 8.338803908034438e-06, + "loss": 0.5194, + "step": 50 + }, + { + "epoch": 0.014309793264765574, + "grad_norm": 0.43617382645606995, + "learning_rate": 8.422406135236239e-06, + "loss": 0.5327, + "step": 52 + }, + { + "epoch": 0.01486016992879502, + "grad_norm": 0.42528873682022095, + "learning_rate": 8.502852797752587e-06, + "loss": 0.5245, + "step": 54 + }, + { + "epoch": 0.015410546592824464, + "grad_norm": 0.40340831875801086, + "learning_rate": 8.58037347002811e-06, + "loss": 0.5059, + "step": 56 + }, + { + "epoch": 0.015960923256853908, + "grad_norm": 0.40859195590019226, + "learning_rate": 8.65517355028691e-06, + "loss": 0.5136, + "step": 58 + }, + { + "epoch": 0.016511299920883356, + "grad_norm": 0.42262887954711914, + "learning_rate": 8.727437540352753e-06, + "loss": 0.5011, + "step": 60 + }, + { + "epoch": 0.0170616765849128, + "grad_norm": 0.4255228638648987, + "learning_rate": 8.79733178747776e-06, + "loss": 0.5218, + "step": 62 + }, + { + "epoch": 0.017612053248942244, + "grad_norm": 0.4350854754447937, + "learning_rate": 8.865006790721166e-06, + "loss": 0.5334, + "step": 64 + }, + { + "epoch": 0.01816242991297169, + "grad_norm": 0.41395291686058044, + "learning_rate": 8.930599152317962e-06, + "loss": 0.5233, + "step": 66 + }, + { + "epoch": 0.018712806577001136, + "grad_norm": 0.4127484858036041, + "learning_rate": 8.99423323765784e-06, + "loss": 0.5143, + "step": 68 + }, + { + "epoch": 0.01926318324103058, + "grad_norm": 0.42464280128479004, + "learning_rate": 9.056022594578175e-06, + "loss": 0.5164, + "step": 70 + }, + { + "epoch": 0.019813559905060025, + "grad_norm": 0.4011682868003845, + "learning_rate": 9.116071172671068e-06, + "loss": 0.5036, + "step": 72 + }, + { + "epoch": 0.02036393656908947, + "grad_norm": 0.39912551641464233, + "learning_rate": 9.174474375494509e-06, + "loss": 0.5038, + "step": 74 + }, + { + "epoch": 0.020914313233118917, + "grad_norm": 0.40526625514030457, + "learning_rate": 9.231319972435167e-06, + "loss": 0.518, + "step": 76 + }, + { + "epoch": 0.02146468989714836, + "grad_norm": 0.4195484220981598, + "learning_rate": 9.28668889210462e-06, + "loss": 0.5065, + "step": 78 + }, + { + "epoch": 0.022015066561177805, + "grad_norm": 0.4483351409435272, + "learning_rate": 9.340655915271231e-06, + "loss": 0.5128, + "step": 80 + }, + { + "epoch": 0.02256544322520725, + "grad_norm": 0.42789894342422485, + "learning_rate": 9.393290282217048e-06, + "loss": 0.525, + "step": 82 + }, + { + "epoch": 0.023115819889236697, + "grad_norm": 0.434644490480423, + "learning_rate": 9.444656226896488e-06, + "loss": 0.5248, + "step": 84 + }, + { + "epoch": 0.02366619655326614, + "grad_norm": 0.4532856345176697, + "learning_rate": 9.494813448234365e-06, + "loss": 0.5226, + "step": 86 + }, + { + "epoch": 0.024216573217295586, + "grad_norm": 0.4103749692440033, + "learning_rate": 9.543817527236444e-06, + "loss": 0.5034, + "step": 88 + }, + { + "epoch": 0.02476694988132503, + "grad_norm": 0.4208613336086273, + "learning_rate": 9.591720297221133e-06, + "loss": 0.5214, + "step": 90 + }, + { + "epoch": 0.025317326545354478, + "grad_norm": 0.4020327627658844, + "learning_rate": 9.638570173355947e-06, + "loss": 0.5047, + "step": 92 + }, + { + "epoch": 0.025867703209383922, + "grad_norm": 0.4074559509754181, + "learning_rate": 9.684412446751251e-06, + "loss": 0.4999, + "step": 94 + }, + { + "epoch": 0.026418079873413366, + "grad_norm": 0.43330731987953186, + "learning_rate": 9.729289547589548e-06, + "loss": 0.5089, + "step": 96 + }, + { + "epoch": 0.02696845653744281, + "grad_norm": 0.42775431275367737, + "learning_rate": 9.773241281121913e-06, + "loss": 0.5169, + "step": 98 + }, + { + "epoch": 0.02751883320147226, + "grad_norm": 0.421403706073761, + "learning_rate": 9.8163050398213e-06, + "loss": 0.5123, + "step": 100 + }, + { + "epoch": 0.028069209865501703, + "grad_norm": 0.42337778210639954, + "learning_rate": 9.858515994526218e-06, + "loss": 0.5116, + "step": 102 + }, + { + "epoch": 0.028619586529531147, + "grad_norm": 0.4156826138496399, + "learning_rate": 9.8999072670231e-06, + "loss": 0.5077, + "step": 104 + }, + { + "epoch": 0.02916996319356059, + "grad_norm": 0.4544354975223541, + "learning_rate": 9.9405100861891e-06, + "loss": 0.5099, + "step": 106 + }, + { + "epoch": 0.02972033985759004, + "grad_norm": 0.4015970528125763, + "learning_rate": 9.980353929539448e-06, + "loss": 0.5049, + "step": 108 + }, + { + "epoch": 0.030270716521619483, + "grad_norm": 0.3907098174095154, + "learning_rate": 1e-05, + "loss": 0.5202, + "step": 110 + }, + { + "epoch": 0.030821093185648928, + "grad_norm": 0.4184499979019165, + "learning_rate": 1e-05, + "loss": 0.5085, + "step": 112 + }, + { + "epoch": 0.031371469849678375, + "grad_norm": 0.47195565700531006, + "learning_rate": 1e-05, + "loss": 0.5161, + "step": 114 + }, + { + "epoch": 0.031921846513707816, + "grad_norm": 0.43992695212364197, + "learning_rate": 1e-05, + "loss": 0.4978, + "step": 116 + }, + { + "epoch": 0.032472223177737264, + "grad_norm": 0.43099331855773926, + "learning_rate": 1e-05, + "loss": 0.5035, + "step": 118 + }, + { + "epoch": 0.03302259984176671, + "grad_norm": 0.44256317615509033, + "learning_rate": 1e-05, + "loss": 0.4991, + "step": 120 + }, + { + "epoch": 0.03357297650579615, + "grad_norm": 0.42082124948501587, + "learning_rate": 1e-05, + "loss": 0.5028, + "step": 122 + }, + { + "epoch": 0.0341233531698256, + "grad_norm": 0.38576358556747437, + "learning_rate": 1e-05, + "loss": 0.5081, + "step": 124 + }, + { + "epoch": 0.03467372983385505, + "grad_norm": 0.3880733251571655, + "learning_rate": 1e-05, + "loss": 0.5001, + "step": 126 + }, + { + "epoch": 0.03522410649788449, + "grad_norm": 0.41802075505256653, + "learning_rate": 1e-05, + "loss": 0.5056, + "step": 128 + }, + { + "epoch": 0.035774483161913936, + "grad_norm": 0.3949527144432068, + "learning_rate": 1e-05, + "loss": 0.5155, + "step": 130 + }, + { + "epoch": 0.03632485982594338, + "grad_norm": 0.4038969576358795, + "learning_rate": 1e-05, + "loss": 0.5056, + "step": 132 + }, + { + "epoch": 0.036875236489972825, + "grad_norm": 0.40195325016975403, + "learning_rate": 1e-05, + "loss": 0.4968, + "step": 134 + }, + { + "epoch": 0.03742561315400227, + "grad_norm": 0.3946043848991394, + "learning_rate": 1e-05, + "loss": 0.4981, + "step": 136 + }, + { + "epoch": 0.037975989818031713, + "grad_norm": 0.3914756774902344, + "learning_rate": 1e-05, + "loss": 0.5, + "step": 138 + }, + { + "epoch": 0.03852636648206116, + "grad_norm": 0.4295148551464081, + "learning_rate": 1e-05, + "loss": 0.5147, + "step": 140 + }, + { + "epoch": 0.03907674314609061, + "grad_norm": 0.40092742443084717, + "learning_rate": 1e-05, + "loss": 0.5196, + "step": 142 + }, + { + "epoch": 0.03962711981012005, + "grad_norm": 0.41200628876686096, + "learning_rate": 1e-05, + "loss": 0.5031, + "step": 144 + }, + { + "epoch": 0.0401774964741495, + "grad_norm": 0.43834391236305237, + "learning_rate": 1e-05, + "loss": 0.5047, + "step": 146 + }, + { + "epoch": 0.04072787313817894, + "grad_norm": 0.3940436542034149, + "learning_rate": 1e-05, + "loss": 0.4912, + "step": 148 + }, + { + "epoch": 0.041278249802208386, + "grad_norm": 0.3873765170574188, + "learning_rate": 1e-05, + "loss": 0.482, + "step": 150 + }, + { + "epoch": 0.041828626466237834, + "grad_norm": 0.4272858798503876, + "learning_rate": 1e-05, + "loss": 0.4923, + "step": 152 + }, + { + "epoch": 0.042379003130267275, + "grad_norm": 0.40542730689048767, + "learning_rate": 1e-05, + "loss": 0.4892, + "step": 154 + }, + { + "epoch": 0.04292937979429672, + "grad_norm": 0.38277357816696167, + "learning_rate": 1e-05, + "loss": 0.517, + "step": 156 + }, + { + "epoch": 0.04347975645832617, + "grad_norm": 0.39421385526657104, + "learning_rate": 1e-05, + "loss": 0.503, + "step": 158 + }, + { + "epoch": 0.04403013312235561, + "grad_norm": 0.3984109163284302, + "learning_rate": 1e-05, + "loss": 0.5074, + "step": 160 + }, + { + "epoch": 0.04458050978638506, + "grad_norm": 0.40513876080513, + "learning_rate": 1e-05, + "loss": 0.5092, + "step": 162 + }, + { + "epoch": 0.0451308864504145, + "grad_norm": 0.45850449800491333, + "learning_rate": 1e-05, + "loss": 0.5086, + "step": 164 + }, + { + "epoch": 0.04568126311444395, + "grad_norm": 0.4050631821155548, + "learning_rate": 1e-05, + "loss": 0.5073, + "step": 166 + }, + { + "epoch": 0.046231639778473395, + "grad_norm": 0.41050952672958374, + "learning_rate": 1e-05, + "loss": 0.5007, + "step": 168 + }, + { + "epoch": 0.046782016442502836, + "grad_norm": 0.39902788400650024, + "learning_rate": 1e-05, + "loss": 0.4941, + "step": 170 + }, + { + "epoch": 0.04733239310653228, + "grad_norm": 0.4421572983264923, + "learning_rate": 1e-05, + "loss": 0.4988, + "step": 172 + }, + { + "epoch": 0.04788276977056173, + "grad_norm": 0.4092646837234497, + "learning_rate": 1e-05, + "loss": 0.5001, + "step": 174 + }, + { + "epoch": 0.04843314643459117, + "grad_norm": 0.4195966124534607, + "learning_rate": 1e-05, + "loss": 0.4964, + "step": 176 + }, + { + "epoch": 0.04898352309862062, + "grad_norm": 0.3937481641769409, + "learning_rate": 1e-05, + "loss": 0.4977, + "step": 178 + }, + { + "epoch": 0.04953389976265006, + "grad_norm": 0.434950590133667, + "learning_rate": 1e-05, + "loss": 0.5054, + "step": 180 + }, + { + "epoch": 0.05008427642667951, + "grad_norm": 0.40112894773483276, + "learning_rate": 1e-05, + "loss": 0.494, + "step": 182 + }, + { + "epoch": 0.050634653090708956, + "grad_norm": 0.42001938819885254, + "learning_rate": 1e-05, + "loss": 0.4744, + "step": 184 + }, + { + "epoch": 0.0511850297547384, + "grad_norm": 0.4066455364227295, + "learning_rate": 1e-05, + "loss": 0.4838, + "step": 186 + }, + { + "epoch": 0.051735406418767844, + "grad_norm": 0.3934157192707062, + "learning_rate": 1e-05, + "loss": 0.5017, + "step": 188 + }, + { + "epoch": 0.05228578308279729, + "grad_norm": 0.38877320289611816, + "learning_rate": 1e-05, + "loss": 0.5018, + "step": 190 + }, + { + "epoch": 0.05283615974682673, + "grad_norm": 0.39771756529808044, + "learning_rate": 1e-05, + "loss": 0.485, + "step": 192 + }, + { + "epoch": 0.05338653641085618, + "grad_norm": 0.3938674330711365, + "learning_rate": 1e-05, + "loss": 0.5034, + "step": 194 + }, + { + "epoch": 0.05393691307488562, + "grad_norm": 0.40473559498786926, + "learning_rate": 1e-05, + "loss": 0.5082, + "step": 196 + }, + { + "epoch": 0.05448728973891507, + "grad_norm": 0.3977149426937103, + "learning_rate": 1e-05, + "loss": 0.4997, + "step": 198 + }, + { + "epoch": 0.05503766640294452, + "grad_norm": 0.39340054988861084, + "learning_rate": 1e-05, + "loss": 0.4859, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_merge_loss": 0.4414624571800232, + "eval_merge_runtime": 600.1539, + "eval_merge_samples_per_second": 56.239, + "eval_merge_steps_per_second": 2.344, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_new_aug_datas_filtered.json_loss": 0.5691156983375549, + "eval_new_aug_datas_filtered.json_runtime": 10.6767, + "eval_new_aug_datas_filtered.json_samples_per_second": 71.839, + "eval_new_aug_datas_filtered.json_steps_per_second": 2.997, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_sharegpt_gpt4.json_loss": 0.8223738670349121, + "eval_sharegpt_gpt4.json_runtime": 31.6183, + "eval_sharegpt_gpt4.json_samples_per_second": 58.858, + "eval_sharegpt_gpt4.json_steps_per_second": 2.467, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_Table_GPT.json_loss": 0.09253557026386261, + "eval_Table_GPT.json_runtime": 24.9748, + "eval_Table_GPT.json_samples_per_second": 83.804, + "eval_Table_GPT.json_steps_per_second": 3.524, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_gpt_4o_200k.json_loss": 0.849287211894989, + "eval_gpt_4o_200k.json_runtime": 48.5339, + "eval_gpt_4o_200k.json_samples_per_second": 129.415, + "eval_gpt_4o_200k.json_steps_per_second": 5.398, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_multi_turn_datas.json_loss": 0.3907540738582611, + "eval_multi_turn_datas.json_runtime": 75.6133, + "eval_multi_turn_datas.json_samples_per_second": 52.927, + "eval_multi_turn_datas.json_steps_per_second": 2.209, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_table_python_code_datas.json_loss": 0.33119720220565796, + "eval_table_python_code_datas.json_runtime": 43.1313, + "eval_table_python_code_datas.json_samples_per_second": 50.056, + "eval_table_python_code_datas.json_steps_per_second": 2.087, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_tabular_llm_data.json_loss": 0.14601922035217285, + "eval_tabular_llm_data.json_runtime": 8.7785, + "eval_tabular_llm_data.json_samples_per_second": 28.023, + "eval_tabular_llm_data.json_steps_per_second": 1.253, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_python_code_critic_21k.json_loss": 0.625038743019104, + "eval_python_code_critic_21k.json_runtime": 3.237, + "eval_python_code_critic_21k.json_samples_per_second": 184.43, + "eval_python_code_critic_21k.json_steps_per_second": 7.723, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_all_merge_table_dataset.json_loss": 0.09772461652755737, + "eval_all_merge_table_dataset.json_runtime": 24.3077, + "eval_all_merge_table_dataset.json_samples_per_second": 29.291, + "eval_all_merge_table_dataset.json_steps_per_second": 1.234, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_code_feedback_multi_turn.json_loss": 0.6093290448188782, + "eval_code_feedback_multi_turn.json_runtime": 32.4589, + "eval_code_feedback_multi_turn.json_samples_per_second": 67.809, + "eval_code_feedback_multi_turn.json_steps_per_second": 2.834, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_ultrainteract_sft.json_loss": 0.4469935894012451, + "eval_ultrainteract_sft.json_runtime": 8.6702, + "eval_ultrainteract_sft.json_samples_per_second": 167.931, + "eval_ultrainteract_sft.json_steps_per_second": 7.036, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_synthetic_text_to_sql.json_loss": 0.11159003525972366, + "eval_synthetic_text_to_sql.json_runtime": 0.1306, + "eval_synthetic_text_to_sql.json_samples_per_second": 260.355, + "eval_synthetic_text_to_sql.json_steps_per_second": 15.315, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_sft_react_sql_datas.json_loss": 0.6847189664840698, + "eval_sft_react_sql_datas.json_runtime": 7.8434, + "eval_sft_react_sql_datas.json_samples_per_second": 40.034, + "eval_sft_react_sql_datas.json_steps_per_second": 1.785, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_all_merge_code.json_loss": 0.32269543409347534, + "eval_all_merge_code.json_runtime": 0.3287, + "eval_all_merge_code.json_samples_per_second": 191.649, + "eval_all_merge_code.json_steps_per_second": 9.126, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_magpie_datas.json_loss": 0.4600640833377838, + "eval_magpie_datas.json_runtime": 2.2095, + "eval_magpie_datas.json_samples_per_second": 77.844, + "eval_magpie_datas.json_steps_per_second": 3.621, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_train_data_for_qwen.json_loss": 0.017207294702529907, + "eval_train_data_for_qwen.json_runtime": 0.2494, + "eval_train_data_for_qwen.json_samples_per_second": 40.095, + "eval_train_data_for_qwen.json_steps_per_second": 4.01, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_alpaca_cleaned.json_loss": 0.9374485015869141, + "eval_alpaca_cleaned.json_runtime": 0.1149, + "eval_alpaca_cleaned.json_samples_per_second": 234.896, + "eval_alpaca_cleaned.json_steps_per_second": 17.4, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_agent_instruct.json_loss": 0.23996739089488983, + "eval_agent_instruct.json_runtime": 0.5126, + "eval_agent_instruct.json_samples_per_second": 93.639, + "eval_agent_instruct.json_steps_per_second": 3.902, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_MathInstruct.json_loss": 0.2269323617219925, + "eval_MathInstruct.json_runtime": 0.3472, + "eval_MathInstruct.json_samples_per_second": 164.184, + "eval_MathInstruct.json_steps_per_second": 8.641, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_tested_143k_python_alpaca.json_loss": 0.4513254165649414, + "eval_tested_143k_python_alpaca.json_runtime": 0.3017, + "eval_tested_143k_python_alpaca.json_samples_per_second": 112.684, + "eval_tested_143k_python_alpaca.json_steps_per_second": 6.628, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_xlam_function_calling_60k.json_loss": 0.011208846233785152, + "eval_xlam_function_calling_60k.json_runtime": 0.1011, + "eval_xlam_function_calling_60k.json_samples_per_second": 227.556, + "eval_xlam_function_calling_60k.json_steps_per_second": 9.894, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_alpaca_data_gpt4_chinese.json_loss": 1.6813441514968872, + "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0523, + "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 306.208, + "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.138, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_alpaca_gpt4_zh.json_loss": 1.0053786039352417, + "eval_alpaca_gpt4_zh.json_runtime": 0.0504, + "eval_alpaca_gpt4_zh.json_samples_per_second": 218.451, + "eval_alpaca_gpt4_zh.json_steps_per_second": 19.859, + "step": 200 + }, + { + "epoch": 0.05503766640294452, + "eval_codefeedback_filtered_instruction.json_loss": 0.5965134501457214, + "eval_codefeedback_filtered_instruction.json_runtime": 0.4841, + "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.316, + "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.066, + "step": 200 + }, + { + "epoch": 0.05558804306697396, + "grad_norm": 0.38687607645988464, + "learning_rate": 1e-05, + "loss": 0.49, + "step": 202 + }, + { + "epoch": 0.056138419731003406, + "grad_norm": 0.39803430438041687, + "learning_rate": 1e-05, + "loss": 0.5047, + "step": 204 + }, + { + "epoch": 0.05668879639503285, + "grad_norm": 0.41770851612091064, + "learning_rate": 1e-05, + "loss": 0.4874, + "step": 206 + }, + { + "epoch": 0.057239173059062294, + "grad_norm": 0.3909968435764313, + "learning_rate": 1e-05, + "loss": 0.4992, + "step": 208 + }, + { + "epoch": 0.05778954972309174, + "grad_norm": 0.3818782866001129, + "learning_rate": 1e-05, + "loss": 0.5006, + "step": 210 + }, + { + "epoch": 0.05833992638712118, + "grad_norm": 0.4179542362689972, + "learning_rate": 1e-05, + "loss": 0.4945, + "step": 212 + }, + { + "epoch": 0.05889030305115063, + "grad_norm": 0.3872973322868347, + "learning_rate": 1e-05, + "loss": 0.4918, + "step": 214 + }, + { + "epoch": 0.05944067971518008, + "grad_norm": 0.4249219298362732, + "learning_rate": 1e-05, + "loss": 0.5039, + "step": 216 + }, + { + "epoch": 0.05999105637920952, + "grad_norm": 0.43381986021995544, + "learning_rate": 1e-05, + "loss": 0.4873, + "step": 218 + }, + { + "epoch": 0.06054143304323897, + "grad_norm": 0.40741005539894104, + "learning_rate": 1e-05, + "loss": 0.4771, + "step": 220 + }, + { + "epoch": 0.061091809707268414, + "grad_norm": 0.37800464034080505, + "learning_rate": 1e-05, + "loss": 0.5015, + "step": 222 + }, + { + "epoch": 0.061642186371297855, + "grad_norm": 0.42365899682044983, + "learning_rate": 1e-05, + "loss": 0.4906, + "step": 224 + }, + { + "epoch": 0.0621925630353273, + "grad_norm": 0.39279666543006897, + "learning_rate": 1e-05, + "loss": 0.51, + "step": 226 + }, + { + "epoch": 0.06274293969935675, + "grad_norm": 0.4037010073661804, + "learning_rate": 1e-05, + "loss": 0.5162, + "step": 228 + }, + { + "epoch": 0.0632933163633862, + "grad_norm": 0.37650179862976074, + "learning_rate": 1e-05, + "loss": 0.4984, + "step": 230 + }, + { + "epoch": 0.06384369302741563, + "grad_norm": 0.42879757285118103, + "learning_rate": 1e-05, + "loss": 0.492, + "step": 232 + }, + { + "epoch": 0.06439406969144508, + "grad_norm": 0.42225000262260437, + "learning_rate": 1e-05, + "loss": 0.5215, + "step": 234 + }, + { + "epoch": 0.06494444635547453, + "grad_norm": 0.3948579430580139, + "learning_rate": 1e-05, + "loss": 0.5045, + "step": 236 + }, + { + "epoch": 0.06549482301950398, + "grad_norm": 0.40142592787742615, + "learning_rate": 1e-05, + "loss": 0.5083, + "step": 238 + }, + { + "epoch": 0.06604519968353342, + "grad_norm": 0.41938111186027527, + "learning_rate": 1e-05, + "loss": 0.5094, + "step": 240 + }, + { + "epoch": 0.06659557634756286, + "grad_norm": 0.4345923066139221, + "learning_rate": 1e-05, + "loss": 0.5076, + "step": 242 + }, + { + "epoch": 0.0671459530115923, + "grad_norm": 0.3985568881034851, + "learning_rate": 1e-05, + "loss": 0.5007, + "step": 244 + }, + { + "epoch": 0.06769632967562175, + "grad_norm": 0.37891215085983276, + "learning_rate": 1e-05, + "loss": 0.513, + "step": 246 + }, + { + "epoch": 0.0682467063396512, + "grad_norm": 0.413566917181015, + "learning_rate": 1e-05, + "loss": 0.493, + "step": 248 + }, + { + "epoch": 0.06879708300368065, + "grad_norm": 0.3980996608734131, + "learning_rate": 1e-05, + "loss": 0.5161, + "step": 250 + }, + { + "epoch": 0.0693474596677101, + "grad_norm": 0.4525178372859955, + "learning_rate": 1e-05, + "loss": 0.5077, + "step": 252 + }, + { + "epoch": 0.06989783633173953, + "grad_norm": 0.3720250427722931, + "learning_rate": 1e-05, + "loss": 0.4809, + "step": 254 + }, + { + "epoch": 0.07044821299576898, + "grad_norm": 0.37366852164268494, + "learning_rate": 1e-05, + "loss": 0.4724, + "step": 256 + }, + { + "epoch": 0.07099858965979843, + "grad_norm": 0.38189247250556946, + "learning_rate": 1e-05, + "loss": 0.5062, + "step": 258 + }, + { + "epoch": 0.07154896632382787, + "grad_norm": 0.39108410477638245, + "learning_rate": 1e-05, + "loss": 0.4894, + "step": 260 + }, + { + "epoch": 0.07209934298785732, + "grad_norm": 0.4071044921875, + "learning_rate": 1e-05, + "loss": 0.4916, + "step": 262 + }, + { + "epoch": 0.07264971965188675, + "grad_norm": 0.38570597767829895, + "learning_rate": 1e-05, + "loss": 0.4925, + "step": 264 + }, + { + "epoch": 0.0732000963159162, + "grad_norm": 0.409600168466568, + "learning_rate": 1e-05, + "loss": 0.4987, + "step": 266 + }, + { + "epoch": 0.07375047297994565, + "grad_norm": 0.3844049274921417, + "learning_rate": 1e-05, + "loss": 0.5011, + "step": 268 + }, + { + "epoch": 0.0743008496439751, + "grad_norm": 0.41260388493537903, + "learning_rate": 1e-05, + "loss": 0.5014, + "step": 270 + }, + { + "epoch": 0.07485122630800455, + "grad_norm": 0.402567982673645, + "learning_rate": 1e-05, + "loss": 0.4926, + "step": 272 + }, + { + "epoch": 0.07540160297203398, + "grad_norm": 0.4058002233505249, + "learning_rate": 1e-05, + "loss": 0.4879, + "step": 274 + }, + { + "epoch": 0.07595197963606343, + "grad_norm": 0.42676812410354614, + "learning_rate": 1e-05, + "loss": 0.5073, + "step": 276 + }, + { + "epoch": 0.07650235630009287, + "grad_norm": 0.3878956735134125, + "learning_rate": 1e-05, + "loss": 0.4831, + "step": 278 + }, + { + "epoch": 0.07705273296412232, + "grad_norm": 0.37560945749282837, + "learning_rate": 1e-05, + "loss": 0.4705, + "step": 280 + }, + { + "epoch": 0.07760310962815177, + "grad_norm": 0.4071865379810333, + "learning_rate": 1e-05, + "loss": 0.489, + "step": 282 + }, + { + "epoch": 0.07815348629218122, + "grad_norm": 0.3832094073295593, + "learning_rate": 1e-05, + "loss": 0.4843, + "step": 284 + }, + { + "epoch": 0.07870386295621065, + "grad_norm": 0.3808830976486206, + "learning_rate": 1e-05, + "loss": 0.5019, + "step": 286 + }, + { + "epoch": 0.0792542396202401, + "grad_norm": 0.40182846784591675, + "learning_rate": 1e-05, + "loss": 0.4921, + "step": 288 + }, + { + "epoch": 0.07980461628426955, + "grad_norm": 0.4483119249343872, + "learning_rate": 1e-05, + "loss": 0.5042, + "step": 290 + }, + { + "epoch": 0.080354992948299, + "grad_norm": 0.3664950132369995, + "learning_rate": 1e-05, + "loss": 0.4758, + "step": 292 + }, + { + "epoch": 0.08090536961232844, + "grad_norm": 0.39573603868484497, + "learning_rate": 1e-05, + "loss": 0.4945, + "step": 294 + }, + { + "epoch": 0.08145574627635788, + "grad_norm": 0.44645532965660095, + "learning_rate": 1e-05, + "loss": 0.4964, + "step": 296 + }, + { + "epoch": 0.08200612294038732, + "grad_norm": 0.39092323184013367, + "learning_rate": 1e-05, + "loss": 0.4947, + "step": 298 + }, + { + "epoch": 0.08255649960441677, + "grad_norm": 0.41762229800224304, + "learning_rate": 1e-05, + "loss": 0.4949, + "step": 300 + }, + { + "epoch": 0.08310687626844622, + "grad_norm": 0.39803358912467957, + "learning_rate": 1e-05, + "loss": 0.4822, + "step": 302 + }, + { + "epoch": 0.08365725293247567, + "grad_norm": 0.39895498752593994, + "learning_rate": 1e-05, + "loss": 0.4893, + "step": 304 + }, + { + "epoch": 0.0842076295965051, + "grad_norm": 0.3883228600025177, + "learning_rate": 1e-05, + "loss": 0.5062, + "step": 306 + }, + { + "epoch": 0.08475800626053455, + "grad_norm": 0.4112294018268585, + "learning_rate": 1e-05, + "loss": 0.4979, + "step": 308 + }, + { + "epoch": 0.085308382924564, + "grad_norm": 0.3851683437824249, + "learning_rate": 1e-05, + "loss": 0.4934, + "step": 310 + }, + { + "epoch": 0.08585875958859344, + "grad_norm": 0.39728567004203796, + "learning_rate": 1e-05, + "loss": 0.4746, + "step": 312 + }, + { + "epoch": 0.08640913625262289, + "grad_norm": 0.3943733274936676, + "learning_rate": 1e-05, + "loss": 0.4904, + "step": 314 + }, + { + "epoch": 0.08695951291665234, + "grad_norm": 0.3954530656337738, + "learning_rate": 1e-05, + "loss": 0.4796, + "step": 316 + }, + { + "epoch": 0.08750988958068177, + "grad_norm": 0.41237205266952515, + "learning_rate": 1e-05, + "loss": 0.4908, + "step": 318 + }, + { + "epoch": 0.08806026624471122, + "grad_norm": 0.3923771381378174, + "learning_rate": 1e-05, + "loss": 0.4988, + "step": 320 + }, + { + "epoch": 0.08861064290874067, + "grad_norm": 0.38542094826698303, + "learning_rate": 1e-05, + "loss": 0.5027, + "step": 322 + }, + { + "epoch": 0.08916101957277012, + "grad_norm": 0.41598251461982727, + "learning_rate": 1e-05, + "loss": 0.4976, + "step": 324 + }, + { + "epoch": 0.08971139623679956, + "grad_norm": 0.40826794505119324, + "learning_rate": 1e-05, + "loss": 0.4929, + "step": 326 + }, + { + "epoch": 0.090261772900829, + "grad_norm": 0.39970022439956665, + "learning_rate": 1e-05, + "loss": 0.4946, + "step": 328 + }, + { + "epoch": 0.09081214956485845, + "grad_norm": 0.3739086985588074, + "learning_rate": 1e-05, + "loss": 0.4678, + "step": 330 + }, + { + "epoch": 0.0913625262288879, + "grad_norm": 0.3746420741081238, + "learning_rate": 1e-05, + "loss": 0.4757, + "step": 332 + }, + { + "epoch": 0.09191290289291734, + "grad_norm": 0.3976924419403076, + "learning_rate": 1e-05, + "loss": 0.487, + "step": 334 + }, + { + "epoch": 0.09246327955694679, + "grad_norm": 0.398971289396286, + "learning_rate": 1e-05, + "loss": 0.5077, + "step": 336 + }, + { + "epoch": 0.09301365622097624, + "grad_norm": 0.3937431871891022, + "learning_rate": 1e-05, + "loss": 0.4885, + "step": 338 + }, + { + "epoch": 0.09356403288500567, + "grad_norm": 0.395084410905838, + "learning_rate": 1e-05, + "loss": 0.4871, + "step": 340 + }, + { + "epoch": 0.09411440954903512, + "grad_norm": 0.3677273690700531, + "learning_rate": 1e-05, + "loss": 0.4813, + "step": 342 + }, + { + "epoch": 0.09466478621306457, + "grad_norm": 0.39645129442214966, + "learning_rate": 1e-05, + "loss": 0.4842, + "step": 344 + }, + { + "epoch": 0.09521516287709401, + "grad_norm": 0.3642916679382324, + "learning_rate": 1e-05, + "loss": 0.504, + "step": 346 + }, + { + "epoch": 0.09576553954112346, + "grad_norm": 0.40385907888412476, + "learning_rate": 1e-05, + "loss": 0.4933, + "step": 348 + }, + { + "epoch": 0.0963159162051529, + "grad_norm": 0.39063799381256104, + "learning_rate": 1e-05, + "loss": 0.4856, + "step": 350 + }, + { + "epoch": 0.09686629286918234, + "grad_norm": 0.38000059127807617, + "learning_rate": 1e-05, + "loss": 0.5001, + "step": 352 + }, + { + "epoch": 0.09741666953321179, + "grad_norm": 0.39380577206611633, + "learning_rate": 1e-05, + "loss": 0.4961, + "step": 354 + }, + { + "epoch": 0.09796704619724124, + "grad_norm": 0.39326363801956177, + "learning_rate": 1e-05, + "loss": 0.498, + "step": 356 + }, + { + "epoch": 0.09851742286127069, + "grad_norm": 0.3775707185268402, + "learning_rate": 1e-05, + "loss": 0.4792, + "step": 358 + }, + { + "epoch": 0.09906779952530012, + "grad_norm": 0.3770863115787506, + "learning_rate": 1e-05, + "loss": 0.4837, + "step": 360 + }, + { + "epoch": 0.09961817618932957, + "grad_norm": 0.41484272480010986, + "learning_rate": 1e-05, + "loss": 0.4739, + "step": 362 + }, + { + "epoch": 0.10016855285335902, + "grad_norm": 0.39758750796318054, + "learning_rate": 1e-05, + "loss": 0.4957, + "step": 364 + }, + { + "epoch": 0.10071892951738846, + "grad_norm": 0.43485164642333984, + "learning_rate": 1e-05, + "loss": 0.492, + "step": 366 + }, + { + "epoch": 0.10126930618141791, + "grad_norm": 0.40296798944473267, + "learning_rate": 1e-05, + "loss": 0.4977, + "step": 368 + }, + { + "epoch": 0.10181968284544736, + "grad_norm": 0.3818409740924835, + "learning_rate": 1e-05, + "loss": 0.481, + "step": 370 + }, + { + "epoch": 0.1023700595094768, + "grad_norm": 0.3949006199836731, + "learning_rate": 1e-05, + "loss": 0.5021, + "step": 372 + }, + { + "epoch": 0.10292043617350624, + "grad_norm": 0.4327391982078552, + "learning_rate": 1e-05, + "loss": 0.5036, + "step": 374 + }, + { + "epoch": 0.10347081283753569, + "grad_norm": 0.4008086025714874, + "learning_rate": 1e-05, + "loss": 0.4854, + "step": 376 + }, + { + "epoch": 0.10402118950156514, + "grad_norm": 0.4146427810192108, + "learning_rate": 1e-05, + "loss": 0.4933, + "step": 378 + }, + { + "epoch": 0.10457156616559458, + "grad_norm": 0.4073733389377594, + "learning_rate": 1e-05, + "loss": 0.4923, + "step": 380 + }, + { + "epoch": 0.10512194282962402, + "grad_norm": 0.40570083260536194, + "learning_rate": 1e-05, + "loss": 0.4806, + "step": 382 + }, + { + "epoch": 0.10567231949365347, + "grad_norm": 0.39516401290893555, + "learning_rate": 1e-05, + "loss": 0.5038, + "step": 384 + }, + { + "epoch": 0.10622269615768291, + "grad_norm": 0.3886268138885498, + "learning_rate": 1e-05, + "loss": 0.4737, + "step": 386 + }, + { + "epoch": 0.10677307282171236, + "grad_norm": 0.3846561014652252, + "learning_rate": 1e-05, + "loss": 0.4852, + "step": 388 + }, + { + "epoch": 0.10732344948574181, + "grad_norm": 0.3952987492084503, + "learning_rate": 1e-05, + "loss": 0.496, + "step": 390 + }, + { + "epoch": 0.10787382614977124, + "grad_norm": 0.3840448558330536, + "learning_rate": 1e-05, + "loss": 0.4976, + "step": 392 + }, + { + "epoch": 0.10842420281380069, + "grad_norm": 0.38074344396591187, + "learning_rate": 1e-05, + "loss": 0.508, + "step": 394 + }, + { + "epoch": 0.10897457947783014, + "grad_norm": 0.4216584861278534, + "learning_rate": 1e-05, + "loss": 0.4841, + "step": 396 + }, + { + "epoch": 0.10952495614185959, + "grad_norm": 0.39932167530059814, + "learning_rate": 1e-05, + "loss": 0.4783, + "step": 398 + }, + { + "epoch": 0.11007533280588903, + "grad_norm": 0.3687106966972351, + "learning_rate": 1e-05, + "loss": 0.4747, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_merge_loss": 0.42975950241088867, + "eval_merge_runtime": 600.4283, + "eval_merge_samples_per_second": 56.213, + "eval_merge_steps_per_second": 2.343, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_new_aug_datas_filtered.json_loss": 0.558424174785614, + "eval_new_aug_datas_filtered.json_runtime": 10.4015, + "eval_new_aug_datas_filtered.json_samples_per_second": 73.74, + "eval_new_aug_datas_filtered.json_steps_per_second": 3.076, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_sharegpt_gpt4.json_loss": 0.8120941519737244, + "eval_sharegpt_gpt4.json_runtime": 31.6378, + "eval_sharegpt_gpt4.json_samples_per_second": 58.822, + "eval_sharegpt_gpt4.json_steps_per_second": 2.465, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_Table_GPT.json_loss": 0.08201506733894348, + "eval_Table_GPT.json_runtime": 24.9859, + "eval_Table_GPT.json_samples_per_second": 83.767, + "eval_Table_GPT.json_steps_per_second": 3.522, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_gpt_4o_200k.json_loss": 0.8391836881637573, + "eval_gpt_4o_200k.json_runtime": 48.456, + "eval_gpt_4o_200k.json_samples_per_second": 129.623, + "eval_gpt_4o_200k.json_steps_per_second": 5.407, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_multi_turn_datas.json_loss": 0.37471804022789, + "eval_multi_turn_datas.json_runtime": 75.4526, + "eval_multi_turn_datas.json_samples_per_second": 53.04, + "eval_multi_turn_datas.json_steps_per_second": 2.213, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_table_python_code_datas.json_loss": 0.3102189898490906, + "eval_table_python_code_datas.json_runtime": 42.9961, + "eval_table_python_code_datas.json_samples_per_second": 50.214, + "eval_table_python_code_datas.json_steps_per_second": 2.093, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_tabular_llm_data.json_loss": 0.16565443575382233, + "eval_tabular_llm_data.json_runtime": 8.5095, + "eval_tabular_llm_data.json_samples_per_second": 28.909, + "eval_tabular_llm_data.json_steps_per_second": 1.293, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_python_code_critic_21k.json_loss": 0.6095640063285828, + "eval_python_code_critic_21k.json_runtime": 3.2106, + "eval_python_code_critic_21k.json_samples_per_second": 185.945, + "eval_python_code_critic_21k.json_steps_per_second": 7.787, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_all_merge_table_dataset.json_loss": 0.09411227703094482, + "eval_all_merge_table_dataset.json_runtime": 23.2459, + "eval_all_merge_table_dataset.json_samples_per_second": 30.629, + "eval_all_merge_table_dataset.json_steps_per_second": 1.291, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_code_feedback_multi_turn.json_loss": 0.6033111810684204, + "eval_code_feedback_multi_turn.json_runtime": 32.3176, + "eval_code_feedback_multi_turn.json_samples_per_second": 68.105, + "eval_code_feedback_multi_turn.json_steps_per_second": 2.847, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_ultrainteract_sft.json_loss": 0.4417967200279236, + "eval_ultrainteract_sft.json_runtime": 8.6225, + "eval_ultrainteract_sft.json_samples_per_second": 168.86, + "eval_ultrainteract_sft.json_steps_per_second": 7.075, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_synthetic_text_to_sql.json_loss": 0.10689640045166016, + "eval_synthetic_text_to_sql.json_runtime": 0.1258, + "eval_synthetic_text_to_sql.json_samples_per_second": 270.238, + "eval_synthetic_text_to_sql.json_steps_per_second": 15.896, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_sft_react_sql_datas.json_loss": 0.6745051145553589, + "eval_sft_react_sql_datas.json_runtime": 7.8354, + "eval_sft_react_sql_datas.json_samples_per_second": 40.074, + "eval_sft_react_sql_datas.json_steps_per_second": 1.787, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_all_merge_code.json_loss": 0.3035649061203003, + "eval_all_merge_code.json_runtime": 0.3282, + "eval_all_merge_code.json_samples_per_second": 191.936, + "eval_all_merge_code.json_steps_per_second": 9.14, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_magpie_datas.json_loss": 0.4511661231517792, + "eval_magpie_datas.json_runtime": 2.2095, + "eval_magpie_datas.json_samples_per_second": 77.847, + "eval_magpie_datas.json_steps_per_second": 3.621, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_train_data_for_qwen.json_loss": 0.012529651634395123, + "eval_train_data_for_qwen.json_runtime": 0.2431, + "eval_train_data_for_qwen.json_samples_per_second": 41.135, + "eval_train_data_for_qwen.json_steps_per_second": 4.113, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_alpaca_cleaned.json_loss": 0.9377387166023254, + "eval_alpaca_cleaned.json_runtime": 0.115, + "eval_alpaca_cleaned.json_samples_per_second": 234.777, + "eval_alpaca_cleaned.json_steps_per_second": 17.391, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_agent_instruct.json_loss": 0.2363067865371704, + "eval_agent_instruct.json_runtime": 0.5121, + "eval_agent_instruct.json_samples_per_second": 93.728, + "eval_agent_instruct.json_steps_per_second": 3.905, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_MathInstruct.json_loss": 0.21367128193378448, + "eval_MathInstruct.json_runtime": 0.3647, + "eval_MathInstruct.json_samples_per_second": 156.291, + "eval_MathInstruct.json_steps_per_second": 8.226, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_tested_143k_python_alpaca.json_loss": 0.4474259316921234, + "eval_tested_143k_python_alpaca.json_runtime": 0.3019, + "eval_tested_143k_python_alpaca.json_samples_per_second": 112.631, + "eval_tested_143k_python_alpaca.json_steps_per_second": 6.625, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_xlam_function_calling_60k.json_loss": 0.011296543292701244, + "eval_xlam_function_calling_60k.json_runtime": 0.1, + "eval_xlam_function_calling_60k.json_samples_per_second": 230.102, + "eval_xlam_function_calling_60k.json_steps_per_second": 10.004, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_alpaca_data_gpt4_chinese.json_loss": 1.645748496055603, + "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0515, + "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 310.905, + "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.432, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_alpaca_gpt4_zh.json_loss": 0.9886136651039124, + "eval_alpaca_gpt4_zh.json_runtime": 0.0503, + "eval_alpaca_gpt4_zh.json_samples_per_second": 218.827, + "eval_alpaca_gpt4_zh.json_steps_per_second": 19.893, + "step": 400 + }, + { + "epoch": 0.11007533280588903, + "eval_codefeedback_filtered_instruction.json_loss": 0.5969922542572021, + "eval_codefeedback_filtered_instruction.json_runtime": 0.4851, + "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.226, + "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.061, + "step": 400 + }, + { + "epoch": 0.11062570946991848, + "grad_norm": 0.40487441420555115, + "learning_rate": 1e-05, + "loss": 0.4811, + "step": 402 + }, + { + "epoch": 0.11117608613394792, + "grad_norm": 0.39143064618110657, + "learning_rate": 1e-05, + "loss": 0.4704, + "step": 404 + }, + { + "epoch": 0.11172646279797736, + "grad_norm": 0.46816787123680115, + "learning_rate": 1e-05, + "loss": 0.4941, + "step": 406 + }, + { + "epoch": 0.11227683946200681, + "grad_norm": 0.37707188725471497, + "learning_rate": 1e-05, + "loss": 0.4839, + "step": 408 + }, + { + "epoch": 0.11282721612603626, + "grad_norm": 0.3780951201915741, + "learning_rate": 1e-05, + "loss": 0.4889, + "step": 410 + }, + { + "epoch": 0.1133775927900657, + "grad_norm": 0.36941519379615784, + "learning_rate": 1e-05, + "loss": 0.4747, + "step": 412 + }, + { + "epoch": 0.11392796945409514, + "grad_norm": 0.39626002311706543, + "learning_rate": 1e-05, + "loss": 0.4872, + "step": 414 + }, + { + "epoch": 0.11447834611812459, + "grad_norm": 0.38315075635910034, + "learning_rate": 1e-05, + "loss": 0.471, + "step": 416 + }, + { + "epoch": 0.11502872278215404, + "grad_norm": 0.37200862169265747, + "learning_rate": 1e-05, + "loss": 0.4891, + "step": 418 + }, + { + "epoch": 0.11557909944618348, + "grad_norm": 0.39199399948120117, + "learning_rate": 1e-05, + "loss": 0.4807, + "step": 420 + }, + { + "epoch": 0.11612947611021293, + "grad_norm": 0.37726107239723206, + "learning_rate": 1e-05, + "loss": 0.4834, + "step": 422 + }, + { + "epoch": 0.11667985277424237, + "grad_norm": 0.38188016414642334, + "learning_rate": 1e-05, + "loss": 0.4853, + "step": 424 + }, + { + "epoch": 0.11723022943827181, + "grad_norm": 0.39772850275039673, + "learning_rate": 1e-05, + "loss": 0.4895, + "step": 426 + }, + { + "epoch": 0.11778060610230126, + "grad_norm": 0.3797503411769867, + "learning_rate": 1e-05, + "loss": 0.4818, + "step": 428 + }, + { + "epoch": 0.11833098276633071, + "grad_norm": 0.39962416887283325, + "learning_rate": 1e-05, + "loss": 0.4802, + "step": 430 + }, + { + "epoch": 0.11888135943036016, + "grad_norm": 0.37405237555503845, + "learning_rate": 1e-05, + "loss": 0.4879, + "step": 432 + }, + { + "epoch": 0.1194317360943896, + "grad_norm": 0.39297720789909363, + "learning_rate": 1e-05, + "loss": 0.4853, + "step": 434 + }, + { + "epoch": 0.11998211275841904, + "grad_norm": 0.3871022164821625, + "learning_rate": 1e-05, + "loss": 0.4845, + "step": 436 + }, + { + "epoch": 0.12053248942244849, + "grad_norm": 0.43845734000205994, + "learning_rate": 1e-05, + "loss": 0.4865, + "step": 438 + }, + { + "epoch": 0.12108286608647793, + "grad_norm": 0.3888757526874542, + "learning_rate": 1e-05, + "loss": 0.4862, + "step": 440 + }, + { + "epoch": 0.12163324275050738, + "grad_norm": 0.3801029920578003, + "learning_rate": 1e-05, + "loss": 0.4751, + "step": 442 + }, + { + "epoch": 0.12218361941453683, + "grad_norm": 0.3861992657184601, + "learning_rate": 1e-05, + "loss": 0.5026, + "step": 444 + }, + { + "epoch": 0.12273399607856626, + "grad_norm": 0.40307343006134033, + "learning_rate": 1e-05, + "loss": 0.4901, + "step": 446 + }, + { + "epoch": 0.12328437274259571, + "grad_norm": 0.36803606152534485, + "learning_rate": 1e-05, + "loss": 0.4927, + "step": 448 + }, + { + "epoch": 0.12383474940662516, + "grad_norm": 0.40266790986061096, + "learning_rate": 1e-05, + "loss": 0.4663, + "step": 450 + }, + { + "epoch": 0.1243851260706546, + "grad_norm": 0.3870522975921631, + "learning_rate": 1e-05, + "loss": 0.475, + "step": 452 + }, + { + "epoch": 0.12493550273468405, + "grad_norm": 0.3978688716888428, + "learning_rate": 1e-05, + "loss": 0.4979, + "step": 454 + }, + { + "epoch": 0.1254858793987135, + "grad_norm": 0.3799881935119629, + "learning_rate": 1e-05, + "loss": 0.4802, + "step": 456 + }, + { + "epoch": 0.12603625606274294, + "grad_norm": 0.3795452415943146, + "learning_rate": 1e-05, + "loss": 0.4878, + "step": 458 + }, + { + "epoch": 0.1265866327267724, + "grad_norm": 0.3865358233451843, + "learning_rate": 1e-05, + "loss": 0.4825, + "step": 460 + }, + { + "epoch": 0.12713700939080183, + "grad_norm": 0.3646644353866577, + "learning_rate": 1e-05, + "loss": 0.4725, + "step": 462 + }, + { + "epoch": 0.12768738605483126, + "grad_norm": 0.3851023018360138, + "learning_rate": 1e-05, + "loss": 0.4849, + "step": 464 + }, + { + "epoch": 0.12823776271886073, + "grad_norm": 0.37587490677833557, + "learning_rate": 1e-05, + "loss": 0.4729, + "step": 466 + }, + { + "epoch": 0.12878813938289016, + "grad_norm": 0.3559257686138153, + "learning_rate": 1e-05, + "loss": 0.4826, + "step": 468 + }, + { + "epoch": 0.12933851604691962, + "grad_norm": 0.3967975974082947, + "learning_rate": 1e-05, + "loss": 0.4917, + "step": 470 + }, + { + "epoch": 0.12988889271094906, + "grad_norm": 0.4064919650554657, + "learning_rate": 1e-05, + "loss": 0.5018, + "step": 472 + }, + { + "epoch": 0.1304392693749785, + "grad_norm": 0.3609434962272644, + "learning_rate": 1e-05, + "loss": 0.4805, + "step": 474 + }, + { + "epoch": 0.13098964603900795, + "grad_norm": 0.4229820668697357, + "learning_rate": 1e-05, + "loss": 0.4756, + "step": 476 + }, + { + "epoch": 0.13154002270303738, + "grad_norm": 0.3882080018520355, + "learning_rate": 1e-05, + "loss": 0.4946, + "step": 478 + }, + { + "epoch": 0.13209039936706685, + "grad_norm": 0.37811529636383057, + "learning_rate": 1e-05, + "loss": 0.495, + "step": 480 + }, + { + "epoch": 0.13264077603109628, + "grad_norm": 0.4139231741428375, + "learning_rate": 1e-05, + "loss": 0.4722, + "step": 482 + }, + { + "epoch": 0.13319115269512571, + "grad_norm": 0.3836536705493927, + "learning_rate": 1e-05, + "loss": 0.4795, + "step": 484 + }, + { + "epoch": 0.13374152935915518, + "grad_norm": 0.39434006810188293, + "learning_rate": 1e-05, + "loss": 0.4783, + "step": 486 + }, + { + "epoch": 0.1342919060231846, + "grad_norm": 0.3847144544124603, + "learning_rate": 1e-05, + "loss": 0.4751, + "step": 488 + }, + { + "epoch": 0.13484228268721407, + "grad_norm": 0.4081107974052429, + "learning_rate": 1e-05, + "loss": 0.4947, + "step": 490 + }, + { + "epoch": 0.1353926593512435, + "grad_norm": 0.3780671954154968, + "learning_rate": 1e-05, + "loss": 0.4932, + "step": 492 + }, + { + "epoch": 0.13594303601527294, + "grad_norm": 0.39522022008895874, + "learning_rate": 1e-05, + "loss": 0.4868, + "step": 494 + }, + { + "epoch": 0.1364934126793024, + "grad_norm": 0.3978594243526459, + "learning_rate": 1e-05, + "loss": 0.4895, + "step": 496 + }, + { + "epoch": 0.13704378934333183, + "grad_norm": 0.40067028999328613, + "learning_rate": 1e-05, + "loss": 0.4841, + "step": 498 + }, + { + "epoch": 0.1375941660073613, + "grad_norm": 0.38525891304016113, + "learning_rate": 1e-05, + "loss": 0.4769, + "step": 500 + }, + { + "epoch": 0.13814454267139073, + "grad_norm": 0.3708615303039551, + "learning_rate": 1e-05, + "loss": 0.4787, + "step": 502 + }, + { + "epoch": 0.1386949193354202, + "grad_norm": 0.3583269417285919, + "learning_rate": 1e-05, + "loss": 0.4905, + "step": 504 + }, + { + "epoch": 0.13924529599944963, + "grad_norm": 0.4004143178462982, + "learning_rate": 1e-05, + "loss": 0.4797, + "step": 506 + }, + { + "epoch": 0.13979567266347906, + "grad_norm": 0.3877711594104767, + "learning_rate": 1e-05, + "loss": 0.4968, + "step": 508 + }, + { + "epoch": 0.14034604932750852, + "grad_norm": 0.394502729177475, + "learning_rate": 1e-05, + "loss": 0.4743, + "step": 510 + }, + { + "epoch": 0.14089642599153795, + "grad_norm": 0.3829086720943451, + "learning_rate": 1e-05, + "loss": 0.4769, + "step": 512 + }, + { + "epoch": 0.14144680265556742, + "grad_norm": 0.3849917948246002, + "learning_rate": 1e-05, + "loss": 0.4763, + "step": 514 + }, + { + "epoch": 0.14199717931959685, + "grad_norm": 0.40810078382492065, + "learning_rate": 1e-05, + "loss": 0.4904, + "step": 516 + }, + { + "epoch": 0.14254755598362628, + "grad_norm": 0.3982490599155426, + "learning_rate": 1e-05, + "loss": 0.4762, + "step": 518 + }, + { + "epoch": 0.14309793264765575, + "grad_norm": 0.36841145157814026, + "learning_rate": 1e-05, + "loss": 0.4745, + "step": 520 + }, + { + "epoch": 0.14364830931168518, + "grad_norm": 0.3805830180644989, + "learning_rate": 1e-05, + "loss": 0.4811, + "step": 522 + }, + { + "epoch": 0.14419868597571464, + "grad_norm": 0.40074169635772705, + "learning_rate": 1e-05, + "loss": 0.4923, + "step": 524 + }, + { + "epoch": 0.14474906263974407, + "grad_norm": 0.42140403389930725, + "learning_rate": 1e-05, + "loss": 0.4972, + "step": 526 + }, + { + "epoch": 0.1452994393037735, + "grad_norm": 0.38489535450935364, + "learning_rate": 1e-05, + "loss": 0.4921, + "step": 528 + }, + { + "epoch": 0.14584981596780297, + "grad_norm": 0.38449668884277344, + "learning_rate": 1e-05, + "loss": 0.4883, + "step": 530 + }, + { + "epoch": 0.1464001926318324, + "grad_norm": 0.38009950518608093, + "learning_rate": 1e-05, + "loss": 0.4808, + "step": 532 + }, + { + "epoch": 0.14695056929586187, + "grad_norm": 0.3916541337966919, + "learning_rate": 1e-05, + "loss": 0.4708, + "step": 534 + }, + { + "epoch": 0.1475009459598913, + "grad_norm": 0.39856135845184326, + "learning_rate": 1e-05, + "loss": 0.4933, + "step": 536 + }, + { + "epoch": 0.14805132262392073, + "grad_norm": 0.3804597556591034, + "learning_rate": 1e-05, + "loss": 0.4772, + "step": 538 + }, + { + "epoch": 0.1486016992879502, + "grad_norm": 0.39584964513778687, + "learning_rate": 1e-05, + "loss": 0.4746, + "step": 540 + }, + { + "epoch": 0.14915207595197963, + "grad_norm": 0.36922863125801086, + "learning_rate": 1e-05, + "loss": 0.4911, + "step": 542 + }, + { + "epoch": 0.1497024526160091, + "grad_norm": 0.38762298226356506, + "learning_rate": 1e-05, + "loss": 0.4744, + "step": 544 + }, + { + "epoch": 0.15025282928003852, + "grad_norm": 0.38803887367248535, + "learning_rate": 1e-05, + "loss": 0.4776, + "step": 546 + }, + { + "epoch": 0.15080320594406796, + "grad_norm": 0.39409226179122925, + "learning_rate": 1e-05, + "loss": 0.4789, + "step": 548 + }, + { + "epoch": 0.15135358260809742, + "grad_norm": 0.4141768217086792, + "learning_rate": 1e-05, + "loss": 0.4752, + "step": 550 + }, + { + "epoch": 0.15190395927212685, + "grad_norm": 0.3770216703414917, + "learning_rate": 1e-05, + "loss": 0.4689, + "step": 552 + }, + { + "epoch": 0.15245433593615632, + "grad_norm": 0.3929697573184967, + "learning_rate": 1e-05, + "loss": 0.4861, + "step": 554 + }, + { + "epoch": 0.15300471260018575, + "grad_norm": 0.3859105706214905, + "learning_rate": 1e-05, + "loss": 0.4799, + "step": 556 + }, + { + "epoch": 0.1535550892642152, + "grad_norm": 0.41044744849205017, + "learning_rate": 1e-05, + "loss": 0.4911, + "step": 558 + }, + { + "epoch": 0.15410546592824464, + "grad_norm": 0.36859771609306335, + "learning_rate": 1e-05, + "loss": 0.4653, + "step": 560 + }, + { + "epoch": 0.15465584259227408, + "grad_norm": 0.39258813858032227, + "learning_rate": 1e-05, + "loss": 0.4769, + "step": 562 + }, + { + "epoch": 0.15520621925630354, + "grad_norm": 0.38241100311279297, + "learning_rate": 1e-05, + "loss": 0.4821, + "step": 564 + }, + { + "epoch": 0.15575659592033297, + "grad_norm": 0.4107513427734375, + "learning_rate": 1e-05, + "loss": 0.4746, + "step": 566 + }, + { + "epoch": 0.15630697258436244, + "grad_norm": 0.3872488737106323, + "learning_rate": 1e-05, + "loss": 0.4817, + "step": 568 + }, + { + "epoch": 0.15685734924839187, + "grad_norm": 0.3712390065193176, + "learning_rate": 1e-05, + "loss": 0.4944, + "step": 570 + }, + { + "epoch": 0.1574077259124213, + "grad_norm": 0.413503497838974, + "learning_rate": 1e-05, + "loss": 0.4774, + "step": 572 + }, + { + "epoch": 0.15795810257645077, + "grad_norm": 0.35706543922424316, + "learning_rate": 1e-05, + "loss": 0.4743, + "step": 574 + }, + { + "epoch": 0.1585084792404802, + "grad_norm": 0.39815768599510193, + "learning_rate": 1e-05, + "loss": 0.4846, + "step": 576 + }, + { + "epoch": 0.15905885590450966, + "grad_norm": 0.38346678018569946, + "learning_rate": 1e-05, + "loss": 0.4633, + "step": 578 + }, + { + "epoch": 0.1596092325685391, + "grad_norm": 0.3905611038208008, + "learning_rate": 1e-05, + "loss": 0.4776, + "step": 580 + }, + { + "epoch": 0.16015960923256853, + "grad_norm": 0.3790382742881775, + "learning_rate": 1e-05, + "loss": 0.4892, + "step": 582 + }, + { + "epoch": 0.160709985896598, + "grad_norm": 0.37033775448799133, + "learning_rate": 1e-05, + "loss": 0.4848, + "step": 584 + }, + { + "epoch": 0.16126036256062742, + "grad_norm": 0.3686079680919647, + "learning_rate": 1e-05, + "loss": 0.4514, + "step": 586 + }, + { + "epoch": 0.16181073922465689, + "grad_norm": 0.3836509883403778, + "learning_rate": 1e-05, + "loss": 0.4859, + "step": 588 + }, + { + "epoch": 0.16236111588868632, + "grad_norm": 0.40387076139450073, + "learning_rate": 1e-05, + "loss": 0.485, + "step": 590 + }, + { + "epoch": 0.16291149255271575, + "grad_norm": 0.3850373327732086, + "learning_rate": 1e-05, + "loss": 0.4843, + "step": 592 + }, + { + "epoch": 0.16346186921674521, + "grad_norm": 0.3814505934715271, + "learning_rate": 1e-05, + "loss": 0.4749, + "step": 594 + }, + { + "epoch": 0.16401224588077465, + "grad_norm": 0.35501739382743835, + "learning_rate": 1e-05, + "loss": 0.4645, + "step": 596 + }, + { + "epoch": 0.1645626225448041, + "grad_norm": 0.34997090697288513, + "learning_rate": 1e-05, + "loss": 0.4687, + "step": 598 + }, + { + "epoch": 0.16511299920883354, + "grad_norm": 0.365212619304657, + "learning_rate": 1e-05, + "loss": 0.4956, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_merge_loss": 0.4216049909591675, + "eval_merge_runtime": 600.0688, + "eval_merge_samples_per_second": 56.247, + "eval_merge_steps_per_second": 2.345, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_new_aug_datas_filtered.json_loss": 0.5500591397285461, + "eval_new_aug_datas_filtered.json_runtime": 10.5295, + "eval_new_aug_datas_filtered.json_samples_per_second": 72.843, + "eval_new_aug_datas_filtered.json_steps_per_second": 3.039, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_sharegpt_gpt4.json_loss": 0.8046284914016724, + "eval_sharegpt_gpt4.json_runtime": 31.7366, + "eval_sharegpt_gpt4.json_samples_per_second": 58.639, + "eval_sharegpt_gpt4.json_steps_per_second": 2.458, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_Table_GPT.json_loss": 0.07981107383966446, + "eval_Table_GPT.json_runtime": 25.0085, + "eval_Table_GPT.json_samples_per_second": 83.691, + "eval_Table_GPT.json_steps_per_second": 3.519, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_gpt_4o_200k.json_loss": 0.8323716521263123, + "eval_gpt_4o_200k.json_runtime": 48.5988, + "eval_gpt_4o_200k.json_samples_per_second": 129.242, + "eval_gpt_4o_200k.json_steps_per_second": 5.391, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_multi_turn_datas.json_loss": 0.36492469906806946, + "eval_multi_turn_datas.json_runtime": 75.8696, + "eval_multi_turn_datas.json_samples_per_second": 52.748, + "eval_multi_turn_datas.json_steps_per_second": 2.201, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_table_python_code_datas.json_loss": 0.29984766244888306, + "eval_table_python_code_datas.json_runtime": 43.1945, + "eval_table_python_code_datas.json_samples_per_second": 49.983, + "eval_table_python_code_datas.json_steps_per_second": 2.084, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_tabular_llm_data.json_loss": 0.13250145316123962, + "eval_tabular_llm_data.json_runtime": 8.5476, + "eval_tabular_llm_data.json_samples_per_second": 28.78, + "eval_tabular_llm_data.json_steps_per_second": 1.287, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_python_code_critic_21k.json_loss": 0.599878191947937, + "eval_python_code_critic_21k.json_runtime": 3.2358, + "eval_python_code_critic_21k.json_samples_per_second": 184.496, + "eval_python_code_critic_21k.json_steps_per_second": 7.726, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_all_merge_table_dataset.json_loss": 0.08627181500196457, + "eval_all_merge_table_dataset.json_runtime": 23.3808, + "eval_all_merge_table_dataset.json_samples_per_second": 30.452, + "eval_all_merge_table_dataset.json_steps_per_second": 1.283, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_code_feedback_multi_turn.json_loss": 0.5982062220573425, + "eval_code_feedback_multi_turn.json_runtime": 32.4617, + "eval_code_feedback_multi_turn.json_samples_per_second": 67.803, + "eval_code_feedback_multi_turn.json_steps_per_second": 2.834, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_ultrainteract_sft.json_loss": 0.4367137849330902, + "eval_ultrainteract_sft.json_runtime": 8.672, + "eval_ultrainteract_sft.json_samples_per_second": 167.896, + "eval_ultrainteract_sft.json_steps_per_second": 7.034, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_synthetic_text_to_sql.json_loss": 0.1079096570611, + "eval_synthetic_text_to_sql.json_runtime": 0.1265, + "eval_synthetic_text_to_sql.json_samples_per_second": 268.769, + "eval_synthetic_text_to_sql.json_steps_per_second": 15.81, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_sft_react_sql_datas.json_loss": 0.6610473990440369, + "eval_sft_react_sql_datas.json_runtime": 7.8536, + "eval_sft_react_sql_datas.json_samples_per_second": 39.982, + "eval_sft_react_sql_datas.json_steps_per_second": 1.783, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_all_merge_code.json_loss": 0.3021065890789032, + "eval_all_merge_code.json_runtime": 0.3377, + "eval_all_merge_code.json_samples_per_second": 186.571, + "eval_all_merge_code.json_steps_per_second": 8.884, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_magpie_datas.json_loss": 0.4455747604370117, + "eval_magpie_datas.json_runtime": 2.2122, + "eval_magpie_datas.json_samples_per_second": 77.751, + "eval_magpie_datas.json_steps_per_second": 3.616, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_train_data_for_qwen.json_loss": 0.009937227703630924, + "eval_train_data_for_qwen.json_runtime": 0.2454, + "eval_train_data_for_qwen.json_samples_per_second": 40.745, + "eval_train_data_for_qwen.json_steps_per_second": 4.075, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_alpaca_cleaned.json_loss": 0.9349167943000793, + "eval_alpaca_cleaned.json_runtime": 0.1148, + "eval_alpaca_cleaned.json_samples_per_second": 235.251, + "eval_alpaca_cleaned.json_steps_per_second": 17.426, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_agent_instruct.json_loss": 0.2310038059949875, + "eval_agent_instruct.json_runtime": 0.5119, + "eval_agent_instruct.json_samples_per_second": 93.766, + "eval_agent_instruct.json_steps_per_second": 3.907, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_MathInstruct.json_loss": 0.21358835697174072, + "eval_MathInstruct.json_runtime": 0.3581, + "eval_MathInstruct.json_samples_per_second": 159.182, + "eval_MathInstruct.json_steps_per_second": 8.378, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_tested_143k_python_alpaca.json_loss": 0.4455429017543793, + "eval_tested_143k_python_alpaca.json_runtime": 0.3013, + "eval_tested_143k_python_alpaca.json_samples_per_second": 112.849, + "eval_tested_143k_python_alpaca.json_steps_per_second": 6.638, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_xlam_function_calling_60k.json_loss": 0.00893339328467846, + "eval_xlam_function_calling_60k.json_runtime": 0.1004, + "eval_xlam_function_calling_60k.json_samples_per_second": 228.974, + "eval_xlam_function_calling_60k.json_steps_per_second": 9.955, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_alpaca_data_gpt4_chinese.json_loss": 1.6295539140701294, + "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0512, + "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 312.613, + "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.538, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_alpaca_gpt4_zh.json_loss": 0.9761592745780945, + "eval_alpaca_gpt4_zh.json_runtime": 0.0499, + "eval_alpaca_gpt4_zh.json_samples_per_second": 220.289, + "eval_alpaca_gpt4_zh.json_steps_per_second": 20.026, + "step": 600 + }, + { + "epoch": 0.16511299920883354, + "eval_codefeedback_filtered_instruction.json_loss": 0.5956905484199524, + "eval_codefeedback_filtered_instruction.json_runtime": 0.4851, + "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.232, + "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.062, + "step": 600 + }, + { + "epoch": 0.16566337587286298, + "grad_norm": 0.38436150550842285, + "learning_rate": 1e-05, + "loss": 0.4609, + "step": 602 + }, + { + "epoch": 0.16621375253689244, + "grad_norm": 0.3946292996406555, + "learning_rate": 1e-05, + "loss": 0.4699, + "step": 604 + }, + { + "epoch": 0.16676412920092187, + "grad_norm": 0.4069615304470062, + "learning_rate": 1e-05, + "loss": 0.4722, + "step": 606 + }, + { + "epoch": 0.16731450586495134, + "grad_norm": 0.371660977602005, + "learning_rate": 1e-05, + "loss": 0.4856, + "step": 608 + }, + { + "epoch": 0.16786488252898077, + "grad_norm": 0.394911527633667, + "learning_rate": 1e-05, + "loss": 0.4804, + "step": 610 + }, + { + "epoch": 0.1684152591930102, + "grad_norm": 0.4873884916305542, + "learning_rate": 1e-05, + "loss": 0.4686, + "step": 612 + }, + { + "epoch": 0.16896563585703966, + "grad_norm": 0.3943842649459839, + "learning_rate": 1e-05, + "loss": 0.4887, + "step": 614 + }, + { + "epoch": 0.1695160125210691, + "grad_norm": 0.3716658055782318, + "learning_rate": 1e-05, + "loss": 0.4898, + "step": 616 + }, + { + "epoch": 0.17006638918509856, + "grad_norm": 0.36271047592163086, + "learning_rate": 1e-05, + "loss": 0.4861, + "step": 618 + }, + { + "epoch": 0.170616765849128, + "grad_norm": 0.3833015263080597, + "learning_rate": 1e-05, + "loss": 0.4814, + "step": 620 + }, + { + "epoch": 0.17116714251315746, + "grad_norm": 0.3661365211009979, + "learning_rate": 1e-05, + "loss": 0.4873, + "step": 622 + }, + { + "epoch": 0.1717175191771869, + "grad_norm": 0.3613869845867157, + "learning_rate": 1e-05, + "loss": 0.4537, + "step": 624 + }, + { + "epoch": 0.17226789584121632, + "grad_norm": 0.34498724341392517, + "learning_rate": 1e-05, + "loss": 0.483, + "step": 626 + }, + { + "epoch": 0.17281827250524578, + "grad_norm": 0.41466256976127625, + "learning_rate": 1e-05, + "loss": 0.4765, + "step": 628 + }, + { + "epoch": 0.17336864916927522, + "grad_norm": 0.36220455169677734, + "learning_rate": 1e-05, + "loss": 0.4842, + "step": 630 + }, + { + "epoch": 0.17391902583330468, + "grad_norm": 0.38009753823280334, + "learning_rate": 1e-05, + "loss": 0.482, + "step": 632 + }, + { + "epoch": 0.17446940249733411, + "grad_norm": 0.3589475452899933, + "learning_rate": 1e-05, + "loss": 0.4714, + "step": 634 + }, + { + "epoch": 0.17501977916136355, + "grad_norm": 0.37625178694725037, + "learning_rate": 1e-05, + "loss": 0.4487, + "step": 636 + }, + { + "epoch": 0.175570155825393, + "grad_norm": 0.3818652331829071, + "learning_rate": 1e-05, + "loss": 0.4757, + "step": 638 + }, + { + "epoch": 0.17612053248942244, + "grad_norm": 0.39498913288116455, + "learning_rate": 1e-05, + "loss": 0.4879, + "step": 640 + }, + { + "epoch": 0.1766709091534519, + "grad_norm": 0.3864663243293762, + "learning_rate": 1e-05, + "loss": 0.4815, + "step": 642 + }, + { + "epoch": 0.17722128581748134, + "grad_norm": 0.37452608346939087, + "learning_rate": 1e-05, + "loss": 0.4773, + "step": 644 + }, + { + "epoch": 0.17777166248151077, + "grad_norm": 0.3754761219024658, + "learning_rate": 1e-05, + "loss": 0.4916, + "step": 646 + }, + { + "epoch": 0.17832203914554023, + "grad_norm": 0.3797055780887604, + "learning_rate": 1e-05, + "loss": 0.4663, + "step": 648 + }, + { + "epoch": 0.17887241580956967, + "grad_norm": 0.3640367090702057, + "learning_rate": 1e-05, + "loss": 0.4737, + "step": 650 + }, + { + "epoch": 0.17942279247359913, + "grad_norm": 0.35961100459098816, + "learning_rate": 1e-05, + "loss": 0.4757, + "step": 652 + }, + { + "epoch": 0.17997316913762856, + "grad_norm": 0.40443646907806396, + "learning_rate": 1e-05, + "loss": 0.4789, + "step": 654 + }, + { + "epoch": 0.180523545801658, + "grad_norm": 0.35993334650993347, + "learning_rate": 1e-05, + "loss": 0.4902, + "step": 656 + }, + { + "epoch": 0.18107392246568746, + "grad_norm": 0.3933318853378296, + "learning_rate": 1e-05, + "loss": 0.4726, + "step": 658 + }, + { + "epoch": 0.1816242991297169, + "grad_norm": 0.3923085033893585, + "learning_rate": 1e-05, + "loss": 0.4714, + "step": 660 + }, + { + "epoch": 0.18217467579374635, + "grad_norm": 0.37387627363204956, + "learning_rate": 1e-05, + "loss": 0.478, + "step": 662 + }, + { + "epoch": 0.1827250524577758, + "grad_norm": 0.3787866532802582, + "learning_rate": 1e-05, + "loss": 0.4849, + "step": 664 + }, + { + "epoch": 0.18327542912180522, + "grad_norm": 0.39361730217933655, + "learning_rate": 1e-05, + "loss": 0.4836, + "step": 666 + }, + { + "epoch": 0.18382580578583468, + "grad_norm": 0.37430262565612793, + "learning_rate": 1e-05, + "loss": 0.4876, + "step": 668 + }, + { + "epoch": 0.18437618244986412, + "grad_norm": 0.3914833068847656, + "learning_rate": 1e-05, + "loss": 0.48, + "step": 670 + }, + { + "epoch": 0.18492655911389358, + "grad_norm": 0.36528506875038147, + "learning_rate": 1e-05, + "loss": 0.4583, + "step": 672 + }, + { + "epoch": 0.185476935777923, + "grad_norm": 0.3779620826244354, + "learning_rate": 1e-05, + "loss": 0.483, + "step": 674 + }, + { + "epoch": 0.18602731244195247, + "grad_norm": 0.3712228834629059, + "learning_rate": 1e-05, + "loss": 0.4833, + "step": 676 + }, + { + "epoch": 0.1865776891059819, + "grad_norm": 0.3959150016307831, + "learning_rate": 1e-05, + "loss": 0.4678, + "step": 678 + }, + { + "epoch": 0.18712806577001134, + "grad_norm": 0.38113903999328613, + "learning_rate": 1e-05, + "loss": 0.4794, + "step": 680 + }, + { + "epoch": 0.1876784424340408, + "grad_norm": 0.3872113525867462, + "learning_rate": 1e-05, + "loss": 0.4627, + "step": 682 + }, + { + "epoch": 0.18822881909807024, + "grad_norm": 0.35678407549858093, + "learning_rate": 1e-05, + "loss": 0.4666, + "step": 684 + }, + { + "epoch": 0.1887791957620997, + "grad_norm": 0.37833312153816223, + "learning_rate": 1e-05, + "loss": 0.4734, + "step": 686 + }, + { + "epoch": 0.18932957242612913, + "grad_norm": 0.3900817930698395, + "learning_rate": 1e-05, + "loss": 0.4834, + "step": 688 + }, + { + "epoch": 0.18987994909015857, + "grad_norm": 0.37114864587783813, + "learning_rate": 1e-05, + "loss": 0.4682, + "step": 690 + }, + { + "epoch": 0.19043032575418803, + "grad_norm": 0.37264662981033325, + "learning_rate": 1e-05, + "loss": 0.4815, + "step": 692 + }, + { + "epoch": 0.19098070241821746, + "grad_norm": 0.3758707344532013, + "learning_rate": 1e-05, + "loss": 0.4847, + "step": 694 + }, + { + "epoch": 0.19153107908224692, + "grad_norm": 0.38832512497901917, + "learning_rate": 1e-05, + "loss": 0.486, + "step": 696 + }, + { + "epoch": 0.19208145574627636, + "grad_norm": 0.382926344871521, + "learning_rate": 1e-05, + "loss": 0.4844, + "step": 698 + }, + { + "epoch": 0.1926318324103058, + "grad_norm": 0.3953557312488556, + "learning_rate": 1e-05, + "loss": 0.472, + "step": 700 + }, + { + "epoch": 0.19318220907433525, + "grad_norm": 0.36295419931411743, + "learning_rate": 1e-05, + "loss": 0.4792, + "step": 702 + }, + { + "epoch": 0.1937325857383647, + "grad_norm": 0.35859328508377075, + "learning_rate": 1e-05, + "loss": 0.4665, + "step": 704 + }, + { + "epoch": 0.19428296240239415, + "grad_norm": 0.3658142685890198, + "learning_rate": 1e-05, + "loss": 0.4724, + "step": 706 + }, + { + "epoch": 0.19483333906642358, + "grad_norm": 0.3860156834125519, + "learning_rate": 1e-05, + "loss": 0.4803, + "step": 708 + }, + { + "epoch": 0.19538371573045302, + "grad_norm": 0.38030922412872314, + "learning_rate": 1e-05, + "loss": 0.4692, + "step": 710 + }, + { + "epoch": 0.19593409239448248, + "grad_norm": 0.417516827583313, + "learning_rate": 1e-05, + "loss": 0.4833, + "step": 712 + }, + { + "epoch": 0.1964844690585119, + "grad_norm": 0.39626750349998474, + "learning_rate": 1e-05, + "loss": 0.4808, + "step": 714 + }, + { + "epoch": 0.19703484572254137, + "grad_norm": 0.3886042535305023, + "learning_rate": 1e-05, + "loss": 0.4716, + "step": 716 + }, + { + "epoch": 0.1975852223865708, + "grad_norm": 0.3816077411174774, + "learning_rate": 1e-05, + "loss": 0.468, + "step": 718 + }, + { + "epoch": 0.19813559905060024, + "grad_norm": 0.39385372400283813, + "learning_rate": 1e-05, + "loss": 0.4671, + "step": 720 + }, + { + "epoch": 0.1986859757146297, + "grad_norm": 0.35457953810691833, + "learning_rate": 1e-05, + "loss": 0.4667, + "step": 722 + }, + { + "epoch": 0.19923635237865914, + "grad_norm": 0.39437657594680786, + "learning_rate": 1e-05, + "loss": 0.4637, + "step": 724 + }, + { + "epoch": 0.1997867290426886, + "grad_norm": 0.41132184863090515, + "learning_rate": 1e-05, + "loss": 0.4723, + "step": 726 + }, + { + "epoch": 0.20033710570671803, + "grad_norm": 0.3640534281730652, + "learning_rate": 1e-05, + "loss": 0.4623, + "step": 728 + }, + { + "epoch": 0.20088748237074747, + "grad_norm": 0.39893659949302673, + "learning_rate": 1e-05, + "loss": 0.4881, + "step": 730 + }, + { + "epoch": 0.20143785903477693, + "grad_norm": 0.3677632212638855, + "learning_rate": 1e-05, + "loss": 0.4572, + "step": 732 + }, + { + "epoch": 0.20198823569880636, + "grad_norm": 0.40594953298568726, + "learning_rate": 1e-05, + "loss": 0.4726, + "step": 734 + }, + { + "epoch": 0.20253861236283582, + "grad_norm": 0.39571645855903625, + "learning_rate": 1e-05, + "loss": 0.4751, + "step": 736 + }, + { + "epoch": 0.20308898902686526, + "grad_norm": 0.3569906949996948, + "learning_rate": 1e-05, + "loss": 0.4855, + "step": 738 + }, + { + "epoch": 0.20363936569089472, + "grad_norm": 0.39166778326034546, + "learning_rate": 1e-05, + "loss": 0.4864, + "step": 740 + }, + { + "epoch": 0.20418974235492415, + "grad_norm": 0.36861687898635864, + "learning_rate": 1e-05, + "loss": 0.4659, + "step": 742 + }, + { + "epoch": 0.2047401190189536, + "grad_norm": 0.3691236078739166, + "learning_rate": 1e-05, + "loss": 0.4688, + "step": 744 + }, + { + "epoch": 0.20529049568298305, + "grad_norm": 0.41912853717803955, + "learning_rate": 1e-05, + "loss": 0.4787, + "step": 746 + }, + { + "epoch": 0.20584087234701248, + "grad_norm": 0.4022221565246582, + "learning_rate": 1e-05, + "loss": 0.4758, + "step": 748 + }, + { + "epoch": 0.20639124901104194, + "grad_norm": 0.402567595243454, + "learning_rate": 1e-05, + "loss": 0.4766, + "step": 750 + }, + { + "epoch": 0.20694162567507138, + "grad_norm": 0.3741600811481476, + "learning_rate": 1e-05, + "loss": 0.4833, + "step": 752 + }, + { + "epoch": 0.2074920023391008, + "grad_norm": 0.3958164155483246, + "learning_rate": 1e-05, + "loss": 0.4786, + "step": 754 + }, + { + "epoch": 0.20804237900313027, + "grad_norm": 0.37908801436424255, + "learning_rate": 1e-05, + "loss": 0.4715, + "step": 756 + }, + { + "epoch": 0.2085927556671597, + "grad_norm": 0.38426473736763, + "learning_rate": 1e-05, + "loss": 0.4874, + "step": 758 + }, + { + "epoch": 0.20914313233118917, + "grad_norm": 0.3873310983181, + "learning_rate": 1e-05, + "loss": 0.4677, + "step": 760 + }, + { + "epoch": 0.2096935089952186, + "grad_norm": 0.4033788740634918, + "learning_rate": 1e-05, + "loss": 0.4815, + "step": 762 + }, + { + "epoch": 0.21024388565924804, + "grad_norm": 0.40875962376594543, + "learning_rate": 1e-05, + "loss": 0.4879, + "step": 764 + }, + { + "epoch": 0.2107942623232775, + "grad_norm": 0.38724496960639954, + "learning_rate": 1e-05, + "loss": 0.4646, + "step": 766 + }, + { + "epoch": 0.21134463898730693, + "grad_norm": 0.39307013154029846, + "learning_rate": 1e-05, + "loss": 0.4762, + "step": 768 + }, + { + "epoch": 0.2118950156513364, + "grad_norm": 0.37346333265304565, + "learning_rate": 1e-05, + "loss": 0.4775, + "step": 770 + }, + { + "epoch": 0.21244539231536583, + "grad_norm": 0.3753449618816376, + "learning_rate": 1e-05, + "loss": 0.4707, + "step": 772 + }, + { + "epoch": 0.21299576897939526, + "grad_norm": 0.3829357922077179, + "learning_rate": 1e-05, + "loss": 0.4825, + "step": 774 + }, + { + "epoch": 0.21354614564342472, + "grad_norm": 0.35514822602272034, + "learning_rate": 1e-05, + "loss": 0.4779, + "step": 776 + }, + { + "epoch": 0.21409652230745416, + "grad_norm": 0.3714098036289215, + "learning_rate": 1e-05, + "loss": 0.4598, + "step": 778 + }, + { + "epoch": 0.21464689897148362, + "grad_norm": 0.3754241168498993, + "learning_rate": 1e-05, + "loss": 0.4736, + "step": 780 + }, + { + "epoch": 0.21519727563551305, + "grad_norm": 0.36637604236602783, + "learning_rate": 1e-05, + "loss": 0.4652, + "step": 782 + }, + { + "epoch": 0.2157476522995425, + "grad_norm": 0.367357075214386, + "learning_rate": 1e-05, + "loss": 0.466, + "step": 784 + }, + { + "epoch": 0.21629802896357195, + "grad_norm": 0.3747154176235199, + "learning_rate": 1e-05, + "loss": 0.4668, + "step": 786 + }, + { + "epoch": 0.21684840562760138, + "grad_norm": 0.3824009895324707, + "learning_rate": 1e-05, + "loss": 0.4748, + "step": 788 + }, + { + "epoch": 0.21739878229163084, + "grad_norm": 0.385030061006546, + "learning_rate": 1e-05, + "loss": 0.4882, + "step": 790 + }, + { + "epoch": 0.21794915895566028, + "grad_norm": 0.35460343956947327, + "learning_rate": 1e-05, + "loss": 0.4664, + "step": 792 + }, + { + "epoch": 0.21849953561968974, + "grad_norm": 0.3792308270931244, + "learning_rate": 1e-05, + "loss": 0.4874, + "step": 794 + }, + { + "epoch": 0.21904991228371917, + "grad_norm": 0.37190011143684387, + "learning_rate": 1e-05, + "loss": 0.4838, + "step": 796 + }, + { + "epoch": 0.2196002889477486, + "grad_norm": 0.3757864832878113, + "learning_rate": 1e-05, + "loss": 0.4538, + "step": 798 + }, + { + "epoch": 0.22015066561177807, + "grad_norm": 0.3677947223186493, + "learning_rate": 1e-05, + "loss": 0.4605, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_merge_loss": 0.4150216579437256, + "eval_merge_runtime": 600.1194, + "eval_merge_samples_per_second": 56.242, + "eval_merge_steps_per_second": 2.345, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_new_aug_datas_filtered.json_loss": 0.5434484481811523, + "eval_new_aug_datas_filtered.json_runtime": 10.4424, + "eval_new_aug_datas_filtered.json_samples_per_second": 73.45, + "eval_new_aug_datas_filtered.json_steps_per_second": 3.064, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_sharegpt_gpt4.json_loss": 0.7981637716293335, + "eval_sharegpt_gpt4.json_runtime": 31.6015, + "eval_sharegpt_gpt4.json_samples_per_second": 58.89, + "eval_sharegpt_gpt4.json_steps_per_second": 2.468, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_Table_GPT.json_loss": 0.0783885195851326, + "eval_Table_GPT.json_runtime": 24.9448, + "eval_Table_GPT.json_samples_per_second": 83.905, + "eval_Table_GPT.json_steps_per_second": 3.528, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_gpt_4o_200k.json_loss": 0.8245088458061218, + "eval_gpt_4o_200k.json_runtime": 48.4135, + "eval_gpt_4o_200k.json_samples_per_second": 129.737, + "eval_gpt_4o_200k.json_steps_per_second": 5.412, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_multi_turn_datas.json_loss": 0.35650402307510376, + "eval_multi_turn_datas.json_runtime": 75.5012, + "eval_multi_turn_datas.json_samples_per_second": 53.006, + "eval_multi_turn_datas.json_steps_per_second": 2.212, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_table_python_code_datas.json_loss": 0.2912423610687256, + "eval_table_python_code_datas.json_runtime": 43.0138, + "eval_table_python_code_datas.json_samples_per_second": 50.193, + "eval_table_python_code_datas.json_steps_per_second": 2.092, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_tabular_llm_data.json_loss": 0.11931464821100235, + "eval_tabular_llm_data.json_runtime": 8.524, + "eval_tabular_llm_data.json_samples_per_second": 28.86, + "eval_tabular_llm_data.json_steps_per_second": 1.29, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_python_code_critic_21k.json_loss": 0.5899094343185425, + "eval_python_code_critic_21k.json_runtime": 3.2108, + "eval_python_code_critic_21k.json_samples_per_second": 185.935, + "eval_python_code_critic_21k.json_steps_per_second": 7.786, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_all_merge_table_dataset.json_loss": 0.08210163563489914, + "eval_all_merge_table_dataset.json_runtime": 23.2334, + "eval_all_merge_table_dataset.json_samples_per_second": 30.646, + "eval_all_merge_table_dataset.json_steps_per_second": 1.291, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_code_feedback_multi_turn.json_loss": 0.5942392349243164, + "eval_code_feedback_multi_turn.json_runtime": 32.3672, + "eval_code_feedback_multi_turn.json_samples_per_second": 68.001, + "eval_code_feedback_multi_turn.json_steps_per_second": 2.842, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_ultrainteract_sft.json_loss": 0.43230774998664856, + "eval_ultrainteract_sft.json_runtime": 8.6469, + "eval_ultrainteract_sft.json_samples_per_second": 168.384, + "eval_ultrainteract_sft.json_steps_per_second": 7.055, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_synthetic_text_to_sql.json_loss": 0.10562511533498764, + "eval_synthetic_text_to_sql.json_runtime": 0.1256, + "eval_synthetic_text_to_sql.json_samples_per_second": 270.776, + "eval_synthetic_text_to_sql.json_steps_per_second": 15.928, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_sft_react_sql_datas.json_loss": 0.6536443829536438, + "eval_sft_react_sql_datas.json_runtime": 7.8424, + "eval_sft_react_sql_datas.json_samples_per_second": 40.039, + "eval_sft_react_sql_datas.json_steps_per_second": 1.785, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_all_merge_code.json_loss": 0.2989647090435028, + "eval_all_merge_code.json_runtime": 0.3335, + "eval_all_merge_code.json_samples_per_second": 188.9, + "eval_all_merge_code.json_steps_per_second": 8.995, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_magpie_datas.json_loss": 0.4389919340610504, + "eval_magpie_datas.json_runtime": 2.209, + "eval_magpie_datas.json_samples_per_second": 77.862, + "eval_magpie_datas.json_steps_per_second": 3.621, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_train_data_for_qwen.json_loss": 0.0057810284197330475, + "eval_train_data_for_qwen.json_runtime": 0.2434, + "eval_train_data_for_qwen.json_samples_per_second": 41.087, + "eval_train_data_for_qwen.json_steps_per_second": 4.109, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_alpaca_cleaned.json_loss": 0.9368440508842468, + "eval_alpaca_cleaned.json_runtime": 0.1149, + "eval_alpaca_cleaned.json_samples_per_second": 234.893, + "eval_alpaca_cleaned.json_steps_per_second": 17.399, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_agent_instruct.json_loss": 0.2261410802602768, + "eval_agent_instruct.json_runtime": 0.5137, + "eval_agent_instruct.json_samples_per_second": 93.432, + "eval_agent_instruct.json_steps_per_second": 3.893, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_MathInstruct.json_loss": 0.208473339676857, + "eval_MathInstruct.json_runtime": 0.3639, + "eval_MathInstruct.json_samples_per_second": 156.645, + "eval_MathInstruct.json_steps_per_second": 8.244, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_tested_143k_python_alpaca.json_loss": 0.44293999671936035, + "eval_tested_143k_python_alpaca.json_runtime": 0.2997, + "eval_tested_143k_python_alpaca.json_samples_per_second": 113.456, + "eval_tested_143k_python_alpaca.json_steps_per_second": 6.674, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_xlam_function_calling_60k.json_loss": 0.010015022940933704, + "eval_xlam_function_calling_60k.json_runtime": 0.1001, + "eval_xlam_function_calling_60k.json_samples_per_second": 229.814, + "eval_xlam_function_calling_60k.json_steps_per_second": 9.992, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_alpaca_data_gpt4_chinese.json_loss": 1.591582179069519, + "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0515, + "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 310.705, + "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.419, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_alpaca_gpt4_zh.json_loss": 0.9911380410194397, + "eval_alpaca_gpt4_zh.json_runtime": 0.0498, + "eval_alpaca_gpt4_zh.json_samples_per_second": 221.019, + "eval_alpaca_gpt4_zh.json_steps_per_second": 20.093, + "step": 800 + }, + { + "epoch": 0.22015066561177807, + "eval_codefeedback_filtered_instruction.json_loss": 0.5947377681732178, + "eval_codefeedback_filtered_instruction.json_runtime": 0.485, + "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.237, + "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.062, + "step": 800 + }, + { + "epoch": 0.2207010422758075, + "grad_norm": 0.36098968982696533, + "learning_rate": 1e-05, + "loss": 0.4646, + "step": 802 + }, + { + "epoch": 0.22125141893983696, + "grad_norm": 0.3653786778450012, + "learning_rate": 1e-05, + "loss": 0.4501, + "step": 804 + }, + { + "epoch": 0.2218017956038664, + "grad_norm": 0.36442849040031433, + "learning_rate": 1e-05, + "loss": 0.4686, + "step": 806 + }, + { + "epoch": 0.22235217226789583, + "grad_norm": 0.3782612383365631, + "learning_rate": 1e-05, + "loss": 0.4598, + "step": 808 + }, + { + "epoch": 0.2229025489319253, + "grad_norm": 0.39521896839141846, + "learning_rate": 1e-05, + "loss": 0.4679, + "step": 810 + }, + { + "epoch": 0.22345292559595473, + "grad_norm": 0.3727470636367798, + "learning_rate": 1e-05, + "loss": 0.4803, + "step": 812 + }, + { + "epoch": 0.2240033022599842, + "grad_norm": 0.3883068263530731, + "learning_rate": 1e-05, + "loss": 0.4773, + "step": 814 + }, + { + "epoch": 0.22455367892401362, + "grad_norm": 0.37147605419158936, + "learning_rate": 1e-05, + "loss": 0.4825, + "step": 816 + }, + { + "epoch": 0.22510405558804306, + "grad_norm": 0.3924333155155182, + "learning_rate": 1e-05, + "loss": 0.4698, + "step": 818 + }, + { + "epoch": 0.22565443225207252, + "grad_norm": 0.38133057951927185, + "learning_rate": 1e-05, + "loss": 0.4842, + "step": 820 + }, + { + "epoch": 0.22620480891610195, + "grad_norm": 0.36132821440696716, + "learning_rate": 1e-05, + "loss": 0.4594, + "step": 822 + }, + { + "epoch": 0.2267551855801314, + "grad_norm": 0.39988580346107483, + "learning_rate": 1e-05, + "loss": 0.4795, + "step": 824 + }, + { + "epoch": 0.22730556224416085, + "grad_norm": 0.38140830397605896, + "learning_rate": 1e-05, + "loss": 0.4649, + "step": 826 + }, + { + "epoch": 0.22785593890819028, + "grad_norm": 0.3726978898048401, + "learning_rate": 1e-05, + "loss": 0.4603, + "step": 828 + }, + { + "epoch": 0.22840631557221974, + "grad_norm": 0.3880995512008667, + "learning_rate": 1e-05, + "loss": 0.4739, + "step": 830 + }, + { + "epoch": 0.22895669223624918, + "grad_norm": 0.4118787944316864, + "learning_rate": 1e-05, + "loss": 0.4733, + "step": 832 + }, + { + "epoch": 0.22950706890027864, + "grad_norm": 0.37878745794296265, + "learning_rate": 1e-05, + "loss": 0.4922, + "step": 834 + }, + { + "epoch": 0.23005744556430807, + "grad_norm": 0.3838474154472351, + "learning_rate": 1e-05, + "loss": 0.4646, + "step": 836 + }, + { + "epoch": 0.2306078222283375, + "grad_norm": 0.37345945835113525, + "learning_rate": 1e-05, + "loss": 0.4798, + "step": 838 + }, + { + "epoch": 0.23115819889236697, + "grad_norm": 0.36341801285743713, + "learning_rate": 1e-05, + "loss": 0.4773, + "step": 840 + }, + { + "epoch": 0.2317085755563964, + "grad_norm": 0.38800522685050964, + "learning_rate": 1e-05, + "loss": 0.4806, + "step": 842 + }, + { + "epoch": 0.23225895222042586, + "grad_norm": 0.38882526755332947, + "learning_rate": 1e-05, + "loss": 0.4765, + "step": 844 + }, + { + "epoch": 0.2328093288844553, + "grad_norm": 0.37744489312171936, + "learning_rate": 1e-05, + "loss": 0.4842, + "step": 846 + }, + { + "epoch": 0.23335970554848473, + "grad_norm": 0.39916718006134033, + "learning_rate": 1e-05, + "loss": 0.467, + "step": 848 + }, + { + "epoch": 0.2339100822125142, + "grad_norm": 0.36556801199913025, + "learning_rate": 1e-05, + "loss": 0.4711, + "step": 850 + }, + { + "epoch": 0.23446045887654363, + "grad_norm": 0.3993853032588959, + "learning_rate": 1e-05, + "loss": 0.4656, + "step": 852 + }, + { + "epoch": 0.2350108355405731, + "grad_norm": 0.39630356431007385, + "learning_rate": 1e-05, + "loss": 0.4734, + "step": 854 + }, + { + "epoch": 0.23556121220460252, + "grad_norm": 0.3797578513622284, + "learning_rate": 1e-05, + "loss": 0.4718, + "step": 856 + }, + { + "epoch": 0.23611158886863198, + "grad_norm": 0.38648873567581177, + "learning_rate": 1e-05, + "loss": 0.4751, + "step": 858 + }, + { + "epoch": 0.23666196553266142, + "grad_norm": 0.3934420347213745, + "learning_rate": 1e-05, + "loss": 0.4653, + "step": 860 + }, + { + "epoch": 0.23721234219669085, + "grad_norm": 0.3899431824684143, + "learning_rate": 1e-05, + "loss": 0.4644, + "step": 862 + }, + { + "epoch": 0.2377627188607203, + "grad_norm": 0.3696826696395874, + "learning_rate": 1e-05, + "loss": 0.482, + "step": 864 + }, + { + "epoch": 0.23831309552474975, + "grad_norm": 0.352923184633255, + "learning_rate": 1e-05, + "loss": 0.4707, + "step": 866 + }, + { + "epoch": 0.2388634721887792, + "grad_norm": 0.36678972840309143, + "learning_rate": 1e-05, + "loss": 0.4687, + "step": 868 + }, + { + "epoch": 0.23941384885280864, + "grad_norm": 0.38986021280288696, + "learning_rate": 1e-05, + "loss": 0.4613, + "step": 870 + }, + { + "epoch": 0.23996422551683808, + "grad_norm": 0.3684535622596741, + "learning_rate": 1e-05, + "loss": 0.4734, + "step": 872 + }, + { + "epoch": 0.24051460218086754, + "grad_norm": 0.36672261357307434, + "learning_rate": 1e-05, + "loss": 0.4796, + "step": 874 + }, + { + "epoch": 0.24106497884489697, + "grad_norm": 0.39910420775413513, + "learning_rate": 1e-05, + "loss": 0.4681, + "step": 876 + }, + { + "epoch": 0.24161535550892643, + "grad_norm": 0.38694077730178833, + "learning_rate": 1e-05, + "loss": 0.4821, + "step": 878 + }, + { + "epoch": 0.24216573217295587, + "grad_norm": 0.4555080831050873, + "learning_rate": 1e-05, + "loss": 0.4882, + "step": 880 + }, + { + "epoch": 0.2427161088369853, + "grad_norm": 0.3934450149536133, + "learning_rate": 1e-05, + "loss": 0.4778, + "step": 882 + }, + { + "epoch": 0.24326648550101476, + "grad_norm": 0.35743412375450134, + "learning_rate": 1e-05, + "loss": 0.4793, + "step": 884 + }, + { + "epoch": 0.2438168621650442, + "grad_norm": 0.3518178462982178, + "learning_rate": 1e-05, + "loss": 0.472, + "step": 886 + }, + { + "epoch": 0.24436723882907366, + "grad_norm": 0.35367751121520996, + "learning_rate": 1e-05, + "loss": 0.4747, + "step": 888 + }, + { + "epoch": 0.2449176154931031, + "grad_norm": 0.3810805678367615, + "learning_rate": 1e-05, + "loss": 0.4834, + "step": 890 + }, + { + "epoch": 0.24546799215713253, + "grad_norm": 0.38103243708610535, + "learning_rate": 1e-05, + "loss": 0.4763, + "step": 892 + }, + { + "epoch": 0.246018368821162, + "grad_norm": 0.3839399218559265, + "learning_rate": 1e-05, + "loss": 0.4696, + "step": 894 + }, + { + "epoch": 0.24656874548519142, + "grad_norm": 0.41292649507522583, + "learning_rate": 1e-05, + "loss": 0.4777, + "step": 896 + }, + { + "epoch": 0.24711912214922088, + "grad_norm": 0.36179229617118835, + "learning_rate": 1e-05, + "loss": 0.4668, + "step": 898 + }, + { + "epoch": 0.24766949881325032, + "grad_norm": 0.3638279139995575, + "learning_rate": 1e-05, + "loss": 0.4645, + "step": 900 + }, + { + "epoch": 0.24821987547727975, + "grad_norm": 0.3458470106124878, + "learning_rate": 1e-05, + "loss": 0.4746, + "step": 902 + }, + { + "epoch": 0.2487702521413092, + "grad_norm": 0.3822806775569916, + "learning_rate": 1e-05, + "loss": 0.4715, + "step": 904 + }, + { + "epoch": 0.24932062880533865, + "grad_norm": 0.3655596077442169, + "learning_rate": 1e-05, + "loss": 0.4659, + "step": 906 + }, + { + "epoch": 0.2498710054693681, + "grad_norm": 0.3868783116340637, + "learning_rate": 1e-05, + "loss": 0.4743, + "step": 908 + }, + { + "epoch": 0.25042138213339754, + "grad_norm": 0.3778232932090759, + "learning_rate": 1e-05, + "loss": 0.4652, + "step": 910 + }, + { + "epoch": 0.250971758797427, + "grad_norm": 0.36664894223213196, + "learning_rate": 1e-05, + "loss": 0.4554, + "step": 912 + }, + { + "epoch": 0.2515221354614564, + "grad_norm": 0.3995139002799988, + "learning_rate": 1e-05, + "loss": 0.4683, + "step": 914 + }, + { + "epoch": 0.25207251212548587, + "grad_norm": 0.40083470940589905, + "learning_rate": 1e-05, + "loss": 0.4673, + "step": 916 + }, + { + "epoch": 0.25262288878951533, + "grad_norm": 0.37919968366622925, + "learning_rate": 1e-05, + "loss": 0.4776, + "step": 918 + }, + { + "epoch": 0.2531732654535448, + "grad_norm": 0.3586704432964325, + "learning_rate": 1e-05, + "loss": 0.4792, + "step": 920 + }, + { + "epoch": 0.2537236421175742, + "grad_norm": 0.3744722902774811, + "learning_rate": 1e-05, + "loss": 0.463, + "step": 922 + }, + { + "epoch": 0.25427401878160366, + "grad_norm": 0.37209680676460266, + "learning_rate": 1e-05, + "loss": 0.4805, + "step": 924 + }, + { + "epoch": 0.2548243954456331, + "grad_norm": 0.40809133648872375, + "learning_rate": 1e-05, + "loss": 0.4781, + "step": 926 + }, + { + "epoch": 0.25537477210966253, + "grad_norm": 0.37261903285980225, + "learning_rate": 1e-05, + "loss": 0.4617, + "step": 928 + }, + { + "epoch": 0.255925148773692, + "grad_norm": 0.37391313910484314, + "learning_rate": 1e-05, + "loss": 0.4617, + "step": 930 + }, + { + "epoch": 0.25647552543772145, + "grad_norm": 0.36610838770866394, + "learning_rate": 1e-05, + "loss": 0.4642, + "step": 932 + }, + { + "epoch": 0.25702590210175086, + "grad_norm": 0.3854142129421234, + "learning_rate": 1e-05, + "loss": 0.4652, + "step": 934 + }, + { + "epoch": 0.2575762787657803, + "grad_norm": 0.365159809589386, + "learning_rate": 1e-05, + "loss": 0.4714, + "step": 936 + }, + { + "epoch": 0.2581266554298098, + "grad_norm": 0.41678836941719055, + "learning_rate": 1e-05, + "loss": 0.4854, + "step": 938 + }, + { + "epoch": 0.25867703209383924, + "grad_norm": 0.380215585231781, + "learning_rate": 1e-05, + "loss": 0.4785, + "step": 940 + }, + { + "epoch": 0.25922740875786865, + "grad_norm": 0.3704361617565155, + "learning_rate": 1e-05, + "loss": 0.4433, + "step": 942 + }, + { + "epoch": 0.2597777854218981, + "grad_norm": 0.34440556168556213, + "learning_rate": 1e-05, + "loss": 0.4642, + "step": 944 + }, + { + "epoch": 0.2603281620859276, + "grad_norm": 0.36701446771621704, + "learning_rate": 1e-05, + "loss": 0.4533, + "step": 946 + }, + { + "epoch": 0.260878538749957, + "grad_norm": 0.3694971799850464, + "learning_rate": 1e-05, + "loss": 0.4942, + "step": 948 + }, + { + "epoch": 0.26142891541398644, + "grad_norm": 0.3697713017463684, + "learning_rate": 1e-05, + "loss": 0.4586, + "step": 950 + }, + { + "epoch": 0.2619792920780159, + "grad_norm": 0.36559173464775085, + "learning_rate": 1e-05, + "loss": 0.4679, + "step": 952 + }, + { + "epoch": 0.26252966874204536, + "grad_norm": 0.3704969584941864, + "learning_rate": 1e-05, + "loss": 0.4624, + "step": 954 + }, + { + "epoch": 0.26308004540607477, + "grad_norm": 0.3804495334625244, + "learning_rate": 1e-05, + "loss": 0.4603, + "step": 956 + }, + { + "epoch": 0.26363042207010423, + "grad_norm": 0.34987303614616394, + "learning_rate": 1e-05, + "loss": 0.4679, + "step": 958 + }, + { + "epoch": 0.2641807987341337, + "grad_norm": 0.3723856508731842, + "learning_rate": 1e-05, + "loss": 0.4631, + "step": 960 + }, + { + "epoch": 0.2647311753981631, + "grad_norm": 0.35623612999916077, + "learning_rate": 1e-05, + "loss": 0.4627, + "step": 962 + }, + { + "epoch": 0.26528155206219256, + "grad_norm": 0.37969711422920227, + "learning_rate": 1e-05, + "loss": 0.4815, + "step": 964 + }, + { + "epoch": 0.265831928726222, + "grad_norm": 0.3889734447002411, + "learning_rate": 1e-05, + "loss": 0.471, + "step": 966 + }, + { + "epoch": 0.26638230539025143, + "grad_norm": 0.39106228947639465, + "learning_rate": 1e-05, + "loss": 0.4542, + "step": 968 + }, + { + "epoch": 0.2669326820542809, + "grad_norm": 0.38163650035858154, + "learning_rate": 1e-05, + "loss": 0.4604, + "step": 970 + }, + { + "epoch": 0.26748305871831035, + "grad_norm": 0.3733852505683899, + "learning_rate": 1e-05, + "loss": 0.4768, + "step": 972 + }, + { + "epoch": 0.2680334353823398, + "grad_norm": 0.3894038796424866, + "learning_rate": 1e-05, + "loss": 0.4816, + "step": 974 + }, + { + "epoch": 0.2685838120463692, + "grad_norm": 0.3697439432144165, + "learning_rate": 1e-05, + "loss": 0.4731, + "step": 976 + }, + { + "epoch": 0.2691341887103987, + "grad_norm": 0.39549171924591064, + "learning_rate": 1e-05, + "loss": 0.473, + "step": 978 + }, + { + "epoch": 0.26968456537442814, + "grad_norm": 0.38712403178215027, + "learning_rate": 1e-05, + "loss": 0.4717, + "step": 980 + }, + { + "epoch": 0.27023494203845755, + "grad_norm": 0.3775619864463806, + "learning_rate": 1e-05, + "loss": 0.4638, + "step": 982 + }, + { + "epoch": 0.270785318702487, + "grad_norm": 0.38664135336875916, + "learning_rate": 1e-05, + "loss": 0.4655, + "step": 984 + }, + { + "epoch": 0.27133569536651647, + "grad_norm": 0.3730804920196533, + "learning_rate": 1e-05, + "loss": 0.4672, + "step": 986 + }, + { + "epoch": 0.2718860720305459, + "grad_norm": 0.36626750230789185, + "learning_rate": 1e-05, + "loss": 0.4562, + "step": 988 + }, + { + "epoch": 0.27243644869457534, + "grad_norm": 0.38708406686782837, + "learning_rate": 1e-05, + "loss": 0.4583, + "step": 990 + }, + { + "epoch": 0.2729868253586048, + "grad_norm": 0.37348565459251404, + "learning_rate": 1e-05, + "loss": 0.4709, + "step": 992 + }, + { + "epoch": 0.27353720202263426, + "grad_norm": 0.39145755767822266, + "learning_rate": 1e-05, + "loss": 0.4667, + "step": 994 + }, + { + "epoch": 0.27408757868666367, + "grad_norm": 0.3615020215511322, + "learning_rate": 1e-05, + "loss": 0.4585, + "step": 996 + }, + { + "epoch": 0.27463795535069313, + "grad_norm": 0.38545548915863037, + "learning_rate": 1e-05, + "loss": 0.472, + "step": 998 + }, + { + "epoch": 0.2751883320147226, + "grad_norm": 0.3605005741119385, + "learning_rate": 1e-05, + "loss": 0.4575, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_merge_loss": 0.4092504382133484, + "eval_merge_runtime": 599.649, + "eval_merge_samples_per_second": 56.286, + "eval_merge_steps_per_second": 2.346, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_new_aug_datas_filtered.json_loss": 0.534787118434906, + "eval_new_aug_datas_filtered.json_runtime": 10.3465, + "eval_new_aug_datas_filtered.json_samples_per_second": 74.131, + "eval_new_aug_datas_filtered.json_steps_per_second": 3.093, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_sharegpt_gpt4.json_loss": 0.7911589741706848, + "eval_sharegpt_gpt4.json_runtime": 31.721, + "eval_sharegpt_gpt4.json_samples_per_second": 58.668, + "eval_sharegpt_gpt4.json_steps_per_second": 2.459, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_Table_GPT.json_loss": 0.07000603526830673, + "eval_Table_GPT.json_runtime": 24.9973, + "eval_Table_GPT.json_samples_per_second": 83.729, + "eval_Table_GPT.json_steps_per_second": 3.52, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_gpt_4o_200k.json_loss": 0.8180866837501526, + "eval_gpt_4o_200k.json_runtime": 48.5388, + "eval_gpt_4o_200k.json_samples_per_second": 129.402, + "eval_gpt_4o_200k.json_steps_per_second": 5.398, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_multi_turn_datas.json_loss": 0.34955134987831116, + "eval_multi_turn_datas.json_runtime": 75.86, + "eval_multi_turn_datas.json_samples_per_second": 52.755, + "eval_multi_turn_datas.json_steps_per_second": 2.201, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_table_python_code_datas.json_loss": 0.285086989402771, + "eval_table_python_code_datas.json_runtime": 43.1585, + "eval_table_python_code_datas.json_samples_per_second": 50.025, + "eval_table_python_code_datas.json_steps_per_second": 2.085, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_tabular_llm_data.json_loss": 0.12198314070701599, + "eval_tabular_llm_data.json_runtime": 8.5654, + "eval_tabular_llm_data.json_samples_per_second": 28.72, + "eval_tabular_llm_data.json_steps_per_second": 1.284, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_python_code_critic_21k.json_loss": 0.5841899514198303, + "eval_python_code_critic_21k.json_runtime": 3.2248, + "eval_python_code_critic_21k.json_samples_per_second": 185.125, + "eval_python_code_critic_21k.json_steps_per_second": 7.752, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_all_merge_table_dataset.json_loss": 0.08110550791025162, + "eval_all_merge_table_dataset.json_runtime": 23.4122, + "eval_all_merge_table_dataset.json_samples_per_second": 30.411, + "eval_all_merge_table_dataset.json_steps_per_second": 1.281, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_code_feedback_multi_turn.json_loss": 0.5908513069152832, + "eval_code_feedback_multi_turn.json_runtime": 32.4627, + "eval_code_feedback_multi_turn.json_samples_per_second": 67.801, + "eval_code_feedback_multi_turn.json_steps_per_second": 2.834, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_ultrainteract_sft.json_loss": 0.42869675159454346, + "eval_ultrainteract_sft.json_runtime": 8.6816, + "eval_ultrainteract_sft.json_samples_per_second": 167.711, + "eval_ultrainteract_sft.json_steps_per_second": 7.026, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_synthetic_text_to_sql.json_loss": 0.10359195619821548, + "eval_synthetic_text_to_sql.json_runtime": 0.1301, + "eval_synthetic_text_to_sql.json_samples_per_second": 261.368, + "eval_synthetic_text_to_sql.json_steps_per_second": 15.375, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_sft_react_sql_datas.json_loss": 0.6493918895721436, + "eval_sft_react_sql_datas.json_runtime": 7.8489, + "eval_sft_react_sql_datas.json_samples_per_second": 40.006, + "eval_sft_react_sql_datas.json_steps_per_second": 1.784, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_all_merge_code.json_loss": 0.29959577322006226, + "eval_all_merge_code.json_runtime": 0.3379, + "eval_all_merge_code.json_samples_per_second": 186.458, + "eval_all_merge_code.json_steps_per_second": 8.879, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_magpie_datas.json_loss": 0.4377444088459015, + "eval_magpie_datas.json_runtime": 2.2091, + "eval_magpie_datas.json_samples_per_second": 77.86, + "eval_magpie_datas.json_steps_per_second": 3.621, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_train_data_for_qwen.json_loss": 0.003975613508373499, + "eval_train_data_for_qwen.json_runtime": 0.2434, + "eval_train_data_for_qwen.json_samples_per_second": 41.087, + "eval_train_data_for_qwen.json_steps_per_second": 4.109, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_alpaca_cleaned.json_loss": 0.9270830750465393, + "eval_alpaca_cleaned.json_runtime": 0.1147, + "eval_alpaca_cleaned.json_samples_per_second": 235.404, + "eval_alpaca_cleaned.json_steps_per_second": 17.437, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_agent_instruct.json_loss": 0.2235051840543747, + "eval_agent_instruct.json_runtime": 0.5147, + "eval_agent_instruct.json_samples_per_second": 93.255, + "eval_agent_instruct.json_steps_per_second": 3.886, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_MathInstruct.json_loss": 0.20924758911132812, + "eval_MathInstruct.json_runtime": 0.3588, + "eval_MathInstruct.json_samples_per_second": 158.853, + "eval_MathInstruct.json_steps_per_second": 8.361, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_tested_143k_python_alpaca.json_loss": 0.44443246722221375, + "eval_tested_143k_python_alpaca.json_runtime": 0.3017, + "eval_tested_143k_python_alpaca.json_samples_per_second": 112.684, + "eval_tested_143k_python_alpaca.json_steps_per_second": 6.628, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_xlam_function_calling_60k.json_loss": 0.008116651326417923, + "eval_xlam_function_calling_60k.json_runtime": 0.1004, + "eval_xlam_function_calling_60k.json_samples_per_second": 229.157, + "eval_xlam_function_calling_60k.json_steps_per_second": 9.963, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_alpaca_data_gpt4_chinese.json_loss": 1.588812232017517, + "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0516, + "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 310.032, + "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.377, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_alpaca_gpt4_zh.json_loss": 0.9696416258811951, + "eval_alpaca_gpt4_zh.json_runtime": 0.0501, + "eval_alpaca_gpt4_zh.json_samples_per_second": 219.488, + "eval_alpaca_gpt4_zh.json_steps_per_second": 19.953, + "step": 1000 + }, + { + "epoch": 0.2751883320147226, + "eval_codefeedback_filtered_instruction.json_loss": 0.5965829491615295, + "eval_codefeedback_filtered_instruction.json_runtime": 0.4872, + "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.049, + "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.052, + "step": 1000 + }, + { + "epoch": 0.275738708678752, + "grad_norm": 0.3598334789276123, + "learning_rate": 1e-05, + "loss": 0.4624, + "step": 1002 + }, + { + "epoch": 0.27628908534278146, + "grad_norm": 0.3716166615486145, + "learning_rate": 1e-05, + "loss": 0.4593, + "step": 1004 + }, + { + "epoch": 0.2768394620068109, + "grad_norm": 0.3814164996147156, + "learning_rate": 1e-05, + "loss": 0.4581, + "step": 1006 + }, + { + "epoch": 0.2773898386708404, + "grad_norm": 0.3595026731491089, + "learning_rate": 1e-05, + "loss": 0.4826, + "step": 1008 + }, + { + "epoch": 0.2779402153348698, + "grad_norm": 0.3587126135826111, + "learning_rate": 1e-05, + "loss": 0.4563, + "step": 1010 + }, + { + "epoch": 0.27849059199889925, + "grad_norm": 0.36048388481140137, + "learning_rate": 1e-05, + "loss": 0.4646, + "step": 1012 + }, + { + "epoch": 0.2790409686629287, + "grad_norm": 0.37650784850120544, + "learning_rate": 1e-05, + "loss": 0.4658, + "step": 1014 + }, + { + "epoch": 0.2795913453269581, + "grad_norm": 0.34934109449386597, + "learning_rate": 1e-05, + "loss": 0.4535, + "step": 1016 + }, + { + "epoch": 0.2801417219909876, + "grad_norm": 0.375130295753479, + "learning_rate": 1e-05, + "loss": 0.4802, + "step": 1018 + }, + { + "epoch": 0.28069209865501704, + "grad_norm": 0.3595198094844818, + "learning_rate": 1e-05, + "loss": 0.4736, + "step": 1020 + }, + { + "epoch": 0.28124247531904645, + "grad_norm": 0.37816157937049866, + "learning_rate": 1e-05, + "loss": 0.4639, + "step": 1022 + }, + { + "epoch": 0.2817928519830759, + "grad_norm": 0.39598193764686584, + "learning_rate": 1e-05, + "loss": 0.4544, + "step": 1024 + }, + { + "epoch": 0.28234322864710537, + "grad_norm": 0.35407206416130066, + "learning_rate": 1e-05, + "loss": 0.4342, + "step": 1026 + }, + { + "epoch": 0.28289360531113483, + "grad_norm": 0.3630298972129822, + "learning_rate": 1e-05, + "loss": 0.48, + "step": 1028 + }, + { + "epoch": 0.28344398197516424, + "grad_norm": 0.35917675495147705, + "learning_rate": 1e-05, + "loss": 0.4647, + "step": 1030 + }, + { + "epoch": 0.2839943586391937, + "grad_norm": 0.36868980526924133, + "learning_rate": 1e-05, + "loss": 0.4633, + "step": 1032 + }, + { + "epoch": 0.28454473530322316, + "grad_norm": 0.38559168577194214, + "learning_rate": 1e-05, + "loss": 0.4786, + "step": 1034 + }, + { + "epoch": 0.28509511196725257, + "grad_norm": 0.3563440442085266, + "learning_rate": 1e-05, + "loss": 0.4703, + "step": 1036 + }, + { + "epoch": 0.28564548863128203, + "grad_norm": 0.3761630654335022, + "learning_rate": 1e-05, + "loss": 0.4712, + "step": 1038 + }, + { + "epoch": 0.2861958652953115, + "grad_norm": 0.3870238661766052, + "learning_rate": 1e-05, + "loss": 0.4622, + "step": 1040 + }, + { + "epoch": 0.2867462419593409, + "grad_norm": 0.36192306876182556, + "learning_rate": 1e-05, + "loss": 0.4619, + "step": 1042 + }, + { + "epoch": 0.28729661862337036, + "grad_norm": 0.3688748776912689, + "learning_rate": 1e-05, + "loss": 0.457, + "step": 1044 + }, + { + "epoch": 0.2878469952873998, + "grad_norm": 0.38211309909820557, + "learning_rate": 1e-05, + "loss": 0.4661, + "step": 1046 + }, + { + "epoch": 0.2883973719514293, + "grad_norm": 0.36421847343444824, + "learning_rate": 1e-05, + "loss": 0.4647, + "step": 1048 + }, + { + "epoch": 0.2889477486154587, + "grad_norm": 0.38917919993400574, + "learning_rate": 1e-05, + "loss": 0.4573, + "step": 1050 + }, + { + "epoch": 0.28949812527948815, + "grad_norm": 0.3668692111968994, + "learning_rate": 1e-05, + "loss": 0.4545, + "step": 1052 + }, + { + "epoch": 0.2900485019435176, + "grad_norm": 0.3869079649448395, + "learning_rate": 1e-05, + "loss": 0.4569, + "step": 1054 + }, + { + "epoch": 0.290598878607547, + "grad_norm": 0.3763209283351898, + "learning_rate": 1e-05, + "loss": 0.451, + "step": 1056 + }, + { + "epoch": 0.2911492552715765, + "grad_norm": 0.37899014353752136, + "learning_rate": 1e-05, + "loss": 0.4658, + "step": 1058 + }, + { + "epoch": 0.29169963193560594, + "grad_norm": 0.38784778118133545, + "learning_rate": 1e-05, + "loss": 0.4589, + "step": 1060 + }, + { + "epoch": 0.2922500085996354, + "grad_norm": 0.38340142369270325, + "learning_rate": 1e-05, + "loss": 0.4644, + "step": 1062 + }, + { + "epoch": 0.2928003852636648, + "grad_norm": 0.3758372962474823, + "learning_rate": 1e-05, + "loss": 0.4597, + "step": 1064 + }, + { + "epoch": 0.29335076192769427, + "grad_norm": 0.36990198493003845, + "learning_rate": 1e-05, + "loss": 0.4577, + "step": 1066 + }, + { + "epoch": 0.29390113859172373, + "grad_norm": 0.35997095704078674, + "learning_rate": 1e-05, + "loss": 0.452, + "step": 1068 + }, + { + "epoch": 0.29445151525575314, + "grad_norm": 0.3728466331958771, + "learning_rate": 1e-05, + "loss": 0.4567, + "step": 1070 + }, + { + "epoch": 0.2950018919197826, + "grad_norm": 0.3471437990665436, + "learning_rate": 1e-05, + "loss": 0.4661, + "step": 1072 + }, + { + "epoch": 0.29555226858381206, + "grad_norm": 0.39197105169296265, + "learning_rate": 1e-05, + "loss": 0.4738, + "step": 1074 + }, + { + "epoch": 0.29610264524784147, + "grad_norm": 0.366745263338089, + "learning_rate": 1e-05, + "loss": 0.4555, + "step": 1076 + }, + { + "epoch": 0.29665302191187093, + "grad_norm": 0.3721451759338379, + "learning_rate": 1e-05, + "loss": 0.4784, + "step": 1078 + }, + { + "epoch": 0.2972033985759004, + "grad_norm": 0.3505246341228485, + "learning_rate": 1e-05, + "loss": 0.4486, + "step": 1080 + }, + { + "epoch": 0.29775377523992985, + "grad_norm": 0.37022680044174194, + "learning_rate": 1e-05, + "loss": 0.4631, + "step": 1082 + }, + { + "epoch": 0.29830415190395926, + "grad_norm": 0.3808286190032959, + "learning_rate": 1e-05, + "loss": 0.472, + "step": 1084 + }, + { + "epoch": 0.2988545285679887, + "grad_norm": 0.3860435485839844, + "learning_rate": 1e-05, + "loss": 0.4541, + "step": 1086 + }, + { + "epoch": 0.2994049052320182, + "grad_norm": 0.35552406311035156, + "learning_rate": 1e-05, + "loss": 0.4565, + "step": 1088 + }, + { + "epoch": 0.2999552818960476, + "grad_norm": 0.3758242428302765, + "learning_rate": 1e-05, + "loss": 0.4803, + "step": 1090 + }, + { + "epoch": 0.30050565856007705, + "grad_norm": 0.3900710940361023, + "learning_rate": 1e-05, + "loss": 0.4658, + "step": 1092 + }, + { + "epoch": 0.3010560352241065, + "grad_norm": 0.38439512252807617, + "learning_rate": 1e-05, + "loss": 0.4677, + "step": 1094 + }, + { + "epoch": 0.3016064118881359, + "grad_norm": 0.3970472812652588, + "learning_rate": 1e-05, + "loss": 0.4751, + "step": 1096 + }, + { + "epoch": 0.3021567885521654, + "grad_norm": 0.36555778980255127, + "learning_rate": 1e-05, + "loss": 0.4556, + "step": 1098 + }, + { + "epoch": 0.30270716521619484, + "grad_norm": 0.3682638108730316, + "learning_rate": 1e-05, + "loss": 0.4504, + "step": 1100 + }, + { + "epoch": 0.3032575418802243, + "grad_norm": 0.4228995442390442, + "learning_rate": 1e-05, + "loss": 0.4736, + "step": 1102 + }, + { + "epoch": 0.3038079185442537, + "grad_norm": 0.35070449113845825, + "learning_rate": 1e-05, + "loss": 0.4589, + "step": 1104 + }, + { + "epoch": 0.30435829520828317, + "grad_norm": 0.40524446964263916, + "learning_rate": 1e-05, + "loss": 0.4616, + "step": 1106 + }, + { + "epoch": 0.30490867187231263, + "grad_norm": 0.3461023271083832, + "learning_rate": 1e-05, + "loss": 0.4679, + "step": 1108 + }, + { + "epoch": 0.30545904853634204, + "grad_norm": 0.3741723299026489, + "learning_rate": 1e-05, + "loss": 0.4618, + "step": 1110 + }, + { + "epoch": 0.3060094252003715, + "grad_norm": 0.37440451979637146, + "learning_rate": 1e-05, + "loss": 0.4638, + "step": 1112 + }, + { + "epoch": 0.30655980186440096, + "grad_norm": 0.34469377994537354, + "learning_rate": 1e-05, + "loss": 0.4426, + "step": 1114 + }, + { + "epoch": 0.3071101785284304, + "grad_norm": 0.35499683022499084, + "learning_rate": 1e-05, + "loss": 0.4548, + "step": 1116 + }, + { + "epoch": 0.30766055519245983, + "grad_norm": 0.3623688220977783, + "learning_rate": 1e-05, + "loss": 0.4574, + "step": 1118 + }, + { + "epoch": 0.3082109318564893, + "grad_norm": 0.3487359583377838, + "learning_rate": 1e-05, + "loss": 0.4632, + "step": 1120 + }, + { + "epoch": 0.30876130852051875, + "grad_norm": 0.36232292652130127, + "learning_rate": 1e-05, + "loss": 0.462, + "step": 1122 + }, + { + "epoch": 0.30931168518454816, + "grad_norm": 0.38301897048950195, + "learning_rate": 1e-05, + "loss": 0.4545, + "step": 1124 + }, + { + "epoch": 0.3098620618485776, + "grad_norm": 0.3788921535015106, + "learning_rate": 1e-05, + "loss": 0.4614, + "step": 1126 + }, + { + "epoch": 0.3104124385126071, + "grad_norm": 0.3723096251487732, + "learning_rate": 1e-05, + "loss": 0.4658, + "step": 1128 + }, + { + "epoch": 0.3109628151766365, + "grad_norm": 0.3926720917224884, + "learning_rate": 1e-05, + "loss": 0.4602, + "step": 1130 + }, + { + "epoch": 0.31151319184066595, + "grad_norm": 0.3565811514854431, + "learning_rate": 1e-05, + "loss": 0.4692, + "step": 1132 + }, + { + "epoch": 0.3120635685046954, + "grad_norm": 0.38179391622543335, + "learning_rate": 1e-05, + "loss": 0.4581, + "step": 1134 + }, + { + "epoch": 0.31261394516872487, + "grad_norm": 0.3732840418815613, + "learning_rate": 1e-05, + "loss": 0.4628, + "step": 1136 + }, + { + "epoch": 0.3131643218327543, + "grad_norm": 0.3934018313884735, + "learning_rate": 1e-05, + "loss": 0.4634, + "step": 1138 + }, + { + "epoch": 0.31371469849678374, + "grad_norm": 0.3575834035873413, + "learning_rate": 1e-05, + "loss": 0.4507, + "step": 1140 + }, + { + "epoch": 0.3142650751608132, + "grad_norm": 0.3623636066913605, + "learning_rate": 1e-05, + "loss": 0.4547, + "step": 1142 + }, + { + "epoch": 0.3148154518248426, + "grad_norm": 0.3794458508491516, + "learning_rate": 1e-05, + "loss": 0.4661, + "step": 1144 + }, + { + "epoch": 0.31536582848887207, + "grad_norm": 0.3896718919277191, + "learning_rate": 1e-05, + "loss": 0.4646, + "step": 1146 + }, + { + "epoch": 0.31591620515290153, + "grad_norm": 0.3608621060848236, + "learning_rate": 1e-05, + "loss": 0.4522, + "step": 1148 + }, + { + "epoch": 0.31646658181693094, + "grad_norm": 0.37019404768943787, + "learning_rate": 1e-05, + "loss": 0.4548, + "step": 1150 + }, + { + "epoch": 0.3170169584809604, + "grad_norm": 0.37957248091697693, + "learning_rate": 1e-05, + "loss": 0.4554, + "step": 1152 + }, + { + "epoch": 0.31756733514498986, + "grad_norm": 0.3605276048183441, + "learning_rate": 1e-05, + "loss": 0.4679, + "step": 1154 + }, + { + "epoch": 0.3181177118090193, + "grad_norm": 0.37218716740608215, + "learning_rate": 1e-05, + "loss": 0.4686, + "step": 1156 + }, + { + "epoch": 0.3186680884730487, + "grad_norm": 0.37037035822868347, + "learning_rate": 1e-05, + "loss": 0.4898, + "step": 1158 + }, + { + "epoch": 0.3192184651370782, + "grad_norm": 0.3569047749042511, + "learning_rate": 1e-05, + "loss": 0.4619, + "step": 1160 + }, + { + "epoch": 0.31976884180110765, + "grad_norm": 0.3728378117084503, + "learning_rate": 1e-05, + "loss": 0.4544, + "step": 1162 + }, + { + "epoch": 0.32031921846513706, + "grad_norm": 0.35970696806907654, + "learning_rate": 1e-05, + "loss": 0.4704, + "step": 1164 + }, + { + "epoch": 0.3208695951291665, + "grad_norm": 0.36476969718933105, + "learning_rate": 1e-05, + "loss": 0.4605, + "step": 1166 + }, + { + "epoch": 0.321419971793196, + "grad_norm": 0.35015928745269775, + "learning_rate": 1e-05, + "loss": 0.4653, + "step": 1168 + }, + { + "epoch": 0.3219703484572254, + "grad_norm": 0.3600417375564575, + "learning_rate": 1e-05, + "loss": 0.4557, + "step": 1170 + }, + { + "epoch": 0.32252072512125485, + "grad_norm": 0.36994755268096924, + "learning_rate": 1e-05, + "loss": 0.4601, + "step": 1172 + }, + { + "epoch": 0.3230711017852843, + "grad_norm": 0.39908286929130554, + "learning_rate": 1e-05, + "loss": 0.472, + "step": 1174 + }, + { + "epoch": 0.32362147844931377, + "grad_norm": 0.3717789947986603, + "learning_rate": 1e-05, + "loss": 0.4646, + "step": 1176 + }, + { + "epoch": 0.3241718551133432, + "grad_norm": 0.3617453873157501, + "learning_rate": 1e-05, + "loss": 0.4606, + "step": 1178 + }, + { + "epoch": 0.32472223177737264, + "grad_norm": 0.35809728503227234, + "learning_rate": 1e-05, + "loss": 0.4548, + "step": 1180 + }, + { + "epoch": 0.3252726084414021, + "grad_norm": 0.3767383396625519, + "learning_rate": 1e-05, + "loss": 0.4785, + "step": 1182 + }, + { + "epoch": 0.3258229851054315, + "grad_norm": 0.3819461166858673, + "learning_rate": 1e-05, + "loss": 0.4695, + "step": 1184 + }, + { + "epoch": 0.32637336176946097, + "grad_norm": 0.3590524196624756, + "learning_rate": 1e-05, + "loss": 0.468, + "step": 1186 + }, + { + "epoch": 0.32692373843349043, + "grad_norm": 0.37356823682785034, + "learning_rate": 1e-05, + "loss": 0.4628, + "step": 1188 + }, + { + "epoch": 0.3274741150975199, + "grad_norm": 0.39389410614967346, + "learning_rate": 1e-05, + "loss": 0.4686, + "step": 1190 + }, + { + "epoch": 0.3280244917615493, + "grad_norm": 0.36901354789733887, + "learning_rate": 1e-05, + "loss": 0.4623, + "step": 1192 + }, + { + "epoch": 0.32857486842557876, + "grad_norm": 0.35733821988105774, + "learning_rate": 1e-05, + "loss": 0.457, + "step": 1194 + }, + { + "epoch": 0.3291252450896082, + "grad_norm": 0.3803520202636719, + "learning_rate": 1e-05, + "loss": 0.4661, + "step": 1196 + }, + { + "epoch": 0.3296756217536376, + "grad_norm": 0.36812326312065125, + "learning_rate": 1e-05, + "loss": 0.453, + "step": 1198 + }, + { + "epoch": 0.3302259984176671, + "grad_norm": 0.37463024258613586, + "learning_rate": 1e-05, + "loss": 0.4611, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_merge_loss": 0.4038620591163635, + "eval_merge_runtime": 600.528, + "eval_merge_samples_per_second": 56.204, + "eval_merge_steps_per_second": 2.343, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_new_aug_datas_filtered.json_loss": 0.5311903953552246, + "eval_new_aug_datas_filtered.json_runtime": 10.3899, + "eval_new_aug_datas_filtered.json_samples_per_second": 73.822, + "eval_new_aug_datas_filtered.json_steps_per_second": 3.08, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_sharegpt_gpt4.json_loss": 0.7848892211914062, + "eval_sharegpt_gpt4.json_runtime": 31.7548, + "eval_sharegpt_gpt4.json_samples_per_second": 58.605, + "eval_sharegpt_gpt4.json_steps_per_second": 2.456, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_Table_GPT.json_loss": 0.07294219732284546, + "eval_Table_GPT.json_runtime": 25.0251, + "eval_Table_GPT.json_samples_per_second": 83.636, + "eval_Table_GPT.json_steps_per_second": 3.516, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_gpt_4o_200k.json_loss": 0.8128483295440674, + "eval_gpt_4o_200k.json_runtime": 48.5727, + "eval_gpt_4o_200k.json_samples_per_second": 129.311, + "eval_gpt_4o_200k.json_steps_per_second": 5.394, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_multi_turn_datas.json_loss": 0.343874990940094, + "eval_multi_turn_datas.json_runtime": 75.779, + "eval_multi_turn_datas.json_samples_per_second": 52.811, + "eval_multi_turn_datas.json_steps_per_second": 2.204, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_table_python_code_datas.json_loss": 0.2791996896266937, + "eval_table_python_code_datas.json_runtime": 43.1703, + "eval_table_python_code_datas.json_samples_per_second": 50.011, + "eval_table_python_code_datas.json_steps_per_second": 2.085, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_tabular_llm_data.json_loss": 0.11510641872882843, + "eval_tabular_llm_data.json_runtime": 8.5754, + "eval_tabular_llm_data.json_samples_per_second": 28.687, + "eval_tabular_llm_data.json_steps_per_second": 1.283, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_python_code_critic_21k.json_loss": 0.5806341171264648, + "eval_python_code_critic_21k.json_runtime": 3.2355, + "eval_python_code_critic_21k.json_samples_per_second": 184.517, + "eval_python_code_critic_21k.json_steps_per_second": 7.727, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_all_merge_table_dataset.json_loss": 0.0781954750418663, + "eval_all_merge_table_dataset.json_runtime": 23.3576, + "eval_all_merge_table_dataset.json_samples_per_second": 30.483, + "eval_all_merge_table_dataset.json_steps_per_second": 1.284, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_code_feedback_multi_turn.json_loss": 0.5880293846130371, + "eval_code_feedback_multi_turn.json_runtime": 32.5337, + "eval_code_feedback_multi_turn.json_samples_per_second": 67.653, + "eval_code_feedback_multi_turn.json_steps_per_second": 2.828, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_ultrainteract_sft.json_loss": 0.42568570375442505, + "eval_ultrainteract_sft.json_runtime": 8.665, + "eval_ultrainteract_sft.json_samples_per_second": 168.033, + "eval_ultrainteract_sft.json_steps_per_second": 7.04, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_synthetic_text_to_sql.json_loss": 0.10025755316019058, + "eval_synthetic_text_to_sql.json_runtime": 0.127, + "eval_synthetic_text_to_sql.json_samples_per_second": 267.683, + "eval_synthetic_text_to_sql.json_steps_per_second": 15.746, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_sft_react_sql_datas.json_loss": 0.6435717344284058, + "eval_sft_react_sql_datas.json_runtime": 7.8854, + "eval_sft_react_sql_datas.json_samples_per_second": 39.82, + "eval_sft_react_sql_datas.json_steps_per_second": 1.775, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_all_merge_code.json_loss": 0.29655295610427856, + "eval_all_merge_code.json_runtime": 0.3333, + "eval_all_merge_code.json_samples_per_second": 189.039, + "eval_all_merge_code.json_steps_per_second": 9.002, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_magpie_datas.json_loss": 0.4353857934474945, + "eval_magpie_datas.json_runtime": 2.22, + "eval_magpie_datas.json_samples_per_second": 77.478, + "eval_magpie_datas.json_steps_per_second": 3.604, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_train_data_for_qwen.json_loss": 0.0036680654156953096, + "eval_train_data_for_qwen.json_runtime": 0.2448, + "eval_train_data_for_qwen.json_samples_per_second": 40.856, + "eval_train_data_for_qwen.json_steps_per_second": 4.086, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_alpaca_cleaned.json_loss": 0.9278478622436523, + "eval_alpaca_cleaned.json_runtime": 0.1139, + "eval_alpaca_cleaned.json_samples_per_second": 237.139, + "eval_alpaca_cleaned.json_steps_per_second": 17.566, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_agent_instruct.json_loss": 0.22283704578876495, + "eval_agent_instruct.json_runtime": 0.5129, + "eval_agent_instruct.json_samples_per_second": 93.582, + "eval_agent_instruct.json_steps_per_second": 3.899, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_MathInstruct.json_loss": 0.20810073614120483, + "eval_MathInstruct.json_runtime": 0.3587, + "eval_MathInstruct.json_samples_per_second": 158.905, + "eval_MathInstruct.json_steps_per_second": 8.363, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_tested_143k_python_alpaca.json_loss": 0.44691047072410583, + "eval_tested_143k_python_alpaca.json_runtime": 0.3024, + "eval_tested_143k_python_alpaca.json_samples_per_second": 112.419, + "eval_tested_143k_python_alpaca.json_steps_per_second": 6.613, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_xlam_function_calling_60k.json_loss": 0.009029570966959, + "eval_xlam_function_calling_60k.json_runtime": 0.1005, + "eval_xlam_function_calling_60k.json_samples_per_second": 228.948, + "eval_xlam_function_calling_60k.json_steps_per_second": 9.954, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_alpaca_data_gpt4_chinese.json_loss": 1.5715256929397583, + "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0514, + "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 311.088, + "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.443, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_alpaca_gpt4_zh.json_loss": 0.9568694233894348, + "eval_alpaca_gpt4_zh.json_runtime": 0.0501, + "eval_alpaca_gpt4_zh.json_samples_per_second": 219.517, + "eval_alpaca_gpt4_zh.json_steps_per_second": 19.956, + "step": 1200 + }, + { + "epoch": 0.3302259984176671, + "eval_codefeedback_filtered_instruction.json_loss": 0.5982481837272644, + "eval_codefeedback_filtered_instruction.json_runtime": 0.487, + "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.068, + "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.053, + "step": 1200 + }, + { + "epoch": 0.33077637508169655, + "grad_norm": 0.3862474262714386, + "learning_rate": 1e-05, + "loss": 0.467, + "step": 1202 + }, + { + "epoch": 0.33132675174572596, + "grad_norm": 0.3586987555027008, + "learning_rate": 1e-05, + "loss": 0.4586, + "step": 1204 + }, + { + "epoch": 0.3318771284097554, + "grad_norm": 0.36768838763237, + "learning_rate": 1e-05, + "loss": 0.4658, + "step": 1206 + }, + { + "epoch": 0.3324275050737849, + "grad_norm": 0.36789608001708984, + "learning_rate": 1e-05, + "loss": 0.4479, + "step": 1208 + }, + { + "epoch": 0.33297788173781434, + "grad_norm": 0.3875747323036194, + "learning_rate": 1e-05, + "loss": 0.4651, + "step": 1210 + }, + { + "epoch": 0.33352825840184375, + "grad_norm": 0.37122058868408203, + "learning_rate": 1e-05, + "loss": 0.4474, + "step": 1212 + }, + { + "epoch": 0.3340786350658732, + "grad_norm": 0.3785482347011566, + "learning_rate": 1e-05, + "loss": 0.4573, + "step": 1214 + }, + { + "epoch": 0.33462901172990267, + "grad_norm": 0.3795594871044159, + "learning_rate": 1e-05, + "loss": 0.4633, + "step": 1216 + }, + { + "epoch": 0.3351793883939321, + "grad_norm": 0.35303714871406555, + "learning_rate": 1e-05, + "loss": 0.4701, + "step": 1218 + }, + { + "epoch": 0.33572976505796154, + "grad_norm": 0.3473946154117584, + "learning_rate": 1e-05, + "loss": 0.4565, + "step": 1220 + }, + { + "epoch": 0.336280141721991, + "grad_norm": 0.36495375633239746, + "learning_rate": 1e-05, + "loss": 0.4528, + "step": 1222 + }, + { + "epoch": 0.3368305183860204, + "grad_norm": 0.3617894649505615, + "learning_rate": 1e-05, + "loss": 0.4756, + "step": 1224 + }, + { + "epoch": 0.33738089505004987, + "grad_norm": 0.36371487379074097, + "learning_rate": 1e-05, + "loss": 0.4606, + "step": 1226 + }, + { + "epoch": 0.33793127171407933, + "grad_norm": 0.39192309975624084, + "learning_rate": 1e-05, + "loss": 0.4435, + "step": 1228 + }, + { + "epoch": 0.3384816483781088, + "grad_norm": 0.3902663588523865, + "learning_rate": 1e-05, + "loss": 0.4699, + "step": 1230 + }, + { + "epoch": 0.3390320250421382, + "grad_norm": 0.3662269115447998, + "learning_rate": 1e-05, + "loss": 0.4627, + "step": 1232 + }, + { + "epoch": 0.33958240170616766, + "grad_norm": 0.3659150004386902, + "learning_rate": 1e-05, + "loss": 0.4663, + "step": 1234 + }, + { + "epoch": 0.3401327783701971, + "grad_norm": 0.3632274568080902, + "learning_rate": 1e-05, + "loss": 0.4499, + "step": 1236 + }, + { + "epoch": 0.3406831550342265, + "grad_norm": 0.38413625955581665, + "learning_rate": 1e-05, + "loss": 0.4516, + "step": 1238 + }, + { + "epoch": 0.341233531698256, + "grad_norm": 0.35747644305229187, + "learning_rate": 1e-05, + "loss": 0.4718, + "step": 1240 + }, + { + "epoch": 0.34178390836228545, + "grad_norm": 0.36938604712486267, + "learning_rate": 1e-05, + "loss": 0.4568, + "step": 1242 + }, + { + "epoch": 0.3423342850263149, + "grad_norm": 0.38448217511177063, + "learning_rate": 1e-05, + "loss": 0.474, + "step": 1244 + }, + { + "epoch": 0.3428846616903443, + "grad_norm": 0.3694998323917389, + "learning_rate": 1e-05, + "loss": 0.4516, + "step": 1246 + }, + { + "epoch": 0.3434350383543738, + "grad_norm": 0.41237321496009827, + "learning_rate": 1e-05, + "loss": 0.4569, + "step": 1248 + }, + { + "epoch": 0.34398541501840324, + "grad_norm": 0.4058983325958252, + "learning_rate": 1e-05, + "loss": 0.4657, + "step": 1250 + }, + { + "epoch": 0.34453579168243265, + "grad_norm": 0.3610474467277527, + "learning_rate": 1e-05, + "loss": 0.4587, + "step": 1252 + }, + { + "epoch": 0.3450861683464621, + "grad_norm": 0.3664454221725464, + "learning_rate": 1e-05, + "loss": 0.4656, + "step": 1254 + }, + { + "epoch": 0.34563654501049157, + "grad_norm": 0.35148540139198303, + "learning_rate": 1e-05, + "loss": 0.4471, + "step": 1256 + }, + { + "epoch": 0.346186921674521, + "grad_norm": 0.35331565141677856, + "learning_rate": 1e-05, + "loss": 0.4674, + "step": 1258 + }, + { + "epoch": 0.34673729833855044, + "grad_norm": 0.35367992520332336, + "learning_rate": 1e-05, + "loss": 0.4572, + "step": 1260 + }, + { + "epoch": 0.3472876750025799, + "grad_norm": 0.36106035113334656, + "learning_rate": 1e-05, + "loss": 0.466, + "step": 1262 + }, + { + "epoch": 0.34783805166660936, + "grad_norm": 0.36034414172172546, + "learning_rate": 1e-05, + "loss": 0.4412, + "step": 1264 + }, + { + "epoch": 0.34838842833063877, + "grad_norm": 0.3532898426055908, + "learning_rate": 1e-05, + "loss": 0.4573, + "step": 1266 + }, + { + "epoch": 0.34893880499466823, + "grad_norm": 0.35383620858192444, + "learning_rate": 1e-05, + "loss": 0.4644, + "step": 1268 + }, + { + "epoch": 0.3494891816586977, + "grad_norm": 0.3757399022579193, + "learning_rate": 1e-05, + "loss": 0.4548, + "step": 1270 + }, + { + "epoch": 0.3500395583227271, + "grad_norm": 0.35997340083122253, + "learning_rate": 1e-05, + "loss": 0.4664, + "step": 1272 + }, + { + "epoch": 0.35058993498675656, + "grad_norm": 0.3761090636253357, + "learning_rate": 1e-05, + "loss": 0.4601, + "step": 1274 + }, + { + "epoch": 0.351140311650786, + "grad_norm": 0.33666959404945374, + "learning_rate": 1e-05, + "loss": 0.4596, + "step": 1276 + }, + { + "epoch": 0.3516906883148154, + "grad_norm": 0.36252304911613464, + "learning_rate": 1e-05, + "loss": 0.4688, + "step": 1278 + }, + { + "epoch": 0.3522410649788449, + "grad_norm": 0.3987884819507599, + "learning_rate": 1e-05, + "loss": 0.4444, + "step": 1280 + }, + { + "epoch": 0.35279144164287435, + "grad_norm": 0.35914021730422974, + "learning_rate": 1e-05, + "loss": 0.4508, + "step": 1282 + }, + { + "epoch": 0.3533418183069038, + "grad_norm": 0.36508429050445557, + "learning_rate": 1e-05, + "loss": 0.4597, + "step": 1284 + }, + { + "epoch": 0.3538921949709332, + "grad_norm": 0.3923473060131073, + "learning_rate": 1e-05, + "loss": 0.4594, + "step": 1286 + }, + { + "epoch": 0.3544425716349627, + "grad_norm": 0.38775792717933655, + "learning_rate": 1e-05, + "loss": 0.4573, + "step": 1288 + }, + { + "epoch": 0.35499294829899214, + "grad_norm": 0.4628289043903351, + "learning_rate": 1e-05, + "loss": 0.4732, + "step": 1290 + }, + { + "epoch": 0.35554332496302155, + "grad_norm": 0.35442307591438293, + "learning_rate": 1e-05, + "loss": 0.4621, + "step": 1292 + }, + { + "epoch": 0.356093701627051, + "grad_norm": 0.3809347152709961, + "learning_rate": 1e-05, + "loss": 0.4696, + "step": 1294 + }, + { + "epoch": 0.35664407829108047, + "grad_norm": 0.3683224618434906, + "learning_rate": 1e-05, + "loss": 0.4649, + "step": 1296 + }, + { + "epoch": 0.35719445495510993, + "grad_norm": 0.3792459964752197, + "learning_rate": 1e-05, + "loss": 0.465, + "step": 1298 + }, + { + "epoch": 0.35774483161913934, + "grad_norm": 0.3704141080379486, + "learning_rate": 1e-05, + "loss": 0.4572, + "step": 1300 + }, + { + "epoch": 0.3582952082831688, + "grad_norm": 0.3618161678314209, + "learning_rate": 1e-05, + "loss": 0.4497, + "step": 1302 + }, + { + "epoch": 0.35884558494719826, + "grad_norm": 0.36538904905319214, + "learning_rate": 1e-05, + "loss": 0.4525, + "step": 1304 + }, + { + "epoch": 0.35939596161122767, + "grad_norm": 0.36815035343170166, + "learning_rate": 1e-05, + "loss": 0.4767, + "step": 1306 + }, + { + "epoch": 0.3599463382752571, + "grad_norm": 0.39006996154785156, + "learning_rate": 1e-05, + "loss": 0.4809, + "step": 1308 + }, + { + "epoch": 0.3604967149392866, + "grad_norm": 0.3829619288444519, + "learning_rate": 1e-05, + "loss": 0.4714, + "step": 1310 + }, + { + "epoch": 0.361047091603316, + "grad_norm": 0.37935730814933777, + "learning_rate": 1e-05, + "loss": 0.4518, + "step": 1312 + }, + { + "epoch": 0.36159746826734546, + "grad_norm": 0.371320903301239, + "learning_rate": 1e-05, + "loss": 0.4437, + "step": 1314 + }, + { + "epoch": 0.3621478449313749, + "grad_norm": 0.35784757137298584, + "learning_rate": 1e-05, + "loss": 0.4579, + "step": 1316 + }, + { + "epoch": 0.3626982215954044, + "grad_norm": 0.36308974027633667, + "learning_rate": 1e-05, + "loss": 0.4561, + "step": 1318 + }, + { + "epoch": 0.3632485982594338, + "grad_norm": 0.3538898825645447, + "learning_rate": 1e-05, + "loss": 0.4574, + "step": 1320 + }, + { + "epoch": 0.36379897492346325, + "grad_norm": 0.3715920150279999, + "learning_rate": 1e-05, + "loss": 0.4649, + "step": 1322 + }, + { + "epoch": 0.3643493515874927, + "grad_norm": 0.3698347806930542, + "learning_rate": 1e-05, + "loss": 0.4648, + "step": 1324 + }, + { + "epoch": 0.3648997282515221, + "grad_norm": 0.3725499212741852, + "learning_rate": 1e-05, + "loss": 0.4669, + "step": 1326 + }, + { + "epoch": 0.3654501049155516, + "grad_norm": 0.37399542331695557, + "learning_rate": 1e-05, + "loss": 0.4615, + "step": 1328 + }, + { + "epoch": 0.36600048157958104, + "grad_norm": 0.35364219546318054, + "learning_rate": 1e-05, + "loss": 0.4573, + "step": 1330 + }, + { + "epoch": 0.36655085824361044, + "grad_norm": 0.3651660084724426, + "learning_rate": 1e-05, + "loss": 0.4485, + "step": 1332 + }, + { + "epoch": 0.3671012349076399, + "grad_norm": 0.3659324645996094, + "learning_rate": 1e-05, + "loss": 0.4492, + "step": 1334 + }, + { + "epoch": 0.36765161157166937, + "grad_norm": 0.35941600799560547, + "learning_rate": 1e-05, + "loss": 0.4727, + "step": 1336 + }, + { + "epoch": 0.36820198823569883, + "grad_norm": 0.35083696246147156, + "learning_rate": 1e-05, + "loss": 0.4453, + "step": 1338 + }, + { + "epoch": 0.36875236489972824, + "grad_norm": 0.3690749406814575, + "learning_rate": 1e-05, + "loss": 0.4582, + "step": 1340 + }, + { + "epoch": 0.3693027415637577, + "grad_norm": 0.3743647038936615, + "learning_rate": 1e-05, + "loss": 0.4383, + "step": 1342 + }, + { + "epoch": 0.36985311822778716, + "grad_norm": 0.37354332208633423, + "learning_rate": 1e-05, + "loss": 0.466, + "step": 1344 + }, + { + "epoch": 0.37040349489181656, + "grad_norm": 0.3735334575176239, + "learning_rate": 1e-05, + "loss": 0.4535, + "step": 1346 + }, + { + "epoch": 0.370953871555846, + "grad_norm": 0.37339311838150024, + "learning_rate": 1e-05, + "loss": 0.454, + "step": 1348 + }, + { + "epoch": 0.3715042482198755, + "grad_norm": 0.35196128487586975, + "learning_rate": 1e-05, + "loss": 0.4685, + "step": 1350 + }, + { + "epoch": 0.37205462488390495, + "grad_norm": 0.4031345546245575, + "learning_rate": 1e-05, + "loss": 0.4689, + "step": 1352 + }, + { + "epoch": 0.37260500154793436, + "grad_norm": 0.363320529460907, + "learning_rate": 1e-05, + "loss": 0.459, + "step": 1354 + }, + { + "epoch": 0.3731553782119638, + "grad_norm": 0.36146363615989685, + "learning_rate": 1e-05, + "loss": 0.4446, + "step": 1356 + }, + { + "epoch": 0.3737057548759933, + "grad_norm": 0.36425283551216125, + "learning_rate": 1e-05, + "loss": 0.468, + "step": 1358 + }, + { + "epoch": 0.3742561315400227, + "grad_norm": 0.3795093894004822, + "learning_rate": 1e-05, + "loss": 0.4513, + "step": 1360 + }, + { + "epoch": 0.37480650820405215, + "grad_norm": 0.37901571393013, + "learning_rate": 1e-05, + "loss": 0.464, + "step": 1362 + }, + { + "epoch": 0.3753568848680816, + "grad_norm": 0.3682788014411926, + "learning_rate": 1e-05, + "loss": 0.4535, + "step": 1364 + }, + { + "epoch": 0.375907261532111, + "grad_norm": 0.38756048679351807, + "learning_rate": 1e-05, + "loss": 0.4421, + "step": 1366 + }, + { + "epoch": 0.3764576381961405, + "grad_norm": 0.3859202563762665, + "learning_rate": 1e-05, + "loss": 0.4601, + "step": 1368 + }, + { + "epoch": 0.37700801486016994, + "grad_norm": 0.3959304392337799, + "learning_rate": 1e-05, + "loss": 0.4427, + "step": 1370 + }, + { + "epoch": 0.3775583915241994, + "grad_norm": 0.3768652379512787, + "learning_rate": 1e-05, + "loss": 0.4483, + "step": 1372 + }, + { + "epoch": 0.3781087681882288, + "grad_norm": 0.37339305877685547, + "learning_rate": 1e-05, + "loss": 0.4605, + "step": 1374 + }, + { + "epoch": 0.37865914485225827, + "grad_norm": 0.4036271572113037, + "learning_rate": 1e-05, + "loss": 0.4546, + "step": 1376 + }, + { + "epoch": 0.37920952151628773, + "grad_norm": 0.35173818469047546, + "learning_rate": 1e-05, + "loss": 0.4675, + "step": 1378 + }, + { + "epoch": 0.37975989818031713, + "grad_norm": 0.3682287335395813, + "learning_rate": 1e-05, + "loss": 0.4676, + "step": 1380 + }, + { + "epoch": 0.3803102748443466, + "grad_norm": 0.37660422921180725, + "learning_rate": 1e-05, + "loss": 0.4483, + "step": 1382 + }, + { + "epoch": 0.38086065150837606, + "grad_norm": 0.37428486347198486, + "learning_rate": 1e-05, + "loss": 0.4537, + "step": 1384 + }, + { + "epoch": 0.38141102817240546, + "grad_norm": 0.36140507459640503, + "learning_rate": 1e-05, + "loss": 0.4542, + "step": 1386 + }, + { + "epoch": 0.3819614048364349, + "grad_norm": 0.3818880021572113, + "learning_rate": 1e-05, + "loss": 0.4546, + "step": 1388 + }, + { + "epoch": 0.3825117815004644, + "grad_norm": 0.3840683698654175, + "learning_rate": 1e-05, + "loss": 0.4419, + "step": 1390 + }, + { + "epoch": 0.38306215816449385, + "grad_norm": 0.36933979392051697, + "learning_rate": 1e-05, + "loss": 0.4632, + "step": 1392 + }, + { + "epoch": 0.38361253482852326, + "grad_norm": 0.3724002540111542, + "learning_rate": 1e-05, + "loss": 0.455, + "step": 1394 + }, + { + "epoch": 0.3841629114925527, + "grad_norm": 0.35783514380455017, + "learning_rate": 1e-05, + "loss": 0.4652, + "step": 1396 + }, + { + "epoch": 0.3847132881565822, + "grad_norm": 0.36758366227149963, + "learning_rate": 1e-05, + "loss": 0.4647, + "step": 1398 + }, + { + "epoch": 0.3852636648206116, + "grad_norm": 0.3690735995769501, + "learning_rate": 1e-05, + "loss": 0.4572, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_merge_loss": 0.39922505617141724, + "eval_merge_runtime": 600.7214, + "eval_merge_samples_per_second": 56.186, + "eval_merge_steps_per_second": 2.342, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_new_aug_datas_filtered.json_loss": 0.5239847898483276, + "eval_new_aug_datas_filtered.json_runtime": 10.3569, + "eval_new_aug_datas_filtered.json_samples_per_second": 74.057, + "eval_new_aug_datas_filtered.json_steps_per_second": 3.09, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_sharegpt_gpt4.json_loss": 0.7791606187820435, + "eval_sharegpt_gpt4.json_runtime": 31.6746, + "eval_sharegpt_gpt4.json_samples_per_second": 58.754, + "eval_sharegpt_gpt4.json_steps_per_second": 2.463, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_Table_GPT.json_loss": 0.0626993179321289, + "eval_Table_GPT.json_runtime": 24.9542, + "eval_Table_GPT.json_samples_per_second": 83.874, + "eval_Table_GPT.json_steps_per_second": 3.526, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_gpt_4o_200k.json_loss": 0.8082922101020813, + "eval_gpt_4o_200k.json_runtime": 48.466, + "eval_gpt_4o_200k.json_samples_per_second": 129.596, + "eval_gpt_4o_200k.json_steps_per_second": 5.406, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_multi_turn_datas.json_loss": 0.3381649851799011, + "eval_multi_turn_datas.json_runtime": 75.5711, + "eval_multi_turn_datas.json_samples_per_second": 52.957, + "eval_multi_turn_datas.json_steps_per_second": 2.21, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_table_python_code_datas.json_loss": 0.2752579152584076, + "eval_table_python_code_datas.json_runtime": 43.0439, + "eval_table_python_code_datas.json_samples_per_second": 50.158, + "eval_table_python_code_datas.json_steps_per_second": 2.091, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_tabular_llm_data.json_loss": 0.11023548245429993, + "eval_tabular_llm_data.json_runtime": 8.5291, + "eval_tabular_llm_data.json_samples_per_second": 28.843, + "eval_tabular_llm_data.json_steps_per_second": 1.29, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_python_code_critic_21k.json_loss": 0.5756029486656189, + "eval_python_code_critic_21k.json_runtime": 3.2275, + "eval_python_code_critic_21k.json_samples_per_second": 184.973, + "eval_python_code_critic_21k.json_steps_per_second": 7.746, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_all_merge_table_dataset.json_loss": 0.08007320761680603, + "eval_all_merge_table_dataset.json_runtime": 23.3, + "eval_all_merge_table_dataset.json_samples_per_second": 30.558, + "eval_all_merge_table_dataset.json_steps_per_second": 1.288, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_code_feedback_multi_turn.json_loss": 0.5849318504333496, + "eval_code_feedback_multi_turn.json_runtime": 32.4131, + "eval_code_feedback_multi_turn.json_samples_per_second": 67.905, + "eval_code_feedback_multi_turn.json_steps_per_second": 2.838, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_ultrainteract_sft.json_loss": 0.4235917031764984, + "eval_ultrainteract_sft.json_runtime": 8.6815, + "eval_ultrainteract_sft.json_samples_per_second": 167.713, + "eval_ultrainteract_sft.json_steps_per_second": 7.026, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_synthetic_text_to_sql.json_loss": 0.10058007389307022, + "eval_synthetic_text_to_sql.json_runtime": 0.1256, + "eval_synthetic_text_to_sql.json_samples_per_second": 270.794, + "eval_synthetic_text_to_sql.json_steps_per_second": 15.929, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_sft_react_sql_datas.json_loss": 0.63919597864151, + "eval_sft_react_sql_datas.json_runtime": 7.8177, + "eval_sft_react_sql_datas.json_samples_per_second": 40.165, + "eval_sft_react_sql_datas.json_steps_per_second": 1.791, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_all_merge_code.json_loss": 0.293491929769516, + "eval_all_merge_code.json_runtime": 0.3331, + "eval_all_merge_code.json_samples_per_second": 189.11, + "eval_all_merge_code.json_steps_per_second": 9.005, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_magpie_datas.json_loss": 0.43307721614837646, + "eval_magpie_datas.json_runtime": 2.214, + "eval_magpie_datas.json_samples_per_second": 77.687, + "eval_magpie_datas.json_steps_per_second": 3.613, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_train_data_for_qwen.json_loss": 0.004504092503339052, + "eval_train_data_for_qwen.json_runtime": 0.2448, + "eval_train_data_for_qwen.json_samples_per_second": 40.845, + "eval_train_data_for_qwen.json_steps_per_second": 4.084, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_alpaca_cleaned.json_loss": 0.9073267579078674, + "eval_alpaca_cleaned.json_runtime": 0.1148, + "eval_alpaca_cleaned.json_samples_per_second": 235.221, + "eval_alpaca_cleaned.json_steps_per_second": 17.424, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_agent_instruct.json_loss": 0.22197985649108887, + "eval_agent_instruct.json_runtime": 0.5129, + "eval_agent_instruct.json_samples_per_second": 93.586, + "eval_agent_instruct.json_steps_per_second": 3.899, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_MathInstruct.json_loss": 0.2014550119638443, + "eval_MathInstruct.json_runtime": 0.3655, + "eval_MathInstruct.json_samples_per_second": 155.953, + "eval_MathInstruct.json_steps_per_second": 8.208, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_tested_143k_python_alpaca.json_loss": 0.44645121693611145, + "eval_tested_143k_python_alpaca.json_runtime": 0.3016, + "eval_tested_143k_python_alpaca.json_samples_per_second": 112.74, + "eval_tested_143k_python_alpaca.json_steps_per_second": 6.632, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_xlam_function_calling_60k.json_loss": 0.009633864276111126, + "eval_xlam_function_calling_60k.json_runtime": 0.0999, + "eval_xlam_function_calling_60k.json_samples_per_second": 230.172, + "eval_xlam_function_calling_60k.json_steps_per_second": 10.007, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_alpaca_data_gpt4_chinese.json_loss": 1.5636402368545532, + "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0503, + "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 318.002, + "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.875, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_alpaca_gpt4_zh.json_loss": 0.9602435231208801, + "eval_alpaca_gpt4_zh.json_runtime": 0.0502, + "eval_alpaca_gpt4_zh.json_samples_per_second": 219.067, + "eval_alpaca_gpt4_zh.json_steps_per_second": 19.915, + "step": 1400 + }, + { + "epoch": 0.3852636648206116, + "eval_codefeedback_filtered_instruction.json_loss": 0.5993592143058777, + "eval_codefeedback_filtered_instruction.json_runtime": 0.4852, + "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.223, + "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.061, + "step": 1400 + }, + { + "epoch": 0.38581404148464105, + "grad_norm": 0.36705121397972107, + "learning_rate": 1e-05, + "loss": 0.4617, + "step": 1402 + }, + { + "epoch": 0.3863644181486705, + "grad_norm": 0.3653152883052826, + "learning_rate": 1e-05, + "loss": 0.4528, + "step": 1404 + }, + { + "epoch": 0.38691479481269997, + "grad_norm": 0.34426313638687134, + "learning_rate": 1e-05, + "loss": 0.4464, + "step": 1406 + }, + { + "epoch": 0.3874651714767294, + "grad_norm": 0.3493911623954773, + "learning_rate": 1e-05, + "loss": 0.4638, + "step": 1408 + }, + { + "epoch": 0.38801554814075884, + "grad_norm": 0.3841487765312195, + "learning_rate": 1e-05, + "loss": 0.4471, + "step": 1410 + }, + { + "epoch": 0.3885659248047883, + "grad_norm": 0.3770912289619446, + "learning_rate": 1e-05, + "loss": 0.4623, + "step": 1412 + }, + { + "epoch": 0.3891163014688177, + "grad_norm": 0.38141822814941406, + "learning_rate": 1e-05, + "loss": 0.4583, + "step": 1414 + }, + { + "epoch": 0.38966667813284717, + "grad_norm": 0.3774464726448059, + "learning_rate": 1e-05, + "loss": 0.4574, + "step": 1416 + }, + { + "epoch": 0.39021705479687663, + "grad_norm": 0.35681846737861633, + "learning_rate": 1e-05, + "loss": 0.4443, + "step": 1418 + }, + { + "epoch": 0.39076743146090603, + "grad_norm": 0.3700469732284546, + "learning_rate": 1e-05, + "loss": 0.4468, + "step": 1420 + }, + { + "epoch": 0.3913178081249355, + "grad_norm": 0.35229384899139404, + "learning_rate": 1e-05, + "loss": 0.456, + "step": 1422 + }, + { + "epoch": 0.39186818478896496, + "grad_norm": 0.3469116687774658, + "learning_rate": 1e-05, + "loss": 0.451, + "step": 1424 + }, + { + "epoch": 0.3924185614529944, + "grad_norm": 0.36313918232917786, + "learning_rate": 1e-05, + "loss": 0.4679, + "step": 1426 + }, + { + "epoch": 0.3929689381170238, + "grad_norm": 0.3543436527252197, + "learning_rate": 1e-05, + "loss": 0.464, + "step": 1428 + }, + { + "epoch": 0.3935193147810533, + "grad_norm": 0.3992765545845032, + "learning_rate": 1e-05, + "loss": 0.486, + "step": 1430 + }, + { + "epoch": 0.39406969144508275, + "grad_norm": 0.36149340867996216, + "learning_rate": 1e-05, + "loss": 0.4426, + "step": 1432 + }, + { + "epoch": 0.39462006810911215, + "grad_norm": 0.37118762731552124, + "learning_rate": 1e-05, + "loss": 0.4531, + "step": 1434 + }, + { + "epoch": 0.3951704447731416, + "grad_norm": 0.3618330955505371, + "learning_rate": 1e-05, + "loss": 0.4621, + "step": 1436 + }, + { + "epoch": 0.3957208214371711, + "grad_norm": 0.37272128462791443, + "learning_rate": 1e-05, + "loss": 0.4616, + "step": 1438 + }, + { + "epoch": 0.3962711981012005, + "grad_norm": 0.3678719997406006, + "learning_rate": 1e-05, + "loss": 0.4477, + "step": 1440 + }, + { + "epoch": 0.39682157476522995, + "grad_norm": 0.342907190322876, + "learning_rate": 1e-05, + "loss": 0.4484, + "step": 1442 + }, + { + "epoch": 0.3973719514292594, + "grad_norm": 0.3722037374973297, + "learning_rate": 1e-05, + "loss": 0.4576, + "step": 1444 + }, + { + "epoch": 0.39792232809328887, + "grad_norm": 0.3829335868358612, + "learning_rate": 1e-05, + "loss": 0.4568, + "step": 1446 + }, + { + "epoch": 0.3984727047573183, + "grad_norm": 0.36857596039772034, + "learning_rate": 1e-05, + "loss": 0.4509, + "step": 1448 + }, + { + "epoch": 0.39902308142134774, + "grad_norm": 0.36784934997558594, + "learning_rate": 1e-05, + "loss": 0.46, + "step": 1450 + }, + { + "epoch": 0.3995734580853772, + "grad_norm": 0.36996331810951233, + "learning_rate": 1e-05, + "loss": 0.4435, + "step": 1452 + }, + { + "epoch": 0.4001238347494066, + "grad_norm": 0.3608056604862213, + "learning_rate": 1e-05, + "loss": 0.4467, + "step": 1454 + }, + { + "epoch": 0.40067421141343607, + "grad_norm": 0.3827229142189026, + "learning_rate": 1e-05, + "loss": 0.4576, + "step": 1456 + }, + { + "epoch": 0.4012245880774655, + "grad_norm": 0.38073116540908813, + "learning_rate": 1e-05, + "loss": 0.4433, + "step": 1458 + }, + { + "epoch": 0.40177496474149493, + "grad_norm": 0.3861468434333801, + "learning_rate": 1e-05, + "loss": 0.4466, + "step": 1460 + }, + { + "epoch": 0.4023253414055244, + "grad_norm": 0.36093631386756897, + "learning_rate": 1e-05, + "loss": 0.4409, + "step": 1462 + }, + { + "epoch": 0.40287571806955386, + "grad_norm": 0.34549927711486816, + "learning_rate": 1e-05, + "loss": 0.4507, + "step": 1464 + }, + { + "epoch": 0.4034260947335833, + "grad_norm": 0.3782083988189697, + "learning_rate": 1e-05, + "loss": 0.4648, + "step": 1466 + }, + { + "epoch": 0.4039764713976127, + "grad_norm": 0.366914302110672, + "learning_rate": 1e-05, + "loss": 0.462, + "step": 1468 + }, + { + "epoch": 0.4045268480616422, + "grad_norm": 0.3604414761066437, + "learning_rate": 1e-05, + "loss": 0.4639, + "step": 1470 + }, + { + "epoch": 0.40507722472567165, + "grad_norm": 0.3806079924106598, + "learning_rate": 1e-05, + "loss": 0.452, + "step": 1472 + }, + { + "epoch": 0.40562760138970105, + "grad_norm": 0.36079150438308716, + "learning_rate": 1e-05, + "loss": 0.4534, + "step": 1474 + }, + { + "epoch": 0.4061779780537305, + "grad_norm": 0.3526926040649414, + "learning_rate": 1e-05, + "loss": 0.4483, + "step": 1476 + }, + { + "epoch": 0.40672835471776, + "grad_norm": 0.36440181732177734, + "learning_rate": 1e-05, + "loss": 0.4445, + "step": 1478 + }, + { + "epoch": 0.40727873138178944, + "grad_norm": 0.3452344238758087, + "learning_rate": 1e-05, + "loss": 0.4531, + "step": 1480 + }, + { + "epoch": 0.40782910804581884, + "grad_norm": 0.3774935007095337, + "learning_rate": 1e-05, + "loss": 0.4644, + "step": 1482 + }, + { + "epoch": 0.4083794847098483, + "grad_norm": 0.3485760986804962, + "learning_rate": 1e-05, + "loss": 0.4489, + "step": 1484 + }, + { + "epoch": 0.40892986137387777, + "grad_norm": 0.3787960708141327, + "learning_rate": 1e-05, + "loss": 0.4682, + "step": 1486 + }, + { + "epoch": 0.4094802380379072, + "grad_norm": 0.38031846284866333, + "learning_rate": 1e-05, + "loss": 0.462, + "step": 1488 + }, + { + "epoch": 0.41003061470193664, + "grad_norm": 0.3756881654262543, + "learning_rate": 1e-05, + "loss": 0.4514, + "step": 1490 + }, + { + "epoch": 0.4105809913659661, + "grad_norm": 0.3663581311702728, + "learning_rate": 1e-05, + "loss": 0.4482, + "step": 1492 + }, + { + "epoch": 0.4111313680299955, + "grad_norm": 0.35938966274261475, + "learning_rate": 1e-05, + "loss": 0.4471, + "step": 1494 + }, + { + "epoch": 0.41168174469402496, + "grad_norm": 0.3561854064464569, + "learning_rate": 1e-05, + "loss": 0.4514, + "step": 1496 + }, + { + "epoch": 0.4122321213580544, + "grad_norm": 0.36052775382995605, + "learning_rate": 1e-05, + "loss": 0.4564, + "step": 1498 + }, + { + "epoch": 0.4127824980220839, + "grad_norm": 0.3753555119037628, + "learning_rate": 1e-05, + "loss": 0.4543, + "step": 1500 + }, + { + "epoch": 0.4133328746861133, + "grad_norm": 0.3747691810131073, + "learning_rate": 1e-05, + "loss": 0.4588, + "step": 1502 + }, + { + "epoch": 0.41388325135014276, + "grad_norm": 0.3654341399669647, + "learning_rate": 1e-05, + "loss": 0.451, + "step": 1504 + }, + { + "epoch": 0.4144336280141722, + "grad_norm": 0.3624642491340637, + "learning_rate": 1e-05, + "loss": 0.4528, + "step": 1506 + }, + { + "epoch": 0.4149840046782016, + "grad_norm": 0.3465966284275055, + "learning_rate": 1e-05, + "loss": 0.45, + "step": 1508 + }, + { + "epoch": 0.4155343813422311, + "grad_norm": 0.38202422857284546, + "learning_rate": 1e-05, + "loss": 0.4459, + "step": 1510 + }, + { + "epoch": 0.41608475800626055, + "grad_norm": 0.3562781512737274, + "learning_rate": 1e-05, + "loss": 0.4375, + "step": 1512 + }, + { + "epoch": 0.41663513467028995, + "grad_norm": 0.36660805344581604, + "learning_rate": 1e-05, + "loss": 0.4511, + "step": 1514 + }, + { + "epoch": 0.4171855113343194, + "grad_norm": 0.36541464924812317, + "learning_rate": 1e-05, + "loss": 0.4618, + "step": 1516 + }, + { + "epoch": 0.4177358879983489, + "grad_norm": 0.3570851981639862, + "learning_rate": 1e-05, + "loss": 0.4568, + "step": 1518 + }, + { + "epoch": 0.41828626466237834, + "grad_norm": 0.3508870005607605, + "learning_rate": 1e-05, + "loss": 0.4492, + "step": 1520 + }, + { + "epoch": 0.41883664132640774, + "grad_norm": 0.35050973296165466, + "learning_rate": 1e-05, + "loss": 0.4481, + "step": 1522 + }, + { + "epoch": 0.4193870179904372, + "grad_norm": 0.3564668297767639, + "learning_rate": 1e-05, + "loss": 0.4461, + "step": 1524 + }, + { + "epoch": 0.41993739465446667, + "grad_norm": 0.3646043539047241, + "learning_rate": 1e-05, + "loss": 0.4554, + "step": 1526 + }, + { + "epoch": 0.4204877713184961, + "grad_norm": 0.3904356360435486, + "learning_rate": 1e-05, + "loss": 0.4731, + "step": 1528 + }, + { + "epoch": 0.42103814798252553, + "grad_norm": 0.37373483180999756, + "learning_rate": 1e-05, + "loss": 0.4679, + "step": 1530 + }, + { + "epoch": 0.421588524646555, + "grad_norm": 0.3704439699649811, + "learning_rate": 1e-05, + "loss": 0.4706, + "step": 1532 + }, + { + "epoch": 0.42213890131058446, + "grad_norm": 0.37894484400749207, + "learning_rate": 1e-05, + "loss": 0.4515, + "step": 1534 + }, + { + "epoch": 0.42268927797461386, + "grad_norm": 0.3871210217475891, + "learning_rate": 1e-05, + "loss": 0.4477, + "step": 1536 + }, + { + "epoch": 0.4232396546386433, + "grad_norm": 0.3755747079849243, + "learning_rate": 1e-05, + "loss": 0.4633, + "step": 1538 + }, + { + "epoch": 0.4237900313026728, + "grad_norm": 0.359764039516449, + "learning_rate": 1e-05, + "loss": 0.4798, + "step": 1540 + }, + { + "epoch": 0.4243404079667022, + "grad_norm": 0.37172380089759827, + "learning_rate": 1e-05, + "loss": 0.4383, + "step": 1542 + }, + { + "epoch": 0.42489078463073165, + "grad_norm": 0.3501332700252533, + "learning_rate": 1e-05, + "loss": 0.442, + "step": 1544 + }, + { + "epoch": 0.4254411612947611, + "grad_norm": 0.3552211821079254, + "learning_rate": 1e-05, + "loss": 0.4539, + "step": 1546 + }, + { + "epoch": 0.4259915379587905, + "grad_norm": 0.35052230954170227, + "learning_rate": 1e-05, + "loss": 0.428, + "step": 1548 + }, + { + "epoch": 0.42654191462282, + "grad_norm": 0.3710823357105255, + "learning_rate": 1e-05, + "loss": 0.4297, + "step": 1550 + }, + { + "epoch": 0.42709229128684945, + "grad_norm": 0.37135034799575806, + "learning_rate": 1e-05, + "loss": 0.4587, + "step": 1552 + }, + { + "epoch": 0.4276426679508789, + "grad_norm": 0.3729698061943054, + "learning_rate": 1e-05, + "loss": 0.4585, + "step": 1554 + }, + { + "epoch": 0.4281930446149083, + "grad_norm": 0.3525015711784363, + "learning_rate": 1e-05, + "loss": 0.459, + "step": 1556 + }, + { + "epoch": 0.4287434212789378, + "grad_norm": 0.38500455021858215, + "learning_rate": 1e-05, + "loss": 0.4469, + "step": 1558 + }, + { + "epoch": 0.42929379794296724, + "grad_norm": 0.3852159380912781, + "learning_rate": 1e-05, + "loss": 0.4421, + "step": 1560 + }, + { + "epoch": 0.42984417460699664, + "grad_norm": 0.3567640781402588, + "learning_rate": 1e-05, + "loss": 0.4538, + "step": 1562 + }, + { + "epoch": 0.4303945512710261, + "grad_norm": 0.36795344948768616, + "learning_rate": 1e-05, + "loss": 0.4432, + "step": 1564 + }, + { + "epoch": 0.43094492793505557, + "grad_norm": 0.37614256143569946, + "learning_rate": 1e-05, + "loss": 0.4631, + "step": 1566 + }, + { + "epoch": 0.431495304599085, + "grad_norm": 0.356991171836853, + "learning_rate": 1e-05, + "loss": 0.4389, + "step": 1568 + }, + { + "epoch": 0.43204568126311443, + "grad_norm": 0.3793700933456421, + "learning_rate": 1e-05, + "loss": 0.4609, + "step": 1570 + }, + { + "epoch": 0.4325960579271439, + "grad_norm": 0.36675581336021423, + "learning_rate": 1e-05, + "loss": 0.4484, + "step": 1572 + }, + { + "epoch": 0.43314643459117336, + "grad_norm": 0.36404114961624146, + "learning_rate": 1e-05, + "loss": 0.45, + "step": 1574 + }, + { + "epoch": 0.43369681125520276, + "grad_norm": 0.3868160843849182, + "learning_rate": 1e-05, + "loss": 0.4652, + "step": 1576 + }, + { + "epoch": 0.4342471879192322, + "grad_norm": 0.3898649215698242, + "learning_rate": 1e-05, + "loss": 0.4612, + "step": 1578 + }, + { + "epoch": 0.4347975645832617, + "grad_norm": 0.36762335896492004, + "learning_rate": 1e-05, + "loss": 0.4543, + "step": 1580 + }, + { + "epoch": 0.4353479412472911, + "grad_norm": 0.3434213101863861, + "learning_rate": 1e-05, + "loss": 0.4423, + "step": 1582 + }, + { + "epoch": 0.43589831791132055, + "grad_norm": 0.3741122782230377, + "learning_rate": 1e-05, + "loss": 0.4638, + "step": 1584 + }, + { + "epoch": 0.43644869457535, + "grad_norm": 0.38991764187812805, + "learning_rate": 1e-05, + "loss": 0.438, + "step": 1586 + }, + { + "epoch": 0.4369990712393795, + "grad_norm": 0.35284510254859924, + "learning_rate": 1e-05, + "loss": 0.4559, + "step": 1588 + }, + { + "epoch": 0.4375494479034089, + "grad_norm": 0.36775341629981995, + "learning_rate": 1e-05, + "loss": 0.4594, + "step": 1590 + }, + { + "epoch": 0.43809982456743835, + "grad_norm": 0.3677217364311218, + "learning_rate": 1e-05, + "loss": 0.451, + "step": 1592 + }, + { + "epoch": 0.4386502012314678, + "grad_norm": 0.35295674204826355, + "learning_rate": 1e-05, + "loss": 0.4506, + "step": 1594 + }, + { + "epoch": 0.4392005778954972, + "grad_norm": 0.3770224452018738, + "learning_rate": 1e-05, + "loss": 0.4506, + "step": 1596 + }, + { + "epoch": 0.4397509545595267, + "grad_norm": 0.3824670612812042, + "learning_rate": 1e-05, + "loss": 0.4633, + "step": 1598 + }, + { + "epoch": 0.44030133122355614, + "grad_norm": 0.38165828585624695, + "learning_rate": 1e-05, + "loss": 0.4458, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_merge_loss": 0.39449170231819153, + "eval_merge_runtime": 599.3899, + "eval_merge_samples_per_second": 56.311, + "eval_merge_steps_per_second": 2.347, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_new_aug_datas_filtered.json_loss": 0.5198476314544678, + "eval_new_aug_datas_filtered.json_runtime": 10.3548, + "eval_new_aug_datas_filtered.json_samples_per_second": 74.072, + "eval_new_aug_datas_filtered.json_steps_per_second": 3.09, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_sharegpt_gpt4.json_loss": 0.7743993997573853, + "eval_sharegpt_gpt4.json_runtime": 31.7173, + "eval_sharegpt_gpt4.json_samples_per_second": 58.675, + "eval_sharegpt_gpt4.json_steps_per_second": 2.459, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_Table_GPT.json_loss": 0.05817935988306999, + "eval_Table_GPT.json_runtime": 25.0301, + "eval_Table_GPT.json_samples_per_second": 83.619, + "eval_Table_GPT.json_steps_per_second": 3.516, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_gpt_4o_200k.json_loss": 0.8023759126663208, + "eval_gpt_4o_200k.json_runtime": 48.5498, + "eval_gpt_4o_200k.json_samples_per_second": 129.372, + "eval_gpt_4o_200k.json_steps_per_second": 5.397, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_multi_turn_datas.json_loss": 0.3328835964202881, + "eval_multi_turn_datas.json_runtime": 75.669, + "eval_multi_turn_datas.json_samples_per_second": 52.888, + "eval_multi_turn_datas.json_steps_per_second": 2.207, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_table_python_code_datas.json_loss": 0.2713072597980499, + "eval_table_python_code_datas.json_runtime": 43.1148, + "eval_table_python_code_datas.json_samples_per_second": 50.076, + "eval_table_python_code_datas.json_steps_per_second": 2.087, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_tabular_llm_data.json_loss": 0.10233539342880249, + "eval_tabular_llm_data.json_runtime": 8.5788, + "eval_tabular_llm_data.json_samples_per_second": 28.675, + "eval_tabular_llm_data.json_steps_per_second": 1.282, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_python_code_critic_21k.json_loss": 0.5702229142189026, + "eval_python_code_critic_21k.json_runtime": 3.2319, + "eval_python_code_critic_21k.json_samples_per_second": 184.719, + "eval_python_code_critic_21k.json_steps_per_second": 7.735, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_all_merge_table_dataset.json_loss": 0.07606548815965652, + "eval_all_merge_table_dataset.json_runtime": 23.3911, + "eval_all_merge_table_dataset.json_samples_per_second": 30.439, + "eval_all_merge_table_dataset.json_steps_per_second": 1.283, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_code_feedback_multi_turn.json_loss": 0.5824379324913025, + "eval_code_feedback_multi_turn.json_runtime": 32.5207, + "eval_code_feedback_multi_turn.json_samples_per_second": 67.68, + "eval_code_feedback_multi_turn.json_steps_per_second": 2.829, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_ultrainteract_sft.json_loss": 0.42119815945625305, + "eval_ultrainteract_sft.json_runtime": 8.677, + "eval_ultrainteract_sft.json_samples_per_second": 167.801, + "eval_ultrainteract_sft.json_steps_per_second": 7.03, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_synthetic_text_to_sql.json_loss": 0.09474331140518188, + "eval_synthetic_text_to_sql.json_runtime": 0.1262, + "eval_synthetic_text_to_sql.json_samples_per_second": 269.361, + "eval_synthetic_text_to_sql.json_steps_per_second": 15.845, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_sft_react_sql_datas.json_loss": 0.6350359320640564, + "eval_sft_react_sql_datas.json_runtime": 7.869, + "eval_sft_react_sql_datas.json_samples_per_second": 39.903, + "eval_sft_react_sql_datas.json_steps_per_second": 1.779, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_all_merge_code.json_loss": 0.2929154634475708, + "eval_all_merge_code.json_runtime": 0.3373, + "eval_all_merge_code.json_samples_per_second": 186.752, + "eval_all_merge_code.json_steps_per_second": 8.893, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_magpie_datas.json_loss": 0.4318141043186188, + "eval_magpie_datas.json_runtime": 2.2195, + "eval_magpie_datas.json_samples_per_second": 77.496, + "eval_magpie_datas.json_steps_per_second": 3.604, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_train_data_for_qwen.json_loss": 0.00419951044023037, + "eval_train_data_for_qwen.json_runtime": 0.2455, + "eval_train_data_for_qwen.json_samples_per_second": 40.733, + "eval_train_data_for_qwen.json_steps_per_second": 4.073, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_alpaca_cleaned.json_loss": 0.910367488861084, + "eval_alpaca_cleaned.json_runtime": 0.1147, + "eval_alpaca_cleaned.json_samples_per_second": 235.312, + "eval_alpaca_cleaned.json_steps_per_second": 17.431, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_agent_instruct.json_loss": 0.21950356662273407, + "eval_agent_instruct.json_runtime": 0.5156, + "eval_agent_instruct.json_samples_per_second": 93.094, + "eval_agent_instruct.json_steps_per_second": 3.879, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_MathInstruct.json_loss": 0.19855839014053345, + "eval_MathInstruct.json_runtime": 0.3654, + "eval_MathInstruct.json_samples_per_second": 155.99, + "eval_MathInstruct.json_steps_per_second": 8.21, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_tested_143k_python_alpaca.json_loss": 0.4433169662952423, + "eval_tested_143k_python_alpaca.json_runtime": 0.3031, + "eval_tested_143k_python_alpaca.json_samples_per_second": 112.164, + "eval_tested_143k_python_alpaca.json_steps_per_second": 6.598, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_xlam_function_calling_60k.json_loss": 0.008965943939983845, + "eval_xlam_function_calling_60k.json_runtime": 0.1008, + "eval_xlam_function_calling_60k.json_samples_per_second": 228.26, + "eval_xlam_function_calling_60k.json_steps_per_second": 9.924, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_alpaca_data_gpt4_chinese.json_loss": 1.560943603515625, + "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0505, + "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 316.662, + "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.791, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_alpaca_gpt4_zh.json_loss": 0.9813264012336731, + "eval_alpaca_gpt4_zh.json_runtime": 0.05, + "eval_alpaca_gpt4_zh.json_samples_per_second": 219.867, + "eval_alpaca_gpt4_zh.json_steps_per_second": 19.988, + "step": 1600 + }, + { + "epoch": 0.44030133122355614, + "eval_codefeedback_filtered_instruction.json_loss": 0.5885769128799438, + "eval_codefeedback_filtered_instruction.json_runtime": 0.4829, + "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.42, + "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.071, + "step": 1600 + }, + { + "epoch": 0.44085170788758554, + "grad_norm": 0.36969345808029175, + "learning_rate": 1e-05, + "loss": 0.4474, + "step": 1602 + }, + { + "epoch": 0.441402084551615, + "grad_norm": 0.3673281967639923, + "learning_rate": 1e-05, + "loss": 0.4566, + "step": 1604 + }, + { + "epoch": 0.44195246121564447, + "grad_norm": 0.3695686459541321, + "learning_rate": 1e-05, + "loss": 0.4602, + "step": 1606 + }, + { + "epoch": 0.4425028378796739, + "grad_norm": 0.3653704822063446, + "learning_rate": 1e-05, + "loss": 0.4489, + "step": 1608 + }, + { + "epoch": 0.44305321454370333, + "grad_norm": 0.37890321016311646, + "learning_rate": 1e-05, + "loss": 0.4588, + "step": 1610 + }, + { + "epoch": 0.4436035912077328, + "grad_norm": 0.34637650847435, + "learning_rate": 1e-05, + "loss": 0.4554, + "step": 1612 + }, + { + "epoch": 0.44415396787176226, + "grad_norm": 0.3733616769313812, + "learning_rate": 1e-05, + "loss": 0.4477, + "step": 1614 + }, + { + "epoch": 0.44470434453579166, + "grad_norm": 0.3740238547325134, + "learning_rate": 1e-05, + "loss": 0.4528, + "step": 1616 + }, + { + "epoch": 0.4452547211998211, + "grad_norm": 0.35610541701316833, + "learning_rate": 1e-05, + "loss": 0.4487, + "step": 1618 + }, + { + "epoch": 0.4458050978638506, + "grad_norm": 0.362763911485672, + "learning_rate": 1e-05, + "loss": 0.4619, + "step": 1620 + }, + { + "epoch": 0.44635547452788, + "grad_norm": 0.3781318962574005, + "learning_rate": 1e-05, + "loss": 0.4481, + "step": 1622 + }, + { + "epoch": 0.44690585119190945, + "grad_norm": 0.40836694836616516, + "learning_rate": 1e-05, + "loss": 0.4597, + "step": 1624 + }, + { + "epoch": 0.4474562278559389, + "grad_norm": 0.3662070035934448, + "learning_rate": 1e-05, + "loss": 0.4466, + "step": 1626 + }, + { + "epoch": 0.4480066045199684, + "grad_norm": 0.37797635793685913, + "learning_rate": 1e-05, + "loss": 0.4589, + "step": 1628 + }, + { + "epoch": 0.4485569811839978, + "grad_norm": 0.3544275462627411, + "learning_rate": 1e-05, + "loss": 0.4549, + "step": 1630 + }, + { + "epoch": 0.44910735784802724, + "grad_norm": 0.36321336030960083, + "learning_rate": 1e-05, + "loss": 0.443, + "step": 1632 + }, + { + "epoch": 0.4496577345120567, + "grad_norm": 0.45478886365890503, + "learning_rate": 1e-05, + "loss": 0.4343, + "step": 1634 + }, + { + "epoch": 0.4502081111760861, + "grad_norm": 0.3670060336589813, + "learning_rate": 1e-05, + "loss": 0.4463, + "step": 1636 + }, + { + "epoch": 0.4507584878401156, + "grad_norm": 0.381145715713501, + "learning_rate": 1e-05, + "loss": 0.4512, + "step": 1638 + }, + { + "epoch": 0.45130886450414504, + "grad_norm": 0.3729204833507538, + "learning_rate": 1e-05, + "loss": 0.451, + "step": 1640 + }, + { + "epoch": 0.4518592411681745, + "grad_norm": 0.36986637115478516, + "learning_rate": 1e-05, + "loss": 0.4622, + "step": 1642 + }, + { + "epoch": 0.4524096178322039, + "grad_norm": 0.37230783700942993, + "learning_rate": 1e-05, + "loss": 0.4377, + "step": 1644 + }, + { + "epoch": 0.45295999449623336, + "grad_norm": 0.3671816885471344, + "learning_rate": 1e-05, + "loss": 0.4433, + "step": 1646 + }, + { + "epoch": 0.4535103711602628, + "grad_norm": 0.359372615814209, + "learning_rate": 1e-05, + "loss": 0.4512, + "step": 1648 + }, + { + "epoch": 0.45406074782429223, + "grad_norm": 0.3682217001914978, + "learning_rate": 1e-05, + "loss": 0.4478, + "step": 1650 + }, + { + "epoch": 0.4546111244883217, + "grad_norm": 0.3779531419277191, + "learning_rate": 1e-05, + "loss": 0.4446, + "step": 1652 + }, + { + "epoch": 0.45516150115235116, + "grad_norm": 0.3579237759113312, + "learning_rate": 1e-05, + "loss": 0.4432, + "step": 1654 + }, + { + "epoch": 0.45571187781638056, + "grad_norm": 0.35086673498153687, + "learning_rate": 1e-05, + "loss": 0.4511, + "step": 1656 + }, + { + "epoch": 0.45626225448041, + "grad_norm": 0.36263635754585266, + "learning_rate": 1e-05, + "loss": 0.4552, + "step": 1658 + }, + { + "epoch": 0.4568126311444395, + "grad_norm": 0.3715769648551941, + "learning_rate": 1e-05, + "loss": 0.4549, + "step": 1660 + }, + { + "epoch": 0.45736300780846895, + "grad_norm": 0.36989322304725647, + "learning_rate": 1e-05, + "loss": 0.4468, + "step": 1662 + }, + { + "epoch": 0.45791338447249835, + "grad_norm": 0.35716795921325684, + "learning_rate": 1e-05, + "loss": 0.4506, + "step": 1664 + }, + { + "epoch": 0.4584637611365278, + "grad_norm": 0.36870133876800537, + "learning_rate": 1e-05, + "loss": 0.4581, + "step": 1666 + }, + { + "epoch": 0.4590141378005573, + "grad_norm": 0.36808547377586365, + "learning_rate": 1e-05, + "loss": 0.4518, + "step": 1668 + }, + { + "epoch": 0.4595645144645867, + "grad_norm": 0.3777028024196625, + "learning_rate": 1e-05, + "loss": 0.4526, + "step": 1670 + }, + { + "epoch": 0.46011489112861614, + "grad_norm": 0.3849789798259735, + "learning_rate": 1e-05, + "loss": 0.452, + "step": 1672 + }, + { + "epoch": 0.4606652677926456, + "grad_norm": 0.38168811798095703, + "learning_rate": 1e-05, + "loss": 0.4408, + "step": 1674 + }, + { + "epoch": 0.461215644456675, + "grad_norm": 0.3601077198982239, + "learning_rate": 1e-05, + "loss": 0.4415, + "step": 1676 + }, + { + "epoch": 0.4617660211207045, + "grad_norm": 0.3658849596977234, + "learning_rate": 1e-05, + "loss": 0.4461, + "step": 1678 + }, + { + "epoch": 0.46231639778473393, + "grad_norm": 0.3822179138660431, + "learning_rate": 1e-05, + "loss": 0.4585, + "step": 1680 + }, + { + "epoch": 0.4628667744487634, + "grad_norm": 0.38321495056152344, + "learning_rate": 1e-05, + "loss": 0.4469, + "step": 1682 + }, + { + "epoch": 0.4634171511127928, + "grad_norm": 0.3911297917366028, + "learning_rate": 1e-05, + "loss": 0.4522, + "step": 1684 + }, + { + "epoch": 0.46396752777682226, + "grad_norm": 0.38053110241889954, + "learning_rate": 1e-05, + "loss": 0.4487, + "step": 1686 + }, + { + "epoch": 0.4645179044408517, + "grad_norm": 0.3704802691936493, + "learning_rate": 1e-05, + "loss": 0.4436, + "step": 1688 + }, + { + "epoch": 0.46506828110488113, + "grad_norm": 0.3804566562175751, + "learning_rate": 1e-05, + "loss": 0.4419, + "step": 1690 + }, + { + "epoch": 0.4656186577689106, + "grad_norm": 0.3807014524936676, + "learning_rate": 1e-05, + "loss": 0.4526, + "step": 1692 + }, + { + "epoch": 0.46616903443294005, + "grad_norm": 0.3678591549396515, + "learning_rate": 1e-05, + "loss": 0.4579, + "step": 1694 + }, + { + "epoch": 0.46671941109696946, + "grad_norm": 0.37586984038352966, + "learning_rate": 1e-05, + "loss": 0.4404, + "step": 1696 + }, + { + "epoch": 0.4672697877609989, + "grad_norm": 0.36084264516830444, + "learning_rate": 1e-05, + "loss": 0.4398, + "step": 1698 + }, + { + "epoch": 0.4678201644250284, + "grad_norm": 0.36694666743278503, + "learning_rate": 1e-05, + "loss": 0.4369, + "step": 1700 + }, + { + "epoch": 0.46837054108905785, + "grad_norm": 0.4061066210269928, + "learning_rate": 1e-05, + "loss": 0.4495, + "step": 1702 + }, + { + "epoch": 0.46892091775308725, + "grad_norm": 0.37329551577568054, + "learning_rate": 1e-05, + "loss": 0.4482, + "step": 1704 + }, + { + "epoch": 0.4694712944171167, + "grad_norm": 0.39072346687316895, + "learning_rate": 1e-05, + "loss": 0.4506, + "step": 1706 + }, + { + "epoch": 0.4700216710811462, + "grad_norm": 0.3565053343772888, + "learning_rate": 1e-05, + "loss": 0.447, + "step": 1708 + }, + { + "epoch": 0.4705720477451756, + "grad_norm": 0.39754360914230347, + "learning_rate": 1e-05, + "loss": 0.4468, + "step": 1710 + }, + { + "epoch": 0.47112242440920504, + "grad_norm": 0.34416159987449646, + "learning_rate": 1e-05, + "loss": 0.4509, + "step": 1712 + }, + { + "epoch": 0.4716728010732345, + "grad_norm": 0.3646188974380493, + "learning_rate": 1e-05, + "loss": 0.4436, + "step": 1714 + }, + { + "epoch": 0.47222317773726397, + "grad_norm": 0.372549831867218, + "learning_rate": 1e-05, + "loss": 0.4622, + "step": 1716 + }, + { + "epoch": 0.47277355440129337, + "grad_norm": 0.34616753458976746, + "learning_rate": 1e-05, + "loss": 0.4513, + "step": 1718 + }, + { + "epoch": 0.47332393106532283, + "grad_norm": 0.39396756887435913, + "learning_rate": 1e-05, + "loss": 0.4464, + "step": 1720 + }, + { + "epoch": 0.4738743077293523, + "grad_norm": 0.3681057095527649, + "learning_rate": 1e-05, + "loss": 0.4514, + "step": 1722 + }, + { + "epoch": 0.4744246843933817, + "grad_norm": 0.38942328095436096, + "learning_rate": 1e-05, + "loss": 0.4603, + "step": 1724 + }, + { + "epoch": 0.47497506105741116, + "grad_norm": 0.380278617143631, + "learning_rate": 1e-05, + "loss": 0.4463, + "step": 1726 + }, + { + "epoch": 0.4755254377214406, + "grad_norm": 0.37930282950401306, + "learning_rate": 1e-05, + "loss": 0.4377, + "step": 1728 + }, + { + "epoch": 0.47607581438547003, + "grad_norm": 0.36719146370887756, + "learning_rate": 1e-05, + "loss": 0.4285, + "step": 1730 + }, + { + "epoch": 0.4766261910494995, + "grad_norm": 0.3802686035633087, + "learning_rate": 1e-05, + "loss": 0.4346, + "step": 1732 + }, + { + "epoch": 0.47717656771352895, + "grad_norm": 0.3655955493450165, + "learning_rate": 1e-05, + "loss": 0.4504, + "step": 1734 + }, + { + "epoch": 0.4777269443775584, + "grad_norm": 0.34403982758522034, + "learning_rate": 1e-05, + "loss": 0.4502, + "step": 1736 + }, + { + "epoch": 0.4782773210415878, + "grad_norm": 0.35954922437667847, + "learning_rate": 1e-05, + "loss": 0.4313, + "step": 1738 + }, + { + "epoch": 0.4788276977056173, + "grad_norm": 0.3489810824394226, + "learning_rate": 1e-05, + "loss": 0.4479, + "step": 1740 + }, + { + "epoch": 0.47937807436964675, + "grad_norm": 0.3789598047733307, + "learning_rate": 1e-05, + "loss": 0.4488, + "step": 1742 + }, + { + "epoch": 0.47992845103367615, + "grad_norm": 0.38226747512817383, + "learning_rate": 1e-05, + "loss": 0.4612, + "step": 1744 + }, + { + "epoch": 0.4804788276977056, + "grad_norm": 0.36648547649383545, + "learning_rate": 1e-05, + "loss": 0.4521, + "step": 1746 + }, + { + "epoch": 0.4810292043617351, + "grad_norm": 0.36434775590896606, + "learning_rate": 1e-05, + "loss": 0.4579, + "step": 1748 + }, + { + "epoch": 0.4815795810257645, + "grad_norm": 0.3805695176124573, + "learning_rate": 1e-05, + "loss": 0.437, + "step": 1750 + }, + { + "epoch": 0.48212995768979394, + "grad_norm": 0.34234747290611267, + "learning_rate": 1e-05, + "loss": 0.4411, + "step": 1752 + }, + { + "epoch": 0.4826803343538234, + "grad_norm": 0.356953501701355, + "learning_rate": 1e-05, + "loss": 0.4563, + "step": 1754 + }, + { + "epoch": 0.48323071101785287, + "grad_norm": 0.35372647643089294, + "learning_rate": 1e-05, + "loss": 0.4506, + "step": 1756 + }, + { + "epoch": 0.48378108768188227, + "grad_norm": 0.3776678442955017, + "learning_rate": 1e-05, + "loss": 0.4517, + "step": 1758 + }, + { + "epoch": 0.48433146434591173, + "grad_norm": 0.336029052734375, + "learning_rate": 1e-05, + "loss": 0.4387, + "step": 1760 + }, + { + "epoch": 0.4848818410099412, + "grad_norm": 0.35482755303382874, + "learning_rate": 1e-05, + "loss": 0.4456, + "step": 1762 + }, + { + "epoch": 0.4854322176739706, + "grad_norm": 0.3713533580303192, + "learning_rate": 1e-05, + "loss": 0.4616, + "step": 1764 + }, + { + "epoch": 0.48598259433800006, + "grad_norm": 0.348069965839386, + "learning_rate": 1e-05, + "loss": 0.4504, + "step": 1766 + }, + { + "epoch": 0.4865329710020295, + "grad_norm": 0.36832061409950256, + "learning_rate": 1e-05, + "loss": 0.45, + "step": 1768 + }, + { + "epoch": 0.487083347666059, + "grad_norm": 0.3665439486503601, + "learning_rate": 1e-05, + "loss": 0.4525, + "step": 1770 + }, + { + "epoch": 0.4876337243300884, + "grad_norm": 0.39572247862815857, + "learning_rate": 1e-05, + "loss": 0.4521, + "step": 1772 + }, + { + "epoch": 0.48818410099411785, + "grad_norm": 0.36583212018013, + "learning_rate": 1e-05, + "loss": 0.4298, + "step": 1774 + }, + { + "epoch": 0.4887344776581473, + "grad_norm": 0.35969898104667664, + "learning_rate": 1e-05, + "loss": 0.4497, + "step": 1776 + }, + { + "epoch": 0.4892848543221767, + "grad_norm": 0.3651510775089264, + "learning_rate": 1e-05, + "loss": 0.4355, + "step": 1778 + }, + { + "epoch": 0.4898352309862062, + "grad_norm": 0.3885847330093384, + "learning_rate": 1e-05, + "loss": 0.4633, + "step": 1780 + }, + { + "epoch": 0.49038560765023564, + "grad_norm": 0.357166588306427, + "learning_rate": 1e-05, + "loss": 0.4512, + "step": 1782 + }, + { + "epoch": 0.49093598431426505, + "grad_norm": 0.34748879075050354, + "learning_rate": 1e-05, + "loss": 0.437, + "step": 1784 + }, + { + "epoch": 0.4914863609782945, + "grad_norm": 0.371999055147171, + "learning_rate": 1e-05, + "loss": 0.4493, + "step": 1786 + }, + { + "epoch": 0.492036737642324, + "grad_norm": 0.3602544665336609, + "learning_rate": 1e-05, + "loss": 0.4413, + "step": 1788 + }, + { + "epoch": 0.49258711430635344, + "grad_norm": 0.38811835646629333, + "learning_rate": 1e-05, + "loss": 0.4406, + "step": 1790 + }, + { + "epoch": 0.49313749097038284, + "grad_norm": 0.366616427898407, + "learning_rate": 1e-05, + "loss": 0.4587, + "step": 1792 + }, + { + "epoch": 0.4936878676344123, + "grad_norm": 0.39588844776153564, + "learning_rate": 1e-05, + "loss": 0.4525, + "step": 1794 + }, + { + "epoch": 0.49423824429844176, + "grad_norm": 0.3641244173049927, + "learning_rate": 1e-05, + "loss": 0.4533, + "step": 1796 + }, + { + "epoch": 0.49478862096247117, + "grad_norm": 0.35738009214401245, + "learning_rate": 1e-05, + "loss": 0.4542, + "step": 1798 + }, + { + "epoch": 0.49533899762650063, + "grad_norm": 0.36343181133270264, + "learning_rate": 1e-05, + "loss": 0.4527, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_merge_loss": 0.3900485932826996, + "eval_merge_runtime": 600.246, + "eval_merge_samples_per_second": 56.23, + "eval_merge_steps_per_second": 2.344, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_new_aug_datas_filtered.json_loss": 0.5161438584327698, + "eval_new_aug_datas_filtered.json_runtime": 10.4655, + "eval_new_aug_datas_filtered.json_samples_per_second": 73.288, + "eval_new_aug_datas_filtered.json_steps_per_second": 3.058, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_sharegpt_gpt4.json_loss": 0.7699668407440186, + "eval_sharegpt_gpt4.json_runtime": 31.6447, + "eval_sharegpt_gpt4.json_samples_per_second": 58.809, + "eval_sharegpt_gpt4.json_steps_per_second": 2.465, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_Table_GPT.json_loss": 0.057397227734327316, + "eval_Table_GPT.json_runtime": 24.974, + "eval_Table_GPT.json_samples_per_second": 83.807, + "eval_Table_GPT.json_steps_per_second": 3.524, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_gpt_4o_200k.json_loss": 0.7959992289543152, + "eval_gpt_4o_200k.json_runtime": 48.4474, + "eval_gpt_4o_200k.json_samples_per_second": 129.646, + "eval_gpt_4o_200k.json_steps_per_second": 5.408, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_multi_turn_datas.json_loss": 0.326607346534729, + "eval_multi_turn_datas.json_runtime": 75.6077, + "eval_multi_turn_datas.json_samples_per_second": 52.931, + "eval_multi_turn_datas.json_steps_per_second": 2.209, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_table_python_code_datas.json_loss": 0.26808008551597595, + "eval_table_python_code_datas.json_runtime": 43.0557, + "eval_table_python_code_datas.json_samples_per_second": 50.144, + "eval_table_python_code_datas.json_steps_per_second": 2.09, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_tabular_llm_data.json_loss": 0.1004142090678215, + "eval_tabular_llm_data.json_runtime": 8.5429, + "eval_tabular_llm_data.json_samples_per_second": 28.796, + "eval_tabular_llm_data.json_steps_per_second": 1.288, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_python_code_critic_21k.json_loss": 0.5654606223106384, + "eval_python_code_critic_21k.json_runtime": 3.2351, + "eval_python_code_critic_21k.json_samples_per_second": 184.538, + "eval_python_code_critic_21k.json_steps_per_second": 7.728, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_all_merge_table_dataset.json_loss": 0.07576768845319748, + "eval_all_merge_table_dataset.json_runtime": 23.2598, + "eval_all_merge_table_dataset.json_samples_per_second": 30.611, + "eval_all_merge_table_dataset.json_steps_per_second": 1.29, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_code_feedback_multi_turn.json_loss": 0.579846203327179, + "eval_code_feedback_multi_turn.json_runtime": 32.4188, + "eval_code_feedback_multi_turn.json_samples_per_second": 67.893, + "eval_code_feedback_multi_turn.json_steps_per_second": 2.838, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_ultrainteract_sft.json_loss": 0.4181068241596222, + "eval_ultrainteract_sft.json_runtime": 8.6461, + "eval_ultrainteract_sft.json_samples_per_second": 168.4, + "eval_ultrainteract_sft.json_steps_per_second": 7.055, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_synthetic_text_to_sql.json_loss": 0.09818249940872192, + "eval_synthetic_text_to_sql.json_runtime": 0.1264, + "eval_synthetic_text_to_sql.json_samples_per_second": 269.092, + "eval_synthetic_text_to_sql.json_steps_per_second": 15.829, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_sft_react_sql_datas.json_loss": 0.6291559338569641, + "eval_sft_react_sql_datas.json_runtime": 7.8451, + "eval_sft_react_sql_datas.json_samples_per_second": 40.025, + "eval_sft_react_sql_datas.json_steps_per_second": 1.785, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_all_merge_code.json_loss": 0.29108163714408875, + "eval_all_merge_code.json_runtime": 0.3447, + "eval_all_merge_code.json_samples_per_second": 182.771, + "eval_all_merge_code.json_steps_per_second": 8.703, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_magpie_datas.json_loss": 0.43020525574684143, + "eval_magpie_datas.json_runtime": 2.2179, + "eval_magpie_datas.json_samples_per_second": 77.551, + "eval_magpie_datas.json_steps_per_second": 3.607, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_train_data_for_qwen.json_loss": 0.0027856978122144938, + "eval_train_data_for_qwen.json_runtime": 0.2444, + "eval_train_data_for_qwen.json_samples_per_second": 40.919, + "eval_train_data_for_qwen.json_steps_per_second": 4.092, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_alpaca_cleaned.json_loss": 0.9129724502563477, + "eval_alpaca_cleaned.json_runtime": 0.1153, + "eval_alpaca_cleaned.json_samples_per_second": 234.093, + "eval_alpaca_cleaned.json_steps_per_second": 17.34, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_agent_instruct.json_loss": 0.22024483978748322, + "eval_agent_instruct.json_runtime": 0.5149, + "eval_agent_instruct.json_samples_per_second": 93.222, + "eval_agent_instruct.json_steps_per_second": 3.884, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_MathInstruct.json_loss": 0.20060402154922485, + "eval_MathInstruct.json_runtime": 0.3648, + "eval_MathInstruct.json_samples_per_second": 156.23, + "eval_MathInstruct.json_steps_per_second": 8.223, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_tested_143k_python_alpaca.json_loss": 0.44536128640174866, + "eval_tested_143k_python_alpaca.json_runtime": 0.3002, + "eval_tested_143k_python_alpaca.json_samples_per_second": 113.24, + "eval_tested_143k_python_alpaca.json_steps_per_second": 6.661, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_xlam_function_calling_60k.json_loss": 0.00967579148709774, + "eval_xlam_function_calling_60k.json_runtime": 0.1002, + "eval_xlam_function_calling_60k.json_samples_per_second": 229.452, + "eval_xlam_function_calling_60k.json_steps_per_second": 9.976, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_alpaca_data_gpt4_chinese.json_loss": 1.5544477701187134, + "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0511, + "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 313.214, + "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.576, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_alpaca_gpt4_zh.json_loss": 0.977000892162323, + "eval_alpaca_gpt4_zh.json_runtime": 0.0508, + "eval_alpaca_gpt4_zh.json_samples_per_second": 216.666, + "eval_alpaca_gpt4_zh.json_steps_per_second": 19.697, + "step": 1800 + }, + { + "epoch": 0.49533899762650063, + "eval_codefeedback_filtered_instruction.json_loss": 0.5895399451255798, + "eval_codefeedback_filtered_instruction.json_runtime": 0.4883, + "eval_codefeedback_filtered_instruction.json_samples_per_second": 40.957, + "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.048, + "step": 1800 + }, + { + "epoch": 0.4958893742905301, + "grad_norm": 0.36430442333221436, + "learning_rate": 1e-05, + "loss": 0.4418, + "step": 1802 + }, + { + "epoch": 0.4964397509545595, + "grad_norm": 0.35012543201446533, + "learning_rate": 1e-05, + "loss": 0.4437, + "step": 1804 + }, + { + "epoch": 0.49699012761858896, + "grad_norm": 0.3726542294025421, + "learning_rate": 1e-05, + "loss": 0.4332, + "step": 1806 + }, + { + "epoch": 0.4975405042826184, + "grad_norm": 0.3564360439777374, + "learning_rate": 1e-05, + "loss": 0.4378, + "step": 1808 + }, + { + "epoch": 0.4980908809466479, + "grad_norm": 0.3730456233024597, + "learning_rate": 1e-05, + "loss": 0.443, + "step": 1810 + }, + { + "epoch": 0.4986412576106773, + "grad_norm": 0.3588622212409973, + "learning_rate": 1e-05, + "loss": 0.4387, + "step": 1812 + }, + { + "epoch": 0.49919163427470675, + "grad_norm": 0.36861783266067505, + "learning_rate": 1e-05, + "loss": 0.4392, + "step": 1814 + }, + { + "epoch": 0.4997420109387362, + "grad_norm": 0.3537515699863434, + "learning_rate": 1e-05, + "loss": 0.4331, + "step": 1816 + }, + { + "epoch": 0.5002923876027656, + "grad_norm": 0.3723071813583374, + "learning_rate": 1e-05, + "loss": 0.4429, + "step": 1818 + }, + { + "epoch": 0.5008427642667951, + "grad_norm": 0.37015634775161743, + "learning_rate": 1e-05, + "loss": 0.4687, + "step": 1820 + }, + { + "epoch": 0.5013931409308245, + "grad_norm": 0.3528953790664673, + "learning_rate": 1e-05, + "loss": 0.4315, + "step": 1822 + }, + { + "epoch": 0.501943517594854, + "grad_norm": 0.357120543718338, + "learning_rate": 1e-05, + "loss": 0.4423, + "step": 1824 + }, + { + "epoch": 0.5024938942588835, + "grad_norm": 0.3655802607536316, + "learning_rate": 1e-05, + "loss": 0.4475, + "step": 1826 + }, + { + "epoch": 0.5030442709229128, + "grad_norm": 0.3676040470600128, + "learning_rate": 1e-05, + "loss": 0.4345, + "step": 1828 + }, + { + "epoch": 0.5035946475869423, + "grad_norm": 0.3427799940109253, + "learning_rate": 1e-05, + "loss": 0.4422, + "step": 1830 + }, + { + "epoch": 0.5041450242509717, + "grad_norm": 0.3482607305049896, + "learning_rate": 1e-05, + "loss": 0.4347, + "step": 1832 + }, + { + "epoch": 0.5046954009150012, + "grad_norm": 0.3690313398838043, + "learning_rate": 1e-05, + "loss": 0.4572, + "step": 1834 + }, + { + "epoch": 0.5052457775790307, + "grad_norm": 0.351601243019104, + "learning_rate": 1e-05, + "loss": 0.4445, + "step": 1836 + }, + { + "epoch": 0.5057961542430601, + "grad_norm": 0.3506658971309662, + "learning_rate": 1e-05, + "loss": 0.4482, + "step": 1838 + }, + { + "epoch": 0.5063465309070896, + "grad_norm": 0.36706456542015076, + "learning_rate": 1e-05, + "loss": 0.4503, + "step": 1840 + }, + { + "epoch": 0.5068969075711189, + "grad_norm": 0.36632585525512695, + "learning_rate": 1e-05, + "loss": 0.4385, + "step": 1842 + }, + { + "epoch": 0.5074472842351484, + "grad_norm": 0.3675621747970581, + "learning_rate": 1e-05, + "loss": 0.4391, + "step": 1844 + }, + { + "epoch": 0.5079976608991779, + "grad_norm": 0.3883734941482544, + "learning_rate": 1e-05, + "loss": 0.4435, + "step": 1846 + }, + { + "epoch": 0.5085480375632073, + "grad_norm": 0.34348422288894653, + "learning_rate": 1e-05, + "loss": 0.4388, + "step": 1848 + }, + { + "epoch": 0.5090984142272368, + "grad_norm": 0.36695536971092224, + "learning_rate": 1e-05, + "loss": 0.4473, + "step": 1850 + }, + { + "epoch": 0.5096487908912662, + "grad_norm": 0.36929944157600403, + "learning_rate": 1e-05, + "loss": 0.4542, + "step": 1852 + }, + { + "epoch": 0.5101991675552957, + "grad_norm": 0.3946716785430908, + "learning_rate": 1e-05, + "loss": 0.4399, + "step": 1854 + }, + { + "epoch": 0.5107495442193251, + "grad_norm": 0.3619132936000824, + "learning_rate": 1e-05, + "loss": 0.4471, + "step": 1856 + }, + { + "epoch": 0.5112999208833545, + "grad_norm": 0.34836745262145996, + "learning_rate": 1e-05, + "loss": 0.4392, + "step": 1858 + }, + { + "epoch": 0.511850297547384, + "grad_norm": 0.37516769766807556, + "learning_rate": 1e-05, + "loss": 0.4579, + "step": 1860 + }, + { + "epoch": 0.5124006742114134, + "grad_norm": 0.35800984501838684, + "learning_rate": 1e-05, + "loss": 0.4479, + "step": 1862 + }, + { + "epoch": 0.5129510508754429, + "grad_norm": 0.3664796054363251, + "learning_rate": 1e-05, + "loss": 0.4556, + "step": 1864 + }, + { + "epoch": 0.5135014275394724, + "grad_norm": 0.3633113503456116, + "learning_rate": 1e-05, + "loss": 0.4405, + "step": 1866 + }, + { + "epoch": 0.5140518042035017, + "grad_norm": 0.3655359447002411, + "learning_rate": 1e-05, + "loss": 0.4486, + "step": 1868 + }, + { + "epoch": 0.5146021808675312, + "grad_norm": 0.36135318875312805, + "learning_rate": 1e-05, + "loss": 0.4473, + "step": 1870 + }, + { + "epoch": 0.5151525575315606, + "grad_norm": 0.4725627601146698, + "learning_rate": 1e-05, + "loss": 0.4579, + "step": 1872 + }, + { + "epoch": 0.5157029341955901, + "grad_norm": 0.37844300270080566, + "learning_rate": 1e-05, + "loss": 0.4502, + "step": 1874 + }, + { + "epoch": 0.5162533108596196, + "grad_norm": 0.35601717233657837, + "learning_rate": 1e-05, + "loss": 0.4392, + "step": 1876 + }, + { + "epoch": 0.516803687523649, + "grad_norm": 0.3960351049900055, + "learning_rate": 1e-05, + "loss": 0.4519, + "step": 1878 + }, + { + "epoch": 0.5173540641876785, + "grad_norm": 0.3775772154331207, + "learning_rate": 1e-05, + "loss": 0.4553, + "step": 1880 + }, + { + "epoch": 0.5179044408517078, + "grad_norm": 0.3815532624721527, + "learning_rate": 1e-05, + "loss": 0.4479, + "step": 1882 + }, + { + "epoch": 0.5184548175157373, + "grad_norm": 0.3661166727542877, + "learning_rate": 1e-05, + "loss": 0.4423, + "step": 1884 + }, + { + "epoch": 0.5190051941797668, + "grad_norm": 0.3378327786922455, + "learning_rate": 1e-05, + "loss": 0.4419, + "step": 1886 + }, + { + "epoch": 0.5195555708437962, + "grad_norm": 0.34638261795043945, + "learning_rate": 1e-05, + "loss": 0.4379, + "step": 1888 + }, + { + "epoch": 0.5201059475078257, + "grad_norm": 0.35764721035957336, + "learning_rate": 1e-05, + "loss": 0.4389, + "step": 1890 + }, + { + "epoch": 0.5206563241718551, + "grad_norm": 0.3674796223640442, + "learning_rate": 1e-05, + "loss": 0.4438, + "step": 1892 + }, + { + "epoch": 0.5212067008358846, + "grad_norm": 0.34744736552238464, + "learning_rate": 1e-05, + "loss": 0.4317, + "step": 1894 + }, + { + "epoch": 0.521757077499914, + "grad_norm": 0.39198940992355347, + "learning_rate": 1e-05, + "loss": 0.4406, + "step": 1896 + }, + { + "epoch": 0.5223074541639434, + "grad_norm": 0.3545363247394562, + "learning_rate": 1e-05, + "loss": 0.4255, + "step": 1898 + }, + { + "epoch": 0.5228578308279729, + "grad_norm": 0.3635193407535553, + "learning_rate": 1e-05, + "loss": 0.4521, + "step": 1900 + }, + { + "epoch": 0.5234082074920023, + "grad_norm": 0.33844560384750366, + "learning_rate": 1e-05, + "loss": 0.4371, + "step": 1902 + }, + { + "epoch": 0.5239585841560318, + "grad_norm": 0.34886521100997925, + "learning_rate": 1e-05, + "loss": 0.4328, + "step": 1904 + }, + { + "epoch": 0.5245089608200613, + "grad_norm": 0.34973517060279846, + "learning_rate": 1e-05, + "loss": 0.4442, + "step": 1906 + }, + { + "epoch": 0.5250593374840907, + "grad_norm": 0.35180777311325073, + "learning_rate": 1e-05, + "loss": 0.4575, + "step": 1908 + }, + { + "epoch": 0.5256097141481201, + "grad_norm": 0.36237335205078125, + "learning_rate": 1e-05, + "loss": 0.4357, + "step": 1910 + }, + { + "epoch": 0.5261600908121495, + "grad_norm": 0.3784085512161255, + "learning_rate": 1e-05, + "loss": 0.4559, + "step": 1912 + }, + { + "epoch": 0.526710467476179, + "grad_norm": 0.3556850254535675, + "learning_rate": 1e-05, + "loss": 0.4563, + "step": 1914 + }, + { + "epoch": 0.5272608441402085, + "grad_norm": 0.3620041310787201, + "learning_rate": 1e-05, + "loss": 0.4458, + "step": 1916 + }, + { + "epoch": 0.5278112208042379, + "grad_norm": 0.3616819679737091, + "learning_rate": 1e-05, + "loss": 0.4304, + "step": 1918 + }, + { + "epoch": 0.5283615974682674, + "grad_norm": 0.3651537597179413, + "learning_rate": 1e-05, + "loss": 0.4463, + "step": 1920 + }, + { + "epoch": 0.5289119741322967, + "grad_norm": 0.3924584686756134, + "learning_rate": 1e-05, + "loss": 0.4418, + "step": 1922 + }, + { + "epoch": 0.5294623507963262, + "grad_norm": 0.353217214345932, + "learning_rate": 1e-05, + "loss": 0.4437, + "step": 1924 + }, + { + "epoch": 0.5300127274603557, + "grad_norm": 0.3897522985935211, + "learning_rate": 1e-05, + "loss": 0.4549, + "step": 1926 + }, + { + "epoch": 0.5305631041243851, + "grad_norm": 0.36462587118148804, + "learning_rate": 1e-05, + "loss": 0.4247, + "step": 1928 + }, + { + "epoch": 0.5311134807884146, + "grad_norm": 0.3874776363372803, + "learning_rate": 1e-05, + "loss": 0.4502, + "step": 1930 + }, + { + "epoch": 0.531663857452444, + "grad_norm": 0.3533260226249695, + "learning_rate": 1e-05, + "loss": 0.4515, + "step": 1932 + }, + { + "epoch": 0.5322142341164735, + "grad_norm": 0.3668268024921417, + "learning_rate": 1e-05, + "loss": 0.4474, + "step": 1934 + }, + { + "epoch": 0.5327646107805029, + "grad_norm": 0.3501083254814148, + "learning_rate": 1e-05, + "loss": 0.4344, + "step": 1936 + }, + { + "epoch": 0.5333149874445323, + "grad_norm": 0.3565337657928467, + "learning_rate": 1e-05, + "loss": 0.4412, + "step": 1938 + }, + { + "epoch": 0.5338653641085618, + "grad_norm": 0.34048742055892944, + "learning_rate": 1e-05, + "loss": 0.4502, + "step": 1940 + }, + { + "epoch": 0.5344157407725912, + "grad_norm": 0.35694393515586853, + "learning_rate": 1e-05, + "loss": 0.4532, + "step": 1942 + }, + { + "epoch": 0.5349661174366207, + "grad_norm": 0.3527338206768036, + "learning_rate": 1e-05, + "loss": 0.4378, + "step": 1944 + }, + { + "epoch": 0.5355164941006502, + "grad_norm": 0.3684084117412567, + "learning_rate": 1e-05, + "loss": 0.4562, + "step": 1946 + }, + { + "epoch": 0.5360668707646796, + "grad_norm": 0.3584345281124115, + "learning_rate": 1e-05, + "loss": 0.4561, + "step": 1948 + }, + { + "epoch": 0.536617247428709, + "grad_norm": 0.35685622692108154, + "learning_rate": 1e-05, + "loss": 0.4532, + "step": 1950 + }, + { + "epoch": 0.5371676240927384, + "grad_norm": 0.36560460925102234, + "learning_rate": 1e-05, + "loss": 0.4529, + "step": 1952 + }, + { + "epoch": 0.5377180007567679, + "grad_norm": 0.36613890528678894, + "learning_rate": 1e-05, + "loss": 0.4536, + "step": 1954 + }, + { + "epoch": 0.5382683774207974, + "grad_norm": 0.3513580858707428, + "learning_rate": 1e-05, + "loss": 0.4496, + "step": 1956 + }, + { + "epoch": 0.5388187540848268, + "grad_norm": 0.38372403383255005, + "learning_rate": 1e-05, + "loss": 0.4506, + "step": 1958 + }, + { + "epoch": 0.5393691307488563, + "grad_norm": 0.35690757632255554, + "learning_rate": 1e-05, + "loss": 0.4371, + "step": 1960 + }, + { + "epoch": 0.5399195074128857, + "grad_norm": 0.36706483364105225, + "learning_rate": 1e-05, + "loss": 0.4292, + "step": 1962 + }, + { + "epoch": 0.5404698840769151, + "grad_norm": 0.35754841566085815, + "learning_rate": 1e-05, + "loss": 0.4543, + "step": 1964 + }, + { + "epoch": 0.5410202607409446, + "grad_norm": 0.35544702410697937, + "learning_rate": 1e-05, + "loss": 0.4522, + "step": 1966 + }, + { + "epoch": 0.541570637404974, + "grad_norm": 0.3689357042312622, + "learning_rate": 1e-05, + "loss": 0.4447, + "step": 1968 + }, + { + "epoch": 0.5421210140690035, + "grad_norm": 0.35911116003990173, + "learning_rate": 1e-05, + "loss": 0.4253, + "step": 1970 + }, + { + "epoch": 0.5426713907330329, + "grad_norm": 0.3458103537559509, + "learning_rate": 1e-05, + "loss": 0.4398, + "step": 1972 + }, + { + "epoch": 0.5432217673970624, + "grad_norm": 0.3606932759284973, + "learning_rate": 1e-05, + "loss": 0.4486, + "step": 1974 + }, + { + "epoch": 0.5437721440610918, + "grad_norm": 0.3759188652038574, + "learning_rate": 1e-05, + "loss": 0.4339, + "step": 1976 + }, + { + "epoch": 0.5443225207251212, + "grad_norm": 0.3803597390651703, + "learning_rate": 1e-05, + "loss": 0.4575, + "step": 1978 + }, + { + "epoch": 0.5448728973891507, + "grad_norm": 0.36220523715019226, + "learning_rate": 1e-05, + "loss": 0.4427, + "step": 1980 + }, + { + "epoch": 0.5454232740531801, + "grad_norm": 0.36756813526153564, + "learning_rate": 1e-05, + "loss": 0.4297, + "step": 1982 + }, + { + "epoch": 0.5459736507172096, + "grad_norm": 0.35930246114730835, + "learning_rate": 1e-05, + "loss": 0.4375, + "step": 1984 + }, + { + "epoch": 0.5465240273812391, + "grad_norm": 0.38998985290527344, + "learning_rate": 1e-05, + "loss": 0.4331, + "step": 1986 + }, + { + "epoch": 0.5470744040452685, + "grad_norm": 0.35975074768066406, + "learning_rate": 1e-05, + "loss": 0.4493, + "step": 1988 + }, + { + "epoch": 0.5476247807092979, + "grad_norm": 0.3618590533733368, + "learning_rate": 1e-05, + "loss": 0.4431, + "step": 1990 + }, + { + "epoch": 0.5481751573733273, + "grad_norm": 0.3768090009689331, + "learning_rate": 1e-05, + "loss": 0.4414, + "step": 1992 + }, + { + "epoch": 0.5487255340373568, + "grad_norm": 0.3526524305343628, + "learning_rate": 1e-05, + "loss": 0.4349, + "step": 1994 + }, + { + "epoch": 0.5492759107013863, + "grad_norm": 0.3426629900932312, + "learning_rate": 1e-05, + "loss": 0.4345, + "step": 1996 + }, + { + "epoch": 0.5498262873654157, + "grad_norm": 0.3500785529613495, + "learning_rate": 1e-05, + "loss": 0.4415, + "step": 1998 + }, + { + "epoch": 0.5503766640294452, + "grad_norm": 0.3602929413318634, + "learning_rate": 1e-05, + "loss": 0.4454, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_merge_loss": 0.3855894207954407, + "eval_merge_runtime": 600.0048, + "eval_merge_samples_per_second": 56.253, + "eval_merge_steps_per_second": 2.345, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_new_aug_datas_filtered.json_loss": 0.5099759697914124, + "eval_new_aug_datas_filtered.json_runtime": 10.3782, + "eval_new_aug_datas_filtered.json_samples_per_second": 73.905, + "eval_new_aug_datas_filtered.json_steps_per_second": 3.083, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_sharegpt_gpt4.json_loss": 0.763576865196228, + "eval_sharegpt_gpt4.json_runtime": 31.7204, + "eval_sharegpt_gpt4.json_samples_per_second": 58.669, + "eval_sharegpt_gpt4.json_steps_per_second": 2.459, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_Table_GPT.json_loss": 0.055675260722637177, + "eval_Table_GPT.json_runtime": 24.9781, + "eval_Table_GPT.json_samples_per_second": 83.793, + "eval_Table_GPT.json_steps_per_second": 3.523, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_gpt_4o_200k.json_loss": 0.7919400334358215, + "eval_gpt_4o_200k.json_runtime": 48.5207, + "eval_gpt_4o_200k.json_samples_per_second": 129.45, + "eval_gpt_4o_200k.json_steps_per_second": 5.4, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_multi_turn_datas.json_loss": 0.321598082780838, + "eval_multi_turn_datas.json_runtime": 75.7401, + "eval_multi_turn_datas.json_samples_per_second": 52.839, + "eval_multi_turn_datas.json_steps_per_second": 2.205, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_table_python_code_datas.json_loss": 0.26337531208992004, + "eval_table_python_code_datas.json_runtime": 43.1695, + "eval_table_python_code_datas.json_samples_per_second": 50.012, + "eval_table_python_code_datas.json_steps_per_second": 2.085, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_tabular_llm_data.json_loss": 0.09393570572137833, + "eval_tabular_llm_data.json_runtime": 8.5822, + "eval_tabular_llm_data.json_samples_per_second": 28.664, + "eval_tabular_llm_data.json_steps_per_second": 1.282, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_python_code_critic_21k.json_loss": 0.5615730285644531, + "eval_python_code_critic_21k.json_runtime": 3.2332, + "eval_python_code_critic_21k.json_samples_per_second": 184.645, + "eval_python_code_critic_21k.json_steps_per_second": 7.732, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_all_merge_table_dataset.json_loss": 0.07384855300188065, + "eval_all_merge_table_dataset.json_runtime": 23.3929, + "eval_all_merge_table_dataset.json_samples_per_second": 30.437, + "eval_all_merge_table_dataset.json_steps_per_second": 1.282, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_code_feedback_multi_turn.json_loss": 0.5769618153572083, + "eval_code_feedback_multi_turn.json_runtime": 32.4541, + "eval_code_feedback_multi_turn.json_samples_per_second": 67.819, + "eval_code_feedback_multi_turn.json_steps_per_second": 2.835, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_ultrainteract_sft.json_loss": 0.41532665491104126, + "eval_ultrainteract_sft.json_runtime": 8.6954, + "eval_ultrainteract_sft.json_samples_per_second": 167.445, + "eval_ultrainteract_sft.json_steps_per_second": 7.015, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_synthetic_text_to_sql.json_loss": 0.09223779290914536, + "eval_synthetic_text_to_sql.json_runtime": 0.1265, + "eval_synthetic_text_to_sql.json_samples_per_second": 268.84, + "eval_synthetic_text_to_sql.json_steps_per_second": 15.814, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_sft_react_sql_datas.json_loss": 0.6254591941833496, + "eval_sft_react_sql_datas.json_runtime": 7.8542, + "eval_sft_react_sql_datas.json_samples_per_second": 39.979, + "eval_sft_react_sql_datas.json_steps_per_second": 1.782, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_all_merge_code.json_loss": 0.2845838665962219, + "eval_all_merge_code.json_runtime": 0.3345, + "eval_all_merge_code.json_samples_per_second": 188.319, + "eval_all_merge_code.json_steps_per_second": 8.968, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_magpie_datas.json_loss": 0.4300972521305084, + "eval_magpie_datas.json_runtime": 2.2166, + "eval_magpie_datas.json_samples_per_second": 77.598, + "eval_magpie_datas.json_steps_per_second": 3.609, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_train_data_for_qwen.json_loss": 0.0036769520957022905, + "eval_train_data_for_qwen.json_runtime": 0.2431, + "eval_train_data_for_qwen.json_samples_per_second": 41.14, + "eval_train_data_for_qwen.json_steps_per_second": 4.114, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_alpaca_cleaned.json_loss": 0.9104709625244141, + "eval_alpaca_cleaned.json_runtime": 0.1148, + "eval_alpaca_cleaned.json_samples_per_second": 235.266, + "eval_alpaca_cleaned.json_steps_per_second": 17.427, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_agent_instruct.json_loss": 0.220087930560112, + "eval_agent_instruct.json_runtime": 0.5143, + "eval_agent_instruct.json_samples_per_second": 93.334, + "eval_agent_instruct.json_steps_per_second": 3.889, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_MathInstruct.json_loss": 0.1989249587059021, + "eval_MathInstruct.json_runtime": 0.3499, + "eval_MathInstruct.json_samples_per_second": 162.904, + "eval_MathInstruct.json_steps_per_second": 8.574, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_tested_143k_python_alpaca.json_loss": 0.4425477683544159, + "eval_tested_143k_python_alpaca.json_runtime": 0.3008, + "eval_tested_143k_python_alpaca.json_samples_per_second": 113.029, + "eval_tested_143k_python_alpaca.json_steps_per_second": 6.649, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_xlam_function_calling_60k.json_loss": 0.008927595801651478, + "eval_xlam_function_calling_60k.json_runtime": 0.1003, + "eval_xlam_function_calling_60k.json_samples_per_second": 229.301, + "eval_xlam_function_calling_60k.json_steps_per_second": 9.97, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_alpaca_data_gpt4_chinese.json_loss": 1.5485728979110718, + "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0512, + "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 312.726, + "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.545, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_alpaca_gpt4_zh.json_loss": 0.9768400192260742, + "eval_alpaca_gpt4_zh.json_runtime": 0.0505, + "eval_alpaca_gpt4_zh.json_samples_per_second": 217.931, + "eval_alpaca_gpt4_zh.json_steps_per_second": 19.812, + "step": 2000 + }, + { + "epoch": 0.5503766640294452, + "eval_codefeedback_filtered_instruction.json_loss": 0.587010383605957, + "eval_codefeedback_filtered_instruction.json_runtime": 0.4876, + "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.015, + "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.051, + "step": 2000 + }, + { + "epoch": 0.5509270406934746, + "grad_norm": 0.3626772463321686, + "learning_rate": 1e-05, + "loss": 0.4442, + "step": 2002 + }, + { + "epoch": 0.551477417357504, + "grad_norm": 0.34878280758857727, + "learning_rate": 1e-05, + "loss": 0.4458, + "step": 2004 + }, + { + "epoch": 0.5520277940215335, + "grad_norm": 0.35377946496009827, + "learning_rate": 1e-05, + "loss": 0.4273, + "step": 2006 + }, + { + "epoch": 0.5525781706855629, + "grad_norm": 0.3649701774120331, + "learning_rate": 1e-05, + "loss": 0.4342, + "step": 2008 + }, + { + "epoch": 0.5531285473495924, + "grad_norm": 0.34736165404319763, + "learning_rate": 1e-05, + "loss": 0.4298, + "step": 2010 + }, + { + "epoch": 0.5536789240136218, + "grad_norm": 0.3697884678840637, + "learning_rate": 1e-05, + "loss": 0.4424, + "step": 2012 + }, + { + "epoch": 0.5542293006776513, + "grad_norm": 0.40290403366088867, + "learning_rate": 1e-05, + "loss": 0.4388, + "step": 2014 + }, + { + "epoch": 0.5547796773416808, + "grad_norm": 0.36797061562538147, + "learning_rate": 1e-05, + "loss": 0.4648, + "step": 2016 + }, + { + "epoch": 0.5553300540057101, + "grad_norm": 0.35621124505996704, + "learning_rate": 1e-05, + "loss": 0.433, + "step": 2018 + }, + { + "epoch": 0.5558804306697396, + "grad_norm": 0.3625437915325165, + "learning_rate": 1e-05, + "loss": 0.441, + "step": 2020 + }, + { + "epoch": 0.556430807333769, + "grad_norm": 0.3642013370990753, + "learning_rate": 1e-05, + "loss": 0.4425, + "step": 2022 + }, + { + "epoch": 0.5569811839977985, + "grad_norm": 0.36053115129470825, + "learning_rate": 1e-05, + "loss": 0.4422, + "step": 2024 + }, + { + "epoch": 0.557531560661828, + "grad_norm": 0.36283549666404724, + "learning_rate": 1e-05, + "loss": 0.4338, + "step": 2026 + }, + { + "epoch": 0.5580819373258574, + "grad_norm": 0.3758421540260315, + "learning_rate": 1e-05, + "loss": 0.439, + "step": 2028 + }, + { + "epoch": 0.5586323139898868, + "grad_norm": 0.33730989694595337, + "learning_rate": 1e-05, + "loss": 0.4446, + "step": 2030 + }, + { + "epoch": 0.5591826906539162, + "grad_norm": 0.36297255754470825, + "learning_rate": 1e-05, + "loss": 0.4358, + "step": 2032 + }, + { + "epoch": 0.5597330673179457, + "grad_norm": 0.3534908890724182, + "learning_rate": 1e-05, + "loss": 0.4257, + "step": 2034 + }, + { + "epoch": 0.5602834439819752, + "grad_norm": 0.3690515160560608, + "learning_rate": 1e-05, + "loss": 0.4383, + "step": 2036 + }, + { + "epoch": 0.5608338206460046, + "grad_norm": 0.3638661503791809, + "learning_rate": 1e-05, + "loss": 0.4452, + "step": 2038 + }, + { + "epoch": 0.5613841973100341, + "grad_norm": 0.3521392047405243, + "learning_rate": 1e-05, + "loss": 0.4342, + "step": 2040 + }, + { + "epoch": 0.5619345739740635, + "grad_norm": 0.3569532632827759, + "learning_rate": 1e-05, + "loss": 0.4507, + "step": 2042 + }, + { + "epoch": 0.5624849506380929, + "grad_norm": 0.37072595953941345, + "learning_rate": 1e-05, + "loss": 0.4354, + "step": 2044 + }, + { + "epoch": 0.5630353273021224, + "grad_norm": 0.38489988446235657, + "learning_rate": 1e-05, + "loss": 0.4528, + "step": 2046 + }, + { + "epoch": 0.5635857039661518, + "grad_norm": 0.38305357098579407, + "learning_rate": 1e-05, + "loss": 0.4428, + "step": 2048 + }, + { + "epoch": 0.5641360806301813, + "grad_norm": 0.3491927981376648, + "learning_rate": 1e-05, + "loss": 0.4242, + "step": 2050 + }, + { + "epoch": 0.5646864572942107, + "grad_norm": 0.35508430004119873, + "learning_rate": 1e-05, + "loss": 0.4556, + "step": 2052 + }, + { + "epoch": 0.5652368339582402, + "grad_norm": 0.36298030614852905, + "learning_rate": 1e-05, + "loss": 0.4337, + "step": 2054 + }, + { + "epoch": 0.5657872106222697, + "grad_norm": 0.3598901629447937, + "learning_rate": 1e-05, + "loss": 0.4378, + "step": 2056 + }, + { + "epoch": 0.566337587286299, + "grad_norm": 0.3838946223258972, + "learning_rate": 1e-05, + "loss": 0.4346, + "step": 2058 + }, + { + "epoch": 0.5668879639503285, + "grad_norm": 0.3986867666244507, + "learning_rate": 1e-05, + "loss": 0.45, + "step": 2060 + }, + { + "epoch": 0.5674383406143579, + "grad_norm": 0.3509708344936371, + "learning_rate": 1e-05, + "loss": 0.4462, + "step": 2062 + }, + { + "epoch": 0.5679887172783874, + "grad_norm": 0.35189950466156006, + "learning_rate": 1e-05, + "loss": 0.4307, + "step": 2064 + }, + { + "epoch": 0.5685390939424169, + "grad_norm": 0.37416207790374756, + "learning_rate": 1e-05, + "loss": 0.4368, + "step": 2066 + }, + { + "epoch": 0.5690894706064463, + "grad_norm": 0.3902382254600525, + "learning_rate": 1e-05, + "loss": 0.4278, + "step": 2068 + }, + { + "epoch": 0.5696398472704758, + "grad_norm": 0.384260892868042, + "learning_rate": 1e-05, + "loss": 0.4449, + "step": 2070 + }, + { + "epoch": 0.5701902239345051, + "grad_norm": 0.367347776889801, + "learning_rate": 1e-05, + "loss": 0.4397, + "step": 2072 + }, + { + "epoch": 0.5707406005985346, + "grad_norm": 0.35011574625968933, + "learning_rate": 1e-05, + "loss": 0.4375, + "step": 2074 + }, + { + "epoch": 0.5712909772625641, + "grad_norm": 0.3609907329082489, + "learning_rate": 1e-05, + "loss": 0.446, + "step": 2076 + }, + { + "epoch": 0.5718413539265935, + "grad_norm": 0.3640425205230713, + "learning_rate": 1e-05, + "loss": 0.4453, + "step": 2078 + }, + { + "epoch": 0.572391730590623, + "grad_norm": 0.3464198112487793, + "learning_rate": 1e-05, + "loss": 0.4489, + "step": 2080 + }, + { + "epoch": 0.5729421072546524, + "grad_norm": 0.3741483688354492, + "learning_rate": 1e-05, + "loss": 0.4515, + "step": 2082 + }, + { + "epoch": 0.5734924839186818, + "grad_norm": 0.37388619780540466, + "learning_rate": 1e-05, + "loss": 0.4632, + "step": 2084 + }, + { + "epoch": 0.5740428605827113, + "grad_norm": 0.37237605452537537, + "learning_rate": 1e-05, + "loss": 0.4425, + "step": 2086 + }, + { + "epoch": 0.5745932372467407, + "grad_norm": 0.35421323776245117, + "learning_rate": 1e-05, + "loss": 0.4474, + "step": 2088 + }, + { + "epoch": 0.5751436139107702, + "grad_norm": 0.33015069365501404, + "learning_rate": 1e-05, + "loss": 0.43, + "step": 2090 + }, + { + "epoch": 0.5756939905747996, + "grad_norm": 0.3670506179332733, + "learning_rate": 1e-05, + "loss": 0.4452, + "step": 2092 + }, + { + "epoch": 0.5762443672388291, + "grad_norm": 0.3514888882637024, + "learning_rate": 1e-05, + "loss": 0.4287, + "step": 2094 + }, + { + "epoch": 0.5767947439028586, + "grad_norm": 0.3714512288570404, + "learning_rate": 1e-05, + "loss": 0.4344, + "step": 2096 + }, + { + "epoch": 0.5773451205668879, + "grad_norm": 0.35363397002220154, + "learning_rate": 1e-05, + "loss": 0.4408, + "step": 2098 + }, + { + "epoch": 0.5778954972309174, + "grad_norm": 0.3529844582080841, + "learning_rate": 1e-05, + "loss": 0.4434, + "step": 2100 + }, + { + "epoch": 0.5784458738949468, + "grad_norm": 0.3400002121925354, + "learning_rate": 1e-05, + "loss": 0.4443, + "step": 2102 + }, + { + "epoch": 0.5789962505589763, + "grad_norm": 0.3620370328426361, + "learning_rate": 1e-05, + "loss": 0.4377, + "step": 2104 + }, + { + "epoch": 0.5795466272230058, + "grad_norm": 0.3476988971233368, + "learning_rate": 1e-05, + "loss": 0.4321, + "step": 2106 + }, + { + "epoch": 0.5800970038870352, + "grad_norm": 0.35739636421203613, + "learning_rate": 1e-05, + "loss": 0.4495, + "step": 2108 + }, + { + "epoch": 0.5806473805510647, + "grad_norm": 0.3718028962612152, + "learning_rate": 1e-05, + "loss": 0.4391, + "step": 2110 + }, + { + "epoch": 0.581197757215094, + "grad_norm": 0.35041627287864685, + "learning_rate": 1e-05, + "loss": 0.454, + "step": 2112 + }, + { + "epoch": 0.5817481338791235, + "grad_norm": 0.36277493834495544, + "learning_rate": 1e-05, + "loss": 0.44, + "step": 2114 + }, + { + "epoch": 0.582298510543153, + "grad_norm": 0.36685582995414734, + "learning_rate": 1e-05, + "loss": 0.4401, + "step": 2116 + }, + { + "epoch": 0.5828488872071824, + "grad_norm": 0.33634135127067566, + "learning_rate": 1e-05, + "loss": 0.4338, + "step": 2118 + }, + { + "epoch": 0.5833992638712119, + "grad_norm": 0.36546674370765686, + "learning_rate": 1e-05, + "loss": 0.4456, + "step": 2120 + }, + { + "epoch": 0.5839496405352413, + "grad_norm": 0.361472487449646, + "learning_rate": 1e-05, + "loss": 0.4368, + "step": 2122 + }, + { + "epoch": 0.5845000171992708, + "grad_norm": 0.36743828654289246, + "learning_rate": 1e-05, + "loss": 0.4464, + "step": 2124 + }, + { + "epoch": 0.5850503938633002, + "grad_norm": 0.35304173827171326, + "learning_rate": 1e-05, + "loss": 0.4407, + "step": 2126 + }, + { + "epoch": 0.5856007705273296, + "grad_norm": 0.35151979327201843, + "learning_rate": 1e-05, + "loss": 0.4532, + "step": 2128 + }, + { + "epoch": 0.5861511471913591, + "grad_norm": 0.34761616587638855, + "learning_rate": 1e-05, + "loss": 0.444, + "step": 2130 + }, + { + "epoch": 0.5867015238553885, + "grad_norm": 0.3763500452041626, + "learning_rate": 1e-05, + "loss": 0.4524, + "step": 2132 + }, + { + "epoch": 0.587251900519418, + "grad_norm": 0.36489951610565186, + "learning_rate": 1e-05, + "loss": 0.4333, + "step": 2134 + }, + { + "epoch": 0.5878022771834475, + "grad_norm": 0.38710853457450867, + "learning_rate": 1e-05, + "loss": 0.4517, + "step": 2136 + }, + { + "epoch": 0.5883526538474768, + "grad_norm": 0.36153027415275574, + "learning_rate": 1e-05, + "loss": 0.438, + "step": 2138 + }, + { + "epoch": 0.5889030305115063, + "grad_norm": 0.3907857835292816, + "learning_rate": 1e-05, + "loss": 0.4429, + "step": 2140 + }, + { + "epoch": 0.5894534071755357, + "grad_norm": 0.3813617527484894, + "learning_rate": 1e-05, + "loss": 0.4392, + "step": 2142 + }, + { + "epoch": 0.5900037838395652, + "grad_norm": 0.3563400208950043, + "learning_rate": 1e-05, + "loss": 0.434, + "step": 2144 + }, + { + "epoch": 0.5905541605035947, + "grad_norm": 0.3556332290172577, + "learning_rate": 1e-05, + "loss": 0.4436, + "step": 2146 + }, + { + "epoch": 0.5911045371676241, + "grad_norm": 0.3623802363872528, + "learning_rate": 1e-05, + "loss": 0.4378, + "step": 2148 + }, + { + "epoch": 0.5916549138316536, + "grad_norm": 0.36329442262649536, + "learning_rate": 1e-05, + "loss": 0.4386, + "step": 2150 + }, + { + "epoch": 0.5922052904956829, + "grad_norm": 0.3771746754646301, + "learning_rate": 1e-05, + "loss": 0.4494, + "step": 2152 + }, + { + "epoch": 0.5927556671597124, + "grad_norm": 0.34596994519233704, + "learning_rate": 1e-05, + "loss": 0.4173, + "step": 2154 + }, + { + "epoch": 0.5933060438237419, + "grad_norm": 0.36507177352905273, + "learning_rate": 1e-05, + "loss": 0.4254, + "step": 2156 + }, + { + "epoch": 0.5938564204877713, + "grad_norm": 0.3519168794155121, + "learning_rate": 1e-05, + "loss": 0.4447, + "step": 2158 + }, + { + "epoch": 0.5944067971518008, + "grad_norm": 0.35316991806030273, + "learning_rate": 1e-05, + "loss": 0.4622, + "step": 2160 + }, + { + "epoch": 0.5949571738158302, + "grad_norm": 0.3529471158981323, + "learning_rate": 1e-05, + "loss": 0.4482, + "step": 2162 + }, + { + "epoch": 0.5955075504798597, + "grad_norm": 0.3722255825996399, + "learning_rate": 1e-05, + "loss": 0.4454, + "step": 2164 + }, + { + "epoch": 0.596057927143889, + "grad_norm": 0.3557456433773041, + "learning_rate": 1e-05, + "loss": 0.4435, + "step": 2166 + }, + { + "epoch": 0.5966083038079185, + "grad_norm": 0.3348141610622406, + "learning_rate": 1e-05, + "loss": 0.436, + "step": 2168 + }, + { + "epoch": 0.597158680471948, + "grad_norm": 0.38193532824516296, + "learning_rate": 1e-05, + "loss": 0.4543, + "step": 2170 + }, + { + "epoch": 0.5977090571359774, + "grad_norm": 0.3672102391719818, + "learning_rate": 1e-05, + "loss": 0.4356, + "step": 2172 + }, + { + "epoch": 0.5982594338000069, + "grad_norm": 0.37538838386535645, + "learning_rate": 1e-05, + "loss": 0.4442, + "step": 2174 + }, + { + "epoch": 0.5988098104640364, + "grad_norm": 0.3512885272502899, + "learning_rate": 1e-05, + "loss": 0.4249, + "step": 2176 + }, + { + "epoch": 0.5993601871280658, + "grad_norm": 0.4028591811656952, + "learning_rate": 1e-05, + "loss": 0.4495, + "step": 2178 + }, + { + "epoch": 0.5999105637920952, + "grad_norm": 0.3539179861545563, + "learning_rate": 1e-05, + "loss": 0.4504, + "step": 2180 + }, + { + "epoch": 0.6004609404561246, + "grad_norm": 0.34848934412002563, + "learning_rate": 1e-05, + "loss": 0.4348, + "step": 2182 + }, + { + "epoch": 0.6010113171201541, + "grad_norm": 0.37469926476478577, + "learning_rate": 1e-05, + "loss": 0.4414, + "step": 2184 + }, + { + "epoch": 0.6015616937841836, + "grad_norm": 0.3511207103729248, + "learning_rate": 1e-05, + "loss": 0.4489, + "step": 2186 + }, + { + "epoch": 0.602112070448213, + "grad_norm": 0.3594874441623688, + "learning_rate": 1e-05, + "loss": 0.4429, + "step": 2188 + }, + { + "epoch": 0.6026624471122425, + "grad_norm": 0.37694159150123596, + "learning_rate": 1e-05, + "loss": 0.4365, + "step": 2190 + }, + { + "epoch": 0.6032128237762718, + "grad_norm": 0.3630627393722534, + "learning_rate": 1e-05, + "loss": 0.4449, + "step": 2192 + }, + { + "epoch": 0.6037632004403013, + "grad_norm": 0.352230042219162, + "learning_rate": 1e-05, + "loss": 0.4382, + "step": 2194 + }, + { + "epoch": 0.6043135771043308, + "grad_norm": 0.369757741689682, + "learning_rate": 1e-05, + "loss": 0.4443, + "step": 2196 + }, + { + "epoch": 0.6048639537683602, + "grad_norm": 0.37120938301086426, + "learning_rate": 1e-05, + "loss": 0.454, + "step": 2198 + }, + { + "epoch": 0.6054143304323897, + "grad_norm": 0.3475727140903473, + "learning_rate": 1e-05, + "loss": 0.4424, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_merge_loss": 0.38126564025878906, + "eval_merge_runtime": 600.3103, + "eval_merge_samples_per_second": 56.224, + "eval_merge_steps_per_second": 2.344, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_new_aug_datas_filtered.json_loss": 0.5048007369041443, + "eval_new_aug_datas_filtered.json_runtime": 10.3514, + "eval_new_aug_datas_filtered.json_samples_per_second": 74.096, + "eval_new_aug_datas_filtered.json_steps_per_second": 3.091, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_sharegpt_gpt4.json_loss": 0.7578977346420288, + "eval_sharegpt_gpt4.json_runtime": 31.6981, + "eval_sharegpt_gpt4.json_samples_per_second": 58.71, + "eval_sharegpt_gpt4.json_steps_per_second": 2.461, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_Table_GPT.json_loss": 0.05305211618542671, + "eval_Table_GPT.json_runtime": 25.0091, + "eval_Table_GPT.json_samples_per_second": 83.69, + "eval_Table_GPT.json_steps_per_second": 3.519, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_gpt_4o_200k.json_loss": 0.7855507135391235, + "eval_gpt_4o_200k.json_runtime": 48.5546, + "eval_gpt_4o_200k.json_samples_per_second": 129.36, + "eval_gpt_4o_200k.json_steps_per_second": 5.396, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_multi_turn_datas.json_loss": 0.3139781355857849, + "eval_multi_turn_datas.json_runtime": 75.6414, + "eval_multi_turn_datas.json_samples_per_second": 52.908, + "eval_multi_turn_datas.json_steps_per_second": 2.208, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_table_python_code_datas.json_loss": 0.2603669762611389, + "eval_table_python_code_datas.json_runtime": 43.0857, + "eval_table_python_code_datas.json_samples_per_second": 50.109, + "eval_table_python_code_datas.json_steps_per_second": 2.089, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_tabular_llm_data.json_loss": 0.0890057235956192, + "eval_tabular_llm_data.json_runtime": 8.5461, + "eval_tabular_llm_data.json_samples_per_second": 28.785, + "eval_tabular_llm_data.json_steps_per_second": 1.287, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_python_code_critic_21k.json_loss": 0.5582770705223083, + "eval_python_code_critic_21k.json_runtime": 3.2316, + "eval_python_code_critic_21k.json_samples_per_second": 184.737, + "eval_python_code_critic_21k.json_steps_per_second": 7.736, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_all_merge_table_dataset.json_loss": 0.07120716571807861, + "eval_all_merge_table_dataset.json_runtime": 23.3637, + "eval_all_merge_table_dataset.json_samples_per_second": 30.475, + "eval_all_merge_table_dataset.json_steps_per_second": 1.284, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_code_feedback_multi_turn.json_loss": 0.5745006799697876, + "eval_code_feedback_multi_turn.json_runtime": 32.5197, + "eval_code_feedback_multi_turn.json_samples_per_second": 67.682, + "eval_code_feedback_multi_turn.json_steps_per_second": 2.829, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_ultrainteract_sft.json_loss": 0.41318273544311523, + "eval_ultrainteract_sft.json_runtime": 8.6602, + "eval_ultrainteract_sft.json_samples_per_second": 168.125, + "eval_ultrainteract_sft.json_steps_per_second": 7.044, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_synthetic_text_to_sql.json_loss": 0.09635543823242188, + "eval_synthetic_text_to_sql.json_runtime": 0.1265, + "eval_synthetic_text_to_sql.json_samples_per_second": 268.832, + "eval_synthetic_text_to_sql.json_steps_per_second": 15.814, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_sft_react_sql_datas.json_loss": 0.6216484904289246, + "eval_sft_react_sql_datas.json_runtime": 7.8599, + "eval_sft_react_sql_datas.json_samples_per_second": 39.949, + "eval_sft_react_sql_datas.json_steps_per_second": 1.781, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_all_merge_code.json_loss": 0.2849319279193878, + "eval_all_merge_code.json_runtime": 0.3296, + "eval_all_merge_code.json_samples_per_second": 191.112, + "eval_all_merge_code.json_steps_per_second": 9.101, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_magpie_datas.json_loss": 0.4269045293331146, + "eval_magpie_datas.json_runtime": 2.2161, + "eval_magpie_datas.json_samples_per_second": 77.615, + "eval_magpie_datas.json_steps_per_second": 3.61, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_train_data_for_qwen.json_loss": 0.005929525941610336, + "eval_train_data_for_qwen.json_runtime": 0.2454, + "eval_train_data_for_qwen.json_samples_per_second": 40.757, + "eval_train_data_for_qwen.json_steps_per_second": 4.076, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_alpaca_cleaned.json_loss": 0.9076781272888184, + "eval_alpaca_cleaned.json_runtime": 0.1144, + "eval_alpaca_cleaned.json_samples_per_second": 236.011, + "eval_alpaca_cleaned.json_steps_per_second": 17.482, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_agent_instruct.json_loss": 0.2231922596693039, + "eval_agent_instruct.json_runtime": 0.5154, + "eval_agent_instruct.json_samples_per_second": 93.136, + "eval_agent_instruct.json_steps_per_second": 3.881, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_MathInstruct.json_loss": 0.19876058399677277, + "eval_MathInstruct.json_runtime": 0.3563, + "eval_MathInstruct.json_samples_per_second": 159.969, + "eval_MathInstruct.json_steps_per_second": 8.419, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_tested_143k_python_alpaca.json_loss": 0.4431252181529999, + "eval_tested_143k_python_alpaca.json_runtime": 0.3026, + "eval_tested_143k_python_alpaca.json_samples_per_second": 112.374, + "eval_tested_143k_python_alpaca.json_steps_per_second": 6.61, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_xlam_function_calling_60k.json_loss": 0.00838847178965807, + "eval_xlam_function_calling_60k.json_runtime": 0.1, + "eval_xlam_function_calling_60k.json_samples_per_second": 230.081, + "eval_xlam_function_calling_60k.json_steps_per_second": 10.004, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_alpaca_data_gpt4_chinese.json_loss": 1.5384413003921509, + "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0514, + "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 311.198, + "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.45, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_alpaca_gpt4_zh.json_loss": 0.969275712966919, + "eval_alpaca_gpt4_zh.json_runtime": 0.0504, + "eval_alpaca_gpt4_zh.json_samples_per_second": 218.311, + "eval_alpaca_gpt4_zh.json_steps_per_second": 19.846, + "step": 2200 + }, + { + "epoch": 0.6054143304323897, + "eval_codefeedback_filtered_instruction.json_loss": 0.5901365876197815, + "eval_codefeedback_filtered_instruction.json_runtime": 0.4874, + "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.032, + "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.052, + "step": 2200 + }, + { + "epoch": 0.6059647070964191, + "grad_norm": 0.37194857001304626, + "learning_rate": 1e-05, + "loss": 0.424, + "step": 2202 + }, + { + "epoch": 0.6065150837604486, + "grad_norm": 0.36095818877220154, + "learning_rate": 1e-05, + "loss": 0.4344, + "step": 2204 + }, + { + "epoch": 0.607065460424478, + "grad_norm": 0.36337706446647644, + "learning_rate": 1e-05, + "loss": 0.4446, + "step": 2206 + }, + { + "epoch": 0.6076158370885074, + "grad_norm": 0.3500390946865082, + "learning_rate": 1e-05, + "loss": 0.4304, + "step": 2208 + }, + { + "epoch": 0.6081662137525369, + "grad_norm": 0.3477112054824829, + "learning_rate": 1e-05, + "loss": 0.4346, + "step": 2210 + }, + { + "epoch": 0.6087165904165663, + "grad_norm": 0.36322692036628723, + "learning_rate": 1e-05, + "loss": 0.4311, + "step": 2212 + }, + { + "epoch": 0.6092669670805958, + "grad_norm": 0.37783941626548767, + "learning_rate": 1e-05, + "loss": 0.4389, + "step": 2214 + }, + { + "epoch": 0.6098173437446253, + "grad_norm": 0.36018887162208557, + "learning_rate": 1e-05, + "loss": 0.4321, + "step": 2216 + }, + { + "epoch": 0.6103677204086547, + "grad_norm": 0.34396857023239136, + "learning_rate": 1e-05, + "loss": 0.4349, + "step": 2218 + }, + { + "epoch": 0.6109180970726841, + "grad_norm": 0.3611605167388916, + "learning_rate": 1e-05, + "loss": 0.4305, + "step": 2220 + }, + { + "epoch": 0.6114684737367135, + "grad_norm": 0.339339941740036, + "learning_rate": 1e-05, + "loss": 0.4338, + "step": 2222 + }, + { + "epoch": 0.612018850400743, + "grad_norm": 0.32705169916152954, + "learning_rate": 1e-05, + "loss": 0.4275, + "step": 2224 + }, + { + "epoch": 0.6125692270647725, + "grad_norm": 0.3551005721092224, + "learning_rate": 1e-05, + "loss": 0.4365, + "step": 2226 + }, + { + "epoch": 0.6131196037288019, + "grad_norm": 0.3826168179512024, + "learning_rate": 1e-05, + "loss": 0.4325, + "step": 2228 + }, + { + "epoch": 0.6136699803928314, + "grad_norm": 0.376407653093338, + "learning_rate": 1e-05, + "loss": 0.4325, + "step": 2230 + }, + { + "epoch": 0.6142203570568608, + "grad_norm": 0.3507418930530548, + "learning_rate": 1e-05, + "loss": 0.4315, + "step": 2232 + }, + { + "epoch": 0.6147707337208902, + "grad_norm": 0.3515014946460724, + "learning_rate": 1e-05, + "loss": 0.4432, + "step": 2234 + }, + { + "epoch": 0.6153211103849197, + "grad_norm": 0.37726324796676636, + "learning_rate": 1e-05, + "loss": 0.4389, + "step": 2236 + }, + { + "epoch": 0.6158714870489491, + "grad_norm": 0.35043272376060486, + "learning_rate": 1e-05, + "loss": 0.4406, + "step": 2238 + }, + { + "epoch": 0.6164218637129786, + "grad_norm": 0.3619838356971741, + "learning_rate": 1e-05, + "loss": 0.4381, + "step": 2240 + }, + { + "epoch": 0.616972240377008, + "grad_norm": 0.3727911114692688, + "learning_rate": 1e-05, + "loss": 0.4261, + "step": 2242 + }, + { + "epoch": 0.6175226170410375, + "grad_norm": 0.35618454217910767, + "learning_rate": 1e-05, + "loss": 0.4353, + "step": 2244 + }, + { + "epoch": 0.6180729937050669, + "grad_norm": 0.3659394681453705, + "learning_rate": 1e-05, + "loss": 0.4281, + "step": 2246 + }, + { + "epoch": 0.6186233703690963, + "grad_norm": 0.35864701867103577, + "learning_rate": 1e-05, + "loss": 0.4409, + "step": 2248 + }, + { + "epoch": 0.6191737470331258, + "grad_norm": 0.36990123987197876, + "learning_rate": 1e-05, + "loss": 0.4424, + "step": 2250 + }, + { + "epoch": 0.6197241236971552, + "grad_norm": 0.36422237753868103, + "learning_rate": 1e-05, + "loss": 0.4516, + "step": 2252 + }, + { + "epoch": 0.6202745003611847, + "grad_norm": 0.34886521100997925, + "learning_rate": 1e-05, + "loss": 0.4299, + "step": 2254 + }, + { + "epoch": 0.6208248770252142, + "grad_norm": 0.3683704137802124, + "learning_rate": 1e-05, + "loss": 0.4379, + "step": 2256 + }, + { + "epoch": 0.6213752536892436, + "grad_norm": 0.3535701334476471, + "learning_rate": 1e-05, + "loss": 0.4292, + "step": 2258 + }, + { + "epoch": 0.621925630353273, + "grad_norm": 0.370959997177124, + "learning_rate": 1e-05, + "loss": 0.4425, + "step": 2260 + }, + { + "epoch": 0.6224760070173024, + "grad_norm": 0.3473008871078491, + "learning_rate": 1e-05, + "loss": 0.4289, + "step": 2262 + }, + { + "epoch": 0.6230263836813319, + "grad_norm": 0.36245644092559814, + "learning_rate": 1e-05, + "loss": 0.4525, + "step": 2264 + }, + { + "epoch": 0.6235767603453614, + "grad_norm": 0.37182751297950745, + "learning_rate": 1e-05, + "loss": 0.4438, + "step": 2266 + }, + { + "epoch": 0.6241271370093908, + "grad_norm": 0.35843655467033386, + "learning_rate": 1e-05, + "loss": 0.4403, + "step": 2268 + }, + { + "epoch": 0.6246775136734203, + "grad_norm": 0.3484828472137451, + "learning_rate": 1e-05, + "loss": 0.429, + "step": 2270 + }, + { + "epoch": 0.6252278903374497, + "grad_norm": 0.35097572207450867, + "learning_rate": 1e-05, + "loss": 0.4435, + "step": 2272 + }, + { + "epoch": 0.6257782670014791, + "grad_norm": 0.35911381244659424, + "learning_rate": 1e-05, + "loss": 0.435, + "step": 2274 + }, + { + "epoch": 0.6263286436655086, + "grad_norm": 0.3544057309627533, + "learning_rate": 1e-05, + "loss": 0.4359, + "step": 2276 + }, + { + "epoch": 0.626879020329538, + "grad_norm": 0.34516793489456177, + "learning_rate": 1e-05, + "loss": 0.4261, + "step": 2278 + }, + { + "epoch": 0.6274293969935675, + "grad_norm": 0.3534994423389435, + "learning_rate": 1e-05, + "loss": 0.4539, + "step": 2280 + }, + { + "epoch": 0.6279797736575969, + "grad_norm": 0.356238454580307, + "learning_rate": 1e-05, + "loss": 0.4321, + "step": 2282 + }, + { + "epoch": 0.6285301503216264, + "grad_norm": 0.37285274267196655, + "learning_rate": 1e-05, + "loss": 0.4515, + "step": 2284 + }, + { + "epoch": 0.6290805269856559, + "grad_norm": 0.3517172336578369, + "learning_rate": 1e-05, + "loss": 0.4268, + "step": 2286 + }, + { + "epoch": 0.6296309036496852, + "grad_norm": 0.35732871294021606, + "learning_rate": 1e-05, + "loss": 0.4363, + "step": 2288 + }, + { + "epoch": 0.6301812803137147, + "grad_norm": 0.3592797815799713, + "learning_rate": 1e-05, + "loss": 0.4424, + "step": 2290 + }, + { + "epoch": 0.6307316569777441, + "grad_norm": 0.3233913481235504, + "learning_rate": 1e-05, + "loss": 0.421, + "step": 2292 + }, + { + "epoch": 0.6312820336417736, + "grad_norm": 0.361591100692749, + "learning_rate": 1e-05, + "loss": 0.4299, + "step": 2294 + }, + { + "epoch": 0.6318324103058031, + "grad_norm": 0.3468184173107147, + "learning_rate": 1e-05, + "loss": 0.4442, + "step": 2296 + }, + { + "epoch": 0.6323827869698325, + "grad_norm": 0.4019412398338318, + "learning_rate": 1e-05, + "loss": 0.4453, + "step": 2298 + }, + { + "epoch": 0.6329331636338619, + "grad_norm": 0.3713074326515198, + "learning_rate": 1e-05, + "loss": 0.435, + "step": 2300 + }, + { + "epoch": 0.6334835402978913, + "grad_norm": 0.35839253664016724, + "learning_rate": 1e-05, + "loss": 0.4449, + "step": 2302 + }, + { + "epoch": 0.6340339169619208, + "grad_norm": 0.33958542346954346, + "learning_rate": 1e-05, + "loss": 0.4433, + "step": 2304 + }, + { + "epoch": 0.6345842936259503, + "grad_norm": 0.3750527501106262, + "learning_rate": 1e-05, + "loss": 0.4297, + "step": 2306 + }, + { + "epoch": 0.6351346702899797, + "grad_norm": 0.35579168796539307, + "learning_rate": 1e-05, + "loss": 0.4307, + "step": 2308 + }, + { + "epoch": 0.6356850469540092, + "grad_norm": 0.3424528241157532, + "learning_rate": 1e-05, + "loss": 0.4451, + "step": 2310 + }, + { + "epoch": 0.6362354236180386, + "grad_norm": 0.3364480137825012, + "learning_rate": 1e-05, + "loss": 0.4251, + "step": 2312 + }, + { + "epoch": 0.636785800282068, + "grad_norm": 0.35307276248931885, + "learning_rate": 1e-05, + "loss": 0.4221, + "step": 2314 + }, + { + "epoch": 0.6373361769460975, + "grad_norm": 0.41354474425315857, + "learning_rate": 1e-05, + "loss": 0.4462, + "step": 2316 + }, + { + "epoch": 0.6378865536101269, + "grad_norm": 0.37485471367836, + "learning_rate": 1e-05, + "loss": 0.4337, + "step": 2318 + }, + { + "epoch": 0.6384369302741564, + "grad_norm": 0.344091534614563, + "learning_rate": 1e-05, + "loss": 0.43, + "step": 2320 + }, + { + "epoch": 0.6389873069381858, + "grad_norm": 0.3772261440753937, + "learning_rate": 1e-05, + "loss": 0.4444, + "step": 2322 + }, + { + "epoch": 0.6395376836022153, + "grad_norm": 0.35307928919792175, + "learning_rate": 1e-05, + "loss": 0.4332, + "step": 2324 + }, + { + "epoch": 0.6400880602662448, + "grad_norm": 0.35815975069999695, + "learning_rate": 1e-05, + "loss": 0.4489, + "step": 2326 + }, + { + "epoch": 0.6406384369302741, + "grad_norm": 0.3731154799461365, + "learning_rate": 1e-05, + "loss": 0.4355, + "step": 2328 + }, + { + "epoch": 0.6411888135943036, + "grad_norm": 0.36875462532043457, + "learning_rate": 1e-05, + "loss": 0.4339, + "step": 2330 + }, + { + "epoch": 0.641739190258333, + "grad_norm": 0.36913126707077026, + "learning_rate": 1e-05, + "loss": 0.4336, + "step": 2332 + }, + { + "epoch": 0.6422895669223625, + "grad_norm": 0.35829678177833557, + "learning_rate": 1e-05, + "loss": 0.4438, + "step": 2334 + }, + { + "epoch": 0.642839943586392, + "grad_norm": 0.36390239000320435, + "learning_rate": 1e-05, + "loss": 0.4405, + "step": 2336 + }, + { + "epoch": 0.6433903202504214, + "grad_norm": 0.34786713123321533, + "learning_rate": 1e-05, + "loss": 0.451, + "step": 2338 + }, + { + "epoch": 0.6439406969144508, + "grad_norm": 0.3522484600543976, + "learning_rate": 1e-05, + "loss": 0.4395, + "step": 2340 + }, + { + "epoch": 0.6444910735784802, + "grad_norm": 0.36442965269088745, + "learning_rate": 1e-05, + "loss": 0.4204, + "step": 2342 + }, + { + "epoch": 0.6450414502425097, + "grad_norm": 0.3635409474372864, + "learning_rate": 1e-05, + "loss": 0.4507, + "step": 2344 + }, + { + "epoch": 0.6455918269065392, + "grad_norm": 0.35682952404022217, + "learning_rate": 1e-05, + "loss": 0.4333, + "step": 2346 + }, + { + "epoch": 0.6461422035705686, + "grad_norm": 0.38101914525032043, + "learning_rate": 1e-05, + "loss": 0.4409, + "step": 2348 + }, + { + "epoch": 0.6466925802345981, + "grad_norm": 0.37273916602134705, + "learning_rate": 1e-05, + "loss": 0.4386, + "step": 2350 + }, + { + "epoch": 0.6472429568986275, + "grad_norm": 0.37394535541534424, + "learning_rate": 1e-05, + "loss": 0.4426, + "step": 2352 + }, + { + "epoch": 0.6477933335626569, + "grad_norm": 0.3374865651130676, + "learning_rate": 1e-05, + "loss": 0.443, + "step": 2354 + }, + { + "epoch": 0.6483437102266864, + "grad_norm": 0.34875357151031494, + "learning_rate": 1e-05, + "loss": 0.4135, + "step": 2356 + }, + { + "epoch": 0.6488940868907158, + "grad_norm": 0.365508109331131, + "learning_rate": 1e-05, + "loss": 0.4455, + "step": 2358 + }, + { + "epoch": 0.6494444635547453, + "grad_norm": 0.36924096941947937, + "learning_rate": 1e-05, + "loss": 0.4327, + "step": 2360 + }, + { + "epoch": 0.6499948402187747, + "grad_norm": 0.3646699786186218, + "learning_rate": 1e-05, + "loss": 0.4324, + "step": 2362 + }, + { + "epoch": 0.6505452168828042, + "grad_norm": 0.34241992235183716, + "learning_rate": 1e-05, + "loss": 0.4414, + "step": 2364 + }, + { + "epoch": 0.6510955935468337, + "grad_norm": 0.3360735774040222, + "learning_rate": 1e-05, + "loss": 0.4228, + "step": 2366 + }, + { + "epoch": 0.651645970210863, + "grad_norm": 0.3782423138618469, + "learning_rate": 1e-05, + "loss": 0.4366, + "step": 2368 + }, + { + "epoch": 0.6521963468748925, + "grad_norm": 0.3839074373245239, + "learning_rate": 1e-05, + "loss": 0.4389, + "step": 2370 + }, + { + "epoch": 0.6527467235389219, + "grad_norm": 0.3636200726032257, + "learning_rate": 1e-05, + "loss": 0.4418, + "step": 2372 + }, + { + "epoch": 0.6532971002029514, + "grad_norm": 0.3629804253578186, + "learning_rate": 1e-05, + "loss": 0.4259, + "step": 2374 + }, + { + "epoch": 0.6538474768669809, + "grad_norm": 0.3819858133792877, + "learning_rate": 1e-05, + "loss": 0.4348, + "step": 2376 + }, + { + "epoch": 0.6543978535310103, + "grad_norm": 0.3597410321235657, + "learning_rate": 1e-05, + "loss": 0.428, + "step": 2378 + }, + { + "epoch": 0.6549482301950398, + "grad_norm": 0.4084703326225281, + "learning_rate": 1e-05, + "loss": 0.4478, + "step": 2380 + }, + { + "epoch": 0.6554986068590691, + "grad_norm": 0.35995879769325256, + "learning_rate": 1e-05, + "loss": 0.4356, + "step": 2382 + }, + { + "epoch": 0.6560489835230986, + "grad_norm": 0.36047980189323425, + "learning_rate": 1e-05, + "loss": 0.4479, + "step": 2384 + }, + { + "epoch": 0.6565993601871281, + "grad_norm": 0.3532986342906952, + "learning_rate": 1e-05, + "loss": 0.424, + "step": 2386 + }, + { + "epoch": 0.6571497368511575, + "grad_norm": 0.3374999761581421, + "learning_rate": 1e-05, + "loss": 0.4338, + "step": 2388 + }, + { + "epoch": 0.657700113515187, + "grad_norm": 0.34645605087280273, + "learning_rate": 1e-05, + "loss": 0.4257, + "step": 2390 + }, + { + "epoch": 0.6582504901792164, + "grad_norm": 0.36470580101013184, + "learning_rate": 1e-05, + "loss": 0.4414, + "step": 2392 + }, + { + "epoch": 0.6588008668432458, + "grad_norm": 0.3823862075805664, + "learning_rate": 1e-05, + "loss": 0.4306, + "step": 2394 + }, + { + "epoch": 0.6593512435072753, + "grad_norm": 0.4070727229118347, + "learning_rate": 1e-05, + "loss": 0.4322, + "step": 2396 + }, + { + "epoch": 0.6599016201713047, + "grad_norm": 0.37519609928131104, + "learning_rate": 1e-05, + "loss": 0.4248, + "step": 2398 + }, + { + "epoch": 0.6604519968353342, + "grad_norm": 0.35447025299072266, + "learning_rate": 1e-05, + "loss": 0.4283, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_merge_loss": 0.37715020775794983, + "eval_merge_runtime": 600.5757, + "eval_merge_samples_per_second": 56.199, + "eval_merge_steps_per_second": 2.343, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_new_aug_datas_filtered.json_loss": 0.5012194514274597, + "eval_new_aug_datas_filtered.json_runtime": 10.4212, + "eval_new_aug_datas_filtered.json_samples_per_second": 73.6, + "eval_new_aug_datas_filtered.json_steps_per_second": 3.071, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_sharegpt_gpt4.json_loss": 0.7534219026565552, + "eval_sharegpt_gpt4.json_runtime": 31.7308, + "eval_sharegpt_gpt4.json_samples_per_second": 58.65, + "eval_sharegpt_gpt4.json_steps_per_second": 2.458, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_Table_GPT.json_loss": 0.050881169736385345, + "eval_Table_GPT.json_runtime": 24.9922, + "eval_Table_GPT.json_samples_per_second": 83.746, + "eval_Table_GPT.json_steps_per_second": 3.521, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_gpt_4o_200k.json_loss": 0.7805712223052979, + "eval_gpt_4o_200k.json_runtime": 48.518, + "eval_gpt_4o_200k.json_samples_per_second": 129.457, + "eval_gpt_4o_200k.json_steps_per_second": 5.4, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_multi_turn_datas.json_loss": 0.3069368898868561, + "eval_multi_turn_datas.json_runtime": 75.8513, + "eval_multi_turn_datas.json_samples_per_second": 52.761, + "eval_multi_turn_datas.json_steps_per_second": 2.202, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_table_python_code_datas.json_loss": 0.2562294006347656, + "eval_table_python_code_datas.json_runtime": 43.1545, + "eval_table_python_code_datas.json_samples_per_second": 50.03, + "eval_table_python_code_datas.json_steps_per_second": 2.086, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_tabular_llm_data.json_loss": 0.09128429740667343, + "eval_tabular_llm_data.json_runtime": 8.5524, + "eval_tabular_llm_data.json_samples_per_second": 28.764, + "eval_tabular_llm_data.json_steps_per_second": 1.286, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_python_code_critic_21k.json_loss": 0.5555644631385803, + "eval_python_code_critic_21k.json_runtime": 3.2271, + "eval_python_code_critic_21k.json_samples_per_second": 184.994, + "eval_python_code_critic_21k.json_steps_per_second": 7.747, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_all_merge_table_dataset.json_loss": 0.07006299495697021, + "eval_all_merge_table_dataset.json_runtime": 23.358, + "eval_all_merge_table_dataset.json_samples_per_second": 30.482, + "eval_all_merge_table_dataset.json_steps_per_second": 1.284, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_code_feedback_multi_turn.json_loss": 0.5720005035400391, + "eval_code_feedback_multi_turn.json_runtime": 32.5016, + "eval_code_feedback_multi_turn.json_samples_per_second": 67.72, + "eval_code_feedback_multi_turn.json_steps_per_second": 2.831, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_ultrainteract_sft.json_loss": 0.4097177982330322, + "eval_ultrainteract_sft.json_runtime": 8.6753, + "eval_ultrainteract_sft.json_samples_per_second": 167.832, + "eval_ultrainteract_sft.json_steps_per_second": 7.031, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_synthetic_text_to_sql.json_loss": 0.09309177845716476, + "eval_synthetic_text_to_sql.json_runtime": 0.1257, + "eval_synthetic_text_to_sql.json_samples_per_second": 270.423, + "eval_synthetic_text_to_sql.json_steps_per_second": 15.907, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_sft_react_sql_datas.json_loss": 0.6212250590324402, + "eval_sft_react_sql_datas.json_runtime": 7.859, + "eval_sft_react_sql_datas.json_samples_per_second": 39.954, + "eval_sft_react_sql_datas.json_steps_per_second": 1.781, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_all_merge_code.json_loss": 0.28449881076812744, + "eval_all_merge_code.json_runtime": 0.3298, + "eval_all_merge_code.json_samples_per_second": 191.001, + "eval_all_merge_code.json_steps_per_second": 9.095, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_magpie_datas.json_loss": 0.426034539937973, + "eval_magpie_datas.json_runtime": 2.2154, + "eval_magpie_datas.json_samples_per_second": 77.638, + "eval_magpie_datas.json_steps_per_second": 3.611, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_train_data_for_qwen.json_loss": 0.005596214439719915, + "eval_train_data_for_qwen.json_runtime": 0.2424, + "eval_train_data_for_qwen.json_samples_per_second": 41.251, + "eval_train_data_for_qwen.json_steps_per_second": 4.125, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_alpaca_cleaned.json_loss": 0.9008170962333679, + "eval_alpaca_cleaned.json_runtime": 0.1147, + "eval_alpaca_cleaned.json_samples_per_second": 235.421, + "eval_alpaca_cleaned.json_steps_per_second": 17.439, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_agent_instruct.json_loss": 0.21443764865398407, + "eval_agent_instruct.json_runtime": 0.5141, + "eval_agent_instruct.json_samples_per_second": 93.36, + "eval_agent_instruct.json_steps_per_second": 3.89, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_MathInstruct.json_loss": 0.1956825852394104, + "eval_MathInstruct.json_runtime": 0.3499, + "eval_MathInstruct.json_samples_per_second": 162.885, + "eval_MathInstruct.json_steps_per_second": 8.573, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_tested_143k_python_alpaca.json_loss": 0.4434005320072174, + "eval_tested_143k_python_alpaca.json_runtime": 0.3023, + "eval_tested_143k_python_alpaca.json_samples_per_second": 112.46, + "eval_tested_143k_python_alpaca.json_steps_per_second": 6.615, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_xlam_function_calling_60k.json_loss": 0.009229443967342377, + "eval_xlam_function_calling_60k.json_runtime": 0.1004, + "eval_xlam_function_calling_60k.json_samples_per_second": 229.185, + "eval_xlam_function_calling_60k.json_steps_per_second": 9.965, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_alpaca_data_gpt4_chinese.json_loss": 1.5269618034362793, + "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0516, + "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 310.215, + "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.388, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_alpaca_gpt4_zh.json_loss": 0.9699357151985168, + "eval_alpaca_gpt4_zh.json_runtime": 0.0505, + "eval_alpaca_gpt4_zh.json_samples_per_second": 217.964, + "eval_alpaca_gpt4_zh.json_steps_per_second": 19.815, + "step": 2400 + }, + { + "epoch": 0.6604519968353342, + "eval_codefeedback_filtered_instruction.json_loss": 0.5749525427818298, + "eval_codefeedback_filtered_instruction.json_runtime": 0.4875, + "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.023, + "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.051, + "step": 2400 + }, + { + "epoch": 0.6610023734993636, + "grad_norm": 0.38521307706832886, + "learning_rate": 1e-05, + "loss": 0.4408, + "step": 2402 + }, + { + "epoch": 0.6615527501633931, + "grad_norm": 0.35963118076324463, + "learning_rate": 1e-05, + "loss": 0.4252, + "step": 2404 + }, + { + "epoch": 0.6621031268274226, + "grad_norm": 0.34755435585975647, + "learning_rate": 1e-05, + "loss": 0.4319, + "step": 2406 + }, + { + "epoch": 0.6626535034914519, + "grad_norm": 0.37133127450942993, + "learning_rate": 1e-05, + "loss": 0.4237, + "step": 2408 + }, + { + "epoch": 0.6632038801554814, + "grad_norm": 0.35870301723480225, + "learning_rate": 1e-05, + "loss": 0.4388, + "step": 2410 + }, + { + "epoch": 0.6637542568195108, + "grad_norm": 0.357415109872818, + "learning_rate": 1e-05, + "loss": 0.4322, + "step": 2412 + }, + { + "epoch": 0.6643046334835403, + "grad_norm": 0.3610486090183258, + "learning_rate": 1e-05, + "loss": 0.434, + "step": 2414 + }, + { + "epoch": 0.6648550101475698, + "grad_norm": 0.35058531165122986, + "learning_rate": 1e-05, + "loss": 0.4325, + "step": 2416 + }, + { + "epoch": 0.6654053868115992, + "grad_norm": 0.3732353448867798, + "learning_rate": 1e-05, + "loss": 0.4266, + "step": 2418 + }, + { + "epoch": 0.6659557634756287, + "grad_norm": 0.3728616535663605, + "learning_rate": 1e-05, + "loss": 0.4373, + "step": 2420 + }, + { + "epoch": 0.666506140139658, + "grad_norm": 0.3697822093963623, + "learning_rate": 1e-05, + "loss": 0.4263, + "step": 2422 + }, + { + "epoch": 0.6670565168036875, + "grad_norm": 0.34242671728134155, + "learning_rate": 1e-05, + "loss": 0.4234, + "step": 2424 + }, + { + "epoch": 0.667606893467717, + "grad_norm": 0.34660401940345764, + "learning_rate": 1e-05, + "loss": 0.4438, + "step": 2426 + }, + { + "epoch": 0.6681572701317464, + "grad_norm": 0.36335524916648865, + "learning_rate": 1e-05, + "loss": 0.447, + "step": 2428 + }, + { + "epoch": 0.6687076467957759, + "grad_norm": 0.39879950881004333, + "learning_rate": 1e-05, + "loss": 0.4328, + "step": 2430 + }, + { + "epoch": 0.6692580234598053, + "grad_norm": 0.3318917453289032, + "learning_rate": 1e-05, + "loss": 0.418, + "step": 2432 + }, + { + "epoch": 0.6698084001238348, + "grad_norm": 0.3548910319805145, + "learning_rate": 1e-05, + "loss": 0.4297, + "step": 2434 + }, + { + "epoch": 0.6703587767878642, + "grad_norm": 0.35431650280952454, + "learning_rate": 1e-05, + "loss": 0.4442, + "step": 2436 + }, + { + "epoch": 0.6709091534518936, + "grad_norm": 0.3501831889152527, + "learning_rate": 1e-05, + "loss": 0.4231, + "step": 2438 + }, + { + "epoch": 0.6714595301159231, + "grad_norm": 0.3664182424545288, + "learning_rate": 1e-05, + "loss": 0.4307, + "step": 2440 + }, + { + "epoch": 0.6720099067799525, + "grad_norm": 0.36051392555236816, + "learning_rate": 1e-05, + "loss": 0.4348, + "step": 2442 + }, + { + "epoch": 0.672560283443982, + "grad_norm": 0.38968268036842346, + "learning_rate": 1e-05, + "loss": 0.44, + "step": 2444 + }, + { + "epoch": 0.6731106601080115, + "grad_norm": 0.34485840797424316, + "learning_rate": 1e-05, + "loss": 0.4387, + "step": 2446 + }, + { + "epoch": 0.6736610367720408, + "grad_norm": 0.36389604210853577, + "learning_rate": 1e-05, + "loss": 0.4279, + "step": 2448 + }, + { + "epoch": 0.6742114134360703, + "grad_norm": 0.3703545331954956, + "learning_rate": 1e-05, + "loss": 0.4498, + "step": 2450 + }, + { + "epoch": 0.6747617901000997, + "grad_norm": 0.34628036618232727, + "learning_rate": 1e-05, + "loss": 0.4145, + "step": 2452 + }, + { + "epoch": 0.6753121667641292, + "grad_norm": 0.3569451570510864, + "learning_rate": 1e-05, + "loss": 0.4308, + "step": 2454 + }, + { + "epoch": 0.6758625434281587, + "grad_norm": 0.3471825122833252, + "learning_rate": 1e-05, + "loss": 0.4299, + "step": 2456 + }, + { + "epoch": 0.6764129200921881, + "grad_norm": 0.37446585297584534, + "learning_rate": 1e-05, + "loss": 0.4417, + "step": 2458 + }, + { + "epoch": 0.6769632967562176, + "grad_norm": 0.355708509683609, + "learning_rate": 1e-05, + "loss": 0.4306, + "step": 2460 + }, + { + "epoch": 0.6775136734202469, + "grad_norm": 0.36398351192474365, + "learning_rate": 1e-05, + "loss": 0.4331, + "step": 2462 + }, + { + "epoch": 0.6780640500842764, + "grad_norm": 0.38390782475471497, + "learning_rate": 1e-05, + "loss": 0.4421, + "step": 2464 + }, + { + "epoch": 0.6786144267483059, + "grad_norm": 0.3586190938949585, + "learning_rate": 1e-05, + "loss": 0.4365, + "step": 2466 + }, + { + "epoch": 0.6791648034123353, + "grad_norm": 0.33874934911727905, + "learning_rate": 1e-05, + "loss": 0.4346, + "step": 2468 + }, + { + "epoch": 0.6797151800763648, + "grad_norm": 0.3699466586112976, + "learning_rate": 1e-05, + "loss": 0.4282, + "step": 2470 + }, + { + "epoch": 0.6802655567403942, + "grad_norm": 0.35685962438583374, + "learning_rate": 1e-05, + "loss": 0.4386, + "step": 2472 + }, + { + "epoch": 0.6808159334044237, + "grad_norm": 0.36509183049201965, + "learning_rate": 1e-05, + "loss": 0.4234, + "step": 2474 + }, + { + "epoch": 0.681366310068453, + "grad_norm": 0.3677407503128052, + "learning_rate": 1e-05, + "loss": 0.4327, + "step": 2476 + }, + { + "epoch": 0.6819166867324825, + "grad_norm": 0.361396849155426, + "learning_rate": 1e-05, + "loss": 0.4282, + "step": 2478 + }, + { + "epoch": 0.682467063396512, + "grad_norm": 0.3637540936470032, + "learning_rate": 1e-05, + "loss": 0.4304, + "step": 2480 + }, + { + "epoch": 0.6830174400605414, + "grad_norm": 0.38396722078323364, + "learning_rate": 1e-05, + "loss": 0.4326, + "step": 2482 + }, + { + "epoch": 0.6835678167245709, + "grad_norm": 0.3760308623313904, + "learning_rate": 1e-05, + "loss": 0.4288, + "step": 2484 + }, + { + "epoch": 0.6841181933886004, + "grad_norm": 0.36777281761169434, + "learning_rate": 1e-05, + "loss": 0.4435, + "step": 2486 + }, + { + "epoch": 0.6846685700526298, + "grad_norm": 0.36967626214027405, + "learning_rate": 1e-05, + "loss": 0.4247, + "step": 2488 + }, + { + "epoch": 0.6852189467166592, + "grad_norm": 0.37309199571609497, + "learning_rate": 1e-05, + "loss": 0.4514, + "step": 2490 + }, + { + "epoch": 0.6857693233806886, + "grad_norm": 0.35478582978248596, + "learning_rate": 1e-05, + "loss": 0.436, + "step": 2492 + }, + { + "epoch": 0.6863197000447181, + "grad_norm": 0.35142141580581665, + "learning_rate": 1e-05, + "loss": 0.4289, + "step": 2494 + }, + { + "epoch": 0.6868700767087476, + "grad_norm": 0.37468215823173523, + "learning_rate": 1e-05, + "loss": 0.4363, + "step": 2496 + }, + { + "epoch": 0.687420453372777, + "grad_norm": 0.3481496572494507, + "learning_rate": 1e-05, + "loss": 0.441, + "step": 2498 + }, + { + "epoch": 0.6879708300368065, + "grad_norm": 0.34628838300704956, + "learning_rate": 1e-05, + "loss": 0.4425, + "step": 2500 + }, + { + "epoch": 0.6885212067008358, + "grad_norm": 0.3759724497795105, + "learning_rate": 1e-05, + "loss": 0.4322, + "step": 2502 + }, + { + "epoch": 0.6890715833648653, + "grad_norm": 0.37153902649879456, + "learning_rate": 1e-05, + "loss": 0.4412, + "step": 2504 + }, + { + "epoch": 0.6896219600288948, + "grad_norm": 0.3601967990398407, + "learning_rate": 1e-05, + "loss": 0.4314, + "step": 2506 + }, + { + "epoch": 0.6901723366929242, + "grad_norm": 0.3510344326496124, + "learning_rate": 1e-05, + "loss": 0.4261, + "step": 2508 + }, + { + "epoch": 0.6907227133569537, + "grad_norm": 0.34007585048675537, + "learning_rate": 1e-05, + "loss": 0.4272, + "step": 2510 + }, + { + "epoch": 0.6912730900209831, + "grad_norm": 0.34424078464508057, + "learning_rate": 1e-05, + "loss": 0.4253, + "step": 2512 + }, + { + "epoch": 0.6918234666850126, + "grad_norm": 0.36498820781707764, + "learning_rate": 1e-05, + "loss": 0.434, + "step": 2514 + }, + { + "epoch": 0.692373843349042, + "grad_norm": 0.3697148859500885, + "learning_rate": 1e-05, + "loss": 0.4358, + "step": 2516 + }, + { + "epoch": 0.6929242200130714, + "grad_norm": 0.36114463210105896, + "learning_rate": 1e-05, + "loss": 0.4177, + "step": 2518 + }, + { + "epoch": 0.6934745966771009, + "grad_norm": 0.3630925714969635, + "learning_rate": 1e-05, + "loss": 0.4438, + "step": 2520 + }, + { + "epoch": 0.6940249733411303, + "grad_norm": 0.36949414014816284, + "learning_rate": 1e-05, + "loss": 0.4281, + "step": 2522 + }, + { + "epoch": 0.6945753500051598, + "grad_norm": 0.36324694752693176, + "learning_rate": 1e-05, + "loss": 0.4253, + "step": 2524 + }, + { + "epoch": 0.6951257266691893, + "grad_norm": 0.3471947908401489, + "learning_rate": 1e-05, + "loss": 0.4215, + "step": 2526 + }, + { + "epoch": 0.6956761033332187, + "grad_norm": 0.33943814039230347, + "learning_rate": 1e-05, + "loss": 0.4546, + "step": 2528 + }, + { + "epoch": 0.6962264799972481, + "grad_norm": 0.34675729274749756, + "learning_rate": 1e-05, + "loss": 0.4191, + "step": 2530 + }, + { + "epoch": 0.6967768566612775, + "grad_norm": 0.3519613742828369, + "learning_rate": 1e-05, + "loss": 0.4272, + "step": 2532 + }, + { + "epoch": 0.697327233325307, + "grad_norm": 0.3635639548301697, + "learning_rate": 1e-05, + "loss": 0.4489, + "step": 2534 + }, + { + "epoch": 0.6978776099893365, + "grad_norm": 0.3636915385723114, + "learning_rate": 1e-05, + "loss": 0.4233, + "step": 2536 + }, + { + "epoch": 0.6984279866533659, + "grad_norm": 0.36174023151397705, + "learning_rate": 1e-05, + "loss": 0.425, + "step": 2538 + }, + { + "epoch": 0.6989783633173954, + "grad_norm": 0.35721176862716675, + "learning_rate": 1e-05, + "loss": 0.4279, + "step": 2540 + }, + { + "epoch": 0.6995287399814248, + "grad_norm": 0.35394319891929626, + "learning_rate": 1e-05, + "loss": 0.4279, + "step": 2542 + }, + { + "epoch": 0.7000791166454542, + "grad_norm": 0.37505972385406494, + "learning_rate": 1e-05, + "loss": 0.423, + "step": 2544 + }, + { + "epoch": 0.7006294933094837, + "grad_norm": 0.3504476249217987, + "learning_rate": 1e-05, + "loss": 0.4212, + "step": 2546 + }, + { + "epoch": 0.7011798699735131, + "grad_norm": 0.39700883626937866, + "learning_rate": 1e-05, + "loss": 0.4257, + "step": 2548 + }, + { + "epoch": 0.7017302466375426, + "grad_norm": 0.36360886693000793, + "learning_rate": 1e-05, + "loss": 0.4276, + "step": 2550 + }, + { + "epoch": 0.702280623301572, + "grad_norm": 0.36123448610305786, + "learning_rate": 1e-05, + "loss": 0.4266, + "step": 2552 + }, + { + "epoch": 0.7028309999656015, + "grad_norm": 0.35183826088905334, + "learning_rate": 1e-05, + "loss": 0.421, + "step": 2554 + }, + { + "epoch": 0.7033813766296309, + "grad_norm": 0.3557921350002289, + "learning_rate": 1e-05, + "loss": 0.4239, + "step": 2556 + }, + { + "epoch": 0.7039317532936603, + "grad_norm": 0.35415929555892944, + "learning_rate": 1e-05, + "loss": 0.4216, + "step": 2558 + }, + { + "epoch": 0.7044821299576898, + "grad_norm": 0.3662279546260834, + "learning_rate": 1e-05, + "loss": 0.4268, + "step": 2560 + }, + { + "epoch": 0.7050325066217192, + "grad_norm": 0.35718172788619995, + "learning_rate": 1e-05, + "loss": 0.4213, + "step": 2562 + }, + { + "epoch": 0.7055828832857487, + "grad_norm": 0.3595860004425049, + "learning_rate": 1e-05, + "loss": 0.4398, + "step": 2564 + }, + { + "epoch": 0.7061332599497782, + "grad_norm": 0.3576621413230896, + "learning_rate": 1e-05, + "loss": 0.4263, + "step": 2566 + }, + { + "epoch": 0.7066836366138076, + "grad_norm": 0.3699706792831421, + "learning_rate": 1e-05, + "loss": 0.4331, + "step": 2568 + }, + { + "epoch": 0.707234013277837, + "grad_norm": 0.38423609733581543, + "learning_rate": 1e-05, + "loss": 0.436, + "step": 2570 + }, + { + "epoch": 0.7077843899418664, + "grad_norm": 0.3747715651988983, + "learning_rate": 1e-05, + "loss": 0.4335, + "step": 2572 + }, + { + "epoch": 0.7083347666058959, + "grad_norm": 0.3554603159427643, + "learning_rate": 1e-05, + "loss": 0.4236, + "step": 2574 + }, + { + "epoch": 0.7088851432699254, + "grad_norm": 0.35446056723594666, + "learning_rate": 1e-05, + "loss": 0.4235, + "step": 2576 + }, + { + "epoch": 0.7094355199339548, + "grad_norm": 0.3770659267902374, + "learning_rate": 1e-05, + "loss": 0.4344, + "step": 2578 + }, + { + "epoch": 0.7099858965979843, + "grad_norm": 0.35676074028015137, + "learning_rate": 1e-05, + "loss": 0.4241, + "step": 2580 + }, + { + "epoch": 0.7105362732620137, + "grad_norm": 0.3687559962272644, + "learning_rate": 1e-05, + "loss": 0.4329, + "step": 2582 + }, + { + "epoch": 0.7110866499260431, + "grad_norm": 0.35311195254325867, + "learning_rate": 1e-05, + "loss": 0.4355, + "step": 2584 + }, + { + "epoch": 0.7116370265900726, + "grad_norm": 0.3590395152568817, + "learning_rate": 1e-05, + "loss": 0.4213, + "step": 2586 + }, + { + "epoch": 0.712187403254102, + "grad_norm": 0.3694981336593628, + "learning_rate": 1e-05, + "loss": 0.4344, + "step": 2588 + }, + { + "epoch": 0.7127377799181315, + "grad_norm": 0.3516077399253845, + "learning_rate": 1e-05, + "loss": 0.4202, + "step": 2590 + }, + { + "epoch": 0.7132881565821609, + "grad_norm": 0.38859254121780396, + "learning_rate": 1e-05, + "loss": 0.4179, + "step": 2592 + }, + { + "epoch": 0.7138385332461904, + "grad_norm": 0.3825247883796692, + "learning_rate": 1e-05, + "loss": 0.4393, + "step": 2594 + }, + { + "epoch": 0.7143889099102199, + "grad_norm": 0.36817750334739685, + "learning_rate": 1e-05, + "loss": 0.4341, + "step": 2596 + }, + { + "epoch": 0.7149392865742492, + "grad_norm": 0.36351174116134644, + "learning_rate": 1e-05, + "loss": 0.4355, + "step": 2598 + }, + { + "epoch": 0.7154896632382787, + "grad_norm": 0.3494237959384918, + "learning_rate": 1e-05, + "loss": 0.4176, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_merge_loss": 0.3735547661781311, + "eval_merge_runtime": 599.6483, + "eval_merge_samples_per_second": 56.286, + "eval_merge_steps_per_second": 2.346, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_new_aug_datas_filtered.json_loss": 0.4953900873661041, + "eval_new_aug_datas_filtered.json_runtime": 10.4567, + "eval_new_aug_datas_filtered.json_samples_per_second": 73.35, + "eval_new_aug_datas_filtered.json_steps_per_second": 3.06, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_sharegpt_gpt4.json_loss": 0.748174786567688, + "eval_sharegpt_gpt4.json_runtime": 31.7349, + "eval_sharegpt_gpt4.json_samples_per_second": 58.642, + "eval_sharegpt_gpt4.json_steps_per_second": 2.458, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_Table_GPT.json_loss": 0.04870549216866493, + "eval_Table_GPT.json_runtime": 25.0368, + "eval_Table_GPT.json_samples_per_second": 83.597, + "eval_Table_GPT.json_steps_per_second": 3.515, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_gpt_4o_200k.json_loss": 0.775393009185791, + "eval_gpt_4o_200k.json_runtime": 48.6152, + "eval_gpt_4o_200k.json_samples_per_second": 129.198, + "eval_gpt_4o_200k.json_steps_per_second": 5.389, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_multi_turn_datas.json_loss": 0.29874685406684875, + "eval_multi_turn_datas.json_runtime": 75.9064, + "eval_multi_turn_datas.json_samples_per_second": 52.723, + "eval_multi_turn_datas.json_steps_per_second": 2.2, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_table_python_code_datas.json_loss": 0.2535416781902313, + "eval_table_python_code_datas.json_runtime": 43.2787, + "eval_table_python_code_datas.json_samples_per_second": 49.886, + "eval_table_python_code_datas.json_steps_per_second": 2.08, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_tabular_llm_data.json_loss": 0.08522781729698181, + "eval_tabular_llm_data.json_runtime": 8.609, + "eval_tabular_llm_data.json_samples_per_second": 28.575, + "eval_tabular_llm_data.json_steps_per_second": 1.278, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_python_code_critic_21k.json_loss": 0.5531289577484131, + "eval_python_code_critic_21k.json_runtime": 3.2416, + "eval_python_code_critic_21k.json_samples_per_second": 184.167, + "eval_python_code_critic_21k.json_steps_per_second": 7.712, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_all_merge_table_dataset.json_loss": 0.07141314446926117, + "eval_all_merge_table_dataset.json_runtime": 23.4197, + "eval_all_merge_table_dataset.json_samples_per_second": 30.402, + "eval_all_merge_table_dataset.json_steps_per_second": 1.281, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_code_feedback_multi_turn.json_loss": 0.5697857737541199, + "eval_code_feedback_multi_turn.json_runtime": 32.4913, + "eval_code_feedback_multi_turn.json_samples_per_second": 67.741, + "eval_code_feedback_multi_turn.json_steps_per_second": 2.832, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_ultrainteract_sft.json_loss": 0.406777024269104, + "eval_ultrainteract_sft.json_runtime": 8.6553, + "eval_ultrainteract_sft.json_samples_per_second": 168.22, + "eval_ultrainteract_sft.json_steps_per_second": 7.048, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_synthetic_text_to_sql.json_loss": 0.09255770593881607, + "eval_synthetic_text_to_sql.json_runtime": 0.1264, + "eval_synthetic_text_to_sql.json_samples_per_second": 268.887, + "eval_synthetic_text_to_sql.json_steps_per_second": 15.817, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_sft_react_sql_datas.json_loss": 0.6156443953514099, + "eval_sft_react_sql_datas.json_runtime": 7.8669, + "eval_sft_react_sql_datas.json_samples_per_second": 39.914, + "eval_sft_react_sql_datas.json_steps_per_second": 1.78, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_all_merge_code.json_loss": 0.2804557681083679, + "eval_all_merge_code.json_runtime": 0.3331, + "eval_all_merge_code.json_samples_per_second": 189.109, + "eval_all_merge_code.json_steps_per_second": 9.005, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_magpie_datas.json_loss": 0.42615047097206116, + "eval_magpie_datas.json_runtime": 2.2188, + "eval_magpie_datas.json_samples_per_second": 77.518, + "eval_magpie_datas.json_steps_per_second": 3.605, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_train_data_for_qwen.json_loss": 0.005531710106879473, + "eval_train_data_for_qwen.json_runtime": 0.2446, + "eval_train_data_for_qwen.json_samples_per_second": 40.888, + "eval_train_data_for_qwen.json_steps_per_second": 4.089, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_alpaca_cleaned.json_loss": 0.8993179202079773, + "eval_alpaca_cleaned.json_runtime": 0.1158, + "eval_alpaca_cleaned.json_samples_per_second": 233.205, + "eval_alpaca_cleaned.json_steps_per_second": 17.274, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_agent_instruct.json_loss": 0.20902203023433685, + "eval_agent_instruct.json_runtime": 0.5148, + "eval_agent_instruct.json_samples_per_second": 93.239, + "eval_agent_instruct.json_steps_per_second": 3.885, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_MathInstruct.json_loss": 0.20088934898376465, + "eval_MathInstruct.json_runtime": 0.3521, + "eval_MathInstruct.json_samples_per_second": 161.889, + "eval_MathInstruct.json_steps_per_second": 8.52, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_tested_143k_python_alpaca.json_loss": 0.44206199049949646, + "eval_tested_143k_python_alpaca.json_runtime": 0.3013, + "eval_tested_143k_python_alpaca.json_samples_per_second": 112.861, + "eval_tested_143k_python_alpaca.json_steps_per_second": 6.639, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_xlam_function_calling_60k.json_loss": 0.00838589109480381, + "eval_xlam_function_calling_60k.json_runtime": 0.1004, + "eval_xlam_function_calling_60k.json_samples_per_second": 229.101, + "eval_xlam_function_calling_60k.json_steps_per_second": 9.961, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_alpaca_data_gpt4_chinese.json_loss": 1.5224987268447876, + "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0517, + "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 309.243, + "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.328, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_alpaca_gpt4_zh.json_loss": 0.9841532111167908, + "eval_alpaca_gpt4_zh.json_runtime": 0.0501, + "eval_alpaca_gpt4_zh.json_samples_per_second": 219.503, + "eval_alpaca_gpt4_zh.json_steps_per_second": 19.955, + "step": 2600 + }, + { + "epoch": 0.7154896632382787, + "eval_codefeedback_filtered_instruction.json_loss": 0.5787987112998962, + "eval_codefeedback_filtered_instruction.json_runtime": 0.4863, + "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.126, + "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.056, + "step": 2600 + }, + { + "epoch": 0.7160400399023081, + "grad_norm": 0.3617021143436432, + "learning_rate": 1e-05, + "loss": 0.4292, + "step": 2602 + }, + { + "epoch": 0.7165904165663376, + "grad_norm": 0.39201030135154724, + "learning_rate": 1e-05, + "loss": 0.4565, + "step": 2604 + }, + { + "epoch": 0.7171407932303671, + "grad_norm": 0.3617227077484131, + "learning_rate": 1e-05, + "loss": 0.4279, + "step": 2606 + }, + { + "epoch": 0.7176911698943965, + "grad_norm": 0.3502630591392517, + "learning_rate": 1e-05, + "loss": 0.4191, + "step": 2608 + }, + { + "epoch": 0.7182415465584259, + "grad_norm": 0.41853633522987366, + "learning_rate": 1e-05, + "loss": 0.4122, + "step": 2610 + }, + { + "epoch": 0.7187919232224553, + "grad_norm": 0.35474300384521484, + "learning_rate": 1e-05, + "loss": 0.4225, + "step": 2612 + }, + { + "epoch": 0.7193422998864848, + "grad_norm": 0.3673190772533417, + "learning_rate": 1e-05, + "loss": 0.4307, + "step": 2614 + }, + { + "epoch": 0.7198926765505143, + "grad_norm": 0.383365273475647, + "learning_rate": 1e-05, + "loss": 0.4335, + "step": 2616 + }, + { + "epoch": 0.7204430532145437, + "grad_norm": 0.35813844203948975, + "learning_rate": 1e-05, + "loss": 0.4462, + "step": 2618 + }, + { + "epoch": 0.7209934298785732, + "grad_norm": 0.7552120685577393, + "learning_rate": 1e-05, + "loss": 0.4209, + "step": 2620 + }, + { + "epoch": 0.7215438065426026, + "grad_norm": 0.365175724029541, + "learning_rate": 1e-05, + "loss": 0.441, + "step": 2622 + }, + { + "epoch": 0.722094183206632, + "grad_norm": 0.3450736701488495, + "learning_rate": 1e-05, + "loss": 0.4302, + "step": 2624 + }, + { + "epoch": 0.7226445598706615, + "grad_norm": 0.34044018387794495, + "learning_rate": 1e-05, + "loss": 0.4265, + "step": 2626 + }, + { + "epoch": 0.7231949365346909, + "grad_norm": 0.36393091082572937, + "learning_rate": 1e-05, + "loss": 0.4226, + "step": 2628 + }, + { + "epoch": 0.7237453131987204, + "grad_norm": 0.3462166488170624, + "learning_rate": 1e-05, + "loss": 0.4236, + "step": 2630 + }, + { + "epoch": 0.7242956898627498, + "grad_norm": 0.4024192988872528, + "learning_rate": 1e-05, + "loss": 0.4377, + "step": 2632 + }, + { + "epoch": 0.7248460665267793, + "grad_norm": 0.354809045791626, + "learning_rate": 1e-05, + "loss": 0.4245, + "step": 2634 + }, + { + "epoch": 0.7253964431908088, + "grad_norm": 0.3701523244380951, + "learning_rate": 1e-05, + "loss": 0.438, + "step": 2636 + }, + { + "epoch": 0.7259468198548381, + "grad_norm": 0.37080636620521545, + "learning_rate": 1e-05, + "loss": 0.4299, + "step": 2638 + }, + { + "epoch": 0.7264971965188676, + "grad_norm": 0.3205287754535675, + "learning_rate": 1e-05, + "loss": 0.4193, + "step": 2640 + }, + { + "epoch": 0.727047573182897, + "grad_norm": 0.3642041087150574, + "learning_rate": 1e-05, + "loss": 0.4259, + "step": 2642 + }, + { + "epoch": 0.7275979498469265, + "grad_norm": 0.34573763608932495, + "learning_rate": 1e-05, + "loss": 0.438, + "step": 2644 + }, + { + "epoch": 0.728148326510956, + "grad_norm": 0.3501754701137543, + "learning_rate": 1e-05, + "loss": 0.4184, + "step": 2646 + }, + { + "epoch": 0.7286987031749854, + "grad_norm": 0.35315144062042236, + "learning_rate": 1e-05, + "loss": 0.4236, + "step": 2648 + }, + { + "epoch": 0.7292490798390149, + "grad_norm": 0.36585912108421326, + "learning_rate": 1e-05, + "loss": 0.4205, + "step": 2650 + }, + { + "epoch": 0.7297994565030442, + "grad_norm": 0.3684290051460266, + "learning_rate": 1e-05, + "loss": 0.4366, + "step": 2652 + }, + { + "epoch": 0.7303498331670737, + "grad_norm": 0.3628571927547455, + "learning_rate": 1e-05, + "loss": 0.4205, + "step": 2654 + }, + { + "epoch": 0.7309002098311032, + "grad_norm": 0.36779502034187317, + "learning_rate": 1e-05, + "loss": 0.4338, + "step": 2656 + }, + { + "epoch": 0.7314505864951326, + "grad_norm": 0.3522249162197113, + "learning_rate": 1e-05, + "loss": 0.4268, + "step": 2658 + }, + { + "epoch": 0.7320009631591621, + "grad_norm": 0.3840633034706116, + "learning_rate": 1e-05, + "loss": 0.425, + "step": 2660 + }, + { + "epoch": 0.7325513398231915, + "grad_norm": 0.3498011529445648, + "learning_rate": 1e-05, + "loss": 0.4269, + "step": 2662 + }, + { + "epoch": 0.7331017164872209, + "grad_norm": 0.36151036620140076, + "learning_rate": 1e-05, + "loss": 0.4215, + "step": 2664 + }, + { + "epoch": 0.7336520931512504, + "grad_norm": 0.37008973956108093, + "learning_rate": 1e-05, + "loss": 0.4468, + "step": 2666 + }, + { + "epoch": 0.7342024698152798, + "grad_norm": 0.3440816104412079, + "learning_rate": 1e-05, + "loss": 0.4349, + "step": 2668 + }, + { + "epoch": 0.7347528464793093, + "grad_norm": 0.3912747800350189, + "learning_rate": 1e-05, + "loss": 0.4188, + "step": 2670 + }, + { + "epoch": 0.7353032231433387, + "grad_norm": 0.3472096025943756, + "learning_rate": 1e-05, + "loss": 0.4344, + "step": 2672 + }, + { + "epoch": 0.7358535998073682, + "grad_norm": 0.3477676510810852, + "learning_rate": 1e-05, + "loss": 0.4226, + "step": 2674 + }, + { + "epoch": 0.7364039764713977, + "grad_norm": 0.3726285696029663, + "learning_rate": 1e-05, + "loss": 0.4263, + "step": 2676 + }, + { + "epoch": 0.736954353135427, + "grad_norm": 0.3610732853412628, + "learning_rate": 1e-05, + "loss": 0.4272, + "step": 2678 + }, + { + "epoch": 0.7375047297994565, + "grad_norm": 0.35711386799812317, + "learning_rate": 1e-05, + "loss": 0.4356, + "step": 2680 + }, + { + "epoch": 0.7380551064634859, + "grad_norm": 0.36050212383270264, + "learning_rate": 1e-05, + "loss": 0.437, + "step": 2682 + }, + { + "epoch": 0.7386054831275154, + "grad_norm": 0.33842894434928894, + "learning_rate": 1e-05, + "loss": 0.4136, + "step": 2684 + }, + { + "epoch": 0.7391558597915449, + "grad_norm": 0.35878267884254456, + "learning_rate": 1e-05, + "loss": 0.4118, + "step": 2686 + }, + { + "epoch": 0.7397062364555743, + "grad_norm": 0.3504185676574707, + "learning_rate": 1e-05, + "loss": 0.4157, + "step": 2688 + }, + { + "epoch": 0.7402566131196038, + "grad_norm": 0.35226139426231384, + "learning_rate": 1e-05, + "loss": 0.4194, + "step": 2690 + }, + { + "epoch": 0.7408069897836331, + "grad_norm": 0.3720513880252838, + "learning_rate": 1e-05, + "loss": 0.4225, + "step": 2692 + }, + { + "epoch": 0.7413573664476626, + "grad_norm": 0.3444679081439972, + "learning_rate": 1e-05, + "loss": 0.433, + "step": 2694 + }, + { + "epoch": 0.741907743111692, + "grad_norm": 0.3685862421989441, + "learning_rate": 1e-05, + "loss": 0.4139, + "step": 2696 + }, + { + "epoch": 0.7424581197757215, + "grad_norm": 0.36269327998161316, + "learning_rate": 1e-05, + "loss": 0.4277, + "step": 2698 + }, + { + "epoch": 0.743008496439751, + "grad_norm": 0.36458590626716614, + "learning_rate": 1e-05, + "loss": 0.4217, + "step": 2700 + }, + { + "epoch": 0.7435588731037804, + "grad_norm": 0.3453613221645355, + "learning_rate": 1e-05, + "loss": 0.4174, + "step": 2702 + }, + { + "epoch": 0.7441092497678099, + "grad_norm": 0.3562467098236084, + "learning_rate": 1e-05, + "loss": 0.4313, + "step": 2704 + }, + { + "epoch": 0.7446596264318392, + "grad_norm": 0.3774909973144531, + "learning_rate": 1e-05, + "loss": 0.432, + "step": 2706 + }, + { + "epoch": 0.7452100030958687, + "grad_norm": 0.3668104112148285, + "learning_rate": 1e-05, + "loss": 0.4236, + "step": 2708 + }, + { + "epoch": 0.7457603797598982, + "grad_norm": 0.38669878244400024, + "learning_rate": 1e-05, + "loss": 0.4432, + "step": 2710 + }, + { + "epoch": 0.7463107564239276, + "grad_norm": 0.3985156714916229, + "learning_rate": 1e-05, + "loss": 0.4422, + "step": 2712 + }, + { + "epoch": 0.7468611330879571, + "grad_norm": 0.3647630512714386, + "learning_rate": 1e-05, + "loss": 0.4273, + "step": 2714 + }, + { + "epoch": 0.7474115097519866, + "grad_norm": 0.37027841806411743, + "learning_rate": 1e-05, + "loss": 0.4166, + "step": 2716 + }, + { + "epoch": 0.7479618864160159, + "grad_norm": 0.3770820200443268, + "learning_rate": 1e-05, + "loss": 0.4461, + "step": 2718 + }, + { + "epoch": 0.7485122630800454, + "grad_norm": 0.35209086537361145, + "learning_rate": 1e-05, + "loss": 0.4473, + "step": 2720 + }, + { + "epoch": 0.7490626397440748, + "grad_norm": 0.38394030928611755, + "learning_rate": 1e-05, + "loss": 0.4353, + "step": 2722 + }, + { + "epoch": 0.7496130164081043, + "grad_norm": 0.3524518311023712, + "learning_rate": 1e-05, + "loss": 0.4277, + "step": 2724 + }, + { + "epoch": 0.7501633930721338, + "grad_norm": 0.35822972655296326, + "learning_rate": 1e-05, + "loss": 0.4277, + "step": 2726 + }, + { + "epoch": 0.7507137697361632, + "grad_norm": 0.3409929573535919, + "learning_rate": 1e-05, + "loss": 0.4172, + "step": 2728 + }, + { + "epoch": 0.7512641464001927, + "grad_norm": 0.3534572422504425, + "learning_rate": 1e-05, + "loss": 0.431, + "step": 2730 + }, + { + "epoch": 0.751814523064222, + "grad_norm": 0.3565024733543396, + "learning_rate": 1e-05, + "loss": 0.4297, + "step": 2732 + }, + { + "epoch": 0.7523648997282515, + "grad_norm": 0.3499157130718231, + "learning_rate": 1e-05, + "loss": 0.4131, + "step": 2734 + }, + { + "epoch": 0.752915276392281, + "grad_norm": 0.37271568179130554, + "learning_rate": 1e-05, + "loss": 0.4224, + "step": 2736 + }, + { + "epoch": 0.7534656530563104, + "grad_norm": 0.38281935453414917, + "learning_rate": 1e-05, + "loss": 0.4366, + "step": 2738 + }, + { + "epoch": 0.7540160297203399, + "grad_norm": 0.35982009768486023, + "learning_rate": 1e-05, + "loss": 0.4384, + "step": 2740 + }, + { + "epoch": 0.7545664063843693, + "grad_norm": 0.3618968427181244, + "learning_rate": 1e-05, + "loss": 0.4484, + "step": 2742 + }, + { + "epoch": 0.7551167830483988, + "grad_norm": 0.35112181305885315, + "learning_rate": 1e-05, + "loss": 0.4132, + "step": 2744 + }, + { + "epoch": 0.7556671597124281, + "grad_norm": 0.35898518562316895, + "learning_rate": 1e-05, + "loss": 0.4234, + "step": 2746 + }, + { + "epoch": 0.7562175363764576, + "grad_norm": 0.36049455404281616, + "learning_rate": 1e-05, + "loss": 0.4254, + "step": 2748 + }, + { + "epoch": 0.7567679130404871, + "grad_norm": 0.3698630630970001, + "learning_rate": 1e-05, + "loss": 0.4387, + "step": 2750 + }, + { + "epoch": 0.7573182897045165, + "grad_norm": 0.36196333169937134, + "learning_rate": 1e-05, + "loss": 0.4242, + "step": 2752 + }, + { + "epoch": 0.757868666368546, + "grad_norm": 0.3553547263145447, + "learning_rate": 1e-05, + "loss": 0.4332, + "step": 2754 + }, + { + "epoch": 0.7584190430325755, + "grad_norm": 0.36536121368408203, + "learning_rate": 1e-05, + "loss": 0.4123, + "step": 2756 + }, + { + "epoch": 0.7589694196966049, + "grad_norm": 0.3394269049167633, + "learning_rate": 1e-05, + "loss": 0.4115, + "step": 2758 + }, + { + "epoch": 0.7595197963606343, + "grad_norm": 0.35857659578323364, + "learning_rate": 1e-05, + "loss": 0.4174, + "step": 2760 + }, + { + "epoch": 0.7600701730246637, + "grad_norm": 0.3676673173904419, + "learning_rate": 1e-05, + "loss": 0.4334, + "step": 2762 + }, + { + "epoch": 0.7606205496886932, + "grad_norm": 0.35949233174324036, + "learning_rate": 1e-05, + "loss": 0.4345, + "step": 2764 + }, + { + "epoch": 0.7611709263527227, + "grad_norm": 0.368569940328598, + "learning_rate": 1e-05, + "loss": 0.4241, + "step": 2766 + }, + { + "epoch": 0.7617213030167521, + "grad_norm": 0.37473535537719727, + "learning_rate": 1e-05, + "loss": 0.4454, + "step": 2768 + }, + { + "epoch": 0.7622716796807816, + "grad_norm": 0.34766483306884766, + "learning_rate": 1e-05, + "loss": 0.4193, + "step": 2770 + }, + { + "epoch": 0.7628220563448109, + "grad_norm": 0.3594741225242615, + "learning_rate": 1e-05, + "loss": 0.4265, + "step": 2772 + }, + { + "epoch": 0.7633724330088404, + "grad_norm": 0.35876014828681946, + "learning_rate": 1e-05, + "loss": 0.4401, + "step": 2774 + }, + { + "epoch": 0.7639228096728699, + "grad_norm": 0.3698675036430359, + "learning_rate": 1e-05, + "loss": 0.4301, + "step": 2776 + }, + { + "epoch": 0.7644731863368993, + "grad_norm": 0.3890196979045868, + "learning_rate": 1e-05, + "loss": 0.4312, + "step": 2778 + }, + { + "epoch": 0.7650235630009288, + "grad_norm": 0.3495800793170929, + "learning_rate": 1e-05, + "loss": 0.4235, + "step": 2780 + }, + { + "epoch": 0.7655739396649582, + "grad_norm": 0.3536211848258972, + "learning_rate": 1e-05, + "loss": 0.4319, + "step": 2782 + }, + { + "epoch": 0.7661243163289877, + "grad_norm": 0.35744360089302063, + "learning_rate": 1e-05, + "loss": 0.419, + "step": 2784 + }, + { + "epoch": 0.766674692993017, + "grad_norm": 0.35292670130729675, + "learning_rate": 1e-05, + "loss": 0.4428, + "step": 2786 + }, + { + "epoch": 0.7672250696570465, + "grad_norm": 0.32827427983283997, + "learning_rate": 1e-05, + "loss": 0.4175, + "step": 2788 + }, + { + "epoch": 0.767775446321076, + "grad_norm": 0.3385542929172516, + "learning_rate": 1e-05, + "loss": 0.4288, + "step": 2790 + }, + { + "epoch": 0.7683258229851054, + "grad_norm": 0.3474958539009094, + "learning_rate": 1e-05, + "loss": 0.4424, + "step": 2792 + }, + { + "epoch": 0.7688761996491349, + "grad_norm": 0.3551865816116333, + "learning_rate": 1e-05, + "loss": 0.4351, + "step": 2794 + }, + { + "epoch": 0.7694265763131644, + "grad_norm": 0.3616306781768799, + "learning_rate": 1e-05, + "loss": 0.4481, + "step": 2796 + }, + { + "epoch": 0.7699769529771938, + "grad_norm": 0.36132022738456726, + "learning_rate": 1e-05, + "loss": 0.4128, + "step": 2798 + }, + { + "epoch": 0.7705273296412232, + "grad_norm": 0.3580198585987091, + "learning_rate": 1e-05, + "loss": 0.4242, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_merge_loss": 0.3696165680885315, + "eval_merge_runtime": 600.0202, + "eval_merge_samples_per_second": 56.251, + "eval_merge_steps_per_second": 2.345, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_new_aug_datas_filtered.json_loss": 0.49126043915748596, + "eval_new_aug_datas_filtered.json_runtime": 10.3252, + "eval_new_aug_datas_filtered.json_samples_per_second": 74.285, + "eval_new_aug_datas_filtered.json_steps_per_second": 3.099, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_sharegpt_gpt4.json_loss": 0.7416729927062988, + "eval_sharegpt_gpt4.json_runtime": 31.6069, + "eval_sharegpt_gpt4.json_samples_per_second": 58.88, + "eval_sharegpt_gpt4.json_steps_per_second": 2.468, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_Table_GPT.json_loss": 0.04911120608448982, + "eval_Table_GPT.json_runtime": 24.9282, + "eval_Table_GPT.json_samples_per_second": 83.961, + "eval_Table_GPT.json_steps_per_second": 3.53, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_gpt_4o_200k.json_loss": 0.7679291367530823, + "eval_gpt_4o_200k.json_runtime": 48.4021, + "eval_gpt_4o_200k.json_samples_per_second": 129.767, + "eval_gpt_4o_200k.json_steps_per_second": 5.413, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_multi_turn_datas.json_loss": 0.2913420498371124, + "eval_multi_turn_datas.json_runtime": 75.4573, + "eval_multi_turn_datas.json_samples_per_second": 53.037, + "eval_multi_turn_datas.json_steps_per_second": 2.213, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_table_python_code_datas.json_loss": 0.25055599212646484, + "eval_table_python_code_datas.json_runtime": 43.009, + "eval_table_python_code_datas.json_samples_per_second": 50.199, + "eval_table_python_code_datas.json_steps_per_second": 2.093, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_tabular_llm_data.json_loss": 0.07946833223104477, + "eval_tabular_llm_data.json_runtime": 8.5236, + "eval_tabular_llm_data.json_samples_per_second": 28.861, + "eval_tabular_llm_data.json_steps_per_second": 1.291, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_python_code_critic_21k.json_loss": 0.5505719184875488, + "eval_python_code_critic_21k.json_runtime": 3.2237, + "eval_python_code_critic_21k.json_samples_per_second": 185.192, + "eval_python_code_critic_21k.json_steps_per_second": 7.755, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_all_merge_table_dataset.json_loss": 0.07032839208841324, + "eval_all_merge_table_dataset.json_runtime": 23.2519, + "eval_all_merge_table_dataset.json_samples_per_second": 30.621, + "eval_all_merge_table_dataset.json_steps_per_second": 1.29, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_code_feedback_multi_turn.json_loss": 0.5668665766716003, + "eval_code_feedback_multi_turn.json_runtime": 32.3765, + "eval_code_feedback_multi_turn.json_samples_per_second": 67.981, + "eval_code_feedback_multi_turn.json_steps_per_second": 2.842, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_ultrainteract_sft.json_loss": 0.405385285615921, + "eval_ultrainteract_sft.json_runtime": 8.6576, + "eval_ultrainteract_sft.json_samples_per_second": 168.176, + "eval_ultrainteract_sft.json_steps_per_second": 7.046, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_synthetic_text_to_sql.json_loss": 0.0894596055150032, + "eval_synthetic_text_to_sql.json_runtime": 0.1263, + "eval_synthetic_text_to_sql.json_samples_per_second": 269.263, + "eval_synthetic_text_to_sql.json_steps_per_second": 15.839, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_sft_react_sql_datas.json_loss": 0.6155156493186951, + "eval_sft_react_sql_datas.json_runtime": 7.8457, + "eval_sft_react_sql_datas.json_samples_per_second": 40.022, + "eval_sft_react_sql_datas.json_steps_per_second": 1.784, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_all_merge_code.json_loss": 0.2757679224014282, + "eval_all_merge_code.json_runtime": 0.3332, + "eval_all_merge_code.json_samples_per_second": 189.076, + "eval_all_merge_code.json_steps_per_second": 9.004, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_magpie_datas.json_loss": 0.42383918166160583, + "eval_magpie_datas.json_runtime": 2.2093, + "eval_magpie_datas.json_samples_per_second": 77.853, + "eval_magpie_datas.json_steps_per_second": 3.621, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_train_data_for_qwen.json_loss": 0.0028582699596881866, + "eval_train_data_for_qwen.json_runtime": 0.244, + "eval_train_data_for_qwen.json_samples_per_second": 40.988, + "eval_train_data_for_qwen.json_steps_per_second": 4.099, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_alpaca_cleaned.json_loss": 0.9000511169433594, + "eval_alpaca_cleaned.json_runtime": 0.1144, + "eval_alpaca_cleaned.json_samples_per_second": 235.991, + "eval_alpaca_cleaned.json_steps_per_second": 17.481, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_agent_instruct.json_loss": 0.21006985008716583, + "eval_agent_instruct.json_runtime": 0.5133, + "eval_agent_instruct.json_samples_per_second": 93.518, + "eval_agent_instruct.json_steps_per_second": 3.897, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_MathInstruct.json_loss": 0.19836944341659546, + "eval_MathInstruct.json_runtime": 0.3623, + "eval_MathInstruct.json_samples_per_second": 157.336, + "eval_MathInstruct.json_steps_per_second": 8.281, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_tested_143k_python_alpaca.json_loss": 0.44593295454978943, + "eval_tested_143k_python_alpaca.json_runtime": 0.303, + "eval_tested_143k_python_alpaca.json_samples_per_second": 112.196, + "eval_tested_143k_python_alpaca.json_steps_per_second": 6.6, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_xlam_function_calling_60k.json_loss": 0.0066245682537555695, + "eval_xlam_function_calling_60k.json_runtime": 0.1016, + "eval_xlam_function_calling_60k.json_samples_per_second": 226.385, + "eval_xlam_function_calling_60k.json_steps_per_second": 9.843, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_alpaca_data_gpt4_chinese.json_loss": 1.5253314971923828, + "eval_alpaca_data_gpt4_chinese.json_runtime": 0.052, + "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 307.853, + "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.241, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_alpaca_gpt4_zh.json_loss": 0.9524829983711243, + "eval_alpaca_gpt4_zh.json_runtime": 0.0499, + "eval_alpaca_gpt4_zh.json_samples_per_second": 220.602, + "eval_alpaca_gpt4_zh.json_steps_per_second": 20.055, + "step": 2800 + }, + { + "epoch": 0.7705273296412232, + "eval_codefeedback_filtered_instruction.json_loss": 0.5769651532173157, + "eval_codefeedback_filtered_instruction.json_runtime": 0.4873, + "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.047, + "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.052, + "step": 2800 + }, + { + "epoch": 0.7710777063052526, + "grad_norm": 0.3490790128707886, + "learning_rate": 1e-05, + "loss": 0.4279, + "step": 2802 + }, + { + "epoch": 0.7716280829692821, + "grad_norm": 0.39200064539909363, + "learning_rate": 1e-05, + "loss": 0.4419, + "step": 2804 + }, + { + "epoch": 0.7721784596333116, + "grad_norm": 0.36754128336906433, + "learning_rate": 1e-05, + "loss": 0.4298, + "step": 2806 + }, + { + "epoch": 0.772728836297341, + "grad_norm": 0.3482655882835388, + "learning_rate": 1e-05, + "loss": 0.4249, + "step": 2808 + }, + { + "epoch": 0.7732792129613705, + "grad_norm": 0.35949841141700745, + "learning_rate": 1e-05, + "loss": 0.4245, + "step": 2810 + }, + { + "epoch": 0.7738295896253999, + "grad_norm": 0.3631410598754883, + "learning_rate": 1e-05, + "loss": 0.4221, + "step": 2812 + }, + { + "epoch": 0.7743799662894293, + "grad_norm": 0.3531825542449951, + "learning_rate": 1e-05, + "loss": 0.415, + "step": 2814 + }, + { + "epoch": 0.7749303429534588, + "grad_norm": 0.3741169571876526, + "learning_rate": 1e-05, + "loss": 0.421, + "step": 2816 + }, + { + "epoch": 0.7754807196174882, + "grad_norm": 0.3431030511856079, + "learning_rate": 1e-05, + "loss": 0.4082, + "step": 2818 + }, + { + "epoch": 0.7760310962815177, + "grad_norm": 0.35572293400764465, + "learning_rate": 1e-05, + "loss": 0.4279, + "step": 2820 + }, + { + "epoch": 0.7765814729455471, + "grad_norm": 0.33715927600860596, + "learning_rate": 1e-05, + "loss": 0.4217, + "step": 2822 + }, + { + "epoch": 0.7771318496095766, + "grad_norm": 0.3827720582485199, + "learning_rate": 1e-05, + "loss": 0.4195, + "step": 2824 + }, + { + "epoch": 0.777682226273606, + "grad_norm": 0.34325775504112244, + "learning_rate": 1e-05, + "loss": 0.4359, + "step": 2826 + }, + { + "epoch": 0.7782326029376354, + "grad_norm": 0.34917858242988586, + "learning_rate": 1e-05, + "loss": 0.4165, + "step": 2828 + }, + { + "epoch": 0.7787829796016649, + "grad_norm": 0.3705228865146637, + "learning_rate": 1e-05, + "loss": 0.4234, + "step": 2830 + }, + { + "epoch": 0.7793333562656943, + "grad_norm": 0.36879298090934753, + "learning_rate": 1e-05, + "loss": 0.4173, + "step": 2832 + }, + { + "epoch": 0.7798837329297238, + "grad_norm": 0.35160768032073975, + "learning_rate": 1e-05, + "loss": 0.427, + "step": 2834 + }, + { + "epoch": 0.7804341095937533, + "grad_norm": 0.35639581084251404, + "learning_rate": 1e-05, + "loss": 0.4342, + "step": 2836 + }, + { + "epoch": 0.7809844862577827, + "grad_norm": 0.3821897804737091, + "learning_rate": 1e-05, + "loss": 0.4143, + "step": 2838 + }, + { + "epoch": 0.7815348629218121, + "grad_norm": 0.35575130581855774, + "learning_rate": 1e-05, + "loss": 0.4052, + "step": 2840 + }, + { + "epoch": 0.7820852395858415, + "grad_norm": 0.367026150226593, + "learning_rate": 1e-05, + "loss": 0.4507, + "step": 2842 + }, + { + "epoch": 0.782635616249871, + "grad_norm": 0.35660848021507263, + "learning_rate": 1e-05, + "loss": 0.4112, + "step": 2844 + }, + { + "epoch": 0.7831859929139005, + "grad_norm": 0.3623476028442383, + "learning_rate": 1e-05, + "loss": 0.4298, + "step": 2846 + }, + { + "epoch": 0.7837363695779299, + "grad_norm": 0.36522987484931946, + "learning_rate": 1e-05, + "loss": 0.4197, + "step": 2848 + }, + { + "epoch": 0.7842867462419594, + "grad_norm": 0.349153608083725, + "learning_rate": 1e-05, + "loss": 0.4179, + "step": 2850 + }, + { + "epoch": 0.7848371229059888, + "grad_norm": 0.3868444263935089, + "learning_rate": 1e-05, + "loss": 0.4309, + "step": 2852 + }, + { + "epoch": 0.7853874995700182, + "grad_norm": 0.3388199210166931, + "learning_rate": 1e-05, + "loss": 0.4255, + "step": 2854 + }, + { + "epoch": 0.7859378762340476, + "grad_norm": 0.3848430812358856, + "learning_rate": 1e-05, + "loss": 0.4248, + "step": 2856 + }, + { + "epoch": 0.7864882528980771, + "grad_norm": 0.34994250535964966, + "learning_rate": 1e-05, + "loss": 0.43, + "step": 2858 + }, + { + "epoch": 0.7870386295621066, + "grad_norm": 0.3475828170776367, + "learning_rate": 1e-05, + "loss": 0.4245, + "step": 2860 + }, + { + "epoch": 0.787589006226136, + "grad_norm": 0.3643713593482971, + "learning_rate": 1e-05, + "loss": 0.4285, + "step": 2862 + }, + { + "epoch": 0.7881393828901655, + "grad_norm": 0.3819843828678131, + "learning_rate": 1e-05, + "loss": 0.4264, + "step": 2864 + }, + { + "epoch": 0.7886897595541948, + "grad_norm": 0.3636263608932495, + "learning_rate": 1e-05, + "loss": 0.4354, + "step": 2866 + }, + { + "epoch": 0.7892401362182243, + "grad_norm": 0.35367467999458313, + "learning_rate": 1e-05, + "loss": 0.4219, + "step": 2868 + }, + { + "epoch": 0.7897905128822538, + "grad_norm": 0.33511704206466675, + "learning_rate": 1e-05, + "loss": 0.427, + "step": 2870 + }, + { + "epoch": 0.7903408895462832, + "grad_norm": 0.3727225363254547, + "learning_rate": 1e-05, + "loss": 0.4325, + "step": 2872 + }, + { + "epoch": 0.7908912662103127, + "grad_norm": 0.35963478684425354, + "learning_rate": 1e-05, + "loss": 0.4331, + "step": 2874 + }, + { + "epoch": 0.7914416428743422, + "grad_norm": 0.3680688440799713, + "learning_rate": 1e-05, + "loss": 0.426, + "step": 2876 + }, + { + "epoch": 0.7919920195383716, + "grad_norm": 0.3594858646392822, + "learning_rate": 1e-05, + "loss": 0.4251, + "step": 2878 + }, + { + "epoch": 0.792542396202401, + "grad_norm": 0.3666832745075226, + "learning_rate": 1e-05, + "loss": 0.4148, + "step": 2880 + }, + { + "epoch": 0.7930927728664304, + "grad_norm": 0.3594750761985779, + "learning_rate": 1e-05, + "loss": 0.424, + "step": 2882 + }, + { + "epoch": 0.7936431495304599, + "grad_norm": 0.34796181321144104, + "learning_rate": 1e-05, + "loss": 0.4188, + "step": 2884 + }, + { + "epoch": 0.7941935261944894, + "grad_norm": 0.3670448958873749, + "learning_rate": 1e-05, + "loss": 0.4184, + "step": 2886 + }, + { + "epoch": 0.7947439028585188, + "grad_norm": 0.38206908106803894, + "learning_rate": 1e-05, + "loss": 0.4333, + "step": 2888 + }, + { + "epoch": 0.7952942795225483, + "grad_norm": 0.3671881854534149, + "learning_rate": 1e-05, + "loss": 0.4117, + "step": 2890 + }, + { + "epoch": 0.7958446561865777, + "grad_norm": 0.33647626638412476, + "learning_rate": 1e-05, + "loss": 0.4098, + "step": 2892 + }, + { + "epoch": 0.7963950328506071, + "grad_norm": 0.3504905700683594, + "learning_rate": 1e-05, + "loss": 0.4227, + "step": 2894 + }, + { + "epoch": 0.7969454095146365, + "grad_norm": 0.3571165204048157, + "learning_rate": 1e-05, + "loss": 0.4126, + "step": 2896 + }, + { + "epoch": 0.797495786178666, + "grad_norm": 0.3529278337955475, + "learning_rate": 1e-05, + "loss": 0.4198, + "step": 2898 + }, + { + "epoch": 0.7980461628426955, + "grad_norm": 0.3688133656978607, + "learning_rate": 1e-05, + "loss": 0.443, + "step": 2900 + }, + { + "epoch": 0.7985965395067249, + "grad_norm": 0.37664586305618286, + "learning_rate": 1e-05, + "loss": 0.4345, + "step": 2902 + }, + { + "epoch": 0.7991469161707544, + "grad_norm": 0.37368759512901306, + "learning_rate": 1e-05, + "loss": 0.4202, + "step": 2904 + }, + { + "epoch": 0.7996972928347839, + "grad_norm": 0.3880954384803772, + "learning_rate": 1e-05, + "loss": 0.4234, + "step": 2906 + }, + { + "epoch": 0.8002476694988132, + "grad_norm": 0.34263235330581665, + "learning_rate": 1e-05, + "loss": 0.4177, + "step": 2908 + }, + { + "epoch": 0.8007980461628427, + "grad_norm": 0.37408214807510376, + "learning_rate": 1e-05, + "loss": 0.4366, + "step": 2910 + }, + { + "epoch": 0.8013484228268721, + "grad_norm": 0.35213685035705566, + "learning_rate": 1e-05, + "loss": 0.411, + "step": 2912 + }, + { + "epoch": 0.8018987994909016, + "grad_norm": 0.3545092046260834, + "learning_rate": 1e-05, + "loss": 0.4378, + "step": 2914 + }, + { + "epoch": 0.802449176154931, + "grad_norm": 0.3618670701980591, + "learning_rate": 1e-05, + "loss": 0.4187, + "step": 2916 + }, + { + "epoch": 0.8029995528189605, + "grad_norm": 0.3392831087112427, + "learning_rate": 1e-05, + "loss": 0.4305, + "step": 2918 + }, + { + "epoch": 0.8035499294829899, + "grad_norm": 0.3700800836086273, + "learning_rate": 1e-05, + "loss": 0.4212, + "step": 2920 + }, + { + "epoch": 0.8041003061470193, + "grad_norm": 0.35381945967674255, + "learning_rate": 1e-05, + "loss": 0.416, + "step": 2922 + }, + { + "epoch": 0.8046506828110488, + "grad_norm": 0.3526875972747803, + "learning_rate": 1e-05, + "loss": 0.4287, + "step": 2924 + }, + { + "epoch": 0.8052010594750783, + "grad_norm": 0.3656879663467407, + "learning_rate": 1e-05, + "loss": 0.4196, + "step": 2926 + }, + { + "epoch": 0.8057514361391077, + "grad_norm": 0.3675120174884796, + "learning_rate": 1e-05, + "loss": 0.419, + "step": 2928 + }, + { + "epoch": 0.8063018128031372, + "grad_norm": 0.34032610058784485, + "learning_rate": 1e-05, + "loss": 0.4301, + "step": 2930 + }, + { + "epoch": 0.8068521894671666, + "grad_norm": 0.39022547006607056, + "learning_rate": 1e-05, + "loss": 0.4347, + "step": 2932 + }, + { + "epoch": 0.807402566131196, + "grad_norm": 0.38301143050193787, + "learning_rate": 1e-05, + "loss": 0.4289, + "step": 2934 + }, + { + "epoch": 0.8079529427952254, + "grad_norm": 0.34974217414855957, + "learning_rate": 1e-05, + "loss": 0.4233, + "step": 2936 + }, + { + "epoch": 0.8085033194592549, + "grad_norm": 0.3554193377494812, + "learning_rate": 1e-05, + "loss": 0.4078, + "step": 2938 + }, + { + "epoch": 0.8090536961232844, + "grad_norm": 0.3496205806732178, + "learning_rate": 1e-05, + "loss": 0.4241, + "step": 2940 + }, + { + "epoch": 0.8096040727873138, + "grad_norm": 0.3549167513847351, + "learning_rate": 1e-05, + "loss": 0.4281, + "step": 2942 + }, + { + "epoch": 0.8101544494513433, + "grad_norm": 0.3635149896144867, + "learning_rate": 1e-05, + "loss": 0.4307, + "step": 2944 + }, + { + "epoch": 0.8107048261153728, + "grad_norm": 0.36100322008132935, + "learning_rate": 1e-05, + "loss": 0.4352, + "step": 2946 + }, + { + "epoch": 0.8112552027794021, + "grad_norm": 0.36892169713974, + "learning_rate": 1e-05, + "loss": 0.4245, + "step": 2948 + }, + { + "epoch": 0.8118055794434316, + "grad_norm": 0.34998342394828796, + "learning_rate": 1e-05, + "loss": 0.4214, + "step": 2950 + }, + { + "epoch": 0.812355956107461, + "grad_norm": 0.36382123827934265, + "learning_rate": 1e-05, + "loss": 0.4342, + "step": 2952 + }, + { + "epoch": 0.8129063327714905, + "grad_norm": 0.361068457365036, + "learning_rate": 1e-05, + "loss": 0.4198, + "step": 2954 + }, + { + "epoch": 0.81345670943552, + "grad_norm": 0.36285367608070374, + "learning_rate": 1e-05, + "loss": 0.4297, + "step": 2956 + }, + { + "epoch": 0.8140070860995494, + "grad_norm": 0.3376438319683075, + "learning_rate": 1e-05, + "loss": 0.3999, + "step": 2958 + }, + { + "epoch": 0.8145574627635789, + "grad_norm": 0.35821884870529175, + "learning_rate": 1e-05, + "loss": 0.4283, + "step": 2960 + }, + { + "epoch": 0.8151078394276082, + "grad_norm": 0.37185990810394287, + "learning_rate": 1e-05, + "loss": 0.4221, + "step": 2962 + }, + { + "epoch": 0.8156582160916377, + "grad_norm": 0.3599165380001068, + "learning_rate": 1e-05, + "loss": 0.4222, + "step": 2964 + }, + { + "epoch": 0.8162085927556672, + "grad_norm": 0.3599473237991333, + "learning_rate": 1e-05, + "loss": 0.4211, + "step": 2966 + }, + { + "epoch": 0.8167589694196966, + "grad_norm": 0.3631754219532013, + "learning_rate": 1e-05, + "loss": 0.4273, + "step": 2968 + }, + { + "epoch": 0.8173093460837261, + "grad_norm": 0.34736868739128113, + "learning_rate": 1e-05, + "loss": 0.4175, + "step": 2970 + }, + { + "epoch": 0.8178597227477555, + "grad_norm": 0.34098127484321594, + "learning_rate": 1e-05, + "loss": 0.4297, + "step": 2972 + }, + { + "epoch": 0.8184100994117849, + "grad_norm": 0.3562553822994232, + "learning_rate": 1e-05, + "loss": 0.4342, + "step": 2974 + }, + { + "epoch": 0.8189604760758143, + "grad_norm": 0.3628046214580536, + "learning_rate": 1e-05, + "loss": 0.4146, + "step": 2976 + }, + { + "epoch": 0.8195108527398438, + "grad_norm": 0.33993610739707947, + "learning_rate": 1e-05, + "loss": 0.4228, + "step": 2978 + }, + { + "epoch": 0.8200612294038733, + "grad_norm": 0.35291528701782227, + "learning_rate": 1e-05, + "loss": 0.4179, + "step": 2980 + }, + { + "epoch": 0.8206116060679027, + "grad_norm": 0.3480774164199829, + "learning_rate": 1e-05, + "loss": 0.4099, + "step": 2982 + }, + { + "epoch": 0.8211619827319322, + "grad_norm": 0.36476173996925354, + "learning_rate": 1e-05, + "loss": 0.4153, + "step": 2984 + }, + { + "epoch": 0.8217123593959617, + "grad_norm": 0.3587859869003296, + "learning_rate": 1e-05, + "loss": 0.4334, + "step": 2986 + }, + { + "epoch": 0.822262736059991, + "grad_norm": 0.38419267535209656, + "learning_rate": 1e-05, + "loss": 0.4357, + "step": 2988 + }, + { + "epoch": 0.8228131127240205, + "grad_norm": 0.3496173024177551, + "learning_rate": 1e-05, + "loss": 0.4156, + "step": 2990 + }, + { + "epoch": 0.8233634893880499, + "grad_norm": 0.36481598019599915, + "learning_rate": 1e-05, + "loss": 0.4108, + "step": 2992 + }, + { + "epoch": 0.8239138660520794, + "grad_norm": 0.36568546295166016, + "learning_rate": 1e-05, + "loss": 0.4329, + "step": 2994 + }, + { + "epoch": 0.8244642427161089, + "grad_norm": 0.3675042390823364, + "learning_rate": 1e-05, + "loss": 0.4301, + "step": 2996 + }, + { + "epoch": 0.8250146193801383, + "grad_norm": 0.3355284035205841, + "learning_rate": 1e-05, + "loss": 0.4162, + "step": 2998 + }, + { + "epoch": 0.8255649960441678, + "grad_norm": 0.34280914068222046, + "learning_rate": 1e-05, + "loss": 0.4168, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_merge_loss": 0.3659045696258545, + "eval_merge_runtime": 599.8368, + "eval_merge_samples_per_second": 56.269, + "eval_merge_steps_per_second": 2.346, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_new_aug_datas_filtered.json_loss": 0.48660770058631897, + "eval_new_aug_datas_filtered.json_runtime": 10.3383, + "eval_new_aug_datas_filtered.json_samples_per_second": 74.19, + "eval_new_aug_datas_filtered.json_steps_per_second": 3.095, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_sharegpt_gpt4.json_loss": 0.7358890175819397, + "eval_sharegpt_gpt4.json_runtime": 31.7081, + "eval_sharegpt_gpt4.json_samples_per_second": 58.692, + "eval_sharegpt_gpt4.json_steps_per_second": 2.46, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_Table_GPT.json_loss": 0.045936468988657, + "eval_Table_GPT.json_runtime": 24.9946, + "eval_Table_GPT.json_samples_per_second": 83.738, + "eval_Table_GPT.json_steps_per_second": 3.521, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_gpt_4o_200k.json_loss": 0.7624426484107971, + "eval_gpt_4o_200k.json_runtime": 48.6264, + "eval_gpt_4o_200k.json_samples_per_second": 129.169, + "eval_gpt_4o_200k.json_steps_per_second": 5.388, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_multi_turn_datas.json_loss": 0.2812780439853668, + "eval_multi_turn_datas.json_runtime": 75.8593, + "eval_multi_turn_datas.json_samples_per_second": 52.756, + "eval_multi_turn_datas.json_steps_per_second": 2.201, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_table_python_code_datas.json_loss": 0.24670127034187317, + "eval_table_python_code_datas.json_runtime": 43.2305, + "eval_table_python_code_datas.json_samples_per_second": 49.942, + "eval_table_python_code_datas.json_steps_per_second": 2.082, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_tabular_llm_data.json_loss": 0.08318436145782471, + "eval_tabular_llm_data.json_runtime": 8.561, + "eval_tabular_llm_data.json_samples_per_second": 28.735, + "eval_tabular_llm_data.json_steps_per_second": 1.285, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_python_code_critic_21k.json_loss": 0.5459744930267334, + "eval_python_code_critic_21k.json_runtime": 3.2232, + "eval_python_code_critic_21k.json_samples_per_second": 185.217, + "eval_python_code_critic_21k.json_steps_per_second": 7.756, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_all_merge_table_dataset.json_loss": 0.07661881297826767, + "eval_all_merge_table_dataset.json_runtime": 23.3773, + "eval_all_merge_table_dataset.json_samples_per_second": 30.457, + "eval_all_merge_table_dataset.json_steps_per_second": 1.283, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_code_feedback_multi_turn.json_loss": 0.5640604496002197, + "eval_code_feedback_multi_turn.json_runtime": 32.4865, + "eval_code_feedback_multi_turn.json_samples_per_second": 67.751, + "eval_code_feedback_multi_turn.json_steps_per_second": 2.832, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_ultrainteract_sft.json_loss": 0.40351128578186035, + "eval_ultrainteract_sft.json_runtime": 8.6435, + "eval_ultrainteract_sft.json_samples_per_second": 168.449, + "eval_ultrainteract_sft.json_steps_per_second": 7.057, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_synthetic_text_to_sql.json_loss": 0.09340357035398483, + "eval_synthetic_text_to_sql.json_runtime": 0.1267, + "eval_synthetic_text_to_sql.json_samples_per_second": 268.437, + "eval_synthetic_text_to_sql.json_steps_per_second": 15.79, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_sft_react_sql_datas.json_loss": 0.614182710647583, + "eval_sft_react_sql_datas.json_runtime": 7.8427, + "eval_sft_react_sql_datas.json_samples_per_second": 40.037, + "eval_sft_react_sql_datas.json_steps_per_second": 1.785, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_all_merge_code.json_loss": 0.2747681736946106, + "eval_all_merge_code.json_runtime": 0.3335, + "eval_all_merge_code.json_samples_per_second": 188.917, + "eval_all_merge_code.json_steps_per_second": 8.996, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_magpie_datas.json_loss": 0.42281365394592285, + "eval_magpie_datas.json_runtime": 2.2171, + "eval_magpie_datas.json_samples_per_second": 77.579, + "eval_magpie_datas.json_steps_per_second": 3.608, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_train_data_for_qwen.json_loss": 0.0027365919668227434, + "eval_train_data_for_qwen.json_runtime": 0.2454, + "eval_train_data_for_qwen.json_samples_per_second": 40.756, + "eval_train_data_for_qwen.json_steps_per_second": 4.076, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_alpaca_cleaned.json_loss": 0.9086716175079346, + "eval_alpaca_cleaned.json_runtime": 0.1143, + "eval_alpaca_cleaned.json_samples_per_second": 236.118, + "eval_alpaca_cleaned.json_steps_per_second": 17.49, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_agent_instruct.json_loss": 0.20960307121276855, + "eval_agent_instruct.json_runtime": 0.5163, + "eval_agent_instruct.json_samples_per_second": 92.971, + "eval_agent_instruct.json_steps_per_second": 3.874, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_MathInstruct.json_loss": 0.20019014179706573, + "eval_MathInstruct.json_runtime": 0.3582, + "eval_MathInstruct.json_samples_per_second": 159.116, + "eval_MathInstruct.json_steps_per_second": 8.375, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_tested_143k_python_alpaca.json_loss": 0.44821104407310486, + "eval_tested_143k_python_alpaca.json_runtime": 0.3022, + "eval_tested_143k_python_alpaca.json_samples_per_second": 112.526, + "eval_tested_143k_python_alpaca.json_steps_per_second": 6.619, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_xlam_function_calling_60k.json_loss": 0.008376230485737324, + "eval_xlam_function_calling_60k.json_runtime": 0.1003, + "eval_xlam_function_calling_60k.json_samples_per_second": 229.41, + "eval_xlam_function_calling_60k.json_steps_per_second": 9.974, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_alpaca_data_gpt4_chinese.json_loss": 1.513078212738037, + "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0516, + "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 310.009, + "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.376, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_alpaca_gpt4_zh.json_loss": 0.9633126258850098, + "eval_alpaca_gpt4_zh.json_runtime": 0.0499, + "eval_alpaca_gpt4_zh.json_samples_per_second": 220.561, + "eval_alpaca_gpt4_zh.json_steps_per_second": 20.051, + "step": 3000 + }, + { + "epoch": 0.8255649960441678, + "eval_codefeedback_filtered_instruction.json_loss": 0.5788259506225586, + "eval_codefeedback_filtered_instruction.json_runtime": 0.4854, + "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.202, + "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.06, + "step": 3000 + }, + { + "epoch": 0.8261153727081971, + "grad_norm": 0.35386523604393005, + "learning_rate": 1e-05, + "loss": 0.4235, + "step": 3002 + }, + { + "epoch": 0.8266657493722266, + "grad_norm": 0.35325145721435547, + "learning_rate": 1e-05, + "loss": 0.4296, + "step": 3004 + }, + { + "epoch": 0.827216126036256, + "grad_norm": 0.35455331206321716, + "learning_rate": 1e-05, + "loss": 0.405, + "step": 3006 + }, + { + "epoch": 0.8277665027002855, + "grad_norm": 0.37510380148887634, + "learning_rate": 1e-05, + "loss": 0.4288, + "step": 3008 + }, + { + "epoch": 0.828316879364315, + "grad_norm": 0.356189489364624, + "learning_rate": 1e-05, + "loss": 0.4145, + "step": 3010 + }, + { + "epoch": 0.8288672560283444, + "grad_norm": 0.36097854375839233, + "learning_rate": 1e-05, + "loss": 0.4247, + "step": 3012 + }, + { + "epoch": 0.8294176326923739, + "grad_norm": 0.3489934802055359, + "learning_rate": 1e-05, + "loss": 0.422, + "step": 3014 + }, + { + "epoch": 0.8299680093564032, + "grad_norm": 0.36287152767181396, + "learning_rate": 1e-05, + "loss": 0.4025, + "step": 3016 + }, + { + "epoch": 0.8305183860204327, + "grad_norm": 0.3664880096912384, + "learning_rate": 1e-05, + "loss": 0.4278, + "step": 3018 + }, + { + "epoch": 0.8310687626844622, + "grad_norm": 0.35230088233947754, + "learning_rate": 1e-05, + "loss": 0.4233, + "step": 3020 + }, + { + "epoch": 0.8316191393484916, + "grad_norm": 0.3595122694969177, + "learning_rate": 1e-05, + "loss": 0.4173, + "step": 3022 + }, + { + "epoch": 0.8321695160125211, + "grad_norm": 0.3618360757827759, + "learning_rate": 1e-05, + "loss": 0.4213, + "step": 3024 + }, + { + "epoch": 0.8327198926765506, + "grad_norm": 0.3699500858783722, + "learning_rate": 1e-05, + "loss": 0.4267, + "step": 3026 + }, + { + "epoch": 0.8332702693405799, + "grad_norm": 0.37343189120292664, + "learning_rate": 1e-05, + "loss": 0.4288, + "step": 3028 + }, + { + "epoch": 0.8338206460046094, + "grad_norm": 0.34580445289611816, + "learning_rate": 1e-05, + "loss": 0.4232, + "step": 3030 + }, + { + "epoch": 0.8343710226686388, + "grad_norm": 0.3410281836986542, + "learning_rate": 1e-05, + "loss": 0.4272, + "step": 3032 + }, + { + "epoch": 0.8349213993326683, + "grad_norm": 0.37444379925727844, + "learning_rate": 1e-05, + "loss": 0.4264, + "step": 3034 + }, + { + "epoch": 0.8354717759966978, + "grad_norm": 0.359546959400177, + "learning_rate": 1e-05, + "loss": 0.4243, + "step": 3036 + }, + { + "epoch": 0.8360221526607272, + "grad_norm": 0.3611339032649994, + "learning_rate": 1e-05, + "loss": 0.4243, + "step": 3038 + }, + { + "epoch": 0.8365725293247567, + "grad_norm": 0.3678295612335205, + "learning_rate": 1e-05, + "loss": 0.4089, + "step": 3040 + }, + { + "epoch": 0.837122905988786, + "grad_norm": 0.37094810605049133, + "learning_rate": 1e-05, + "loss": 0.4191, + "step": 3042 + }, + { + "epoch": 0.8376732826528155, + "grad_norm": 0.354481041431427, + "learning_rate": 1e-05, + "loss": 0.4235, + "step": 3044 + }, + { + "epoch": 0.838223659316845, + "grad_norm": 0.3498587906360626, + "learning_rate": 1e-05, + "loss": 0.431, + "step": 3046 + }, + { + "epoch": 0.8387740359808744, + "grad_norm": 0.35214436054229736, + "learning_rate": 1e-05, + "loss": 0.4132, + "step": 3048 + }, + { + "epoch": 0.8393244126449039, + "grad_norm": 0.35119178891181946, + "learning_rate": 1e-05, + "loss": 0.4161, + "step": 3050 + }, + { + "epoch": 0.8398747893089333, + "grad_norm": 0.3671429753303528, + "learning_rate": 1e-05, + "loss": 0.4276, + "step": 3052 + }, + { + "epoch": 0.8404251659729628, + "grad_norm": 0.3626399636268616, + "learning_rate": 1e-05, + "loss": 0.417, + "step": 3054 + }, + { + "epoch": 0.8409755426369921, + "grad_norm": 0.3819148540496826, + "learning_rate": 1e-05, + "loss": 0.4261, + "step": 3056 + }, + { + "epoch": 0.8415259193010216, + "grad_norm": 0.3481554687023163, + "learning_rate": 1e-05, + "loss": 0.4339, + "step": 3058 + }, + { + "epoch": 0.8420762959650511, + "grad_norm": 0.3603340983390808, + "learning_rate": 1e-05, + "loss": 0.406, + "step": 3060 + }, + { + "epoch": 0.8426266726290805, + "grad_norm": 0.3565911650657654, + "learning_rate": 1e-05, + "loss": 0.4245, + "step": 3062 + }, + { + "epoch": 0.84317704929311, + "grad_norm": 0.36305105686187744, + "learning_rate": 1e-05, + "loss": 0.4255, + "step": 3064 + }, + { + "epoch": 0.8437274259571395, + "grad_norm": 0.33078432083129883, + "learning_rate": 1e-05, + "loss": 0.4045, + "step": 3066 + }, + { + "epoch": 0.8442778026211689, + "grad_norm": 0.346562922000885, + "learning_rate": 1e-05, + "loss": 0.4279, + "step": 3068 + }, + { + "epoch": 0.8448281792851983, + "grad_norm": 0.36170172691345215, + "learning_rate": 1e-05, + "loss": 0.4139, + "step": 3070 + }, + { + "epoch": 0.8453785559492277, + "grad_norm": 0.360568106174469, + "learning_rate": 1e-05, + "loss": 0.4276, + "step": 3072 + }, + { + "epoch": 0.8459289326132572, + "grad_norm": 0.38023245334625244, + "learning_rate": 1e-05, + "loss": 0.4317, + "step": 3074 + }, + { + "epoch": 0.8464793092772867, + "grad_norm": 0.344732403755188, + "learning_rate": 1e-05, + "loss": 0.4109, + "step": 3076 + }, + { + "epoch": 0.8470296859413161, + "grad_norm": 0.35157695412635803, + "learning_rate": 1e-05, + "loss": 0.4192, + "step": 3078 + }, + { + "epoch": 0.8475800626053456, + "grad_norm": 0.36455512046813965, + "learning_rate": 1e-05, + "loss": 0.4247, + "step": 3080 + }, + { + "epoch": 0.8481304392693749, + "grad_norm": 0.39768150448799133, + "learning_rate": 1e-05, + "loss": 0.4383, + "step": 3082 + }, + { + "epoch": 0.8486808159334044, + "grad_norm": 0.38052836060523987, + "learning_rate": 1e-05, + "loss": 0.4199, + "step": 3084 + }, + { + "epoch": 0.8492311925974338, + "grad_norm": 0.3625752925872803, + "learning_rate": 1e-05, + "loss": 0.4161, + "step": 3086 + }, + { + "epoch": 0.8497815692614633, + "grad_norm": 0.3708571493625641, + "learning_rate": 1e-05, + "loss": 0.4402, + "step": 3088 + }, + { + "epoch": 0.8503319459254928, + "grad_norm": 0.3581870496273041, + "learning_rate": 1e-05, + "loss": 0.4376, + "step": 3090 + }, + { + "epoch": 0.8508823225895222, + "grad_norm": 0.33589842915534973, + "learning_rate": 1e-05, + "loss": 0.4144, + "step": 3092 + }, + { + "epoch": 0.8514326992535517, + "grad_norm": 0.35838133096694946, + "learning_rate": 1e-05, + "loss": 0.4128, + "step": 3094 + }, + { + "epoch": 0.851983075917581, + "grad_norm": 0.3660927712917328, + "learning_rate": 1e-05, + "loss": 0.42, + "step": 3096 + }, + { + "epoch": 0.8525334525816105, + "grad_norm": 0.3606925904750824, + "learning_rate": 1e-05, + "loss": 0.4288, + "step": 3098 + }, + { + "epoch": 0.85308382924564, + "grad_norm": 0.3437570333480835, + "learning_rate": 1e-05, + "loss": 0.4213, + "step": 3100 + }, + { + "epoch": 0.8536342059096694, + "grad_norm": 0.35351496934890747, + "learning_rate": 1e-05, + "loss": 0.4238, + "step": 3102 + }, + { + "epoch": 0.8541845825736989, + "grad_norm": 0.3595280051231384, + "learning_rate": 1e-05, + "loss": 0.4107, + "step": 3104 + }, + { + "epoch": 0.8547349592377284, + "grad_norm": 0.3546600937843323, + "learning_rate": 1e-05, + "loss": 0.4105, + "step": 3106 + }, + { + "epoch": 0.8552853359017578, + "grad_norm": 0.3654036819934845, + "learning_rate": 1e-05, + "loss": 0.4158, + "step": 3108 + }, + { + "epoch": 0.8558357125657872, + "grad_norm": 0.3742349445819855, + "learning_rate": 1e-05, + "loss": 0.4217, + "step": 3110 + }, + { + "epoch": 0.8563860892298166, + "grad_norm": 0.35527029633522034, + "learning_rate": 1e-05, + "loss": 0.414, + "step": 3112 + }, + { + "epoch": 0.8569364658938461, + "grad_norm": 0.3408162295818329, + "learning_rate": 1e-05, + "loss": 0.4245, + "step": 3114 + }, + { + "epoch": 0.8574868425578756, + "grad_norm": 0.3608722686767578, + "learning_rate": 1e-05, + "loss": 0.4346, + "step": 3116 + }, + { + "epoch": 0.858037219221905, + "grad_norm": 0.36163628101348877, + "learning_rate": 1e-05, + "loss": 0.422, + "step": 3118 + }, + { + "epoch": 0.8585875958859345, + "grad_norm": 0.35417988896369934, + "learning_rate": 1e-05, + "loss": 0.4101, + "step": 3120 + }, + { + "epoch": 0.8591379725499639, + "grad_norm": 0.3626682162284851, + "learning_rate": 1e-05, + "loss": 0.4147, + "step": 3122 + }, + { + "epoch": 0.8596883492139933, + "grad_norm": 0.34313321113586426, + "learning_rate": 1e-05, + "loss": 0.4215, + "step": 3124 + }, + { + "epoch": 0.8602387258780227, + "grad_norm": 0.3839293122291565, + "learning_rate": 1e-05, + "loss": 0.4173, + "step": 3126 + }, + { + "epoch": 0.8607891025420522, + "grad_norm": 0.3548083007335663, + "learning_rate": 1e-05, + "loss": 0.4153, + "step": 3128 + }, + { + "epoch": 0.8613394792060817, + "grad_norm": 0.35141652822494507, + "learning_rate": 1e-05, + "loss": 0.4066, + "step": 3130 + }, + { + "epoch": 0.8618898558701111, + "grad_norm": 0.3777351975440979, + "learning_rate": 1e-05, + "loss": 0.4128, + "step": 3132 + }, + { + "epoch": 0.8624402325341406, + "grad_norm": 0.3580491840839386, + "learning_rate": 1e-05, + "loss": 0.4048, + "step": 3134 + }, + { + "epoch": 0.86299060919817, + "grad_norm": 0.373532772064209, + "learning_rate": 1e-05, + "loss": 0.4173, + "step": 3136 + }, + { + "epoch": 0.8635409858621994, + "grad_norm": 0.35365086793899536, + "learning_rate": 1e-05, + "loss": 0.4076, + "step": 3138 + }, + { + "epoch": 0.8640913625262289, + "grad_norm": 0.3887852728366852, + "learning_rate": 1e-05, + "loss": 0.418, + "step": 3140 + }, + { + "epoch": 0.8646417391902583, + "grad_norm": 0.35862478613853455, + "learning_rate": 1e-05, + "loss": 0.4234, + "step": 3142 + }, + { + "epoch": 0.8651921158542878, + "grad_norm": 0.3472420275211334, + "learning_rate": 1e-05, + "loss": 0.4132, + "step": 3144 + }, + { + "epoch": 0.8657424925183173, + "grad_norm": 0.344862163066864, + "learning_rate": 1e-05, + "loss": 0.41, + "step": 3146 + }, + { + "epoch": 0.8662928691823467, + "grad_norm": 0.35329338908195496, + "learning_rate": 1e-05, + "loss": 0.4152, + "step": 3148 + }, + { + "epoch": 0.8668432458463761, + "grad_norm": 0.3792724907398224, + "learning_rate": 1e-05, + "loss": 0.4307, + "step": 3150 + }, + { + "epoch": 0.8673936225104055, + "grad_norm": 0.3611691892147064, + "learning_rate": 1e-05, + "loss": 0.4166, + "step": 3152 + }, + { + "epoch": 0.867943999174435, + "grad_norm": 0.35675716400146484, + "learning_rate": 1e-05, + "loss": 0.4309, + "step": 3154 + }, + { + "epoch": 0.8684943758384644, + "grad_norm": 0.37591055035591125, + "learning_rate": 1e-05, + "loss": 0.4174, + "step": 3156 + }, + { + "epoch": 0.8690447525024939, + "grad_norm": 0.34695202112197876, + "learning_rate": 1e-05, + "loss": 0.4067, + "step": 3158 + }, + { + "epoch": 0.8695951291665234, + "grad_norm": 0.36810246109962463, + "learning_rate": 1e-05, + "loss": 0.4236, + "step": 3160 + }, + { + "epoch": 0.8701455058305528, + "grad_norm": 0.3910383880138397, + "learning_rate": 1e-05, + "loss": 0.4344, + "step": 3162 + }, + { + "epoch": 0.8706958824945822, + "grad_norm": 0.3465210497379303, + "learning_rate": 1e-05, + "loss": 0.4156, + "step": 3164 + }, + { + "epoch": 0.8712462591586116, + "grad_norm": 0.39839833974838257, + "learning_rate": 1e-05, + "loss": 0.417, + "step": 3166 + }, + { + "epoch": 0.8717966358226411, + "grad_norm": 0.33419859409332275, + "learning_rate": 1e-05, + "loss": 0.4131, + "step": 3168 + }, + { + "epoch": 0.8723470124866706, + "grad_norm": 0.3657875955104828, + "learning_rate": 1e-05, + "loss": 0.4243, + "step": 3170 + }, + { + "epoch": 0.8728973891507, + "grad_norm": 0.35600635409355164, + "learning_rate": 1e-05, + "loss": 0.4251, + "step": 3172 + }, + { + "epoch": 0.8734477658147295, + "grad_norm": 0.3642902374267578, + "learning_rate": 1e-05, + "loss": 0.4334, + "step": 3174 + }, + { + "epoch": 0.873998142478759, + "grad_norm": 0.35452064871788025, + "learning_rate": 1e-05, + "loss": 0.4244, + "step": 3176 + }, + { + "epoch": 0.8745485191427883, + "grad_norm": 0.372953861951828, + "learning_rate": 1e-05, + "loss": 0.421, + "step": 3178 + }, + { + "epoch": 0.8750988958068178, + "grad_norm": 0.3428981900215149, + "learning_rate": 1e-05, + "loss": 0.4172, + "step": 3180 + }, + { + "epoch": 0.8756492724708472, + "grad_norm": 0.36314892768859863, + "learning_rate": 1e-05, + "loss": 0.4008, + "step": 3182 + }, + { + "epoch": 0.8761996491348767, + "grad_norm": 0.35167455673217773, + "learning_rate": 1e-05, + "loss": 0.4114, + "step": 3184 + }, + { + "epoch": 0.8767500257989062, + "grad_norm": 0.3496149778366089, + "learning_rate": 1e-05, + "loss": 0.4053, + "step": 3186 + }, + { + "epoch": 0.8773004024629356, + "grad_norm": 0.351510226726532, + "learning_rate": 1e-05, + "loss": 0.4173, + "step": 3188 + }, + { + "epoch": 0.877850779126965, + "grad_norm": 0.35172203183174133, + "learning_rate": 1e-05, + "loss": 0.4202, + "step": 3190 + }, + { + "epoch": 0.8784011557909944, + "grad_norm": 0.36200663447380066, + "learning_rate": 1e-05, + "loss": 0.3987, + "step": 3192 + }, + { + "epoch": 0.8789515324550239, + "grad_norm": 0.36070528626441956, + "learning_rate": 1e-05, + "loss": 0.4202, + "step": 3194 + }, + { + "epoch": 0.8795019091190533, + "grad_norm": 0.37506040930747986, + "learning_rate": 1e-05, + "loss": 0.4125, + "step": 3196 + }, + { + "epoch": 0.8800522857830828, + "grad_norm": 0.3433153033256531, + "learning_rate": 1e-05, + "loss": 0.4185, + "step": 3198 + }, + { + "epoch": 0.8806026624471123, + "grad_norm": 0.3672421872615814, + "learning_rate": 1e-05, + "loss": 0.4227, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_merge_loss": 0.36182981729507446, + "eval_merge_runtime": 600.3542, + "eval_merge_samples_per_second": 56.22, + "eval_merge_steps_per_second": 2.344, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_new_aug_datas_filtered.json_loss": 0.4833287298679352, + "eval_new_aug_datas_filtered.json_runtime": 10.3758, + "eval_new_aug_datas_filtered.json_samples_per_second": 73.922, + "eval_new_aug_datas_filtered.json_steps_per_second": 3.084, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_sharegpt_gpt4.json_loss": 0.7305224537849426, + "eval_sharegpt_gpt4.json_runtime": 31.7036, + "eval_sharegpt_gpt4.json_samples_per_second": 58.7, + "eval_sharegpt_gpt4.json_steps_per_second": 2.46, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_Table_GPT.json_loss": 0.04232589527964592, + "eval_Table_GPT.json_runtime": 24.994, + "eval_Table_GPT.json_samples_per_second": 83.74, + "eval_Table_GPT.json_steps_per_second": 3.521, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_gpt_4o_200k.json_loss": 0.7571491003036499, + "eval_gpt_4o_200k.json_runtime": 48.5629, + "eval_gpt_4o_200k.json_samples_per_second": 129.337, + "eval_gpt_4o_200k.json_steps_per_second": 5.395, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_multi_turn_datas.json_loss": 0.2720319330692291, + "eval_multi_turn_datas.json_runtime": 75.6646, + "eval_multi_turn_datas.json_samples_per_second": 52.891, + "eval_multi_turn_datas.json_steps_per_second": 2.207, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_table_python_code_datas.json_loss": 0.24331320822238922, + "eval_table_python_code_datas.json_runtime": 43.0385, + "eval_table_python_code_datas.json_samples_per_second": 50.164, + "eval_table_python_code_datas.json_steps_per_second": 2.091, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_tabular_llm_data.json_loss": 0.0862693339586258, + "eval_tabular_llm_data.json_runtime": 8.5454, + "eval_tabular_llm_data.json_samples_per_second": 28.788, + "eval_tabular_llm_data.json_steps_per_second": 1.287, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_python_code_critic_21k.json_loss": 0.5425785183906555, + "eval_python_code_critic_21k.json_runtime": 3.2194, + "eval_python_code_critic_21k.json_samples_per_second": 185.437, + "eval_python_code_critic_21k.json_steps_per_second": 7.765, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_all_merge_table_dataset.json_loss": 0.06970688700675964, + "eval_all_merge_table_dataset.json_runtime": 23.3201, + "eval_all_merge_table_dataset.json_samples_per_second": 30.532, + "eval_all_merge_table_dataset.json_steps_per_second": 1.286, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_code_feedback_multi_turn.json_loss": 0.5619133114814758, + "eval_code_feedback_multi_turn.json_runtime": 32.4257, + "eval_code_feedback_multi_turn.json_samples_per_second": 67.878, + "eval_code_feedback_multi_turn.json_steps_per_second": 2.837, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_ultrainteract_sft.json_loss": 0.4016903042793274, + "eval_ultrainteract_sft.json_runtime": 8.6472, + "eval_ultrainteract_sft.json_samples_per_second": 168.378, + "eval_ultrainteract_sft.json_steps_per_second": 7.054, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_synthetic_text_to_sql.json_loss": 0.09171026945114136, + "eval_synthetic_text_to_sql.json_runtime": 0.1264, + "eval_synthetic_text_to_sql.json_samples_per_second": 268.914, + "eval_synthetic_text_to_sql.json_steps_per_second": 15.818, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_sft_react_sql_datas.json_loss": 0.6105172038078308, + "eval_sft_react_sql_datas.json_runtime": 7.8425, + "eval_sft_react_sql_datas.json_samples_per_second": 40.038, + "eval_sft_react_sql_datas.json_steps_per_second": 1.785, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_all_merge_code.json_loss": 0.264506459236145, + "eval_all_merge_code.json_runtime": 0.3347, + "eval_all_merge_code.json_samples_per_second": 188.208, + "eval_all_merge_code.json_steps_per_second": 8.962, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_magpie_datas.json_loss": 0.4236694872379303, + "eval_magpie_datas.json_runtime": 2.213, + "eval_magpie_datas.json_samples_per_second": 77.723, + "eval_magpie_datas.json_steps_per_second": 3.615, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_train_data_for_qwen.json_loss": 0.0027615067083388567, + "eval_train_data_for_qwen.json_runtime": 0.2435, + "eval_train_data_for_qwen.json_samples_per_second": 41.06, + "eval_train_data_for_qwen.json_steps_per_second": 4.106, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_alpaca_cleaned.json_loss": 0.9028782844543457, + "eval_alpaca_cleaned.json_runtime": 0.1145, + "eval_alpaca_cleaned.json_samples_per_second": 235.866, + "eval_alpaca_cleaned.json_steps_per_second": 17.472, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_agent_instruct.json_loss": 0.20418775081634521, + "eval_agent_instruct.json_runtime": 0.5123, + "eval_agent_instruct.json_samples_per_second": 93.693, + "eval_agent_instruct.json_steps_per_second": 3.904, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_MathInstruct.json_loss": 0.2024046629667282, + "eval_MathInstruct.json_runtime": 0.35, + "eval_MathInstruct.json_samples_per_second": 162.858, + "eval_MathInstruct.json_steps_per_second": 8.571, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_tested_143k_python_alpaca.json_loss": 0.4462108612060547, + "eval_tested_143k_python_alpaca.json_runtime": 0.3037, + "eval_tested_143k_python_alpaca.json_samples_per_second": 111.97, + "eval_tested_143k_python_alpaca.json_steps_per_second": 6.586, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_xlam_function_calling_60k.json_loss": 0.008976898156106472, + "eval_xlam_function_calling_60k.json_runtime": 0.1004, + "eval_xlam_function_calling_60k.json_samples_per_second": 229.083, + "eval_xlam_function_calling_60k.json_steps_per_second": 9.96, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_alpaca_data_gpt4_chinese.json_loss": 1.512216567993164, + "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0511, + "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 313.242, + "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.578, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_alpaca_gpt4_zh.json_loss": 0.9716835021972656, + "eval_alpaca_gpt4_zh.json_runtime": 0.0498, + "eval_alpaca_gpt4_zh.json_samples_per_second": 220.769, + "eval_alpaca_gpt4_zh.json_steps_per_second": 20.07, + "step": 3200 + }, + { + "epoch": 0.8806026624471123, + "eval_codefeedback_filtered_instruction.json_loss": 0.5663765072822571, + "eval_codefeedback_filtered_instruction.json_runtime": 0.4857, + "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.176, + "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.059, + "step": 3200 + } + ], + "logging_steps": 2, + "max_steps": 3633, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 200, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.4654685450828094e+20, + "train_batch_size": 3, + "trial_name": null, + "trial_params": null +}