{ "best_metric": 0.0033526704646646976, "best_model_checkpoint": "saves/CADICA_qwenvl_direction_scale4/lora/sft/checkpoint-550", "epoch": 0.43780582024208087, "eval_steps": 50, "global_step": 850, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0025753283543651817, "grad_norm": 14.230540018977528, "learning_rate": 2.9411764705882355e-06, "loss": 1.0954, "num_input_tokens_seen": 49920, "step": 5 }, { "epoch": 0.0051506567087303634, "grad_norm": 12.725175128208063, "learning_rate": 5.882352941176471e-06, "loss": 0.9793, "num_input_tokens_seen": 99840, "step": 10 }, { "epoch": 0.007725985063095545, "grad_norm": 11.908760913870685, "learning_rate": 8.823529411764707e-06, "loss": 1.0964, "num_input_tokens_seen": 149760, "step": 15 }, { "epoch": 0.010301313417460727, "grad_norm": 7.5714179545602835, "learning_rate": 1.1764705882352942e-05, "loss": 0.7079, "num_input_tokens_seen": 199680, "step": 20 }, { "epoch": 0.012876641771825908, "grad_norm": 2.273110551179123, "learning_rate": 1.4705882352941177e-05, "loss": 0.4213, "num_input_tokens_seen": 249600, "step": 25 }, { "epoch": 0.01545197012619109, "grad_norm": 1.4511084067844011, "learning_rate": 1.7647058823529414e-05, "loss": 0.3359, "num_input_tokens_seen": 299520, "step": 30 }, { "epoch": 0.018027298480556272, "grad_norm": 1.7462649262033438, "learning_rate": 2.058823529411765e-05, "loss": 0.3463, "num_input_tokens_seen": 349440, "step": 35 }, { "epoch": 0.020602626834921454, "grad_norm": 1.2143595820366577, "learning_rate": 2.3529411764705884e-05, "loss": 0.28, "num_input_tokens_seen": 399360, "step": 40 }, { "epoch": 0.023177955189286635, "grad_norm": 0.8468719125045373, "learning_rate": 2.647058823529412e-05, "loss": 0.3442, "num_input_tokens_seen": 449280, "step": 45 }, { "epoch": 0.025753283543651816, "grad_norm": 0.9726867932660042, "learning_rate": 2.9411764705882354e-05, "loss": 0.3441, "num_input_tokens_seen": 499200, "step": 50 }, { "epoch": 0.025753283543651816, "eval_loss": 0.33834776282310486, "eval_runtime": 47.4467, "eval_samples_per_second": 1.265, "eval_steps_per_second": 0.316, "num_input_tokens_seen": 499200, "step": 50 }, { "epoch": 0.028328611898016998, "grad_norm": 1.4016556961184263, "learning_rate": 3.235294117647059e-05, "loss": 0.3182, "num_input_tokens_seen": 549120, "step": 55 }, { "epoch": 0.03090394025238218, "grad_norm": 0.6437613769459606, "learning_rate": 3.529411764705883e-05, "loss": 0.3294, "num_input_tokens_seen": 599040, "step": 60 }, { "epoch": 0.03347926860674736, "grad_norm": 0.7389008951321312, "learning_rate": 3.8235294117647055e-05, "loss": 0.3097, "num_input_tokens_seen": 648960, "step": 65 }, { "epoch": 0.036054596961112545, "grad_norm": 0.771553860801019, "learning_rate": 4.11764705882353e-05, "loss": 0.3008, "num_input_tokens_seen": 698880, "step": 70 }, { "epoch": 0.03862992531547772, "grad_norm": 0.6965369148334918, "learning_rate": 4.411764705882353e-05, "loss": 0.3278, "num_input_tokens_seen": 748800, "step": 75 }, { "epoch": 0.04120525366984291, "grad_norm": 0.912943461315541, "learning_rate": 4.705882352941177e-05, "loss": 0.3074, "num_input_tokens_seen": 798720, "step": 80 }, { "epoch": 0.043780582024208085, "grad_norm": 0.8407481737577445, "learning_rate": 5e-05, "loss": 0.3423, "num_input_tokens_seen": 848640, "step": 85 }, { "epoch": 0.04635591037857327, "grad_norm": 0.9112879058417015, "learning_rate": 5.294117647058824e-05, "loss": 0.3008, "num_input_tokens_seen": 898560, "step": 90 }, { "epoch": 0.04893123873293845, "grad_norm": 2.391489040464162, "learning_rate": 5.588235294117647e-05, "loss": 0.2815, "num_input_tokens_seen": 948480, "step": 95 }, { "epoch": 0.05150656708730363, "grad_norm": 2.155211791607199, "learning_rate": 5.882352941176471e-05, "loss": 0.2274, "num_input_tokens_seen": 998400, "step": 100 }, { "epoch": 0.05150656708730363, "eval_loss": 0.18663176894187927, "eval_runtime": 18.9199, "eval_samples_per_second": 3.171, "eval_steps_per_second": 0.793, "num_input_tokens_seen": 998400, "step": 100 }, { "epoch": 0.05408189544166881, "grad_norm": 2.2181531422996716, "learning_rate": 6.176470588235295e-05, "loss": 0.168, "num_input_tokens_seen": 1048320, "step": 105 }, { "epoch": 0.056657223796033995, "grad_norm": 3.1829920225573236, "learning_rate": 6.470588235294118e-05, "loss": 0.0709, "num_input_tokens_seen": 1098240, "step": 110 }, { "epoch": 0.05923255215039917, "grad_norm": 4.337350477588576, "learning_rate": 6.764705882352942e-05, "loss": 0.1609, "num_input_tokens_seen": 1148160, "step": 115 }, { "epoch": 0.06180788050476436, "grad_norm": 2.1010046045637365, "learning_rate": 7.058823529411765e-05, "loss": 0.0354, "num_input_tokens_seen": 1198080, "step": 120 }, { "epoch": 0.06438320885912954, "grad_norm": 2.232308844812103, "learning_rate": 7.352941176470589e-05, "loss": 0.1133, "num_input_tokens_seen": 1248000, "step": 125 }, { "epoch": 0.06695853721349472, "grad_norm": 5.641631090993415, "learning_rate": 7.647058823529411e-05, "loss": 0.0867, "num_input_tokens_seen": 1297920, "step": 130 }, { "epoch": 0.0695338655678599, "grad_norm": 1.5031437609685787, "learning_rate": 7.941176470588235e-05, "loss": 0.1352, "num_input_tokens_seen": 1347840, "step": 135 }, { "epoch": 0.07210919392222509, "grad_norm": 3.2992644431188465, "learning_rate": 8.23529411764706e-05, "loss": 0.101, "num_input_tokens_seen": 1397760, "step": 140 }, { "epoch": 0.07468452227659027, "grad_norm": 3.494236832758233, "learning_rate": 8.529411764705883e-05, "loss": 0.0334, "num_input_tokens_seen": 1447680, "step": 145 }, { "epoch": 0.07725985063095545, "grad_norm": 0.0602113869322109, "learning_rate": 8.823529411764706e-05, "loss": 0.0667, "num_input_tokens_seen": 1497600, "step": 150 }, { "epoch": 0.07725985063095545, "eval_loss": 0.09665286540985107, "eval_runtime": 19.2745, "eval_samples_per_second": 3.113, "eval_steps_per_second": 0.778, "num_input_tokens_seen": 1497600, "step": 150 }, { "epoch": 0.07983517898532062, "grad_norm": 6.096163706748617, "learning_rate": 9.11764705882353e-05, "loss": 0.06, "num_input_tokens_seen": 1547520, "step": 155 }, { "epoch": 0.08241050733968582, "grad_norm": 4.278069142242893, "learning_rate": 9.411764705882353e-05, "loss": 0.0265, "num_input_tokens_seen": 1597440, "step": 160 }, { "epoch": 0.08498583569405099, "grad_norm": 1.7183097652953412, "learning_rate": 9.705882352941177e-05, "loss": 0.1743, "num_input_tokens_seen": 1647360, "step": 165 }, { "epoch": 0.08756116404841617, "grad_norm": 4.95401899568707, "learning_rate": 0.0001, "loss": 0.051, "num_input_tokens_seen": 1697280, "step": 170 }, { "epoch": 0.09013649240278135, "grad_norm": 0.34558354886099124, "learning_rate": 9.999940874631277e-05, "loss": 0.0584, "num_input_tokens_seen": 1747200, "step": 175 }, { "epoch": 0.09271182075714654, "grad_norm": 3.6326401692458146, "learning_rate": 9.999763499923432e-05, "loss": 0.0704, "num_input_tokens_seen": 1797120, "step": 180 }, { "epoch": 0.09528714911151172, "grad_norm": 0.4999720825488852, "learning_rate": 9.999467880071402e-05, "loss": 0.0278, "num_input_tokens_seen": 1847040, "step": 185 }, { "epoch": 0.0978624774658769, "grad_norm": 0.5211964505880501, "learning_rate": 9.999054022066641e-05, "loss": 0.0862, "num_input_tokens_seen": 1896960, "step": 190 }, { "epoch": 0.10043780582024209, "grad_norm": 0.8767038751832389, "learning_rate": 9.998521935696953e-05, "loss": 0.0565, "num_input_tokens_seen": 1946880, "step": 195 }, { "epoch": 0.10301313417460727, "grad_norm": 0.1848827361202722, "learning_rate": 9.997871633546257e-05, "loss": 0.0459, "num_input_tokens_seen": 1996800, "step": 200 }, { "epoch": 0.10301313417460727, "eval_loss": 0.09957947582006454, "eval_runtime": 19.2651, "eval_samples_per_second": 3.114, "eval_steps_per_second": 0.779, "num_input_tokens_seen": 1996800, "step": 200 }, { "epoch": 0.10558846252897244, "grad_norm": 1.0969393658164421, "learning_rate": 9.997103130994296e-05, "loss": 0.0539, "num_input_tokens_seen": 2046720, "step": 205 }, { "epoch": 0.10816379088333762, "grad_norm": 2.885869194934028, "learning_rate": 9.996216446216267e-05, "loss": 0.0654, "num_input_tokens_seen": 2096640, "step": 210 }, { "epoch": 0.11073911923770281, "grad_norm": 0.5225257245731217, "learning_rate": 9.995211600182397e-05, "loss": 0.0316, "num_input_tokens_seen": 2146560, "step": 215 }, { "epoch": 0.11331444759206799, "grad_norm": 2.1553510734212797, "learning_rate": 9.994088616657444e-05, "loss": 0.1169, "num_input_tokens_seen": 2196480, "step": 220 }, { "epoch": 0.11588977594643317, "grad_norm": 1.1133884703723633, "learning_rate": 9.992847522200133e-05, "loss": 0.0382, "num_input_tokens_seen": 2246400, "step": 225 }, { "epoch": 0.11846510430079835, "grad_norm": 0.8875243341616034, "learning_rate": 9.99148834616253e-05, "loss": 0.0406, "num_input_tokens_seen": 2296320, "step": 230 }, { "epoch": 0.12104043265516354, "grad_norm": 1.81283533812695, "learning_rate": 9.990011120689351e-05, "loss": 0.0182, "num_input_tokens_seen": 2346240, "step": 235 }, { "epoch": 0.12361576100952872, "grad_norm": 3.873083258671571, "learning_rate": 9.988415880717194e-05, "loss": 0.0881, "num_input_tokens_seen": 2396160, "step": 240 }, { "epoch": 0.1261910893638939, "grad_norm": 3.427761103620865, "learning_rate": 9.986702663973722e-05, "loss": 0.0565, "num_input_tokens_seen": 2446080, "step": 245 }, { "epoch": 0.12876641771825909, "grad_norm": 1.531943599765959, "learning_rate": 9.98487151097676e-05, "loss": 0.0805, "num_input_tokens_seen": 2496000, "step": 250 }, { "epoch": 0.12876641771825909, "eval_loss": 0.055875860154628754, "eval_runtime": 19.5106, "eval_samples_per_second": 3.075, "eval_steps_per_second": 0.769, "num_input_tokens_seen": 2496000, "step": 250 }, { "epoch": 0.13134174607262425, "grad_norm": 0.845514751132877, "learning_rate": 9.98292246503335e-05, "loss": 0.0235, "num_input_tokens_seen": 2545920, "step": 255 }, { "epoch": 0.13391707442698944, "grad_norm": 1.7441538057875345, "learning_rate": 9.980855572238714e-05, "loss": 0.0137, "num_input_tokens_seen": 2595840, "step": 260 }, { "epoch": 0.13649240278135463, "grad_norm": 0.6628270146260998, "learning_rate": 9.978670881475172e-05, "loss": 0.0605, "num_input_tokens_seen": 2645760, "step": 265 }, { "epoch": 0.1390677311357198, "grad_norm": 4.49926360488578, "learning_rate": 9.976368444410985e-05, "loss": 0.0585, "num_input_tokens_seen": 2695680, "step": 270 }, { "epoch": 0.141643059490085, "grad_norm": 0.7159498856832776, "learning_rate": 9.973948315499126e-05, "loss": 0.0096, "num_input_tokens_seen": 2745600, "step": 275 }, { "epoch": 0.14421838784445018, "grad_norm": 4.09842502301765, "learning_rate": 9.971410551976002e-05, "loss": 0.1383, "num_input_tokens_seen": 2795520, "step": 280 }, { "epoch": 0.14679371619881534, "grad_norm": 1.0773897808674016, "learning_rate": 9.968755213860094e-05, "loss": 0.049, "num_input_tokens_seen": 2845440, "step": 285 }, { "epoch": 0.14936904455318054, "grad_norm": 1.7224091076306651, "learning_rate": 9.96598236395054e-05, "loss": 0.0336, "num_input_tokens_seen": 2895360, "step": 290 }, { "epoch": 0.1519443729075457, "grad_norm": 0.24158350585209873, "learning_rate": 9.96309206782565e-05, "loss": 0.019, "num_input_tokens_seen": 2945280, "step": 295 }, { "epoch": 0.1545197012619109, "grad_norm": 0.6614271496466305, "learning_rate": 9.960084393841355e-05, "loss": 0.0381, "num_input_tokens_seen": 2995200, "step": 300 }, { "epoch": 0.1545197012619109, "eval_loss": 0.030908752232789993, "eval_runtime": 19.8702, "eval_samples_per_second": 3.02, "eval_steps_per_second": 0.755, "num_input_tokens_seen": 2995200, "step": 300 }, { "epoch": 0.15709502961627608, "grad_norm": 0.2314383428097956, "learning_rate": 9.956959413129585e-05, "loss": 0.0245, "num_input_tokens_seen": 3045120, "step": 305 }, { "epoch": 0.15967035797064125, "grad_norm": 0.2035043692790888, "learning_rate": 9.953717199596598e-05, "loss": 0.0121, "num_input_tokens_seen": 3095040, "step": 310 }, { "epoch": 0.16224568632500644, "grad_norm": 0.014152925203425593, "learning_rate": 9.95035782992122e-05, "loss": 0.054, "num_input_tokens_seen": 3144960, "step": 315 }, { "epoch": 0.16482101467937163, "grad_norm": 1.0162363271291994, "learning_rate": 9.94688138355304e-05, "loss": 0.0849, "num_input_tokens_seen": 3194880, "step": 320 }, { "epoch": 0.1673963430337368, "grad_norm": 3.92714662464831, "learning_rate": 9.943287942710527e-05, "loss": 0.0503, "num_input_tokens_seen": 3244800, "step": 325 }, { "epoch": 0.16997167138810199, "grad_norm": 0.47450325213422423, "learning_rate": 9.939577592379088e-05, "loss": 0.0302, "num_input_tokens_seen": 3294720, "step": 330 }, { "epoch": 0.17254699974246718, "grad_norm": 1.6706834486788253, "learning_rate": 9.935750420309055e-05, "loss": 0.082, "num_input_tokens_seen": 3344640, "step": 335 }, { "epoch": 0.17512232809683234, "grad_norm": 0.16856461555999938, "learning_rate": 9.931806517013612e-05, "loss": 0.0411, "num_input_tokens_seen": 3394560, "step": 340 }, { "epoch": 0.17769765645119753, "grad_norm": 8.475859938133665, "learning_rate": 9.927745975766654e-05, "loss": 0.1158, "num_input_tokens_seen": 3444480, "step": 345 }, { "epoch": 0.1802729848055627, "grad_norm": 0.7891476176494399, "learning_rate": 9.923568892600578e-05, "loss": 0.1761, "num_input_tokens_seen": 3494400, "step": 350 }, { "epoch": 0.1802729848055627, "eval_loss": 0.043948542326688766, "eval_runtime": 19.3589, "eval_samples_per_second": 3.099, "eval_steps_per_second": 0.775, "num_input_tokens_seen": 3494400, "step": 350 }, { "epoch": 0.1828483131599279, "grad_norm": 0.3217035938609153, "learning_rate": 9.91927536630402e-05, "loss": 0.0267, "num_input_tokens_seen": 3544320, "step": 355 }, { "epoch": 0.18542364151429308, "grad_norm": 1.9146681345367045, "learning_rate": 9.91486549841951e-05, "loss": 0.0642, "num_input_tokens_seen": 3594240, "step": 360 }, { "epoch": 0.18799896986865824, "grad_norm": 0.053067921944857786, "learning_rate": 9.91033939324107e-05, "loss": 0.0302, "num_input_tokens_seen": 3644160, "step": 365 }, { "epoch": 0.19057429822302344, "grad_norm": 0.07277339857890938, "learning_rate": 9.905697157811761e-05, "loss": 0.0099, "num_input_tokens_seen": 3694080, "step": 370 }, { "epoch": 0.19314962657738863, "grad_norm": 4.117705571576552, "learning_rate": 9.900938901921131e-05, "loss": 0.0573, "num_input_tokens_seen": 3744000, "step": 375 }, { "epoch": 0.1957249549317538, "grad_norm": 2.4479374171708304, "learning_rate": 9.896064738102635e-05, "loss": 0.0302, "num_input_tokens_seen": 3793920, "step": 380 }, { "epoch": 0.19830028328611898, "grad_norm": 2.1751552571638153, "learning_rate": 9.891074781630966e-05, "loss": 0.0374, "num_input_tokens_seen": 3843840, "step": 385 }, { "epoch": 0.20087561164048418, "grad_norm": 1.4138134932494708, "learning_rate": 9.885969150519331e-05, "loss": 0.0227, "num_input_tokens_seen": 3893760, "step": 390 }, { "epoch": 0.20345093999484934, "grad_norm": 0.3312762825665537, "learning_rate": 9.88074796551666e-05, "loss": 0.0104, "num_input_tokens_seen": 3943680, "step": 395 }, { "epoch": 0.20602626834921453, "grad_norm": 0.06774783590394957, "learning_rate": 9.875411350104744e-05, "loss": 0.0146, "num_input_tokens_seen": 3993600, "step": 400 }, { "epoch": 0.20602626834921453, "eval_loss": 0.024433813989162445, "eval_runtime": 18.8143, "eval_samples_per_second": 3.189, "eval_steps_per_second": 0.797, "num_input_tokens_seen": 3993600, "step": 400 }, { "epoch": 0.2086015967035797, "grad_norm": 0.06416132017781931, "learning_rate": 9.86995943049533e-05, "loss": 0.0074, "num_input_tokens_seen": 4043520, "step": 405 }, { "epoch": 0.2111769250579449, "grad_norm": 0.04809215529224866, "learning_rate": 9.864392335627117e-05, "loss": 0.0375, "num_input_tokens_seen": 4093440, "step": 410 }, { "epoch": 0.21375225341231008, "grad_norm": 0.40943321947836153, "learning_rate": 9.858710197162721e-05, "loss": 0.0217, "num_input_tokens_seen": 4143360, "step": 415 }, { "epoch": 0.21632758176667524, "grad_norm": 0.08069011480339518, "learning_rate": 9.852913149485556e-05, "loss": 0.016, "num_input_tokens_seen": 4193280, "step": 420 }, { "epoch": 0.21890291012104043, "grad_norm": 1.806585526467194, "learning_rate": 9.847001329696653e-05, "loss": 0.0832, "num_input_tokens_seen": 4243200, "step": 425 }, { "epoch": 0.22147823847540563, "grad_norm": 3.122914329597603, "learning_rate": 9.840974877611422e-05, "loss": 0.0444, "num_input_tokens_seen": 4293120, "step": 430 }, { "epoch": 0.2240535668297708, "grad_norm": 1.0101510644418257, "learning_rate": 9.834833935756344e-05, "loss": 0.0465, "num_input_tokens_seen": 4343040, "step": 435 }, { "epoch": 0.22662889518413598, "grad_norm": 1.7770164472545809, "learning_rate": 9.828578649365601e-05, "loss": 0.0428, "num_input_tokens_seen": 4392960, "step": 440 }, { "epoch": 0.22920422353850115, "grad_norm": 0.9798524063329249, "learning_rate": 9.822209166377635e-05, "loss": 0.02, "num_input_tokens_seen": 4442880, "step": 445 }, { "epoch": 0.23177955189286634, "grad_norm": 0.22305429096692395, "learning_rate": 9.815725637431662e-05, "loss": 0.0157, "num_input_tokens_seen": 4492800, "step": 450 }, { "epoch": 0.23177955189286634, "eval_loss": 0.006690301466733217, "eval_runtime": 18.7725, "eval_samples_per_second": 3.196, "eval_steps_per_second": 0.799, "num_input_tokens_seen": 4492800, "step": 450 }, { "epoch": 0.23435488024723153, "grad_norm": 2.8137229294991064, "learning_rate": 9.809128215864097e-05, "loss": 0.0596, "num_input_tokens_seen": 4542720, "step": 455 }, { "epoch": 0.2369302086015967, "grad_norm": 0.045182426772616975, "learning_rate": 9.802417057704931e-05, "loss": 0.0238, "num_input_tokens_seen": 4592640, "step": 460 }, { "epoch": 0.23950553695596188, "grad_norm": 2.1659990503080424, "learning_rate": 9.795592321674045e-05, "loss": 0.0331, "num_input_tokens_seen": 4642560, "step": 465 }, { "epoch": 0.24208086531032708, "grad_norm": 0.3208179936186945, "learning_rate": 9.788654169177453e-05, "loss": 0.0177, "num_input_tokens_seen": 4692480, "step": 470 }, { "epoch": 0.24465619366469224, "grad_norm": 1.4150868022870595, "learning_rate": 9.781602764303487e-05, "loss": 0.028, "num_input_tokens_seen": 4742400, "step": 475 }, { "epoch": 0.24723152201905743, "grad_norm": 0.21015675310797993, "learning_rate": 9.774438273818911e-05, "loss": 0.0561, "num_input_tokens_seen": 4792320, "step": 480 }, { "epoch": 0.24980685037342262, "grad_norm": 2.6788940269709287, "learning_rate": 9.767160867164979e-05, "loss": 0.0242, "num_input_tokens_seen": 4842240, "step": 485 }, { "epoch": 0.2523821787277878, "grad_norm": 0.03933517764434036, "learning_rate": 9.759770716453436e-05, "loss": 0.0286, "num_input_tokens_seen": 4892160, "step": 490 }, { "epoch": 0.254957507082153, "grad_norm": 2.640299598300008, "learning_rate": 9.752267996462434e-05, "loss": 0.0394, "num_input_tokens_seen": 4942080, "step": 495 }, { "epoch": 0.25753283543651817, "grad_norm": 0.03143215150024672, "learning_rate": 9.744652884632406e-05, "loss": 0.0122, "num_input_tokens_seen": 4992000, "step": 500 }, { "epoch": 0.25753283543651817, "eval_loss": 0.007972972467541695, "eval_runtime": 18.8509, "eval_samples_per_second": 3.183, "eval_steps_per_second": 0.796, "num_input_tokens_seen": 4992000, "step": 500 }, { "epoch": 0.26010816379088336, "grad_norm": 0.3421379771468272, "learning_rate": 9.736925561061871e-05, "loss": 0.0184, "num_input_tokens_seen": 5041920, "step": 505 }, { "epoch": 0.2626834921452485, "grad_norm": 0.16336037200955245, "learning_rate": 9.729086208503174e-05, "loss": 0.0046, "num_input_tokens_seen": 5091840, "step": 510 }, { "epoch": 0.2652588204996137, "grad_norm": 0.7318312729236729, "learning_rate": 9.721135012358156e-05, "loss": 0.0226, "num_input_tokens_seen": 5141760, "step": 515 }, { "epoch": 0.2678341488539789, "grad_norm": 2.1703341770378333, "learning_rate": 9.713072160673777e-05, "loss": 0.0475, "num_input_tokens_seen": 5191680, "step": 520 }, { "epoch": 0.2704094772083441, "grad_norm": 1.6372493711976897, "learning_rate": 9.704897844137673e-05, "loss": 0.0572, "num_input_tokens_seen": 5241600, "step": 525 }, { "epoch": 0.27298480556270927, "grad_norm": 3.5840623375725804, "learning_rate": 9.696612256073633e-05, "loss": 0.025, "num_input_tokens_seen": 5291520, "step": 530 }, { "epoch": 0.2755601339170744, "grad_norm": 0.06141348661076509, "learning_rate": 9.688215592437039e-05, "loss": 0.0122, "num_input_tokens_seen": 5341440, "step": 535 }, { "epoch": 0.2781354622714396, "grad_norm": 0.04980492840656424, "learning_rate": 9.679708051810221e-05, "loss": 0.0045, "num_input_tokens_seen": 5391360, "step": 540 }, { "epoch": 0.2807107906258048, "grad_norm": 3.153079438616822, "learning_rate": 9.67108983539777e-05, "loss": 0.0431, "num_input_tokens_seen": 5441280, "step": 545 }, { "epoch": 0.28328611898017, "grad_norm": 1.6764625232464418, "learning_rate": 9.662361147021779e-05, "loss": 0.0339, "num_input_tokens_seen": 5491200, "step": 550 }, { "epoch": 0.28328611898017, "eval_loss": 0.0033526704646646976, "eval_runtime": 18.8387, "eval_samples_per_second": 3.185, "eval_steps_per_second": 0.796, "num_input_tokens_seen": 5491200, "step": 550 }, { "epoch": 0.28586144733453517, "grad_norm": 0.05427488923085365, "learning_rate": 9.653522193117013e-05, "loss": 0.0341, "num_input_tokens_seen": 5541120, "step": 555 }, { "epoch": 0.28843677568890036, "grad_norm": 1.538376531568411, "learning_rate": 9.644573182726035e-05, "loss": 0.0341, "num_input_tokens_seen": 5591040, "step": 560 }, { "epoch": 0.2910121040432655, "grad_norm": 0.04402865983757039, "learning_rate": 9.63551432749426e-05, "loss": 0.0453, "num_input_tokens_seen": 5640960, "step": 565 }, { "epoch": 0.2935874323976307, "grad_norm": 0.1469885874710427, "learning_rate": 9.626345841664953e-05, "loss": 0.0227, "num_input_tokens_seen": 5690880, "step": 570 }, { "epoch": 0.2961627607519959, "grad_norm": 0.3459204695983954, "learning_rate": 9.617067942074153e-05, "loss": 0.0135, "num_input_tokens_seen": 5740800, "step": 575 }, { "epoch": 0.29873808910636107, "grad_norm": 1.7604264611882452, "learning_rate": 9.607680848145558e-05, "loss": 0.0439, "num_input_tokens_seen": 5790720, "step": 580 }, { "epoch": 0.30131341746072626, "grad_norm": 2.1661707703583204, "learning_rate": 9.598184781885318e-05, "loss": 0.0344, "num_input_tokens_seen": 5840640, "step": 585 }, { "epoch": 0.3038887458150914, "grad_norm": 0.19674934066326408, "learning_rate": 9.588579967876806e-05, "loss": 0.0244, "num_input_tokens_seen": 5890560, "step": 590 }, { "epoch": 0.3064640741694566, "grad_norm": 4.463551986915384, "learning_rate": 9.578866633275288e-05, "loss": 0.022, "num_input_tokens_seen": 5940480, "step": 595 }, { "epoch": 0.3090394025238218, "grad_norm": 0.08358626804613765, "learning_rate": 9.569045007802559e-05, "loss": 0.0217, "num_input_tokens_seen": 5990400, "step": 600 }, { "epoch": 0.3090394025238218, "eval_loss": 0.013290103524923325, "eval_runtime": 18.8576, "eval_samples_per_second": 3.182, "eval_steps_per_second": 0.795, "num_input_tokens_seen": 5990400, "step": 600 }, { "epoch": 0.311614730878187, "grad_norm": 0.06328082378687645, "learning_rate": 9.55911532374151e-05, "loss": 0.0207, "num_input_tokens_seen": 6040320, "step": 605 }, { "epoch": 0.31419005923255217, "grad_norm": 1.846961607001629, "learning_rate": 9.549077815930636e-05, "loss": 0.023, "num_input_tokens_seen": 6090240, "step": 610 }, { "epoch": 0.31676538758691736, "grad_norm": 0.07122251919235507, "learning_rate": 9.538932721758474e-05, "loss": 0.017, "num_input_tokens_seen": 6140160, "step": 615 }, { "epoch": 0.3193407159412825, "grad_norm": 1.9564639611342651, "learning_rate": 9.528680281157999e-05, "loss": 0.009, "num_input_tokens_seen": 6190080, "step": 620 }, { "epoch": 0.3219160442956477, "grad_norm": 0.43628623054662674, "learning_rate": 9.518320736600943e-05, "loss": 0.0421, "num_input_tokens_seen": 6240000, "step": 625 }, { "epoch": 0.3244913726500129, "grad_norm": 0.7626652394047067, "learning_rate": 9.507854333092063e-05, "loss": 0.0324, "num_input_tokens_seen": 6289920, "step": 630 }, { "epoch": 0.32706670100437807, "grad_norm": 0.5823990007842583, "learning_rate": 9.497281318163346e-05, "loss": 0.0139, "num_input_tokens_seen": 6339840, "step": 635 }, { "epoch": 0.32964202935874326, "grad_norm": 1.6726620150276597, "learning_rate": 9.486601941868154e-05, "loss": 0.0375, "num_input_tokens_seen": 6389760, "step": 640 }, { "epoch": 0.3322173577131084, "grad_norm": 0.08456122780891835, "learning_rate": 9.475816456775313e-05, "loss": 0.0618, "num_input_tokens_seen": 6439680, "step": 645 }, { "epoch": 0.3347926860674736, "grad_norm": 0.3287256358007373, "learning_rate": 9.464925117963133e-05, "loss": 0.0327, "num_input_tokens_seen": 6489600, "step": 650 }, { "epoch": 0.3347926860674736, "eval_loss": 0.02095886506140232, "eval_runtime": 19.1993, "eval_samples_per_second": 3.125, "eval_steps_per_second": 0.781, "num_input_tokens_seen": 6489600, "step": 650 }, { "epoch": 0.3373680144218388, "grad_norm": 1.642912250249591, "learning_rate": 9.453928183013385e-05, "loss": 0.037, "num_input_tokens_seen": 6539520, "step": 655 }, { "epoch": 0.33994334277620397, "grad_norm": 1.6829194140719153, "learning_rate": 9.442825912005202e-05, "loss": 0.034, "num_input_tokens_seen": 6589440, "step": 660 }, { "epoch": 0.34251867113056916, "grad_norm": 1.7722553129499397, "learning_rate": 9.431618567508933e-05, "loss": 0.0584, "num_input_tokens_seen": 6639360, "step": 665 }, { "epoch": 0.34509399948493436, "grad_norm": 1.8398383207694327, "learning_rate": 9.420306414579925e-05, "loss": 0.059, "num_input_tokens_seen": 6689280, "step": 670 }, { "epoch": 0.3476693278392995, "grad_norm": 0.2451231053960763, "learning_rate": 9.408889720752266e-05, "loss": 0.023, "num_input_tokens_seen": 6739200, "step": 675 }, { "epoch": 0.3502446561936647, "grad_norm": 0.13166014546250449, "learning_rate": 9.397368756032445e-05, "loss": 0.0327, "num_input_tokens_seen": 6789120, "step": 680 }, { "epoch": 0.3528199845480299, "grad_norm": 1.9392864055798256, "learning_rate": 9.385743792892982e-05, "loss": 0.0241, "num_input_tokens_seen": 6839040, "step": 685 }, { "epoch": 0.35539531290239507, "grad_norm": 0.7574757741852987, "learning_rate": 9.374015106265968e-05, "loss": 0.0227, "num_input_tokens_seen": 6888960, "step": 690 }, { "epoch": 0.35797064125676026, "grad_norm": 1.1649279233149774, "learning_rate": 9.362182973536569e-05, "loss": 0.0171, "num_input_tokens_seen": 6938880, "step": 695 }, { "epoch": 0.3605459696111254, "grad_norm": 0.04561814221227265, "learning_rate": 9.35024767453647e-05, "loss": 0.0267, "num_input_tokens_seen": 6988800, "step": 700 }, { "epoch": 0.3605459696111254, "eval_loss": 0.005286735948175192, "eval_runtime": 19.2824, "eval_samples_per_second": 3.112, "eval_steps_per_second": 0.778, "num_input_tokens_seen": 6988800, "step": 700 }, { "epoch": 0.3631212979654906, "grad_norm": 1.7432145241229078, "learning_rate": 9.338209491537257e-05, "loss": 0.0335, "num_input_tokens_seen": 7038720, "step": 705 }, { "epoch": 0.3656966263198558, "grad_norm": 1.821635752664896, "learning_rate": 9.326068709243727e-05, "loss": 0.0108, "num_input_tokens_seen": 7088640, "step": 710 }, { "epoch": 0.36827195467422097, "grad_norm": 0.5782878283043246, "learning_rate": 9.313825614787177e-05, "loss": 0.0296, "num_input_tokens_seen": 7138560, "step": 715 }, { "epoch": 0.37084728302858616, "grad_norm": 2.0568373189070908, "learning_rate": 9.301480497718593e-05, "loss": 0.042, "num_input_tokens_seen": 7188480, "step": 720 }, { "epoch": 0.37342261138295135, "grad_norm": 0.20623581315280795, "learning_rate": 9.289033650001817e-05, "loss": 0.0083, "num_input_tokens_seen": 7238400, "step": 725 }, { "epoch": 0.3759979397373165, "grad_norm": 1.424505040979842, "learning_rate": 9.276485366006634e-05, "loss": 0.0109, "num_input_tokens_seen": 7288320, "step": 730 }, { "epoch": 0.3785732680916817, "grad_norm": 2.1248350485820584, "learning_rate": 9.263835942501807e-05, "loss": 0.0122, "num_input_tokens_seen": 7338240, "step": 735 }, { "epoch": 0.3811485964460469, "grad_norm": 0.357888255955218, "learning_rate": 9.251085678648072e-05, "loss": 0.017, "num_input_tokens_seen": 7388160, "step": 740 }, { "epoch": 0.38372392480041206, "grad_norm": 3.859418562917624, "learning_rate": 9.238234875991046e-05, "loss": 0.0236, "num_input_tokens_seen": 7438080, "step": 745 }, { "epoch": 0.38629925315477726, "grad_norm": 2.660896728773248, "learning_rate": 9.225283838454111e-05, "loss": 0.014, "num_input_tokens_seen": 7488000, "step": 750 }, { "epoch": 0.38629925315477726, "eval_loss": 0.005334165878593922, "eval_runtime": 19.3986, "eval_samples_per_second": 3.093, "eval_steps_per_second": 0.773, "num_input_tokens_seen": 7488000, "step": 750 }, { "epoch": 0.3888745815091424, "grad_norm": 3.511127189174474, "learning_rate": 9.21223287233121e-05, "loss": 0.0426, "num_input_tokens_seen": 7537920, "step": 755 }, { "epoch": 0.3914499098635076, "grad_norm": 2.858952603531337, "learning_rate": 9.199082286279622e-05, "loss": 0.0094, "num_input_tokens_seen": 7587840, "step": 760 }, { "epoch": 0.3940252382178728, "grad_norm": 1.9568584051828664, "learning_rate": 9.185832391312644e-05, "loss": 0.008, "num_input_tokens_seen": 7637760, "step": 765 }, { "epoch": 0.39660056657223797, "grad_norm": 1.7485145115139782, "learning_rate": 9.172483500792244e-05, "loss": 0.0297, "num_input_tokens_seen": 7687680, "step": 770 }, { "epoch": 0.39917589492660316, "grad_norm": 0.05161945290023201, "learning_rate": 9.159035930421658e-05, "loss": 0.0239, "num_input_tokens_seen": 7737600, "step": 775 }, { "epoch": 0.40175122328096835, "grad_norm": 5.343080934932771, "learning_rate": 9.145489998237902e-05, "loss": 0.0162, "num_input_tokens_seen": 7787520, "step": 780 }, { "epoch": 0.4043265516353335, "grad_norm": 0.020924688334471742, "learning_rate": 9.131846024604274e-05, "loss": 0.0309, "num_input_tokens_seen": 7837440, "step": 785 }, { "epoch": 0.4069018799896987, "grad_norm": 0.5607043400469908, "learning_rate": 9.11810433220276e-05, "loss": 0.0432, "num_input_tokens_seen": 7887360, "step": 790 }, { "epoch": 0.40947720834406387, "grad_norm": 0.05221845282486231, "learning_rate": 9.104265246026415e-05, "loss": 0.0022, "num_input_tokens_seen": 7937280, "step": 795 }, { "epoch": 0.41205253669842906, "grad_norm": 0.09825330413354653, "learning_rate": 9.090329093371666e-05, "loss": 0.0065, "num_input_tokens_seen": 7987200, "step": 800 }, { "epoch": 0.41205253669842906, "eval_loss": 0.006797688081860542, "eval_runtime": 19.3386, "eval_samples_per_second": 3.103, "eval_steps_per_second": 0.776, "num_input_tokens_seen": 7987200, "step": 800 }, { "epoch": 0.41462786505279425, "grad_norm": 0.016984614729716928, "learning_rate": 9.076296203830579e-05, "loss": 0.0025, "num_input_tokens_seen": 8037120, "step": 805 }, { "epoch": 0.4172031934071594, "grad_norm": 8.87156169406028, "learning_rate": 9.062166909283062e-05, "loss": 0.0338, "num_input_tokens_seen": 8087040, "step": 810 }, { "epoch": 0.4197785217615246, "grad_norm": 0.07123492673307173, "learning_rate": 9.047941543889014e-05, "loss": 0.0505, "num_input_tokens_seen": 8136960, "step": 815 }, { "epoch": 0.4223538501158898, "grad_norm": 0.3486611879887143, "learning_rate": 9.033620444080428e-05, "loss": 0.0335, "num_input_tokens_seen": 8186880, "step": 820 }, { "epoch": 0.42492917847025496, "grad_norm": 0.2658900406741178, "learning_rate": 9.019203948553422e-05, "loss": 0.0196, "num_input_tokens_seen": 8236800, "step": 825 }, { "epoch": 0.42750450682462016, "grad_norm": 0.15327530188348007, "learning_rate": 9.004692398260244e-05, "loss": 0.018, "num_input_tokens_seen": 8286720, "step": 830 }, { "epoch": 0.43007983517898535, "grad_norm": 5.292901102191953, "learning_rate": 8.9900861364012e-05, "loss": 0.012, "num_input_tokens_seen": 8336640, "step": 835 }, { "epoch": 0.4326551635333505, "grad_norm": 3.2113529029102375, "learning_rate": 8.975385508416532e-05, "loss": 0.0062, "num_input_tokens_seen": 8386560, "step": 840 }, { "epoch": 0.4352304918877157, "grad_norm": 0.3685880212421972, "learning_rate": 8.960590861978265e-05, "loss": 0.0232, "num_input_tokens_seen": 8436480, "step": 845 }, { "epoch": 0.43780582024208087, "grad_norm": 1.4858234531184304, "learning_rate": 8.945702546981969e-05, "loss": 0.0306, "num_input_tokens_seen": 8486400, "step": 850 }, { "epoch": 0.43780582024208087, "eval_loss": 0.007198736071586609, "eval_runtime": 19.7019, "eval_samples_per_second": 3.045, "eval_steps_per_second": 0.761, "num_input_tokens_seen": 8486400, "step": 850 } ], "logging_steps": 5, "max_steps": 3400, "num_input_tokens_seen": 8486400, "num_train_epochs": 2, "save_steps": 50, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 559924481097728.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }