{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.9255575164447296, "global_step": 4000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "learning_rate": 1.8e-07, "loss": 0.6934, "step": 10 }, { "epoch": 0.01, "learning_rate": 3.8e-07, "loss": 0.6932, "step": 20 }, { "epoch": 0.01, "learning_rate": 5.800000000000001e-07, "loss": 0.6941, "step": 30 }, { "epoch": 0.02, "learning_rate": 7.8e-07, "loss": 0.6938, "step": 40 }, { "epoch": 0.02, "learning_rate": 9.800000000000001e-07, "loss": 0.693, "step": 50 }, { "epoch": 0.03, "learning_rate": 1.1800000000000001e-06, "loss": 0.6943, "step": 60 }, { "epoch": 0.03, "learning_rate": 1.3800000000000001e-06, "loss": 0.6945, "step": 70 }, { "epoch": 0.04, "learning_rate": 1.5800000000000001e-06, "loss": 0.693, "step": 80 }, { "epoch": 0.04, "learning_rate": 1.7800000000000001e-06, "loss": 0.6935, "step": 90 }, { "epoch": 0.05, "learning_rate": 1.98e-06, "loss": 0.6937, "step": 100 }, { "epoch": 0.05, "learning_rate": 2.1800000000000003e-06, "loss": 0.6932, "step": 110 }, { "epoch": 0.06, "learning_rate": 2.38e-06, "loss": 0.6936, "step": 120 }, { "epoch": 0.06, "learning_rate": 2.5800000000000003e-06, "loss": 0.6932, "step": 130 }, { "epoch": 0.07, "learning_rate": 2.7800000000000005e-06, "loss": 0.6938, "step": 140 }, { "epoch": 0.07, "learning_rate": 2.9800000000000003e-06, "loss": 0.6937, "step": 150 }, { "epoch": 0.08, "learning_rate": 3.1800000000000005e-06, "loss": 0.6925, "step": 160 }, { "epoch": 0.08, "learning_rate": 3.3800000000000007e-06, "loss": 0.6923, "step": 170 }, { "epoch": 0.09, "learning_rate": 3.58e-06, "loss": 0.6928, "step": 180 }, { "epoch": 0.09, "learning_rate": 3.7800000000000002e-06, "loss": 0.6899, "step": 190 }, { "epoch": 0.1, "learning_rate": 3.980000000000001e-06, "loss": 0.691, "step": 200 }, { "epoch": 0.1, "learning_rate": 4.18e-06, "loss": 0.6897, "step": 210 }, { "epoch": 0.11, "learning_rate": 4.38e-06, "loss": 0.6874, "step": 220 }, { "epoch": 0.11, "learning_rate": 4.58e-06, "loss": 0.6784, "step": 230 }, { "epoch": 0.12, "learning_rate": 4.78e-06, "loss": 0.6619, "step": 240 }, { "epoch": 0.12, "learning_rate": 4.980000000000001e-06, "loss": 0.6517, "step": 250 }, { "epoch": 0.13, "learning_rate": 5.18e-06, "loss": 0.6444, "step": 260 }, { "epoch": 0.13, "learning_rate": 5.380000000000001e-06, "loss": 0.6422, "step": 270 }, { "epoch": 0.13, "learning_rate": 5.580000000000001e-06, "loss": 0.6389, "step": 280 }, { "epoch": 0.14, "learning_rate": 5.78e-06, "loss": 0.6255, "step": 290 }, { "epoch": 0.14, "learning_rate": 5.98e-06, "loss": 0.6281, "step": 300 }, { "epoch": 0.15, "learning_rate": 6.18e-06, "loss": 0.6368, "step": 310 }, { "epoch": 0.15, "learning_rate": 6.380000000000001e-06, "loss": 0.6291, "step": 320 }, { "epoch": 0.16, "learning_rate": 6.5800000000000005e-06, "loss": 0.6103, "step": 330 }, { "epoch": 0.16, "learning_rate": 6.780000000000001e-06, "loss": 0.611, "step": 340 }, { "epoch": 0.17, "learning_rate": 6.98e-06, "loss": 0.6003, "step": 350 }, { "epoch": 0.17, "learning_rate": 7.180000000000001e-06, "loss": 0.607, "step": 360 }, { "epoch": 0.18, "learning_rate": 7.3800000000000005e-06, "loss": 0.6062, "step": 370 }, { "epoch": 0.18, "learning_rate": 7.5600000000000005e-06, "loss": 0.6014, "step": 380 }, { "epoch": 0.19, "learning_rate": 7.76e-06, "loss": 0.6156, "step": 390 }, { "epoch": 0.19, "learning_rate": 7.960000000000002e-06, "loss": 0.6034, "step": 400 }, { "epoch": 0.2, "learning_rate": 8.16e-06, "loss": 0.5974, "step": 410 }, { "epoch": 0.2, "learning_rate": 8.36e-06, "loss": 0.6065, "step": 420 }, { "epoch": 0.21, "learning_rate": 8.560000000000001e-06, "loss": 0.5785, "step": 430 }, { "epoch": 0.21, "learning_rate": 8.76e-06, "loss": 0.601, "step": 440 }, { "epoch": 0.22, "learning_rate": 8.96e-06, "loss": 0.5838, "step": 450 }, { "epoch": 0.22, "learning_rate": 9.16e-06, "loss": 0.6045, "step": 460 }, { "epoch": 0.23, "learning_rate": 9.360000000000002e-06, "loss": 0.5914, "step": 470 }, { "epoch": 0.23, "learning_rate": 9.56e-06, "loss": 0.585, "step": 480 }, { "epoch": 0.24, "learning_rate": 9.760000000000001e-06, "loss": 0.5845, "step": 490 }, { "epoch": 0.24, "learning_rate": 9.960000000000001e-06, "loss": 0.5926, "step": 500 }, { "epoch": 0.24, "eval_webgpt_accuracy": 0.5497957099080695, "eval_webgpt_loss": 0.7101226449012756, "eval_webgpt_runtime": 39.1748, "eval_webgpt_samples_per_second": 99.962, "eval_webgpt_steps_per_second": 1.327, "step": 500 }, { "epoch": 0.24, "eval_hfsummary_accuracy": 0.6480669830426503, "eval_hfsummary_loss": 0.6175686717033386, "eval_hfsummary_runtime": 973.4644, "eval_hfsummary_samples_per_second": 33.985, "eval_hfsummary_steps_per_second": 0.448, "step": 500 }, { "epoch": 0.24, "eval_anthropic_rlhf_accuracy": 0.5922591206735267, "eval_anthropic_rlhf_loss": 0.6707426905632019, "eval_anthropic_rlhf_runtime": 90.7482, "eval_anthropic_rlhf_samples_per_second": 94.239, "eval_anthropic_rlhf_steps_per_second": 1.245, "step": 500 }, { "epoch": 0.24, "eval_gptsynthetic_accuracy": 0.9969834087481146, "eval_gptsynthetic_loss": 0.03679310530424118, "eval_gptsynthetic_runtime": 31.6689, "eval_gptsynthetic_samples_per_second": 104.677, "eval_gptsynthetic_steps_per_second": 1.389, "step": 500 }, { "epoch": 0.25, "learning_rate": 9.999974108670129e-06, "loss": 0.582, "step": 510 }, { "epoch": 0.25, "learning_rate": 9.999868925602092e-06, "loss": 0.5802, "step": 520 }, { "epoch": 0.26, "learning_rate": 9.99968283428854e-06, "loss": 0.5811, "step": 530 }, { "epoch": 0.26, "learning_rate": 9.99941583774082e-06, "loss": 0.5818, "step": 540 }, { "epoch": 0.26, "learning_rate": 9.999067940279488e-06, "loss": 0.5585, "step": 550 }, { "epoch": 0.27, "learning_rate": 9.998639147534247e-06, "loss": 0.5645, "step": 560 }, { "epoch": 0.27, "learning_rate": 9.998129466443849e-06, "loss": 0.584, "step": 570 }, { "epoch": 0.28, "learning_rate": 9.997538905255994e-06, "loss": 0.577, "step": 580 }, { "epoch": 0.28, "learning_rate": 9.996867473527188e-06, "loss": 0.5629, "step": 590 }, { "epoch": 0.29, "learning_rate": 9.996115182122585e-06, "loss": 0.5721, "step": 600 }, { "epoch": 0.29, "learning_rate": 9.995282043215824e-06, "loss": 0.5599, "step": 610 }, { "epoch": 0.3, "learning_rate": 9.994368070288814e-06, "loss": 0.5631, "step": 620 }, { "epoch": 0.3, "learning_rate": 9.993373278131538e-06, "loss": 0.5652, "step": 630 }, { "epoch": 0.31, "learning_rate": 9.99229768284179e-06, "loss": 0.557, "step": 640 }, { "epoch": 0.31, "learning_rate": 9.991141301824936e-06, "loss": 0.554, "step": 650 }, { "epoch": 0.32, "learning_rate": 9.989904153793614e-06, "loss": 0.5763, "step": 660 }, { "epoch": 0.32, "learning_rate": 9.98858625876745e-06, "loss": 0.5628, "step": 670 }, { "epoch": 0.33, "learning_rate": 9.987187638072715e-06, "loss": 0.5632, "step": 680 }, { "epoch": 0.33, "learning_rate": 9.98570831434199e-06, "loss": 0.5896, "step": 690 }, { "epoch": 0.34, "learning_rate": 9.984148311513805e-06, "loss": 0.5666, "step": 700 }, { "epoch": 0.34, "learning_rate": 9.982507654832232e-06, "loss": 0.5489, "step": 710 }, { "epoch": 0.35, "learning_rate": 9.980786370846504e-06, "loss": 0.5392, "step": 720 }, { "epoch": 0.35, "learning_rate": 9.978984487410564e-06, "loss": 0.5703, "step": 730 }, { "epoch": 0.36, "learning_rate": 9.977102033682621e-06, "loss": 0.5619, "step": 740 }, { "epoch": 0.36, "learning_rate": 9.97513904012468e-06, "loss": 0.5554, "step": 750 }, { "epoch": 0.37, "learning_rate": 9.973095538502044e-06, "loss": 0.5583, "step": 760 }, { "epoch": 0.37, "learning_rate": 9.97097156188281e-06, "loss": 0.5588, "step": 770 }, { "epoch": 0.38, "learning_rate": 9.96876714463732e-06, "loss": 0.5786, "step": 780 }, { "epoch": 0.38, "learning_rate": 9.966482322437616e-06, "loss": 0.5515, "step": 790 }, { "epoch": 0.39, "learning_rate": 9.96411713225686e-06, "loss": 0.5532, "step": 800 }, { "epoch": 0.39, "learning_rate": 9.961671612368731e-06, "loss": 0.5526, "step": 810 }, { "epoch": 0.39, "learning_rate": 9.959145802346815e-06, "loss": 0.5495, "step": 820 }, { "epoch": 0.4, "learning_rate": 9.95653974306395e-06, "loss": 0.5531, "step": 830 }, { "epoch": 0.4, "learning_rate": 9.95385347669158e-06, "loss": 0.5465, "step": 840 }, { "epoch": 0.41, "learning_rate": 9.951087046699066e-06, "loss": 0.5606, "step": 850 }, { "epoch": 0.41, "learning_rate": 9.94824049785298e-06, "loss": 0.546, "step": 860 }, { "epoch": 0.42, "learning_rate": 9.945313876216383e-06, "loss": 0.5503, "step": 870 }, { "epoch": 0.42, "learning_rate": 9.942307229148083e-06, "loss": 0.5395, "step": 880 }, { "epoch": 0.43, "learning_rate": 9.939220605301861e-06, "loss": 0.5466, "step": 890 }, { "epoch": 0.43, "learning_rate": 9.936054054625692e-06, "loss": 0.5672, "step": 900 }, { "epoch": 0.44, "learning_rate": 9.932807628360933e-06, "loss": 0.5491, "step": 910 }, { "epoch": 0.44, "learning_rate": 9.929481379041486e-06, "loss": 0.5549, "step": 920 }, { "epoch": 0.45, "learning_rate": 9.926075360492967e-06, "loss": 0.544, "step": 930 }, { "epoch": 0.45, "learning_rate": 9.922589627831814e-06, "loss": 0.5427, "step": 940 }, { "epoch": 0.46, "learning_rate": 9.919024237464411e-06, "loss": 0.5515, "step": 950 }, { "epoch": 0.46, "learning_rate": 9.915379247086166e-06, "loss": 0.5402, "step": 960 }, { "epoch": 0.47, "learning_rate": 9.911654715680578e-06, "loss": 0.5262, "step": 970 }, { "epoch": 0.47, "learning_rate": 9.907850703518292e-06, "loss": 0.5327, "step": 980 }, { "epoch": 0.48, "learning_rate": 9.903967272156108e-06, "loss": 0.5355, "step": 990 }, { "epoch": 0.48, "learning_rate": 9.900004484435997e-06, "loss": 0.5385, "step": 1000 }, { "epoch": 0.48, "eval_webgpt_accuracy": 0.5771195097037793, "eval_webgpt_loss": 0.6884612441062927, "eval_webgpt_runtime": 38.7402, "eval_webgpt_samples_per_second": 101.084, "eval_webgpt_steps_per_second": 1.342, "step": 1000 }, { "epoch": 0.48, "eval_hfsummary_accuracy": 0.6676843091618052, "eval_hfsummary_loss": 0.62443608045578, "eval_hfsummary_runtime": 978.4924, "eval_hfsummary_samples_per_second": 33.81, "eval_hfsummary_steps_per_second": 0.446, "step": 1000 }, { "epoch": 0.48, "eval_anthropic_rlhf_accuracy": 0.6214920486435921, "eval_anthropic_rlhf_loss": 0.6546406149864197, "eval_anthropic_rlhf_runtime": 91.0623, "eval_anthropic_rlhf_samples_per_second": 93.914, "eval_anthropic_rlhf_steps_per_second": 1.241, "step": 1000 }, { "epoch": 0.48, "eval_gptsynthetic_accuracy": 0.9963800904977376, "eval_gptsynthetic_loss": 0.019633520394563675, "eval_gptsynthetic_runtime": 32.5041, "eval_gptsynthetic_samples_per_second": 101.987, "eval_gptsynthetic_steps_per_second": 1.354, "step": 1000 }, { "epoch": 0.49, "learning_rate": 9.895962404484083e-06, "loss": 0.5368, "step": 1010 }, { "epoch": 0.49, "learning_rate": 9.891841097709599e-06, "loss": 0.5293, "step": 1020 }, { "epoch": 0.5, "learning_rate": 9.88764063080383e-06, "loss": 0.5373, "step": 1030 }, { "epoch": 0.5, "learning_rate": 9.883361071739045e-06, "loss": 0.5253, "step": 1040 }, { "epoch": 0.51, "learning_rate": 9.87900248976738e-06, "loss": 0.5406, "step": 1050 }, { "epoch": 0.51, "learning_rate": 9.874564955419725e-06, "loss": 0.5194, "step": 1060 }, { "epoch": 0.52, "learning_rate": 9.870048540504592e-06, "loss": 0.533, "step": 1070 }, { "epoch": 0.52, "learning_rate": 9.865453318106934e-06, "loss": 0.5138, "step": 1080 }, { "epoch": 0.52, "learning_rate": 9.860779362586978e-06, "loss": 0.5455, "step": 1090 }, { "epoch": 0.53, "learning_rate": 9.856026749579014e-06, "loss": 0.5369, "step": 1100 }, { "epoch": 0.53, "learning_rate": 9.851195555990178e-06, "loss": 0.5299, "step": 1110 }, { "epoch": 0.54, "learning_rate": 9.846285859999194e-06, "loss": 0.5544, "step": 1120 }, { "epoch": 0.54, "learning_rate": 9.841297741055124e-06, "loss": 0.5408, "step": 1130 }, { "epoch": 0.55, "learning_rate": 9.836231279876076e-06, "loss": 0.52, "step": 1140 }, { "epoch": 0.55, "learning_rate": 9.831086558447895e-06, "loss": 0.5383, "step": 1150 }, { "epoch": 0.56, "learning_rate": 9.82586366002284e-06, "loss": 0.5239, "step": 1160 }, { "epoch": 0.56, "learning_rate": 9.820562669118237e-06, "loss": 0.5239, "step": 1170 }, { "epoch": 0.57, "learning_rate": 9.815183671515107e-06, "loss": 0.5248, "step": 1180 }, { "epoch": 0.57, "learning_rate": 9.809726754256788e-06, "loss": 0.5193, "step": 1190 }, { "epoch": 0.58, "learning_rate": 9.804192005647512e-06, "loss": 0.5258, "step": 1200 }, { "epoch": 0.58, "learning_rate": 9.798579515250988e-06, "loss": 0.5361, "step": 1210 }, { "epoch": 0.59, "learning_rate": 9.792889373888946e-06, "loss": 0.5342, "step": 1220 }, { "epoch": 0.59, "learning_rate": 9.78712167363967e-06, "loss": 0.5263, "step": 1230 }, { "epoch": 0.6, "learning_rate": 9.781276507836513e-06, "loss": 0.5339, "step": 1240 }, { "epoch": 0.6, "learning_rate": 9.77535397106637e-06, "loss": 0.5457, "step": 1250 }, { "epoch": 0.61, "learning_rate": 9.76935415916817e-06, "loss": 0.5322, "step": 1260 }, { "epoch": 0.61, "learning_rate": 9.763277169231307e-06, "loss": 0.535, "step": 1270 }, { "epoch": 0.62, "learning_rate": 9.757123099594078e-06, "loss": 0.5216, "step": 1280 }, { "epoch": 0.62, "learning_rate": 9.750892049842088e-06, "loss": 0.5189, "step": 1290 }, { "epoch": 0.63, "learning_rate": 9.744584120806644e-06, "loss": 0.5265, "step": 1300 }, { "epoch": 0.63, "learning_rate": 9.738199414563113e-06, "loss": 0.5274, "step": 1310 }, { "epoch": 0.64, "learning_rate": 9.731738034429281e-06, "loss": 0.5196, "step": 1320 }, { "epoch": 0.64, "learning_rate": 9.725200084963676e-06, "loss": 0.5204, "step": 1330 }, { "epoch": 0.64, "learning_rate": 9.71858567196387e-06, "loss": 0.5323, "step": 1340 }, { "epoch": 0.65, "learning_rate": 9.711894902464789e-06, "loss": 0.5406, "step": 1350 }, { "epoch": 0.65, "learning_rate": 9.705127884736947e-06, "loss": 0.5274, "step": 1360 }, { "epoch": 0.66, "learning_rate": 9.698284728284723e-06, "loss": 0.5152, "step": 1370 }, { "epoch": 0.66, "learning_rate": 9.691365543844578e-06, "loss": 0.5235, "step": 1380 }, { "epoch": 0.67, "learning_rate": 9.684370443383262e-06, "loss": 0.5234, "step": 1390 }, { "epoch": 0.67, "learning_rate": 9.677299540096002e-06, "loss": 0.5128, "step": 1400 }, { "epoch": 0.68, "learning_rate": 9.670152948404674e-06, "loss": 0.5438, "step": 1410 }, { "epoch": 0.68, "learning_rate": 9.662930783955948e-06, "loss": 0.5242, "step": 1420 }, { "epoch": 0.69, "learning_rate": 9.65563316361942e-06, "loss": 0.5387, "step": 1430 }, { "epoch": 0.69, "learning_rate": 9.648260205485718e-06, "loss": 0.5148, "step": 1440 }, { "epoch": 0.7, "learning_rate": 9.640812028864588e-06, "loss": 0.51, "step": 1450 }, { "epoch": 0.7, "learning_rate": 9.633288754282972e-06, "loss": 0.4974, "step": 1460 }, { "epoch": 0.71, "learning_rate": 9.625690503483051e-06, "loss": 0.5009, "step": 1470 }, { "epoch": 0.71, "learning_rate": 9.618017399420277e-06, "loss": 0.5155, "step": 1480 }, { "epoch": 0.72, "learning_rate": 9.610269566261382e-06, "loss": 0.5268, "step": 1490 }, { "epoch": 0.72, "learning_rate": 9.602447129382368e-06, "loss": 0.5118, "step": 1500 }, { "epoch": 0.72, "eval_webgpt_accuracy": 0.5863125638406538, "eval_webgpt_loss": 0.6783966422080994, "eval_webgpt_runtime": 39.014, "eval_webgpt_samples_per_second": 100.374, "eval_webgpt_steps_per_second": 1.333, "step": 1500 }, { "epoch": 0.72, "eval_hfsummary_accuracy": 0.6642082036091044, "eval_hfsummary_loss": 0.6211047172546387, "eval_hfsummary_runtime": 981.3434, "eval_hfsummary_samples_per_second": 33.712, "eval_hfsummary_steps_per_second": 0.444, "step": 1500 }, { "epoch": 0.72, "eval_anthropic_rlhf_accuracy": 0.6421889616463985, "eval_anthropic_rlhf_loss": 0.6418160796165466, "eval_anthropic_rlhf_runtime": 91.7717, "eval_anthropic_rlhf_samples_per_second": 93.188, "eval_anthropic_rlhf_steps_per_second": 1.231, "step": 1500 }, { "epoch": 0.72, "eval_gptsynthetic_accuracy": 0.9969834087481146, "eval_gptsynthetic_loss": 0.01542825810611248, "eval_gptsynthetic_runtime": 31.9538, "eval_gptsynthetic_samples_per_second": 103.743, "eval_gptsynthetic_steps_per_second": 1.377, "step": 1500 }, { "epoch": 0.73, "learning_rate": 9.594550215366485e-06, "loss": 0.517, "step": 1510 }, { "epoch": 0.73, "learning_rate": 9.586578952002173e-06, "loss": 0.527, "step": 1520 }, { "epoch": 0.74, "learning_rate": 9.578533468281001e-06, "loss": 0.518, "step": 1530 }, { "epoch": 0.74, "learning_rate": 9.570413894395577e-06, "loss": 0.5312, "step": 1540 }, { "epoch": 0.75, "learning_rate": 9.562220361737437e-06, "loss": 0.5052, "step": 1550 }, { "epoch": 0.75, "learning_rate": 9.553953002894933e-06, "loss": 0.5263, "step": 1560 }, { "epoch": 0.76, "learning_rate": 9.545611951651068e-06, "loss": 0.5323, "step": 1570 }, { "epoch": 0.76, "learning_rate": 9.537197342981346e-06, "loss": 0.5091, "step": 1580 }, { "epoch": 0.77, "learning_rate": 9.52870931305158e-06, "loss": 0.5316, "step": 1590 }, { "epoch": 0.77, "learning_rate": 9.520147999215697e-06, "loss": 0.5067, "step": 1600 }, { "epoch": 0.77, "learning_rate": 9.511513540013498e-06, "loss": 0.5102, "step": 1610 }, { "epoch": 0.78, "learning_rate": 9.502806075168441e-06, "loss": 0.5119, "step": 1620 }, { "epoch": 0.78, "learning_rate": 9.494025745585357e-06, "loss": 0.5256, "step": 1630 }, { "epoch": 0.79, "learning_rate": 9.485172693348186e-06, "loss": 0.5119, "step": 1640 }, { "epoch": 0.79, "learning_rate": 9.476247061717666e-06, "loss": 0.5071, "step": 1650 }, { "epoch": 0.8, "learning_rate": 9.467248995129022e-06, "loss": 0.5241, "step": 1660 }, { "epoch": 0.8, "learning_rate": 9.45817863918963e-06, "loss": 0.5155, "step": 1670 }, { "epoch": 0.81, "learning_rate": 9.449036140676652e-06, "loss": 0.4948, "step": 1680 }, { "epoch": 0.81, "learning_rate": 9.43982164753467e-06, "loss": 0.5125, "step": 1690 }, { "epoch": 0.82, "learning_rate": 9.43053530887329e-06, "loss": 0.5095, "step": 1700 }, { "epoch": 0.82, "learning_rate": 9.42117727496472e-06, "loss": 0.5122, "step": 1710 }, { "epoch": 0.83, "learning_rate": 9.411747697241356e-06, "loss": 0.4946, "step": 1720 }, { "epoch": 0.83, "learning_rate": 9.402246728293313e-06, "loss": 0.5244, "step": 1730 }, { "epoch": 0.84, "learning_rate": 9.392674521865968e-06, "loss": 0.4982, "step": 1740 }, { "epoch": 0.84, "learning_rate": 9.383031232857466e-06, "loss": 0.52, "step": 1750 }, { "epoch": 0.85, "learning_rate": 9.373317017316218e-06, "loss": 0.5022, "step": 1760 }, { "epoch": 0.85, "learning_rate": 9.363532032438372e-06, "loss": 0.5131, "step": 1770 }, { "epoch": 0.86, "learning_rate": 9.353676436565271e-06, "loss": 0.5072, "step": 1780 }, { "epoch": 0.86, "learning_rate": 9.343750389180887e-06, "loss": 0.5201, "step": 1790 }, { "epoch": 0.87, "learning_rate": 9.333754050909251e-06, "loss": 0.4917, "step": 1800 }, { "epoch": 0.87, "learning_rate": 9.32368758351184e-06, "loss": 0.4938, "step": 1810 }, { "epoch": 0.88, "learning_rate": 9.313551149884968e-06, "loss": 0.5183, "step": 1820 }, { "epoch": 0.88, "learning_rate": 9.303344914057145e-06, "loss": 0.4944, "step": 1830 }, { "epoch": 0.89, "learning_rate": 9.293069041186426e-06, "loss": 0.5088, "step": 1840 }, { "epoch": 0.89, "learning_rate": 9.282723697557746e-06, "loss": 0.5025, "step": 1850 }, { "epoch": 0.9, "learning_rate": 9.272309050580207e-06, "loss": 0.5302, "step": 1860 }, { "epoch": 0.9, "learning_rate": 9.261825268784395e-06, "loss": 0.4862, "step": 1870 }, { "epoch": 0.9, "learning_rate": 9.25127252181963e-06, "loss": 0.5122, "step": 1880 }, { "epoch": 0.91, "learning_rate": 9.240650980451244e-06, "loss": 0.512, "step": 1890 }, { "epoch": 0.91, "learning_rate": 9.229960816557793e-06, "loss": 0.4975, "step": 1900 }, { "epoch": 0.92, "learning_rate": 9.219202203128293e-06, "loss": 0.4917, "step": 1910 }, { "epoch": 0.92, "learning_rate": 9.20837531425942e-06, "loss": 0.4892, "step": 1920 }, { "epoch": 0.93, "learning_rate": 9.197480325152677e-06, "loss": 0.495, "step": 1930 }, { "epoch": 0.93, "learning_rate": 9.18651741211158e-06, "loss": 0.5009, "step": 1940 }, { "epoch": 0.94, "learning_rate": 9.175486752538792e-06, "loss": 0.526, "step": 1950 }, { "epoch": 0.94, "learning_rate": 9.164388524933253e-06, "loss": 0.505, "step": 1960 }, { "epoch": 0.95, "learning_rate": 9.1532229088873e-06, "loss": 0.4911, "step": 1970 }, { "epoch": 0.95, "learning_rate": 9.141990085083746e-06, "loss": 0.4796, "step": 1980 }, { "epoch": 0.96, "learning_rate": 9.130690235292969e-06, "loss": 0.5006, "step": 1990 }, { "epoch": 0.96, "learning_rate": 9.119323542369969e-06, "loss": 0.4968, "step": 2000 }, { "epoch": 0.96, "eval_webgpt_accuracy": 0.5903983656792645, "eval_webgpt_loss": 0.6728653311729431, "eval_webgpt_runtime": 39.1014, "eval_webgpt_samples_per_second": 100.15, "eval_webgpt_steps_per_second": 1.33, "step": 2000 }, { "epoch": 0.96, "eval_hfsummary_accuracy": 0.6710999607048937, "eval_hfsummary_loss": 0.646239697933197, "eval_hfsummary_runtime": 984.6253, "eval_hfsummary_samples_per_second": 33.6, "eval_hfsummary_steps_per_second": 0.443, "step": 2000 }, { "epoch": 0.96, "eval_anthropic_rlhf_accuracy": 0.6413704396632367, "eval_anthropic_rlhf_loss": 0.6394137740135193, "eval_anthropic_rlhf_runtime": 92.1998, "eval_anthropic_rlhf_samples_per_second": 92.755, "eval_anthropic_rlhf_steps_per_second": 1.226, "step": 2000 }, { "epoch": 0.96, "eval_gptsynthetic_accuracy": 0.9972850678733032, "eval_gptsynthetic_loss": 0.022552527487277985, "eval_gptsynthetic_runtime": 32.0814, "eval_gptsynthetic_samples_per_second": 103.331, "eval_gptsynthetic_steps_per_second": 1.372, "step": 2000 }, { "epoch": 0.97, "learning_rate": 9.107890190251403e-06, "loss": 0.4929, "step": 2010 }, { "epoch": 0.97, "learning_rate": 9.096390363952614e-06, "loss": 0.498, "step": 2020 }, { "epoch": 0.98, "learning_rate": 9.084824249564635e-06, "loss": 0.5041, "step": 2030 }, { "epoch": 0.98, "learning_rate": 9.073192034251174e-06, "loss": 0.4995, "step": 2040 }, { "epoch": 0.99, "learning_rate": 9.061493906245596e-06, "loss": 0.5, "step": 2050 }, { "epoch": 0.99, "learning_rate": 9.049730054847865e-06, "loss": 0.4851, "step": 2060 }, { "epoch": 1.0, "learning_rate": 9.037900670421484e-06, "loss": 0.5055, "step": 2070 }, { "epoch": 1.0, "learning_rate": 9.02600594439042e-06, "loss": 0.5374, "step": 2080 }, { "epoch": 1.01, "learning_rate": 9.014046069236005e-06, "loss": 0.4931, "step": 2090 }, { "epoch": 1.01, "learning_rate": 9.00202123849381e-06, "loss": 0.4837, "step": 2100 }, { "epoch": 1.02, "learning_rate": 8.989931646750533e-06, "loss": 0.4903, "step": 2110 }, { "epoch": 1.02, "learning_rate": 8.97777748964083e-06, "loss": 0.4727, "step": 2120 }, { "epoch": 1.03, "learning_rate": 8.965558963844165e-06, "loss": 0.479, "step": 2130 }, { "epoch": 1.03, "learning_rate": 8.953276267081618e-06, "loss": 0.4769, "step": 2140 }, { "epoch": 1.04, "learning_rate": 8.940929598112687e-06, "loss": 0.4967, "step": 2150 }, { "epoch": 1.04, "learning_rate": 8.928519156732074e-06, "loss": 0.4734, "step": 2160 }, { "epoch": 1.04, "learning_rate": 8.916045143766455e-06, "loss": 0.5009, "step": 2170 }, { "epoch": 1.05, "learning_rate": 8.90350776107122e-06, "loss": 0.469, "step": 2180 }, { "epoch": 1.05, "learning_rate": 8.890907211527213e-06, "loss": 0.4942, "step": 2190 }, { "epoch": 1.06, "learning_rate": 8.878243699037452e-06, "loss": 0.468, "step": 2200 }, { "epoch": 1.06, "learning_rate": 8.865517428523826e-06, "loss": 0.4829, "step": 2210 }, { "epoch": 1.07, "learning_rate": 8.852728605923774e-06, "loss": 0.501, "step": 2220 }, { "epoch": 1.07, "learning_rate": 8.839877438186966e-06, "loss": 0.4659, "step": 2230 }, { "epoch": 1.08, "learning_rate": 8.826964133271937e-06, "loss": 0.4634, "step": 2240 }, { "epoch": 1.08, "learning_rate": 8.813988900142734e-06, "loss": 0.4656, "step": 2250 }, { "epoch": 1.09, "learning_rate": 8.800951948765531e-06, "loss": 0.4709, "step": 2260 }, { "epoch": 1.09, "learning_rate": 8.787853490105233e-06, "loss": 0.4657, "step": 2270 }, { "epoch": 1.1, "learning_rate": 8.774693736122055e-06, "loss": 0.4773, "step": 2280 }, { "epoch": 1.1, "learning_rate": 8.761472899768101e-06, "loss": 0.4796, "step": 2290 }, { "epoch": 1.11, "learning_rate": 8.748191194983919e-06, "loss": 0.4823, "step": 2300 }, { "epoch": 1.11, "learning_rate": 8.734848836695023e-06, "loss": 0.4655, "step": 2310 }, { "epoch": 1.12, "learning_rate": 8.721446040808438e-06, "loss": 0.4838, "step": 2320 }, { "epoch": 1.12, "learning_rate": 8.707983024209186e-06, "loss": 0.4927, "step": 2330 }, { "epoch": 1.13, "learning_rate": 8.694460004756791e-06, "loss": 0.4772, "step": 2340 }, { "epoch": 1.13, "learning_rate": 8.680877201281746e-06, "loss": 0.4792, "step": 2350 }, { "epoch": 1.14, "learning_rate": 8.66723483358197e-06, "loss": 0.4739, "step": 2360 }, { "epoch": 1.14, "learning_rate": 8.653533122419257e-06, "loss": 0.4876, "step": 2370 }, { "epoch": 1.15, "learning_rate": 8.639772289515705e-06, "loss": 0.4471, "step": 2380 }, { "epoch": 1.15, "learning_rate": 8.625952557550119e-06, "loss": 0.4641, "step": 2390 }, { "epoch": 1.16, "learning_rate": 8.612074150154412e-06, "loss": 0.468, "step": 2400 }, { "epoch": 1.16, "learning_rate": 8.598137291909996e-06, "loss": 0.4665, "step": 2410 }, { "epoch": 1.17, "learning_rate": 8.584142208344128e-06, "loss": 0.452, "step": 2420 }, { "epoch": 1.17, "learning_rate": 8.570089125926279e-06, "loss": 0.4811, "step": 2430 }, { "epoch": 1.17, "learning_rate": 8.555978272064456e-06, "loss": 0.4823, "step": 2440 }, { "epoch": 1.18, "learning_rate": 8.541809875101535e-06, "loss": 0.4692, "step": 2450 }, { "epoch": 1.18, "learning_rate": 8.527584164311548e-06, "loss": 0.4704, "step": 2460 }, { "epoch": 1.19, "learning_rate": 8.513301369895998e-06, "loss": 0.4721, "step": 2470 }, { "epoch": 1.19, "learning_rate": 8.498961722980106e-06, "loss": 0.4416, "step": 2480 }, { "epoch": 1.2, "learning_rate": 8.48456545560909e-06, "loss": 0.4708, "step": 2490 }, { "epoch": 1.2, "learning_rate": 8.470112800744406e-06, "loss": 0.4855, "step": 2500 }, { "epoch": 1.2, "eval_webgpt_accuracy": 0.5863125638406538, "eval_webgpt_loss": 0.6755186319351196, "eval_webgpt_runtime": 39.3033, "eval_webgpt_samples_per_second": 99.635, "eval_webgpt_steps_per_second": 1.323, "step": 2500 }, { "epoch": 1.2, "eval_hfsummary_accuracy": 0.6743342502191458, "eval_hfsummary_loss": 0.6676438450813293, "eval_hfsummary_runtime": 986.7916, "eval_hfsummary_samples_per_second": 33.526, "eval_hfsummary_steps_per_second": 0.442, "step": 2500 }, { "epoch": 1.2, "eval_anthropic_rlhf_accuracy": 0.6465154349859682, "eval_anthropic_rlhf_loss": 0.635563850402832, "eval_anthropic_rlhf_runtime": 92.2813, "eval_anthropic_rlhf_samples_per_second": 92.673, "eval_anthropic_rlhf_steps_per_second": 1.225, "step": 2500 }, { "epoch": 1.2, "eval_gptsynthetic_accuracy": 0.9975867269984917, "eval_gptsynthetic_loss": 0.016244370490312576, "eval_gptsynthetic_runtime": 32.0562, "eval_gptsynthetic_samples_per_second": 103.412, "eval_gptsynthetic_steps_per_second": 1.373, "step": 2500 }, { "epoch": 1.21, "learning_rate": 8.455603992259972e-06, "loss": 0.4741, "step": 2510 }, { "epoch": 1.21, "learning_rate": 8.441039264938393e-06, "loss": 0.4763, "step": 2520 }, { "epoch": 1.22, "learning_rate": 8.426418854467154e-06, "loss": 0.4687, "step": 2530 }, { "epoch": 1.22, "learning_rate": 8.411742997434809e-06, "loss": 0.4707, "step": 2540 }, { "epoch": 1.23, "learning_rate": 8.39701193132715e-06, "loss": 0.4658, "step": 2550 }, { "epoch": 1.23, "learning_rate": 8.382225894523373e-06, "loss": 0.4709, "step": 2560 }, { "epoch": 1.24, "learning_rate": 8.367385126292207e-06, "loss": 0.475, "step": 2570 }, { "epoch": 1.24, "learning_rate": 8.352489866788051e-06, "loss": 0.4764, "step": 2580 }, { "epoch": 1.25, "learning_rate": 8.337540357047087e-06, "loss": 0.4954, "step": 2590 }, { "epoch": 1.25, "learning_rate": 8.32253683898338e-06, "loss": 0.4626, "step": 2600 }, { "epoch": 1.26, "learning_rate": 8.307479555384954e-06, "loss": 0.4724, "step": 2610 }, { "epoch": 1.26, "learning_rate": 8.29236874990988e-06, "loss": 0.4564, "step": 2620 }, { "epoch": 1.27, "learning_rate": 8.27720466708232e-06, "loss": 0.4582, "step": 2630 }, { "epoch": 1.27, "learning_rate": 8.261987552288573e-06, "loss": 0.482, "step": 2640 }, { "epoch": 1.28, "learning_rate": 8.246717651773108e-06, "loss": 0.4907, "step": 2650 }, { "epoch": 1.28, "learning_rate": 8.231395212634572e-06, "loss": 0.4576, "step": 2660 }, { "epoch": 1.29, "learning_rate": 8.216020482821797e-06, "loss": 0.4544, "step": 2670 }, { "epoch": 1.29, "learning_rate": 8.200593711129796e-06, "loss": 0.4665, "step": 2680 }, { "epoch": 1.3, "learning_rate": 8.18511514719571e-06, "loss": 0.4729, "step": 2690 }, { "epoch": 1.3, "learning_rate": 8.1695850414948e-06, "loss": 0.4891, "step": 2700 }, { "epoch": 1.3, "learning_rate": 8.154003645336378e-06, "loss": 0.4538, "step": 2710 }, { "epoch": 1.31, "learning_rate": 8.138371210859732e-06, "loss": 0.476, "step": 2720 }, { "epoch": 1.31, "learning_rate": 8.122687991030067e-06, "loss": 0.4519, "step": 2730 }, { "epoch": 1.32, "learning_rate": 8.106954239634394e-06, "loss": 0.4644, "step": 2740 }, { "epoch": 1.32, "learning_rate": 8.091170211277437e-06, "loss": 0.4551, "step": 2750 }, { "epoch": 1.33, "learning_rate": 8.075336161377492e-06, "loss": 0.4765, "step": 2760 }, { "epoch": 1.33, "learning_rate": 8.059452346162321e-06, "loss": 0.4481, "step": 2770 }, { "epoch": 1.34, "learning_rate": 8.043519022664984e-06, "loss": 0.47, "step": 2780 }, { "epoch": 1.34, "learning_rate": 8.027536448719692e-06, "loss": 0.4955, "step": 2790 }, { "epoch": 1.35, "learning_rate": 8.011504882957626e-06, "loss": 0.4688, "step": 2800 }, { "epoch": 1.35, "learning_rate": 7.99542458480276e-06, "loss": 0.447, "step": 2810 }, { "epoch": 1.36, "learning_rate": 7.979295814467656e-06, "loss": 0.4581, "step": 2820 }, { "epoch": 1.36, "learning_rate": 7.963118832949259e-06, "loss": 0.4857, "step": 2830 }, { "epoch": 1.37, "learning_rate": 7.94689390202467e-06, "loss": 0.458, "step": 2840 }, { "epoch": 1.37, "learning_rate": 7.930621284246914e-06, "loss": 0.4797, "step": 2850 }, { "epoch": 1.38, "learning_rate": 7.914301242940678e-06, "loss": 0.4722, "step": 2860 }, { "epoch": 1.38, "learning_rate": 7.897934042198073e-06, "loss": 0.4499, "step": 2870 }, { "epoch": 1.39, "learning_rate": 7.881519946874336e-06, "loss": 0.4457, "step": 2880 }, { "epoch": 1.39, "learning_rate": 7.865059222583565e-06, "loss": 0.4563, "step": 2890 }, { "epoch": 1.4, "learning_rate": 7.848552135694407e-06, "loss": 0.4753, "step": 2900 }, { "epoch": 1.4, "learning_rate": 7.83199895332575e-06, "loss": 0.4595, "step": 2910 }, { "epoch": 1.41, "learning_rate": 7.815399943342408e-06, "loss": 0.4529, "step": 2920 }, { "epoch": 1.41, "learning_rate": 7.798755374350774e-06, "loss": 0.4634, "step": 2930 }, { "epoch": 1.42, "learning_rate": 7.782065515694486e-06, "loss": 0.4671, "step": 2940 }, { "epoch": 1.42, "learning_rate": 7.765330637450057e-06, "loss": 0.4921, "step": 2950 }, { "epoch": 1.42, "learning_rate": 7.748551010422518e-06, "loss": 0.4841, "step": 2960 }, { "epoch": 1.43, "learning_rate": 7.731726906141024e-06, "loss": 0.4348, "step": 2970 }, { "epoch": 1.43, "learning_rate": 7.71485859685446e-06, "loss": 0.4612, "step": 2980 }, { "epoch": 1.44, "learning_rate": 7.697946355527044e-06, "loss": 0.4517, "step": 2990 }, { "epoch": 1.44, "learning_rate": 7.680990455833907e-06, "loss": 0.4514, "step": 3000 }, { "epoch": 1.44, "eval_webgpt_accuracy": 0.5827374872318692, "eval_webgpt_loss": 0.6752701997756958, "eval_webgpt_runtime": 39.1231, "eval_webgpt_samples_per_second": 100.094, "eval_webgpt_steps_per_second": 1.329, "step": 3000 }, { "epoch": 1.44, "eval_hfsummary_accuracy": 0.6794123870265696, "eval_hfsummary_loss": 0.6801344156265259, "eval_hfsummary_runtime": 986.8465, "eval_hfsummary_samples_per_second": 33.524, "eval_hfsummary_steps_per_second": 0.442, "step": 3000 }, { "epoch": 1.44, "eval_anthropic_rlhf_accuracy": 0.657156220767072, "eval_anthropic_rlhf_loss": 0.6261478662490845, "eval_anthropic_rlhf_runtime": 92.2987, "eval_anthropic_rlhf_samples_per_second": 92.656, "eval_anthropic_rlhf_steps_per_second": 1.224, "step": 3000 }, { "epoch": 1.44, "eval_gptsynthetic_accuracy": 0.9969834087481146, "eval_gptsynthetic_loss": 0.009839520789682865, "eval_gptsynthetic_runtime": 32.2002, "eval_gptsynthetic_samples_per_second": 102.95, "eval_gptsynthetic_steps_per_second": 1.366, "step": 3000 }, { "epoch": 1.45, "learning_rate": 7.663991172156662e-06, "loss": 0.4668, "step": 3010 }, { "epoch": 1.45, "learning_rate": 7.646948779578961e-06, "loss": 0.4689, "step": 3020 }, { "epoch": 1.46, "learning_rate": 7.62986355388205e-06, "loss": 0.4544, "step": 3030 }, { "epoch": 1.46, "learning_rate": 7.612735771540308e-06, "loss": 0.4549, "step": 3040 }, { "epoch": 1.47, "learning_rate": 7.59556570971676e-06, "loss": 0.4738, "step": 3050 }, { "epoch": 1.47, "learning_rate": 7.5783536462586076e-06, "loss": 0.4585, "step": 3060 }, { "epoch": 1.48, "learning_rate": 7.561099859692722e-06, "loss": 0.4782, "step": 3070 }, { "epoch": 1.48, "learning_rate": 7.543804629221143e-06, "loss": 0.475, "step": 3080 }, { "epoch": 1.49, "learning_rate": 7.526468234716557e-06, "loss": 0.465, "step": 3090 }, { "epoch": 1.49, "learning_rate": 7.509090956717775e-06, "loss": 0.4689, "step": 3100 }, { "epoch": 1.5, "learning_rate": 7.49167307642518e-06, "loss": 0.4652, "step": 3110 }, { "epoch": 1.5, "learning_rate": 7.474214875696188e-06, "loss": 0.4409, "step": 3120 }, { "epoch": 1.51, "learning_rate": 7.456716637040683e-06, "loss": 0.4637, "step": 3130 }, { "epoch": 1.51, "learning_rate": 7.439178643616447e-06, "loss": 0.4402, "step": 3140 }, { "epoch": 1.52, "learning_rate": 7.421601179224574e-06, "loss": 0.4522, "step": 3150 }, { "epoch": 1.52, "learning_rate": 7.403984528304884e-06, "loss": 0.4457, "step": 3160 }, { "epoch": 1.53, "learning_rate": 7.386328975931311e-06, "loss": 0.4591, "step": 3170 }, { "epoch": 1.53, "learning_rate": 7.368634807807298e-06, "loss": 0.4509, "step": 3180 }, { "epoch": 1.54, "learning_rate": 7.3509023102611745e-06, "loss": 0.4404, "step": 3190 }, { "epoch": 1.54, "learning_rate": 7.33313177024151e-06, "loss": 0.461, "step": 3200 }, { "epoch": 1.55, "learning_rate": 7.315323475312489e-06, "loss": 0.4479, "step": 3210 }, { "epoch": 1.55, "learning_rate": 7.297477713649247e-06, "loss": 0.4604, "step": 3220 }, { "epoch": 1.55, "learning_rate": 7.279594774033205e-06, "loss": 0.4488, "step": 3230 }, { "epoch": 1.56, "learning_rate": 7.261674945847406e-06, "loss": 0.4634, "step": 3240 }, { "epoch": 1.56, "learning_rate": 7.2437185190718205e-06, "loss": 0.4503, "step": 3250 }, { "epoch": 1.57, "learning_rate": 7.225725784278663e-06, "loss": 0.4605, "step": 3260 }, { "epoch": 1.57, "learning_rate": 7.207697032627689e-06, "loss": 0.446, "step": 3270 }, { "epoch": 1.58, "learning_rate": 7.18963255586148e-06, "loss": 0.4385, "step": 3280 }, { "epoch": 1.58, "learning_rate": 7.17153264630072e-06, "loss": 0.45, "step": 3290 }, { "epoch": 1.59, "learning_rate": 7.153397596839475e-06, "loss": 0.4611, "step": 3300 }, { "epoch": 1.59, "learning_rate": 7.1352277009404454e-06, "loss": 0.4495, "step": 3310 }, { "epoch": 1.6, "learning_rate": 7.117023252630216e-06, "loss": 0.4377, "step": 3320 }, { "epoch": 1.6, "learning_rate": 7.098784546494507e-06, "loss": 0.4517, "step": 3330 }, { "epoch": 1.61, "learning_rate": 7.080511877673397e-06, "loss": 0.4461, "step": 3340 }, { "epoch": 1.61, "learning_rate": 7.062205541856552e-06, "loss": 0.4587, "step": 3350 }, { "epoch": 1.62, "learning_rate": 7.04386583527844e-06, "loss": 0.4473, "step": 3360 }, { "epoch": 1.62, "learning_rate": 7.025493054713542e-06, "loss": 0.4516, "step": 3370 }, { "epoch": 1.63, "learning_rate": 7.007087497471532e-06, "loss": 0.4698, "step": 3380 }, { "epoch": 1.63, "learning_rate": 6.988649461392491e-06, "loss": 0.4415, "step": 3390 }, { "epoch": 1.64, "learning_rate": 6.970179244842074e-06, "loss": 0.4593, "step": 3400 }, { "epoch": 1.64, "learning_rate": 6.9516771467066726e-06, "loss": 0.4428, "step": 3410 }, { "epoch": 1.65, "learning_rate": 6.933143466388599e-06, "loss": 0.4606, "step": 3420 }, { "epoch": 1.65, "learning_rate": 6.914578503801228e-06, "loss": 0.457, "step": 3430 }, { "epoch": 1.66, "learning_rate": 6.8959825593641425e-06, "loss": 0.4408, "step": 3440 }, { "epoch": 1.66, "learning_rate": 6.87735593399828e-06, "loss": 0.4636, "step": 3450 }, { "epoch": 1.67, "learning_rate": 6.858698929121058e-06, "loss": 0.4519, "step": 3460 }, { "epoch": 1.67, "learning_rate": 6.840011846641497e-06, "loss": 0.4586, "step": 3470 }, { "epoch": 1.68, "learning_rate": 6.821294988955334e-06, "loss": 0.4316, "step": 3480 }, { "epoch": 1.68, "learning_rate": 6.802548658940136e-06, "loss": 0.4685, "step": 3490 }, { "epoch": 1.68, "learning_rate": 6.783773159950386e-06, "loss": 0.4427, "step": 3500 }, { "epoch": 1.68, "eval_webgpt_accuracy": 0.5957609805924413, "eval_webgpt_loss": 0.674083948135376, "eval_webgpt_runtime": 39.2401, "eval_webgpt_samples_per_second": 99.796, "eval_webgpt_steps_per_second": 1.325, "step": 3500 }, { "epoch": 1.68, "eval_hfsummary_accuracy": 0.6743040232143397, "eval_hfsummary_loss": 0.6820210814476013, "eval_hfsummary_runtime": 984.942, "eval_hfsummary_samples_per_second": 33.589, "eval_hfsummary_steps_per_second": 0.443, "step": 3500 }, { "epoch": 1.68, "eval_anthropic_rlhf_accuracy": 0.6555191768007483, "eval_anthropic_rlhf_loss": 0.6293210983276367, "eval_anthropic_rlhf_runtime": 92.3544, "eval_anthropic_rlhf_samples_per_second": 92.6, "eval_anthropic_rlhf_steps_per_second": 1.224, "step": 3500 }, { "epoch": 1.68, "eval_gptsynthetic_accuracy": 0.9975867269984917, "eval_gptsynthetic_loss": 0.01304171234369278, "eval_gptsynthetic_runtime": 32.2744, "eval_gptsynthetic_samples_per_second": 102.713, "eval_gptsynthetic_steps_per_second": 1.363, "step": 3500 }, { "epoch": 1.69, "learning_rate": 6.764968795812587e-06, "loss": 0.4489, "step": 3510 }, { "epoch": 1.69, "learning_rate": 6.746135870820338e-06, "loss": 0.4492, "step": 3520 }, { "epoch": 1.7, "learning_rate": 6.727274689729411e-06, "loss": 0.4482, "step": 3530 }, { "epoch": 1.7, "learning_rate": 6.7083855577528234e-06, "loss": 0.455, "step": 3540 }, { "epoch": 1.71, "learning_rate": 6.689468780555892e-06, "loss": 0.4285, "step": 3550 }, { "epoch": 1.71, "learning_rate": 6.670524664251291e-06, "loss": 0.4485, "step": 3560 }, { "epoch": 1.72, "learning_rate": 6.651553515394102e-06, "loss": 0.4534, "step": 3570 }, { "epoch": 1.72, "learning_rate": 6.632555640976846e-06, "loss": 0.4551, "step": 3580 }, { "epoch": 1.73, "learning_rate": 6.613531348424518e-06, "loss": 0.4409, "step": 3590 }, { "epoch": 1.73, "learning_rate": 6.5944809455896145e-06, "loss": 0.444, "step": 3600 }, { "epoch": 1.74, "learning_rate": 6.575404740747155e-06, "loss": 0.4487, "step": 3610 }, { "epoch": 1.74, "learning_rate": 6.55630304258968e-06, "loss": 0.4339, "step": 3620 }, { "epoch": 1.75, "learning_rate": 6.537176160222271e-06, "loss": 0.4464, "step": 3630 }, { "epoch": 1.75, "learning_rate": 6.518024403157544e-06, "loss": 0.4342, "step": 3640 }, { "epoch": 1.76, "learning_rate": 6.4988480813106305e-06, "loss": 0.4467, "step": 3650 }, { "epoch": 1.76, "learning_rate": 6.47964750499418e-06, "loss": 0.4536, "step": 3660 }, { "epoch": 1.77, "learning_rate": 6.460422984913326e-06, "loss": 0.455, "step": 3670 }, { "epoch": 1.77, "learning_rate": 6.44117483216066e-06, "loss": 0.4372, "step": 3680 }, { "epoch": 1.78, "learning_rate": 6.4219033582112026e-06, "loss": 0.431, "step": 3690 }, { "epoch": 1.78, "learning_rate": 6.402608874917358e-06, "loss": 0.438, "step": 3700 }, { "epoch": 1.79, "learning_rate": 6.383291694503868e-06, "loss": 0.4606, "step": 3710 }, { "epoch": 1.79, "learning_rate": 6.3639521295627635e-06, "loss": 0.4256, "step": 3720 }, { "epoch": 1.8, "learning_rate": 6.344590493048304e-06, "loss": 0.4561, "step": 3730 }, { "epoch": 1.8, "learning_rate": 6.325207098271907e-06, "loss": 0.4479, "step": 3740 }, { "epoch": 1.81, "learning_rate": 6.30580225889709e-06, "loss": 0.4541, "step": 3750 }, { "epoch": 1.81, "learning_rate": 6.286376288934388e-06, "loss": 0.4371, "step": 3760 }, { "epoch": 1.81, "learning_rate": 6.266929502736268e-06, "loss": 0.4312, "step": 3770 }, { "epoch": 1.82, "learning_rate": 6.24746221499205e-06, "loss": 0.4311, "step": 3780 }, { "epoch": 1.82, "learning_rate": 6.227974740722815e-06, "loss": 0.4313, "step": 3790 }, { "epoch": 1.83, "learning_rate": 6.208467395276298e-06, "loss": 0.4704, "step": 3800 }, { "epoch": 1.83, "learning_rate": 6.188940494321794e-06, "loss": 0.4417, "step": 3810 }, { "epoch": 1.84, "learning_rate": 6.169394353845045e-06, "loss": 0.4341, "step": 3820 }, { "epoch": 1.84, "learning_rate": 6.149829290143129e-06, "loss": 0.4558, "step": 3830 }, { "epoch": 1.85, "learning_rate": 6.1302456198193415e-06, "loss": 0.4511, "step": 3840 }, { "epoch": 1.85, "learning_rate": 6.110643659778073e-06, "loss": 0.4284, "step": 3850 }, { "epoch": 1.86, "learning_rate": 6.091023727219675e-06, "loss": 0.4459, "step": 3860 }, { "epoch": 1.86, "learning_rate": 6.0713861396353355e-06, "loss": 0.4321, "step": 3870 }, { "epoch": 1.87, "learning_rate": 6.051731214801939e-06, "loss": 0.447, "step": 3880 }, { "epoch": 1.87, "learning_rate": 6.032059270776917e-06, "loss": 0.4342, "step": 3890 }, { "epoch": 1.88, "learning_rate": 6.012370625893112e-06, "loss": 0.4391, "step": 3900 }, { "epoch": 1.88, "learning_rate": 5.992665598753619e-06, "loss": 0.4408, "step": 3910 }, { "epoch": 1.89, "learning_rate": 5.972944508226629e-06, "loss": 0.4514, "step": 3920 }, { "epoch": 1.89, "learning_rate": 5.953207673440278e-06, "loss": 0.4748, "step": 3930 }, { "epoch": 1.9, "learning_rate": 5.933455413777473e-06, "loss": 0.4508, "step": 3940 }, { "epoch": 1.9, "learning_rate": 5.913688048870725e-06, "loss": 0.4583, "step": 3950 }, { "epoch": 1.91, "learning_rate": 5.893905898596986e-06, "loss": 0.4331, "step": 3960 }, { "epoch": 1.91, "learning_rate": 5.87410928307246e-06, "loss": 0.451, "step": 3970 }, { "epoch": 1.92, "learning_rate": 5.85429852264743e-06, "loss": 0.4601, "step": 3980 }, { "epoch": 1.92, "learning_rate": 5.834473937901075e-06, "loss": 0.4388, "step": 3990 }, { "epoch": 1.93, "learning_rate": 5.814635849636281e-06, "loss": 0.4362, "step": 4000 }, { "epoch": 1.93, "eval_webgpt_accuracy": 0.5914198161389173, "eval_webgpt_loss": 0.672228991985321, "eval_webgpt_runtime": 39.1644, "eval_webgpt_samples_per_second": 99.989, "eval_webgpt_steps_per_second": 1.328, "step": 4000 }, { "epoch": 1.93, "eval_hfsummary_accuracy": 0.676631502584409, "eval_hfsummary_loss": 0.7084701657295227, "eval_hfsummary_runtime": 988.7214, "eval_hfsummary_samples_per_second": 33.46, "eval_hfsummary_steps_per_second": 0.441, "step": 4000 }, { "epoch": 1.93, "eval_anthropic_rlhf_accuracy": 0.6534144059869037, "eval_anthropic_rlhf_loss": 0.6274996995925903, "eval_anthropic_rlhf_runtime": 92.1951, "eval_anthropic_rlhf_samples_per_second": 92.76, "eval_anthropic_rlhf_steps_per_second": 1.226, "step": 4000 }, { "epoch": 1.93, "eval_gptsynthetic_accuracy": 0.9969834087481146, "eval_gptsynthetic_loss": 0.012532506138086319, "eval_gptsynthetic_runtime": 32.2302, "eval_gptsynthetic_samples_per_second": 102.854, "eval_gptsynthetic_steps_per_second": 1.365, "step": 4000 } ], "max_steps": 4154, "num_train_epochs": 2, "total_flos": 4.526360616326298e+17, "trial_name": null, "trial_params": null }