File size: 32,949 Bytes
563066e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 5.0,
  "eval_steps": 1000,
  "global_step": 625,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.008,
      "grad_norm": 7.598883452709097,
      "learning_rate": 1.5873015873015874e-07,
      "logits/chosen": -1.7671998739242554,
      "logits/rejected": -2.2639822959899902,
      "logps/chosen": -46.430763244628906,
      "logps/rejected": -102.85381317138672,
      "loss": 0.6931,
      "rewards/accuracies": 0.0,
      "rewards/chosen": 0.0,
      "rewards/margins": 0.0,
      "rewards/rejected": 0.0,
      "step": 1
    },
    {
      "epoch": 0.08,
      "grad_norm": 9.021711112045313,
      "learning_rate": 1.5873015873015873e-06,
      "logits/chosen": -1.6467827558517456,
      "logits/rejected": -2.05173921585083,
      "logps/chosen": -50.05971145629883,
      "logps/rejected": -110.39069366455078,
      "loss": 0.6403,
      "rewards/accuracies": 0.7777777910232544,
      "rewards/chosen": 0.0030777277424931526,
      "rewards/margins": 0.11931005865335464,
      "rewards/rejected": -0.11623234301805496,
      "step": 10
    },
    {
      "epoch": 0.16,
      "grad_norm": 1.8190332435996819,
      "learning_rate": 3.1746031746031746e-06,
      "logits/chosen": -1.3254443407058716,
      "logits/rejected": -1.779170036315918,
      "logps/chosen": -37.68037033081055,
      "logps/rejected": -313.1996154785156,
      "loss": 0.2472,
      "rewards/accuracies": 0.9937499761581421,
      "rewards/chosen": 0.09366801381111145,
      "rewards/margins": 2.184483051300049,
      "rewards/rejected": -2.0908150672912598,
      "step": 20
    },
    {
      "epoch": 0.24,
      "grad_norm": 8.393638030400462,
      "learning_rate": 4.761904761904762e-06,
      "logits/chosen": -0.7945032715797424,
      "logits/rejected": -1.1047414541244507,
      "logps/chosen": -55.5176887512207,
      "logps/rejected": -406.99609375,
      "loss": 0.1799,
      "rewards/accuracies": 0.925000011920929,
      "rewards/chosen": -0.04824959114193916,
      "rewards/margins": 3.0284643173217773,
      "rewards/rejected": -3.076713800430298,
      "step": 30
    },
    {
      "epoch": 0.32,
      "grad_norm": 3.6170805565428488,
      "learning_rate": 6.349206349206349e-06,
      "logits/chosen": -0.3710061311721802,
      "logits/rejected": -0.7456581592559814,
      "logps/chosen": -50.16427230834961,
      "logps/rejected": -541.6275634765625,
      "loss": 0.1374,
      "rewards/accuracies": 0.9750000238418579,
      "rewards/chosen": -0.01431644894182682,
      "rewards/margins": 4.384586334228516,
      "rewards/rejected": -4.398902416229248,
      "step": 40
    },
    {
      "epoch": 0.4,
      "grad_norm": 3.795764554921293,
      "learning_rate": 7.936507936507936e-06,
      "logits/chosen": -0.581534206867218,
      "logits/rejected": -0.8648965954780579,
      "logps/chosen": -80.1789779663086,
      "logps/rejected": -672.813720703125,
      "loss": 0.0553,
      "rewards/accuracies": 0.987500011920929,
      "rewards/chosen": -0.3146038353443146,
      "rewards/margins": 5.392758846282959,
      "rewards/rejected": -5.707362174987793,
      "step": 50
    },
    {
      "epoch": 0.48,
      "grad_norm": 1.397363269705544,
      "learning_rate": 9.523809523809525e-06,
      "logits/chosen": -0.7689735293388367,
      "logits/rejected": -0.9561458826065063,
      "logps/chosen": -145.53469848632812,
      "logps/rejected": -831.8748779296875,
      "loss": 0.0371,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.9438420534133911,
      "rewards/margins": 6.384782314300537,
      "rewards/rejected": -7.328624725341797,
      "step": 60
    },
    {
      "epoch": 0.56,
      "grad_norm": 0.13618082387486424,
      "learning_rate": 9.996172565322375e-06,
      "logits/chosen": -1.2459557056427002,
      "logits/rejected": -1.4872663021087646,
      "logps/chosen": -132.92080688476562,
      "logps/rejected": -911.9078979492188,
      "loss": 0.0409,
      "rewards/accuracies": 0.9937499761581421,
      "rewards/chosen": -0.8237069249153137,
      "rewards/margins": 7.28142786026001,
      "rewards/rejected": -8.105135917663574,
      "step": 70
    },
    {
      "epoch": 0.64,
      "grad_norm": 1.3606359582288166,
      "learning_rate": 9.97744005136599e-06,
      "logits/chosen": -1.503824234008789,
      "logits/rejected": -1.8992702960968018,
      "logps/chosen": -115.5097885131836,
      "logps/rejected": -916.9987182617188,
      "loss": 0.013,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6468590497970581,
      "rewards/margins": 7.513754367828369,
      "rewards/rejected": -8.160614013671875,
      "step": 80
    },
    {
      "epoch": 0.72,
      "grad_norm": 8.135027509581384,
      "learning_rate": 9.943157907471825e-06,
      "logits/chosen": -1.5570601224899292,
      "logits/rejected": -1.9588285684585571,
      "logps/chosen": -122.7470703125,
      "logps/rejected": -1007.5989379882812,
      "loss": 0.0185,
      "rewards/accuracies": 0.9937499761581421,
      "rewards/chosen": -0.71613609790802,
      "rewards/margins": 8.348057746887207,
      "rewards/rejected": -9.064192771911621,
      "step": 90
    },
    {
      "epoch": 0.8,
      "grad_norm": 0.07850356571109195,
      "learning_rate": 9.893433231795864e-06,
      "logits/chosen": -2.478653907775879,
      "logits/rejected": -2.867173671722412,
      "logps/chosen": -142.598388671875,
      "logps/rejected": -989.5442504882812,
      "loss": 0.0079,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.9251037836074829,
      "rewards/margins": 7.951480865478516,
      "rewards/rejected": -8.876585006713867,
      "step": 100
    },
    {
      "epoch": 0.88,
      "grad_norm": 0.010143597010992798,
      "learning_rate": 9.828421365296023e-06,
      "logits/chosen": -1.9893125295639038,
      "logits/rejected": -2.527003526687622,
      "logps/chosen": -127.56657409667969,
      "logps/rejected": -1111.281005859375,
      "loss": 0.0025,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.7765494585037231,
      "rewards/margins": 9.32172679901123,
      "rewards/rejected": -10.098276138305664,
      "step": 110
    },
    {
      "epoch": 0.96,
      "grad_norm": 0.005924828079353313,
      "learning_rate": 9.748325406443647e-06,
      "logits/chosen": -1.330843448638916,
      "logits/rejected": -2.0016732215881348,
      "logps/chosen": -80.08639526367188,
      "logps/rejected": -1047.385009765625,
      "loss": 0.0004,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.33333706855773926,
      "rewards/margins": 9.099153518676758,
      "rewards/rejected": -9.43249225616455,
      "step": 120
    },
    {
      "epoch": 1.04,
      "grad_norm": 0.006907045924266673,
      "learning_rate": 9.653395576739504e-06,
      "logits/chosen": -1.3165078163146973,
      "logits/rejected": -1.999441385269165,
      "logps/chosen": -142.83404541015625,
      "logps/rejected": -1252.3265380859375,
      "loss": 0.004,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.9162995219230652,
      "rewards/margins": 10.60672378540039,
      "rewards/rejected": -11.52302360534668,
      "step": 130
    },
    {
      "epoch": 1.12,
      "grad_norm": 0.11065788109459221,
      "learning_rate": 9.543928439016445e-06,
      "logits/chosen": -2.337254047393799,
      "logits/rejected": -2.8029887676239014,
      "logps/chosen": -141.67652893066406,
      "logps/rejected": -1141.8990478515625,
      "loss": 0.0013,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.9172345399856567,
      "rewards/margins": 9.48987102508545,
      "rewards/rejected": -10.407105445861816,
      "step": 140
    },
    {
      "epoch": 1.2,
      "grad_norm": 0.08367928566939553,
      "learning_rate": 9.42026597097071e-06,
      "logits/chosen": -2.5381298065185547,
      "logits/rejected": -3.0234386920928955,
      "logps/chosen": -162.30075073242188,
      "logps/rejected": -1239.561767578125,
      "loss": 0.0006,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -1.1074072122573853,
      "rewards/margins": 10.28199291229248,
      "rewards/rejected": -11.389399528503418,
      "step": 150
    },
    {
      "epoch": 1.28,
      "grad_norm": 0.0036761092850006146,
      "learning_rate": 9.282794496816244e-06,
      "logits/chosen": -2.3002023696899414,
      "logits/rejected": -2.8048605918884277,
      "logps/chosen": -136.76681518554688,
      "logps/rejected": -1181.3677978515625,
      "loss": 0.0008,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.8667460680007935,
      "rewards/margins": 9.932957649230957,
      "rewards/rejected": -10.799702644348145,
      "step": 160
    },
    {
      "epoch": 1.3599999999999999,
      "grad_norm": 0.0021196612287394884,
      "learning_rate": 9.131943480399531e-06,
      "logits/chosen": -2.4460995197296143,
      "logits/rejected": -2.9539546966552734,
      "logps/chosen": -154.1061553955078,
      "logps/rejected": -1166.7919921875,
      "loss": 0.0017,
      "rewards/accuracies": 0.9937499761581421,
      "rewards/chosen": -1.0605894327163696,
      "rewards/margins": 9.586730003356934,
      "rewards/rejected": -10.647318840026855,
      "step": 170
    },
    {
      "epoch": 1.44,
      "grad_norm": 0.004351477647807626,
      "learning_rate": 8.968184183545285e-06,
      "logits/chosen": -1.3524295091629028,
      "logits/rejected": -2.148332357406616,
      "logps/chosen": -146.11656188964844,
      "logps/rejected": -1293.1431884765625,
      "loss": 0.0002,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.959304928779602,
      "rewards/margins": 10.965489387512207,
      "rewards/rejected": -11.92479419708252,
      "step": 180
    },
    {
      "epoch": 1.52,
      "grad_norm": 0.004066031914054124,
      "learning_rate": 8.792028193824364e-06,
      "logits/chosen": -0.9521867036819458,
      "logits/rejected": -1.7849302291870117,
      "logps/chosen": -146.82675170898438,
      "logps/rejected": -1224.1842041015625,
      "loss": 0.0011,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.9709192514419556,
      "rewards/margins": 10.264577865600586,
      "rewards/rejected": -11.235495567321777,
      "step": 190
    },
    {
      "epoch": 1.6,
      "grad_norm": 2.8254638303864446,
      "learning_rate": 8.604025826343167e-06,
      "logits/chosen": -1.4731212854385376,
      "logits/rejected": -2.2158398628234863,
      "logps/chosen": -166.7774200439453,
      "logps/rejected": -1286.593505859375,
      "loss": 0.0007,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -1.1668568849563599,
      "rewards/margins": 10.681219100952148,
      "rewards/rejected": -11.848076820373535,
      "step": 200
    },
    {
      "epoch": 1.6800000000000002,
      "grad_norm": 0.027929253424062123,
      "learning_rate": 8.404764404547404e-06,
      "logits/chosen": -1.5285543203353882,
      "logits/rejected": -2.207693099975586,
      "logps/chosen": -145.4683837890625,
      "logps/rejected": -1240.819091796875,
      "loss": 0.0003,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.9586501121520996,
      "rewards/margins": 10.435168266296387,
      "rewards/rejected": -11.393818855285645,
      "step": 210
    },
    {
      "epoch": 1.76,
      "grad_norm": 0.00891305422535769,
      "learning_rate": 8.194866425410984e-06,
      "logits/chosen": -1.2906019687652588,
      "logits/rejected": -1.9918200969696045,
      "logps/chosen": -109.07038879394531,
      "logps/rejected": -1155.131591796875,
      "loss": 0.0002,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6084994077682495,
      "rewards/margins": 9.912607192993164,
      "rewards/rejected": -10.521106719970703,
      "step": 220
    },
    {
      "epoch": 1.8399999999999999,
      "grad_norm": 0.0018316293380985671,
      "learning_rate": 7.974987614742066e-06,
      "logits/chosen": -0.9763646125793457,
      "logits/rejected": -1.7541002035140991,
      "logps/chosen": -96.57362365722656,
      "logps/rejected": -1161.369384765625,
      "loss": 0.0004,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.4819282591342926,
      "rewards/margins": 10.10207462310791,
      "rewards/rejected": -10.584001541137695,
      "step": 230
    },
    {
      "epoch": 1.92,
      "grad_norm": 0.003080239253269156,
      "learning_rate": 7.745814878681516e-06,
      "logits/chosen": -1.5342729091644287,
      "logits/rejected": -2.221407175064087,
      "logps/chosen": -136.1558074951172,
      "logps/rejected": -1264.690673828125,
      "loss": 0.0047,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.8548258543014526,
      "rewards/margins": 10.780150413513184,
      "rewards/rejected": -11.634977340698242,
      "step": 240
    },
    {
      "epoch": 2.0,
      "grad_norm": 0.029482148279890704,
      "learning_rate": 7.50806415779332e-06,
      "logits/chosen": -2.258669853210449,
      "logits/rejected": -2.8217015266418457,
      "logps/chosen": -140.51853942871094,
      "logps/rejected": -1197.581787109375,
      "loss": 0.0003,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.8967872858047485,
      "rewards/margins": 10.075855255126953,
      "rewards/rejected": -10.97264289855957,
      "step": 250
    },
    {
      "epoch": 2.08,
      "grad_norm": 0.020782306025551353,
      "learning_rate": 7.262478190450834e-06,
      "logits/chosen": -2.3182389736175537,
      "logits/rejected": -2.92055082321167,
      "logps/chosen": -120.69600677490234,
      "logps/rejected": -1217.408447265625,
      "loss": 0.0003,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.7067109942436218,
      "rewards/margins": 10.442468643188477,
      "rewards/rejected": -11.14918041229248,
      "step": 260
    },
    {
      "epoch": 2.16,
      "grad_norm": 0.014378242296594932,
      "learning_rate": 7.0098241925061215e-06,
      "logits/chosen": -2.2274346351623535,
      "logits/rejected": -2.8495638370513916,
      "logps/chosen": -143.10069274902344,
      "logps/rejected": -1256.431640625,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.919533371925354,
      "rewards/margins": 10.646241188049316,
      "rewards/rejected": -11.565774917602539,
      "step": 270
    },
    {
      "epoch": 2.24,
      "grad_norm": 0.0011703826093326713,
      "learning_rate": 6.750891460491093e-06,
      "logits/chosen": -1.953466773033142,
      "logits/rejected": -2.550438642501831,
      "logps/chosen": -119.64122009277344,
      "logps/rejected": -1253.8546142578125,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6891164779663086,
      "rewards/margins": 10.844141960144043,
      "rewards/rejected": -11.533258438110352,
      "step": 280
    },
    {
      "epoch": 2.32,
      "grad_norm": 0.016444776166175196,
      "learning_rate": 6.486488905838143e-06,
      "logits/chosen": -1.8186432123184204,
      "logits/rejected": -2.431183338165283,
      "logps/chosen": -114.60746002197266,
      "logps/rejected": -1220.325927734375,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6494568586349487,
      "rewards/margins": 10.547185897827148,
      "rewards/rejected": -11.196642875671387,
      "step": 290
    },
    {
      "epoch": 2.4,
      "grad_norm": 0.0005496505009132721,
      "learning_rate": 6.2174425278234115e-06,
      "logits/chosen": -1.7246840000152588,
      "logits/rejected": -2.2873055934906006,
      "logps/chosen": -108.48016357421875,
      "logps/rejected": -1233.37744140625,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.5782238245010376,
      "rewards/margins": 10.742240905761719,
      "rewards/rejected": -11.320464134216309,
      "step": 300
    },
    {
      "epoch": 2.48,
      "grad_norm": 0.0006580056763243573,
      "learning_rate": 5.944592833127253e-06,
      "logits/chosen": -1.8141095638275146,
      "logits/rejected": -2.429810047149658,
      "logps/chosen": -119.1697006225586,
      "logps/rejected": -1326.865966796875,
      "loss": 0.0002,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6833556890487671,
      "rewards/margins": 11.57683277130127,
      "rewards/rejected": -12.260188102722168,
      "step": 310
    },
    {
      "epoch": 2.56,
      "grad_norm": 0.0011989998803146951,
      "learning_rate": 5.668792210073255e-06,
      "logits/chosen": -1.7204673290252686,
      "logits/rejected": -2.3327724933624268,
      "logps/chosen": -103.1290512084961,
      "logps/rejected": -1265.806640625,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.5410266518592834,
      "rewards/margins": 11.101877212524414,
      "rewards/rejected": -11.642904281616211,
      "step": 320
    },
    {
      "epoch": 2.64,
      "grad_norm": 0.01089396369261857,
      "learning_rate": 5.39090226574877e-06,
      "logits/chosen": -1.52898371219635,
      "logits/rejected": -2.140092134475708,
      "logps/chosen": -91.94349670410156,
      "logps/rejected": -1207.033447265625,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.43181926012039185,
      "rewards/margins": 10.611051559448242,
      "rewards/rejected": -11.042871475219727,
      "step": 330
    },
    {
      "epoch": 2.7199999999999998,
      "grad_norm": 0.0015016734818494304,
      "learning_rate": 5.111791134325793e-06,
      "logits/chosen": -1.6063101291656494,
      "logits/rejected": -2.3003313541412354,
      "logps/chosen": -126.36214447021484,
      "logps/rejected": -1331.2056884765625,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.7603046298027039,
      "rewards/margins": 11.541264533996582,
      "rewards/rejected": -12.301568984985352,
      "step": 340
    },
    {
      "epoch": 2.8,
      "grad_norm": 0.044167012422386985,
      "learning_rate": 4.832330764991131e-06,
      "logits/chosen": -1.6678760051727295,
      "logits/rejected": -2.332516670227051,
      "logps/chosen": -119.7556381225586,
      "logps/rejected": -1301.3248291015625,
      "loss": 0.0004,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6866098642349243,
      "rewards/margins": 11.313838958740234,
      "rewards/rejected": -12.000448226928711,
      "step": 350
    },
    {
      "epoch": 2.88,
      "grad_norm": 0.0005385533629396381,
      "learning_rate": 4.553394197958339e-06,
      "logits/chosen": -1.9416593313217163,
      "logits/rejected": -2.6904163360595703,
      "logps/chosen": -121.2122802734375,
      "logps/rejected": -1295.510498046875,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.7095136642456055,
      "rewards/margins": 11.224905014038086,
      "rewards/rejected": -11.934419631958008,
      "step": 360
    },
    {
      "epoch": 2.96,
      "grad_norm": 0.0006203949257092619,
      "learning_rate": 4.275852837071309e-06,
      "logits/chosen": -1.8644075393676758,
      "logits/rejected": -2.7058677673339844,
      "logps/chosen": -114.23481750488281,
      "logps/rejected": -1273.7774658203125,
      "loss": 0.0014,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6469079256057739,
      "rewards/margins": 11.084717750549316,
      "rewards/rejected": -11.731626510620117,
      "step": 370
    },
    {
      "epoch": 3.04,
      "grad_norm": 0.0018701198703410247,
      "learning_rate": 4.000573727519868e-06,
      "logits/chosen": -1.6911453008651733,
      "logits/rejected": -2.5057501792907715,
      "logps/chosen": -115.5822982788086,
      "logps/rejected": -1267.122314453125,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6596534848213196,
      "rewards/margins": 11.005937576293945,
      "rewards/rejected": -11.665590286254883,
      "step": 380
    },
    {
      "epoch": 3.12,
      "grad_norm": 0.004226077469779532,
      "learning_rate": 3.7284168471719527e-06,
      "logits/chosen": -1.646296501159668,
      "logits/rejected": -2.4523534774780273,
      "logps/chosen": -110.4627914428711,
      "logps/rejected": -1280.544921875,
      "loss": 0.0003,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6036595106124878,
      "rewards/margins": 11.18139934539795,
      "rewards/rejected": -11.785058975219727,
      "step": 390
    },
    {
      "epoch": 3.2,
      "grad_norm": 0.0006904055330191614,
      "learning_rate": 3.4602324199842026e-06,
      "logits/chosen": -1.7251287698745728,
      "logits/rejected": -2.5812900066375732,
      "logps/chosen": -138.866943359375,
      "logps/rejected": -1301.4326171875,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.8702784776687622,
      "rewards/margins": 11.145977973937988,
      "rewards/rejected": -12.016257286071777,
      "step": 400
    },
    {
      "epoch": 3.2800000000000002,
      "grad_norm": 0.0026124260205581465,
      "learning_rate": 3.1968582598840234e-06,
      "logits/chosen": -1.5730335712432861,
      "logits/rejected": -2.383654832839966,
      "logps/chosen": -109.4728012084961,
      "logps/rejected": -1262.404541015625,
      "loss": 0.0,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6025822758674622,
      "rewards/margins": 11.004800796508789,
      "rewards/rejected": -11.607383728027344,
      "step": 410
    },
    {
      "epoch": 3.36,
      "grad_norm": 0.016129813594014157,
      "learning_rate": 2.9391171534208185e-06,
      "logits/chosen": -1.7119560241699219,
      "logits/rejected": -2.51015043258667,
      "logps/chosen": -122.90666198730469,
      "logps/rejected": -1313.09228515625,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.7348455190658569,
      "rewards/margins": 11.394600868225098,
      "rewards/rejected": -12.129446983337402,
      "step": 420
    },
    {
      "epoch": 3.44,
      "grad_norm": 0.0024401290515470016,
      "learning_rate": 2.6878142893630904e-06,
      "logits/chosen": -1.6607134342193604,
      "logits/rejected": -2.4547019004821777,
      "logps/chosen": -137.77719116210938,
      "logps/rejected": -1349.2073974609375,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.8648022413253784,
      "rewards/margins": 11.622838020324707,
      "rewards/rejected": -12.487640380859375,
      "step": 430
    },
    {
      "epoch": 3.52,
      "grad_norm": 0.001375330125575793,
      "learning_rate": 2.4437347432713838e-06,
      "logits/chosen": -1.8045127391815186,
      "logits/rejected": -2.586901903152466,
      "logps/chosen": -156.62245178222656,
      "logps/rejected": -1351.8687744140625,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -1.0343620777130127,
      "rewards/margins": 11.499258995056152,
      "rewards/rejected": -12.533620834350586,
      "step": 440
    },
    {
      "epoch": 3.6,
      "grad_norm": 0.0016289824782741704,
      "learning_rate": 2.207641024905322e-06,
      "logits/chosen": -1.6454054117202759,
      "logits/rejected": -2.4403579235076904,
      "logps/chosen": -108.79969787597656,
      "logps/rejected": -1255.9583740234375,
      "loss": 0.0,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.5948039293289185,
      "rewards/margins": 10.957615852355957,
      "rewards/rejected": -11.552419662475586,
      "step": 450
    },
    {
      "epoch": 3.68,
      "grad_norm": 0.0003909667399708806,
      "learning_rate": 1.9802706961266936e-06,
      "logits/chosen": -1.5957154035568237,
      "logits/rejected": -2.4980740547180176,
      "logps/chosen": -124.23823547363281,
      "logps/rejected": -1321.4207763671875,
      "loss": 0.0002,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.7479596138000488,
      "rewards/margins": 11.454780578613281,
      "rewards/rejected": -12.202739715576172,
      "step": 460
    },
    {
      "epoch": 3.76,
      "grad_norm": 0.0006870240069107201,
      "learning_rate": 1.7623340667403089e-06,
      "logits/chosen": -1.5271575450897217,
      "logits/rejected": -2.3313746452331543,
      "logps/chosen": -115.49822998046875,
      "logps/rejected": -1287.483642578125,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6483306288719177,
      "rewards/margins": 11.213019371032715,
      "rewards/rejected": -11.861350059509277,
      "step": 470
    },
    {
      "epoch": 3.84,
      "grad_norm": 0.03279067348481457,
      "learning_rate": 1.5545119754708682e-06,
      "logits/chosen": -1.719366431236267,
      "logits/rejected": -2.5721096992492676,
      "logps/chosen": -141.73440551757812,
      "logps/rejected": -1358.951416015625,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.9100968241691589,
      "rewards/margins": 11.68384838104248,
      "rewards/rejected": -12.59394645690918,
      "step": 480
    },
    {
      "epoch": 3.92,
      "grad_norm": 0.004304506433089471,
      "learning_rate": 1.3574536630081208e-06,
      "logits/chosen": -1.5175102949142456,
      "logits/rejected": -2.292325496673584,
      "logps/chosen": -100.6703872680664,
      "logps/rejected": -1273.519775390625,
      "loss": 0.0,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.5087094902992249,
      "rewards/margins": 11.203500747680664,
      "rewards/rejected": -11.71220874786377,
      "step": 490
    },
    {
      "epoch": 4.0,
      "grad_norm": 0.0004662700999661582,
      "learning_rate": 1.1717747437649657e-06,
      "logits/chosen": -1.5407252311706543,
      "logits/rejected": -2.3491063117980957,
      "logps/chosen": -122.54182434082031,
      "logps/rejected": -1335.7579345703125,
      "loss": 0.0,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.7224875092506409,
      "rewards/margins": 11.629507064819336,
      "rewards/rejected": -12.351993560791016,
      "step": 500
    },
    {
      "epoch": 4.08,
      "grad_norm": 0.0017439323111767312,
      "learning_rate": 9.980552826847635e-07,
      "logits/chosen": -1.6905838251113892,
      "logits/rejected": -2.487473487854004,
      "logps/chosen": -166.93624877929688,
      "logps/rejected": -1409.1981201171875,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -1.1514984369277954,
      "rewards/margins": 11.959856033325195,
      "rewards/rejected": -13.111355781555176,
      "step": 510
    },
    {
      "epoch": 4.16,
      "grad_norm": 0.0024902235420863005,
      "learning_rate": 8.368379831059592e-07,
      "logits/chosen": -1.5832383632659912,
      "logits/rejected": -2.3622446060180664,
      "logps/chosen": -130.75253295898438,
      "logps/rejected": -1372.863037109375,
      "loss": 0.0,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.806990921497345,
      "rewards/margins": 11.918740272521973,
      "rewards/rejected": -12.725730895996094,
      "step": 520
    },
    {
      "epoch": 4.24,
      "grad_norm": 0.0004030809278175782,
      "learning_rate": 6.886264913451635e-07,
      "logits/chosen": -1.5754905939102173,
      "logits/rejected": -2.314927101135254,
      "logps/chosen": -109.8852310180664,
      "logps/rejected": -1284.387451171875,
      "loss": 0.0,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6023492813110352,
      "rewards/margins": 11.232891082763672,
      "rewards/rejected": -11.835241317749023,
      "step": 530
    },
    {
      "epoch": 4.32,
      "grad_norm": 0.014990693875413815,
      "learning_rate": 5.538838232952104e-07,
      "logits/chosen": -1.6247650384902954,
      "logits/rejected": -2.395631790161133,
      "logps/chosen": -112.5528335571289,
      "logps/rejected": -1308.465087890625,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6267929673194885,
      "rewards/margins": 11.434622764587402,
      "rewards/rejected": -12.06141471862793,
      "step": 540
    },
    {
      "epoch": 4.4,
      "grad_norm": 0.0008931683311901964,
      "learning_rate": 4.3303091795353024e-07,
      "logits/chosen": -1.7327635288238525,
      "logits/rejected": -2.521054744720459,
      "logps/chosen": -193.12649536132812,
      "logps/rejected": -1453.8939208984375,
      "loss": 0.0,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -1.3928953409194946,
      "rewards/margins": 12.179082870483398,
      "rewards/rejected": -13.571980476379395,
      "step": 550
    },
    {
      "epoch": 4.48,
      "grad_norm": 0.0006090223103858115,
      "learning_rate": 3.2644532239966444e-07,
      "logits/chosen": -1.540307879447937,
      "logits/rejected": -2.274130344390869,
      "logps/chosen": -116.36529541015625,
      "logps/rejected": -1331.5732421875,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6516658663749695,
      "rewards/margins": 11.651314735412598,
      "rewards/rejected": -12.302980422973633,
      "step": 560
    },
    {
      "epoch": 4.5600000000000005,
      "grad_norm": 0.0012600919102292293,
      "learning_rate": 2.3446001233004333e-07,
      "logits/chosen": -1.545082926750183,
      "logits/rejected": -2.3070833683013916,
      "logps/chosen": -118.43846130371094,
      "logps/rejected": -1310.844482421875,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6805299520492554,
      "rewards/margins": 11.414255142211914,
      "rewards/rejected": -12.094785690307617,
      "step": 570
    },
    {
      "epoch": 4.64,
      "grad_norm": 0.0033895825817679913,
      "learning_rate": 1.573623518347517e-07,
      "logits/chosen": -1.3955062627792358,
      "logits/rejected": -2.1139488220214844,
      "logps/chosen": -95.18065643310547,
      "logps/rejected": -1255.7984619140625,
      "loss": 0.0,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.46435871720314026,
      "rewards/margins": 11.06248664855957,
      "rewards/rejected": -11.52684497833252,
      "step": 580
    },
    {
      "epoch": 4.72,
      "grad_norm": 0.0005277953893460651,
      "learning_rate": 9.539319566590766e-08,
      "logits/chosen": -1.4713037014007568,
      "logits/rejected": -2.1673684120178223,
      "logps/chosen": -103.91761779785156,
      "logps/rejected": -1269.9681396484375,
      "loss": 0.0,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.5508452653884888,
      "rewards/margins": 11.134601593017578,
      "rewards/rejected": -11.685447692871094,
      "step": 590
    },
    {
      "epoch": 4.8,
      "grad_norm": 0.0038828625096458437,
      "learning_rate": 4.8746136802240716e-08,
      "logits/chosen": -1.60639226436615,
      "logits/rejected": -2.2908573150634766,
      "logps/chosen": -111.3029556274414,
      "logps/rejected": -1284.128173828125,
      "loss": 0.0001,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6181883811950684,
      "rewards/margins": 11.210031509399414,
      "rewards/rejected": -11.82822036743164,
      "step": 600
    },
    {
      "epoch": 4.88,
      "grad_norm": 0.0016152355239910333,
      "learning_rate": 1.75669016604485e-08,
      "logits/chosen": -1.5895562171936035,
      "logits/rejected": -2.336843490600586,
      "logps/chosen": -118.6875991821289,
      "logps/rejected": -1293.158935546875,
      "loss": 0.0,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.6965440511703491,
      "rewards/margins": 11.21881103515625,
      "rewards/rejected": -11.915353775024414,
      "step": 610
    },
    {
      "epoch": 4.96,
      "grad_norm": 0.0004233911082118195,
      "learning_rate": 1.952894842735531e-09,
      "logits/chosen": -1.5527576208114624,
      "logits/rejected": -2.2691588401794434,
      "logps/chosen": -98.73721313476562,
      "logps/rejected": -1254.2822265625,
      "loss": 0.0,
      "rewards/accuracies": 1.0,
      "rewards/chosen": -0.4978357255458832,
      "rewards/margins": 11.023382186889648,
      "rewards/rejected": -11.52121639251709,
      "step": 620
    },
    {
      "epoch": 5.0,
      "step": 625,
      "total_flos": 0.0,
      "train_loss": 0.022511796173290348,
      "train_runtime": 65461.2558,
      "train_samples_per_second": 0.611,
      "train_steps_per_second": 0.01
    }
  ],
  "logging_steps": 10,
  "max_steps": 625,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 5,
  "save_steps": 1000,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 0.0,
  "train_batch_size": 4,
  "trial_name": null,
  "trial_params": null
}