ShenaoZ commited on
Commit
ff21c5e
1 Parent(s): ac7244e

Model save

Browse files
README.md CHANGED
@@ -2,15 +2,9 @@
2
  license: mit
3
  base_model: HuggingFaceH4/mistral-7b-sft-beta
4
  tags:
5
- - alignment-handbook
6
  - trl
7
  - dpo
8
  - generated_from_trainer
9
- - trl
10
- - dpo
11
- - generated_from_trainer
12
- datasets:
13
- - HuggingFaceH4/ultrafeedback_binarized
14
  model-index:
15
  - name: DPO-Zephyr-7B
16
  results: []
@@ -21,7 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  # DPO-Zephyr-7B
23
 
24
- This model is a fine-tuned version of [HuggingFaceH4/mistral-7b-sft-beta](https://huggingface.co/HuggingFaceH4/mistral-7b-sft-beta) on the HuggingFaceH4/ultrafeedback_binarized dataset.
25
 
26
  ## Model description
27
 
 
2
  license: mit
3
  base_model: HuggingFaceH4/mistral-7b-sft-beta
4
  tags:
 
5
  - trl
6
  - dpo
7
  - generated_from_trainer
 
 
 
 
 
8
  model-index:
9
  - name: DPO-Zephyr-7B
10
  results: []
 
15
 
16
  # DPO-Zephyr-7B
17
 
18
+ This model is a fine-tuned version of [HuggingFaceH4/mistral-7b-sft-beta](https://huggingface.co/HuggingFaceH4/mistral-7b-sft-beta) on the None dataset.
19
 
20
  ## Model description
21
 
all_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 0.9874476987447699,
3
  "total_flos": 0.0,
4
- "train_loss": 0.6490178754774191,
5
- "train_runtime": 1630.8514,
6
  "train_samples": 15283,
7
- "train_samples_per_second": 9.371,
8
  "train_steps_per_second": 0.036
9
  }
 
1
  {
2
  "epoch": 0.9874476987447699,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.6490362135030455,
5
+ "train_runtime": 1628.642,
6
  "train_samples": 15283,
7
+ "train_samples_per_second": 9.384,
8
  "train_steps_per_second": 0.036
9
  }
config.json CHANGED
@@ -21,6 +21,6 @@
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
  "transformers_version": "4.40.2",
24
- "use_cache": true,
25
  "vocab_size": 32000
26
  }
 
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
  "transformers_version": "4.40.2",
24
+ "use_cache": false,
25
  "vocab_size": 32000
26
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c102c7c24a4ffc9699e03c2bef20903e0d393c638be088dd4c10475225b2f91c
3
  size 4943162336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d655c22edf89a46ecc9b8ab8f1dd1f136951064119abcf1b7e2f3f2a47138f9
3
  size 4943162336
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5eec78610bcb027f2d9a6a0dbf2b4b18aa303873090b7bbac8f6a4193cd50c5b
3
  size 4999819336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c2831fb26b09c22f51c2c3908253744b2fa7bd5252b25175e23dc337bb1dd70
3
  size 4999819336
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc676b5fe1e6987a2d21ed941fe4d6ec069fa2a0ee9df004a9da8eb1462a4229
3
  size 4540516344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e911e70c1914212070f5a2b59d5003a2daf2522f3a2346faffd311c91d5cbbb
3
  size 4540516344
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 0.9874476987447699,
3
  "total_flos": 0.0,
4
- "train_loss": 0.6490178754774191,
5
- "train_runtime": 1630.8514,
6
  "train_samples": 15283,
7
- "train_samples_per_second": 9.371,
8
  "train_steps_per_second": 0.036
9
  }
 
1
  {
2
  "epoch": 0.9874476987447699,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.6490362135030455,
5
+ "train_runtime": 1628.642,
6
  "train_samples": 15283,
7
+ "train_samples_per_second": 9.384,
8
  "train_steps_per_second": 0.036
9
  }
trainer_state.json CHANGED
@@ -10,7 +10,7 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.016736401673640166,
13
- "grad_norm": 7.772482099050406,
14
  "learning_rate": 8.333333333333333e-08,
15
  "logits/chosen": -2.807276487350464,
16
  "logits/rejected": -2.7759768962860107,
@@ -25,86 +25,86 @@
25
  },
26
  {
27
  "epoch": 0.16736401673640167,
28
- "grad_norm": 7.4630481490893485,
29
  "learning_rate": 4.930057285201027e-07,
30
- "logits/chosen": -2.7555835247039795,
31
- "logits/rejected": -2.746177911758423,
32
- "logps/chosen": -271.9296569824219,
33
- "logps/rejected": -260.67138671875,
34
- "loss": 0.6914,
35
- "rewards/accuracies": 0.5347222089767456,
36
- "rewards/chosen": 0.0054025910794734955,
37
- "rewards/margins": 0.0035073086619377136,
38
- "rewards/rejected": 0.0018952824175357819,
39
  "step": 10
40
  },
41
  {
42
  "epoch": 0.33472803347280333,
43
- "grad_norm": 6.708300371266294,
44
  "learning_rate": 4.187457503795526e-07,
45
- "logits/chosen": -2.787097930908203,
46
- "logits/rejected": -2.7682840824127197,
47
- "logps/chosen": -261.45843505859375,
48
- "logps/rejected": -250.8958740234375,
49
  "loss": 0.6745,
50
- "rewards/accuracies": 0.699999988079071,
51
- "rewards/chosen": 0.03223257511854172,
52
- "rewards/margins": 0.040766604244709015,
53
- "rewards/rejected": -0.008534022606909275,
54
  "step": 20
55
  },
56
  {
57
  "epoch": 0.502092050209205,
58
- "grad_norm": 7.409956341692702,
59
  "learning_rate": 2.8691164100062034e-07,
60
- "logits/chosen": -2.7880468368530273,
61
- "logits/rejected": -2.7689902782440186,
62
- "logps/chosen": -293.8247985839844,
63
- "logps/rejected": -254.217529296875,
64
- "loss": 0.6487,
65
- "rewards/accuracies": 0.6812499761581421,
66
- "rewards/chosen": -0.023687291890382767,
67
- "rewards/margins": 0.12690749764442444,
68
- "rewards/rejected": -0.1505947858095169,
69
  "step": 30
70
  },
71
  {
72
  "epoch": 0.6694560669456067,
73
- "grad_norm": 8.167332409118467,
74
  "learning_rate": 1.4248369943086995e-07,
75
- "logits/chosen": -2.760864734649658,
76
- "logits/rejected": -2.7383790016174316,
77
- "logps/chosen": -264.1250915527344,
78
- "logps/rejected": -255.5350799560547,
79
- "loss": 0.6354,
80
  "rewards/accuracies": 0.684374988079071,
81
- "rewards/chosen": -0.11205615103244781,
82
- "rewards/margins": 0.16018818318843842,
83
- "rewards/rejected": -0.27224433422088623,
84
  "step": 40
85
  },
86
  {
87
  "epoch": 0.8368200836820083,
88
- "grad_norm": 8.545155858110496,
89
  "learning_rate": 3.473909705816111e-08,
90
- "logits/chosen": -2.770779848098755,
91
- "logits/rejected": -2.750847339630127,
92
- "logps/chosen": -279.1131286621094,
93
- "logps/rejected": -287.27044677734375,
94
  "loss": 0.6202,
95
  "rewards/accuracies": 0.628125011920929,
96
- "rewards/chosen": -0.22104480862617493,
97
- "rewards/margins": 0.13653476536273956,
98
- "rewards/rejected": -0.3575795590877533,
99
  "step": 50
100
  },
101
  {
102
  "epoch": 0.9874476987447699,
103
  "step": 59,
104
  "total_flos": 0.0,
105
- "train_loss": 0.6490178754774191,
106
- "train_runtime": 1630.8514,
107
- "train_samples_per_second": 9.371,
108
  "train_steps_per_second": 0.036
109
  }
110
  ],
 
10
  "log_history": [
11
  {
12
  "epoch": 0.016736401673640166,
13
+ "grad_norm": 7.775175087490474,
14
  "learning_rate": 8.333333333333333e-08,
15
  "logits/chosen": -2.807276487350464,
16
  "logits/rejected": -2.7759768962860107,
 
25
  },
26
  {
27
  "epoch": 0.16736401673640167,
28
+ "grad_norm": 7.448045779799194,
29
  "learning_rate": 4.930057285201027e-07,
30
+ "logits/chosen": -2.7555572986602783,
31
+ "logits/rejected": -2.746171474456787,
32
+ "logps/chosen": -271.9293212890625,
33
+ "logps/rejected": -260.67010498046875,
34
+ "loss": 0.6915,
35
+ "rewards/accuracies": 0.53125,
36
+ "rewards/chosen": 0.005406347569078207,
37
+ "rewards/margins": 0.0034980433993041515,
38
+ "rewards/rejected": 0.0019083041697740555,
39
  "step": 10
40
  },
41
  {
42
  "epoch": 0.33472803347280333,
43
+ "grad_norm": 6.734236245718351,
44
  "learning_rate": 4.187457503795526e-07,
45
+ "logits/chosen": -2.7869458198547363,
46
+ "logits/rejected": -2.768073320388794,
47
+ "logps/chosen": -261.4452209472656,
48
+ "logps/rejected": -250.89541625976562,
49
  "loss": 0.6745,
50
+ "rewards/accuracies": 0.715624988079071,
51
+ "rewards/chosen": 0.03236500173807144,
52
+ "rewards/margins": 0.04089399054646492,
53
+ "rewards/rejected": -0.008528990671038628,
54
  "step": 20
55
  },
56
  {
57
  "epoch": 0.502092050209205,
58
+ "grad_norm": 7.437539615124712,
59
  "learning_rate": 2.8691164100062034e-07,
60
+ "logits/chosen": -2.7878143787384033,
61
+ "logits/rejected": -2.7688212394714355,
62
+ "logps/chosen": -293.8128662109375,
63
+ "logps/rejected": -254.24874877929688,
64
+ "loss": 0.6485,
65
+ "rewards/accuracies": 0.675000011920929,
66
+ "rewards/chosen": -0.02356739528477192,
67
+ "rewards/margins": 0.12733949720859528,
68
+ "rewards/rejected": -0.15090689063072205,
69
  "step": 30
70
  },
71
  {
72
  "epoch": 0.6694560669456067,
73
+ "grad_norm": 8.137783776888698,
74
  "learning_rate": 1.4248369943086995e-07,
75
+ "logits/chosen": -2.760474681854248,
76
+ "logits/rejected": -2.7379024028778076,
77
+ "logps/chosen": -264.1443176269531,
78
+ "logps/rejected": -255.5771026611328,
79
+ "loss": 0.6353,
80
  "rewards/accuracies": 0.684374988079071,
81
+ "rewards/chosen": -0.11224844306707382,
82
+ "rewards/margins": 0.16041621565818787,
83
+ "rewards/rejected": -0.2726646959781647,
84
  "step": 40
85
  },
86
  {
87
  "epoch": 0.8368200836820083,
88
+ "grad_norm": 8.560922020463133,
89
  "learning_rate": 3.473909705816111e-08,
90
+ "logits/chosen": -2.7703192234039307,
91
+ "logits/rejected": -2.7502920627593994,
92
+ "logps/chosen": -279.1363525390625,
93
+ "logps/rejected": -287.27764892578125,
94
  "loss": 0.6202,
95
  "rewards/accuracies": 0.628125011920929,
96
+ "rewards/chosen": -0.22127707302570343,
97
+ "rewards/margins": 0.13637453317642212,
98
+ "rewards/rejected": -0.35765162110328674,
99
  "step": 50
100
  },
101
  {
102
  "epoch": 0.9874476987447699,
103
  "step": 59,
104
  "total_flos": 0.0,
105
+ "train_loss": 0.6490362135030455,
106
+ "train_runtime": 1628.642,
107
+ "train_samples_per_second": 9.384,
108
  "train_steps_per_second": 0.036
109
  }
110
  ],
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cd4aa07115a1b0094354c437b369487eb676a555745cf463b02b3eaa8496dd8f
3
  size 6264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a96c5176af1b3e2b03b4b13bd4c023071e4d91fb081150a429e0a08207ff8287
3
  size 6264