Model save
Browse files- README.md +1 -5
- all_results.json +6 -6
- eval_results.json +3 -3
- runs/Jan09_22-10-39_ip-26-0-161-123/events.out.tfevents.1704838276.ip-26-0-161-123.15622.0 +2 -2
- runs/Jan09_22-10-39_ip-26-0-161-123/events.out.tfevents.1704843559.ip-26-0-161-123.15622.1 +3 -0
- train_results.json +3 -3
- trainer_state.json +15 -15
README.md
CHANGED
@@ -2,13 +2,9 @@
|
|
2 |
license: apache-2.0
|
3 |
base_model: alignment-handbook/zephyr-7b-sft-full
|
4 |
tags:
|
5 |
-
- alignment-handbook
|
6 |
-
- generated_from_trainer
|
7 |
- trl
|
8 |
- dpo
|
9 |
- generated_from_trainer
|
10 |
-
datasets:
|
11 |
-
- HuggingFaceH4/ultrafeedback_binarized
|
12 |
model-index:
|
13 |
- name: zephyr-7b-dpo-full
|
14 |
results: []
|
@@ -19,7 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
|
|
19 |
|
20 |
# zephyr-7b-dpo-full
|
21 |
|
22 |
-
This model is a fine-tuned version of [alignment-handbook/zephyr-7b-sft-full](https://huggingface.co/alignment-handbook/zephyr-7b-sft-full) on the
|
23 |
It achieves the following results on the evaluation set:
|
24 |
- Loss: 0.5028
|
25 |
- Rewards/chosen: -0.9469
|
|
|
2 |
license: apache-2.0
|
3 |
base_model: alignment-handbook/zephyr-7b-sft-full
|
4 |
tags:
|
|
|
|
|
5 |
- trl
|
6 |
- dpo
|
7 |
- generated_from_trainer
|
|
|
|
|
8 |
model-index:
|
9 |
- name: zephyr-7b-dpo-full
|
10 |
results: []
|
|
|
15 |
|
16 |
# zephyr-7b-dpo-full
|
17 |
|
18 |
+
This model is a fine-tuned version of [alignment-handbook/zephyr-7b-sft-full](https://huggingface.co/alignment-handbook/zephyr-7b-sft-full) on the None dataset.
|
19 |
It achieves the following results on the evaluation set:
|
20 |
- Loss: 0.5028
|
21 |
- Rewards/chosen: -0.9469
|
all_results.json
CHANGED
@@ -9,13 +9,13 @@
|
|
9 |
"eval_rewards/chosen": -0.9468507170677185,
|
10 |
"eval_rewards/margins": 0.946345865726471,
|
11 |
"eval_rewards/rejected": -1.8931965827941895,
|
12 |
-
"eval_runtime":
|
13 |
"eval_samples": 2000,
|
14 |
-
"eval_samples_per_second":
|
15 |
-
"eval_steps_per_second": 0.
|
16 |
"train_loss": 0.5366686437918052,
|
17 |
-
"train_runtime":
|
18 |
"train_samples": 61135,
|
19 |
-
"train_samples_per_second": 11.
|
20 |
-
"train_steps_per_second": 0.
|
21 |
}
|
|
|
9 |
"eval_rewards/chosen": -0.9468507170677185,
|
10 |
"eval_rewards/margins": 0.946345865726471,
|
11 |
"eval_rewards/rejected": -1.8931965827941895,
|
12 |
+
"eval_runtime": 86.4354,
|
13 |
"eval_samples": 2000,
|
14 |
+
"eval_samples_per_second": 23.139,
|
15 |
+
"eval_steps_per_second": 0.37,
|
16 |
"train_loss": 0.5366686437918052,
|
17 |
+
"train_runtime": 5196.7487,
|
18 |
"train_samples": 61135,
|
19 |
+
"train_samples_per_second": 11.764,
|
20 |
+
"train_steps_per_second": 0.092
|
21 |
}
|
eval_results.json
CHANGED
@@ -9,8 +9,8 @@
|
|
9 |
"eval_rewards/chosen": -0.9468507170677185,
|
10 |
"eval_rewards/margins": 0.946345865726471,
|
11 |
"eval_rewards/rejected": -1.8931965827941895,
|
12 |
-
"eval_runtime":
|
13 |
"eval_samples": 2000,
|
14 |
-
"eval_samples_per_second":
|
15 |
-
"eval_steps_per_second": 0.
|
16 |
}
|
|
|
9 |
"eval_rewards/chosen": -0.9468507170677185,
|
10 |
"eval_rewards/margins": 0.946345865726471,
|
11 |
"eval_rewards/rejected": -1.8931965827941895,
|
12 |
+
"eval_runtime": 86.4354,
|
13 |
"eval_samples": 2000,
|
14 |
+
"eval_samples_per_second": 23.139,
|
15 |
+
"eval_steps_per_second": 0.37
|
16 |
}
|
runs/Jan09_22-10-39_ip-26-0-161-123/events.out.tfevents.1704838276.ip-26-0-161-123.15622.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:aa992e831700f770bddff3a401b7041798cbf0c8d045f8b2c3a00e6466847ef0
|
3 |
+
size 38122
|
runs/Jan09_22-10-39_ip-26-0-161-123/events.out.tfevents.1704843559.ip-26-0-161-123.15622.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:52db6ce34acd137452951519c58c3a279488a7192f40fc17a7a1f0feee27a8ec
|
3 |
+
size 828
|
train_results.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"epoch": 1.0,
|
3 |
"train_loss": 0.5366686437918052,
|
4 |
-
"train_runtime":
|
5 |
"train_samples": 61135,
|
6 |
-
"train_samples_per_second": 11.
|
7 |
-
"train_steps_per_second": 0.
|
8 |
}
|
|
|
1 |
{
|
2 |
"epoch": 1.0,
|
3 |
"train_loss": 0.5366686437918052,
|
4 |
+
"train_runtime": 5196.7487,
|
5 |
"train_samples": 61135,
|
6 |
+
"train_samples_per_second": 11.764,
|
7 |
+
"train_steps_per_second": 0.092
|
8 |
}
|
trainer_state.json
CHANGED
@@ -173,9 +173,9 @@
|
|
173 |
"eval_rewards/chosen": -0.49532508850097656,
|
174 |
"eval_rewards/margins": 0.6263692378997803,
|
175 |
"eval_rewards/rejected": -1.1216944456100464,
|
176 |
-
"eval_runtime":
|
177 |
-
"eval_samples_per_second":
|
178 |
-
"eval_steps_per_second": 0.
|
179 |
"step": 100
|
180 |
},
|
181 |
{
|
@@ -329,9 +329,9 @@
|
|
329 |
"eval_rewards/chosen": -0.899510383605957,
|
330 |
"eval_rewards/margins": 0.8722902536392212,
|
331 |
"eval_rewards/rejected": -1.7718006372451782,
|
332 |
-
"eval_runtime":
|
333 |
-
"eval_samples_per_second":
|
334 |
-
"eval_steps_per_second": 0.
|
335 |
"step": 200
|
336 |
},
|
337 |
{
|
@@ -485,9 +485,9 @@
|
|
485 |
"eval_rewards/chosen": -0.7945692539215088,
|
486 |
"eval_rewards/margins": 0.8339425921440125,
|
487 |
"eval_rewards/rejected": -1.628511905670166,
|
488 |
-
"eval_runtime":
|
489 |
-
"eval_samples_per_second":
|
490 |
-
"eval_steps_per_second": 0.
|
491 |
"step": 300
|
492 |
},
|
493 |
{
|
@@ -641,9 +641,9 @@
|
|
641 |
"eval_rewards/chosen": -0.9776647090911865,
|
642 |
"eval_rewards/margins": 0.9803228974342346,
|
643 |
"eval_rewards/rejected": -1.9579875469207764,
|
644 |
-
"eval_runtime":
|
645 |
-
"eval_samples_per_second":
|
646 |
-
"eval_steps_per_second": 0.
|
647 |
"step": 400
|
648 |
},
|
649 |
{
|
@@ -749,9 +749,9 @@
|
|
749 |
"step": 478,
|
750 |
"total_flos": 0.0,
|
751 |
"train_loss": 0.5366686437918052,
|
752 |
-
"train_runtime":
|
753 |
-
"train_samples_per_second": 11.
|
754 |
-
"train_steps_per_second": 0.
|
755 |
}
|
756 |
],
|
757 |
"logging_steps": 10,
|
|
|
173 |
"eval_rewards/chosen": -0.49532508850097656,
|
174 |
"eval_rewards/margins": 0.6263692378997803,
|
175 |
"eval_rewards/rejected": -1.1216944456100464,
|
176 |
+
"eval_runtime": 85.4149,
|
177 |
+
"eval_samples_per_second": 23.415,
|
178 |
+
"eval_steps_per_second": 0.375,
|
179 |
"step": 100
|
180 |
},
|
181 |
{
|
|
|
329 |
"eval_rewards/chosen": -0.899510383605957,
|
330 |
"eval_rewards/margins": 0.8722902536392212,
|
331 |
"eval_rewards/rejected": -1.7718006372451782,
|
332 |
+
"eval_runtime": 85.7368,
|
333 |
+
"eval_samples_per_second": 23.327,
|
334 |
+
"eval_steps_per_second": 0.373,
|
335 |
"step": 200
|
336 |
},
|
337 |
{
|
|
|
485 |
"eval_rewards/chosen": -0.7945692539215088,
|
486 |
"eval_rewards/margins": 0.8339425921440125,
|
487 |
"eval_rewards/rejected": -1.628511905670166,
|
488 |
+
"eval_runtime": 86.1066,
|
489 |
+
"eval_samples_per_second": 23.227,
|
490 |
+
"eval_steps_per_second": 0.372,
|
491 |
"step": 300
|
492 |
},
|
493 |
{
|
|
|
641 |
"eval_rewards/chosen": -0.9776647090911865,
|
642 |
"eval_rewards/margins": 0.9803228974342346,
|
643 |
"eval_rewards/rejected": -1.9579875469207764,
|
644 |
+
"eval_runtime": 86.2141,
|
645 |
+
"eval_samples_per_second": 23.198,
|
646 |
+
"eval_steps_per_second": 0.371,
|
647 |
"step": 400
|
648 |
},
|
649 |
{
|
|
|
749 |
"step": 478,
|
750 |
"total_flos": 0.0,
|
751 |
"train_loss": 0.5366686437918052,
|
752 |
+
"train_runtime": 5196.7487,
|
753 |
+
"train_samples_per_second": 11.764,
|
754 |
+
"train_steps_per_second": 0.092
|
755 |
}
|
756 |
],
|
757 |
"logging_steps": 10,
|