Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,9 @@
|
|
1 |
*This model was trained as part of a series of experiments testing the performance of pure DPO vs SFT vs ORPO, all supported by Unsloth/Huggingface TRL.*
|
2 |
|
|
|
|
|
|
|
|
|
3 |
**Training Details**
|
4 |
|
5 |
Dataset: https://huggingface.co/datasets/argilla/dpo-mix-7k
|
|
|
1 |
*This model was trained as part of a series of experiments testing the performance of pure DPO vs SFT vs ORPO, all supported by Unsloth/Huggingface TRL.*
|
2 |
|
3 |
+
**Benchmarks**
|
4 |
+
|
5 |
+
TBA
|
6 |
+
|
7 |
**Training Details**
|
8 |
|
9 |
Dataset: https://huggingface.co/datasets/argilla/dpo-mix-7k
|