Update README.md
Browse files
README.md
CHANGED
@@ -10,7 +10,7 @@ inference: false
|
|
10 |
---
|
11 |
# Model Card for Mixtral-8x7B DPO
|
12 |
|
13 |
-
* [
|
14 |
|
15 |
* [DPO Trainer](https://huggingface.co/docs/trl/main/en/dpo_trainer)
|
16 |
|
|
|
10 |
---
|
11 |
# Model Card for Mixtral-8x7B DPO
|
12 |
|
13 |
+
* [DPO training try to improve mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1)
|
14 |
|
15 |
* [DPO Trainer](https://huggingface.co/docs/trl/main/en/dpo_trainer)
|
16 |
|