--- license: apache-2.0 language: - fr - it - de - es - en tag: - moe - DPO --- # Model Card for cloudyu/Mixtral-8x7B-Instruct-v0.1-DPO * [DPO training try to improve mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) * [DPO Trainer](https://huggingface.co/docs/trl/main/en/dpo_trainer) * metrics not test