--- inference: false language: - en - es library_name: transformers license: apache-2.0 model_type: mixtral pipeline_tag: text-generation tags: - mistral - mixtral --- Barcenas Mixtral 8x7b based on argilla/notux-8x7b-v1 It is a 4-bit version of this model to make it more accessible to users Trained with DPO and using MoE Technology makes it a powerful and innovative model. Made with ❤️ in Guadalupe, Nuevo Leon, Mexico 🇲🇽