--- base_model: nbeerbower/flammen17-py-DPO-v1-7B datasets: - jondurbin/py-dpo-v0.1 inference: false library_name: transformers license: apache-2.0 merged_models: - nbeerbower/flammen17-mistral-7B pipeline_tag: text-generation quantized_by: Suparious tags: - 4-bit - AWQ - text-generation - autotrain_compatible - endpoints_compatible - experimental --- # nbeerbower/flammen17-py-DPO-v1-7B AWQ - Model creator: [nbeerbower](https://huggingface.co/nbeerbower) - Original model: [flammen17-py-DPO-v1-7B](https://huggingface.co/nbeerbower/flammen17-py-DPO-v1-7B) ![image/png](https://huggingface.co/nbeerbower/flammen13X-mistral-7B/resolve/main/flammen13x.png) ## Model Summary A Mistral 7B LLM built from merging pretrained models and finetuning on [Jon Durbin](https://huggingface.co/jondurbin)'s [py-dpo-v0.1](https://huggingface.co/datasets/jondurbin/py-dpo-v0.1). Finetuned using an A100 on Google Colab. 🙏 [Fine-tune a Mistral-7b model with Direct Preference Optimization](https://towardsdatascience.com/fine-tune-a-mistral-7b-model-with-direct-preference-optimization-708042745aac) - [Maxime Labonne](https://huggingface.co/mlabonne)