Good Robot π€
β There is an updated version of this model available, please see Good Robot 2 β.
The model "Good Robot" had one simple goal in mind: to be a good instruction-following model that doesn't talk like ChatGPT.
Built upon the Mistral 7b base, this model aims to provide responses that are as human-like as possible, thanks to some DPO training using the (for now, private) minerva-ai/yes-robots-dpo
dataset.
HuggingFaceH4/no-robots was used as the base for generating a custom dataset to create DPO pairs.
It should follow instructions and be generally as smart as a typical Mistral model - just not as soulless and full of GPT slop.
Prompt Format:
Alpaca, my beloved β€οΈ
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{your prompt goes here}
### Response:
Huge Thanks:
- Gryphe for DPO scripts and all the patience π
Training Data:
- HuggingFaceH4/no_robots
- MinervaAI/yes-robots-dpo
- private datasets with common GPTisms
Limitations:
While I did my best to minimize GPTisms, no model is perfect, and there may still be instances where the generated content has GPT's common phrases - I have a suspicion that's due to them being engrained into Mistral model itself.
License:
cc-by-nc-4.0
- Downloads last month
- 18