--- license: cc-by-nc-4.0 datasets: - Dahoas/instruct-synthetic-prompt-responses language: - en pipeline_tag: text-generation --- Question answering model finetuned from [GPT4All-J v1.3](https://huggingface.co/nomic-ai/gpt4all-j) with [Direct Preference Optimization](https://arxiv.org/abs/2305.18290). \ Dataset: [Dahoas/instruct-synthetic-prompt-responses](https://huggingface.co/datasets/Dahoas/instruct-synthetic-prompt-responses). The model was finetuned with the following promt: \ ``"Answer the following question in context:\n\nQuestion: " + samples["prompt"] + " Answer: "`` \ It should be benefical to use the same or a similar prompt for inference. An increase in performance compared to [GPT4All-J v1.3](https://huggingface.co/nomic-ai/gpt4all-j) was observed when using two-shot Chain-of-Thought prompting. | HellaSwag | WinoGrande | BooLQ | ARC-c | |:------:|:------:|:------:|:------:| | 62.37% | 63.3% | 65.2% | 32.76% |