metadata
language:
- en
tags:
- text-generation-inference
Model Card for Mistral-7B pre-trained
Model Description
This model is a fine-tuned Mistral-7B model on a collection of books.
- Language(s) (NLP): English
- Finetuned from model: Mistral-7B
- Dataset used for fine-tuning: Public Domain Gutenberg Corpus fiction books
This model was pre-trained on a collection of public domain books from Gutenberg. It is meant to be further fine-tuned on task specific data related to narrative content.