# FLAN-T5-Base pre-trained on Historical Text Completion | |
This model is a pre-trained version of [google/flan-t5-base](https://huggingface.co/google/flan-t5-base) fine-tuned on historical text completion tasks. | |
## Model description | |
This model has been trained to complete historical texts, enhancing its understanding of historical contexts and language. | |
## Intended uses & limitations | |
This model is intended for further fine-tuning on specific historical NLP tasks or for generating historically-aware text completions. | |
## Training and evaluation data | |
The model was trained on a subset of the dataset [ambrosfitz/just_history_xl_masked](https://huggingface.co/datasets/ambrosfitz/just_history_xl_masked), | |
limited by available GPU memory. | |
## Training procedure | |
The model was trained using the following hyperparameters: | |
- Number of epochs: 3 | |
- Batch size: 2 | |
- Learning rate: 5e-05 | |
- Gradient Accumulation Steps: 8 | |
- Weight Decay: 0.01 | |
## Results | |
Evaluation results: {'eval_loss': nan, 'eval_runtime': 0.7866, 'eval_samples_per_second': 38.139, 'eval_steps_per_second': 19.069, 'epoch': 2.909090909090909} | |