mt5-small-spanish-summarization
Model description
This is a mt5-small model finetuned for generating headlines from the body of the news in Spanish.
Training data
The model was trained with 58425 news extracted from the La Razón (31477) and Público (26948) newspapers. These news belong to the following categories: "España", "Cultura", "Economía", "Igualdad" and "Política".
Training procedure
It was trained with Google Colab's GPU Tesla P100-PCIE-16GB for 2 epochs.
Hyperparameters
{evaluation_strategy = "epoch", learning_rate = 2e-4, per_device_train_batch_size = 6, per_device_eval_batch_size = 6, weight_decay = 0.01, save_total_limi t= 3, num_train_epochs = 2, predict_with_generate = True, fp16 = False}
Eval results
metric | score |
---|---|
rouge1 | 44.03 |
rouge2 | 28.2900 |
rougeL | 40.54 |
rougeLsum | 40.5587 |
BibTeX entry and citation info
@inproceedings{ mt5lrpjosmunpen,
year={2020},
}
- Downloads last month
- 90
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.