politibeto / README.md
villacu's picture
Update README.md
fa286a1
---
language:
- es
tags:
- masked-lm
widget:
- text: "La mayor ventaja de la democracia es su [MASK]."
example_title: "Ejemplo 1"
---
# PolitiBETO: A Spanish BERT adapted to a language domain of Political Tweets
PolitiBETO is a [BERT model](https://github.com/google-research/bert) tailored for political tasks in social media corpora. It is a Domain Adaptation on top of [BETO](https://huggingface.co/dccuchile/bert-base-spanish-wwm-uncased), a pretrained BERT in Spanish.
This model is meant to be fine-tuned for downstream tasks.
## Citation
[NLP-CIMAT at PoliticEs 2022: PolitiBETO, a Domain-Adapted Transformer for Multi-class Political Author Profiling](https://ceur-ws.org/Vol-3202/politices-paper2.pdf)
To cite this in a publication please use the following:
```
@inproceedings{PolitiBeto2022,
title={{NLP-CIMAT} at {P}olitic{E}s 2022: {P}oliti{BETO}, a {D}omain-{A}dapted {T}ransformer for {M}ulti-class {P}olitical {A}uthor {P}rofiling},
author={Emilio Villa-Cueva and Ivan Gonz{\'a}lez-Franco and Fernando Sanchez-Vega and Adri{\'a}n Pastor L{\'o}pez-Monroy},
booktitle={Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2022)},
series = {{CEUR} Workshop Proceedings},
publisher = {CEUR-WS},
year={2022}
}
```