DanskBERT / README.md
vesteinn's picture
Update README.md
7e282ae
|
raw
history blame
1.31 kB
metadata
language: da
widget:
  - text: En trend, der kan blive ligeså hot som<mask>.
tags:
  - roberta
  - danish
  - masked-lm
  - pytorch
license: cc-by-4.0

DanskBERT

This is DanskBERT, a Danish language model. Note that you should not prepend the mask with a space when using it directly!

The model is the best performing base-size model on the ScandEval benchmark for Danish.

DanskBERT was trained on the Danish Gigaword Corpus (Strømberg-Derczynski et al., 2021).

DanskBERT was trained using fairseq using the RoBERTa-base configuration. The model was trained with a batch size of 2k, and was trained to convergence for 500k steps using 16 V100 cards for approximately two weeks.

If you find this model useful, please cite

@inproceedings{snaebjarnarson-etal-2023-transfer,
    title = "{T}ransfer to a Low-Resource Language via Close Relatives: The Case Study on Faroese",
    author = "Snæbjarnarson, Vésteinn  and
      Simonsen, Annika  and
      Glavaš, Goran  and
      Vulić, Ivan",
    booktitle = "Proceedings of the 24th Nordic Conference on Computational Linguistics (NoDaLiDa)",
    month = "may 22--24",
    year = "2023",
    address = "Tórshavn, Faroe Islands",
    publisher = {Link{\"o}ping University Electronic Press, Sweden},
}