File size: 707 Bytes
caeb89f a1c969e |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 |
---
license: mit
language:
- nl
---
# GysBERT v1
This model is a Historical Language Model for Dutch coming from the [MacBERTh project](https://macberth.netlify.app/).
The architecture is based on BERT base uncased from the original BERT pre-training codebase.
The training material comes mostly from the DBNL and the Delpher newspaper dump.
The details can be found in the accompanying publication: [Non-Parametric Word Sense Disambiguation for Historical Languages](https://aclanthology.org/2022.nlp4dh-1.16.pdf)
The model has been successfully tested on Word Sense Disambiguation tasks as discussed in the referenced paper above.
An updated version with an enlarged pre-training dataset is due soon. |