PyTorch
Latin
French
Spanish
roberta
Edit model card

Model Details

This is a RoBERTa model trained from scratch on medieval texts. The model is intended to be used as a fondation for other ML tasks on NLP and HTR environments.

The train dataset entails 650M of tokens coming from texts on classical and medieval latin; old french and old Spanish from a period ranging from 5th BC to 16th centuries.

Several big corpora were cleaned and transformed to be used during the training process :

dataset size Lang dates
CC100 [1] 3,2Gb la 5th BC - 18th
Corpus Corporum [2] 3,0Gb la 5th BC - 16th
CEMA [3] 320Mb la+fro 9th - 15th
HOME-Alcar [4] 38Mb la+fro 12th - 15th
BFM [5] 34Mb fro 13th - 15th
AND [6] 19Mb fro 13th - 15th
CODEA [7] 13Mb spa 12th - 16th
~6,5Gb
650M tokens (4,5Gb)*
  • A significant overlapped quantity of text was detected across the corpora, specially on medieval collections. Besides, synthetic text ("Lorem ipsum dolorem...") was iteratively deleted.

[1] CC-NET Repository : https://huggingface.co/datasets/cc100

[2] Repositorium operum lationorum apud universitatem Turicensem : https://mlat.uzh.ch/

[3] Cartae Europae Medii Aevi (5th-15th c.) : https://cema.lamop.fr/

[4] History of Medieval Europe : https://doi.org/10.5281/zenodo.5600884

[5] Base du Français Médieval : https://txm-bfm.huma-num.fr/txm/

[6] Anglo-Normand Dictionary : https://anglo-norman.net/

[7] Corpus de Docuemntos Españoles anteriores a 1900: https://www.corpuscodea.es/

Downloads last month
6
Inference API
Unable to determine this model's library. Check the docs .

Model tree for magistermilitum/Roberta_Historical

Finetunes
1 model

Datasets used to train magistermilitum/Roberta_Historical