Datasets:
pretty_name: Korpus Malti
language:
- mt
multilinguality:
- monolingual
size_categories:
- 10M<n<100M
annotations_creators:
- no-annotation
language_creators:
- found
source_datasets:
- original
task_categories:
- text-generation
- fill-mask
task_ids:
- language-modeling
- masked-language-modeling
license:
- cc-by-nc-sa-4.0
Korpus Malti 🇲🇹
General Corpora for the Maltese Language.
This dataset is composed of texts from various genres/domains written in Maltese.
Configurations
Shuffled data
The default configuration ("shuffled"
) yields the the entire corpus from all genres:
import datasets
dataset = datasets.load_dataset("MLRS/korpus_malti")
All sentences are combined together and shuffled, without preserving the sentence order. No other annotations are present, so an instance would be of the following form:
{
"text": "Din hija sentenza."
}
The training/validation/testing split is what was used to train the BERTu model.
Domain-split data
All other configurations contain a subset of the data. For instance, this loads the Wikipedia portion:
import datasets
dataset = datasets.load_dataset("MLRS/korpus_malti", "wiki")
For these configurations the data is not shuffled, so the sentence order on a document level is preserved. An instance from these configurations would take the following form:
{
"text": ["Din hija sentenza.", "U hawn oħra!"],
}
The raw data files contain additional metadata. Its structure differs from one instance to another, depending on what's available from the source. This information was typically scraped from the source itself & minimal processing is performed on such data.
Additional Information
Dataset Curators
The dataset was created by Albert Gatt, Kurt Micallef, Marc Tanti, Lonneke van der Plas and Claudia Borg.
Licensing Information
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License]cc-by-nc-sa. Permissions beyond the scope of this license may be available at https://mlrs.research.um.edu.mt/.
Citation Information
This work was first presented in Pre-training Data Quality and Quantity for a Low-Resource Language: New Corpus and BERT Models for Maltese. Cite it as follows:
@inproceedings{BERTu,
title = "Pre-training Data Quality and Quantity for a Low-Resource Language: New Corpus and {BERT} Models for {M}altese",
author = "Micallef, Kurt and
Gatt, Albert and
Tanti, Marc and
van der Plas, Lonneke and
Borg, Claudia",
booktitle = "Proceedings of the Third Workshop on Deep Learning for Low-Resource Natural Language Processing",
month = jul,
year = "2022",
address = "Hybrid",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.deeplo-1.10",
doi = "10.18653/v1/2022.deeplo-1.10",
pages = "90--101",
}