|
--- |
|
language: |
|
- om |
|
- am |
|
- rw |
|
- rn |
|
- ha |
|
- ig |
|
- pcm |
|
- so |
|
- sw |
|
- ti |
|
- yo |
|
- multilingual |
|
|
|
license: "Apache License 2.0" |
|
|
|
task_categories: |
|
- sequence-modeling |
|
task_ids: |
|
- language-modeling |
|
--- |
|
|
|
# Dataset Card for AfriBERTa's Corpus |
|
|
|
## Table of Contents |
|
- [Dataset Description](#dataset-description) |
|
- [Dataset Summary](#dataset-summary) |
|
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) |
|
- [Languages](#languages) |
|
- [Loading Dataset](#loading-dataset) |
|
- [Dataset Structure](#dataset-structure) |
|
- [Data Instances](#data-instances) |
|
- [Data Fields](#data-fields) |
|
- [Data Splits](#data-splits) |
|
- [Considerations for Using the Data](#considerations-for-using-the-data) |
|
- [Discussion of Biases](#discussion-of-biases) |
|
- [Additional Information](#additional-information) |
|
- [Citation Information](#citation-information) |
|
- [Contributions](#contributions) |
|
|
|
## Dataset Description |
|
|
|
### Dataset Summary |
|
This is the corpus on which AfriBERTa was trained on. |
|
The dataset is mostly from the BBC news website, but some languages also have data from Common Crawl. |
|
|
|
- **Homepage:** https://github.com/keleog/afriberta |
|
- **Models:** |
|
- https://huggingface.co/castorini/afriberta_small |
|
- https://huggingface.co/castorini/afriberta_base |
|
- https://huggingface.co/castorini/afriberta_large |
|
- **Paper:** https://aclanthology.org/2021.mrl-1.11/ |
|
- **Point of Contact:** kelechi.ogueji@uwaterloo.ca |
|
|
|
|
|
### Supported Tasks and Leaderboards |
|
The AfriBERTa corpus was mostly intended to pre-train language models. |
|
|
|
### Languages |
|
``` |
|
afaanoromoo |
|
amharic |
|
gahuza |
|
hausa |
|
igbo |
|
pidgin |
|
somali |
|
swahili |
|
tigrinya |
|
yoruba |
|
``` |
|
### Loading Dataset |
|
An example to load the train split of the Somali corpus: |
|
``` |
|
dataset = load_dataset("castorini/afriberta-corpus", "somali", split="train") |
|
``` |
|
|
|
An example to load the test split of the Pidgin corpus: |
|
``` |
|
dataset = load_dataset("castorini/afriberta-corpus", "pidgin", split="test") |
|
``` |
|
|
|
## Dataset Structure |
|
|
|
### Data Instances |
|
Each data point is a line of text. |
|
An example from the `igbo` dataset: |
|
``` |
|
{"id": "6", "text": "Ngwá ọrụ na-echebe ma na-ebuli gị na kọmputa."} |
|
``` |
|
|
|
### Data Fields |
|
|
|
The data fields are: |
|
|
|
- id: id of the example |
|
- text: content as a string |
|
|
|
### Data Splits |
|
Each language has a train and test split, with varying sizes. |
|
|
|
## Considerations for Using the Data |
|
|
|
### Discussion of Biases |
|
Since majority of the data is obtained from the BBC's news website, models trained on this dataset are likely going to |
|
be biased towards the news domain. |
|
|
|
Also, since some of the data is obtained from Common Crawl, care should be taken (especially for text generation models) since personal and sensitive information might be present. |
|
|
|
## Additional Information |
|
### Citation Information |
|
``` |
|
@inproceedings{ogueji-etal-2021-small, |
|
title = "Small Data? No Problem! Exploring the Viability of Pretrained Multilingual Language Models for Low-resourced Languages", |
|
author = "Ogueji, Kelechi and |
|
Zhu, Yuxin and |
|
Lin, Jimmy", |
|
booktitle = "Proceedings of the 1st Workshop on Multilingual Representation Learning", |
|
month = nov, |
|
year = "2021", |
|
address = "Punta Cana, Dominican Republic", |
|
publisher = "Association for Computational Linguistics", |
|
url = "https://aclanthology.org/2021.mrl-1.11", |
|
pages = "116--126", |
|
} |
|
``` |
|
|
|
### Contributions |
|
Thanks to [Kelechi Ogueji](https://github.com/keleog) for adding this dataset. |