|
--- |
|
license: cc-by-sa-4.0 |
|
dataset_info: |
|
- config_name: default |
|
features: |
|
- name: text |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 203133024 |
|
num_examples: 109486 |
|
- name: validation |
|
num_bytes: 11424453 |
|
num_examples: 6173 |
|
- name: test |
|
num_bytes: 11808744 |
|
num_examples: 6219 |
|
download_size: 143418920 |
|
dataset_size: 226366221 |
|
- config_name: sentences |
|
features: |
|
- name: text |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 202232488.28022403 |
|
num_examples: 1572268 |
|
- name: validation |
|
num_bytes: 11383118.592627235 |
|
num_examples: 88647 |
|
- name: test |
|
num_bytes: 11756845.828945814 |
|
num_examples: 90769 |
|
download_size: 149698561 |
|
dataset_size: 225372452.70179707 |
|
configs: |
|
- config_name: default |
|
data_files: |
|
- split: train |
|
path: data/train-* |
|
- split: validation |
|
path: data/validation-* |
|
- split: test |
|
path: data/test-* |
|
- config_name: sentences |
|
data_files: |
|
- split: train |
|
path: sentences/train-* |
|
- split: validation |
|
path: sentences/validation-* |
|
- split: test |
|
path: sentences/test-* |
|
--- |
|
|
|
# Dataset Card for "wiki40b-da-clean" |
|
|
|
|
|
### Dataset Summary |
|
|
|
This dataset is an slightly modified and filtered version of [Wiki40b-da daset](https://huggingface.co/datasets/alexandrainst/wiki40b-da/) which is a fork of [this dataset on the Hugging Face Hub](https://huggingface.co/datasets/wiki40b). |
|
|
|
The dataset contains two sub-sets, for which the original columns "wikidata_id" and "version_id" are removed from both: |
|
- "**text**": Contains the filtered text of the Wikipedia paragraphs, with formatting removed (_START_ARTICLE_, _START_PARAGRAPH_ and \n removed) |
|
- "**sentences**" Contains the sentences from all the 'text' dataset, filtered to only include sentences >5 and <100 words (split after all punctuations (!,?,.) that is followed by a space and a capital letter) |
|
|
|
The dataset is curated to use the "text" config for masked next token prediction (MNTP) and the sentences config for SimCSE in relation to training encoder and decoder models. |
|
|
|
The training, validation and test splits are the original ones. |
|
|
|
|
|
### Languages |
|
|
|
The dataset is available in Danish (`da`). |
|
|
|
|
|
## Dataset |
|
|
|
**text** (default) |
|
An example from the text dataset looks as follows. |
|
``` |
|
{ |
|
'text': "Tekstiler havde mange forskellige formål i oldtidens Ægypten, og blev brugt af (...)", |
|
} |
|
``` |
|
|
|
**sentences** |
|
An example from the sentences dataset looks as follows. |
|
``` |
|
{ |
|
'text': "Det tog tre måneder, før hørren kunne høstes.", |
|
} |
|
``` |
|
|
|
## Additional Information |
|
|
|
### Dataset Curators |
|
|
|
[Jesper Alkestrup](https://github.com/jalkestrup) from the [The Tech Collective](https://thetechcollective.eu/) filtered and uploaded the dataset to the Hugging Face Hub. |
|
|
|
Thanks to [Dan Saattrup Nielsen](https://saattrupdan.github.io/) from the [The Alexandra |
|
Institute](https://alexandra.dk/) for uploading [Wiki40b-da daset](https://huggingface.co/datasets/alexandrainst/wiki40b-da/). |
|
|
|
### Licensing Information |
|
|
|
The dataset is licensed under the [CC-BY-SA |
|
license](https://creativecommons.org/licenses/by-sa/4.0/). |