|
"""EusCrawl dataset.""" |
|
|
|
import json |
|
import datasets |
|
|
|
|
|
_DESCRIPTION = """\ |
|
EusCrawl (http://www.ixa.eus/euscrawl/) is a high-quality corpus for |
|
Basque comprising 12.5 million documents and 423 million tokens, |
|
totalling 2.1 GiB of uncompressed text. EusCrawl was built using |
|
ad-hoc scrapers to extract text from 33 Basque websites with |
|
high-quality content, resulting in cleaner text compared to general |
|
purpose approaches. |
|
|
|
We do not claim ownership of any document in the corpus. All documents |
|
we collected were published under a Creative Commons license in their |
|
original website, and the specific variant can be found in the |
|
"license" field of each document. Should you consider |
|
that our data contains material that is owned by you and you would not |
|
like to be reproduced here, please contact Aitor Soroa at |
|
a.soroa@ehu.eus. |
|
|
|
For more details about the corpus, refer to our paper "Artetxe M., |
|
Aldabe I., Agerri R., Perez-de-Viñaspre O, Soroa A. (2022). Does |
|
Corpus Quality Really Matter for Low-Resource Languages?" |
|
https://arxiv.org/abs/2203.08111 |
|
|
|
If you use our corpus or models for academic research, please cite the paper in question: |
|
@misc{artetxe2022euscrawl, |
|
title={Does corpus quality really matter for low-resource languages?}, |
|
author={Mikel Artetxe, Itziar Aldabe, Rodrigo Agerri, Olatz Perez-de-Viñaspre, Aitor Soroa}, |
|
year={2022}, |
|
eprint={2203.08111}, |
|
archivePrefix={arXiv}, |
|
primaryClass={cs.CL} |
|
} |
|
|
|
For questions please contact Aitor Soroa at a.soroa@ehu.eus. |
|
""" |
|
_HOMEPAGE_URL = "https://ixa.ehu.eus/euscrawl/" |
|
_CITATION = """\ |
|
@misc{artetxe2022euscrawl, |
|
title={Does corpus quality really matter for low-resource languages?}, |
|
author={Mikel Artetxe, Itziar Aldabe, Rodrigo Agerri, |
|
Olatz Perez-de-Viñaspre, Aitor Soroa}, |
|
year={2022}, |
|
eprint={2203.08111}, |
|
archivePrefix={arXiv}, |
|
primaryClass={cs.CL} |
|
} |
|
""" |
|
|
|
_URL = "http://ixa.ehu.eus/euscrawl/files/euscrawl-v1-free-jsonl.tar.bz2" |
|
_FILEPATH = "euscrawl-v1-free-jsonl/euscrawl-v1.free.jsonl" |
|
KEYS = [ |
|
"plain_text", |
|
"title", |
|
"opening", |
|
"text", |
|
"extra", |
|
"license", |
|
"source", |
|
"url", |
|
"author", |
|
"type", |
|
"lang", |
|
"heading", |
|
"category", |
|
"tags", |
|
"id", |
|
"revid", |
|
"year", |
|
"month", |
|
"day", |
|
"hour", |
|
"minute", |
|
"second", |
|
] |
|
|
|
|
|
class EusCrawl(datasets.GeneratorBasedBuilder): |
|
def _info(self): |
|
return datasets.DatasetInfo( |
|
description=_DESCRIPTION, |
|
features=datasets.Features( |
|
{ |
|
"plain_text": datasets.Value("string"), |
|
"title": datasets.Value("string"), |
|
"opening": datasets.Value("string"), |
|
"text": datasets.Value("string"), |
|
"extra": datasets.Sequence( |
|
{ |
|
"title": datasets.Value("string"), |
|
"text": datasets.Value("string"), |
|
} |
|
), |
|
"license": datasets.Value("string"), |
|
"source": datasets.Value("string"), |
|
"url": datasets.Value("string"), |
|
"author": datasets.Value("string"), |
|
"type": datasets.Value("string"), |
|
"lang": datasets.Value("string"), |
|
"heading": datasets.Value("string"), |
|
"category": datasets.Sequence(datasets.Value("string")), |
|
"tags": datasets.Sequence(datasets.Value("string")), |
|
"id": datasets.Value("int32"), |
|
"revid": datasets.Value("int32"), |
|
"year": datasets.Value("int32"), |
|
"month": datasets.Value("int32"), |
|
"day": datasets.Value("int32"), |
|
"hour": datasets.Value("int32"), |
|
"minute": datasets.Value("int32"), |
|
"second": datasets.Value("int32"), |
|
} |
|
), |
|
supervised_keys=None, |
|
homepage=_HOMEPAGE_URL, |
|
citation=_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
path = dl_manager.download(_URL) |
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
gen_kwargs={"filepaths": dl_manager.iter_archive(path)}, |
|
) |
|
] |
|
|
|
def _generate_examples(self, filepaths): |
|
for filepath, file in filepaths: |
|
if filepath == _FILEPATH: |
|
for id, line in enumerate(file): |
|
data = json.loads(line) |
|
plain_text_lines = [] |
|
plain_text_lines += data.get("title", "").splitlines() |
|
plain_text_lines += data.get("opening", "").splitlines() |
|
plain_text_lines += data.get("text", "").splitlines() |
|
plain_text_lines += [ |
|
line |
|
for extra in data.get("extra", []) |
|
for line in extra["title"].splitlines() |
|
+ extra["text"].splitlines() |
|
] |
|
plain_text_lines = [ |
|
line.strip() for line in plain_text_lines if line.strip() |
|
] |
|
data["plain_text"] = "\n".join(plain_text_lines) |
|
|
|
yield id, {key: data.get(key, None) for key in KEYS} |
|
|