ParaDocs
Data availability limits the scope of any given task. In machine translation, historical models were incapable of handling longer contexts, so the lack of document-level datasets was less noticeable. Now, despite the emergence of long-sequence methods, we remain within a sentence-level paradigm and without data to adequately approach context-aware machine translation. Most large-scale datasets have been processed through a pipeline that discards document-level metadata.
ParaDocs is a publicly available dataset that produces parallel annotations for the document-level metadata of three large publicly available corpora (ParaCrawl, Europal, and News Commentary) in many languages. Using this data and the following scripts, you can download parallel document contexts for the purpose of training context-aware machine translation systems.
If you have questions about this data or use of the following scripts, please do not hesitate to contact the maintainer at rewicks@jhu.edu.
Quick Start
The scripts to download and process the data can be found here:
Clone these scripts:
git clone https://github.com/rewicks/ParaDocs.git
From this directory, you can stream a specific language and split from huggingface with:
paradocs/paradocs-hf --name en-de-strict --minimum_size 2 --frequency_cutoff 100 --lid_cutoff 0.5
It may alternatively be faster to download the *.gz
files of your desired split and then pipe them through the paradocs/paradocs
file for filtering.
zcat data/en-de/strict* | paradocs/paradocs --minimum_size 2 --frequency_cutoff 100 --lid_cutoff 0.5
The filtering commandline arguments are explained in more detial in Section 3.2.
The Paper
If you use this dataset in your research. Please cite our paper.