de-francophones's picture
Upload 162 files
942801f verified
# eng.rst.gum
The Georgetown University Multilayer (GUM) corpus
To cite this corpus, please refer to the following article:
Zeldes, Amir (2017) "The GUM Corpus: Creating Multilayer Resources in the Classroom". Language Resources and Evaluation 51(3), 581–612.
## Introduction
a corpus of English texts from twelve text types:
Interviews from Wikimedia
News from Wikinews
Travel guides from Wikivoyage
How-to guides from wikiHow
Academic writing from Creative Commons sources
Biographies from Wikipedia
Fiction from Creative Commons sources
Online forum discussions from Reddit
Face-to-face conversations from the Santa Barbara Corpus
Political speeches
OpenStax open access textbooks
YouTube Creative Commons vlogs
The corpus is created as part of the course LING-367 (Computational Corpus Linguistics) at Georgetown University.
For more details see: https://gucorpling.org/gum
## DISRPT 2021 shared task information
For the DISRPT 2021 shared task on elementary discourse unit segmentation, only 11 open text genres are included with plain text, while the remaining genre, containing Reddit forum discussions, **must be reconstructed** using the script in `utils/process_underscores.py` (see main repository README). The data follows the normal division into train, test and dev partitions used for other tasks (e.g. for the conll shared task on UD parsing).
POS tags and syntactic parses are manually annotated gold data.
### Notes on segmentation
GUM RST guidelines follow the RST-DT segmentation guidelines for English, according to which most clauses, including adnominal and nested clauses are discourse units. This dataset contains discontinuous discourse units (split 'same-unit'). Note that the .conllu data contains some reconstructed ellipsis tokens with decimal IDs (e.g. 12.1); these do not appear in the other formats and are ignored in token index spans.