|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""The SuperGLUE benchmark.""" |
|
|
|
|
|
import json |
|
import os |
|
|
|
import datasets |
|
|
|
|
|
_SUPER_GLUE_CITATION = """\ |
|
@article{wang2019superglue, |
|
title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, |
|
author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, |
|
journal={arXiv preprint arXiv:1905.00537}, |
|
year={2019} |
|
} |
|
|
|
Note that each SuperGLUE dataset has its own citation. Please see the source to |
|
get the correct citation for each contained dataset. |
|
""" |
|
|
|
_GLUE_DESCRIPTION = """\ |
|
SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after |
|
GLUE with a new set of more difficult language understanding tasks, improved |
|
resources, and a new public leaderboard. |
|
|
|
""" |
|
|
|
_BOOLQ_DESCRIPTION = """\ |
|
BoolQ (Boolean Questions, Clark et al., 2019a) is a QA task where each example consists of a short |
|
passage and a yes/no question about the passage. The questions are provided anonymously and |
|
unsolicited by users of the Google search engine, and afterwards paired with a paragraph from a |
|
Wikipedia article containing the answer. Following the original work, we evaluate with accuracy.""" |
|
|
|
_CB_DESCRIPTION = """\ |
|
The CommitmentBank (De Marneffe et al., 2019) is a corpus of short texts in which at least |
|
one sentence contains an embedded clause. Each of these embedded clauses is annotated with the |
|
degree to which we expect that the person who wrote the text is committed to the truth of the clause. |
|
The resulting task framed as three-class textual entailment on examples that are drawn from the Wall |
|
Street Journal, fiction from the British National Corpus, and Switchboard. Each example consists |
|
of a premise containing an embedded clause and the corresponding hypothesis is the extraction of |
|
that clause. We use a subset of the data that had inter-annotator agreement above 0.85. The data is |
|
imbalanced (relatively fewer neutral examples), so we evaluate using accuracy and F1, where for |
|
multi-class F1 we compute the unweighted average of the F1 per class.""" |
|
|
|
_COPA_DESCRIPTION = """\ |
|
The Choice Of Plausible Alternatives (COPA, Roemmele et al., 2011) dataset is a causal |
|
reasoning task in which a system is given a premise sentence and two possible alternatives. The |
|
system must choose the alternative which has the more plausible causal relationship with the premise. |
|
The method used for the construction of the alternatives ensures that the task requires causal reasoning |
|
to solve. Examples either deal with alternative possible causes or alternative possible effects of the |
|
premise sentence, accompanied by a simple question disambiguating between the two instance |
|
types for the model. All examples are handcrafted and focus on topics from online blogs and a |
|
photography-related encyclopedia. Following the recommendation of the authors, we evaluate using |
|
accuracy.""" |
|
|
|
_RECORD_DESCRIPTION = """\ |
|
(Reading Comprehension with Commonsense Reasoning Dataset, Zhang et al., 2018) is a |
|
multiple-choice QA task. Each example consists of a news article and a Cloze-style question about |
|
the article in which one entity is masked out. The system must predict the masked out entity from a |
|
given list of possible entities in the provided passage, where the same entity may be expressed using |
|
multiple different surface forms, all of which are considered correct. Articles are drawn from CNN |
|
and Daily Mail. Following the original work, we evaluate with max (over all mentions) token-level |
|
F1 and exact match (EM).""" |
|
|
|
_RTE_DESCRIPTION = """\ |
|
The Recognizing Textual Entailment (RTE) datasets come from a series of annual competitions |
|
on textual entailment, the problem of predicting whether a given premise sentence entails a given |
|
hypothesis sentence (also known as natural language inference, NLI). RTE was previously included |
|
in GLUE, and we use the same data and format as before: We merge data from RTE1 (Dagan |
|
et al., 2006), RTE2 (Bar Haim et al., 2006), RTE3 (Giampiccolo et al., 2007), and RTE5 (Bentivogli |
|
et al., 2009). All datasets are combined and converted to two-class classification: entailment and |
|
not_entailment. Of all the GLUE tasks, RTE was among those that benefited from transfer learning |
|
the most, jumping from near random-chance performance (~56%) at the time of GLUE's launch to |
|
85% accuracy (Liu et al., 2019c) at the time of writing. Given the eight point gap with respect to |
|
human performance, however, the task is not yet solved by machines, and we expect the remaining |
|
gap to be difficult to close.""" |
|
|
|
_MULTIRC_DESCRIPTION = """\ |
|
The Multi-Sentence Reading Comprehension dataset (MultiRC, Khashabi et al., 2018) |
|
is a true/false question-answering task. Each example consists of a context paragraph, a question |
|
about that paragraph, and a list of possible answers to that question which must be labeled as true or |
|
false. Question-answering (QA) is a popular problem with many datasets. We use MultiRC because |
|
of a number of desirable properties: (i) each question can have multiple possible correct answers, |
|
so each question-answer pair must be evaluated independent of other pairs, (ii) the questions are |
|
designed such that answering each question requires drawing facts from multiple context sentences, |
|
and (iii) the question-answer pair format more closely matches the API of other SuperGLUE tasks |
|
than span-based extractive QA does. The paragraphs are drawn from seven domains including news, |
|
fiction, and historical text.""" |
|
|
|
_WIC_DESCRIPTION = """\ |
|
The Word-in-Context (WiC, Pilehvar and Camacho-Collados, 2019) dataset supports a word |
|
sense disambiguation task cast as binary classification over sentence pairs. Given two sentences and a |
|
polysemous (sense-ambiguous) word that appears in both sentences, the task is to determine whether |
|
the word is used with the same sense in both sentences. Sentences are drawn from WordNet (Miller, |
|
1995), VerbNet (Schuler, 2005), and Wiktionary. We follow the original work and evaluate using |
|
accuracy.""" |
|
|
|
_WSC_DESCRIPTION = """\ |
|
The Winograd Schema Challenge (WSC, Levesque et al., 2012) is a reading comprehension |
|
task in which a system must read a sentence with a pronoun and select the referent of that pronoun |
|
from a list of choices. Given the difficulty of this task and the headroom still left, we have included |
|
WSC in SuperGLUE and recast the dataset into its coreference form. The task is cast as a binary |
|
classification problem, as opposed to N-multiple choice, in order to isolate the model's ability to |
|
understand the coreference links within a sentence as opposed to various other strategies that may |
|
come into play in multiple choice conditions. With that in mind, we create a split with 65% negative |
|
majority class in the validation set, reflecting the distribution of the hidden test set, and 52% negative |
|
class in the training set. The training and validation examples are drawn from the original Winograd |
|
Schema dataset (Levesque et al., 2012), as well as those distributed by the affiliated organization |
|
Commonsense Reasoning. The test examples are derived from fiction books and have been shared |
|
with us by the authors of the original dataset. Previously, a version of WSC recast as NLI as included |
|
in GLUE, known as WNLI. No substantial progress was made on WNLI, with many submissions |
|
opting to submit only majority class predictions. WNLI was made especially difficult due to an |
|
adversarial train/dev split: Premise sentences that appeared in the training set sometimes appeared |
|
in the development set with a different hypothesis and a flipped label. If a system memorized the |
|
training set without meaningfully generalizing, which was easy due to the small size of the training |
|
set, it could perform far below chance on the development set. We remove this adversarial design |
|
in the SuperGLUE version of WSC by ensuring that no sentences are shared between the training, |
|
validation, and test sets. |
|
|
|
However, the validation and test sets come from different domains, with the validation set consisting |
|
of ambiguous examples such that changing one non-noun phrase word will change the coreference |
|
dependencies in the sentence. The test set consists only of more straightforward examples, with a |
|
high number of noun phrases (and thus more choices for the model), but low to no ambiguity.""" |
|
|
|
_AXB_DESCRIPTION = """\ |
|
An expert-constructed, |
|
diagnostic dataset that automatically tests models for a broad range of linguistic, commonsense, and |
|
world knowledge. Each example in this broad-coverage diagnostic is a sentence pair labeled with |
|
a three-way entailment relation (entailment, neutral, or contradiction) and tagged with labels that |
|
indicate the phenomena that characterize the relationship between the two sentences. Submissions |
|
to the GLUE leaderboard are required to include predictions from the submission's MultiNLI |
|
classifier on the diagnostic dataset, and analyses of the results were shown alongside the main |
|
leaderboard. Since this broad-coverage diagnostic task has proved difficult for top models, we retain |
|
it in SuperGLUE. However, since MultiNLI is not part of SuperGLUE, we collapse contradiction |
|
and neutral into a single not_entailment label, and request that submissions include predictions |
|
on the resulting set from the model used for the RTE task. |
|
""" |
|
|
|
_AXG_DESCRIPTION = """\ |
|
Winogender is designed to measure gender |
|
bias in coreference resolution systems. We use the Diverse Natural Language Inference Collection |
|
(DNC; Poliak et al., 2018) version that casts Winogender as a textual entailment task. Each example |
|
consists of a premise sentence with a male or female pronoun and a hypothesis giving a possible |
|
antecedent of the pronoun. Examples occur in minimal pairs, where the only difference between |
|
an example and its pair is the gender of the pronoun in the premise. Performance on Winogender |
|
is measured with both accuracy and the gender parity score: the percentage of minimal pairs for |
|
which the predictions are the same. We note that a system can trivially obtain a perfect gender parity |
|
score by guessing the same class for all examples, so a high gender parity score is meaningless unless |
|
accompanied by high accuracy. As a diagnostic test of gender bias, we view the schemas as having high |
|
positive predictive value and low negative predictive value; that is, they may demonstrate the presence |
|
of gender bias in a system, but not prove its absence. |
|
""" |
|
|
|
_BOOLQ_CITATION = """\ |
|
@inproceedings{clark2019boolq, |
|
title={BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions}, |
|
author={Clark, Christopher and Lee, Kenton and Chang, Ming-Wei, and Kwiatkowski, Tom and Collins, Michael, and Toutanova, Kristina}, |
|
booktitle={NAACL}, |
|
year={2019} |
|
}""" |
|
|
|
_CB_CITATION = """\ |
|
@article{de marneff_simons_tonhauser_2019, |
|
title={The CommitmentBank: Investigating projection in naturally occurring discourse}, |
|
journal={proceedings of Sinn und Bedeutung 23}, |
|
author={De Marneff, Marie-Catherine and Simons, Mandy and Tonhauser, Judith}, |
|
year={2019} |
|
}""" |
|
|
|
_COPA_CITATION = """\ |
|
@inproceedings{roemmele2011choice, |
|
title={Choice of plausible alternatives: An evaluation of commonsense causal reasoning}, |
|
author={Roemmele, Melissa and Bejan, Cosmin Adrian and Gordon, Andrew S}, |
|
booktitle={2011 AAAI Spring Symposium Series}, |
|
year={2011} |
|
}""" |
|
|
|
_RECORD_CITATION = """\ |
|
@article{zhang2018record, |
|
title={Record: Bridging the gap between human and machine commonsense reading comprehension}, |
|
author={Zhang, Sheng and Liu, Xiaodong and Liu, Jingjing and Gao, Jianfeng and Duh, Kevin and Van Durme, Benjamin}, |
|
journal={arXiv preprint arXiv:1810.12885}, |
|
year={2018} |
|
}""" |
|
|
|
_RTE_CITATION = """\ |
|
@inproceedings{dagan2005pascal, |
|
title={The PASCAL recognising textual entailment challenge}, |
|
author={Dagan, Ido and Glickman, Oren and Magnini, Bernardo}, |
|
booktitle={Machine Learning Challenges Workshop}, |
|
pages={177--190}, |
|
year={2005}, |
|
organization={Springer} |
|
} |
|
@inproceedings{bar2006second, |
|
title={The second pascal recognising textual entailment challenge}, |
|
author={Bar-Haim, Roy and Dagan, Ido and Dolan, Bill and Ferro, Lisa and Giampiccolo, Danilo and Magnini, Bernardo and Szpektor, Idan}, |
|
booktitle={Proceedings of the second PASCAL challenges workshop on recognising textual entailment}, |
|
volume={6}, |
|
number={1}, |
|
pages={6--4}, |
|
year={2006}, |
|
organization={Venice} |
|
} |
|
@inproceedings{giampiccolo2007third, |
|
title={The third pascal recognizing textual entailment challenge}, |
|
author={Giampiccolo, Danilo and Magnini, Bernardo and Dagan, Ido and Dolan, Bill}, |
|
booktitle={Proceedings of the ACL-PASCAL workshop on textual entailment and paraphrasing}, |
|
pages={1--9}, |
|
year={2007}, |
|
organization={Association for Computational Linguistics} |
|
} |
|
@inproceedings{bentivogli2009fifth, |
|
title={The Fifth PASCAL Recognizing Textual Entailment Challenge.}, |
|
author={Bentivogli, Luisa and Clark, Peter and Dagan, Ido and Giampiccolo, Danilo}, |
|
booktitle={TAC}, |
|
year={2009} |
|
}""" |
|
|
|
_MULTIRC_CITATION = """\ |
|
@inproceedings{MultiRC2018, |
|
author = {Daniel Khashabi and Snigdha Chaturvedi and Michael Roth and Shyam Upadhyay and Dan Roth}, |
|
title = {Looking Beyond the Surface:A Challenge Set for Reading Comprehension over Multiple Sentences}, |
|
booktitle = {Proceedings of North American Chapter of the Association for Computational Linguistics (NAACL)}, |
|
year = {2018} |
|
}""" |
|
|
|
_WIC_CITATION = """\ |
|
@article{DBLP:journals/corr/abs-1808-09121, |
|
author={Mohammad Taher Pilehvar and os{\'{e}} Camacho{-}Collados}, |
|
title={WiC: 10, 000 Example Pairs for Evaluating Context-Sensitive Representations}, |
|
journal={CoRR}, |
|
volume={abs/1808.09121}, |
|
year={2018}, |
|
url={http://arxiv.org/abs/1808.09121}, |
|
archivePrefix={arXiv}, |
|
eprint={1808.09121}, |
|
timestamp={Mon, 03 Sep 2018 13:36:40 +0200}, |
|
biburl={https://dblp.org/rec/bib/journals/corr/abs-1808-09121}, |
|
bibsource={dblp computer science bibliography, https://dblp.org} |
|
}""" |
|
|
|
_WSC_CITATION = """\ |
|
@inproceedings{levesque2012winograd, |
|
title={The winograd schema challenge}, |
|
author={Levesque, Hector and Davis, Ernest and Morgenstern, Leora}, |
|
booktitle={Thirteenth International Conference on the Principles of Knowledge Representation and Reasoning}, |
|
year={2012} |
|
}""" |
|
|
|
_AXG_CITATION = """\ |
|
@inproceedings{rudinger-EtAl:2018:N18, |
|
author = {Rudinger, Rachel and Naradowsky, Jason and Leonard, Brian and {Van Durme}, Benjamin}, |
|
title = {Gender Bias in Coreference Resolution}, |
|
booktitle = {Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies}, |
|
month = {June}, |
|
year = {2018}, |
|
address = {New Orleans, Louisiana}, |
|
publisher = {Association for Computational Linguistics} |
|
} |
|
""" |
|
|
|
|
|
class SuperGlueConfig(datasets.BuilderConfig): |
|
"""BuilderConfig for SuperGLUE.""" |
|
|
|
def __init__(self, features, data_url, citation, url, label_classes=("False", "True"), **kwargs): |
|
"""BuilderConfig for SuperGLUE. |
|
|
|
Args: |
|
features: `list[string]`, list of the features that will appear in the |
|
feature dict. Should not include "label". |
|
data_url: `string`, url to download the zip file from. |
|
citation: `string`, citation for the data set. |
|
url: `string`, url for information about the data set. |
|
label_classes: `list[string]`, the list of classes for the label if the |
|
label is present as a string. Non-string labels will be cast to either |
|
'False' or 'True'. |
|
**kwargs: keyword arguments forwarded to super. |
|
""" |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
super(SuperGlueConfig, self).__init__(version=datasets.Version("1.0.3"), **kwargs) |
|
self.features = features |
|
self.label_classes = label_classes |
|
self.data_url = data_url |
|
self.citation = citation |
|
self.url = url |
|
|
|
|
|
class SuperGlue(datasets.GeneratorBasedBuilder): |
|
"""The SuperGLUE benchmark.""" |
|
|
|
BUILDER_CONFIGS = [ |
|
SuperGlueConfig( |
|
name="boolq", |
|
description=_BOOLQ_DESCRIPTION, |
|
features=["question", "passage"], |
|
data_url="https://dl.fbaipublicfiles.com/glue/superglue/data/v2/BoolQ.zip", |
|
citation=_BOOLQ_CITATION, |
|
url="https://github.com/google-research-datasets/boolean-questions", |
|
), |
|
SuperGlueConfig( |
|
name="cb", |
|
description=_CB_DESCRIPTION, |
|
features=["premise", "hypothesis"], |
|
label_classes=["entailment", "contradiction", "neutral"], |
|
data_url="https://dl.fbaipublicfiles.com/glue/superglue/data/v2/CB.zip", |
|
citation=_CB_CITATION, |
|
url="https://github.com/mcdm/CommitmentBank", |
|
), |
|
SuperGlueConfig( |
|
name="copa", |
|
description=_COPA_DESCRIPTION, |
|
label_classes=["choice1", "choice2"], |
|
|
|
|
|
features=["premise", "choice1", "choice2", "question"], |
|
data_url="https://dl.fbaipublicfiles.com/glue/superglue/data/v2/COPA.zip", |
|
citation=_COPA_CITATION, |
|
url="http://people.ict.usc.edu/~gordon/copa.html", |
|
), |
|
SuperGlueConfig( |
|
name="multirc", |
|
description=_MULTIRC_DESCRIPTION, |
|
features=["paragraph", "question", "answer"], |
|
data_url="https://dl.fbaipublicfiles.com/glue/superglue/data/v2/MultiRC.zip", |
|
citation=_MULTIRC_CITATION, |
|
url="https://cogcomp.org/multirc/", |
|
), |
|
SuperGlueConfig( |
|
name="record", |
|
description=_RECORD_DESCRIPTION, |
|
|
|
|
|
|
|
features=["passage", "query", "entities", "entity_spans", "answers"], |
|
data_url="https://dl.fbaipublicfiles.com/glue/superglue/data/v2/ReCoRD.zip", |
|
citation=_RECORD_CITATION, |
|
url="https://sheng-z.github.io/ReCoRD-explorer/", |
|
), |
|
SuperGlueConfig( |
|
name="rte", |
|
description=_RTE_DESCRIPTION, |
|
features=["premise", "hypothesis"], |
|
label_classes=["entailment", "not_entailment"], |
|
data_url="https://dl.fbaipublicfiles.com/glue/superglue/data/v2/RTE.zip", |
|
citation=_RTE_CITATION, |
|
url="https://aclweb.org/aclwiki/Recognizing_Textual_Entailment", |
|
), |
|
SuperGlueConfig( |
|
name="wic", |
|
description=_WIC_DESCRIPTION, |
|
|
|
|
|
features=["word", "sentence1", "sentence2", "start1", "start2", "end1", "end2"], |
|
data_url="https://dl.fbaipublicfiles.com/glue/superglue/data/v2/WiC.zip", |
|
citation=_WIC_CITATION, |
|
url="https://pilehvar.github.io/wic/", |
|
), |
|
SuperGlueConfig( |
|
name="wsc", |
|
description=_WSC_DESCRIPTION, |
|
|
|
|
|
features=["text", "span1_index", "span2_index", "span1_text", "span2_text"], |
|
data_url="https://dl.fbaipublicfiles.com/glue/superglue/data/v2/WSC.zip", |
|
citation=_WSC_CITATION, |
|
url="https://cs.nyu.edu/faculty/davise/papers/WinogradSchemas/WS.html", |
|
), |
|
SuperGlueConfig( |
|
name="wsc.fixed", |
|
description=( |
|
_WSC_DESCRIPTION + "\n\nThis version fixes issues where the spans are not actually " |
|
"substrings of the text." |
|
), |
|
|
|
|
|
features=["text", "span1_index", "span2_index", "span1_text", "span2_text"], |
|
data_url="https://dl.fbaipublicfiles.com/glue/superglue/data/v2/WSC.zip", |
|
citation=_WSC_CITATION, |
|
url="https://cs.nyu.edu/faculty/davise/papers/WinogradSchemas/WS.html", |
|
), |
|
SuperGlueConfig( |
|
name="axb", |
|
description=_AXB_DESCRIPTION, |
|
features=["sentence1", "sentence2"], |
|
label_classes=["entailment", "not_entailment"], |
|
data_url="https://dl.fbaipublicfiles.com/glue/superglue/data/v2/AX-b.zip", |
|
citation="", |
|
url="https://gluebenchmark.com/diagnostics", |
|
), |
|
SuperGlueConfig( |
|
name="axg", |
|
description=_AXG_DESCRIPTION, |
|
features=["premise", "hypothesis"], |
|
label_classes=["entailment", "not_entailment"], |
|
data_url="https://dl.fbaipublicfiles.com/glue/superglue/data/v2/AX-g.zip", |
|
citation=_AXG_CITATION, |
|
url="https://github.com/rudinger/winogender-schemas", |
|
), |
|
] |
|
|
|
def _info(self): |
|
features = {feature: datasets.Value("string") for feature in self.config.features} |
|
if self.config.name.startswith("wsc"): |
|
features["span1_index"] = datasets.Value("int32") |
|
features["span2_index"] = datasets.Value("int32") |
|
if self.config.name == "wic": |
|
features["start1"] = datasets.Value("int32") |
|
features["start2"] = datasets.Value("int32") |
|
features["end1"] = datasets.Value("int32") |
|
features["end2"] = datasets.Value("int32") |
|
if self.config.name == "multirc": |
|
features["idx"] = dict( |
|
{ |
|
"paragraph": datasets.Value("int32"), |
|
"question": datasets.Value("int32"), |
|
"answer": datasets.Value("int32"), |
|
} |
|
) |
|
elif self.config.name == "record": |
|
features["idx"] = dict( |
|
{ |
|
"passage": datasets.Value("int32"), |
|
"query": datasets.Value("int32"), |
|
} |
|
) |
|
else: |
|
features["idx"] = datasets.Value("int32") |
|
|
|
if self.config.name == "record": |
|
|
|
features["entities"] = datasets.features.Sequence(datasets.Value("string")) |
|
|
|
features["entity_spans"] = datasets.features.Sequence( |
|
{ |
|
"text": datasets.Value("string"), |
|
"start": datasets.Value("int32"), |
|
"end": datasets.Value("int32"), |
|
} |
|
) |
|
|
|
features["answers"] = datasets.features.Sequence(datasets.Value("string")) |
|
else: |
|
features["label"] = datasets.features.ClassLabel(names=self.config.label_classes) |
|
|
|
return datasets.DatasetInfo( |
|
description=_GLUE_DESCRIPTION + self.config.description, |
|
features=datasets.Features(features), |
|
homepage=self.config.url, |
|
citation=self.config.citation + "\n" + _SUPER_GLUE_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
dl_dir = dl_manager.download_and_extract(self.config.data_url) or "" |
|
task_name = _get_task_name_from_data_url(self.config.data_url) |
|
dl_dir = os.path.join(dl_dir, task_name) |
|
if self.config.name in ["axb", "axg"]: |
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TEST, |
|
gen_kwargs={ |
|
"data_file": os.path.join(dl_dir, f"{task_name}.jsonl"), |
|
"split": datasets.Split.TEST, |
|
}, |
|
), |
|
] |
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
gen_kwargs={ |
|
"data_file": os.path.join(dl_dir, "train.jsonl"), |
|
"split": datasets.Split.TRAIN, |
|
}, |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.VALIDATION, |
|
gen_kwargs={ |
|
"data_file": os.path.join(dl_dir, "val.jsonl"), |
|
"split": datasets.Split.VALIDATION, |
|
}, |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TEST, |
|
gen_kwargs={ |
|
"data_file": os.path.join(dl_dir, "test.jsonl"), |
|
"split": datasets.Split.TEST, |
|
}, |
|
), |
|
] |
|
|
|
def _generate_examples(self, data_file, split): |
|
with open(data_file, encoding="utf-8") as f: |
|
for line in f: |
|
row = json.loads(line) |
|
|
|
if self.config.name == "multirc": |
|
paragraph = row["passage"] |
|
for question in paragraph["questions"]: |
|
for answer in question["answers"]: |
|
label = answer.get("label") |
|
key = "%s_%s_%s" % (row["idx"], question["idx"], answer["idx"]) |
|
yield key, { |
|
"paragraph": paragraph["text"], |
|
"question": question["question"], |
|
"answer": answer["text"], |
|
"label": -1 if label is None else _cast_label(bool(label)), |
|
"idx": {"paragraph": row["idx"], "question": question["idx"], "answer": answer["idx"]}, |
|
} |
|
elif self.config.name == "record": |
|
passage = row["passage"] |
|
entity_texts, entity_spans = _get_record_entities(passage) |
|
for qa in row["qas"]: |
|
yield qa["idx"], { |
|
"passage": passage["text"], |
|
"query": qa["query"], |
|
"entities": entity_texts, |
|
"entity_spans": entity_spans, |
|
"answers": _get_record_answers(qa), |
|
"idx": {"passage": row["idx"], "query": qa["idx"]}, |
|
} |
|
else: |
|
if self.config.name.startswith("wsc"): |
|
row.update(row["target"]) |
|
example = {feature: row[feature] for feature in self.config.features} |
|
if self.config.name == "wsc.fixed": |
|
example = _fix_wst(example) |
|
example["idx"] = row["idx"] |
|
|
|
if "label" in row: |
|
if self.config.name == "copa": |
|
example["label"] = "choice2" if row["label"] else "choice1" |
|
else: |
|
example["label"] = _cast_label(row["label"]) |
|
else: |
|
assert split == datasets.Split.TEST, row |
|
example["label"] = -1 |
|
yield example["idx"], example |
|
|
|
|
|
def _fix_wst(ex): |
|
"""Fixes most cases where spans are not actually substrings of text.""" |
|
|
|
def _fix_span_text(k): |
|
"""Fixes a single span.""" |
|
text = ex[k + "_text"] |
|
index = ex[k + "_index"] |
|
|
|
if text in ex["text"]: |
|
return |
|
|
|
if text in ("Kamenev and Zinoviev", "Kamenev, Zinoviev, and Stalin"): |
|
|
|
|
|
return |
|
|
|
if "theyscold" in text: |
|
ex["text"].replace("theyscold", "they scold") |
|
ex["span2_index"] = 10 |
|
|
|
first_word = ex["text"].split()[index] |
|
if first_word[0].islower(): |
|
text = text[0].lower() + text[1:] |
|
else: |
|
text = text[0].upper() + text[1:] |
|
|
|
text = text.rstrip(".") |
|
|
|
text = text.replace("\n", " ") |
|
ex[k + "_text"] = text |
|
assert ex[k + "_text"] in ex["text"], ex |
|
|
|
_fix_span_text("span1") |
|
_fix_span_text("span2") |
|
return ex |
|
|
|
|
|
def _cast_label(label): |
|
"""Converts the label into the appropriate string version.""" |
|
if isinstance(label, str): |
|
return label |
|
elif isinstance(label, bool): |
|
return "True" if label else "False" |
|
elif isinstance(label, int): |
|
assert label in (0, 1) |
|
return str(label) |
|
else: |
|
raise ValueError("Invalid label format.") |
|
|
|
|
|
def _get_record_entities(passage): |
|
"""Returns the unique set of entities.""" |
|
text = passage["text"] |
|
entity_spans = list() |
|
for entity in passage["entities"]: |
|
entity_text = text[entity["start"] : entity["end"] + 1] |
|
entity_spans.append({"text": entity_text, "start": entity["start"], "end": entity["end"] + 1}) |
|
entity_spans = sorted(entity_spans, key=lambda e: e["start"]) |
|
entity_texts = set(e["text"] for e in entity_spans) |
|
return entity_texts, entity_spans |
|
|
|
|
|
def _get_record_answers(qa): |
|
"""Returns the unique set of answers.""" |
|
if "answers" not in qa: |
|
return [] |
|
answers = set() |
|
for answer in qa["answers"]: |
|
answers.add(answer["text"]) |
|
return sorted(answers) |
|
|
|
|
|
def _get_task_name_from_data_url(data_url): |
|
return data_url.split("/")[-1].split(".")[0] |
|
|