scrolls-quality-mcq / README.md
rbiswasfc's picture
Update README.md
d994a40 verified
|
raw
history blame
3.4 kB
---
license: mit
dataset_info:
features:
- name: id
dtype: string
- name: question
dtype: string
- name: context
dtype: string
- name: A
dtype: string
- name: B
dtype: string
- name: C
dtype: string
- name: D
dtype: string
- name: label
dtype: int64
splits:
- name: train
num_bytes: 63759933.69322235
num_examples: 2517
- name: test
num_bytes: 52057383.0
num_examples: 2086
download_size: 19849080
dataset_size: 115817316.69322234
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
---
This dataset is derived from `tau/scrolls` [dataset](tau/scrolls) by running the following script:
```python
import re
from datasets import load_dataset
quality_dataset = load_dataset("tau/scrolls", "quality")
def parse_example(example):
text = example["input"]
options = dict(re.findall(r"\((A|B|C|D)\) ([^\n]+)", text))
question_part, context = re.split(r"\(D\) [^\n]+\n", text, maxsplit=1)
question = re.sub(r"\([A-D]\) [^\n]+\n?", "", question_part).strip()
result = {"question": question, "context": context.strip(), **options}
if not all(key in result for key in ["A", "B", "C", "D"]):
raise ValueError("One or more options (A, B, C, D) are missing!")
# get label
label = -1
answer = example["output"]
if answer is None:
answer = ""
for idx, option in enumerate([options["A"], options["B"], options["C"], options["D"]]):
if answer.strip() == option.strip():
label = idx
result["label"] = label
return result
quality_dataset = quality_dataset.map(parse_example)
quality_dataset = quality_dataset.filter(lambda x: x["label"] >= 0)
train_ds = quality_dataset["train"].remove_columns(["pid", "input", "output"])
test_ds = quality_dataset["validation"].remove_columns(["pid", "input", "output"])
```
Specifically, only `quality` subset is kept and processed into MCQ format. The `test` split from original dataset is removed since it doesn't have ground truth labels.
Instead, validation split is assigned as test.
Number of examples in train: ~2.5k
Number of examples in test: ~2.1k
This dataset can be used to test performance of a model focusing on long contexts.
Input Tokens as per [llama2](bclavie/bert24_32k_tok_llama2) tokenizer: Mean -> 7.4k, SD: 2.3k, Max -> 11.6k
---
Relevant sections from the [SCROLLS: Standardized CompaRison Over Long Language Sequences paper](https://arxiv.org/pdf/2201.03533)
```
QuALITY (Pang et al., 2021): A multiplechoice question answering dataset over stories
and articles sourced from Project Gutenberg,10 the
Open American National Corpus (Fillmore et al.,
1998; Ide and Suderman, 2004), and more. Experienced writers wrote questions and distractors, and
were incentivized to write answerable, unambiguous questions such that in order to correctly answer
them, human annotators must read large portions
of the given document. To measure the difficulty
of their questions, Pang et al. conducted a speed
validation process, where another set of annotators
were asked to answer questions given only a short
period of time to skim through the document. As
a result, 50% of the questions in QuALITY are
labeled as hard, i.e. the majority of the annotators in the speed validation setting chose the wrong
answer.
```