Datasets:
Tasks:
Question Answering
Modalities:
Text
Formats:
parquet
Sub-tasks:
open-domain-qa
Languages:
English
Size:
100K - 1M
ArXiv:
License:
parquet-converter
commited on
Commit
•
82e6946
1
Parent(s):
d91c63f
Update parquet files
Browse files- .gitattributes +0 -27
- README.md +0 -473
- answer_selection_analysis/selqa-test.parquet +3 -0
- answer_selection_analysis/selqa-train.parquet +3 -0
- answer_selection_analysis/selqa-validation.parquet +3 -0
- answer_selection_experiments/selqa-test.parquet +3 -0
- answer_selection_experiments/selqa-train.parquet +3 -0
- answer_selection_experiments/selqa-validation.parquet +3 -0
- answer_triggering_analysis/selqa-test.parquet +3 -0
- answer_triggering_analysis/selqa-train.parquet +3 -0
- answer_triggering_analysis/selqa-validation.parquet +3 -0
- answer_triggering_experiments/selqa-test.parquet +3 -0
- answer_triggering_experiments/selqa-train.parquet +3 -0
- answer_triggering_experiments/selqa-validation.parquet +3 -0
- dataset_infos.json +0 -1
- selqa.py +0 -300
.gitattributes
DELETED
@@ -1,27 +0,0 @@
|
|
1 |
-
*.7z filter=lfs diff=lfs merge=lfs -text
|
2 |
-
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
-
*.bin filter=lfs diff=lfs merge=lfs -text
|
4 |
-
*.bin.* filter=lfs diff=lfs merge=lfs -text
|
5 |
-
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
6 |
-
*.ftz filter=lfs diff=lfs merge=lfs -text
|
7 |
-
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
-
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
-
*.joblib filter=lfs diff=lfs merge=lfs -text
|
10 |
-
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
11 |
-
*.model filter=lfs diff=lfs merge=lfs -text
|
12 |
-
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
13 |
-
*.onnx filter=lfs diff=lfs merge=lfs -text
|
14 |
-
*.ot filter=lfs diff=lfs merge=lfs -text
|
15 |
-
*.parquet filter=lfs diff=lfs merge=lfs -text
|
16 |
-
*.pb filter=lfs diff=lfs merge=lfs -text
|
17 |
-
*.pt filter=lfs diff=lfs merge=lfs -text
|
18 |
-
*.pth filter=lfs diff=lfs merge=lfs -text
|
19 |
-
*.rar filter=lfs diff=lfs merge=lfs -text
|
20 |
-
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
21 |
-
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
22 |
-
*.tflite filter=lfs diff=lfs merge=lfs -text
|
23 |
-
*.tgz filter=lfs diff=lfs merge=lfs -text
|
24 |
-
*.xz filter=lfs diff=lfs merge=lfs -text
|
25 |
-
*.zip filter=lfs diff=lfs merge=lfs -text
|
26 |
-
*.zstandard filter=lfs diff=lfs merge=lfs -text
|
27 |
-
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
README.md
DELETED
@@ -1,473 +0,0 @@
|
|
1 |
-
---
|
2 |
-
annotations_creators:
|
3 |
-
- crowdsourced
|
4 |
-
language_creators:
|
5 |
-
- found
|
6 |
-
language:
|
7 |
-
- en
|
8 |
-
license:
|
9 |
-
- apache-2.0
|
10 |
-
multilinguality:
|
11 |
-
- monolingual
|
12 |
-
size_categories:
|
13 |
-
- 1K<n<10K
|
14 |
-
source_datasets:
|
15 |
-
- original
|
16 |
-
task_categories:
|
17 |
-
- question-answering
|
18 |
-
task_ids:
|
19 |
-
- open-domain-qa
|
20 |
-
paperswithcode_id: selqa
|
21 |
-
pretty_name: SelQA
|
22 |
-
dataset_info:
|
23 |
-
- config_name: answer_selection_analysis
|
24 |
-
features:
|
25 |
-
- name: section
|
26 |
-
dtype: string
|
27 |
-
- name: question
|
28 |
-
dtype: string
|
29 |
-
- name: article
|
30 |
-
dtype: string
|
31 |
-
- name: is_paraphrase
|
32 |
-
dtype: bool
|
33 |
-
- name: topic
|
34 |
-
dtype:
|
35 |
-
class_label:
|
36 |
-
names:
|
37 |
-
0: MUSIC
|
38 |
-
1: TV
|
39 |
-
2: TRAVEL
|
40 |
-
3: ART
|
41 |
-
4: SPORT
|
42 |
-
5: COUNTRY
|
43 |
-
6: MOVIES
|
44 |
-
7: HISTORICAL EVENTS
|
45 |
-
8: SCIENCE
|
46 |
-
9: FOOD
|
47 |
-
- name: answers
|
48 |
-
sequence: int32
|
49 |
-
- name: candidates
|
50 |
-
sequence: string
|
51 |
-
- name: q_types
|
52 |
-
sequence:
|
53 |
-
class_label:
|
54 |
-
names:
|
55 |
-
0: what
|
56 |
-
1: why
|
57 |
-
2: when
|
58 |
-
3: who
|
59 |
-
4: where
|
60 |
-
5: how
|
61 |
-
6: ''
|
62 |
-
splits:
|
63 |
-
- name: train
|
64 |
-
num_bytes: 9676758
|
65 |
-
num_examples: 5529
|
66 |
-
- name: test
|
67 |
-
num_bytes: 2798537
|
68 |
-
num_examples: 1590
|
69 |
-
- name: validation
|
70 |
-
num_bytes: 1378407
|
71 |
-
num_examples: 785
|
72 |
-
download_size: 14773444
|
73 |
-
dataset_size: 13853702
|
74 |
-
- config_name: answer_selection_experiments
|
75 |
-
features:
|
76 |
-
- name: question
|
77 |
-
dtype: string
|
78 |
-
- name: candidate
|
79 |
-
dtype: string
|
80 |
-
- name: label
|
81 |
-
dtype:
|
82 |
-
class_label:
|
83 |
-
names:
|
84 |
-
0: '0'
|
85 |
-
1: '1'
|
86 |
-
splits:
|
87 |
-
- name: train
|
88 |
-
num_bytes: 13782826
|
89 |
-
num_examples: 66438
|
90 |
-
- name: test
|
91 |
-
num_bytes: 4008077
|
92 |
-
num_examples: 19435
|
93 |
-
- name: validation
|
94 |
-
num_bytes: 1954877
|
95 |
-
num_examples: 9377
|
96 |
-
download_size: 18602700
|
97 |
-
dataset_size: 19745780
|
98 |
-
- config_name: answer_triggering_analysis
|
99 |
-
features:
|
100 |
-
- name: section
|
101 |
-
dtype: string
|
102 |
-
- name: question
|
103 |
-
dtype: string
|
104 |
-
- name: article
|
105 |
-
dtype: string
|
106 |
-
- name: is_paraphrase
|
107 |
-
dtype: bool
|
108 |
-
- name: topic
|
109 |
-
dtype:
|
110 |
-
class_label:
|
111 |
-
names:
|
112 |
-
0: MUSIC
|
113 |
-
1: TV
|
114 |
-
2: TRAVEL
|
115 |
-
3: ART
|
116 |
-
4: SPORT
|
117 |
-
5: COUNTRY
|
118 |
-
6: MOVIES
|
119 |
-
7: HISTORICAL EVENTS
|
120 |
-
8: SCIENCE
|
121 |
-
9: FOOD
|
122 |
-
- name: q_types
|
123 |
-
sequence:
|
124 |
-
class_label:
|
125 |
-
names:
|
126 |
-
0: what
|
127 |
-
1: why
|
128 |
-
2: when
|
129 |
-
3: who
|
130 |
-
4: where
|
131 |
-
5: how
|
132 |
-
6: ''
|
133 |
-
- name: candidate_list
|
134 |
-
sequence:
|
135 |
-
- name: article
|
136 |
-
dtype: string
|
137 |
-
- name: section
|
138 |
-
dtype: string
|
139 |
-
- name: candidates
|
140 |
-
sequence: string
|
141 |
-
- name: answers
|
142 |
-
sequence: int32
|
143 |
-
splits:
|
144 |
-
- name: train
|
145 |
-
num_bytes: 30176650
|
146 |
-
num_examples: 5529
|
147 |
-
- name: test
|
148 |
-
num_bytes: 8766787
|
149 |
-
num_examples: 1590
|
150 |
-
- name: validation
|
151 |
-
num_bytes: 4270904
|
152 |
-
num_examples: 785
|
153 |
-
download_size: 46149676
|
154 |
-
dataset_size: 43214341
|
155 |
-
- config_name: answer_triggering_experiments
|
156 |
-
features:
|
157 |
-
- name: question
|
158 |
-
dtype: string
|
159 |
-
- name: candidate
|
160 |
-
dtype: string
|
161 |
-
- name: label
|
162 |
-
dtype:
|
163 |
-
class_label:
|
164 |
-
names:
|
165 |
-
0: '0'
|
166 |
-
1: '1'
|
167 |
-
splits:
|
168 |
-
- name: train
|
169 |
-
num_bytes: 42956518
|
170 |
-
num_examples: 205075
|
171 |
-
- name: test
|
172 |
-
num_bytes: 12504961
|
173 |
-
num_examples: 59845
|
174 |
-
- name: validation
|
175 |
-
num_bytes: 6055616
|
176 |
-
num_examples: 28798
|
177 |
-
download_size: 57992239
|
178 |
-
dataset_size: 61517095
|
179 |
-
---
|
180 |
-
|
181 |
-
# Dataset Card for SelQA
|
182 |
-
|
183 |
-
## Table of Contents
|
184 |
-
- [Dataset Description](#dataset-description)
|
185 |
-
- [Dataset Summary](#dataset-summary)
|
186 |
-
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
|
187 |
-
- [Languages](#languages)
|
188 |
-
- [Dataset Structure](#dataset-structure)
|
189 |
-
- [Data Instances](#data-instances)
|
190 |
-
- [Data Fields](#data-fields)
|
191 |
-
- [Data Splits](#data-splits)
|
192 |
-
- [Dataset Creation](#dataset-creation)
|
193 |
-
- [Curation Rationale](#curation-rationale)
|
194 |
-
- [Source Data](#source-data)
|
195 |
-
- [Annotations](#annotations)
|
196 |
-
- [Personal and Sensitive Information](#personal-and-sensitive-information)
|
197 |
-
- [Considerations for Using the Data](#considerations-for-using-the-data)
|
198 |
-
- [Social Impact of Dataset](#social-impact-of-dataset)
|
199 |
-
- [Discussion of Biases](#discussion-of-biases)
|
200 |
-
- [Other Known Limitations](#other-known-limitations)
|
201 |
-
- [Additional Information](#additional-information)
|
202 |
-
- [Dataset Curators](#dataset-curators)
|
203 |
-
- [Licensing Information](#licensing-information)
|
204 |
-
- [Citation Information](#citation-information)
|
205 |
-
- [Contributions](#contributions)
|
206 |
-
|
207 |
-
## Dataset Description
|
208 |
-
|
209 |
-
- **Homepage:** https://github.com/emorynlp/selqa
|
210 |
-
- **Repository:** https://github.com/emorynlp/selqa
|
211 |
-
- **Paper:** https://arxiv.org/abs/1606.00851
|
212 |
-
- **Leaderboard:** [Needs More Information]
|
213 |
-
- **Point of Contact:** Tomasz Jurczyk <http://tomaszjurczyk.com/>, Jinho D. Choi <http://www.mathcs.emory.edu/~choi/home.html>
|
214 |
-
|
215 |
-
### Dataset Summary
|
216 |
-
|
217 |
-
SelQA: A New Benchmark for Selection-Based Question Answering
|
218 |
-
|
219 |
-
|
220 |
-
### Supported Tasks and Leaderboards
|
221 |
-
|
222 |
-
Question Answering
|
223 |
-
|
224 |
-
### Languages
|
225 |
-
|
226 |
-
English
|
227 |
-
|
228 |
-
## Dataset Structure
|
229 |
-
|
230 |
-
### Data Instances
|
231 |
-
|
232 |
-
An example from the `answer selection` set:
|
233 |
-
```
|
234 |
-
{
|
235 |
-
"section": "Museums",
|
236 |
-
"question": "Where are Rockefeller Museum and LA Mayer Institute for Islamic Art?",
|
237 |
-
"article": "Israel",
|
238 |
-
"is_paraphrase": true,
|
239 |
-
"topic": "COUNTRY",
|
240 |
-
"answers": [
|
241 |
-
5
|
242 |
-
],
|
243 |
-
"candidates": [
|
244 |
-
"The Israel Museum in Jerusalem is one of Israel's most important cultural institutions and houses the Dead Sea scrolls, along with an extensive collection of Judaica and European art.",
|
245 |
-
"Israel's national Holocaust museum, Yad Vashem, is the world central archive of Holocaust-related information.",
|
246 |
-
"Beth Hatefutsoth (the Diaspora Museum), on the campus of Tel Aviv University, is an interactive museum devoted to the history of Jewish communities around the world.",
|
247 |
-
"Apart from the major museums in large cities, there are high-quality artspaces in many towns and \"kibbutzim\".",
|
248 |
-
"\"Mishkan Le'Omanut\" on Kibbutz Ein Harod Meuhad is the largest art museum in the north of the country.",
|
249 |
-
"Several Israeli museums are devoted to Islamic culture, including the Rockefeller Museum and the L. A. Mayer Institute for Islamic Art, both in Jerusalem.",
|
250 |
-
"The Rockefeller specializes in archaeological remains from the Ottoman and other periods of Middle East history.",
|
251 |
-
"It is also the home of the first hominid fossil skull found in Western Asia called Galilee Man.",
|
252 |
-
"A cast of the skull is on display at the Israel Museum."
|
253 |
-
],
|
254 |
-
"q_types": [
|
255 |
-
"where"
|
256 |
-
]
|
257 |
-
}
|
258 |
-
```
|
259 |
-
|
260 |
-
An example from the `answer triggering` set:
|
261 |
-
```
|
262 |
-
{
|
263 |
-
"section": "Museums",
|
264 |
-
"question": "Where are Rockefeller Museum and LA Mayer Institute for Islamic Art?",
|
265 |
-
"article": "Israel",
|
266 |
-
"is_paraphrase": true,
|
267 |
-
"topic": "COUNTRY",
|
268 |
-
"candidate_list": [
|
269 |
-
{
|
270 |
-
"article": "List of places in Jerusalem",
|
271 |
-
"section": "List_of_places_in_Jerusalem-Museums",
|
272 |
-
"answers": [],
|
273 |
-
"candidates": [
|
274 |
-
" Israel Museum *Shrine of the Book *Rockefeller Museum of Archeology Bible Lands Museum Jerusalem Yad Vashem Holocaust Museum L.A. Mayer Institute for Islamic Art Bloomfield Science Museum Natural History Museum Museum of Italian Jewish Art Ticho House Tower of David Jerusalem Tax Museum Herzl Museum Siebenberg House Museums.",
|
275 |
-
"Museum on the Seam "
|
276 |
-
]
|
277 |
-
},
|
278 |
-
{
|
279 |
-
"article": "Israel",
|
280 |
-
"section": "Israel-Museums",
|
281 |
-
"answers": [
|
282 |
-
5
|
283 |
-
],
|
284 |
-
"candidates": [
|
285 |
-
"The Israel Museum in Jerusalem is one of Israel's most important cultural institutions and houses the Dead Sea scrolls, along with an extensive collection of Judaica and European art.",
|
286 |
-
"Israel's national Holocaust museum, Yad Vashem, is the world central archive of Holocaust-related information.",
|
287 |
-
"Beth Hatefutsoth (the Diaspora Museum), on the campus of Tel Aviv University, is an interactive museum devoted to the history of Jewish communities around the world.",
|
288 |
-
"Apart from the major museums in large cities, there are high-quality artspaces in many towns and \"kibbutzim\".",
|
289 |
-
"\"Mishkan Le'Omanut\" on Kibbutz Ein Harod Meuhad is the largest art museum in the north of the country.",
|
290 |
-
"Several Israeli museums are devoted to Islamic culture, including the Rockefeller Museum and the L. A. Mayer Institute for Islamic Art, both in Jerusalem.",
|
291 |
-
"The Rockefeller specializes in archaeological remains from the Ottoman and other periods of Middle East history.",
|
292 |
-
"It is also the home of the first hominid fossil skull found in Western Asia called Galilee Man.",
|
293 |
-
"A cast of the skull is on display at the Israel Museum."
|
294 |
-
]
|
295 |
-
},
|
296 |
-
{
|
297 |
-
"article": "L. A. Mayer Institute for Islamic Art",
|
298 |
-
"section": "L._A._Mayer_Institute_for_Islamic_Art-Abstract",
|
299 |
-
"answers": [],
|
300 |
-
"candidates": [
|
301 |
-
"The L.A. Mayer Institute for Islamic Art (Hebrew: \u05de\u05d5\u05d6\u05d9\u05d0\u05d5\u05df \u05dc.",
|
302 |
-
"\u05d0.",
|
303 |
-
"\u05de\u05d0\u05d9\u05e8 \u05dc\u05d0\u05de\u05e0\u05d5\u05ea \u05d4\u05d0\u05e1\u05dc\u05d0\u05dd) is a museum in Jerusalem, Israel, established in 1974.",
|
304 |
-
"It is located in Katamon, down the road from the Jerusalem Theater.",
|
305 |
-
"The museum houses Islamic pottery, textiles, jewelry, ceremonial objects and other Islamic cultural artifacts.",
|
306 |
-
"It is not to be confused with the Islamic Museum, Jerusalem. "
|
307 |
-
]
|
308 |
-
},
|
309 |
-
{
|
310 |
-
"article": "Islamic Museum, Jerusalem",
|
311 |
-
"section": "Islamic_Museum,_Jerusalem-Abstract",
|
312 |
-
"answers": [],
|
313 |
-
"candidates": [
|
314 |
-
"The Islamic Museum is a museum on the Temple Mount in the Old City section of Jerusalem.",
|
315 |
-
"On display are exhibits from ten periods of Islamic history encompassing several Muslim regions.",
|
316 |
-
"The museum is located adjacent to al-Aqsa Mosque.",
|
317 |
-
"It is not to be confused with the L. A. Mayer Institute for Islamic Art, also a museum in Jerusalem. "
|
318 |
-
]
|
319 |
-
},
|
320 |
-
{
|
321 |
-
"article": "L. A. Mayer Institute for Islamic Art",
|
322 |
-
"section": "L._A._Mayer_Institute_for_Islamic_Art-Contemporary_Arab_art",
|
323 |
-
"answers": [],
|
324 |
-
"candidates": [
|
325 |
-
"In 2008, a group exhibit of contemporary Arab art opened at L.A. Mayer Institute, the first show of local Arab art in an Israeli museum and the first to be mounted by an Arab curator.",
|
326 |
-
"Thirteen Arab artists participated in the show. "
|
327 |
-
]
|
328 |
-
}
|
329 |
-
],
|
330 |
-
"q_types": [
|
331 |
-
"where"
|
332 |
-
]
|
333 |
-
}
|
334 |
-
```
|
335 |
-
|
336 |
-
An example from any of the `experiments` data:
|
337 |
-
```
|
338 |
-
Where are Rockefeller Museum and LA Mayer Institute for Islamic Art ? The Israel Museum in Jerusalem is one of Israel 's most important cultural institutions and houses the Dead Sea scrolls , along with an extensive collection of Judaica and European art . 0
|
339 |
-
Where are Rockefeller Museum and LA Mayer Institute for Islamic Art ? Israel 's national Holocaust museum , Yad Vashem , is the world central archive of Holocaust - related information . 0
|
340 |
-
Where are Rockefeller Museum and LA Mayer Institute for Islamic Art ? Beth Hatefutsoth ( the Diaspora Museum ) , on the campus of Tel Aviv University , is an interactive museum devoted to the history of Jewish communities around the world . 0
|
341 |
-
Where are Rockefeller Museum and LA Mayer Institute for Islamic Art ? Apart from the major museums in large cities , there are high - quality artspaces in many towns and " kibbutzim " . 0
|
342 |
-
Where are Rockefeller Museum and LA Mayer Institute for Islamic Art ? " Mishkan Le'Omanut " on Kibbutz Ein Harod Meuhad is the largest art museum in the north of the country . 0
|
343 |
-
Where are Rockefeller Museum and LA Mayer Institute for Islamic Art ? Several Israeli museums are devoted to Islamic culture , including the Rockefeller Museum and the L. A. Mayer Institute for Islamic Art , both in Jerusalem . 1
|
344 |
-
Where are Rockefeller Museum and LA Mayer Institute for Islamic Art ? The Rockefeller specializes in archaeological remains from the Ottoman and other periods of Middle East history . 0
|
345 |
-
Where are Rockefeller Museum and LA Mayer Institute for Islamic Art ? It is also the home of the first hominid fossil skull found in Western Asia called Galilee Man . 0
|
346 |
-
Where are Rockefeller Museum and LA Mayer Institute for Islamic Art ? A cast of the skull is on display at the Israel Museum . 0
|
347 |
-
```
|
348 |
-
|
349 |
-
### Data Fields
|
350 |
-
|
351 |
-
#### Answer Selection
|
352 |
-
##### Data for Analysis
|
353 |
-
|
354 |
-
for analysis, the columns are:
|
355 |
-
|
356 |
-
* `question`: the question.
|
357 |
-
* `article`: the Wikipedia article related to this question.
|
358 |
-
* `section`: the section in the Wikipedia article related to this question.
|
359 |
-
* `topic`: the topic of this question, where the topics are *MUSIC*, *TV*, *TRAVEL*, *ART*, *SPORT*, *COUNTRY*, *MOVIES*, *HISTORICAL EVENTS*, *SCIENCE*, *FOOD*.
|
360 |
-
* `q_types`: the list of question types, where the types are *what*, *why*, *when*, *who*, *where*, and *how*. If empty, none of the those types are recognized in this question.
|
361 |
-
* `is_paraphrase`: *True* if this question is a paragraph of some other question in this dataset; otherwise, *False*.
|
362 |
-
* `candidates`: the list of sentences in the related section.
|
363 |
-
* `answers`: the list of candidate indices containing the answer context of this question.
|
364 |
-
|
365 |
-
##### Data for Experiments
|
366 |
-
|
367 |
-
for experiments, each column gives:
|
368 |
-
|
369 |
-
* `0`: a question where all tokens are separated.
|
370 |
-
* `1`: a candidate of the question where all tokens are separated.
|
371 |
-
* `2`: the label where `0` implies no answer to the question is found in this candidate and `1` implies the answer is found.
|
372 |
-
|
373 |
-
#### Answer Triggering
|
374 |
-
##### Data for Analysis
|
375 |
-
|
376 |
-
for analysis, the columns are:
|
377 |
-
|
378 |
-
* `question`: the question.
|
379 |
-
* `article`: the Wikipedia article related to this question.
|
380 |
-
* `section`: the section in the Wikipedia article related to this question.
|
381 |
-
* `topic`: the topic of this question, where the topics are *MUSIC*, *TV*, *TRAVEL*, *ART*, *SPORT*, *COUNTRY*, *MOVIES*, *HISTORICAL EVENTS*, *SCIENCE*, *FOOD*.
|
382 |
-
* `q_types`: the list of question types, where the types are *what*, *why*, *when*, *who*, *where*, and *how*. If empty, none of the those types are recognized in this question.
|
383 |
-
* `is_paraphrase`: *True* if this question is a paragraph of some other question in this dataset; otherwise, *False*.
|
384 |
-
* `candidate_list`: the list of 5 candidate sections:
|
385 |
-
* `article`: the title of the candidate article.
|
386 |
-
* `section`: the section in the candidate article.
|
387 |
-
* `candidates`: the list of sentences in this candidate section.
|
388 |
-
* `answers`: the list of candidate indices containing the answer context of this question (can be empty).
|
389 |
-
|
390 |
-
##### Data for Experiments
|
391 |
-
|
392 |
-
for experiments, each column gives:
|
393 |
-
|
394 |
-
* `0`: a question where all tokens are separated.
|
395 |
-
* `1`: a candidate of the question where all tokens are separated.
|
396 |
-
* `2`: the label where `0` implies no answer to the question is found in this candidate and `1` implies the answer is found.
|
397 |
-
|
398 |
-
### Data Splits
|
399 |
-
|
400 |
-
| |Train| Valid| Test|
|
401 |
-
| --- | --- | --- | --- |
|
402 |
-
| Answer Selection | 5529 | 785 | 1590 |
|
403 |
-
| Answer Triggering | 27645 | 3925 | 7950 |
|
404 |
-
|
405 |
-
## Dataset Creation
|
406 |
-
|
407 |
-
### Curation Rationale
|
408 |
-
|
409 |
-
To encourage research and provide an initial benchmark for selection based question answering and answer triggering tasks
|
410 |
-
|
411 |
-
### Source Data
|
412 |
-
|
413 |
-
#### Initial Data Collection and Normalization
|
414 |
-
|
415 |
-
[Needs More Information]
|
416 |
-
|
417 |
-
#### Who are the source language producers?
|
418 |
-
|
419 |
-
[Needs More Information]
|
420 |
-
|
421 |
-
### Annotations
|
422 |
-
|
423 |
-
#### Annotation process
|
424 |
-
|
425 |
-
Crowdsourced
|
426 |
-
|
427 |
-
#### Who are the annotators?
|
428 |
-
|
429 |
-
[Needs More Information]
|
430 |
-
|
431 |
-
### Personal and Sensitive Information
|
432 |
-
|
433 |
-
[Needs More Information]
|
434 |
-
|
435 |
-
## Considerations for Using the Data
|
436 |
-
|
437 |
-
### Social Impact of Dataset
|
438 |
-
|
439 |
-
The purpose of this dataset is to help develop better selection-based question answering systems.
|
440 |
-
|
441 |
-
### Discussion of Biases
|
442 |
-
|
443 |
-
[Needs More Information]
|
444 |
-
|
445 |
-
### Other Known Limitations
|
446 |
-
|
447 |
-
[Needs More Information]
|
448 |
-
|
449 |
-
## Additional Information
|
450 |
-
|
451 |
-
### Dataset Curators
|
452 |
-
|
453 |
-
[Needs More Information]
|
454 |
-
|
455 |
-
### Licensing Information
|
456 |
-
|
457 |
-
Apache License 2.0
|
458 |
-
|
459 |
-
### Citation Information
|
460 |
-
@InProceedings{7814688,
|
461 |
-
author={T. {Jurczyk} and M. {Zhai} and J. D. {Choi}},
|
462 |
-
booktitle={2016 IEEE 28th International Conference on Tools with Artificial Intelligence (ICTAI)},
|
463 |
-
title={SelQA: A New Benchmark for Selection-Based Question Answering},
|
464 |
-
year={2016},
|
465 |
-
volume={},
|
466 |
-
number={},
|
467 |
-
pages={820-827},
|
468 |
-
doi={10.1109/ICTAI.2016.0128}
|
469 |
-
}
|
470 |
-
|
471 |
-
### Contributions
|
472 |
-
|
473 |
-
Thanks to [@Bharat123rox](https://github.com/Bharat123rox) for adding this dataset.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
answer_selection_analysis/selqa-test.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:303313c2f7d230b437e15f9eec318d3142556bf2d5fea66b051972969f3b3c36
|
3 |
+
size 1625035
|
answer_selection_analysis/selqa-train.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:68df038218905dc98f9f4778a674317d3d6981fa964b4cac133772956bb6b3c9
|
3 |
+
size 5575769
|
answer_selection_analysis/selqa-validation.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b3000b1e4624a52e8f121cf4a2a8141a024c6cacabfd2dd4ded7214996c1ff01
|
3 |
+
size 781513
|
answer_selection_experiments/selqa-test.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5a4269c3975b8d0f5828c6281c7a609cdbe62ff38a738eb61b5852af038da332
|
3 |
+
size 1803902
|
answer_selection_experiments/selqa-train.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e072414e92bef8a07ec2db04f2cd2d522f23ca057481a6bc04924296413b54c4
|
3 |
+
size 6198250
|
answer_selection_experiments/selqa-validation.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dd37d9b351e31030abca2b0ac036317d581b54d1931396b1924deef370b1441e
|
3 |
+
size 887819
|
answer_triggering_analysis/selqa-test.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4ceab302edb5e4ddc381c0a103aa7a1cb97aaa6c66dbcaf5313205f15cfdc5fb
|
3 |
+
size 5311393
|
answer_triggering_analysis/selqa-train.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:55e631ec310e25afd81e93a2de34625e128357c27dffa1adb5565f33b6b741b3
|
3 |
+
size 18176470
|
answer_triggering_analysis/selqa-validation.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8ded1d427e96883bf87dbbd0d9a8847f807b3686aa3178cd08be6ae77ed87005
|
3 |
+
size 2562055
|
answer_triggering_experiments/selqa-test.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fa3d825df8729c4a63c220bee718b78a43cbd180204a595c42e1f88d5930d1e4
|
3 |
+
size 5165742
|
answer_triggering_experiments/selqa-train.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d4428aa2e854f289ae2de777ca6c90b1f8023e0faf15e4a2b75d52b09eb38fe5
|
3 |
+
size 17708407
|
answer_triggering_experiments/selqa-validation.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d2d1ab407ab34bf9135eb093b5f02e625e4707d997cbddcba6bd46d9649fc68b
|
3 |
+
size 2494266
|
dataset_infos.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"answer_selection_analysis": {"description": "The SelQA dataset provides crowdsourced annotation for two selection-based question answer tasks, \nanswer sentence selection and answer triggering.\n", "citation": "@InProceedings{7814688,\n author={T. {Jurczyk} and M. {Zhai} and J. D. {Choi}},\n booktitle={2016 IEEE 28th International Conference on Tools with Artificial Intelligence (ICTAI)}, \n title={SelQA: A New Benchmark for Selection-Based Question Answering}, \n year={2016},\n volume={},\n number={},\n pages={820-827},\n doi={10.1109/ICTAI.2016.0128}\n}\n", "homepage": "", "license": "", "features": {"section": {"dtype": "string", "id": null, "_type": "Value"}, "question": {"dtype": "string", "id": null, "_type": "Value"}, "article": {"dtype": "string", "id": null, "_type": "Value"}, "is_paraphrase": {"dtype": "bool", "id": null, "_type": "Value"}, "topic": {"num_classes": 10, "names": ["MUSIC", "TV", "TRAVEL", "ART", "SPORT", "COUNTRY", "MOVIES", "HISTORICAL EVENTS", "SCIENCE", "FOOD"], "names_file": null, "id": null, "_type": "ClassLabel"}, "answers": {"feature": {"dtype": "int32", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "candidates": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "q_types": {"feature": {"num_classes": 7, "names": ["what", "why", "when", "who", "where", "how", ""], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "builder_name": "selqa", "config_name": "answer_selection_analysis", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 9676758, "num_examples": 5529, "dataset_name": "selqa"}, "test": {"name": "test", "num_bytes": 2798537, "num_examples": 1590, "dataset_name": "selqa"}, "validation": {"name": "validation", "num_bytes": 1378407, "num_examples": 785, "dataset_name": "selqa"}}, "download_checksums": {"https://raw.githubusercontent.com/emorynlp/selqa/master/ass/selqa-ass-train.json": {"num_bytes": 10320158, "checksum": "30622b7820bb2fa8e766d0ad3c7cf29dac658772cd763a9dabf81d9cab1fd534"}, "https://raw.githubusercontent.com/emorynlp/selqa/master/ass/selqa-ass-dev.json": {"num_bytes": 1470163, "checksum": "b4e6687e44a30b486e24d2b06aa3012ec07d61145f3521f35b7d49daae3e0ca4"}, "https://raw.githubusercontent.com/emorynlp/selqa/master/ass/selqa-ass-test.json": {"num_bytes": 2983123, "checksum": "ca1184d94cc9030883723fab76ef8180b3cf5fb142549a5648d22f59fe7c6fc6"}}, "download_size": 14773444, "post_processing_size": null, "dataset_size": 13853702, "size_in_bytes": 28627146}, "answer_selection_experiments": {"description": "The SelQA dataset provides crowdsourced annotation for two selection-based question answer tasks, \nanswer sentence selection and answer triggering.\n", "citation": "@InProceedings{7814688,\n author={T. {Jurczyk} and M. {Zhai} and J. D. {Choi}},\n booktitle={2016 IEEE 28th International Conference on Tools with Artificial Intelligence (ICTAI)}, \n title={SelQA: A New Benchmark for Selection-Based Question Answering}, \n year={2016},\n volume={},\n number={},\n pages={820-827},\n doi={10.1109/ICTAI.2016.0128}\n}\n", "homepage": "", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "candidate": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 2, "names": ["0", "1"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "builder_name": "selqa", "config_name": "answer_selection_experiments", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 13782826, "num_examples": 66438, "dataset_name": "selqa"}, "test": {"name": "test", "num_bytes": 4008077, "num_examples": 19435, "dataset_name": "selqa"}, "validation": {"name": "validation", "num_bytes": 1954877, "num_examples": 9377, "dataset_name": "selqa"}}, "download_checksums": {"https://raw.githubusercontent.com/emorynlp/selqa/master/ass/selqa-ass-train.tsv": {"num_bytes": 12985514, "checksum": "9f40017c0bf97f2f5816fba5ac18c7eafb847a9e351d85584afaecd1296010db"}, "https://raw.githubusercontent.com/emorynlp/selqa/master/ass/selqa-ass-dev.tsv": {"num_bytes": 1842345, "checksum": "0f0d73b379bb4efc6e678e36b122ea17c957998a1d002e3c480b3bc7854f77a9"}, "https://raw.githubusercontent.com/emorynlp/selqa/master/ass/selqa-ass-test.tsv": {"num_bytes": 3774841, "checksum": "4129ffa31237eb7f673baf6313bdd7d01658000c253b45e195d235493a435b91"}}, "download_size": 18602700, "post_processing_size": null, "dataset_size": 19745780, "size_in_bytes": 38348480}, "answer_triggering_analysis": {"description": "The SelQA dataset provides crowdsourced annotation for two selection-based question answer tasks, \nanswer sentence selection and answer triggering.\n", "citation": "@InProceedings{7814688,\n author={T. {Jurczyk} and M. {Zhai} and J. D. {Choi}},\n booktitle={2016 IEEE 28th International Conference on Tools with Artificial Intelligence (ICTAI)}, \n title={SelQA: A New Benchmark for Selection-Based Question Answering}, \n year={2016},\n volume={},\n number={},\n pages={820-827},\n doi={10.1109/ICTAI.2016.0128}\n}\n", "homepage": "", "license": "", "features": {"section": {"dtype": "string", "id": null, "_type": "Value"}, "question": {"dtype": "string", "id": null, "_type": "Value"}, "article": {"dtype": "string", "id": null, "_type": "Value"}, "is_paraphrase": {"dtype": "bool", "id": null, "_type": "Value"}, "topic": {"num_classes": 10, "names": ["MUSIC", "TV", "TRAVEL", "ART", "SPORT", "COUNTRY", "MOVIES", "HISTORICAL EVENTS", "SCIENCE", "FOOD"], "names_file": null, "id": null, "_type": "ClassLabel"}, "q_types": {"feature": {"num_classes": 7, "names": ["what", "why", "when", "who", "where", "how", ""], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}, "candidate_list": {"feature": {"article": {"dtype": "string", "id": null, "_type": "Value"}, "section": {"dtype": "string", "id": null, "_type": "Value"}, "candidates": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answers": {"feature": {"dtype": "int32", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "builder_name": "selqa", "config_name": "answer_triggering_analysis", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 30176650, "num_examples": 5529, "dataset_name": "selqa"}, "test": {"name": "test", "num_bytes": 8766787, "num_examples": 1590, "dataset_name": "selqa"}, "validation": {"name": "validation", "num_bytes": 4270904, "num_examples": 785, "dataset_name": "selqa"}}, "download_checksums": {"https://raw.githubusercontent.com/emorynlp/selqa/master/at/selqa-at-train.json": {"num_bytes": 32230643, "checksum": "6af1e82dbec94d2c87c0cd6463a0d7eba1dd746cbdc72f481697843c466f4952"}, "https://raw.githubusercontent.com/emorynlp/selqa/master/at/selqa-at-dev.json": {"num_bytes": 4562321, "checksum": "8cf266e9b8404e9ba1c062a1dbf43c79ae9bd2da929cb11351872c4f221815ac"}, "https://raw.githubusercontent.com/emorynlp/selqa/master/at/selqa-at-test.json": {"num_bytes": 9356712, "checksum": "38971e74506b74c808756fefb1816453eb1a3c3989f2feb77d864c93da468905"}}, "download_size": 46149676, "post_processing_size": null, "dataset_size": 43214341, "size_in_bytes": 89364017}, "answer_triggering_experiments": {"description": "The SelQA dataset provides crowdsourced annotation for two selection-based question answer tasks, \nanswer sentence selection and answer triggering.\n", "citation": "@InProceedings{7814688,\n author={T. {Jurczyk} and M. {Zhai} and J. D. {Choi}},\n booktitle={2016 IEEE 28th International Conference on Tools with Artificial Intelligence (ICTAI)}, \n title={SelQA: A New Benchmark for Selection-Based Question Answering}, \n year={2016},\n volume={},\n number={},\n pages={820-827},\n doi={10.1109/ICTAI.2016.0128}\n}\n", "homepage": "", "license": "", "features": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "candidate": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 2, "names": ["0", "1"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "builder_name": "selqa", "config_name": "answer_triggering_experiments", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 42956518, "num_examples": 205075, "dataset_name": "selqa"}, "test": {"name": "test", "num_bytes": 12504961, "num_examples": 59845, "dataset_name": "selqa"}, "validation": {"name": "validation", "num_bytes": 6055616, "num_examples": 28798, "dataset_name": "selqa"}}, "download_checksums": {"https://raw.githubusercontent.com/emorynlp/selqa/master/at/selqa-at-train.tsv": {"num_bytes": 40495450, "checksum": "9cf58039e30583187e7e93e19043dceb2540d72fc13eb4eb09fd8147b3022346"}, "https://raw.githubusercontent.com/emorynlp/selqa/master/at/selqa-at-dev.tsv": {"num_bytes": 5710016, "checksum": "76466b282ab62353e029af4292acb658c0659860c716c637c3e5f3faa9c693d1"}, "https://raw.githubusercontent.com/emorynlp/selqa/master/at/selqa-at-test.tsv": {"num_bytes": 11786773, "checksum": "4151fa580983f7d3903ea70e71d5d86f20abe75cb975b7d77434ea2e978fc132"}}, "download_size": 57992239, "post_processing_size": null, "dataset_size": 61517095, "size_in_bytes": 119509334}}
|
|
|
|
selqa.py
DELETED
@@ -1,300 +0,0 @@
|
|
1 |
-
# coding=utf-8
|
2 |
-
# Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
|
3 |
-
#
|
4 |
-
# Licensed under the Apache License, Version 2.0 (the "License");
|
5 |
-
# you may not use this file except in compliance with the License.
|
6 |
-
# You may obtain a copy of the License at
|
7 |
-
#
|
8 |
-
# http://www.apache.org/licenses/LICENSE-2.0
|
9 |
-
#
|
10 |
-
# Unless required by applicable law or agreed to in writing, software
|
11 |
-
# distributed under the License is distributed on an "AS IS" BASIS,
|
12 |
-
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
13 |
-
# See the License for the specific language governing permissions and
|
14 |
-
# limitations under the License.
|
15 |
-
"""SelQA: A New Benchmark for Selection-Based Question Answering"""
|
16 |
-
|
17 |
-
|
18 |
-
import csv
|
19 |
-
import json
|
20 |
-
|
21 |
-
import datasets
|
22 |
-
|
23 |
-
|
24 |
-
# TODO: Add BibTeX citation
|
25 |
-
# Find for instance the citation on arxiv or on the dataset repo/website
|
26 |
-
_CITATION = """\
|
27 |
-
@InProceedings{7814688,
|
28 |
-
author={T. {Jurczyk} and M. {Zhai} and J. D. {Choi}},
|
29 |
-
booktitle={2016 IEEE 28th International Conference on Tools with Artificial Intelligence (ICTAI)},
|
30 |
-
title={SelQA: A New Benchmark for Selection-Based Question Answering},
|
31 |
-
year={2016},
|
32 |
-
volume={},
|
33 |
-
number={},
|
34 |
-
pages={820-827},
|
35 |
-
doi={10.1109/ICTAI.2016.0128}
|
36 |
-
}
|
37 |
-
"""
|
38 |
-
|
39 |
-
# TODO: Add description of the dataset here
|
40 |
-
# You can copy an official description
|
41 |
-
_DESCRIPTION = """\
|
42 |
-
The SelQA dataset provides crowdsourced annotation for two selection-based question answer tasks,
|
43 |
-
answer sentence selection and answer triggering.
|
44 |
-
"""
|
45 |
-
|
46 |
-
# TODO: Add a link to an official homepage for the dataset here
|
47 |
-
_HOMEPAGE = ""
|
48 |
-
|
49 |
-
# TODO: Add the licence for the dataset here if you can find it
|
50 |
-
_LICENSE = ""
|
51 |
-
|
52 |
-
# TODO: Add link to the official dataset URLs here
|
53 |
-
# The HuggingFace dataset library don't host the datasets but only point to the original files
|
54 |
-
# This can be an arbitrary nested dict/list of URLs (see below in `_split_generators` method)
|
55 |
-
types = {
|
56 |
-
"answer_selection": "ass",
|
57 |
-
"answer_triggering": "at",
|
58 |
-
}
|
59 |
-
|
60 |
-
modes = {"analysis": "json", "experiments": "tsv"}
|
61 |
-
|
62 |
-
|
63 |
-
class SelqaConfig(datasets.BuilderConfig):
|
64 |
-
""" "BuilderConfig for SelQA Dataset"""
|
65 |
-
|
66 |
-
def __init__(self, mode, type_, **kwargs):
|
67 |
-
super(SelqaConfig, self).__init__(**kwargs)
|
68 |
-
self.mode = mode
|
69 |
-
self.type_ = type_
|
70 |
-
|
71 |
-
|
72 |
-
# TODO: Name of the dataset usually match the script name with CamelCase instead of snake_case
|
73 |
-
class Selqa(datasets.GeneratorBasedBuilder):
|
74 |
-
"""A New Benchmark for Selection-based Question Answering."""
|
75 |
-
|
76 |
-
VERSION = datasets.Version("1.1.0")
|
77 |
-
|
78 |
-
# This is an example of a dataset with multiple configurations.
|
79 |
-
# If you don't want/need to define several sub-sets in your dataset,
|
80 |
-
# just remove the BUILDER_CONFIG_CLASS and the BUILDER_CONFIGS attributes.
|
81 |
-
|
82 |
-
# If you need to make complex sub-parts in the datasets with configurable options
|
83 |
-
# You can create your own builder configuration class to store attribute, inheriting from datasets.BuilderConfig
|
84 |
-
BUILDER_CONFIG_CLASS = SelqaConfig
|
85 |
-
|
86 |
-
# You will be able to load one or the other configurations in the following list with
|
87 |
-
# data = datasets.load_dataset('my_dataset', 'first_domain')
|
88 |
-
# data = datasets.load_dataset('my_dataset', 'second_domain')
|
89 |
-
BUILDER_CONFIGS = [
|
90 |
-
SelqaConfig(
|
91 |
-
name="answer_selection_analysis",
|
92 |
-
mode="analysis",
|
93 |
-
type_="answer_selection",
|
94 |
-
version=VERSION,
|
95 |
-
description="This part covers answer selection analysis",
|
96 |
-
),
|
97 |
-
SelqaConfig(
|
98 |
-
name="answer_selection_experiments",
|
99 |
-
mode="experiments",
|
100 |
-
type_="answer_selection",
|
101 |
-
version=VERSION,
|
102 |
-
description="This part covers answer selection experiments",
|
103 |
-
),
|
104 |
-
SelqaConfig(
|
105 |
-
name="answer_triggering_analysis",
|
106 |
-
mode="analysis",
|
107 |
-
type_="answer_triggering",
|
108 |
-
version=VERSION,
|
109 |
-
description="This part covers answer triggering analysis",
|
110 |
-
),
|
111 |
-
SelqaConfig(
|
112 |
-
name="answer_triggering_experiments",
|
113 |
-
mode="experiments",
|
114 |
-
type_="answer_triggering",
|
115 |
-
version=VERSION,
|
116 |
-
description="This part covers answer triggering experiments",
|
117 |
-
),
|
118 |
-
]
|
119 |
-
|
120 |
-
DEFAULT_CONFIG_NAME = "answer_selection_analysis" # It's not mandatory to have a default configuration. Just use one if it make sense.
|
121 |
-
|
122 |
-
def _info(self):
|
123 |
-
if (
|
124 |
-
self.config.mode == "experiments"
|
125 |
-
): # This is the name of the configuration selected in BUILDER_CONFIGS above
|
126 |
-
features = datasets.Features(
|
127 |
-
{
|
128 |
-
"question": datasets.Value("string"),
|
129 |
-
"candidate": datasets.Value("string"),
|
130 |
-
"label": datasets.ClassLabel(names=["0", "1"]),
|
131 |
-
}
|
132 |
-
)
|
133 |
-
else:
|
134 |
-
if self.config.type_ == "answer_selection":
|
135 |
-
features = datasets.Features(
|
136 |
-
{
|
137 |
-
"section": datasets.Value("string"),
|
138 |
-
"question": datasets.Value("string"),
|
139 |
-
"article": datasets.Value("string"),
|
140 |
-
"is_paraphrase": datasets.Value("bool"),
|
141 |
-
"topic": datasets.ClassLabel(
|
142 |
-
names=[
|
143 |
-
"MUSIC",
|
144 |
-
"TV",
|
145 |
-
"TRAVEL",
|
146 |
-
"ART",
|
147 |
-
"SPORT",
|
148 |
-
"COUNTRY",
|
149 |
-
"MOVIES",
|
150 |
-
"HISTORICAL EVENTS",
|
151 |
-
"SCIENCE",
|
152 |
-
"FOOD",
|
153 |
-
]
|
154 |
-
),
|
155 |
-
"answers": datasets.Sequence(datasets.Value("int32")),
|
156 |
-
"candidates": datasets.Sequence(datasets.Value("string")),
|
157 |
-
"q_types": datasets.Sequence(
|
158 |
-
datasets.ClassLabel(names=["what", "why", "when", "who", "where", "how", ""])
|
159 |
-
),
|
160 |
-
}
|
161 |
-
)
|
162 |
-
else:
|
163 |
-
features = datasets.Features(
|
164 |
-
{
|
165 |
-
"section": datasets.Value("string"),
|
166 |
-
"question": datasets.Value("string"),
|
167 |
-
"article": datasets.Value("string"),
|
168 |
-
"is_paraphrase": datasets.Value("bool"),
|
169 |
-
"topic": datasets.ClassLabel(
|
170 |
-
names=[
|
171 |
-
"MUSIC",
|
172 |
-
"TV",
|
173 |
-
"TRAVEL",
|
174 |
-
"ART",
|
175 |
-
"SPORT",
|
176 |
-
"COUNTRY",
|
177 |
-
"MOVIES",
|
178 |
-
"HISTORICAL EVENTS",
|
179 |
-
"SCIENCE",
|
180 |
-
"FOOD",
|
181 |
-
]
|
182 |
-
),
|
183 |
-
"q_types": datasets.Sequence(
|
184 |
-
datasets.ClassLabel(names=["what", "why", "when", "who", "where", "how", ""])
|
185 |
-
),
|
186 |
-
"candidate_list": datasets.Sequence(
|
187 |
-
{
|
188 |
-
"article": datasets.Value("string"),
|
189 |
-
"section": datasets.Value("string"),
|
190 |
-
"candidates": datasets.Sequence(datasets.Value("string")),
|
191 |
-
"answers": datasets.Sequence(datasets.Value("int32")),
|
192 |
-
}
|
193 |
-
),
|
194 |
-
}
|
195 |
-
)
|
196 |
-
return datasets.DatasetInfo(
|
197 |
-
# This is the description that will appear on the datasets page.
|
198 |
-
description=_DESCRIPTION,
|
199 |
-
# This defines the different columns of the dataset and their types
|
200 |
-
features=features, # Here we define them above because they are different between the two configurations
|
201 |
-
# If there's a common (input, target) tuple from the features,
|
202 |
-
# specify them here. They'll be used if as_supervised=True in
|
203 |
-
# builder.as_dataset.
|
204 |
-
supervised_keys=None,
|
205 |
-
# Homepage of the dataset for documentation
|
206 |
-
homepage=_HOMEPAGE,
|
207 |
-
# License for the dataset if available
|
208 |
-
license=_LICENSE,
|
209 |
-
# Citation for the dataset
|
210 |
-
citation=_CITATION,
|
211 |
-
)
|
212 |
-
|
213 |
-
def _split_generators(self, dl_manager):
|
214 |
-
"""Returns SplitGenerators."""
|
215 |
-
# TODO: This method is tasked with downloading/extracting the data and defining the splits depending on the configuration
|
216 |
-
# If several configurations are possible (listed in BUILDER_CONFIGS), the configuration selected by the user is in self.config.name
|
217 |
-
|
218 |
-
# dl_manager is a datasets.download.DownloadManager that can be used to download and extract URLs
|
219 |
-
# It can accept any type or nested list/dict and will give back the same structure with the url replaced with path to local files.
|
220 |
-
# By default the archives will be extracted and a path to a cached folder where they are extracted is returned instead of the archive
|
221 |
-
urls = {
|
222 |
-
"train": f"https://raw.githubusercontent.com/emorynlp/selqa/master/{types[self.config.type_]}/selqa-{types[self.config.type_]}-train.{modes[self.config.mode]}",
|
223 |
-
"dev": f"https://raw.githubusercontent.com/emorynlp/selqa/master/{types[self.config.type_]}/selqa-{types[self.config.type_]}-dev.{modes[self.config.mode]}",
|
224 |
-
"test": f"https://raw.githubusercontent.com/emorynlp/selqa/master/{types[self.config.type_]}/selqa-{types[self.config.type_]}-test.{modes[self.config.mode]}",
|
225 |
-
}
|
226 |
-
data_dir = dl_manager.download_and_extract(urls)
|
227 |
-
return [
|
228 |
-
datasets.SplitGenerator(
|
229 |
-
name=datasets.Split.TRAIN,
|
230 |
-
# These kwargs will be passed to _generate_examples
|
231 |
-
gen_kwargs={
|
232 |
-
"filepath": data_dir["train"],
|
233 |
-
"split": "train",
|
234 |
-
},
|
235 |
-
),
|
236 |
-
datasets.SplitGenerator(
|
237 |
-
name=datasets.Split.TEST,
|
238 |
-
# These kwargs will be passed to _generate_examples
|
239 |
-
gen_kwargs={"filepath": data_dir["test"], "split": "test"},
|
240 |
-
),
|
241 |
-
datasets.SplitGenerator(
|
242 |
-
name=datasets.Split.VALIDATION,
|
243 |
-
# These kwargs will be passed to _generate_examples
|
244 |
-
gen_kwargs={
|
245 |
-
"filepath": data_dir["dev"],
|
246 |
-
"split": "dev",
|
247 |
-
},
|
248 |
-
),
|
249 |
-
]
|
250 |
-
|
251 |
-
def _generate_examples(self, filepath, split):
|
252 |
-
"""Yields examples."""
|
253 |
-
# TODO: This method will receive as arguments the `gen_kwargs` defined in the previous `_split_generators` method.
|
254 |
-
# It is in charge of opening the given file and yielding (key, example) tuples from the dataset
|
255 |
-
# The key is not important, it's more here for legacy reason (legacy from tfds)
|
256 |
-
with open(filepath, encoding="utf-8") as f:
|
257 |
-
if self.config.mode == "experiments":
|
258 |
-
csv_reader = csv.DictReader(
|
259 |
-
f, delimiter="\t", quoting=csv.QUOTE_NONE, fieldnames=["question", "candidate", "label"]
|
260 |
-
)
|
261 |
-
for id_, row in enumerate(csv_reader):
|
262 |
-
yield id_, row
|
263 |
-
else:
|
264 |
-
if self.config.type_ == "answer_selection":
|
265 |
-
for row in f:
|
266 |
-
data = json.loads(row)
|
267 |
-
for id_, item in enumerate(data):
|
268 |
-
yield id_, {
|
269 |
-
"section": item["section"],
|
270 |
-
"question": item["question"],
|
271 |
-
"article": item["article"],
|
272 |
-
"is_paraphrase": item["is_paraphrase"],
|
273 |
-
"topic": item["topic"],
|
274 |
-
"answers": item["answers"],
|
275 |
-
"candidates": item["candidates"],
|
276 |
-
"q_types": item["q_types"],
|
277 |
-
}
|
278 |
-
else:
|
279 |
-
for row in f:
|
280 |
-
data = json.loads(row)
|
281 |
-
for id_, item in enumerate(data):
|
282 |
-
candidate_list = []
|
283 |
-
for entity in item["candidate_list"]:
|
284 |
-
candidate_list.append(
|
285 |
-
{
|
286 |
-
"article": entity["article"],
|
287 |
-
"section": entity["section"],
|
288 |
-
"answers": entity["answers"],
|
289 |
-
"candidates": entity["candidates"],
|
290 |
-
}
|
291 |
-
)
|
292 |
-
yield id_, {
|
293 |
-
"section": item["section"],
|
294 |
-
"question": item["question"],
|
295 |
-
"article": item["article"],
|
296 |
-
"is_paraphrase": item["is_paraphrase"],
|
297 |
-
"topic": item["topic"],
|
298 |
-
"q_types": item["q_types"],
|
299 |
-
"candidate_list": candidate_list,
|
300 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|