Datasets:
tau
/

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
pandas
License:
commonsense_qa / dataset_infos.json
albertvillanova's picture
Convert dataset to Parquet
34986e8
raw
history blame
3.23 kB
{
"default": {
"description": "CommonsenseQA is a new multiple-choice question answering dataset that requires different types of commonsense knowledge\nto predict the correct answers . It contains 12,102 questions with one correct answer and four distractor answers.\nThe dataset is provided in two major training/validation/testing set splits: \"Random split\" which is the main evaluation\nsplit, and \"Question token split\", see paper for details.\n",
"citation": "@inproceedings{talmor-etal-2019-commonsenseqa,\n title = \"{C}ommonsense{QA}: A Question Answering Challenge Targeting Commonsense Knowledge\",\n author = \"Talmor, Alon and\n Herzig, Jonathan and\n Lourie, Nicholas and\n Berant, Jonathan\",\n booktitle = \"Proceedings of the 2019 Conference of the North {A}merican Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)\",\n month = jun,\n year = \"2019\",\n address = \"Minneapolis, Minnesota\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/N19-1421\",\n doi = \"10.18653/v1/N19-1421\",\n pages = \"4149--4158\",\n archivePrefix = \"arXiv\",\n eprint = \"1811.00937\",\n primaryClass = \"cs\",\n}\n",
"homepage": "https://www.tau-nlp.org/commonsenseqa",
"license": "",
"features": {
"id": {
"dtype": "string",
"_type": "Value"
},
"question": {
"dtype": "string",
"_type": "Value"
},
"question_concept": {
"dtype": "string",
"_type": "Value"
},
"choices": {
"feature": {
"label": {
"dtype": "string",
"_type": "Value"
},
"text": {
"dtype": "string",
"_type": "Value"
}
},
"_type": "Sequence"
},
"answerKey": {
"dtype": "string",
"_type": "Value"
}
},
"builder_name": "commonsense_qa",
"dataset_name": "commonsense_qa",
"config_name": "default",
"version": {
"version_str": "1.0.0",
"major": 1,
"minor": 0,
"patch": 0
},
"splits": {
"train": {
"name": "train",
"num_bytes": 2207794,
"num_examples": 9741,
"dataset_name": null
},
"validation": {
"name": "validation",
"num_bytes": 273848,
"num_examples": 1221,
"dataset_name": null
},
"test": {
"name": "test",
"num_bytes": 257842,
"num_examples": 1140,
"dataset_name": null
}
},
"download_size": 1558570,
"dataset_size": 2739484,
"size_in_bytes": 4298054
}
}