qa4mre / dataset_infos.json
system's picture
system HF staff
Update files from the datasets library (from 1.0.0)
4cb7603
raw
history blame
77.7 kB
{"2011.main.DE": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2011.main.DE", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 1747118, "num_examples": 120, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2011/Training_Data/Goldstandard/QA4MRE-2011-DE_GS.xml": {"num_bytes": 222289, "checksum": "9271e5f0aba731914fda18117b8255807d85b4cd7b0835ddca67f86ea9cb9381"}}, "download_size": 222289, "dataset_size": 1747118, "size_in_bytes": 1969407}, "2011.main.EN": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2011.main.EN", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 1569676, "num_examples": 120, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2011/Training_Data/Goldstandard/QA4MRE-2011-EN_GS.xml": {"num_bytes": 202490, "checksum": "6d2524952a3a015f2a82df785b85b5578681e3602ec276b4e72c01f4ebc50034"}}, "download_size": 202490, "dataset_size": 1569676, "size_in_bytes": 1772166}, "2011.main.ES": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2011.main.ES", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 1694460, "num_examples": 120, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2011/Training_Data/Goldstandard/QA4MRE-2011-ES_GS.xml": {"num_bytes": 217617, "checksum": "750b907bccf485cb9fa6e34c4f2fc2dbb4e87123ebf8ee123feaa5e178cdd956"}}, "download_size": 217617, "dataset_size": 1694460, "size_in_bytes": 1912077}, "2011.main.IT": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2011.main.IT", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 1667188, "num_examples": 120, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2011/Training_Data/Goldstandard/QA4MRE-2011-IT_GS.xml": {"num_bytes": 214764, "checksum": "8dad14ffc087092d8de2351ca99cd219f492b860ec29c3c2b89ddbecc5f9f19a"}}, "download_size": 214764, "dataset_size": 1667188, "size_in_bytes": 1881952}, "2011.main.RO": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2011.main.RO", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 1740419, "num_examples": 120, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2011/Training_Data/Goldstandard/QA4MRE-2011-RO_GS.xml": {"num_bytes": 221510, "checksum": "8e6abab8105c95dca29dd6db29e08fc361ee5538383b443a5f8e0abcf04a3483"}}, "download_size": 221510, "dataset_size": 1740419, "size_in_bytes": 1961929}, "2012.main.AR": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2012.main.AR", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 2710656, "num_examples": 160, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2012/Main_Task/Training_Data/Goldstandard/Used_in_Evaluation/QA4MRE-2012-AR_GS.xml": {"num_bytes": 356178, "checksum": "b175b9284d6bdb9a5f64a2a4058994fecdc910e35daba07a854e7583eec50ddd"}}, "download_size": 356178, "dataset_size": 2710656, "size_in_bytes": 3066834}, "2012.main.BG": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2012.main.BG", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 3454215, "num_examples": 160, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2012/Main_Task/Training_Data/Goldstandard/Used_in_Evaluation/QA4MRE-2012-BG_GS.xml": {"num_bytes": 445060, "checksum": "4c4d0548a0ad58136d2801934da206a3744a5fb0e16df820bc0010fcf71968ea"}}, "download_size": 445060, "dataset_size": 3454215, "size_in_bytes": 3899275}, "2012.main.DE": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2012.main.DE", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 2087466, "num_examples": 160, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2012/Main_Task/Training_Data/Goldstandard/Used_in_Evaluation/QA4MRE-2012-DE_GS.xml": {"num_bytes": 281600, "checksum": "5a1384187712efc15c525ac3440dd73e3885bd8f4bfa0e2f5b482b84392c841c"}}, "download_size": 281600, "dataset_size": 2087466, "size_in_bytes": 2369066}, "2012.main.EN": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2012.main.EN", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 1757586, "num_examples": 160, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2012/Main_Task/Training_Data/Goldstandard/Used_in_Evaluation/QA4MRE-2012-EN_GS.xml": {"num_bytes": 243467, "checksum": "f9edaf408f8ac93f89a643a0d0b19263a1bb5ce64f19b2af10df279a656dfb24"}}, "download_size": 243467, "dataset_size": 1757586, "size_in_bytes": 2001053}, "2012.main.ES": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2012.main.ES", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 2057402, "num_examples": 160, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2012/Main_Task/Training_Data/Goldstandard/Used_in_Evaluation/QA4MRE-2012-ES_GS.xml": {"num_bytes": 278445, "checksum": "8f8d1074da323553f4b26597d6fb96da7896f5434558b2889a6dda1ff0eb12d5"}}, "download_size": 278445, "dataset_size": 2057402, "size_in_bytes": 2335847}, "2012.main.IT": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2012.main.IT", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 2071710, "num_examples": 160, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2012/Main_Task/Training_Data/Goldstandard/Used_in_Evaluation/QA4MRE-2012-IT_GS.xml": {"num_bytes": 280051, "checksum": "6337c0db0001b70c9c463a0b8346599b4cb8a561a05ec094c3ab7e027130b195"}}, "download_size": 280051, "dataset_size": 2071710, "size_in_bytes": 2351761}, "2012.main.RO": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2012.main.RO", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 2074930, "num_examples": 160, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2012/Main_Task/Training_Data/Goldstandard/Used_in_Evaluation/QA4MRE-2012-RO_GS.xml": {"num_bytes": 279541, "checksum": "3077e3b1b8a941a1b8c48943fe91bef904c35e041b7bdf944e166abbc76ed6c4"}}, "download_size": 279541, "dataset_size": 2074930, "size_in_bytes": 2354471}, "2012.alzheimers.EN": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2012.alzheimers.EN", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 1637988, "num_examples": 40, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2012/Pilot_Tasks/Biomedical_About_Alzheimer/Training_Data/Goldstandard/QA4MRE-2012_BIOMEDICAL_GS.xml": {"num_bytes": 177345, "checksum": "fc7384029559f37c0b8ea33432a9d4a8dbe16056fc371f48a3efcebf87d187cb"}}, "download_size": 177345, "dataset_size": 1637988, "size_in_bytes": 1815333}, "2013.main.AR": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2013.main.AR", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 4180979, "num_examples": 284, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2013/Main_Task/Training_Data/Goldstandard/QA4MRE-2013-AR_GS.xml": {"num_bytes": 378302, "checksum": "392f979fdc644a312551c8c8bf218aa28079478d34eb3ce35456734b46f3d039"}}, "download_size": 378302, "dataset_size": 4180979, "size_in_bytes": 4559281}, "2013.main.BG": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2013.main.BG", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 5403246, "num_examples": 284, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2013/Main_Task/Training_Data/Goldstandard/QA4MRE-2013-BG_GS.xml": {"num_bytes": 463605, "checksum": "b8cd10a31299530aa47c0cd73b06d503c72e25e8b420f93b9c26660b64f4e3b4"}}, "download_size": 463605, "dataset_size": 5403246, "size_in_bytes": 5866851}, "2013.main.EN": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2013.main.EN", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 2887866, "num_examples": 284, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2013/Main_Task/Training_Data/Goldstandard/QA4MRE-2013-EN_GS.xml": {"num_bytes": 274969, "checksum": "c60e5aa4ec77e0493ef0b11d46bd1d74d58a499a3a2f871b8cf3af9536f0f094"}}, "download_size": 274969, "dataset_size": 2887866, "size_in_bytes": 3162835}, "2013.main.ES": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2013.main.ES", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 3449693, "num_examples": 284, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2013/Main_Task/Training_Data/Goldstandard/QA4MRE-2013-ES_GS.xml": {"num_bytes": 315166, "checksum": "60d92a4d8542a038ad1552afeee58860dc90c6ad5a2b2eda7f68f8788b5e704f"}}, "download_size": 315166, "dataset_size": 3449693, "size_in_bytes": 3764859}, "2013.main.RO": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2013.main.RO", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 3363049, "num_examples": 284, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2013/Main_Task/Training_Data/Goldstandard/QA4MRE-2013-RO_GS.xml": {"num_bytes": 313510, "checksum": "d3e6ecb8563a9ebb07419e1e2c402daf7e3fd232d14eda54386c108855623f5e"}}, "download_size": 313510, "dataset_size": 3363049, "size_in_bytes": 3676559}, "2013.alzheimers.EN": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2013.alzheimers.EN", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 2614812, "num_examples": 40, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2013/Biomedical_About_Alzheimer/Training_Data/Goldstandard/QA4MRE-2013_BIO_GS-RUN.xml": {"num_bytes": 274413, "checksum": "81b106eb14ec482a96ff646c6661bc34b2939392b858036458f80e7691082f72"}}, "download_size": 274413, "dataset_size": 2614812, "size_in_bytes": 2889225}, "2013.entrance_exam.EN": {"description": "\nQA4MRE dataset was created for the CLEF 2011/2012/2013 shared tasks to promote research in \nquestion answering and reading comprehension. The dataset contains a supporting \npassage and a set of questions corresponding to the passage. Multiple options \nfor answers are provided for each question, of which only one is correct. The \ntraining and test datasets are available for the main track.\nAdditional gold standard documents are available for two pilot studies: one on \nalzheimers data, and the other on entrance exams data.\n", "citation": "\n@InProceedings{10.1007/978-3-642-40802-1_29,\nauthor=\"Pe{\\~{n}}as, Anselmo\nand Hovy, Eduard\nand Forner, Pamela\nand Rodrigo, {\\'A}lvaro\nand Sutcliffe, Richard\nand Morante, Roser\",\neditor=\"Forner, Pamela\nand M{\\\"u}ller, Henning\nand Paredes, Roberto\nand Rosso, Paolo\nand Stein, Benno\",\ntitle=\"QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation\",\nbooktitle=\"Information Access Evaluation. Multilinguality, Multimodality, and Visualization\",\nyear=\"2013\",\npublisher=\"Springer Berlin Heidelberg\",\naddress=\"Berlin, Heidelberg\",\npages=\"303--320\",\nabstract=\"This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011--2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimer's disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.\",\nisbn=\"978-3-642-40802-1\"\n}\n", "homepage": "http://nlp.uned.es/clef-qa/repository/pastCampaigns.php", "license": "", "features": {"topic_id": {"dtype": "string", "id": null, "_type": "Value"}, "topic_name": {"dtype": "string", "id": null, "_type": "Value"}, "test_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_id": {"dtype": "string", "id": null, "_type": "Value"}, "document_str": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_str": {"dtype": "string", "id": null, "_type": "Value"}, "answer_options": {"feature": {"answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}, "correct_answer_id": {"dtype": "string", "id": null, "_type": "Value"}, "correct_answer_str": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "qa4mre", "config_name": "2013.entrance_exam.EN", "version": {"version_str": "0.1.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 180827, "num_examples": 46, "dataset_name": "qa4mre"}}, "download_checksums": {"http://nlp.uned.es/clef-qa/repository/js/scripts/downloadFile.php?file=/var/www/html/nlp/clef-qa/repository/resources/QA4MRE/2013/Entrance_Exams/Training_Data/Goldstandard/qa4mre-exam-test-withanswer.xml": {"num_bytes": 54598, "checksum": "fac91b06e6a8b0ac17e0c1b94b1296051020384b4fb1ba9a1d416da1a5ed9e0b"}}, "download_size": 54598, "dataset_size": 180827, "size_in_bytes": 235425}}