|
{ |
|
"wikitext-103-v1": { |
|
"description": " The WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified\n Good and Featured articles on Wikipedia. The dataset is available under the Creative Commons Attribution-ShareAlike\n License.\n", |
|
"citation": "@misc{merity2016pointer,\n title={Pointer Sentinel Mixture Models},\n author={Stephen Merity and Caiming Xiong and James Bradbury and Richard Socher},\n year={2016},\n eprint={1609.07843},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", |
|
"homepage": "https://blog.einstein.ai/the-wikitext-long-term-dependency-language-modeling-dataset/", |
|
"license": "Creative Commons Attribution-ShareAlike 4.0 International (CC BY-SA 4.0)", |
|
"features": { |
|
"text": { |
|
"dtype": "string", |
|
"_type": "Value" |
|
} |
|
}, |
|
"builder_name": "wikitext", |
|
"dataset_name": "wikitext", |
|
"config_name": "wikitext-103-v1", |
|
"version": { |
|
"version_str": "1.0.0", |
|
"major": 1, |
|
"minor": 0, |
|
"patch": 0 |
|
}, |
|
"splits": { |
|
"test": { |
|
"name": "test", |
|
"num_bytes": 1295575, |
|
"num_examples": 4358, |
|
"dataset_name": null |
|
}, |
|
"train": { |
|
"name": "train", |
|
"num_bytes": 545141915, |
|
"num_examples": 1801350, |
|
"dataset_name": null |
|
}, |
|
"validation": { |
|
"name": "validation", |
|
"num_bytes": 1154751, |
|
"num_examples": 3760, |
|
"dataset_name": null |
|
} |
|
}, |
|
"download_size": 313093838, |
|
"dataset_size": 547592241, |
|
"size_in_bytes": 860686079 |
|
}, |
|
"wikitext-2-v1": { |
|
"description": " The WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified\n Good and Featured articles on Wikipedia. The dataset is available under the Creative Commons Attribution-ShareAlike\n License.\n", |
|
"citation": "@misc{merity2016pointer,\n title={Pointer Sentinel Mixture Models},\n author={Stephen Merity and Caiming Xiong and James Bradbury and Richard Socher},\n year={2016},\n eprint={1609.07843},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", |
|
"homepage": "https://blog.einstein.ai/the-wikitext-long-term-dependency-language-modeling-dataset/", |
|
"license": "Creative Commons Attribution-ShareAlike 4.0 International (CC BY-SA 4.0)", |
|
"features": { |
|
"text": { |
|
"dtype": "string", |
|
"_type": "Value" |
|
} |
|
}, |
|
"builder_name": "wikitext", |
|
"dataset_name": "wikitext", |
|
"config_name": "wikitext-2-v1", |
|
"version": { |
|
"version_str": "1.0.0", |
|
"major": 1, |
|
"minor": 0, |
|
"patch": 0 |
|
}, |
|
"splits": { |
|
"test": { |
|
"name": "test", |
|
"num_bytes": 1270947, |
|
"num_examples": 4358, |
|
"dataset_name": null |
|
}, |
|
"train": { |
|
"name": "train", |
|
"num_bytes": 10918118, |
|
"num_examples": 36718, |
|
"dataset_name": null |
|
}, |
|
"validation": { |
|
"name": "validation", |
|
"num_bytes": 1134123, |
|
"num_examples": 3760, |
|
"dataset_name": null |
|
} |
|
}, |
|
"download_size": 7371282, |
|
"dataset_size": 13323188, |
|
"size_in_bytes": 20694470 |
|
}, |
|
"wikitext-103-raw-v1": { |
|
"description": " The WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified\n Good and Featured articles on Wikipedia. The dataset is available under the Creative Commons Attribution-ShareAlike\n License.\n", |
|
"citation": "@misc{merity2016pointer,\n title={Pointer Sentinel Mixture Models},\n author={Stephen Merity and Caiming Xiong and James Bradbury and Richard Socher},\n year={2016},\n eprint={1609.07843},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", |
|
"homepage": "https://blog.einstein.ai/the-wikitext-long-term-dependency-language-modeling-dataset/", |
|
"license": "Creative Commons Attribution-ShareAlike 4.0 International (CC BY-SA 4.0)", |
|
"features": { |
|
"text": { |
|
"dtype": "string", |
|
"_type": "Value" |
|
} |
|
}, |
|
"builder_name": "wikitext", |
|
"dataset_name": "wikitext", |
|
"config_name": "wikitext-103-raw-v1", |
|
"version": { |
|
"version_str": "1.0.0", |
|
"major": 1, |
|
"minor": 0, |
|
"patch": 0 |
|
}, |
|
"splits": { |
|
"test": { |
|
"name": "test", |
|
"num_bytes": 1305088, |
|
"num_examples": 4358, |
|
"dataset_name": null |
|
}, |
|
"train": { |
|
"name": "train", |
|
"num_bytes": 546500949, |
|
"num_examples": 1801350, |
|
"dataset_name": null |
|
}, |
|
"validation": { |
|
"name": "validation", |
|
"num_bytes": 1159288, |
|
"num_examples": 3760, |
|
"dataset_name": null |
|
} |
|
}, |
|
"download_size": 315466397, |
|
"dataset_size": 548965325, |
|
"size_in_bytes": 864431722 |
|
}, |
|
"wikitext-2-raw-v1": { |
|
"description": " The WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified\n Good and Featured articles on Wikipedia. The dataset is available under the Creative Commons Attribution-ShareAlike\n License.\n", |
|
"citation": "@misc{merity2016pointer,\n title={Pointer Sentinel Mixture Models},\n author={Stephen Merity and Caiming Xiong and James Bradbury and Richard Socher},\n year={2016},\n eprint={1609.07843},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", |
|
"homepage": "https://blog.einstein.ai/the-wikitext-long-term-dependency-language-modeling-dataset/", |
|
"license": "Creative Commons Attribution-ShareAlike 4.0 International (CC BY-SA 4.0)", |
|
"features": { |
|
"text": { |
|
"dtype": "string", |
|
"id": null, |
|
"_type": "Value" |
|
} |
|
}, |
|
"post_processed": null, |
|
"supervised_keys": null, |
|
"task_templates": null, |
|
"builder_name": "wikitext", |
|
"config_name": "wikitext-2-raw-v1", |
|
"version": { |
|
"version_str": "1.0.0", |
|
"description": null, |
|
"major": 1, |
|
"minor": 0, |
|
"patch": 0 |
|
}, |
|
"splits": { |
|
"test": { |
|
"name": "test", |
|
"num_bytes": 1305092, |
|
"num_examples": 4358, |
|
"dataset_name": "wikitext" |
|
}, |
|
"train": { |
|
"name": "train", |
|
"num_bytes": 11061733, |
|
"num_examples": 36718, |
|
"dataset_name": "wikitext" |
|
}, |
|
"validation": { |
|
"name": "validation", |
|
"num_bytes": 1159292, |
|
"num_examples": 3760, |
|
"dataset_name": "wikitext" |
|
} |
|
}, |
|
"download_checksums": { |
|
"https://s3.amazonaws.com/research.metamind.io/wikitext/wikitext-2-raw-v1.zip": { |
|
"num_bytes": 4721645, |
|
"checksum": "ef7edb566e3e2b2d31b29c1fdb0c89a4cc683597484c3dc2517919c615435a11" |
|
} |
|
}, |
|
"download_size": 4721645, |
|
"post_processing_size": null, |
|
"dataset_size": 13526117, |
|
"size_in_bytes": 18247762 |
|
} |
|
} |