--- language: - en --- # Dataset description Number of articles for the English wiki: - 2014: 4.599.592 - 2016: 5.144.403 - 2018: 5.599.764 - 2020: 6.037.287 - 2022: 6.291.973 - 2024: 6.629.861 with 8k documents per gzipped json, we get approx. 800 files per year, so aprrox. 3500 files in total. Approx. 190-200 MB per bin of 8k examples, 63 in gzipped mode. # Dataset sizes There are different splits and dataset sizes, which are all subsets of the `full` set. | Config | Train | Validation | Estimate | |----------|-------|------------|----------| | tiny | 16k docs (2 shards) | 8k docs (1 shard) | 0.1GB | | small | 800k docs (100 shards) | 16k docs (2 shards) | 4GB | | medium | 6M docs (750 shards) | 16k docs (2 shards) | 30GB | | large | 12M docs (1500 shards) | 24k docs (3 shards) | 59GB | | full | 28M docs (3497 shards) | 32k docs (4 shards) | 137GB |