File size: 24,295 Bytes
41ef435 fb49936 6a7ed79 fb49936 6a7ed79 41ef435 80908ce 41ef435 48ff894 41ef435 48ff894 41ef435 48ff894 41ef435 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 |
---
license:
- cc-by-sa-4.0
- gfdl
size_categories:
- 1M<n<10M
task_ids:
- language-modeling
- masked-language-modeling
language:
- en
- fr
tags:
- wikimedia
- wikipedia
configs:
- config_name: 20240916.en
data_files:
- split: train
path: 20240916.en/enwiki_*
- config_name: 20240916.fr
data_files:
- split: train
path: 20240916.fr/frwiki_*
---
# Dataset Card for Wikimedia Structured Wikipedia
## Dataset Description
- **Homepage:** https://enterprise.wikimedia.com/
- **Point of Contact:** [sdelbecque@wikimedia.org](mailto:sdelbecque@wikimedia.org)
## Dataset Description
### Dataset Summary
Early beta release of pre-parsed English and French Wikipedia articles including infoboxes. Inviting feedback.
This dataset contains all articles of the [English](https://en.wikipedia.org/) and [French](https://fr.wikipedia.org/) language editions of Wikipedia, pre-parsed and outputted as structured JSON files with a consistent schema (JSONL compressed as zip). Each JSON line holds the content of one full Wikipedia article stripped of extra markdown and non-prose sections (references, etc.).
### Invitation for Feedback
The dataset is built as part of the Structured Contents initiative and based on the Wikimedia Enterprise html [snapshots](https://enterprise.wikimedia.com/docs/snapshot/). It is an early beta release to improve transparency in the development process and request feedback. This first version includes pre-parsed Wikipedia abstracts, short descriptions, main images links, infoboxes and article sections, excluding non-prose sections (e.g. references). More elements (such as lists and tables) may be added over time. For updates follow the project’s [blog](https://enterprise.wikimedia.com/blog/) and our [Mediawiki Quarterly software updates](https://www.mediawiki.org/wiki/Wikimedia_Enterprise#Updates) on MediaWiki.
As this is an early beta release, we highly value your feedback to help us refine and improve this dataset. Please share your thoughts, suggestions, and any issues you encounter either [on the discussion page](https://meta.wikimedia.org/wiki/Talk:Wikimedia_Enterprise) of Wikimedia Enterprise’s homepage on Meta wiki, or on the discussion page for this dataset here on HuggingFace.
The contents of this dataset of Wikipedia articles is collectively written and curated by a global volunteer community. All original textual content is licensed under the [GNU Free Documentation License](https://www.gnu.org/licenses/fdl-1.3.html) (GFDL) and the [Creative Commons Attribution-Share-Alike 4.0 License](https://creativecommons.org/licenses/by-sa/4.0/). Some text may be available only under the Creative Commons license; see the Wikimedia [Terms of Use](https://foundation.wikimedia.org/wiki/Policy:Terms_of_Use) for details. Text written by some authors may be released under additional licenses or into the public domain.
### Supported Tasks and Leaderboards
The dataset in its structured form is generally helpful for a wide variety of tasks, including all phases of model development, from pre-training to alignment, fine-tuning, updating/RAG as well as testing/benchmarking.
We would love to hear more about your use cases.
### Languages
English <br/>
BCP 47 Language Code: EN
French/Français <br/>
BCP 47 Language Code: FR
There is only one language edition for English, and one for French. They encompass national and cultural variations of spelling, vocabulary, grammar etc. Within a Wikipedia language edition, no national variety is officially preferred over others. The rule of thumb is that the conventions of a particular variety of language should be followed consistently within a given article.
The relevant Manual of Style policy documents are available [for English Wikipedia](https://en.wikipedia.org/wiki/Wikipedia:Manual_of_Style#National_varieties_of_English), and [for French Wikipedia](https://fr.wikipedia.org/wiki/Wikip%C3%A9dia:Conventions_de_style#Variantes_g%C3%A9ographiques).
As part of this Beta Wikimedia Enterprise Structured Contents project, the team is working to including more language editions: [https://enterprise.wikimedia.com/docs/snapshot/#available-structured-contents-snapshots-beta](https://enterprise.wikimedia.com/docs/snapshot/#available-structured-contents-snapshots-beta)
## Dataset Structure
### Data Instances
An example of each line of JSON looks as follows (abbreviated data):
```json
{
"name":"JosephineBaker",
"identifier":255083,
"url":"https://en.wikipedia.org/wiki/Josephine_Baker",
"date_created":"...",
"date_modified":"...",
"is_part_of":{"..."},
"in_language":{"..."},
"main_entity":{"identifier":"Q151972",...},
"additional_entities":[...],
"version":{...},
"description":"American-bornFrenchdancer...",
"abstract":"FredaJosephineBaker,naturalizedas...",
"image":{"content_url":"https://upload.wikimedia.org/wikipedia/...",...},
"infobox":[{"name":"Infoboxperson",
"type":"infobox",
"has_parts":[
{ "name":"JosephineBaker",
"type":"section",
"has_parts":[
{"name":"Born",
"type":"field",
"value":"FredaJosephineMcDonaldJune3,1906 St.Louis,Missouri,US",
"links":[{"url": "https://en.wikipedia.org/wiki/St._Louis", "text":"St.Louis"},}],
]
}
]
}],
"sections": [{"name": "Abstract",
"type": "section",
"has_parts": [
{ "type": "paragraph",
"value": "Freda Josephine Baker (née McDonald; June 3, 1906 - April 12, 1975), naturalized as Joséphine Baker...",
"links": [{"url": "https://en.wikipedia.org/wiki/Siren_...","text": "Siren of the Tropics"...}]
}
],
...
}],
"license":[...],
}
```
### Timestamp
Dataset extracted 16 September 2024
### Size
English: enwiki_namespace_0.zip
Size of compressed dataset files: 17.91 GB <br/>
Size of uncompressed dataset: 79.57 GB
French: frwiki_namespace_0.zip
Size of compressed dataset files: 6.95 GB <br/>
Size of the uncompressed dataset: 34.01 GB
JSONL files compressed in zip, once uncompressed they are chunked by max 2.15GB.
For more guidelines on how to process the snapshots, see [SDKs](https://enterprise.wikimedia.com/docs/#sdk).
### Data Fields
The data fields are the same among all, noteworthy included fields:
* [name](https://enterprise.wikimedia.com/docs/data-dictionary/#name) - title of the article.
* [identifier](https://enterprise.wikimedia.com/docs/data-dictionary/#identifier) - ID of the article.
* [url](https://enterprise.wikimedia.com/docs/data-dictionary/#url) - URL of the article.
* [version](https://enterprise.wikimedia.com/docs/data-dictionary/#version): metadata related to the latest specific revision of the article
* [version.editor](https://enterprise.wikimedia.com/docs/data-dictionary/#version_editor) - editor-specific signals that can help contextualize the revision
* [version.scores](https://enterprise.wikimedia.com/docs/data-dictionary/#version_scores) - returns assessments by ML models on the likelihood of a revision being reverted.
* [main entity](https://enterprise.wikimedia.com/docs/data-dictionary/#main_entity) - Wikidata QID the article is related to.
* [abstract](https://enterprise.wikimedia.com/docs/data-dictionary/#abstract) - lead section, summarizing what the article is about.
* [description](https://enterprise.wikimedia.com/docs/data-dictionary/#description) - one-sentence description of the article for quick reference.
* [image](https://enterprise.wikimedia.com/docs/data-dictionary/#image) - main image representing the article's subject.
* [infobox](https://enterprise.wikimedia.com/docs/data-dictionary/#infobox) - parsed information from the side panel (infobox) on the Wikipedia article.
* [sections](https://enterprise.wikimedia.com/docs/data-dictionary/#article_sections) - parsed sections of the article, including links.
Note: excludes other media/images, lists, tables and references or similar non-prose sections.
Full data dictionary is available here: [https://enterprise.wikimedia.com/docs/data-dictionary/](https://enterprise.wikimedia.com/docs/data-dictionary/)
### Data Splits
[N/A]
## Dataset Creation
### Curation Rationale
This dataset has been created as part of the larger [Structured Contents initiative at Wikimedia Enterprise](https://meta.wikimedia.org/wiki/Wikimedia_Enterprise/FAQ#What_is_Structured_Contents) with the aim of making Wikimedia data more machine readable. These efforts are both focused on pre-parsing Wikipedia snippets as well as connecting the different projects closer together.
Even if Wikipedia is very structured to the human eye, it is a non-trivial task to extract the knowledge lying within in a machine readable manner. Projects, languages, domains all have their own specific community experts and way of structuring data, bolstered by various templates and best practices. A specific example we’ve addressed in this release is article infoboxes. Infoboxes are panels that commonly appear in the top right corner of a Wikipedia article and summarize key facts and statistics of the article’s subject. The editorial community works hard to keep infoboxes populated with the article’s most pertinent and current metadata, and we’d like to lower the barrier of entry significantly so that this data is also accessible at scale without the need for bespoke parsing systems.
We also include the link to the [Wikidata Q](https://www.wikidata.org/wiki/Wikidata:Identifiers) Identifier (corresponding Wikidata entity), and the link to (main and infobox) images to facilitate easier access to additional information on the specific topics.
You will also find [Credibility Signals fields](https://enterprise.wikimedia.com/blog/understanding-credibility-signals-in-wikimedia-enterprise-api/) included. These can help you decide when, how, and why to use what is in the dataset. These fields mirror the over 20 years of editorial policies created and kept by the Wikipedia editing communities, taking publicly available information and structuring it. Like with article structures, because this information is not centralized (neither on a single project nor across them), it is hard to access. Credibility signals shine a light on that blind spot. You will find most of these signals under the ‘version’ object, but other objects like ‘protection’ and ‘watchers_count’ offer similar insight.
This is an early beta release of pre-parsed Wikipedia articles in bulk, as a means to improve transparency in the development process and gather insights of current use cases to follow where the AI community needs us most; as well as feedback points, to develop this further through collaboration. There will be limitations (see ‘known limitations’ section below), but in line with [our values](https://wikimediafoundation.org/about/values/#a3-we-welcome-everyone-who-shares-our-vision-and-values), we believe it is better to share early, often, and respond to feedback.
You can also test out more languages on an article by article basis through our [beta Structured Contents On-demand endpoint](https://enterprise.wikimedia.com/docs/on-demand/#article-structured-contents-beta) with a [free account](https://enterprise.wikimedia.com/blog/enhanced-free-api/).
Attribution is core to the sustainability of the Wikimedia projects. It is what drives new editors and donors to Wikipedia. With consistent attribution, this cycle of content creation and reuse ensures encyclopedic content of high-quality, reliability, and verifiability will continue being written on Wikipedia and ultimately remain available for reuse via datasets such as these.
As such, we require all users of this dataset to conform to our expectations for proper attribution. Detailed attribution requirements for use of this dataset are outlined below.
Beyond attribution, there are many ways of contributing to and supporting the Wikimedia movement. and various other ways of supporting and participating in the Wikimedia movement below. To discuss your specific circumstances please contact Nicholas Perry from the Wikimedia Foundation technical partnerships team at [nperry@wikimedia.org](mailto:nperry@wikimedia.org). You can also contact us on either the [discussion page](https://meta.wikimedia.org/wiki/Talk:Wikimedia_Enterprise) of Wikimedia Enterprise’s homepage on Meta wiki, or on the discussion page for this dataset here on HuggingFace.
### Source Data
### Initial Data Collection and Normalization
The dataset is built from the Wikimedia Enterprise HTML “snapshots”: [https://enterprise.wikimedia.com/docs/snapshot/](https://enterprise.wikimedia.com/docs/snapshot/) and focusses on the Wikipedia article namespace ([namespace 0 (main))](https://en.wikipedia.org/wiki/Wikipedia:What_is_an_article%3F#Namespace).
### Who are the source language producers?
Wikipedia is a human generated corpus of free knowledge, written, edited, and curated by a [global community of editors](https://meta.wikimedia.org/wiki/Community_Insights/Community_Insights_2023_Report) since 2001.
It is the largest and most accessed educational resource in history, accessed over 20 billion times by half a billion people each month. Wikipedia represents almost 25 years of work by its community; the creation, curation, and maintenance of millions of articles on distinct topics.
This dataset includes the complete article contents of two Wikipedia language editions: English [https://en.wikipedia.org/](https://en.wikipedia.org/) and French [https://fr.wikipedia.org/](https://fr.wikipedia.org/), written by the respective communities.
### Annotations
### Annotation process
[N/A]
The dataset doesn't contain any additional annotations
### Who are the annotators?
[N/A]
### Personal and Sensitive Information
The Wikipedia community and the Wikimedia Foundation - which operates Wikipedia - establish robust policy and guidelines around personal and sensitive information; both to avoid personal/sensitive information within articles as well as strict privacy policies around information of their contributors.
The Wikimedia Foundation’s privacy policy is available at: [https://foundation.wikimedia.org/wiki/Policy:Privacy_policy](https://foundation.wikimedia.org/wiki/Policy:Privacy_policy)
Transparency reports covering the Wikimedia Foundation’s responses to requests received to alter or remove content from the projects, and to provide nonpublic information about users, are found here: [https://wikimediafoundation.org/about/transparency/](https://wikimediafoundation.org/about/transparency/)
Among many editorial policies regarding personal and sensitive information, particular care is paid by the community to biographies of living people. Details for each language community’s responses and norms can be found here: [https://www.wikidata.org/wiki/Q4663389#sitelinks-wikipedia](https://www.wikidata.org/wiki/Q4663389#sitelinks-wikipedia)
## Considerations for Using the Data
### Social Impact of Dataset
Wikipedia’s articles are read over 20 billion times by half a billion people each month. It does not belong to, or come from a single culture or language. It is an example of mass international cooperation, across languages and continents. Wikipedia is the only website among the world’s most visited that is not operated by a commercial organization.
Wikimedia projects have been used (and upsampled) as a core source of qualitative data in AI/ML/LLM . The Wikimedia Foundation has published an article on [the value of Wikipedia in age of generative AI](https://wikimediafoundation.org/news/2023/07/12/wikipedias-value-in-the-age-of-generative-ai).
There is also a Community Article about [why Wikimedia data matters for ML](https://huggingface.co/blog/frimelle/wikipedias-treasure-trove-ml-data#why-wikimedia-data-for-ml) on the HuggingFace blog. This highlights that it is: rich, diverse content; multimodal; and both community-curated and openly licensed.
### Discussion of Biases
While consciously trying to present an editorially neutral point of view, Wikipedia’s content reflects [the biases of the society it comes from](https://wikimediafoundation.org/our-work/open-the-knowledge/otk-change-the-stats/). This includes various “gaps” (notably in both the [proportion of biographies of, and editors who identify as, women](https://wikimediafoundation.org/our-work/open-the-knowledge/wikipedia-needs-more-women/). Other significant gaps include in linguistic and technical accessibility of the websites, and censorship. Because the content is written by its readers, ensuring the widest possible access to the content is crucial to reducing the biases of the content itself. There is continuous work to redress these biases through various social and technical efforts – both centrally and at the grassroots around the world.
### Other Known Limitations
This is an early beta version, the following limitations may apply:
* A small percentage of duplicated, deleted or missed articles may be part of the snapshot. Duplicates can be filtered out by looking at the highest "version.identifier", being the most up-to-date revision of the article
* Revision discrepancies may happen due to limitations with long articles
* On occasion empty sections or values may be returned. This is either because the section contains references or similar; or is made out of structured elements like lists and tables; or the section was left empty by editors.
* Images: only main and infobox image links are supported at the moment, we encourage you to take additional information and licensing by following the image link, while we are evaluating adding this data directly.
Please let us know if there are any other limitations that aren't covered above.
## Additional Information
### Dataset Curators
This dataset was created by the [Wikimedia Enterprise](https://enterprise.wikimedia.com/about/) team of the [Wikimedia Foundation](https://wikimediafoundation.org/) as part of the Structured Contents initiative.
### Attribution Information
Wikimedia Enterprise provides this dataset under the assumption that downstream users will adhere to the relevant free culture licenses when the data is reused. In situations where attribution is required, reusers should identify the Wikimedia project from which the content was retrieved as the source of the content. Any attribution should adhere to Wikimedia’s trademark policy (available at [https://foundation.wikimedia.org/wiki/Trademark_policy](https://foundation.wikimedia.org/wiki/Trademark_policy)) and visual identity guidelines (available at [https://foundation.wikimedia.org/wiki/Visual_identity_guidelines](https://foundation.wikimedia.org/wiki/Visual_identity_guidelines)) when identifying Wikimedia as the source of content.
#### How To Attribute Wikipedia
We ask that all content re-users attribute Wikipedia in a way that supports our model. In the spirit of reciprocity the framework allows you to leverage our brand to signal trust, reliability and recency whilst also communicating that our dataset is written entirely by human contributors who have volunteered their time in the spirit of knowledge for all.
For example, for Generative AI, what to include in all attribution cases:
<img src="https://huggingface.co/datasets/wikimedia/structured-wikipedia/resolve/main/images/attrib_table.png" style="width: 1000px; height:483px;" />
Attribution UI Example: Outputs using Wikipedia non-specifically
![](https://huggingface.co/datasets/wikimedia/structured-wikipedia/resolve/main/images/attr1.png)
Attribution UI Example: Outputs using Wikipedia in-line
![](https://huggingface.co/datasets/wikimedia/structured-wikipedia/resolve/main/images/attr2.png)
#### Tools & Resources for Attribution
[W Favicons](https://commons.wikimedia.org/wiki/Category:Wikimedia_Attribution_Guide_Favicons): The Wikipedia ‘W’ icon is a distinctive graphic element derived from the W of Wikipedia, set in the font Hoefler combined with various framing devices.
<img src="https://upload.wikimedia.org/wikipedia/commons/thumb/1/1e/Wikipedia_%22W%22_Square_Black.svg/240px-Wikipedia_%22W%22_Square_Black.svg.png" style="display: inline-block; width: 50px; height:50px;" /><img src="https://upload.wikimedia.org/wikipedia/commons/thumb/c/cf/Wikipedia_%22W%22_Rounded_Black.svg/240px-Wikipedia_%22W%22_Rounded_Black.svg.png" style="width: 50px; height:50px;display: inline-block;" /><img src="https://upload.wikimedia.org/wikipedia/commons/thumb/6/67/Wikipedia_%22W%22_Square_Tech.svg/240px-Wikipedia_%22W%22_Square_Tech.svg.png" style="width: 50px; height:50px;display: inline-block;" /><img src="https://upload.wikimedia.org/wikipedia/commons/thumb/a/aa/Wikipedia_%22W%22_Rounded_Tech.svg/240px-Wikipedia_%22W%22_Rounded_Tech.svg.png" style="width: 50px; height:50px;display: inline-block;" /><img src="https://upload.wikimedia.org/wikipedia/commons/thumb/2/23/Wikipedia_%22W%22_Square_White.svg/240px-Wikipedia_%22W%22_Square_White.svg.png" style="width: 50px; height:50px;display: inline-block;" /><img src="https://upload.wikimedia.org/wikipedia/commons/thumb/2/23/Wikipedia_%22W%22_Rounded_White.svg/240px-Wikipedia_%22W%22_Rounded_White.svg.png" style="width: 50px; height:50px;display: inline-block;" />
[W Wordmarks](https://foundation.wikimedia.org/wiki/Legal:Visual_identity_guidelines): The distinctive wordmark, with larger case W and A characters. The wordmark alone is the most identifiable and clearly understood version of all of the official marks representing Wikipedia.
![](https://huggingface.co/datasets/wikimedia/structured-wikipedia/resolve/main/images/wikipedia1.png)![](https://huggingface.co/datasets/wikimedia/structured-wikipedia/resolve/main/images/wikipedia2.png)
Wikimedia Product Icons: You can find all currently available icons in the [Assets library](https://www.figma.com/file/1lT9LKOK6wiHLnpraMjP3E/%E2%9D%96-Assets-(Icons%2C-Logos%2C-Illustrations)?node-id=3295-13631&t=XsJ03mZaUOTNMw9j-0) in Figma. We provide a listing of all icons with their ids for implementation in the [Codex demo](https://doc.wikimedia.org/codex/latest/icons/all-icons.html). Additionally you also find all icons as single [SVGO production optimized](https://www.mediawiki.org/wiki/Manual:Coding_conventions/SVG) SVGs for usage outside of MediaWiki.
<img src="https://huggingface.co/datasets/wikimedia/structured-wikipedia/resolve/main/images/svgs.png" style="width: 400px; height:79px"/>
More logo assets and guidance can be found on the [Wikipedia](https://foundation.wikimedia.org/wiki/Legal:Visual_identity_guidelines) and [Wikimedia](https://meta.wikimedia.org/wiki/Brand) Brand portals.
Wikipedia favicons can be found on [Wikimedia Commons](https://commons.wikimedia.org/wiki/Category:Wikimedia_Attribution_Guide_Favicons).
To schedule a 30-minute brand attribution walkthrough or to request a customized solution, please email us at [brandattribution@wikimedia.org](mailto:brandattribution@wikimedia.org)
### Licensing Information
All original textual content is licensed under the [GNU Free Documentation License](https://www.gnu.org/licenses/fdl-1.3.html) (GFDL) and the [Creative Commons Attribution-Share-Alike 4.0 License](https://creativecommons.org/licenses/by-sa/4.0/). Some text may be available only under the Creative Commons license; see the Wikimedia [Terms of Use](https://foundation.wikimedia.org/wiki/Policy:Terms_of_Use) for details. Text written by some authors may be released under additional licenses or into the public domain.
Derived dataset prepared by[ Wikimedia Enterprise](https://enterprise.wikimedia.com/). Wikipedia and Wikimedia Enterprise are registered trademarks of the[ Wikimedia Foundation, Inc.](https://www.wikimediafoundation.org/), a non-profit organization.
For more information, see [dumps licensing information](https://dumps.wikimedia.org/legal.html) and [Wikimedia Enterprise Terms](https://enterprise.wikimedia.com/terms/).
### Citation Information
```
@ONLINE{structured-wikipedia,
author = {Wikimedia Enterprise, Wikimedia Foundation},
title = {Structured Wikipedia},
month = {sep},
year= {2024}
}
``` |