Documentation
This contains some really quick docs and notes on filtering with SuperWIKI NEXT.
wikipedia_soup.py
...Is the main class that handles the bulk of the filtering.
Each filter has code documentation to explain what each function generally does. So I'd suggest you to read those instead.
Usage for wikipedia_soup.py
probably the most important bit.
wikipedia_soup takes in *.ndjson
files directly from html wikipedia dumps. via the process-root
command.
Note: there are 3 publicly exposed commands via typer, process-root
, process-folder
, process-file
process-root
is probably what you want to use. It takes in the following folder structure:
dumps <- Input folder for [process-root]
|-afwiki-NS0-20240420-ENTERPRISE-HTML <- Input folder for [process-folder]
|-afwiki_namespace_0_0.ndjson <- Input file for [process-file]
|-afwiki_namespace_0_1.ndjson
|-afwiki_namespace_0_2.ndjson
...
|-arwiki-NS0-20240420-ENTERPRISE-HTML
|-arwiki_namespace_0_0.ndjson
|-arwiki_namespace_0_1.ndjson
|-arwiki_namespace_0_2.ndjson
...
... And so on...
Downloading and filtering the files is relatively easy.
- Get a list of http urls (Whichever way you prefer)
- Download said list (wget, curl, aria2c, etc)
- Extract tar files into their own folder as shown above
- Run
process-root
command. - Patience.
- ???
- Finished!
wikipedia_template.py
This file contains templates used in Wikipedia articles.
If you do need to update a template, follow these steps:
- Open the file in your web browser.
- Paste the following URL, replacing
<ID>
with the relevant Wikidata entry ID.
https://www.wikidata.org/w/api.php?action=wbgetentities&ids=<ID>&format=json&props=labels
As for the related templates:
- Stubs:
Q4663261
- Citation needed:
Q7106262
- Redirect:
Q6042392
Note: For Sections, there are currently no templates available. These must be added manually.
mediawiki_soup.py
Before the introduction of Hugging Face's Datatrove and for the sake of simpler code development, this module implemented the MediaWikiSoup
class.
This class processes HTML content into markdown format and performs additional post-processing steps on the resulting markdown.
MediaWikiSoup
leverages a "filter chain" architecture. You can extend its functionalities by adding filter functions using either add_markdown_filter
(for markdown processing) or add_soup_filter
(for BeautifulSoup processing).
html2markdown.py
Contains a customized markdownify instance. Since this is mainly carried over from 1.5, details on it are a bit hazy.
For <a>
elements, I only use the text contained. That is to say, I don't include the href.
<a href="//example.com">This is an example</a>
Will be md'd into:
This is an example
For image elements:
<img src="//example.com" alt="Alt Text"/>
Will be md'd into:
Alt Text
For <li>
elements, I'm unsure what was the reason behind it. Now, God/LLM/Model/??? only knows.
folders2jsonl.py
Is a simple script converting chunked ndjson files into 1 singular file for ease of processing.
Tools
Extra tools unrelated to main filtering. But used in some shape or form.
tools/wikipedia_eligablewiki.py
As the title says, it unbiasedly selects groups of wikipedia with high enough content. Refer to Selection of Wikipedia
for how it was computed.
the stats .json file can be fetched from the source page here: https://commons.wikimedia.org/w/index.php?title=Data:Wikipedia_statistics/data.tab&action=edit
Copy the json in the source into a .json file and you should be good to go.
if you have to filter a list of URLS, it should look like this:
Using mirror.accum.se as mirror:
https://mirror.accum.se/mirror/wikimedia.org/other/enterprise_html/runs/20240420/amiwiki-NS0-20240420-ENTERPRISE-HTML.json.tar.gz
https://mirror.accum.se/mirror/wikimedia.org/other/enterprise_html/runs/20240420/amwiki-NS0-20240420-ENTERPRISE-HTML.json.tar.gz
https://mirror.accum.se/mirror/wikimedia.org/other/enterprise_html/runs/20240420/angwiki-NS0-20240420-ENTERPRISE-HTML.json.tar.gz
https://mirror.accum.se/mirror/wikimedia.org/other/enterprise_html/runs/20240420/anwiki-NS0-20240420-ENTERPRISE-HTML.json.tar.gz
Or with the official dumps:
https://dumps.wikimedia.org/other/enterprise_html/runs/20240420/amiwiki-NS0-20240420-ENTERPRISE-HTML.json.tar.gz
https://dumps.wikimedia.org/other/enterprise_html/runs/20240420/amwiki-NS0-20240420-ENTERPRISE-HTML.json.tar.gz
https://dumps.wikimedia.org/other/enterprise_html/runs/20240420/angwiki-NS0-20240420-ENTERPRISE-HTML.json.tar.gz
https://dumps.wikimedia.org/other/enterprise_html/runs/20240420/anwiki-NS0-20240420-ENTERPRISE-HTML.json.tar.gz
tools/wikipedia_pageview.py
Not used in NEXT, but included. The idea is to accumulate all pageviews and filter each article based on pageviews. While it's a neat idea, I just didn't use it.
tools/wikipedia_mediaalias.py
Pretty sure it's unfinished. I didn't use it in the end. Similar to pageview. Though someone could improvise and use it.