File size: 6,105 Bytes
ceba5f9
 
 
 
 
 
 
e240108
 
126f86f
e240108
7f1b406
c23c545
ceba5f9
1896c7c
ceba5f9
 
 
 
 
 
 
e240108
 
fa3521c
 
 
 
7f1b406
 
c23c545
 
1896c7c
4963e7a
197443a
4963e7a
ceba5f9
c23c545
 
ceba5f9
c23c545
ceba5f9
 
e240108
 
 
 
7f1b406
c23c545
4963e7a
ceba5f9
c23c545
ceba5f9
 
 
 
 
126f86f
 
ceba5f9
 
 
 
 
 
 
 
fa3521c
 
 
e240108
fa3521c
 
7f1b406
c23c545
ceba5f9
 
 
 
 
 
 
 
 
 
 
 
 
197443a
ceba5f9
197443a
ceba5f9
197443a
 
 
 
 
 
7f1b406
c23c545
4963e7a
197443a
4963e7a
197443a
4963e7a
c23c545
 
 
 
 
 
 
 
1896c7c
c23c545
 
ceba5f9
 
 
1896c7c
 
ceba5f9
1896c7c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ceba5f9
 
 
 
 
 
 
 
 
 
 
 
 
 
e240108
ceba5f9
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
---
license: apache-2.0
pretty_name: English Wiktionary Data in JSONL
language:
  - en
  - de
  - la
  - grc
  - ko
  - peo
  - akk
  - elx
  - sa
configs:
  - config_name: Wiktionary
    data_files:
    - split: German
      path: german-wiktextract-data.jsonl
    - split: Latin
      path: latin-wiktextract-data.jsonl
    - split: AncientGreek
      path: ancient-greek-wiktextract-data.jsonl
    - split: Korean
      path: korean-wiktextract-data.jsonl
    - split: OldPersian
      path: old-persian-wiktextract-data.jsonl
    - split: Akkadian
      path: akkadian-wiktextract-data.jsonl
    - split: Elamite
      path: elamite-wiktextract-data.jsonl
    - split: Sanskrit
      path: sanskrit-wiktextract-data.jsonl
  - config_name: Knowledge Graph
    data_files:
    - split: AllLanguage
      path: word-definition-graph-data.jsonl
tags:
  - Natural Language Processing
  - NLP
  - Wiktionary
  - Vocabulary
  - German
  - Latin
  - Ancient Greek
  - Korean
  - Old Persian
  - Akkadian
  - Elamite
  - Sanskrit
  - Knowledge Graph
size_categories:
  - 100M<n<1B
---

Wiktionary Data on Hugging Face Datasets
========================================

[![Hugging Face dataset badge]][Hugging Face dataset URL]

![Python Version Badge]
[![GitHub workflow status badge][GitHub workflow status badge]][GitHub workflow status URL]
[![Hugging Face sync status badge]][Hugging Face sync status URL]
[![Apache License Badge]][Apache License, Version 2.0]

[wiktionary-data]() is a sub-data extraction of the [English Wiktionary](https://en.wiktionary.org) that currently
supports the following languages:

- __Deutsch__ - German
- __Latinum__ - Latin
- __Ἑλληνική__ - Ancient Greek
- __한국어__ - Korean
- __𐎠𐎼𐎹__ - [Old Persian](https://en.wikipedia.org/wiki/Old_Persian_cuneiform)
- __𒀝𒅗𒁺𒌑(𒌝)__ - [Akkadian](https://en.wikipedia.org/wiki/Akkadian_language)
- [Elamite](https://en.wikipedia.org/wiki/Elamite_language)
- __संस्कृतम्__ - Sanskrit, or Classical Sanskrit

[wiktionary-data]() was originally a sub-module of [wilhelm-graphdb](https://github.com/QubitPi/wilhelm-graphdb). While
the dataset it's getting bigger, I noticed a wave of more exciting potentials this dataset can bring about that
stretches beyond the scope of the containing project. Therefore I decided to promote it to a dedicated module; and here
comes this repo.

The Wiktionary language data is available on 🤗 [Hugging Face Datasets][Hugging Face dataset URL].

```python
from datasets import load_dataset
dataset = load_dataset("QubitPi/wiktionary-data", split="German")
```

There are __two__ data subsets:

1. __Languages__ subset that contains the sub-data extraction of the following splits:

   - `German`
   - `Latin`
   - `AncientGreek`
   - `Korean`
   - `OldPersian`
   - `Akkadian`
   - `Elamite`
   - `Sanskrit`

2. __Graph__ subset that is useful for constructing knowledge graphs:

   - `AllLanguage`: all the languages in a giant graph

   The _Graph_ data ontology is the following:

   <div align="center">
       <img src="ontology.png" size="50%" alt="Error loading ontology.png"/>
   </div>

> [!TIP]
>
> Two words are structurally similar if and only if the two shares the same
> [stem](https://en.wikipedia.org/wiki/Word_stem)

Development
-----------

### Data Source

Although [the original Wiktionary dump](https://dumps.wikimedia.org/) is available, parsing it from scratch involves
rather complicated process. For example,
[acquiring the inflection data of most Indo-European languages on Wiktionary has already triggered some research-level efforts](https://stackoverflow.com/a/62977327).
We would probably do it in the future. At present, however, we would simply take the awesome works by
[tatuylonen](https://github.com/tatuylonen/wiktextract) which has already processed it and presented it in
[in JSONL format](https://kaikki.org/dictionary/rawdata.html). __wiktionary-data sources the data from
__raw Wiktextract data (JSONL, one object per line)__ option there.

### Environment Setup

Get the source code:

```console
git@github.com:QubitPi/wiktionary-data.git
cd wiktionary-data
```

It is strongly recommended to work in an isolated environment. Install virtualenv and create an isolated Python
environment by

```console
python3 -m pip install --user -U virtualenv
python3 -m virtualenv .venv
```

To activate this environment:

```console
source .venv/bin/activate
```

or, on Windows

```console
./venv\Scripts\activate
```

> [!TIP]
> 
> To deactivate this environment, use
> 
> ```console
> deactivate
> ```

### Installing Dependencies

```console
pip3 install -r requirements.txt
```

License
-------

The use and distribution terms for [wiktionary-data]() are covered by the [Apache License, Version 2.0].

[Apache License Badge]: https://img.shields.io/badge/Apache%202.0-F25910.svg?style=for-the-badge&logo=Apache&logoColor=white
[Apache License, Version 2.0]: https://www.apache.org/licenses/LICENSE-2.0

[Docker login command]: https://docker.qubitpi.org//reference/cli/docker/login/#options

[GitHub workflow status badge]: https://img.shields.io/github/actions/workflow/status/QubitPi/wiktionary-data/ci-cd.yaml?branch=master&style=for-the-badge&logo=github&logoColor=white&label=CI/CD
[GitHub workflow status URL]: https://github.com/QubitPi/wiktionary-data/actions/workflows/ci-cd.yaml

[Hugging Face dataset badge]: https://img.shields.io/badge/Hugging%20Face%20Dataset-wiktionary--data-FF9D00?style=for-the-badge&logo=huggingface&logoColor=white&labelColor=6B7280
[Hugging Face dataset URL]: https://huggingface.co/datasets/QubitPi/wiktionary-data

[Hugging Face sync status badge]: https://img.shields.io/github/actions/workflow/status/QubitPi/wiktionary-data/ci-cd.yaml?branch=master&style=for-the-badge&logo=github&logoColor=white&label=Hugging%20Face%20Sync%20Up
[Hugging Face sync status URL]: https://github.com/QubitPi/wiktionary-data/actions/workflows/ci-cd.yaml

[Python Version Badge]: https://img.shields.io/badge/Python-3.10-FFD845?labelColor=498ABC&style=for-the-badge&logo=python&logoColor=white