|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""A dictionary of events in Wikidata""" |
|
|
|
|
|
import json |
|
|
|
import datasets |
|
|
|
_CITATION = """\ |
|
@article{pratapa-etal-2022-multilingual, |
|
title = {Multilingual Event Linking to Wikidata}, |
|
author = {Pratapa, Adithya and Gupta, Rishubh and Mitamura, Teruko}, |
|
publisher = {arXiv}, |
|
year = {2022}, |
|
url = {https://arxiv.org/abs/2204.06535}, |
|
} |
|
""" |
|
|
|
_DESCRIPTION = """\ |
|
XLEL-WD is a multilingual event linking dataset. \ |
|
This sub-dataset contains a dictionary of events from Wikidata. \ |
|
The multilingual descriptions for Wikidata event items are taken from the corresponding Wikipedia articles. |
|
""" |
|
|
|
_HOMEPAGE = "https://github.com/adithya7/xlel-wd" |
|
|
|
_LICENSE = "CC-BY-4.0" |
|
|
|
_URLS = { |
|
"wikidata": "label_dict.jsonl", |
|
} |
|
|
|
_WIKIDATA_LANGS = [ |
|
"af", |
|
"ar", |
|
"be", |
|
"bg", |
|
"bn", |
|
"ca", |
|
"cs", |
|
"da", |
|
"de", |
|
"el", |
|
"en", |
|
"es", |
|
"fa", |
|
"fi", |
|
"fr", |
|
"he", |
|
"hi", |
|
"hu", |
|
"id", |
|
"it", |
|
"ja", |
|
"ko", |
|
"ml", |
|
"mr", |
|
"ms", |
|
"nl", |
|
"no", |
|
"pl", |
|
"pt", |
|
"ro", |
|
"ru", |
|
"si", |
|
"sk", |
|
"sl", |
|
"sr", |
|
"sv", |
|
"sw", |
|
"ta", |
|
"te", |
|
"th", |
|
"tr", |
|
"uk", |
|
"vi", |
|
"zh", |
|
] |
|
|
|
_DATASET_NAMES = [f"wikidata"] |
|
_DATASET_DESCRIPTIONS = [ |
|
f"Events from Wikidata, with text descriptions from corresponding multilingual Wikipedia." |
|
] |
|
|
|
for lang in _WIKIDATA_LANGS: |
|
_DATASET_NAMES += [f"wikidata.{lang}"] |
|
_DATASET_DESCRIPTIONS += [ |
|
f"Events from Wikidata, with text descriptions from corresponding {lang}-wikipedia." |
|
] |
|
|
|
|
|
class XlelWdDictionary(datasets.GeneratorBasedBuilder): |
|
"""A dictionary of events from Wikidata, with text descriptions from multilingual Wikipedia.""" |
|
|
|
BUILDER_CONFIGS = [ |
|
datasets.BuilderConfig( |
|
name=name, version=datasets.Version("1.0.0"), description=desc |
|
) |
|
for name, desc in zip(_DATASET_NAMES, _DATASET_DESCRIPTIONS) |
|
] |
|
|
|
|
|
DEFAULT_CONFIG_NAME = "wikidata" |
|
|
|
def _info(self): |
|
|
|
features = { |
|
"label_id": datasets.Value("string"), |
|
"label_title": datasets.Value("string"), |
|
"label_desc": datasets.Value("string"), |
|
"label_lang": datasets.Value("string"), |
|
} |
|
return datasets.DatasetInfo( |
|
description=_DESCRIPTION + self.config.description, |
|
features=datasets.Features(features), |
|
homepage=_HOMEPAGE, |
|
license=_LICENSE, |
|
citation=_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
|
|
urls = _URLS[self.config.name.split(".")[0]] |
|
data_file_path = dl_manager.download_and_extract(urls) |
|
return [ |
|
datasets.SplitGenerator( |
|
name="dictionary", |
|
gen_kwargs={"filepath": data_file_path}, |
|
) |
|
] |
|
|
|
def _generate_examples(self, filepath): |
|
|
|
_, *langs = self.config.name.split(".") |
|
with open(filepath, encoding="utf-8") as f: |
|
for key, row in enumerate(f): |
|
data = json.loads(row) |
|
|
|
|
|
if len(langs) == 0 or langs[0] == data["label_lang"]: |
|
yield key, data |
|
|