|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""NKJP-POS tagging dataset.""" |
|
|
|
import json |
|
from typing import List, Tuple, Dict, Generator |
|
|
|
import datasets |
|
|
|
_DESCRIPTION = """NKJP-POS tagging dataset.""" |
|
|
|
_URLS = { |
|
"train": "https://huggingface.co/datasets/clarin-pl/nkjp-pos/resolve/main/data/train.jsonl", |
|
"test": "https://huggingface.co/datasets/clarin-pl/nkjp-pos/resolve/main/data/test.jsonl", |
|
} |
|
|
|
_HOMEPAGE = "http://clip.ipipan.waw.pl/NationalCorpusOfPolish" |
|
|
|
_POS_TAGS = { |
|
'adj', |
|
'adja', |
|
'adjc', |
|
'adjp', |
|
'adv', |
|
'aglt', |
|
'bedzie', |
|
'brev', |
|
'burk', |
|
'comp', |
|
'conj', |
|
'depr', |
|
'fin', |
|
'ger', |
|
'imps', |
|
'impt', |
|
'inf', |
|
'interj', |
|
'interp', |
|
'num', |
|
'numcol', |
|
'pact', |
|
'pant', |
|
'pcon', |
|
'ppas', |
|
'ppron12', |
|
'ppron3', |
|
'praet', |
|
'pred', |
|
'prep', |
|
'qub', |
|
'siebie', |
|
'subst', |
|
'winien', |
|
'xxx' |
|
} |
|
|
|
|
|
class NKJPPOS(datasets.GeneratorBasedBuilder): |
|
VERSION = datasets.Version("1.1.0") |
|
|
|
def _info(self) -> datasets.DatasetInfo: |
|
return datasets.DatasetInfo( |
|
description=_DESCRIPTION, |
|
features=datasets.Features( |
|
{ |
|
"id": datasets.Value("string"), |
|
"tokens": datasets.Sequence(datasets.Value("string")), |
|
"pos_tags": datasets.Sequence(datasets.features.ClassLabel( |
|
names=list(_POS_TAGS), |
|
num_classes=len(_POS_TAGS) |
|
)), |
|
} |
|
), |
|
homepage=_HOMEPAGE, |
|
version=self.VERSION, |
|
) |
|
|
|
def _split_generators( |
|
self, dl_manager: datasets.DownloadManager |
|
) -> List[datasets.SplitGenerator]: |
|
urls_to_download = _URLS |
|
downloaded_files = dl_manager.download_and_extract(urls_to_download) |
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
gen_kwargs={"filepath": downloaded_files["train"]}, |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TEST, |
|
gen_kwargs={"filepath": downloaded_files["test"]}, |
|
), |
|
] |
|
|
|
@staticmethod |
|
def _clean_line(data_line: Dict): |
|
new_tokens = [] |
|
new_pos_tags = [] |
|
for token, pos_tag in zip(data_line["tokens"], data_line["pos_tags"]): |
|
if pos_tag in _POS_TAGS: |
|
new_tokens.append(token) |
|
new_pos_tags.append(pos_tag) |
|
data_line["tokens"] = new_tokens |
|
data_line["pos_tags"] = new_pos_tags |
|
assert len(data_line["tokens"]) == len(data_line["pos_tags"]) |
|
return data_line |
|
|
|
def _generate_examples( |
|
self, filepath: str |
|
) -> Generator[Tuple[str, Dict[str, str]], None, None]: |
|
with open(filepath, 'r') as f: |
|
for line in f: |
|
data_line = self._clean_line(json.loads(line)) |
|
yield data_line["id"], data_line |
|
|