from __future__ import annotations from pathlib import Path import numpy as np import datasets _HF_AFFIX = { "ara": "arabic", "cmn": "mandarin", "eng": "", "deu": "german", "fra": "french", "hin": "hindi", "ita": "italian", "nld": "dutch", "pol": "polish", "por": "portuguese", "spa": "spanish", } _HF_AFFIX_REV = {v:k for k,v in _HF_AFFIX.items()} _REVISION_DICT = { "ara": "65eb7455a05cb77b3ae0c69d444569a8eee54628", "cmn": "617d3e9fccd186277297cc305f6588af7384b008", "eng": "9d2ac89df04254e5c427bcc8d61b6d6c83a1f59b", "deu": "5229a5cc475f36c08d03ca52f0ccb005705e60d2", "fra": "5d3085f2129139abc10d2b58becd4d4f2978e5d5", "hin": "e9e68e1a4db04726b9278192377049d0f9693012", "ita": "21e3d5c827cb60619a89988b24979850a7af85a5", "nld": "d622427417d37a8d74e110e6289bc29af4ba4056", "pol": "28d7098e2e5a211c4810d0a4d8deccc5889e55b6", "por": "323bdf67e0fbd3d7f8086fad0971b5bd5a62524b", "spa": "a7ea759535bb9fad6361cca151cf94a46e88edf3", } def _transform(dataset): target_cols = ["test_case", "label_gold"] new_cols = ['text', 'is_hateful'] rename_dict = dict(zip(target_cols, ["text", "is_hateful"])) dataset = dataset.rename_columns(rename_dict) keep_cols = new_cols + ["functionality"] remove_cols = [col for col in dataset["test"].column_names if col not in keep_cols] dataset = dataset.remove_columns(remove_cols) return dataset def make_dataset(): """ Load dataset from HuggingFace hub """ ds = {} for lang in _HF_AFFIX.values(): lcode = _HF_AFFIX_REV[lang] path = f'Paul/hatecheck-{lang}'.rstrip('-') dataset = datasets.load_dataset( path=path, revision=_REVISION_DICT[lcode] ) dataset = _transform(dataset) out_path = Path('..') / lcode / 'test.jsonl' dataset['test'].to_json(out_path) ds[lcode] = dataset return ds if __name__ == '__main__': dataset = make_dataset() AVG_CHAR = 0 for lang in _HF_AFFIX: AVG_CHAR += np.mean([len(x['text']) for x in dataset[lang]['test']]) print(f'avg char: {AVG_CHAR / len(_HF_AFFIX)}')