Commit
·
53bd002
1
Parent(s):
fff9c79
Support streaming xtreme dataset for udpos config (#4131)
Browse files* Support streaming xtreme dataset for udpos config
* Fix style
Commit from https://github.com/huggingface/datasets/commit/f87f79a864775d8f9e779073d8bace6c66230f60
xtreme.py
CHANGED
@@ -2,7 +2,6 @@
|
|
2 |
|
3 |
|
4 |
import csv
|
5 |
-
import glob
|
6 |
import json
|
7 |
import os
|
8 |
import textwrap
|
@@ -477,34 +476,7 @@ class Xtreme(datasets.GeneratorBasedBuilder):
|
|
477 |
features["gold_label"] = datasets.Value("string")
|
478 |
|
479 |
if self.config.name.startswith("udpos"):
|
480 |
-
features =
|
481 |
-
{
|
482 |
-
"tokens": datasets.Sequence(datasets.Value("string")),
|
483 |
-
"pos_tags": datasets.Sequence(
|
484 |
-
datasets.features.ClassLabel(
|
485 |
-
names=[
|
486 |
-
"ADJ",
|
487 |
-
"ADP",
|
488 |
-
"ADV",
|
489 |
-
"AUX",
|
490 |
-
"CCONJ",
|
491 |
-
"DET",
|
492 |
-
"INTJ",
|
493 |
-
"NOUN",
|
494 |
-
"NUM",
|
495 |
-
"PART",
|
496 |
-
"PRON",
|
497 |
-
"PROPN",
|
498 |
-
"PUNCT",
|
499 |
-
"SCONJ",
|
500 |
-
"SYM",
|
501 |
-
"VERB",
|
502 |
-
"X",
|
503 |
-
]
|
504 |
-
)
|
505 |
-
),
|
506 |
-
}
|
507 |
-
)
|
508 |
|
509 |
if self.config.name.startswith("PAN-X"):
|
510 |
features = datasets.Features(
|
@@ -676,95 +648,7 @@ class Xtreme(datasets.GeneratorBasedBuilder):
|
|
676 |
),
|
677 |
]
|
678 |
if self.config.name.startswith("udpos"):
|
679 |
-
|
680 |
-
data_dir = os.path.join(udpos_downloaded_files, "ud-treebanks-v2.5")
|
681 |
-
|
682 |
-
lang = self.config.name.split(".")[1]
|
683 |
-
data_dir = os.path.join(data_dir, "*_" + lang + "*")
|
684 |
-
folders = sorted(glob.glob(data_dir))
|
685 |
-
|
686 |
-
if lang == "Kazakh":
|
687 |
-
return [
|
688 |
-
datasets.SplitGenerator(
|
689 |
-
name=datasets.Split.TEST,
|
690 |
-
# These kwargs will be passed to _generate_examples
|
691 |
-
gen_kwargs={
|
692 |
-
"filepath": [
|
693 |
-
os.path.join(folder, file)
|
694 |
-
for folder in folders
|
695 |
-
for file in sorted(os.listdir(folder))
|
696 |
-
if "test" in file and file.endswith(".conllu")
|
697 |
-
]
|
698 |
-
},
|
699 |
-
),
|
700 |
-
datasets.SplitGenerator(
|
701 |
-
name=datasets.Split.TRAIN,
|
702 |
-
# These kwargs will be passed to _generate_examples
|
703 |
-
gen_kwargs={
|
704 |
-
"filepath": [
|
705 |
-
os.path.join(folder, file)
|
706 |
-
for folder in folders
|
707 |
-
for file in sorted(os.listdir(folder))
|
708 |
-
if "train" in file and file.endswith(".conllu")
|
709 |
-
]
|
710 |
-
},
|
711 |
-
),
|
712 |
-
]
|
713 |
-
elif lang == "Tagalog" or lang == "Thai" or lang == "Yoruba":
|
714 |
-
return [
|
715 |
-
datasets.SplitGenerator(
|
716 |
-
name=datasets.Split.TEST,
|
717 |
-
# These kwargs will be passed to _generate_examples
|
718 |
-
gen_kwargs={
|
719 |
-
"filepath": [
|
720 |
-
os.path.join(folder, file)
|
721 |
-
for folder in folders
|
722 |
-
for file in sorted(os.listdir(folder))
|
723 |
-
if "test" in file and file.endswith(".conllu")
|
724 |
-
]
|
725 |
-
},
|
726 |
-
)
|
727 |
-
]
|
728 |
-
else:
|
729 |
-
return [
|
730 |
-
# We exclude Arabic-NYUAD which does not contains any words, only _
|
731 |
-
datasets.SplitGenerator(
|
732 |
-
name=datasets.Split.VALIDATION,
|
733 |
-
# These kwargs will be passed to _generate_examples
|
734 |
-
gen_kwargs={
|
735 |
-
"filepath": [
|
736 |
-
os.path.join(folder, file)
|
737 |
-
for folder in folders
|
738 |
-
for file in sorted(os.listdir(folder))
|
739 |
-
if "NYUAD" not in folder and "dev" in file and file.endswith(".conllu")
|
740 |
-
]
|
741 |
-
},
|
742 |
-
),
|
743 |
-
datasets.SplitGenerator(
|
744 |
-
name=datasets.Split.TEST,
|
745 |
-
# These kwargs will be passed to _generate_examples
|
746 |
-
gen_kwargs={
|
747 |
-
"filepath": [
|
748 |
-
os.path.join(folder, file)
|
749 |
-
for folder in folders
|
750 |
-
for file in sorted(os.listdir(folder))
|
751 |
-
if "NYUAD" not in folder and "test" in file and file.endswith(".conllu")
|
752 |
-
]
|
753 |
-
},
|
754 |
-
),
|
755 |
-
datasets.SplitGenerator(
|
756 |
-
name=datasets.Split.TRAIN,
|
757 |
-
# These kwargs will be passed to _generate_examples
|
758 |
-
gen_kwargs={
|
759 |
-
"filepath": [
|
760 |
-
os.path.join(folder, file)
|
761 |
-
for folder in folders
|
762 |
-
for file in sorted(os.listdir(folder))
|
763 |
-
if "NYUAD" not in folder and "train" in file and file.endswith(".conllu")
|
764 |
-
]
|
765 |
-
},
|
766 |
-
),
|
767 |
-
]
|
768 |
|
769 |
if self.config.name == "SQuAD":
|
770 |
|
@@ -808,7 +692,7 @@ class Xtreme(datasets.GeneratorBasedBuilder):
|
|
808 |
),
|
809 |
]
|
810 |
|
811 |
-
def _generate_examples(self, filepath):
|
812 |
"""Yields examples."""
|
813 |
# TODO(xtreme): Yields (key, example) tuples from the dataset
|
814 |
|
@@ -935,22 +819,7 @@ class Xtreme(datasets.GeneratorBasedBuilder):
|
|
935 |
"target_lang": "eng",
|
936 |
}
|
937 |
if self.config.name.startswith("udpos"):
|
938 |
-
|
939 |
-
with open(file, encoding="utf-8") as f:
|
940 |
-
data = csv.reader(f, delimiter="\t", quoting=csv.QUOTE_NONE)
|
941 |
-
tokens = []
|
942 |
-
pos_tags = []
|
943 |
-
for id_row, row in enumerate(data):
|
944 |
-
if len(row) >= 10 and row[1] != "_" and row[3] != "_":
|
945 |
-
tokens.append(row[1])
|
946 |
-
pos_tags.append(row[3])
|
947 |
-
if len(row) == 0 and len(tokens) > 0:
|
948 |
-
yield str(id_file) + "_" + str(id_row), {
|
949 |
-
"tokens": tokens,
|
950 |
-
"pos_tags": pos_tags,
|
951 |
-
}
|
952 |
-
tokens = []
|
953 |
-
pos_tags = []
|
954 |
if self.config.name.startswith("PAN-X"):
|
955 |
guid_index = 1
|
956 |
with open(filepath, encoding="utf-8") as f:
|
@@ -986,3 +855,82 @@ class Xtreme(datasets.GeneratorBasedBuilder):
|
|
986 |
"ner_tags": ner_tags,
|
987 |
"langs": langs,
|
988 |
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
|
3 |
|
4 |
import csv
|
|
|
5 |
import json
|
6 |
import os
|
7 |
import textwrap
|
|
|
476 |
features["gold_label"] = datasets.Value("string")
|
477 |
|
478 |
if self.config.name.startswith("udpos"):
|
479 |
+
features = UdposParser.features
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
480 |
|
481 |
if self.config.name.startswith("PAN-X"):
|
482 |
features = datasets.Features(
|
|
|
648 |
),
|
649 |
]
|
650 |
if self.config.name.startswith("udpos"):
|
651 |
+
return UdposParser.split_generators(dl_manager=dl_manager, config=self.config)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
652 |
|
653 |
if self.config.name == "SQuAD":
|
654 |
|
|
|
692 |
),
|
693 |
]
|
694 |
|
695 |
+
def _generate_examples(self, filepath=None, **kwargs):
|
696 |
"""Yields examples."""
|
697 |
# TODO(xtreme): Yields (key, example) tuples from the dataset
|
698 |
|
|
|
819 |
"target_lang": "eng",
|
820 |
}
|
821 |
if self.config.name.startswith("udpos"):
|
822 |
+
yield from UdposParser.generate_examples(config=self.config, filepath=filepath, **kwargs)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
823 |
if self.config.name.startswith("PAN-X"):
|
824 |
guid_index = 1
|
825 |
with open(filepath, encoding="utf-8") as f:
|
|
|
855 |
"ner_tags": ner_tags,
|
856 |
"langs": langs,
|
857 |
}
|
858 |
+
|
859 |
+
|
860 |
+
class UdposParser:
|
861 |
+
|
862 |
+
features = datasets.Features(
|
863 |
+
{
|
864 |
+
"tokens": datasets.Sequence(datasets.Value("string")),
|
865 |
+
"pos_tags": datasets.Sequence(
|
866 |
+
datasets.features.ClassLabel(
|
867 |
+
names=[
|
868 |
+
"ADJ",
|
869 |
+
"ADP",
|
870 |
+
"ADV",
|
871 |
+
"AUX",
|
872 |
+
"CCONJ",
|
873 |
+
"DET",
|
874 |
+
"INTJ",
|
875 |
+
"NOUN",
|
876 |
+
"NUM",
|
877 |
+
"PART",
|
878 |
+
"PRON",
|
879 |
+
"PROPN",
|
880 |
+
"PUNCT",
|
881 |
+
"SCONJ",
|
882 |
+
"SYM",
|
883 |
+
"VERB",
|
884 |
+
"X",
|
885 |
+
]
|
886 |
+
)
|
887 |
+
),
|
888 |
+
}
|
889 |
+
)
|
890 |
+
|
891 |
+
@staticmethod
|
892 |
+
def split_generators(dl_manager=None, config=None):
|
893 |
+
archive = dl_manager.download(config.data_url)
|
894 |
+
split_names = {datasets.Split.TRAIN: "train", datasets.Split.VALIDATION: "dev", datasets.Split.TEST: "test"}
|
895 |
+
split_generators = {
|
896 |
+
split: datasets.SplitGenerator(
|
897 |
+
name=split,
|
898 |
+
gen_kwargs={
|
899 |
+
"filepath": dl_manager.iter_archive(archive),
|
900 |
+
"split": split_names[split],
|
901 |
+
},
|
902 |
+
)
|
903 |
+
for split in split_names
|
904 |
+
}
|
905 |
+
lang = config.name.split(".")[1]
|
906 |
+
if lang in ["Tagalog", "Thai", "Yoruba"]:
|
907 |
+
return [split_generators["test"]]
|
908 |
+
elif lang == "Kazakh":
|
909 |
+
return [split_generators["train"], split_generators["test"]]
|
910 |
+
else:
|
911 |
+
return [split_generators["train"], split_generators["validation"], split_generators["test"]]
|
912 |
+
|
913 |
+
@staticmethod
|
914 |
+
def generate_examples(config=None, filepath=None, split=None):
|
915 |
+
lang = config.name.split(".")[1]
|
916 |
+
idx = 0
|
917 |
+
for path, file in filepath:
|
918 |
+
if f"_{lang}" in path and split in path and path.endswith(".conllu"):
|
919 |
+
# For lang other than [see below], we exclude Arabic-NYUAD which does not contains any words, only _
|
920 |
+
if lang in ["Kazakh", "Tagalog", "Thai", "Yoruba"] or "NYUAD" not in path:
|
921 |
+
lines = (line.decode("utf-8") for line in file)
|
922 |
+
data = csv.reader(lines, delimiter="\t", quoting=csv.QUOTE_NONE)
|
923 |
+
tokens = []
|
924 |
+
pos_tags = []
|
925 |
+
for id_row, row in enumerate(data):
|
926 |
+
if len(row) >= 10 and row[1] != "_" and row[3] != "_":
|
927 |
+
tokens.append(row[1])
|
928 |
+
pos_tags.append(row[3])
|
929 |
+
if len(row) == 0 and len(tokens) > 0:
|
930 |
+
yield idx, {
|
931 |
+
"tokens": tokens,
|
932 |
+
"pos_tags": pos_tags,
|
933 |
+
}
|
934 |
+
idx += 1
|
935 |
+
tokens = []
|
936 |
+
pos_tags = []
|