wiki_sentences / wiki_sentences.py
Fraser's picture
final fixes
d65a7c3
raw
history blame
1.43 kB
from __future__ import absolute_import, division, print_function
from typing import List
import datasets
import pandas as pd
DESCRIPTION = '''
A dataset of all the sentences in Wikipedia.
Filtered to only include sentences <=64 characters.
Taken from the OPTIMUS project. https://github.com/ChunyuanLI/Optimus/blob/master/download_datasets.md
'''
DOWNLOAD_URL = "https://textae.blob.core.windows.net/optimus/data/datasets/wikipedia.segmented.nltk.txt"
class WikiSentences(datasets.GeneratorBasedBuilder):
def _info(self):
return datasets.DatasetInfo(
description=DESCRIPTION,
features=datasets.Features(
{
'text': datasets.Value("string"),
}
),
homepage="https://github.com/ChunyuanLI/Optimus/blob/master/download_datasets.md",
)
def _split_generators(self, dl_manager: datasets.DownloadManager) -> List[datasets.SplitGenerator]:
path = dl_manager.download(DOWNLOAD_URL)
return [
datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": path}),
]
def _generate_examples(self, filepath):
with open(filepath, encoding="utf-8") as txt_file:
i = 0
for line in txt_file:
line = line.strip()
if line:
yield i, {"text": line}
i += 1