Commit
•
213288b
0
Parent(s):
Update files from the datasets library (from 1.0.0)
Browse filesRelease notes: https://github.com/huggingface/datasets/releases/tag/1.0.0
- .gitattributes +27 -0
- c4.py +331 -0
- c4_utils.py +489 -0
.gitattributes
ADDED
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
4 |
+
*.bin.* filter=lfs diff=lfs merge=lfs -text
|
5 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
11 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
12 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
13 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
14 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
15 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
16 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
17 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
18 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
19 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
20 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
21 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
22 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
23 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
24 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
25 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
26 |
+
*.zstandard filter=lfs diff=lfs merge=lfs -text
|
27 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
c4.py
ADDED
@@ -0,0 +1,331 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# coding=utf-8
|
2 |
+
# Copyright 2020 The TensorFlow Datasets Authors and the HuggingFace Datasets Authors.
|
3 |
+
#
|
4 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
5 |
+
# you may not use this file except in compliance with the License.
|
6 |
+
# You may obtain a copy of the License at
|
7 |
+
#
|
8 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
9 |
+
#
|
10 |
+
# Unless required by applicable law or agreed to in writing, software
|
11 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
12 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
13 |
+
# See the License for the specific language governing permissions and
|
14 |
+
# limitations under the License.
|
15 |
+
|
16 |
+
# Lint as: python3
|
17 |
+
"""C4 dataset based on Common Crawl."""
|
18 |
+
|
19 |
+
from __future__ import absolute_import, division, print_function
|
20 |
+
|
21 |
+
import json
|
22 |
+
import logging
|
23 |
+
import os
|
24 |
+
|
25 |
+
import datasets
|
26 |
+
|
27 |
+
from .c4_utils import (
|
28 |
+
dedupe_urls,
|
29 |
+
filter_by_webtextlike,
|
30 |
+
get_clean_page_fn,
|
31 |
+
get_counter_inc_fn,
|
32 |
+
get_hashed_url_filter_fn,
|
33 |
+
is_language,
|
34 |
+
is_realnews_domain,
|
35 |
+
is_valid_length,
|
36 |
+
normalize_url,
|
37 |
+
remove_duplicate_text,
|
38 |
+
split_wet_file,
|
39 |
+
)
|
40 |
+
|
41 |
+
|
42 |
+
_DESCRIPTION = """\
|
43 |
+
A colossal, cleaned version of Common Crawl's web crawl corpus.
|
44 |
+
|
45 |
+
Based on Common Crawl dataset: "https://commoncrawl.org"
|
46 |
+
|
47 |
+
Due to the overhead of cleaning the dataset, it is recommend you prepare it with
|
48 |
+
a distributed service like Cloud Dataflow. More info at
|
49 |
+
https://www.tensorflow.org/datasets/beam_datasets.
|
50 |
+
"""
|
51 |
+
_CITATION = """
|
52 |
+
@article{2019t5,
|
53 |
+
author = {Colin Raffel and Noam Shazeer and Adam Roberts and Katherine Lee and Sharan Narang and Michael Matena and Yanqi Zhou and Wei Li and Peter J. Liu},
|
54 |
+
title = {Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer},
|
55 |
+
journal = {arXiv e-prints},
|
56 |
+
year = {2019},
|
57 |
+
archivePrefix = {arXiv},
|
58 |
+
eprint = {1910.10683},
|
59 |
+
}
|
60 |
+
"""
|
61 |
+
_VERSION = datasets.Version("2.3.0", "Deduplicate lines within a page.")
|
62 |
+
|
63 |
+
_DOWNLOAD_HOST = "https://commoncrawl.s3.amazonaws.com"
|
64 |
+
_WET_PATH_URL = "https://commoncrawl.s3.amazonaws.com/crawl-data/CC-MAIN-{cc_version}/wet.paths.gz"
|
65 |
+
_REALNEWS_DOMAINS_URL = "https://raw.githubusercontent.com/rowanz/grover/38f7184bd87237ae2d3bc330b99f1e2e246f6d51/realnews/domain_to_allowed_subdomains.json"
|
66 |
+
_BADWORDS_URL = "https://raw.githubusercontent.com/LDNOOBW/List-of-Dirty-Naughty-Obscene-and-Otherwise-Bad-Words/25e679f03d96baa721cde20db9944649e8d0a844/{lang}"
|
67 |
+
_CHECKSUMS_URL = "https://storage.googleapis.com/tfds-data/manual_checksums/c4.txt"
|
68 |
+
_OPENWEBTEXT_URLS_ZIP = "OpenWebText.zip"
|
69 |
+
_OPENWEBTEXT_URLS_URL = "https://mega.nz/#F!EZZD0YwJ!9_PlEQzdMVLaNdKv_ICNVQ"
|
70 |
+
_OPENWEBTEXT_URLS_FILE_PATTERN = "OpenWebText/Version 1/URLs/*.txt"
|
71 |
+
|
72 |
+
_DEFAULT_CC_VERSIONS = ("2019-18",) # April 2019
|
73 |
+
_DEFAULT_WEBTEXTLIKE_CC_VERSIONS = ( # August 2018 - July 2019
|
74 |
+
"2018-34",
|
75 |
+
"2018-39",
|
76 |
+
"2018-43",
|
77 |
+
"2018-47",
|
78 |
+
"2018-51",
|
79 |
+
"2019-04",
|
80 |
+
"2019-09",
|
81 |
+
"2019-13",
|
82 |
+
"2019-18",
|
83 |
+
"2019-22",
|
84 |
+
"2019-26",
|
85 |
+
"2019-30",
|
86 |
+
)
|
87 |
+
|
88 |
+
|
89 |
+
class C4Config(datasets.BuilderConfig):
|
90 |
+
"""BuilderConfig for C4 dataset."""
|
91 |
+
|
92 |
+
def __init__(self, language, cc_versions=None, clean=True, realnewslike=False, webtextlike=False, **kwargs):
|
93 |
+
"""BuilderConfig for C4.
|
94 |
+
|
95 |
+
Args:
|
96 |
+
language: string, the language code, or "all" to disable language
|
97 |
+
filtering.
|
98 |
+
cc_versions: tuple(string), a collection of versions of Common Crawl to
|
99 |
+
use as the raw source text. Set to None to use defaults.
|
100 |
+
clean: bool, whether to clean the dataset for badwords, duplications, etc.
|
101 |
+
realnewslike: bool, whether to limit to news domains as compiled by
|
102 |
+
RealNews.
|
103 |
+
webtextlike: bool, whether to limit to WebText-like URLs.
|
104 |
+
**kwargs: keyword arguments forwarded to super.
|
105 |
+
"""
|
106 |
+
name_parts = [language]
|
107 |
+
if cc_versions:
|
108 |
+
name_parts.append("_".join(cc_versions))
|
109 |
+
if not clean:
|
110 |
+
name_parts.append("noclean")
|
111 |
+
if realnewslike:
|
112 |
+
name_parts.append("realnewslike")
|
113 |
+
if webtextlike:
|
114 |
+
name_parts.append("webtextlike")
|
115 |
+
name = ".".join(name_parts)
|
116 |
+
super(C4Config, self).__init__(name=name, version=_VERSION, **kwargs)
|
117 |
+
self.lang = language
|
118 |
+
self.cc_versions = cc_versions or (_DEFAULT_WEBTEXTLIKE_CC_VERSIONS if webtextlike else _DEFAULT_CC_VERSIONS)
|
119 |
+
self.clean = clean
|
120 |
+
self.realnewslike = realnewslike
|
121 |
+
self.webtextlike = webtextlike
|
122 |
+
|
123 |
+
|
124 |
+
class C4(datasets.BeamBasedBuilder):
|
125 |
+
"""C4 dataset based on Common Crawl."""
|
126 |
+
|
127 |
+
BUILDER_CONFIGS = [
|
128 |
+
C4Config(language="en", description="English C4 dataset."),
|
129 |
+
C4Config(
|
130 |
+
language="en",
|
131 |
+
clean=False,
|
132 |
+
description="Disables all cleaning (deduplication, removal based on bad words, " "etc.)",
|
133 |
+
),
|
134 |
+
C4Config(
|
135 |
+
language="en",
|
136 |
+
realnewslike=True,
|
137 |
+
description="Filters from the default config to only include content from the "
|
138 |
+
"domains used in the 'RealNews' dataset (Zellers et al., 2019).",
|
139 |
+
),
|
140 |
+
C4Config(
|
141 |
+
language="en",
|
142 |
+
webtextlike=True,
|
143 |
+
description="Filters from the default config to only include content from the "
|
144 |
+
"URLs in OpenWebText (https://github.com/jcpeterson/openwebtext).",
|
145 |
+
),
|
146 |
+
]
|
147 |
+
|
148 |
+
def manual_download_instructions(self):
|
149 |
+
return """\
|
150 |
+
For the WebText-like config, you must manually download 'OpenWebText.zip'
|
151 |
+
(from https://mega.nz/#F!EZZD0YwJ!9_PlEQzdMVLaNdKv_ICNVQ) and the Common Crawl
|
152 |
+
WET files from August 2018 to July 2019
|
153 |
+
(https://commoncrawl.org/the-data/get-started/) and place them in the
|
154 |
+
`data_dir`.
|
155 |
+
"""
|
156 |
+
|
157 |
+
def _info(self):
|
158 |
+
features = {
|
159 |
+
"text": datasets.Value("string"),
|
160 |
+
"url": datasets.Value("string"),
|
161 |
+
"content-type": datasets.Value("string"),
|
162 |
+
"content-length": datasets.Value("string"),
|
163 |
+
"timestamp": datasets.Value("string"),
|
164 |
+
}
|
165 |
+
return datasets.DatasetInfo(
|
166 |
+
description=_DESCRIPTION,
|
167 |
+
features=datasets.Features(features),
|
168 |
+
citation=_CITATION,
|
169 |
+
homepage="https://github.com/google-research/text-to-text-transfer-transformer#datasets",
|
170 |
+
)
|
171 |
+
|
172 |
+
def _split_generators(self, dl_manager, pipeline):
|
173 |
+
import apache_beam as beam
|
174 |
+
|
175 |
+
# We will automatically down the default CC version(s), but others need to
|
176 |
+
# be manually downloaded.
|
177 |
+
cc_versions = set(self.config.cc_versions)
|
178 |
+
auto_cc_versions = cc_versions & set(_DEFAULT_CC_VERSIONS)
|
179 |
+
manual_cc_versions = cc_versions - set(_DEFAULT_CC_VERSIONS)
|
180 |
+
|
181 |
+
files_to_download = {}
|
182 |
+
files_to_download["wet_path_urls"] = [
|
183 |
+
_WET_PATH_URL.format(cc_version=cc_version) for cc_version in auto_cc_versions
|
184 |
+
]
|
185 |
+
if self.config.clean:
|
186 |
+
files_to_download["badwords"] = _BADWORDS_URL.format(lang=self.config.lang)
|
187 |
+
if self.config.realnewslike:
|
188 |
+
files_to_download["realnews_domains"] = _REALNEWS_DOMAINS_URL
|
189 |
+
file_paths = dl_manager.download_and_extract(files_to_download)
|
190 |
+
|
191 |
+
if self.config.webtextlike:
|
192 |
+
owt_path = os.path.join(dl_manager.manual_dir, _OPENWEBTEXT_URLS_ZIP)
|
193 |
+
if not os.path.exists(owt_path):
|
194 |
+
raise FileNotFoundError(
|
195 |
+
"{} does not exist. Make sure you insert a manual dir via `datasets.load_dataset('c4', data_dir=...)` that includes a file name {}. Manual download instructions: {})".format(
|
196 |
+
owt_path, _OPENWEBTEXT_URLS_ZIP, self.manual_download_instructions
|
197 |
+
)
|
198 |
+
)
|
199 |
+
file_paths["openwebtext_urls_zip"] = dl_manager.extract(owt_path)
|
200 |
+
|
201 |
+
wet_urls = []
|
202 |
+
for wet_path_url in file_paths["wet_path_urls"]:
|
203 |
+
with open(wet_path_url, "r", encoding="utf-8") as f:
|
204 |
+
wet_urls.extend(["%s/%s" % (_DOWNLOAD_HOST, l.strip()) for l in f])
|
205 |
+
file_paths["wet_urls"] = wet_urls
|
206 |
+
file_paths["wet_files"] = []
|
207 |
+
|
208 |
+
for cc_version in manual_cc_versions:
|
209 |
+
cc_dir = os.path.join(dl_manager.manual_dir, cc_version)
|
210 |
+
wet_files = beam.io.filesystems.FileSystems.match(os.path.join(cc_dir, "*.warc.wet.gz"))
|
211 |
+
if not os.path.exists(cc_dir):
|
212 |
+
raise FileNotFoundError(
|
213 |
+
"{} does not exist. Make sure you insert a manual dir via `datasets.load_dataset('c4', data_dir=...)` that includes the files {}. Manual download instructions: {})".format(
|
214 |
+
cc_dir, "*.warc.wet.gz", self.manual_download_instructions
|
215 |
+
)
|
216 |
+
)
|
217 |
+
logging.info("Adding %d WET files for manually downloaded version %s.", len(wet_files), cc_version)
|
218 |
+
file_paths["wet_files"].extend(wet_files)
|
219 |
+
|
220 |
+
page_content_pcollection = self._get_page_content(pipeline, file_paths, dl_manager)
|
221 |
+
return [
|
222 |
+
datasets.SplitGenerator(
|
223 |
+
name=datasets.Split.TRAIN,
|
224 |
+
gen_kwargs=dict(
|
225 |
+
split="train",
|
226 |
+
page_content=page_content_pcollection,
|
227 |
+
hashed_url_predicate=lambda x: x % 1000 != 0, # 99.9%
|
228 |
+
),
|
229 |
+
),
|
230 |
+
datasets.SplitGenerator(
|
231 |
+
name=datasets.Split.VALIDATION,
|
232 |
+
gen_kwargs=dict(
|
233 |
+
split="validation",
|
234 |
+
page_content=page_content_pcollection,
|
235 |
+
hashed_url_predicate=lambda x: x % 1000 == 0, # 0.01%
|
236 |
+
),
|
237 |
+
),
|
238 |
+
]
|
239 |
+
|
240 |
+
def _get_page_content(self, pipeline, file_paths, dl_manager):
|
241 |
+
"""Build PCollection of un-split page content."""
|
242 |
+
import apache_beam as beam
|
243 |
+
|
244 |
+
wet_file_paths = pipeline | "create_wet_files" >> beam.Create(file_paths["wet_files"])
|
245 |
+
if "wet_urls" in file_paths:
|
246 |
+
|
247 |
+
def download_url(url, downloader, pipeline):
|
248 |
+
path = downloader.download(url)
|
249 |
+
if not pipeline.is_local():
|
250 |
+
path = downloader.ship_files_with_pipeline(path, pipeline)
|
251 |
+
return path
|
252 |
+
|
253 |
+
dl_wet_file_paths = (
|
254 |
+
pipeline
|
255 |
+
| "create_wet_urls" >> beam.Create(file_paths["wet_urls"])
|
256 |
+
| beam.Map(download_url, downloader=dl_manager, pipeline=pipeline)
|
257 |
+
)
|
258 |
+
wet_file_paths = (wet_file_paths, dl_wet_file_paths) | beam.Flatten()
|
259 |
+
|
260 |
+
# Parse WET files and filter by length.
|
261 |
+
# Output: url, text
|
262 |
+
page_content = wet_file_paths | beam.FlatMap(split_wet_file) | beam.Filter(is_valid_length)
|
263 |
+
|
264 |
+
# Optionally filter for RealNews domains.
|
265 |
+
# Output: url, text
|
266 |
+
if self.config.realnewslike:
|
267 |
+
with open(file_paths["realnews_domains"], "r", encoding="utf-8") as f:
|
268 |
+
realnews_domains = json.load(f)
|
269 |
+
page_content = page_content | beam.Filter(is_realnews_domain, realnews_domains)
|
270 |
+
|
271 |
+
# Normalize and deduplicate by URL.
|
272 |
+
# Output: url, text
|
273 |
+
page_content = (
|
274 |
+
page_content
|
275 |
+
| "normalize_url" >> beam.Map(normalize_url)
|
276 |
+
| "group_url" >> beam.GroupByKey()
|
277 |
+
| beam.Map(dedupe_urls)
|
278 |
+
)
|
279 |
+
|
280 |
+
# Optionally filter for WebText-like URLs.
|
281 |
+
# Output: url, text
|
282 |
+
if self.config.webtextlike:
|
283 |
+
webtextlike_urls = (
|
284 |
+
pipeline
|
285 |
+
| "read_webtextlike_urls"
|
286 |
+
>> beam.io.ReadFromText(
|
287 |
+
os.path.join(file_paths["openwebtext_urls_zip"], _OPENWEBTEXT_URLS_FILE_PATTERN)
|
288 |
+
)
|
289 |
+
| "add_dummy_page" >> beam.Map(lambda x: (x, ""))
|
290 |
+
| "normal_webtext_url" >> beam.Map(normalize_url)
|
291 |
+
)
|
292 |
+
page_content = (
|
293 |
+
{"text": page_content, "webtextlike_urls": webtextlike_urls}
|
294 |
+
| "group_webtextlike_urls" >> beam.CoGroupByKey()
|
295 |
+
| beam.FlatMap(filter_by_webtextlike)
|
296 |
+
)
|
297 |
+
|
298 |
+
# Optionally clean pages of badwords, boilerpolate text, and duplicate
|
299 |
+
# spans of sentences.
|
300 |
+
# Output: url, text
|
301 |
+
if self.config.clean:
|
302 |
+
with open(file_paths["badwords"], "r", encoding="utf-8") as f:
|
303 |
+
badwords = [l.strip() for l in f]
|
304 |
+
page_content = page_content | "clean_pages" >> beam.FlatMap(get_clean_page_fn(badwords))
|
305 |
+
page_content = remove_duplicate_text(page_content)
|
306 |
+
|
307 |
+
# Optionally filter out non-`language` pages. We do this after cleaning
|
308 |
+
# since it may change the predominate language.
|
309 |
+
if self.config.lang != "all":
|
310 |
+
page_content |= beam.Filter(is_language, language=self.config.lang)
|
311 |
+
|
312 |
+
return page_content
|
313 |
+
|
314 |
+
def _build_pcollection(self, unused_pipeline, split, page_content, hashed_url_predicate):
|
315 |
+
import apache_beam as beam
|
316 |
+
|
317 |
+
def _emit_examples(el):
|
318 |
+
get_counter_inc_fn(split)("examples")
|
319 |
+
_, features = el
|
320 |
+
return (
|
321 |
+
features["url"],
|
322 |
+
{
|
323 |
+
"url": features["url"],
|
324 |
+
"text": features["text"],
|
325 |
+
"content-type": features["content-type"],
|
326 |
+
"content-length": features["content-length"],
|
327 |
+
"timestamp": features["timestamp"],
|
328 |
+
},
|
329 |
+
)
|
330 |
+
|
331 |
+
return page_content | beam.Filter(get_hashed_url_filter_fn(hashed_url_predicate)) | beam.Map(_emit_examples)
|
c4_utils.py
ADDED
@@ -0,0 +1,489 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# coding=utf-8
|
2 |
+
# Copyright 2020 The TensorFlow Datasets Authors and the HuggingFace Datasets Authors.
|
3 |
+
#
|
4 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
5 |
+
# you may not use this file except in compliance with the License.
|
6 |
+
# You may obtain a copy of the License at
|
7 |
+
#
|
8 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
9 |
+
#
|
10 |
+
# Unless required by applicable law or agreed to in writing, software
|
11 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
12 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
13 |
+
# See the License for the specific language governing permissions and
|
14 |
+
# limitations under the License.
|
15 |
+
|
16 |
+
# Lint as: python3
|
17 |
+
"""Utilities for generating the C4 dataset."""
|
18 |
+
|
19 |
+
from __future__ import absolute_import, division, print_function
|
20 |
+
|
21 |
+
import functools
|
22 |
+
import gzip
|
23 |
+
import hashlib
|
24 |
+
import io
|
25 |
+
import re
|
26 |
+
import threading
|
27 |
+
|
28 |
+
|
29 |
+
# WET file constants
|
30 |
+
_PAGE_DELIMITER = "WARC/1.0"
|
31 |
+
_URL_KEY = "WARC-Target-URI:"
|
32 |
+
_URL_DATE = "WARC-Date:"
|
33 |
+
_CONTENT_TYPE = "Content-Type:"
|
34 |
+
_CONTENT_LEN = "Content-Length:"
|
35 |
+
_METADATA_PREFIXES = ("WARC", "CONTENT-", "Content-")
|
36 |
+
|
37 |
+
# Filters
|
38 |
+
_MIN_WORDS_PER_LINE = 5
|
39 |
+
_MIN_NUM_SENTENCES = 3
|
40 |
+
_MAX_WORD_LENGTH = 1000
|
41 |
+
_END_MARKS = (".", "?", "!", '"')
|
42 |
+
_ELLIPSIS = "..."
|
43 |
+
_POLICY_SUBSTRINGS = [
|
44 |
+
"terms of use",
|
45 |
+
"privacy policy",
|
46 |
+
"cookie policy",
|
47 |
+
"uses cookies",
|
48 |
+
"use of cookies",
|
49 |
+
"use cookies",
|
50 |
+
]
|
51 |
+
|
52 |
+
# Memoized sentence tokenizer.
|
53 |
+
_SENTENCE_TOKENIZER = None
|
54 |
+
|
55 |
+
|
56 |
+
def get_counter_inc_fn(namespace):
|
57 |
+
import apache_beam as beam
|
58 |
+
|
59 |
+
def counter_inc_fn(counter, amt=1):
|
60 |
+
beam.metrics.Metrics.counter(namespace, counter).inc(amt)
|
61 |
+
|
62 |
+
return counter_inc_fn
|
63 |
+
|
64 |
+
|
65 |
+
def get_hashed_url_filter_fn(predicate_fn):
|
66 |
+
import tensorflow.compat.v2 as tf
|
67 |
+
|
68 |
+
def filter_fn(el):
|
69 |
+
url, _ = el
|
70 |
+
val = int(hashlib.md5(tf.compat.as_text(url).encode("utf-8")).hexdigest(), 16)
|
71 |
+
return predicate_fn(val)
|
72 |
+
|
73 |
+
return filter_fn
|
74 |
+
|
75 |
+
|
76 |
+
def _load_sentence_tokenizer():
|
77 |
+
"""Returns a sentence tokenization function."""
|
78 |
+
# Lock to avoid a race-condition in the creation of the download directory.
|
79 |
+
with threading.Lock():
|
80 |
+
import nltk
|
81 |
+
|
82 |
+
nltk.download("punkt")
|
83 |
+
return nltk.data.load("nltk:tokenizers/punkt/english.pickle")
|
84 |
+
|
85 |
+
|
86 |
+
def _get_sentences(text):
|
87 |
+
import tensorflow.compat.v2 as tf
|
88 |
+
|
89 |
+
global _SENTENCE_TOKENIZER
|
90 |
+
if not _SENTENCE_TOKENIZER:
|
91 |
+
_SENTENCE_TOKENIZER = _load_sentence_tokenizer()
|
92 |
+
return list(_SENTENCE_TOKENIZER.tokenize(tf.compat.as_text(text)))
|
93 |
+
|
94 |
+
|
95 |
+
def _get_sentences_by_line(text, lower=False):
|
96 |
+
sentences = []
|
97 |
+
for line in text.splitlines():
|
98 |
+
sentences.append([s.lower() if lower else s for s in _get_sentences(line)])
|
99 |
+
return sentences
|
100 |
+
|
101 |
+
|
102 |
+
def is_language(page, language, min_probability=0.99):
|
103 |
+
"""Returns True iff text is in `language` with at least `min_probability`."""
|
104 |
+
unused_url, features = page
|
105 |
+
text = features["text"]
|
106 |
+
|
107 |
+
counter_inc_fn = get_counter_inc_fn("detected-lang")
|
108 |
+
|
109 |
+
# Make langdetect predictions deterministic.
|
110 |
+
import langdetect
|
111 |
+
|
112 |
+
langdetect.DetectorFactory.seed = 0
|
113 |
+
try:
|
114 |
+
predictions = langdetect.detect_langs(text)
|
115 |
+
except langdetect.lang_detect_exception.LangDetectException:
|
116 |
+
counter_inc_fn("langdetect-exception")
|
117 |
+
return False
|
118 |
+
if not predictions:
|
119 |
+
counter_inc_fn("page-filtered-nolangpredictions")
|
120 |
+
return False
|
121 |
+
best_prediction = predictions[0]
|
122 |
+
if best_prediction.prob < min_probability:
|
123 |
+
counter_inc_fn("page-filtered-lowlangdetectconf")
|
124 |
+
return False
|
125 |
+
if best_prediction.lang != language:
|
126 |
+
counter_inc_fn("page-filtered-ignoredlang")
|
127 |
+
counter_inc_fn("page-filtered-ignoredlang-%s" % (best_prediction.lang))
|
128 |
+
return False
|
129 |
+
counter_inc_fn("page-emited-%s" % best_prediction.lang)
|
130 |
+
return True
|
131 |
+
|
132 |
+
|
133 |
+
def get_clean_page_fn(badwords=None):
|
134 |
+
"""Returns `clean_page` with pre-compiled badword and citation regexes."""
|
135 |
+
# Used to filter citation from Wikipedia pages (among others).
|
136 |
+
citation_regex = re.compile(r"\[\d*\]|\[edit\]|\[citation needed\]")
|
137 |
+
if badwords:
|
138 |
+
badwords_regex = re.compile("[^a-z]({})[^a-z]".format("|".join(badwords or [])))
|
139 |
+
else:
|
140 |
+
badwords_regex = None
|
141 |
+
return functools.partial(clean_page, citation_regex=citation_regex, badwords_regex=badwords_regex)
|
142 |
+
|
143 |
+
|
144 |
+
def clean_page(
|
145 |
+
url_and_features,
|
146 |
+
citation_regex,
|
147 |
+
badwords_regex=None,
|
148 |
+
counter_inc_fn=None,
|
149 |
+
min_words_per_line=_MIN_WORDS_PER_LINE,
|
150 |
+
min_num_sentences=_MIN_NUM_SENTENCES,
|
151 |
+
max_word_length=_MAX_WORD_LENGTH,
|
152 |
+
):
|
153 |
+
"""Cleans a CommonCrawl page, yielding nothing if it should be skipped.
|
154 |
+
|
155 |
+
Cleaning removes lines with no end marks or with too few words. After line
|
156 |
+
filtering, pages are filtered out if they have too few sentences based on a
|
157 |
+
simple count of end marks.
|
158 |
+
|
159 |
+
Args:
|
160 |
+
url_and_features: tuple(string, dict), the url and features of the page.
|
161 |
+
citation_regex: Regex to use for finding Wikipedia-like citations to filter.
|
162 |
+
badwords_regex: Regex to use for finding badwords. Default None, which means
|
163 |
+
don't apply badwords filtering.
|
164 |
+
counter_inc_fn: function, a function taking the name of a counter to be
|
165 |
+
incremented and the (optional) amount. Defaults to a beam Metric counter.
|
166 |
+
min_words_per_line: int, the minimum number of words a line needs to not be
|
167 |
+
removed.
|
168 |
+
min_num_sentences: int, the minimum number of sentences a page needs to not
|
169 |
+
be skipped.
|
170 |
+
max_word_length: int, the maximum number of characters allowed in a word.
|
171 |
+
Lines containing a word with too many characters are removed.
|
172 |
+
Yields:
|
173 |
+
The url and cleaned text for the page.
|
174 |
+
"""
|
175 |
+
url, features = url_and_features
|
176 |
+
text = features["text"]
|
177 |
+
|
178 |
+
if not counter_inc_fn:
|
179 |
+
counter_inc_fn = get_counter_inc_fn("clean-page")
|
180 |
+
|
181 |
+
lines = text.splitlines()
|
182 |
+
valid_lines = []
|
183 |
+
num_sentences = 0
|
184 |
+
|
185 |
+
def line_has_too_long_word(line):
|
186 |
+
for word in line.split():
|
187 |
+
if len(word) > max_word_length:
|
188 |
+
return True
|
189 |
+
return False
|
190 |
+
|
191 |
+
for line in lines:
|
192 |
+
line = line.strip()
|
193 |
+
if line_has_too_long_word(line):
|
194 |
+
counter_inc_fn("lines-with-too-long-word")
|
195 |
+
continue
|
196 |
+
line = citation_regex.sub("", line)
|
197 |
+
if not line.endswith(_END_MARKS) or line.endswith(_ELLIPSIS):
|
198 |
+
counter_inc_fn("lines-no-endmark")
|
199 |
+
continue
|
200 |
+
if len(line.split()) < min_words_per_line:
|
201 |
+
counter_inc_fn("lines-too-short")
|
202 |
+
continue
|
203 |
+
line_lower = line.lower()
|
204 |
+
# Remove documents which contain lorem ipsum
|
205 |
+
if "lorem ipsum" in line_lower:
|
206 |
+
counter_inc_fn("filtered-page-loremipsum")
|
207 |
+
return
|
208 |
+
# Remove "javascript must be enabled" notices
|
209 |
+
if "javascript" in line_lower:
|
210 |
+
counter_inc_fn("lines-javascript")
|
211 |
+
continue
|
212 |
+
# Remove docs which probably contain javascript code
|
213 |
+
if "{" in line:
|
214 |
+
counter_inc_fn("filtered-page-squigglybracket")
|
215 |
+
return
|
216 |
+
# Remove policy lines
|
217 |
+
if any(p in line_lower for p in _POLICY_SUBSTRINGS):
|
218 |
+
counter_inc_fn("lines-policy")
|
219 |
+
continue
|
220 |
+
# If any badword appears on its own in the line, skip this doc
|
221 |
+
if badwords_regex:
|
222 |
+
badwords_found = badwords_regex.search(line_lower)
|
223 |
+
if badwords_found is not None:
|
224 |
+
counter_inc_fn("filtered-page-badword")
|
225 |
+
return
|
226 |
+
num_sentences += len(_get_sentences(line))
|
227 |
+
valid_lines.append(line)
|
228 |
+
counter_inc_fn("lines-valid")
|
229 |
+
|
230 |
+
if num_sentences < min_num_sentences:
|
231 |
+
counter_inc_fn("filtered-page-toofewsentences")
|
232 |
+
return
|
233 |
+
counter_inc_fn("emitted-clean-pages")
|
234 |
+
features["text"] = "\n".join(valid_lines).strip()
|
235 |
+
yield url, features
|
236 |
+
|
237 |
+
|
238 |
+
def _hash_line(line):
|
239 |
+
import tensorflow.compat.v2 as tf
|
240 |
+
|
241 |
+
m = hashlib.md5()
|
242 |
+
m.update(tf.compat.as_text(line).encode("utf-8").strip().lower())
|
243 |
+
return m.hexdigest()
|
244 |
+
|
245 |
+
|
246 |
+
def _emit_url_to_lines(page):
|
247 |
+
"""Emits url to all (lower-cased, hashed) lines."""
|
248 |
+
url, features = page
|
249 |
+
text = features["text"]
|
250 |
+
for line in text.split("\n"):
|
251 |
+
yield _hash_line(line), url
|
252 |
+
|
253 |
+
|
254 |
+
def _emit_line_to_urls(el, counter_inc_fn):
|
255 |
+
"""Emits (hashed) line to all but one url."""
|
256 |
+
import tensorflow.compat.v2 as tf
|
257 |
+
|
258 |
+
line, urls = el
|
259 |
+
# Materialize urls as a list.
|
260 |
+
urls = list(urls)
|
261 |
+
# Hash urls and sort to have a consistent, but unbiased, selection when the
|
262 |
+
# same urls exist for multiple lines.
|
263 |
+
skip_url = min(urls, key=lambda x: hashlib.md5(tf.compat.as_text(x).encode("utf-8")).hexdigest())
|
264 |
+
for url in urls:
|
265 |
+
if url != skip_url:
|
266 |
+
yield url, line
|
267 |
+
counter_inc_fn("emitted-line-duplicate", amt=len(urls) - 1)
|
268 |
+
|
269 |
+
|
270 |
+
def _remove_lines_from_text(el, counter_inc_fn, min_num_sentences=_MIN_NUM_SENTENCES):
|
271 |
+
"""Removes matching lines from the page.
|
272 |
+
|
273 |
+
Process the result of a join containing a single value for 'features' and zero
|
274 |
+
or more values for 'lines'. Each value in 'lines' is a lower-cased, hashed
|
275 |
+
line.
|
276 |
+
|
277 |
+
If a line has fewer sentences than `max_window_size`, the full line is
|
278 |
+
compared for a match.
|
279 |
+
|
280 |
+
Args:
|
281 |
+
el: `(string, {'features': features_dict, 'lines': [string]})`,
|
282 |
+
element containing the result of a join on key with both the page text
|
283 |
+
and lower-cased, hashed lines to remove.
|
284 |
+
counter_inc_fn: function, a function taking the name of a counter to be
|
285 |
+
incremented and the (optional) amount.
|
286 |
+
min_num_sentences: int, the minimum number of sentences a page needs to not
|
287 |
+
be skipped.
|
288 |
+
|
289 |
+
Yields:
|
290 |
+
url: The URL of the page.
|
291 |
+
features: The page features with lines removed from text.
|
292 |
+
"""
|
293 |
+
url, join_values = el
|
294 |
+
features = join_values["features"]
|
295 |
+
|
296 |
+
assert len(features) == 1, "Invalid page count (%d) for %s" % (len(features), url)
|
297 |
+
features = features[0]
|
298 |
+
text = features["text"]
|
299 |
+
lines_to_remove = set(join_values["lines"])
|
300 |
+
new_lines = []
|
301 |
+
hashed_lines = set()
|
302 |
+
for line in text.split("\n"):
|
303 |
+
hashed_line = _hash_line(line)
|
304 |
+
if hashed_line in lines_to_remove:
|
305 |
+
counter_inc_fn("filtered-lines-duplicate")
|
306 |
+
elif hashed_line not in hashed_lines:
|
307 |
+
new_lines.append(line)
|
308 |
+
hashed_lines.add(hashed_line)
|
309 |
+
new_text = "\n".join(new_lines)
|
310 |
+
if len(_get_sentences(new_text)) < min_num_sentences:
|
311 |
+
counter_inc_fn("filtered-doc-toofewsentences")
|
312 |
+
return
|
313 |
+
new_features = features.copy()
|
314 |
+
new_features["text"] = new_text
|
315 |
+
yield (url, new_features)
|
316 |
+
|
317 |
+
|
318 |
+
def remove_duplicate_text(pages):
|
319 |
+
"""Utility to remove duplicate lines across text documents."""
|
320 |
+
# Output: url, lines
|
321 |
+
import apache_beam as beam
|
322 |
+
|
323 |
+
counter_inc_fn = get_counter_inc_fn("dedupe-lines")
|
324 |
+
lines_to_remove = (
|
325 |
+
pages
|
326 |
+
| beam.FlatMap(_emit_url_to_lines)
|
327 |
+
| "group_sentences" >> beam.GroupByKey()
|
328 |
+
| beam.FlatMap(_emit_line_to_urls, counter_inc_fn=counter_inc_fn)
|
329 |
+
)
|
330 |
+
|
331 |
+
# Output: url, text
|
332 |
+
final_docs = (
|
333 |
+
{"features": pages, "lines": lines_to_remove}
|
334 |
+
| "group_features_and_lines_by_url" >> beam.CoGroupByKey()
|
335 |
+
| beam.FlatMap(_remove_lines_from_text, counter_inc_fn=counter_inc_fn)
|
336 |
+
)
|
337 |
+
|
338 |
+
return final_docs
|
339 |
+
|
340 |
+
|
341 |
+
def split_wet_file(wet_file_path, counter_inc_fn=None):
|
342 |
+
"""Split a WET file into separate pages."""
|
343 |
+
from absl import logging
|
344 |
+
|
345 |
+
logging.info("Splitting file: %s", wet_file_path)
|
346 |
+
if not counter_inc_fn:
|
347 |
+
counter_inc_fn = get_counter_inc_fn("split-wet-file")
|
348 |
+
counter_inc_fn("wet-file")
|
349 |
+
|
350 |
+
import apache_beam as beam
|
351 |
+
|
352 |
+
with beam.io.filesystems.FileSystems.open(wet_file_path) as f, gzip.GzipFile(fileobj=f) as g:
|
353 |
+
url = None
|
354 |
+
content = None
|
355 |
+
content_len = None
|
356 |
+
content_type = None
|
357 |
+
timestamp = None
|
358 |
+
|
359 |
+
def _maybe_get_page():
|
360 |
+
"""Generate a (url, {features}) page."""
|
361 |
+
if not url and url is not None:
|
362 |
+
counter_inc_fn("page-filtered-nourl")
|
363 |
+
if not content and content is not None:
|
364 |
+
counter_inc_fn("page-filtered-nocontent")
|
365 |
+
if not content_type and content_type is not None:
|
366 |
+
counter_inc_fn("page-nocontenttype")
|
367 |
+
if not content_len and content_len is not None:
|
368 |
+
counter_inc_fn("page-nocontentlen")
|
369 |
+
if not timestamp and timestamp is not None:
|
370 |
+
counter_inc_fn("page-notimestamp")
|
371 |
+
if content and url:
|
372 |
+
counter_inc_fn("page-emitted")
|
373 |
+
return (
|
374 |
+
url,
|
375 |
+
{
|
376 |
+
"text": "\n".join(content),
|
377 |
+
"content-type": content_type,
|
378 |
+
"content-length": content_len,
|
379 |
+
"timestamp": timestamp,
|
380 |
+
"url": url,
|
381 |
+
},
|
382 |
+
)
|
383 |
+
return None
|
384 |
+
|
385 |
+
for line in io.TextIOWrapper(g, encoding="utf-8"):
|
386 |
+
line = line.strip()
|
387 |
+
if not line:
|
388 |
+
continue
|
389 |
+
if line == _PAGE_DELIMITER:
|
390 |
+
page = _maybe_get_page()
|
391 |
+
if page:
|
392 |
+
yield page
|
393 |
+
url = ""
|
394 |
+
content = []
|
395 |
+
content_len = ""
|
396 |
+
content_type = ""
|
397 |
+
timestamp = ""
|
398 |
+
|
399 |
+
if line.startswith(_URL_KEY):
|
400 |
+
url = line[len(_URL_KEY) :].strip()
|
401 |
+
|
402 |
+
if line.startswith(_URL_DATE):
|
403 |
+
timestamp = line[len(_URL_DATE) :].strip()
|
404 |
+
|
405 |
+
if line.startswith(_CONTENT_TYPE):
|
406 |
+
content_type = line[len(_CONTENT_TYPE) :].strip()
|
407 |
+
|
408 |
+
if line.startswith(_CONTENT_LEN):
|
409 |
+
content_len = line[len(_CONTENT_LEN) :].strip()
|
410 |
+
|
411 |
+
if line.startswith(_METADATA_PREFIXES):
|
412 |
+
continue
|
413 |
+
|
414 |
+
content.append(line)
|
415 |
+
|
416 |
+
page = _maybe_get_page()
|
417 |
+
if page:
|
418 |
+
yield page
|
419 |
+
|
420 |
+
|
421 |
+
def dedupe_urls(el):
|
422 |
+
"""Returns the first value for a given URL."""
|
423 |
+
counter_inc_fn = get_counter_inc_fn("dedupe-urls")
|
424 |
+
url, vals = el
|
425 |
+
cnt = 0
|
426 |
+
v = None
|
427 |
+
for v in vals:
|
428 |
+
cnt += 1
|
429 |
+
counter_inc_fn("filtered-url-duplicate", cnt - 1)
|
430 |
+
counter_inc_fn("unique-url")
|
431 |
+
return url, v
|
432 |
+
|
433 |
+
|
434 |
+
def is_valid_length(el, max_length=1.9e5):
|
435 |
+
"""Returns False iff page's text is too long."""
|
436 |
+
counter_inc_fn = get_counter_inc_fn("is-valid-length")
|
437 |
+
_, page = el
|
438 |
+
if len(page["text"]) > max_length:
|
439 |
+
counter_inc_fn("filtered-page-contenttoolong")
|
440 |
+
return False
|
441 |
+
counter_inc_fn("valid-length")
|
442 |
+
return True
|
443 |
+
|
444 |
+
|
445 |
+
def is_realnews_domain(el, realnews_domains):
|
446 |
+
"""Returns False iff page's (sub)domain is not allowed."""
|
447 |
+
import tldextract
|
448 |
+
|
449 |
+
counter_inc_fn = get_counter_inc_fn("is-realnews-domain")
|
450 |
+
url, _ = el
|
451 |
+
ext = tldextract.extract(url)
|
452 |
+
main_domain = ext.domain + "." + ext.suffix
|
453 |
+
if main_domain not in realnews_domains:
|
454 |
+
counter_inc_fn("filtered-url-invaliddomain")
|
455 |
+
return False
|
456 |
+
allowed_subdomains = realnews_domains[main_domain]
|
457 |
+
if isinstance(allowed_subdomains, list) and ext.subdomain not in allowed_subdomains:
|
458 |
+
counter_inc_fn("filtered-url-invalidsubdomain")
|
459 |
+
return False
|
460 |
+
counter_inc_fn("realnews-domain")
|
461 |
+
return True
|
462 |
+
|
463 |
+
|
464 |
+
def filter_by_webtextlike(el):
|
465 |
+
"""Yields only pages with a matching WebText-like URL."""
|
466 |
+
counter_inc_fn = get_counter_inc_fn("filter-by-webtextlike")
|
467 |
+
url, join_values = el
|
468 |
+
text = join_values["text"]
|
469 |
+
webtextlike = join_values["webtextlike_urls"]
|
470 |
+
if not webtextlike:
|
471 |
+
counter_inc_fn("filtered-url-notwebtextlike")
|
472 |
+
return
|
473 |
+
if not text:
|
474 |
+
counter_inc_fn("missing-webtextlike")
|
475 |
+
return
|
476 |
+
assert len(text) == 1
|
477 |
+
counter_inc_fn("found-webtextlike")
|
478 |
+
yield url, text[0]
|
479 |
+
|
480 |
+
|
481 |
+
def normalize_url(el):
|
482 |
+
import tensorflow.compat.v2 as tf
|
483 |
+
|
484 |
+
url, val = el
|
485 |
+
url = tf.compat.as_text(url)
|
486 |
+
url = re.sub(r"https?:\/\/(www\.)?", "", url)
|
487 |
+
url = re.sub(r"\?(utm_|ref|feed).*", "", url)
|
488 |
+
url = url.rstrip("/")
|
489 |
+
return url, val
|