|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
""" |
|
DocLayNet small is a about 1% of the dataset DocLayNet (more information at https://huggingface.co/datasets/pierreguillou/DocLayNet-small) |
|
DocLayNet: A Large Human-Annotated Dataset for Document-Layout Analysis |
|
DocLayNet dataset: |
|
- DocLayNet core dataset: https://codait-cos-dax.s3.us.cloud-object-storage.appdomain.cloud/dax-doclaynet/1.0.0/DocLayNet_core.zip |
|
- DocLayNet extra dataset: https://codait-cos-dax.s3.us.cloud-object-storage.appdomain.cloud/dax-doclaynet/1.0.0/DocLayNet_extra.zip |
|
""" |
|
|
|
import json |
|
import os |
|
import base64 |
|
from PIL import Image |
|
import datasets |
|
|
|
|
|
_CITATION = """\ |
|
@article{doclaynet2022, |
|
title = {DocLayNet: A Large Human-Annotated Dataset for Document-Layout Analysis}, |
|
doi = {10.1145/3534678.353904}, |
|
url = {https://arxiv.org/abs/2206.01062}, |
|
author = {Pfitzmann, Birgit and Auer, Christoph and Dolfi, Michele and Nassar, Ahmed S and Staar, Peter W J}, |
|
year = {2022} |
|
} |
|
""" |
|
|
|
|
|
_DESCRIPTION = """\ |
|
Accurate document layout analysis is a key requirement for high-quality PDF document conversion. With the recent availability of public, large ground-truth datasets such as PubLayNet and DocBank, deep-learning models have proven to be very effective at layout detection and segmentation. While these datasets are of adequate size to train such models, they severely lack in layout variability since they are sourced from scientific article repositories such as PubMed and arXiv only. Consequently, the accuracy of the layout segmentation drops significantly when these models are applied on more challenging and diverse layouts. In this paper, we present \textit{DocLayNet}, a new, publicly available, document-layout annotation dataset in COCO format. It contains 80863 manually annotated pages from diverse data sources to represent a wide variability in layouts. For each PDF page, the layout annotations provide labelled bounding-boxes with a choice of 11 distinct classes. DocLayNet also provides a subset of double- and triple-annotated pages to determine the inter-annotator agreement. In multiple experiments, we provide smallline accuracy scores (in mAP) for a set of popular object detection models. We also demonstrate that these models fall approximately 10\% behind the inter-annotator agreement. Furthermore, we provide evidence that DocLayNet is of sufficient size. Lastly, we compare models trained on PubLayNet, DocBank and DocLayNet, showing that layout predictions of the DocLayNet-trained models are more robust and thus the preferred choice for general-purpose document-layout analysis. |
|
""" |
|
|
|
_HOMEPAGE = "https://developer.ibm.com/exchanges/data/all/doclaynet/" |
|
|
|
_LICENSE = "https://github.com/DS4SD/DocLayNet/blob/main/LICENSE" |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def load_image(image_path): |
|
image = Image.open(image_path).convert("RGB") |
|
w, h = image.size |
|
return image, (w, h) |
|
|
|
logger = datasets.logging.get_logger(__name__) |
|
|
|
|
|
class DocLayNetConfig(datasets.BuilderConfig): |
|
"""BuilderConfig for DocLayNet small""" |
|
|
|
def __init__(self, **kwargs): |
|
"""BuilderConfig for DocLayNet small. |
|
Args: |
|
**kwargs: keyword arguments forwarded to super. |
|
""" |
|
super(DocLayNetConfig, self).__init__(**kwargs) |
|
|
|
|
|
class DocLayNet(datasets.GeneratorSmalldBuilder): |
|
""" |
|
DocLayNet small is a about 1% of the dataset DocLayNet (more information at https://huggingface.co/datasets/pierreguillou/DocLayNet-small) |
|
DocLayNet: A Large Human-Annotated Dataset for Document-Layout Analysis |
|
DocLayNet dataset: |
|
- DocLayNet core dataset: https://codait-cos-dax.s3.us.cloud-object-storage.appdomain.cloud/dax-doclaynet/1.0.0/DocLayNet_core.zip |
|
- DocLayNet extra dataset: https://codait-cos-dax.s3.us.cloud-object-storage.appdomain.cloud/dax-doclaynet/1.0.0/DocLayNet_extra.zip |
|
""" |
|
|
|
VERSION = datasets.Version("1.1.0") |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
BUILDER_CONFIGS = [ |
|
DocLayNetConfig(name="DocLayNet", version=datasets.Version("1.0.0"), description="DocLayNeT small dataset"), |
|
] |
|
|
|
|
|
|
|
def _info(self): |
|
|
|
features = datasets.Features( |
|
{ |
|
"id": datasets.Value("string"), |
|
"texts": datasets.Sequence(datasets.Value("string")), |
|
"bboxes_block": datasets.Sequence(datasets.Sequence(datasets.Value("int64"))), |
|
"bboxes_line": datasets.Sequence(datasets.Sequence(datasets.Value("int64"))), |
|
"categories": datasets.Sequence( |
|
datasets.features.ClassLabel( |
|
names=["Caption", "Footnote", "Formula", "List-item", "Page-footer", "Page-header", "Picture", "Section-header", "Table", "Text", "Title"] |
|
) |
|
), |
|
"image": datasets.features.Image(), |
|
"pdf": datasets.Value("string"), |
|
"page_hash": datasets.Value("string"), |
|
"original_filename": datasets.Value("string"), |
|
"page_no": datasets.Value("int32"), |
|
"num_pages": datasets.Value("int32"), |
|
"original_width": datasets.Value("int32"), |
|
"original_height": datasets.Value("int32"), |
|
"coco_width": datasets.Value("int32"), |
|
"coco_height": datasets.Value("int32"), |
|
"collection": datasets.Value("string"), |
|
"doc_category": datasets.Value("string"), |
|
|
|
} |
|
) |
|
|
|
return datasets.DatasetInfo( |
|
|
|
description=_DESCRIPTION, |
|
|
|
features=features, |
|
|
|
|
|
|
|
|
|
homepage=_HOMEPAGE, |
|
|
|
license=_LICENSE, |
|
|
|
citation=_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
downloaded_file = dl_manager.download_and_extract("https://huggingface.co/datasets/pierreguillou/DocLayNet-small/resolve/main/data/dataset_small.zip") |
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
|
|
gen_kwargs={ |
|
"filepath": f"{downloaded_file}/dataset_small/train/", |
|
"split": "train", |
|
}, |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.VALIDATION, |
|
|
|
gen_kwargs={ |
|
"filepath": f"{downloaded_file}/dataset_small/val/", |
|
"split": "dev", |
|
}, |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TEST, |
|
|
|
gen_kwargs={ |
|
"filepath": f"{downloaded_file}/dataset_small/test/", |
|
"split": "test" |
|
}, |
|
), |
|
] |
|
|
|
def _generate_examples(self, filepath, split): |
|
logger.info("⏳ Generating examples from = %s", filepath) |
|
ann_dir = os.path.join(filepath, "annotations") |
|
img_dir = os.path.join(filepath, "images") |
|
pdf_dir = os.path.join(filepath, "pdfs") |
|
|
|
for guid, file in enumerate(sorted(os.listdir(ann_dir))): |
|
texts = [] |
|
bboxes_block = [] |
|
bboxes_line = [] |
|
categories = [] |
|
|
|
|
|
file_path = os.path.join(ann_dir, file) |
|
with open(file_path, "r", encoding="utf8") as f: |
|
data = json.load(f) |
|
|
|
|
|
image_path = os.path.join(img_dir, file) |
|
image_path = image_path.replace("json", "png") |
|
image, size = load_image(image_path) |
|
|
|
|
|
pdf_path = os.path.join(pdf_dir, file) |
|
pdf_path = pdf_path.replace("json", "pdf") |
|
with open(pdf_path, "rb") as pdf_file: |
|
pdf_bytes = pdf_file.read() |
|
pdf_encoded_string = base64.b64encode(pdf_bytes) |
|
|
|
|
|
page_hash = data["metadata"]["page_hash"] |
|
original_filename = data["metadata"]["original_filename"] |
|
page_no = data["metadata"]["page_no"] |
|
num_pages = data["metadata"]["num_pages"] |
|
original_width = data["metadata"]["original_width"] |
|
original_height = data["metadata"]["original_height"] |
|
coco_width = data["metadata"]["coco_width"] |
|
coco_height = data["metadata"]["coco_height"] |
|
collection = data["metadata"]["collection"] |
|
doc_category = data["metadata"]["doc_category"] |
|
|
|
for item in data["form"]: |
|
text_example, category_example, bbox_block_example, bbox_line_example = item["text"], item["category"], item["box"], item["box_line"] |
|
texts.append(text_example) |
|
categories.append(category_example) |
|
bboxes_block.append(bbox_block_example) |
|
bboxes_line.append(bbox_line_example) |
|
|
|
yield guid, {"id": str(guid), "texts": texts, "bboxes_block": bboxes_block, "bboxes_line": bboxes_line, "categories": categories, "image": image, "pdf": pdf_encoded_string, "page_hash": page_hash, "original_filename": original_filename, "page_no": page_no, "num_pages": num_pages, "original_width": original_width, "original_height": original_height, "coco_width": coco_width, "coco_height": coco_height, "collection": collection, "doc_category": doc_category} |