Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
Dask
License:
OBELICS / README.md
HugoLaurencon's picture
Update README.md
6f833eb
|
raw
history blame
4.86 kB
metadata
language:
  - en
license: cc-by-4.0
size_categories:
  - 100M<n<1B
pretty_name: OBELISC
configs:
  - config_name: default
    data_files:
      - split: train
        path: data/train-*
  - config_name: opt_out_docs_removed
    data_files:
      - split: train
        path: opt_out_docs_removed/train-*
dataset_info:
  - config_name: default
    features:
      - name: images
        sequence: string
      - name: metadata
        dtype: string
      - name: general_metadata
        dtype: string
      - name: texts
        sequence: string
    splits:
      - name: train
        num_bytes: 715724717192
        num_examples: 141047697
    download_size: 71520629655
    dataset_size: 715724717192
  - config_name: opt_out_docs_removed
    features:
      - name: images
        sequence: string
      - name: metadata
        dtype: string
      - name: general_metadata
        dtype: string
      - name: texts
        sequence: string
    splits:
      - name: train
        num_bytes: 684638314215
        num_examples: 134648855
    download_size: 266501092920
    dataset_size: 684638314215

Dataset Card for OBELISC

Dataset Description

Dataset Summary

OBELISC is an open, massive and curated collection of interleaved image-text web documents, containing 141M documents, 115B text tokens and 353M images.

This dataset can be used to train large multimodal models, significantly improving their reasoning abilities compared to models trained solely on image/text pairs. Please refer to our paper for further details about the construction of the dataset, quantitative and qualitative analyses of OBELISC, and experiments we conducted.

Languages

English

Data Fields

There are 4 fields: images, texts, metadata and general_metadata.

For each example, the data in the columns images and texts are two lists of the same size, where for each index, one element and only one is not None.

For example, for the web document <image_1>text<image_2>, in images, we have [image_1,None,image_2] and in texts we have [None,text,None].

The images are replaced by their URLs, and the users have to download them themselves, for example with the library img2dataset.

In metadata, there is a string that can be transformed into a list with json.loads(example["metadata"]). This list will have the same size as the lists of images and texts, and will have a dictionary for each index where there is an image, and a None value when there is a text. This dictionary will contain the metadata of the image (original source document, unformatted source, alt-text if present, ...).

Finally, in general_metadata, there is a string that can be transformed into a dictionary, containing the URL of the document, and information about its location in the Common Crawl data.

Data Splits

There is only one split, train, that contains 141,047,697 examples.

Size

OBELISC with images replaced by their URLs weighs 666.6 GB (unwanted!) in arrow format and 377 GB in this uploaded parquet format.

Configs

The default config, downloaded when nothing is specified in the config argument, with

from datasets import load_dataset

ds = load_dataset("HuggingFaceM4/OBELISC")

corresponds to the original version of the dataset.

When building the dataset, we sent every image URL to the Spawning AI API and removed all the opted-out images. However, we noticed afterward that some images might not be opted-out, but the whole web page containing them is. This is why we created another config of the dataset to additionally filter out opted-out web pages, that can be loaded with ds = load_dataset("HuggingFaceM4/OBELISC", config_name="opt_out_docs_removed").

Visualization of OBELISC documents

https://huggingface.co/spaces/HuggingFaceM4/obelisc_visualization

Research paper

https://arxiv.org/abs/2306.16527

GitHub repository

https://github.com/huggingface/OBELISC

Terms of Use

By using the dataset, you agree to comply with the original licenses of the source content as well as the dataset license (CC-BY-4.0). Additionally, if you use this dataset to train a Machine Learning model, you agree to disclose your use of the dataset when releasing the model or an ML application using the model.

Licensing Information

License CC-BY-4.0.

Citation Information

If you are using this dataset, please cite

@inproceedings{
lauren{\c{c}}on2023obe,
title={OBELISC: An Open Web-Scale Filtered Dataset of Interleaved Image-Text Documents},
author={Hugo Lauren{\c{c}}on and Lucile Saulnier and L{\'e}o Tronchon and Stas Bekman and Amanpreet Singh and Anton Lozhkov and Thomas Wang and Siddharth Karamcheti and Alexander M Rush and Douwe Kiela and Matthieu Cord and Victor Sanh},
year={2023}
}