Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
Dask
License:
VictorSanh commited on
Commit
892d4f5
1 Parent(s): b17df95

date of CC

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -56,7 +56,7 @@ dataset_info:
56
  - **Paper: OBELISC: An Open Web-Scale Filtered Dataset of Interleaved Image-Text Documents**
57
  - **Point of Contact: hugo@huggingface.co**
58
 
59
- `OBELISC` is an open, massive and curated collection of interleaved image-text web documents, containing 141M English documents, 115B text tokens and 353M images, extracted from Common Crawl. The collection and filtering steps are described in our [paper](https://huggingface.co/papers/2306.16527).
60
 
61
  Interleaved image-text web documents are a succession of text paragraphs interleaved by images, such as web pages that contain images. Models trained on these web documents outperform vision and language models trained solely on image-text pairs on various benchmarks. They can also generate long and coherent text about a set of multiple images. As an example, we trained [IDEFICS](https://huggingface.co/HuggingFaceM4/idefics-80b), a visual language model that accepts arbitrary sequences of image and text inputs and produces text outputs.
62
 
 
56
  - **Paper: OBELISC: An Open Web-Scale Filtered Dataset of Interleaved Image-Text Documents**
57
  - **Point of Contact: hugo@huggingface.co**
58
 
59
+ `OBELISC` is an open, massive and curated collection of interleaved image-text web documents, containing 141M English documents, 115B text tokens and 353M images, extracted from Common Crawl dumps between February 2020 and February 2023. The collection and filtering steps are described in our [paper](https://huggingface.co/papers/2306.16527).
60
 
61
  Interleaved image-text web documents are a succession of text paragraphs interleaved by images, such as web pages that contain images. Models trained on these web documents outperform vision and language models trained solely on image-text pairs on various benchmarks. They can also generate long and coherent text about a set of multiple images. As an example, we trained [IDEFICS](https://huggingface.co/HuggingFaceM4/idefics-80b), a visual language model that accepts arbitrary sequences of image and text inputs and produces text outputs.
62