|
--- |
|
license: cc-by-4.0 |
|
dataset_info: |
|
features: |
|
- name: SAMPLE_ID |
|
dtype: int64 |
|
- name: URL |
|
dtype: string |
|
- name: TEXT |
|
dtype: string |
|
- name: HEIGHT |
|
dtype: float64 |
|
- name: WIDTH |
|
dtype: float64 |
|
- name: LICENSE |
|
dtype: string |
|
- name: NSFW |
|
dtype: string |
|
- name: similarity |
|
dtype: float64 |
|
- name: ase_scores |
|
dtype: float64 |
|
- name: kmeans |
|
dtype: int64 |
|
- name: __index_level_0__ |
|
dtype: int64 |
|
splits: |
|
- name: train |
|
num_bytes: 28506248899 |
|
num_examples: 107166507 |
|
download_size: 16353125308 |
|
dataset_size: 28506248899 |
|
configs: |
|
- config_name: default |
|
data_files: |
|
- split: train |
|
path: data/train-* |
|
--- |
|
# 100M Text Debiased Subset from LAION 2B |
|
|
|
- Captions in LAION-2B have a significant bias towards describing visual text content embedded in the images. |
|
- Released CLIP models have strong text spotting bias in almost every style of web images, resulting in the CLIP-filtering datasets inherently biased towards visual text dominant data. |
|
- CLIP models easily learn text spotting capacity from parrot captions while failing to connect the vision-language semantics, just like a text spotting parrot. |
|
|
|
For more details, please see our [paper](https://arxiv.org/abs/2312.14232). |
|
|
|
## Filtering Details |
|
|
|
We provide an alternative solution by releasing a less biased filtered LAION-2B 100M(107,166,507) subset. |
|
|
|
We construct a less biased 100M subset from the LAION-2B subset with Empty OCR results, CLIP score > 0.3, and Aesthetics score > 4.5. |
|
|
|
We add the ase_scores and K-means labels (4000 total) for each image-text pair. |
|
|
|
*We also released the dataset on [OpenDataLab](https://openxlab.org.cn/datasets/opendatalab-linyiqi/LAION-text-debiased-100M).* |
|
|
|
The pre-trained CLIP model is released on [github](https://github.com/opendatalab/CLIP-Parrot-Bias). |
|
|
|
## Reference |
|
``` |
|
@article{lin2023parrot, |
|
title={Parrot Captions Teach CLIP to Spot Text}, |
|
author={Yiqi Lin and Conghui He and Alex Jinpeng Wang and Bin Wang and Weijia Li and Mike Zheng Shou}, |
|
journal={arXiv preprint arXiv:2312.14232}, |
|
year={2023} |
|
} |
|
@misc{conghui2022opendatalab, |
|
author={He, Conghui and Li, Wei and Jin, Zhenjiang and Wang, Bin and Xu, Chao and Lin, Dahua}, |
|
title={OpenDataLab: Empowering General Artificial Intelligence with Open Datasets}, |
|
howpublished = {\url{https://opendatalab.com}}, |
|
year={2022} |
|
} |
|
``` |
|
|