--- license: apache-2.0 dataset_info: features: - name: image dtype: image - name: difficulty_in_understanding dtype: string - name: left_image dtype: string - name: overall_description dtype: string - name: right_image dtype: string - name: stage dtype: string splits: - name: train num_bytes: 142007936.672 num_examples: 1084 download_size: 139447610 dataset_size: 142007936.672 configs: - config_name: default data_files: - split: train path: data/train-* language: - en pretty_name: 'Y' size_categories: - 1K Download non-satirical images from the following Google Drive Links - > https://drive.google.com/file/d/1Tzs4OcEJK469myApGqOUKPQNUtVyTRDy/view?usp=sharing - Non-Satirical Images annotated in Stage 3 > https://drive.google.com/file/d/1i4Fy01uBZ_2YGPzyVArZjijleNbt8xRu/view?usp=sharing - Non-Satirical Images annotated in Stage 4 ### Dataset Description The YesBut dataset is a high-quality annotated dataset designed to evaluate the satire comprehension capabilities of vision-language models. It consists of 2547 multimodal images, 1084 of which are satirical, while 1463 are non-satirical. The dataset covers a variety of artistic styles, such as colorized sketches, 2D stick figures, and 3D stick figures, making it highly diverse. The satirical images follow a "Yes, But" structure, where the left half depicts a normal scenario, and the right half contains an ironic or humorous twist. The dataset is curated to challenge models in three main tasks: Satirical Image Detection, Understanding, and Completion. An additional dataset of 119 real, satirical photographs is also provided to further assess real-world satire comprehension. - **Curated by:** Annotators who met the qualification criteria of being undergraduate sophomore students or above, enrolled in English-medium colleges. - **Language(s) (NLP):** English - **License:** Apache license 2.0 ### Dataset Sources - **Repository:** https://github.com/abhi1nandy2/yesbut_dataset - **Paper:** - HuggingFace: https://huggingface.co/papers/2409.13592 - ArXiv - https://arxiv.org/abs/2409.13592 - This work has been accepted **as a long paper in EMNLP Main 2024** ## Uses ### Direct Use The YesBut dataset is intended for benchmarking the performance of vision-language models on satire and humor comprehension tasks. Researchers and developers can use this dataset to test their models on Satirical Image Detection (classifying an image as satirical or non-satirical), Satirical Image Understanding (generating the reason behind the satire), and Satirical Image Completion (choosing the correct half of an image that makes the completed image satirical). This dataset is particularly suitable for models developed for image-text reasoning and cross-modal understanding. ### Out-of-Scope Use YesBut Dataset is not recommended for tasks unrelated to multimodal understanding, such as basic image classification without the context of humor or satire. ## Dataset Structure The YesBut dataset contains two types of images: satirical and non-satirical. Satirical images are annotated with textual descriptions for both the left and right sub-images, as well as an overall description containing the punchline. Non-satirical images are also annotated but lack the element of irony or contradiction. The dataset is divided into multiple annotation stages (Stage 2, Stage 3, Stage 4), with each stage including a mix of original and generated sub-images. The data is stored in image and metadata formats, with the annotations being key to the benchmarking tasks. ## Dataset Creation ### Curation Rationale The YesBut dataset was curated to address the gap in the ability of existing vision-language models to comprehend satire and humor. Satire comprehension involves a higher level of reasoning and understanding of context, irony, and social cues, making it a challenging task for models. By including a diverse range of images and satirical scenarios, the dataset aims to push the boundaries of multimodal understanding in artificial intelligence. ### Source Data #### Data Collection and Processing The images in the YesBut dataset were collected from social media platforms, particularly from the "X" (formerly Twitter) handle @_yesbut_. The original satirical images were manually annotated, and additional synthetic images were generated using DALL-E 3. These images were then manually labeled as satirical or non-satirical. The annotations include textual descriptions, binary features (e.g., presence of text), and difficulty ratings based on the difficulty of understanding satire/irony in the images. Data processing involved adding new variations of sub-images to expand the dataset's diversity. #### Who are the source data producers? Prior to annotation and expansion of the dataset, images were downloaded from the posts in ‘X’ (erstwhile known as Twitter) handle @_yesbut_ (with proper consent). #### Annotation process The annotation process was carried out in four stages: (1) collecting satirical images from social media, (2) manual annotation of the images, (3) generating additional 2D stick figure images using DALL-E 3, and (4) generating 3D stick figure images. #### Who are the annotators? YesBut was curated by annotators who met the qualification criteria of being undergraduate sophomore students or above, enrolled in English-medium colleges. #### Personal and Sensitive Information The images in YesBut Dataset do not include any personal identifiable information, and the annotations are general descriptions related to the satirical content of the images. ## Bias, Risks, and Limitations - Subjectivity of annotations: The annotation task involves utilizing background knowledge that may differ among annotators. Consequently, we manually reviewed the annotations to minimize the number of incorrect annotations in the dataset. However, some subjectivity still remains. - Extension to languages other than English: This work is in the English Language. However, we plan to extend our work to languages other than English. ## Citation **BibTeX:** @article{nandy2024yesbut, title={YesBut: A High-Quality Annotated Multimodal Dataset for evaluating Satire Comprehension capability of Vision-Language Models}, author={Nandy, Abhilash and Agarwal, Yash and Patwa, Ashish and Das, Millon Madhur and Bansal, Aman and Raj, Ankit and Goyal, Pawan and Ganguly, Niloy}, journal={arXiv preprint arXiv:2409.13592}, year={2024} } **APA:** Nandy, A., Agarwal, Y., Patwa, A., Das, M. M., Bansal, A., Raj, A., ... & Ganguly, N. (2024). YesBut: A High-Quality Annotated Multimodal Dataset for evaluating Satire Comprehension capability of Vision-Language Models. arXiv preprint arXiv:2409.13592. ## Dataset Card Contact Get in touch at nandyabhilash@gmail.com