Datasets:

Modalities:
Image
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
pandas
License:
abhi1nandy2 commited on
Commit
cd945e7
1 Parent(s): bba5166

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +24 -7
README.md CHANGED
@@ -128,16 +128,33 @@ The images in YesBut Dataset do not include any personal identifiable informatio
128
 
129
  **BibTeX:**
130
 
131
- @article{nandy2024yesbut,
132
- title={YesBut: A High-Quality Annotated Multimodal Dataset for evaluating Satire Comprehension capability of Vision-Language Models},
133
- author={Nandy, Abhilash and Agarwal, Yash and Patwa, Ashish and Das, Millon Madhur and Bansal, Aman and Raj, Ankit and Goyal, Pawan and Ganguly, Niloy},
134
- journal={arXiv preprint arXiv:2409.13592},
135
- year={2024}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
136
  }
137
 
138
- **APA:**
139
 
140
- Nandy, A., Agarwal, Y., Patwa, A., Das, M. M., Bansal, A., Raj, A., ... & Ganguly, N. (2024). YesBut: A High-Quality Annotated Multimodal Dataset for evaluating Satire Comprehension capability of Vision-Language Models. arXiv preprint arXiv:2409.13592.
141
 
142
  ## Dataset Card Contact
143
 
 
128
 
129
  **BibTeX:**
130
 
131
+ @inproceedings{nandy-etal-2024-yesbut,
132
+ title = "***{Y}es{B}ut***: A High-Quality Annotated Multimodal Dataset for evaluating Satire Comprehension capability of Vision-Language Models",
133
+ author = "Nandy, Abhilash and
134
+ Agarwal, Yash and
135
+ Patwa, Ashish and
136
+ Das, Millon Madhur and
137
+ Bansal, Aman and
138
+ Raj, Ankit and
139
+ Goyal, Pawan and
140
+ Ganguly, Niloy",
141
+ editor = "Al-Onaizan, Yaser and
142
+ Bansal, Mohit and
143
+ Chen, Yun-Nung",
144
+ booktitle = "Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing",
145
+ month = nov,
146
+ year = "2024",
147
+ address = "Miami, Florida, USA",
148
+ publisher = "Association for Computational Linguistics",
149
+ url = "https://aclanthology.org/2024.emnlp-main.937",
150
+ doi = "10.18653/v1/2024.emnlp-main.937",
151
+ pages = "16878--16895",
152
+ abstract = "Understanding satire and humor is a challenging task for even current Vision-Language models. In this paper, we propose the challenging tasks of Satirical Image Detection (detecting whether an image is satirical), Understanding (generating the reason behind the image being satirical), and Completion (given one half of the image, selecting the other half from 2 given options, such that the complete image is satirical) and release a high-quality dataset ***YesBut***, consisting of 2547 images, 1084 satirical and 1463 non-satirical, containing different artistic styles, to evaluate those tasks. Each satirical image in the dataset depicts a normal scenario, along with a conflicting scenario which is funny or ironic. Despite the success of current Vision-Language Models on multimodal tasks such as Visual QA and Image Captioning, our benchmarking experiments show that such models perform poorly on the proposed tasks on the ***YesBut*** Dataset in Zero-Shot Settings w.r.t both automated as well as human evaluation. Additionally, we release a dataset of 119 real, satirical photographs for further research.",
153
  }
154
 
155
+ **ACL:**
156
 
157
+ Abhilash Nandy, Yash Agarwal, Ashish Patwa, Millon Madhur Das, Aman Bansal, Ankit Raj, Pawan Goyal, and Niloy Ganguly. 2024. ***YesBut***: A High-Quality Annotated Multimodal Dataset for evaluating Satire Comprehension capability of Vision-Language Models. In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, pages 16878–16895, Miami, Florida, USA. Association for Computational Linguistics.
158
 
159
  ## Dataset Card Contact
160