Datasets:

Modalities:
Image
Text
Formats:
parquet
ArXiv:
Libraries:
Datasets
pandas
PY007 commited on
Commit
ecd23e0
โ€ข
1 Parent(s): 8e50eeb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -1
README.md CHANGED
@@ -47,5 +47,25 @@ configs:
47
  path: data/test-*
48
  ---
49
  # Dataset Card for "MMBench_EN"
 
 
 
50
 
51
- [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
  path: data/test-*
48
  ---
49
  # Dataset Card for "MMBench_EN"
50
+ <p align="center" width="100%">
51
+ <img src="https://i.postimg.cc/g0QRgMVv/WX20240228-113337-2x.png" width="100%" height="80%">
52
+ </p>
53
 
54
+ # Large-scale Multi-modality Models Evaluation Suite
55
+
56
+ > Accelerating the development of large-scale multi-modality models (LMMs) with `lmms-eval`
57
+
58
+ ๐Ÿ  [Homepage](https://lmms-lab.github.io/) | ๐Ÿ“š [Documentation](docs/README.md) | ๐Ÿค— [Huggingface Datasets](https://huggingface.co/lmms-lab)
59
+
60
+ # This Dataset
61
+
62
+ This is a formatted version of the English subset of [MMBench](https://arxiv.org/abs/2307.06281). It is used in our `lmms-eval` pipeline to allow for one-click evaluations of large multi-modality models.
63
+
64
+ ```
65
+ @article{MMBench,
66
+ author = {Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin},
67
+ journal = {arXiv:2307.06281},
68
+ title = {MMBench: Is Your Multi-modal Model an All-around Player?},
69
+ year = {2023},
70
+ }
71
+ ```