๐ Excited to announce the release of InfiMM-WebMath-40B โ the largest open-source multimodal pretraining dataset designed to advance mathematical reasoning in AI! ๐งฎโจ
With 40 billion tokens, this dataset aims for enhancing the reasoning capabilities of multimodal large language models in the domain of mathematics.
If you're interested in MLLMs, AI, and math reasoning, check out our work and dataset:
๐ ๐ ๐ Happy to share our recent work. We noticed that image resolution plays an important role, either in improving multi-modal large language models (MLLM) performance or in Sora style any resolution encoder decoder, we hope this work can help lift restriction of 224x224 resolution limit in ViT.