Settings used for these benchmarks
I'm a big fan of OpenGVLab's work, particularly InternVL2, which I'm interested in using for an AI assistant project I'm developing. Currently, I'm using GPT-4o and Gemini Pro 1.5, but their closed-source nature makes them quite expensive.
I've seen the InternVL2-pro benchmark results showing it outperforming GPT-4o and Gemini Pro 1.5, which is impressive. However, when I tested the model for my specific use case, which requires strong reasoning skills and image understanding, it didn't perform as well as I expected.
I'm curious about the settings used for these benchmarks. Could you provide more information on the testing parameters and conditions?
Thank you for your work guys, I appreciate it!
Hi, thank you for your interest.
Our latest eval code is open-sourced in the InternVL repository. The specific test setup is consistent with InternVL 1.5, and you can refer to the tutorial here or here.
@czczup I tried using InternVL2-pro and it can't answer some basic questions, I'd like to share you some examples. what's the best way to reach out to you?
My email is wztxy89@163.com
@czczup I tried using InternVL2-pro and it can't answer some basic questions, I'd like to share you some examples. what's the best way to reach out to you?
My email is wztxy89@163.com
@czczup I sent an email please check :)