Chatbot Arena meets multi-modality! Multi-Modality Arena allows you to benchmark vision-language models side-by-side while providing images as inputs. Supports MiniGPT-4, LLaMA-Adapter V2, LLaVA, BLIP-2, and many more!
And specify the version of the data used
Such as the ./OmniMedVQA/QA_information/Restricted-access/3D Modality.json.
The AMOS2022 dataset is used.
However, the AMOS2022 dataset has multiple versions.
Can you share the application link for these datasets?