Chatbot Arena meets multi-modality! Multi-Modality Arena allows you to benchmark vision-language models side-by-side while providing images as inputs. Supports MiniGPT-4, LLaMA-Adapter V2, LLaVA, BLIP-2, and many more!
I run the scripts on ScienceQA but it raises error:
'''
File "./Multi-Modality-Arena/LVLM_evaluation/task_datasets/vqa_datasets.py", line 140, in load_save_dataset
self.image_list.append(sample['image'].convert('RGB'))
^^^^^^^^^^^^^^^^^^^^^^^
AttributeError: 'dict' object has no attribute 'convert'
'''
I run the scripts on ScienceQA but it raises error: ''' File "./Multi-Modality-Arena/LVLM_evaluation/task_datasets/vqa_datasets.py", line 140, in load_save_dataset self.image_list.append(sample['image'].convert('RGB')) ^^^^^^^^^^^^^^^^^^^^^^^ AttributeError: 'dict' object has no attribute 'convert' '''