Added MMAlaya2 implementation based on vlmeval/vlm/internvl_chat.py.
Checked code to download model weights from https://huggingface.co/DataCanvas/MMAlaya2 and infer mmbench_test_cn_20231003.tsv, achieving A_Overall (test) score of 0.8211883408071748.
Added some packages in requirements.txt.
MMAlaya2 (LoRA-merging), with the company name still being DataCanvas, reuses other components from InternVL-Chat-V1.5. The model has 26 billion parameters, with the language model being InternLM2-20B and the vision model being InternViT-6B.
vlmeval/vlm/internvl_chat.py
.mmbench_test_cn_20231003.tsv
, achieving A_Overall (test) score of 0.8211883408071748.MMAlaya2 (LoRA-merging), with the company name still being DataCanvas, reuses other components from InternVL-Chat-V1.5. The model has 26 billion parameters, with the language model being InternLM2-20B and the vision model being InternViT-6B.