-
I think it's a great work for MLLM. Nice idea.
I'm sorry but I don't find evaluation code. Do you release the evaluation code? If not, will you release it?
-
Hi, we recently finished a paper "[Towards End-to-End Embodied Decision Making via Multi-modal Large Language Model: Explorations with GPT4-Vision and Beyond](https://arxiv.org/abs/2310.02071)" ,we t…
-
I feel confused about those two branches since it seems the evaluation branch is out of sync with the main branch...
Anyone could confirm which branch I should refer to if I want to commit?
-
Thanks for your great work!
1. have you tested _instructblip-flant5_ based on _CHEF_? For the same task, why the result of flant5 is quite different from vicuna? For example, with "SRC/config/ChEF/sc…
-
Environment
GPUs: 8x4090
**Package Version**
Package Version
------------------------ ---------------------
absl-py 2.1.0
aiohttp 3.9.3
ai…
-
佬请教一下,怎么评估的aid这个分类数据集?
上图来自LHRS-Bot,下图是GeoChat,两个论文都测试了llava在aid上精度,但是差距很大。
-
### Describe the issue
Issue: As shon in this [issue](https://github.com/haotian-liu/LLaVA/issues/62), the training loss in coonvergence should be lower than 2 for `llava-vicuna-chat-hf-pretrain`. Ho…
-
Firstly, thank you for your contributions to the multi-modal large language model (MLLM) research with MiniGPT-5. I'm experiencing an issue while testing the model's image comprehension capabilities.
…
-
1. There are error in "The text-only loss corresponds to training only on training only RefinedWeb", double " training only "
2. which dataset is used when "text-only loss, w/o RefinedWeb"
3. Why…
-
I used Lora to fine tune my own dataset, but the model only replied to the content I had trained on, and I didn't know any other common sense content but Bunny-v1_0-2B-zh is ok
Do you have any train…