Open Gumpest opened 2 weeks ago
Can you try using lmsys/vicuna-7b-v1.5 as the base model?
Thanks a lot! I have trained the LLaVA with S^2 w/o Lora, and its accuracy on textvqa is higher (43.99-->45.72). However, the inference time costs 4x (08:38 --> 37:57). I wonder about that. @bfshi
Interesting. When you say the accuracy is higher and inference is slower, are you comparing to llava w/ S2 and w/ lora, or llava w/o S2? If you are comparing with llava w/ S2 and w/ Lora, then using Lora or not shouldn't affect inference speed. One possible reason is the model w/o Lora you trained tends to output longer responses than the model w/ Lora. This will make the time of answering each question in textvqa longer and increase the inference time. Maybe worth to check that if that's the reason?
It seems not work.