WisconsinAIVision / ViP-LLaVA

[CVPR2024] ViP-LLaVA: Making Large Multimodal Models Understand Arbitrary Visual Prompts
https://vip-llava.github.io/
Apache License 2.0
212 stars 15 forks source link

Validation loss on pretraining?[Feature request] #20

Open james20141606 opened 1 week ago

james20141606 commented 1 week ago

feature

Hi, I am trying to redo the pretrain step as you described in the readme doc. The training loss converges pretty fast. I find the logs in wandb and it turned out to be only containing the training loss. I wonder if you could add other metrics, like validation loss and perplexity.

Thanks a lot!

mu-cai commented 4 days ago

Thanks for the question. However, I do not have validation dataset incorporated during training. Feel free to try it by your own!

james20141606 commented 4 days ago

Thanks for your reply! By the way do you have validation data in the finetuning stage?

james20141606 commented 4 days ago

And I have two extra questions which I am confused with:

mu-cai commented 1 day ago
  1. Yes, LLMs's loss decrease very fast.
  2. I think either works, and all of those depends on the quality and quantity of your data!