OpenGVLab / Ask-Anything

[CVPR2024 Highlight][VideoChatGPT] ChatGPT with video understanding! And many more supported LMs such as miniGPT4, StableLM, and MOSS.
https://vchat.opengvlab.com/
MIT License
2.86k stars 230 forks source link

Dear author, How much time does it cost to train this model? With what type of GPU cards? #93

Open zhangyuereal opened 6 months ago

Andy1621 commented 6 months ago

Thanks for your question. In our experiments, we use 32 A100 for faster training. However, if you want to fine-tune the model on your own dataset, 1 A100 or V100 is enough.

rovebot commented 4 months ago

Thanks for your question. In our experiments, we use 32 A100 for faster training. However, if you want to fine-tune the model on your own dataset, 1 A100 or V100 is enough.

Hi, how many hours do 32*A100 for total training?

adeobootpin commented 3 months ago

Knowing how many hours is very helpful (and is often left out of papers). This information allows folks that don't have access to dozens of GPUs know if they have a chance of training models in a reasonable time on one or two GPUs.

Andy1621 commented 3 months ago

Actually, for instruction tuning, the data size is often small. For example, for small data with thousands of videos, we only use 1 GPU (>40G) to train for a few hours.

However, in our paper, we hope to verify the data scale and diversity, so we collect millions of videos, which requires many more sources to train it. That's why we release different stages of models. We hope researchers can fine-tune models based on our pretraining.

Besides, in the current codebase, we do not conduct in-depth optimization for low-cost training. Researchers can follow other repos for efficient training strategies like QLoRA and low-bit training, like LAVIN, Otter or others.