PKU-YuanGroup / LanguageBind

【ICLR 2024🔥】 Extending Video-Language Pretraining to N-modality by Language-based Semantic Alignment
https://arxiv.org/abs/2310.01852
MIT License
549 stars 44 forks source link

gpu资源 #47

Open letaozhang opened 1 month ago

letaozhang commented 1 month ago

Thanks for your wonderful work. I am very excited about your idea. May I ask the computation budget used to train the largest Imagebind model? How many GPU hour do you use?

noah003 commented 2 weeks ago

same question