[ECCV 2024] Official PyTorch implementation code for realizing the technical part of Mixture of All Intelligence (MoAI) to improve performance of numerous zero-shot vision language tasks.
Thanks for sharing your great work!
I have a question about the paper. As said in the paper, "'Word Embed' represents the word embedding dictionary of MLM". It seems to be a fixed module. Where did you get the module weights of Word Embed? From other MLM model? How to find it?
Thanks for sharing your great work! I have a question about the paper. As said in the paper, "'Word Embed' represents the word embedding dictionary of MLM". It seems to be a fixed module. Where did you get the module weights of Word Embed? From other MLM model? How to find it?