-
Hello, the following problems occurred while the code was running, but I am not sure where this/img/msrvtt is or where the address was passed. In the initial get_ video_ retrieval_ args() function als…
-
We are unable to reproduce your experimental results on the MSRVTT QA dataset, with an accuracy rate of around 31. May I ask for the reason, or can you provide some checkpoints
-
Can you give me a link to obtain MSRVTT traininglabel.json in your code ?
-
Hi, nice work!
Do you have a plan to release the evaluation code of SHOW-1 in UCF-101 and MSRVTT? If you can open source the evaluation code, I believe that future work can be fairly compared to sh…
-
Thank you for your excellent work! I'd like to express my gratitude for your efforts in contributing to open-source data and models. I encountered a minor issue when loading a dataset from Hugging Fac…
-
According to the ReadMe at [https://github.com/OpenGVLab/InternVideo/tree/main/InternVideo1/Downstream/Video-Text-Retrieval](url), the zero-shot retrieval results will be obtained after running the co…
-
Hi! Thank you for your amazing work! I want to ask how can I generate the _msrvtt_train_evalscores.pkl_? I see the shape is [6513,20], so which caption is respectively corresponding to the every 20 sc…
-
Hello, I want to reproduce your code results, but many imported packages are missing in the code header, such as YouTube_ dataloader、youcook_ dataloader、msrvtt_ Dataloader、lsmdc_dataloader、model_kmean…
-
I had some trouble reproducing InstructBlip model results on the msvd_qa and msrvtt_qa datasets. Could you please tell me what prompt template and hyperparameters were used for these datasets ? It wou…
-
### Discussion
Hello, esteemed LLaVA developer, thank you for contributing such robust code and data to the community.
We have extended LLaVA to [Video-LLaVA](https://github.com/PKU-YuanGroup/Vide…