-
Dear author, I hope you have a good day.
With your deep help, I almost succeeded to train, but, I think that there is another issues on it.
Before start begging your help again, I commnet my e…
-
Thank you for sharing this code.
I have reproduced pre-training by following GitHub script.
I got below results which are much lower than those of the paper.
(zero-shot results on MSRVTT with …
-
Hi, I evaluated the CLIP zero-shot result on the MSRVTT dataset used your evaluation code (the dot production using text_embeds and vid_embeds (torch.mean on dim=1 ([N, T, H] --> [N, H])), and your si…
-
Hi Shvedova,
Congratulations on your team's paper being accepted by CVPR. I have a slight question. What hardware resources do you need to run your model and data? Because I have three 2080Ti and it …
-
When I directly finetune CLIP4CLIP on msrvtt, I get NaN loss after about 100 iters.
After checking the log, the reason may be that some missed videos in msr-vtt produce many zero input tensors.
T…
-
Hello, thank you for the code of MCQ! We utilize the released weights and follow the data settings, trying to reproduce MSRVTT ZS results. But our result(R@1) is about four points lower than the repor…
-
Can you provide your code about the video feature extraction? Nice Work in CVPR 😄
-
Really great job, but how can I get the _MSVD_splits_ files : **MSVD_test_list.pkl MSVD_train_list.pkl
MSVD_valid_list.pkl**
I get this error while running **main.py**
and the …
-
I notice that butd output a 'npz' file corresponding to a single image.
When i want to extract video caption based on xmodaler, it requires a global video feature.
How to extract the final video f…
-
Hi,
when I reproduced the test process, I found that the results were inconsistent with the results reported in your paper. Especially on MSVD dataset, there was a large gap.
When I use **test.yam…