-
Hi, sorry to bother you, could you please provide the generated story of your model on visual storytelling on the test set so that I can compare it with my model.
Best
-
||link|
|----|---|
|paper| [HowTo100M: Learning a Text-Video Embedding by Watching Hundred Million Narrated Video Clips](https://openaccess.thecvf.com/content_ICCV_2019/papers/Miech_HowTo100M_Learni…
-
||link|
|----|---|
|paper| [CLIP4Clip: An Empirical Study of CLIP for End to End Video Clip Retrieval](https://arxiv.org/pdf/2104.08860v2.pdf) |
|code| [papers with code](https://paperswithcode.com…
-
Dear authors, can you share the annotation files that are used for LSMDC retrieval? I am trying to reproduce the results using ViClip models but it's not clear which annotation files were used. Thanks…
-
Thank you for your outstanding work on the Storyboard20K dataset and I have recently been working with your publicly available Storyboard20K dataset.
However, it appears that for samples containin…
-
||link|
|----|---|
|paper| [Cross Modal Retrieval with Querybank Normalisation](https://arxiv.org/pdf/2112.12777v3.pdf) |
|code| [papers with code](https://paperswithcode.com/paper/cross-modal-retr…
-
Thank you to the author for sharing the open-source code. I noticed that the official training scripts have slightly different settings for different datasets. For example, the MSRVTT dataset uses `su…
-
(ms_sl) root@acd91bbbb0c2:/home/ms-sl-main# ./do_test.sh $DATASET $FEATURE $ROOTPATH $MODELDIR
2023-06-22 08:24:36.436:INFO:__main__ - Setup config, data and model...
------------ Options ----------…
-
I found the statistics in LSMDC's readme.txt is following:
= Statistics
- Training: 101,079
- Validation: 7,408
- Public Test: 10,053
- Blind Test: 9,578
But the paper shows 118081 videos and …
-
Hello, thank you very much for your work and code, I have a few questions that I am puzzled about, and I hope you can help me.
I trained the small diet many times according to the parameters you gave…