-
Thank you for your contribution.
Under the huggingface `lmms-lab/LLaVA-OneVision-Data` repo, I find that there are only single-image data, and in your `scripts/train/README.md`, you say that the vide…
-
Thanks for your awesome work.
I have two questions about running the code on the YouCook2 dataset.
(1) It seems that to run on the YouCook2 dataset, we need to trim the downloaded YC2 videos with …
-
Hi,
I am unable to evaluate SwinBert on the YouCook2 dataset due to missing coco captions files. The evaluation expects the following files:
```
validation.caption_coco_format.json
testing.ca…
-
Hi, thanks for sharing the awesome work! I am really exciting to test on your code. However, while trying to download the original videos for training datasets following the DATA.md, I faced the follo…
-
is there a way to download only cooking videos from this dataset
-
Hello, thank you very much for your outstanding work. Could you please share the feature files of the YouCook2 dataset, similar to the structure of the ActivityNet dataset?
-
hi @crodriguezo,
Thank you for sharing the code. I am trying to download the features and pretrained weights, however, the link is not working.
I would appreciate if you can help or check the li…
-
I followed the instructions in [README](https://github.com/google-research/scenic/tree/main/scenic/projects/vid2seq#training) to evaluate the released checkpoints, but I could not reproduce the result…
-
Hello, I am trying to use the model to generate captions for external .mp4 videos and I was wondering if you could give me any pointers about how one would go about it and which functions are relevant…
-
Besides, how can we prepare the data files like *.label.tsv / *.caption.tsv / *.caption.linelist.tsv to train SwinBert on our own dataset? Thank you very much ~