-
So far the only way to run inference on OtterHD I've seen is with a script from `shared_scripts/Demo_OtterHD.sh`.
This script accepts multiple args, including `--training_data_yaml=./Demo_Data.yaml`…
-
您好,在进行视频字幕生成的时候出现的问题。
推理代码:
from modelscope.pipelines import pipeline
from modelscope.utils.constant import Tasks
model_id = 'damo/multi-modal_hitea_video-captioning_base_en'
input_caption = 'h…
-
Hi, Dr. Jian:
Thanks for this video repo. I tried to reproduce the report result but still have two problems:
1. In "lavis/projects/blip2/train/caption_vatex_stage1.yaml", I gave the param…
-
When running the powershell file
docker build -t py-worker .
it fails on
#7 [2/7] ADD ./checkpoints /app/checkpoints
#7 sha256:66385eb2b9e14ebca5144f4dc9ed2f54ac7cfa4adfaee182aaf2035af0103d3b
#7 …
-
# MediaElement
* [x] Proposed
* [x] Prototype: Not Started
* [x] Implementation: Not Started
* [x] iOS Support
* [x] Android Support
* [x] macOS Support
* [x] Windows Support
* [x] U…
-
Hi authors,
Amazing paper and thanks for providing this nice code base. I have a question regarding the **finetuned model**, specifically for **video-text retrieval task**. Do you have plans to rel…
-
I reproduce the result of VL-PET. Actually, I only reproduce the result of video-text: yc2c dataset. Because my experiment conditions are poor. I only have 2080Ti GPU with 12G.
So I download the y…
-
Hi, Dr. Cho, there is an error when I run bash scripts/run.sh for inference. When implementing Moment_segmentation Prediction, I met an error of KeyError: 'video_fnames'. It seems that the list of 'vi…
-
Introduction
I think the PeerTube users and project itself would benefit significantly from having an official default instance with an associated official Patreon creator page to cover the infra…
-
Hi,
Thanks for releasing the code and data.
I have checked the provided caption data and I found that there are two additional keys in the dataset, 'title' and 'titles'.
Can you provide some expl…