pochih / Video-Cap

🎬 Video Captioning: ICCV '15 paper implementation
48 stars 20 forks source link
attention-mechanism computer-vision deep-learning nlp seq2seq tnesorflow video-captioning

Open Source Love

Video-Captioning

Image src

performance

method BLEU@1 score
seq2seq* 0.28

*seq2seq is the reproduction of paper's model

run the code

pip install -r requirements.txt
./run.sh data/testing_id.txt data/test_features

for details, run.sh needs two parameters

./run.sh <video_id_file> <path_to_video_features>

a txt file with video id

you can use data/testing_id.txt for convience

a path contains video features, each video feature should be a *.npy file

take a look at data/test_features

you can use "data/test_features" directory for convience

train the code

pip install -r requirements.txt
./train.sh

test the code

./test.sh <path_to_model>

the path to trained model

type "models/model-2380" to use pre-trained model

Environment

Author

Po-Chih Huang / @pochih