WingsBrokenAngel / delving-deeper-into-the-decoder-for-video-captioning

Source code for Delving Deeper into the Decoder for Video Captioning
MIT License
38 stars 14 forks source link
decoder msr-vtt msvd professional-learning semantics state-of-the-art tensorflow video-captioning

Delving Deeper into the Decoder for Video Captioning

PRs Welcome DeepLearning Github Watchers GitHub stars GitHub forks License

Table of Contents

  1. Description
  2. Requirement
  3. Manual
  4. Results
    1. Comparison on Youtube2Text
    2. Comparison on MSR-VTT
  5. Data
  6. Citation

Description

This repository is the source code for the paper named Delving Deeper into the Decoder for Video Captioning.
The paper has been accepted by ECAI 2020. The encoder-decoder framework is the most popular paradigm for video captioning task. There still exist some non-negligible problems in the decoder of a video captioning model. We propose three methods to improve the performance of the model.

  1. A combination of variational dropout and layer normalization is embeded into semantic compositional gated recurrent unit to alleviate the problem of overfitting.
  2. A unified, flexible method is proposed to evaluate the model performance on a validation set so as to select the best checkpoint for testing.
  3. A new training strategy called professional learning is proposed which develops the strong points of a captioning model and bypasses its weaknesses.

It is demonstrated in the experiments of MSVD and MSR-VTT datasets that our model has achieved the best results evaluated by BLEU, CIDEr, METEOR and ROUGE-L metrics with significant gains of up to 11.7% on MSVD and 5% on MSR-VTT compared with the previous state-of-the-art models.


If you need more information about how to generate training, validating and testing data for the datasets, please refer to Semantics-AssistedVideoCaptioning.


Professional Learning

Requirement

  1. Python 3.6
  2. TensorFlow-GPU 1.13
  3. pycocoevalcap (Python3)
  4. NumPy

Manual

  1. Make sure you have installed all the required packages.
  2. Download files in the Data section.
  3. cd path_to_directory_of_model; mkdir saves
  4. run_model.sh is used for training or testing models. Specify the GPU you want to use by modifying CUDA_VISIBLE_DEVICES value. name will be used in the name of saved model during training. Specify the needed data paths by modifying corpus, ecores, tag and ref values. test refers to the path of the saved model which is to be tested. Do not give a parameter to test if you want to train a model.
  5. After completing the configuration of the bash file, then bash run_model.sh for training or testing.

Results

Comparison on Youtube2Text

MSVD Results

Comparison on MSR-VTT

MSR-VTT Results


Data

MSVD

MSRVTT


Citation

@article{chen2020delving,
    title={Delving Deeper into the Decoder for Video Captioning},
    author={Haoran Chen and Jianmin Li and Xiaolin Hu},
    journal={CoRR},
        archivePrefix={arXiv},
    primaryClass={cs.CV},
    url={https://arxiv.org/abs/2001.05614},
    eprint={2001.05614},
    year={2020}
}