We propose a Lightweight Visual-Linguistic Reasoning framework (LiVLR) for VideoQA. The overview of LiVLR:
Comparison with SoTA | Trainable parameters |
---|---|
conda create -n livlr_qa python=3.6
conda activate livlr_qa
conda install -c conda-forge ffmpeg
conda install -c conda-forge scikit-video
pip install -r requirements.txt
CUDA_VISIBLE_DEVICES=1,2 python train.py --exp_name Exp_DiVS/all
Setting the correct file path and run the following code:
python test.py