Codes are tested on RTX 3090 and I not sure you can get the same results on different GPUs or different python environment.
1. conda create -n PAMFN python=3.8 -y
2. conda activate PAMFN
3. conda install pytorch==1.9.0 torchvision==0.10.0 torchaudio==0.9.0 cudatoolkit=11.1 -c pytorch -c conda-forge -y
4. pip install -r requires.txt
The extracted features and pretrained models can be downloaded from here and should be placed in the current directory.
./
├── data/
└── pretrained_models
Using the following command to evaluate the pretrained model:
python main.py --gpu {gpu_id} --feats {feature_type} --action {action_type} --multi_modality --test
Using the following command to train a modality-specific branch:
python main.py --gpu {gpu_id} --feats {feature_type} --action {action_type} --modality {modality_type}
An Example:
python main.py --gpu 0 --feats 2 --action Ball --modality V
Using the following command to train the mixed-modality branch:
python main.py --gpu {gpu_id} --feats {feature_type} --action {action_type} --multi_modality
An Example:
python main.py --gpu 0 --feats 2 --action Ball --multi_modality
Please cite this work if you find it useful:
@article{zeng2024multimodal,
title={Multimodal Action Quality Assessment},
author={Zeng, Ling-An and Zheng, Wei-Shi},
journal={IEEE Transactions on Image Processing},
year={2024},
publisher={IEEE}
}