About
This is the source code for the paper
Chen Li, Gim Hee Lee. Generating Multiple Hypotheses for 3D Human Pose Estimation with Mixture Density Network. In CVPR2019.
We argue that 3D human pose estimation from a monocular/2D-joint input is an iverse problem where multiple solutions can exist.
We use a two-stage approach to generate multiple 3D pose hypotheses. The 2D joints are firstly detected from the input images in the first stage, followed by a feature extractor and hypotheses generator to generate 3D pose hypotheses.
For more details, please refer to our paper on arXiv.
Bibtex:
@inproceedings{li2019generating,
title={Generating multiple hypotheses for 3d human pose estimation with mixture density network},
author={Li, Chen and Lee, Gim Hee},
booktitle={Proceedings of the IEEE/CVF conference on computer vision and pattern recognition},
pages={9887--9895},
year={2019}
}
Dependencies
Train
Get this code:
git clone https://github.com/chaneyddtt/Generating-Multiple-Hypotheses-for-3D-Human-Pose-Estimation-with-Mixture-Density-Network.git
Download the 2D detections of Human3.6 dataset.
Run:
python predict_3dpose_mdm.py --train_dir <specify your training folder>
You can also change the arguments during training. For example,you can train with one or two missing joint(s) randomly selected from the limbs by run:
python predict_3dpose_mdm.py --miss_num <specify how many joints are missing>
You can also change other arguments in the predict_3dpose_mdm.py in a similar way.
Test
Down our pretrained model
To test our pretrained model, run:
python predict_3dpose_mdm.py --test True --load 4338038 --load_dir ../Models/mdm_5_prior/ (model with the dirichlet conjucate prior)
or run:
python predict_3dpose_mdm.py --test True --load 4679232 --load_dir ../Models/mdm_5/ (model without the dirichlet conjucate prior)
Visualize
To visualze all the five 3D pose hypotheses generated by our model, run:
python predict_3dpose_mdm.py --sample True --load 4338038 --load_dir ../Models/mdm_5_prior/
Acknowledgments
The pre-processed human3.6 dataset and the feature extractor of our model was ported or adapted from the code by @una-dinosauria.