KHB1698 / DC-GCT

20 stars 1 forks source link

Double-chain Constraints for 3D Human Pose Estimation in Images and Videos

Double-chain Constraints for 3D Human Pose Estimation in Images and Videos,
Hongbo Kang, Yong Wang, Mengyuan Liu, Doudou Wu, Peng Liu, Wenming Yang
arXiv, 2023

This link contains the code that supports our latest work: DRPose

Results on Human3.6M

Protocol 1 (mean per-joint position error) when 2D keypoints detected by CPN, HRNet and the ground truth of 2D poses.

Method 2D Pose MPJPE
DC-GCT GT 32.4 mm
DC-GCT CPN 48.4 mm
DC-GCT (w/refine) CPN 47.4 mm
DC-GCT HRNet 47.2 mm
DC-GCT (w/refine) HRNet 46.1 mm

Dependencies

pip install -r requirement.txt

Dataset setup

Please download the dataset here and refer to VideoPose3D to set up the Human3.6M dataset ('./dataset' directory).

${POSE_ROOT}/
|-- dataset
|   |-- data_3d_h36m.npz
|   |-- data_2d_h36m_gt.npz
|   |-- data_2d_h36m_cpn_ft_h36m_dbb.npz

Download pretrained model

The pretrained model is here, please download it and put it in the './ckpt/pretrained' directory.

Test the model

To test on Human3.6M on single frame, run:

python main.py --reload --previous_dir "ckpt/pretrained" 

Train the model

To train on Human3.6M with single frame, run:

python main.py --train -n 'name'

Demo

To begin, download the YOLOv3 and HRNet pretrained models here and put it in the './demo/lib/checkpoint' directory. Next, download the pretrained model and put it in the './ckpt/pretrained' directory. Lastly, Put your own images in the './demo/figure', and run:

python demo/vis.py

Citation

If you find our work useful in your research, please consider citing:

@article{kang2023double,
title={Double-chain Constraints for 3D Human Pose Estimation in Images and Videos},
author={Kang, Hongbo and Wang, Yong and Liu, Mengyuan and Wu, Doudou and Liu, Peng and Yang, Wenming},
journal={arXiv preprint arXiv:2308.05298},
year={2023}
}

Acknowledgement

Our code is extended from the following repositories. We thank the authors for releasing the codes.