Jeff-sjtu / HybrIK

Official code of "HybrIK: A Hybrid Analytical-Neural Inverse Kinematics Solution for 3D Human Pose and Shape Estimation", CVPR 2021
MIT License
1.25k stars 149 forks source link
3d-pose-estimation cvpr cvpr21 inverse-kinematics pose-estimation pytorch smpl

HybrIK: Hybrid Analytical-Neural Inverse Kinematics for Body Mesh Recovery


PyTorch Google Colab
Paper PDF Paper PDF youtube views youtube views

This repo contains the code of our papers:

HybrIK: A Hybrid Analytical-Neural Inverse Kinematics Solution for 3D Human Pose and Shape Estimation, In CVPR 2021

HybrIK-X: Hybrid Analytical-Neural Inverse Kinematics for Whole-body Mesh Recovery, ArXiv 2023

News :triangular_flag_on_post:

[2023/06/02] Demo code for whole-body HybrIK-X is released.

[2022/12/03] HybrIK for Blender add-on is now available for download. The output of HybrIK can be imported to Blender and saved as fbx.

[2022/08/16] Pretrained model with HRNet-W48 backbone is available.

[2022/07/31] Training code with predicted camera is released.

[2022/07/25] HybrIK is now supported in Alphapose! Multi-person demo with pose-tracking is available.

[2022/04/27] Google Colab is ready to use.

[2022/04/26] Achieve SOTA results by adding the 3DPW dataset for training.

[2022/04/25] The demo code is released!

Key idea: Inverse Kinematics

HybrIK and HybrIK-X are based on a hybrid inverse kinematics (IK) to convert accurate 3D keypoints to parametric body meshes.


Twist-and-Swing Decomposition

Installation instructions

# 1. Create a conda virtual environment.
conda create -n hybrik python=3.8 -y
conda activate hybrik

# 2. Install PyTorch
conda install pytorch==1.9.1 torchvision==0.10.1 -c pytorch

# 3. Install PyTorch3D (Optional, only for visualization)
conda install -c fvcore -c iopath -c conda-forge fvcore iopath
conda install -c bottler nvidiacub
pip install git+ssh://git@github.com/facebookresearch/pytorch3d.git@stable

# 4. Pull our code
git clone https://github.com/Jeff-sjtu/HybrIK.git
cd HybrIK

# 5. Install
pip install pycocotools
python setup.py develop  # or "pip install -e ."

Download necessary model files from [Google Drive | Baidu (code: 2u3c) ] and un-zip them in the ${ROOT} directory.

MODEL ZOO

HybrIK (SMPL)

Backbone Training Data PA-MPJPE (3DPW) MPJPE (3DPW) PA-MPJPE (Human3.6M) MPJPE (Human3.6M) Download Config
ResNet-34 w/ 3DPW 44.5 72.4 33.8 55.5 model cfg
HRNet-W48 w/o 3DPW 48.6 88.0 29.5 50.4 model cfg
HRNet-W48 w/ 3DPW 41.8 71.3 29.8 47.1 model cfg

HybrIK-X (SMPL-X)

Backbone MVE (AGORA Test) MPJPE (AGORA Test) Download Config
HRNet-W48 134.1 127.5 model cfg
HRNet-W48 + RLE 112.1 107.6 model cfg

Demo

First make sure you download the pretrained model (with predicted camera) and place it in the ${ROOT}/pretrained_models directory, i.e., ./pretrained_models/hybrik_hrnet.pth and ./pretrained_models/hybrikx_rle_hrnet.pth.

SMPL

python scripts/demo_video.py --video-name examples/dance.mp4 --out-dir res_dance --save-pk --save-img

The saved results in ./res_dance/res.pk can be imported to Blender with our add-on.

python scripts/demo_image.py --img-dir examples --out-dir res

SMPL-X

python scripts/demo_video_x.py --video-name examples/dance.mp4 --out-dir res_dance --save-pk --save-img

Fetch data

Download Human3.6M, MPI-INF-3DHP, 3DPW and MSCOCO datasets. You need to follow directory structure of the data as below. Thanks to the great job done by Moon et al., we use the Human3.6M images provided in PoseNet.

|-- data
`-- |-- h36m
    `-- |-- annotations
        `-- images
`-- |-- pw3d
    `-- |-- json
        `-- imageFiles
`-- |-- 3dhp
    `-- |-- annotation_mpi_inf_3dhp_train.json
        |-- annotation_mpi_inf_3dhp_test.json
        |-- mpi_inf_3dhp_train_set
        `-- mpi_inf_3dhp_test_set
`-- |-- coco
    `-- |-- annotations
        |   |-- person_keypoints_train2017.json
        |   `-- person_keypoints_val2017.json
        |-- train2017
        `-- val2017

Train from scratch

./scripts/train_smpl_cam.sh test_3dpw configs/256x192_adam_lr1e-3-res34_smpl_3d_cam_2x_mix_w_pw3d.yaml

Evaluation

Download the pretrained model (ResNet-34 or HRNet-W48).

./scripts/validate_smpl_cam.sh ./configs/256x192_adam_lr1e-3-hrw48_cam_2x_w_pw3d_3dhp.yaml ./pretrained_hrnet.pth

Citing

If our code helps your research, please consider citing the following paper:

@inproceedings{li2021hybrik,
    title={Hybrik: A hybrid analytical-neural inverse kinematics solution for 3d human pose and shape estimation},
    author={Li, Jiefeng and Xu, Chao and Chen, Zhicun and Bian, Siyuan and Yang, Lixin and Lu, Cewu},
    booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
    pages={3383--3393},
    year={2021}
}

@article{li2023hybrik,
    title={HybrIK-X: Hybrid Analytical-Neural Inverse Kinematics for Whole-body Mesh Recovery},
    author={Li, Jiefeng and Bian, Siyuan and Xu, Chao and Chen, Zhicun and Yang, Lixin and Lu, Cewu},
    journal={arXiv preprint arXiv:2304.05690},
    year={2023}
}