Official code will be released soon for ICCV 2023 paper AG3D: Learning to Generate 3D Avatars from 2D Image Collections. Learned from 2D image collections, AG3D synthesizes novel 3D humans with high-quality appearance and geometry, different identities and clothing styles including loose clothing such as skirts.
Clone this repo:
git clone https://github.com/zj-dong/AG3D.git
cd AG3D
We suggest to use anaconda to manage the python environments:
conda env create -f env.yml
conda activate ag3d
python setup.py install
Download SMPL models (1.0.0 for Python 2.7 (10 shape PCs)) and move them to the corresponding locations:
mkdir training/deformers/smplx/SMPLX
mv /path/to/smpl/models/basicModel_f_lbs_10_207_0_v1.0.0.pkl training/deformers/smplx/SMPLX/SMPL_NEUTRAL.pkl
Download our pretrained models:
sh ./scripts/download_model.sh
Download pose distribution of training data:
sh ./scripts/download_pose.sh
DeepFashion:
python test.py --network=./model/deep_fashion.pkl --pose_dist=./data/dp_pose_dist.npy --output_path './result/deepfashion' --res=512 --truncation=0.7 --number=100 --type=gen_samples
UBCFashion:
python test.py --network=./model/ubc_fashion.pkl --pose_dist=./data/ubc_pose_dist.npy --output_path './result/ubc' --res=512 --truncation=0.7 --number=100 --type=gen_samples
Generate results of novel view synthesis:
python test.py --network=./model/deep_fashion.pkl --pose_dist=./data/dp_pose_dist.npy --output_path='./result/gen_novel_view' --res=512 --truncation=0.7 --number=100 --type=gen_novel_view
Some sampled generated results are shown below.
Generate interpolation results:
python test.py --network=./model/deep_fashion.pkl --pose_dist=./data/dp_pose_dist.npy --output_path='./result/result_interp' --res=512 --truncation=0.7 --number=100 --type=gen_interp --is_mesh=True
Some sampled interpolation results:
Generate animations (please download motion sequence from AMASS):
python test.py --network=./model/deep_fashion.pkl --pose_dist=./data/dp_pose_dist.npy --output_path='./result/result_anim' --res=512 --truncation=0.7 --number=100 --type=gen_anim --motion_path=./data/animation/motion_seq
Some sampled animation results:
Calculate metrics in the paper:
python evaluate.py --metrics=fid5k_full --network=./model/deep_fashion.pkl --data=./data/eva3d_icon.zip --res=512
If you find our code or paper useful, please cite as
@inproceedings{dong2023ag3d,
title={{AG3D}: {L}earning to {G}enerate {3D} {A}vatars from {2D} {I}mage {C}ollections},
author={Dong, Zijian and Chen, Xu and Yang, Jinlong and Black, Michael J and Hilliges, Otmar and Geiger, Andreas},
booktitle = {International Conference on Computer Vision (ICCV)},
year = {2023}
}
We thank for their feedback and discussions.
Here are some great resources we benefit from:
This project was supported by the ERC Starting Grant LEGO-3D (850533), the BMWi project KI Delta Learning (project number 19A19013O) and the DFG EXC number 2064/1 - project number 390727645.