Here is the official PyTorch implementation for our paper "Deep Facial Synthesis: A New Challenge".
This project achieve the translation between face photos and artistic portrait drawings using a GAN-based model. You may find useful information in training/testing tips.
📕Find our paper on Springer, Machine Intelligence Research, arXiv.
✨Try our online Colab demo to generate your own facial sketches.
Find the introduction video made by MIR:
Style-1 | Style-2 | Style-3 |
---|---|---|
pip install -r requirements.txt
We conduct all the experiments on the currently largest Facial Sketch Synthesis (FSS) dataset FS2K. For more details about this dataset, please visit its repo.
In this project, we follow the APDrawingGAN to do some preprocessing on original images, including aligning photo by key points (MTCNN), segment human portrait regions (U-2-Net). You can download the preprocessed FS2K dataset here.
If you want to conduct the preprocessing on other images, please refer to the preprocessing section in APDrawingGAN and our own codes for more smooth preprocessing with MTCNN and U-2-Net.
Run python -m visdom.server
python train.py --dataroot /home/pz1/datasets/fss/FS2K_data/train/photo/ --checkpoints_dir checkpoints --name ckpt_0 \
--use_local --discriminator_local --niter 150 --niter_decay 0 --save_epoch_freq 1
If you run on DGX-server, you can use sub_by_id.sh
to set up many experiments one time.
To see losses in training, please refer to log file slurm.out.
Download the weights of pretrained models from the folder for this FSS task on google-drive and specify the path of weights in train/test shell script.
single_model_test.sh
.test_ours.sh
.Best practice for training and testing your models.
Thanks to the great codebase of APDrawingGAN.
If you find our code and metric useful in your research, please cite our papers.
@aticle{Fan2022FS2K,
title={Facial-Sketch Synthesis: A New Challenge},
author={Deng-Ping, Fan and Ziling, Huang and Peng, Zheng and Hong, Liu and Xuebin, Qin and Luc, Van Gool},
journal={Machine Intelligence Research},
year={2022}
}
@inproceedings{fan2019scoot,
title={Scoot: A perceptual metric for facial sketches},
author={Fan, Deng-Ping and Zhang, ShengChuan and Wu, Yu-Huan and Liu, Yun and Cheng, Ming-Ming and Ren, Bo and Rosin, Paul L and Ji, Rongrong},
booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
pages={5612--5622},
year={2019}
}