Scott Reed, Zeynep Akata, Santosh Mohan, Samuel Tenka, Bernt Schiele, Honglak Lee
This is the code for our NIPS 2016 paper on text- and location-controllable image synthesis using conditional GANs. Much of the code is adapted from reedscot/icml2016 and dcgan.torch.
You will need to install Torch, CuDNN, stnbhwd and the display package.
CONFIG
file to point to your data../scripts/train_cub_keypoints.sh
./scripts/run_all_demos.sh
.Moving the bird's position via bounding box:
Moving the bird's position via keypoints:
Birds text to image with ground-truth keypoints:
Birds text to image with generated keypoints:
Humans text to image with ground-truth keypoints:
Humans text to image with generated keypoints:
If you find this useful, please cite our work as follows:
@inproceedings{reed2016learning,
title={Learning What and Where to Draw},
author={Scott Reed and Zeynep Akata and Santosh Mohan and Samuel Tenka and Bernt Schiele and Honglak Lee},
booktitle={Advances in Neural Information Processing Systems},
year={2016}
}