Zheng Zeng, Valentin Deschaintre, Iliyan Georgiev, Yannick Hold-Geoffroy, Yiwei Hu, Fujun Luan, Ling-Qi Yan, Miloš Hašan
ACM SIGGRAPH 2024
The three areas of realistic forward rendering, per-pixel inverse rendering, and generative image synthesis may seem like separate and unrelated sub-fields of graphics and vision. However, recent work has demonstrated improved estimation of per-pixel intrinsic channels (albedo, roughness, metallicity) based on a diffusion architecture; we call this the RGB→X problem. We further show that the reverse problem of synthesizing realistic images given intrinsic channels, X→RGB, can also be addressed in a diffusion framework.
Focusing on the image domain of interior scenes, we introduce an improved diffusion model for RGB→X, which also estimates lighting, as well as the first diffusion X→RGB model capable of synthesizing realistic images from (full or partial) intrinsic channels. Our X→RGB model explores a middle ground between traditional rendering and generative models: we can specify only certain appearance properties that should be followed, and give freedom to the model to hallucinate a plausible version of the rest.
This flexibility makes it possible to use a mix of heterogeneous training datasets, which differ in the available channels. We use multiple existing datasets and extend them with our own synthetic and real data, resulting in a model capable of extracting scene properties better than previous work and of generating highly realistic images of interior scenes.
├── assets <- Assets used by the README.md
├── rgb2x <- Code for the RGB→X model
│ ├── example <- Example photo
│ └── model_cache <- Model weights (automatically downloaded when running the inference script)
├── x2rgb <- Code for the X→RGB model
│ ├── example <- Example photo
│ └── model_cache <- Model weights (automatically downloaded when running the inference script)
├── environment.yaml <- Env file for creating conda environment
├── LICENSE
└── README.md
You don't need to manually download the model weights. The weights will be downloaded automatically to /rgb2x/model_cache/
and /x2rgb/model_cache/
when you run the inference scripts.
You can manually acquire the weights by cloning the models from Hugging Face:
git-lfs install
git clone https://huggingface.co/zheng95z/x-to-rgb
git clone https://huggingface.co/zheng95z/rgb-to-x
Create a conda environment using the provided environment.yaml
file.
conda env create -n rgbx -f environment.yaml
conda activate rgbx
Note that this environment is only compatible with NVIDIA GPUs. Additionally, we recommend using a GPU with a minimum of 12GB of memory.
When you run the inference scripts, gradio demos will be hosted on your local machine. You can access the demos by opening the URLs (shown in the terminal) in your browser.
cd rgb2x
python gradio_demo_rgb2x.py
Please note that the metallicity channel prediction might behave differently between the demo and the paper. This is because the demo utilizes a checkpoint that predicts roughness and metallicity separately, whereas in the paper, we used a checkpoint where the roughness and metallicity channels were combined into a single RGB image (with the blue channel set to 0). Unfortunately, the latter checkpoint was lost during the transition between computing platforms, and we apologize for the inconvenience. We plan to resolve this issue and will provide an updated demo in the near future.
cd x2rgb
python gradio_demo_x2rgb.py
This implementation builds upon Hugging Face’s Diffusers library. We also acknowledge Gradio for providing an easy-to-use interface that allowed us to create the inference demos for our models.