PyTorch implementation of our IJCV paper:
https://arxiv.org/pdf/1910.04099.pdf
Original Torch implementation for LayoutNet is here.
python t72pkl.py
python train_PC.py
python train_stanford.py
python train_matterport.py
python test_PC.py
python test_stanford.py
On Matterport3D (3D IoU, 2D IoU on under top-down view, RMSE for depth and delta_1 for depth)
python test_matterport.py
For depth related evaluation, we need to render depth map from predicted corner position on equirectangualr view (you can skip this step as we've provided pre-computed depth maps from our approach)
First, uncomment L313-L314 in test_matterport.py, and comment out lines related to evaluation for depth. Run test_matterport.py and save intermediate corner predictions to folder ./result_gen. Then open matlab:
cd matlab
cor2depth
cd ..
Rendered depth maps will be saved to folder ./result_gen_depth/. Then comment out L313-L314 in test_matterport.py, uncomment lines related to evaluation for depth, and run test_matterport.py again
Please cite our paper for any purpose of usage.
@article{zou2021manhattan,
title={Manhattan Room Layout Reconstruction from a Single $ $360\^{}$\{$$\backslash$circ$\}$ $$360∘ Image: A Comparative Study of State-of-the-Art Methods},
author={Zou, Chuhang and Su, Jheng-Wei and Peng, Chi-Han and Colburn, Alex and Shan, Qi and Wonka, Peter and Chu, Hung-Kuo and Hoiem, Derek},
journal={International Journal of Computer Vision},
volume={129},
number={5},
pages={1410--1431},
year={2021},
publisher={Springer}
}