By Zhuo Deng, Longin Jan Latecki (Temple University). This paper was published in CVPR 2017.
Code is released under the GNU GENERAL PUBLIC LICENSE (refer to the LICENSE file for details).
If you use this project for your research, please consider citing:
@inproceedings{zhuo17amodal3det,
author = {Zhuo Deng and Longin Jan Latecki},
booktitle = {Conference on Computer Vision and Pattern Recognition (CVPR)},
title = {Amodal Detection of 3D Objects: Inferring 3D Bounding Boxes from 2D Ones in RGB-Depth Images},
year = {2017}
}
The code is tested on the following system:
clone the Amodal3Det repository:
git clone https://github.com/phoenixnn/Amodal3Det.git
# assume you clone the repo into the local your_root_dir
cd your_root_dir
make all -j8 && make pycaffe
sudo cp cudnn_folder/include/cudnn.h /usr/local/cuda-7.5/include/
sudo cp cudnn_folder/lib64/*.so* /usr/local/cuda-7.5/lib64/
Download NYUV2 dataset with 3D annotations and unzip:
wget 'https://cis.temple.edu/~latecki/TestData/DengCVPR2017/NYUV2_3D_dataset.zip' -P your_root_dir/dataset/NYUV2/
Download precomputed 2D segment proposals based on MCG3D and unzip:
wget 'https://cis.temple.edu/~latecki/TestData/DengCVPR2017/Segs.zip' -P your_root_dir/matlab/NYUV2/
Download pretrained models and unzip:
wget 'https://cis.temple.edu/~latecki/TestData/DengCVPR2017/pretrained.zip' -P your_root_dir/rgbd_3det/
VGG16 comes from the Caffe Model Zoo, but is provided here for your convenience.
Run "your_root_dir/matlab/NYUV2/pipeline.m" in Matlab to extact required data.
Set up training/test data:
run "setup_training_data.py" and "setup_testing_data.py" under your_root_dir/rgbd_3det/data respectively
Train model:
cd your_root_dir
./trainNet.sh
Test model: run "test_cnn.py"