In the project, we propose to replicate the depth map prediction from a single image using the multi-scale deep network introduced by Eigen et al. [1] (2014) with pytorch implementation and compare our work with existing literature.
These instructions will get you a copy of the project up and running on your local machine for development and testing purposes.
The main function of the project.
The CNN model with coarse net and fine net.
Handles the input data.
What things you need to install in your python environment
import pytorch
import numpy as np
example
We use NYU Depth V2 provided by Nathan Siberman and Rob Fergus. The NYU-Depth data set is comprised of video sequences from a variety of indoor scenes as recorded by both the RGB and Depth cameras from the Microsoft Kinect. The dataset contains 464 different indoor scenes with 26 scene types. There are 1446 densely labeled pairs of aligned RGB and depth frames. In addition to the labeled images, the dataset also contains a large number of new unlabeled images.
This project is licensed under the MIT License - see the LICENSE.md file for details