Sbrunoberenguel / FreDSNet

Code to test FreDSNet: Frequential Depth estimation and Semantic segmentation Network
GNU General Public License v3.0
39 stars 3 forks source link
computer-vision deep-learning monocular-depth-estimation omnidirectional-vision scene-understanding semantic-segmentation

FreDSNet

We are glad to present FreDSNet (Frequential Depth and Semantics Network). This is the first network which takes advantage of the Fast Fourier Convolutions for Scene Understanding purposes. Also, this is the first network that jointly obtains Semantic Segmentation and monocular Depth from a single equirectangular panorama.

The presented code belongs to the investigation from the paper accepted on ICRA2023. The conference paper can be found as: FreDSNet: Joint Monocular Depth and Semantic Segmentation with Fast Fourier Convolutions from Single Panoramas.

PWC PWC

Code

We provide our original implementation of FreDSNet and the code for obtaining the corresponding depth and semantic maps from equirectangular panoramas. We also provide a little code for 3D reconstruction of the environment taking as input the depth maps generated with our network. To download the weigths used in the article, click Here

First, we set up a virtual environment with Anaconda and activate it as:

conda env create -f FreDSEnv.yaml

conda activate FreDS

An example of use of our network is:

python inference.py 

And, in case you do not have a GPU on your computer:

python inference.py --no_cuda

The input images are in Example and the output information will be stored in Results.

To generate the 3D reconstruction (no GPU is needed), run:

python room_viewer.py 

Obtaining these 3D reconstruction from the images from Example:

Note from the authors

This code has not been thoroughly tested, which means it may have some bugs. Please use with caution.

The authors and developers of this code want the best for the users and have gone to great lengths to make it easy to use and accessible. Be nice to them and enjoy their work.

If any problem may appear, do not hesitate and we will do our best to solve it (at least we will try).

License

This software is under GNU General Public License Version 3 (GPLv3), please see GNU License

For commercial purposes, please contact the authors: Bruno Berenguel-Baeta (berenguel@unizar.es), Jesús Bermudez-Cameo (bermudez@unizar.es) and Josechu Guerrero (josechu.guerrero@unizar.es)