XiaohanLei / GaussNav

7 stars 1 forks source link

GaussNav

PyTorch implementation of paper: GaussNav: Gaussian Splatting for Visual Navigation

Project Page

Overview:

Our GaussNav framework consists of three stages, including Frontier Exploration, Semantic Gaussian Construction and Gaussian Navigation. First, the agent employs Frontier Exploration to collect observations of the unknown environment. Second, the collected observations are used to construct Semantic Gaussian. By leveraging semantic segmentation algorithms, we assign semantic labels to each Gaussian. We then cluster Gaussians with their semantic labels and 3D positions, segmenting objects in the scene into different instances under various semantic categories. This representation is capable of preserving not only the 3D geometry of the scene and the semantic labels of each Gaussian, but also the texture details of the scene, thereby enabling novel view synthesis. Third, we render descriptive images for object instances, matching them with the goal image to effectively locate the target object. Upon determining the predicted goal object’s position, we can efficiently transform our Semantic Gaussian into grid map and employ path planning algorithms to accomplish the navigation.

This repository contains:

The Gaussian Navigation module of GaussNav (which can be used for code base for either PointGoal task or the upper bound of planning-based methods of Instance ImageNav task).

We will soon release the code of the Semantic Gaussian Construction module and previously constructed 3D model.

Installing Dependencies

Downloading scene dataset and episode dataset

Test setup

To verify that the data is setup correctly, run:

cd GaussianNavigation
python run.py

Some tips

TODO list:

This project is still under development. Please feel free to raise issues or submit pull requests to contribute to our codebase.

Bibtex:

@misc{lei2024gaussnavgaussiansplattingvisual,
      title={GaussNav: Gaussian Splatting for Visual Navigation}, 
      author={Xiaohan Lei and Min Wang and Wengang Zhou and Houqiang Li},
      year={2024},
      eprint={2403.11625},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2403.11625}, 
}