ethz-asl / COIN-LIO

Other
198 stars 18 forks source link

Complementary Intensity-Augmented LiDAR Inertial Odometry

Patrick Pfreundschuh, Helen Oleynikova, Cesar Cadena, Roland Siegwart, and Olov Andersson. "COIN-LIO: Complementary Intensity-Augmented LiDAR Inertial Odometry" accepted at ICRA 2024. [ ArXiv | Video ]

Abstract
We present COIN-LIO, a LiDAR Inertial Odometry pipeline that tightly couples information from LiDAR intensity with geometry-based point cloud registration. The focus of our work is to improve the robustness of LiDAR-inertial odometry in geometrically degenerate scenarios, like tunnels or flat fields. We project LiDAR intensity returns into an intensity image, and propose an image processing pipeline that produces filtered images with improved brightness consistency within the image as well as across different scenes. To effectively leverage intensity as an additional modality, we present a novel feature selection scheme that detects uninformative directions in the point cloud registration and explicitly selects patches with complementary image information. Photometric error minimization in the image patches is then fused with inertial measurements and point-to-plane registration in an iterated Extended Kalman Filter. The proposed approach improves accuracy and robustness on a public dataset. We additionally publish a new dataset, that captures five real-world environments in challenging, geometrically degenerate scenes. By using the additional photometric information, our approach shows drastically improved robustness against geometric degeneracy in environments where all compared baseline approaches fail.

Please cite our work if you are using COIN-LIO in your research.

  @article{pfreundschuh2023coin,
  title={COIN-LIO: Complementary Intensity-Augmented LiDAR Inertial Odometry},
  author={Pfreundschuh, Patrick and Oleynikova, Helen and Cadena, Cesar and Siegwart, Roland and Andersson, Olov},
  journal={arXiv preprint arXiv:2310.01235},
  year={2023}
  }

Setup

Installation

This package was developed on Ubuntu 20.04 using ROS Noetic. Other versions should also work but have not been tested and we do not guarantee support.

  1. If not done yet, please install ROS, install the proposed system dependencies. Install some additional system dependencies:
    sudo apt-get install python3-catkin-tools libgoogle-glog-dev
  2. Then create a catkin workspace:
    mkdir -p ~/catkin_ws/src
    cd ~/catkin_ws
    catkin init
    catkin config --extend /opt/ros/$ROS_DISTRO
    catkin config --cmake-args -DCMAKE_BUILD_TYPE=RelWithDebInfo
    catkin config --merge-devel
  3. Clone COIN-LIO into your workspace:
    cd ~/catkin_ws/src
    git clone git@github.com:ethz-asl/coin-lio.git
    cd COIN-LIO
  4. Build COIN-LIO:
    catkin build coin_lio

Alternative Installation: Docker

To instead use docker, check out the repository locally, navigate to it, and:

    cd docker/
    ./run_docker.sh -b

Which will build a docker image with a copy of the code checked out inside. Your ~/data folder will be mounted to /root/data within the docker, so you can download datasets and follow the rest of the tutorial below. On future runs, you can simply use ./run_docker.sh (without -b) to not re-build the image.

Running ENWIDE Dataset Sequences

The ENWIDE dataset sequences can be downloaded here. Run a sequence:

  roslaunch coin_lio mapping_enwide.launch bag_file:=<example_bag_path.bag>

Running Newer College Dataset Sequences

The Newer College Dataset sequences can be downloaded here. Run a sequence:

  roslaunch coin_lio mapping_newer_college.launch bag_file:=<example_bag_path.bag>

Running COIN-LIO on your own data:

Note on LiDAR type: COIN-LIO currently only supports data from Ouster LiDARs, as we use the calibration in the metadata file for the image projection model. Implementing different sensors is theoretically possible but requires a proper implementation of a projection model that works for the specific sensor. Contributions are welcome.

Sensor Calibration

We used ascii-image-converter for our ascii animation.