Tsinghua-MARS-Lab / Occ3D

MIT License
404 stars 26 forks source link
# Occ3D: A Large-Scale 3D Occupancy Prediction Benchmark for Autonomous Driving [Paper](https://arxiv.org/abs/2304.14365) | [Webpage](https://tsinghua-mars-lab.github.io/Occ3D/)

News

Table of Contents

Introduction

This paper propose 3D Occupancy Prediction, a general and comprehensive 3D perception task for vision-based robotic applications. Occupancy prediction can represent both the semantics and geometry of any scene effectively. We develop a rigorous label generation pipeline for occupancy prediction, construct two challenging datasets (Occ3D-Waymo and Occ3d-nuScenes) and establish a benchmark together with evaluation metrics to facilitate future research. In addition, we propose a novel CTF-Occ network that achieves outstanding occupancy prediction performance. For more information, please refer to the Paper.

Data

Basic Information

Figure 1. Semantic labels (left), visibility masks in the LiDAR (middle) and the camera (right) view. Grey voxels are unobserved in LiDAR view and white voxels are observed in the accumulative LiDAR view but unobserved in the current camera view.

Occ3D-Waymo

| Type | Info | | :----: | :----: | | train | 798 | | val | 202 | | test | 150 | | cameras | 5 | | voxel size | [0.1m, 0.1m, 0.2m] / [0.4m, 0.4m, 0.4m] | | range | [-80m, -80m, -5m, 80m, 80m, 7.8m] / [-40m, -40m, -1m, 40m, 40m, 5.4m] | | volume size | [1600, 1600, 64] / [200, 200, 16]| | classes | 0 - 15 |

(back to top)

Occ3D-nuScenes

| Type | Info | | :----: | :----: | | train | 600 | | val | 150 | | test | 250 | | cameras | 6 | | voxel size | [0.4m, 0.4m, 0.4m] | | range | [-40m, -40m, -1m, 40m, 40m, 5.4m] | | volume size | [200, 200, 16]| | classes | 0 - 17 |

(back to top)

Getting Started

We will release the code soon.

Citation

If you find our work useful for your research, please consider citing the paper:

@article{tian2023occ3d,
  title={Occ3D: A Large-Scale 3D Occupancy Prediction Benchmark for Autonomous Driving},
  author={Tian, Xiaoyu and Jiang, Tao and Yun, Longfei and Wang, Yue and Wang, Yilun and Zhao, Hang},
  journal={arXiv preprint arXiv:2304.14365},
  year={2023}
}

License

Before using the dataset, you should register on the website and agree to the terms of use of the nuScenes and Waymo. The code used to generate the data and the generated data are both subject to the MIT License.

(back to top)