PointBeV: A Sparse Approach to BeV Predictions
Loick Chambon, Eloi Zablocki, Mickael Chen, Florent Bartoccioni, Patrick Perez, Matthieu Cord.
Valeo AI, Sorbonne University
PointBeV reaches state-of-the-art on several segmentation tasks (vehicle without filtering above) while allowing a trade-off between performance and memory consumption. | PointBeV reaches state-of-the-art on several segmentation tasks (vehicle with filtering above). It can also be used using different pattern strategies, for instance a LiDAR pattern. |
Illustration of different sampling patterns, respectively: a full, a regular, a drivable hdmap, a lane hdmap, a front camera and a LiDAR pattern. PointBeV is flexible to any pattern. |
We propose PointBeV, a novel sparse BeV segmentation model operating on sparse BeV features instead of dense grids. This approach offers precise control over memory usage, enabling the use of long temporal contexts and accommodating memory-constrained platforms. PointBeV employs an efficient two-pass strategy for training, enabling focused computation on regions of interest. At inference time, it can be used with various memory/performance trade-offs and flexibly adjusts to new specific use cases. PointBeV achieves state-of-the-art results on the nuScenes dataset for vehicle, pedestrian, and lane segmentation, showcasing superior performance in static and temporal settings despite being trained solely with sparse signals.
PointBeV architecture is an architecture dealing with sparse representations. It uses an efficient Sparse Feature Pulling module to propagate features from images to BeV and a Sparse Attention module for temporal aggregation. |
If this work is helpful for your research, please consider citing the following BibTeX entry and putting a star on this repository.
@inproceedings{chambon2024pointbev,
title={PointBeV: A Sparse Approach to BeV Predictions},
author={Loick Chambon and Eloi Zablocki and Mickael Chen and Florent Bartoccioni and Patrick Perez and Matthieu Cord},
year={2024},
booktitle={CVPR}
}
PointBeV is originally designed for vehicle segmentation. It can be used with different sampling patterns and different memory/performance trade-offs. It can also be used with temporal context to improve the segmentation.
Models | PointBeV (ours) | BAEFormer | SimpleBeV | BEVFormer | CVT |
---|---|---|---|---|---|
IoU | 47.6 | 41.0 | 46.6 | 45.5 | 37.7 |
Below we illustrate the model output. On the ground truth, we distinguish vehicle with low visibility (vis < 40%) in light blue from those with higher visibility (vis > 40%) in dark blue. We can see that PointBeV is able to segment vehicles with low visibility, which is a challenging task for other models. They often correspond to occluded vehicles.
We also illustrate the results of a temporal model on random samples taken from the NuScenes validation set. The model used for the visualisation is trained without filtering, at resolution 448x800.
PointBeV can be used to perform inference with fewer points than other models. We illustrate this below with a vehicle segmentation model. We can see that PointBeV is able to perform inference with 1/10 of the points used by other models while maintaining a similar performance. This is possible thanks to the sparse approach of PointBeV. In green is represented the sampling mask. Predictions are only performed on the sampled points.
PointBeV can also be used for different segmentation tasks such as pedestrians or hdmap segmentation.
Models | PointBeV (ours) | TBP-Former | ST-P3 | FIERY | LSS |
---|---|---|---|---|---|
IoU | 18.5 | 17.2 | 14.5 | 17.2 | 15.0 |
Models | PointBeV (ours) | MatrixVT | M2BeV | PeTRv2 | BeVFormer |
---|---|---|---|---|---|
IoU | 49.6 | 44.8 | 38.0 | 44.8 | 25.7 |
➡️ Create the environment.
git clone https://github.com/...
cd PointBeV
micromamba create -f environment.yaml -y
micromamba activate pointbev
➡️ Install cuda dependencies.
cd pointbev/ops/gs; python setup.py build install; cd -
➡️ Datasets.
We used nuScenes dataset for our experiments. You can download it from the official website: https://www.nuscenes.org/nuscenes.
mkdir data
ln -s $PATH/nuscenes data/nuScenes
pytest tests/test_datasets.py
➡️ Backbones:
Backbones are downloaded the first time the code is run. We've moved them to a folder so that we can retrieve the weights quickly for other runs.
wget https://download.pytorch.org/models/resnet50-0676ba61.pth -P backbones
wget https://download.pytorch.org/models/resnet101-63fe2227.pth -P backbones
wget https://github.com/lukemelas/EfficientNet-PyTorch/releases/download/1.0/efficientnet-b4-6ed6700e.pth -P backbones
Optional: Preprocess the dataset to train HDmaps model. Building hdmaps 'on the fly' can slow down the dataloader, so we strongly advise you to save the preprocessed dataset.
python pointbev/data/dataset/create_maps.py --split val train --version=trainval
python pointbev/data/dataset/create_maps.py --split mini_val mini_train --version=mini
The directory will be as follows.
PointBeV
├── data
│ ├── nuScenes
│ │ ├── samples
│ │ ├── sweeps
│ │ ├── v1.0-mini
| | ├── v1.0-trainval
| |── nuscenes_processed_map
| | ├── label
| | | ├── mini_train
| | | ├── mini_val
| | | ├── train
| | | ├── val
| | ├── map_0.1
Sanity check.
pytest tests/test_model.py
Overfitting.
python pointbev/train.py flags.debug=True task_name=debug
Training with simple options:
python pointbev/train.py \
model/net/backbone=efficientnet \ # Specifiy the backbone
data.batch_size=8 \ # Select a batch size
data.valid_batch_size=24 \ # Can be a different batch size to faster validation
data.img_params.min_visibility=1 \ # With or without the visibility filtering
data/augs@data.img_params=scale_0_3 \ # Image resolution
task_name=folder # Where to save the experiment in the logs folder.
If you want to train with the reproduced code of BeVFormer static (by specifying model=BeVFormer
), do not forget to compile the CUDA dependency.
cd pointbev/ops/defattn; python setup.py build install; cd -
Then select BeVFormer model when running code:
python pointbev/train.py \
model=BeVFormer
To evaluate a checkpoint, do not forget to specify the actual resolution and the visibility filtering applied.
python pointbev/train.py train=False test=True task_name=eval \
ckpt.path=PATH_TO_CKPT \
model/net/backbone=efficientnet \
data/augs@data.img_params=scale_0_5 \
data.img_params.min_visibility=1
If you evaluate a pedestrian or an hdmap model do not forget to change the annotations.
python pointbev/train.py train=False test=True task_name=eval \
ckpt.path=PATH_TO_CKPT \
model/net/backbone=resnet50 \
data/augs@data.img_params=scale_0_3 \
data.img_params.min_visibility=2 \
data.filters_cat="[pedestrian]" # Instead of filtering vehicles, we filter pedestrians for GT.
If you evaluate a temporal model do not forget to change the model and the temporal frames.
python pointbev/train.py train=False test=True task_name=eval \
model=PointBeV_T \
data.cam_T_P='[[-8,0],[-7,0],[-6,0],[-5,0],[-4,0],[-3,0],[-2,0],[-1,0],[0,0]]' \
ckpt.path=PATH_TO_CKPT \
model/net/backbone=resnet50 \
data/augs@data.img_params=scale_0_3 \
data.img_params.min_visibility=2 \
data.filters_cat="[pedestrian]"
About the temporal frames, T_P means 'Time_Pose'. For instance:
Backbone | Resolution | Visibility | IoU |
---|---|---|---|
Eff-b4 | 224x480 | 1 | 38.69 |
Eff-b4 | 448x800 | 1 | 42.09 |
Eff-b4 | 224x480 | 2 | 43.97 |
Eff-b4 | 448x800 | 2 | 47.58 |
Many thanks to these excellent open source projects:
To structure our code we used this template: https://github.com/ashleve/lightning-hydra-template