Fsoft-AIC / Open-Vocabulary-Affordance-Detection-in-3D-Point-Clouds

[IROS 2023] Open-Vocabulary Affordance Detection in 3d Point Clouds
https://openad2023.github.io
MIT License
49 stars 1 forks source link
3d-point-clouds affordance-learning computer-vision-ai iros2023 robotics
# Open-Vocabulary Affordance Detection in 3D Point Clouds [![Conference](https://img.shields.io/badge/IROS-2023-FGD94D.svg)](https://ieee-iros.org/) [![Paper](https://img.shields.io/badge/Paper-arxiv.2303.02401-FF6B6B.svg)](https://arxiv.org/abs/2303.02401) Official code for the IROS 2023 paper "Open-Vocabulary Affordance Detection in 3D Point Clouds".

Best Overall and Best Student Paper Awards Finalist

We present OpenAD for a new task of open-vocabulary affordance detection in 3D point clouds. Different from traditional methods that are restricted to a predefined affordance labels set, OpenAD can detect unlimited affordances conveyed through the form of natural language. ![image](demo/method.jpg) Our key idea is to learn collaboratively the mapping between the language labels and the visual features of the point cloud.

1. Getting Started

We strongly encourage you to create a separate CONDA environment.

conda create -n openad python=3.8
conda activate openad
conda install pytorch pytorch-cuda=11.6 -c pytorch -c nvidia
conda install pip
pip install -r requirements.txt

2. Data

Download data from this drive folder.

Currently, we support 2 models (OpenAD with backbones of PointNet++ and DGCNN) and 2 settings (full-shape and partial-view).

3. Training

Please train the model on a single GPU for the best performance. Below are the steps for training the model with PointNet++ backbone on the full-shape setting, those of other combinations are equivalent.

4. Open-Vocabulary Testing

The followings are steps for open-vocabulary testing a trained model with PointNet++ backbone on the full-shape setting, those of other combinations are equivalent.

We provide the pretrained models at this drive.

5. Citation

If you find our work useful for your research, please cite:

@inproceedings{Nguyen2023open,
      title={Open-vocabulary affordance detection in 3d point clouds},
      author={Nguyen, Toan and Vu, Minh Nhat and Vuong, An and Nguyen, Dzung and Vo, Thieu and Le, Ngan and Nguyen, Anh},
      booktitle = IROS,
      year      = 2023
}

6. Acknowledgement

Our source code is built with the heavy support from 3D AffordaceNet. We express a huge thank to them.