UMass-Foundation-Model / 3D-VLA

[ICML 2024] 3D-VLA: A 3D Vision-Language-Action Generative World Model
https://vis-www.cs.umass.edu/3dvla/
342 stars 13 forks source link


3D-VLA: A 3D Vision-Language-Action Generative World Model

ICML 2024

Haoyu Zhen, Xiaowen Qiu, Peihao Chen, Jincheng Yang, Xin Yan, Yilun Du, Yining Hong, Chuang Gan

Paper PDF Project Page

Tabel of Contents
  1. Method
  2. Installation
  3. Embodied Diffusion Models
  4. Multimodal Large Language Model
  5. Citation
  6. Acknowledgement

News 📢

Method

3D-VLA is a framework that connects vision-language-action (VLA) models to the 3D physical world. Unlike traditional 2D models, 3D-VLA integrates 3D perception, reasoning, and action through a generative world model, similar to human cognitive processes. It is built on the 3D-LLM and uses interaction tokens to engage with the environment. Embodied diffusion models are trained and aligned with the LLM to predict goal images and point clouds.

Logo

Installation

conda create -n 3dvla python=3.9
conda activate 3dvla
pip install -r requirements.txt

We will update the file structure and the installation process in the future.

We provide a model card for the 3D-VLA model. The model card includes the task description, model description, and training datasets.

Embodied Diffusion Models

Goal Image Generation

Train the goal image latent diffusion model with the following command. If you want to include depth information, you could add --include_depth to the command in the train_ldm.sh file.

bash launcher/train_ldm.sh [NUM_GPUS] [NUM_NODES]

Then you could generate the goal images. The results will be saved in the lavis/output/LDM/pix2pix/results folder.

python inference_ldm_goal_image.py \
    --ckpt_folder lavis/output/LDM/pix2pix/runs (--include_depth)

We have released our model on Hugging Face: goal-image and goal-depth. A simple demo can be run using the following command:

python inference_ldm_goal_image.py \
    --ckpt_folder anyezhy/3dvla-diffusion \
    --image docs/cans.png --text "knock pepsi can over" \
    --save_path result.png

python inference_ldm_goal_image.py \
    --ckpt_folder anyezhy/3dvla-diffusion-depth --include_depth \
    --image docs/bottle.png --text "move water bottle near sponge" \
    --save_path result.png

Goal Point Cloud Generation

We have implemented xFormers for the goal point cloud diffusion model. You could install it and accelerate the training and inference process.

Train the goal point cloud diffusion model (finetuning the pretrained Point-E model).

bash launcher/train_pe.sh [NUM_GPUS] [NUM_NODES]

We have released our model on Hugging Face: goal-point-cloud. Inferece the goal point cloud with the following command. If you want to use multiple GPUs, use torchrun --nproc_per_node=[NUM_GPUS] --master_port=[PORT] inference_pe_goal_pcd.py instead.

  python inference_pe_goal_pcd.py \
    --input_npy docs/point_cloud.npy --text "close bottom drawer" \
    --output_dir SAVE_PATH

  python inference_pe_goal_pcd.py \
    --input_npy docs/money.npy \
    --text "put the money away in the safe on the bottom shelf"

Multimodal Large Language Model

Pretrain 3D-VLA

Train our 3D-VLA model:

bash launcher/train_llm.sh [NUM_GPUS] [NUM_NODES]

Citation

@article{zhen20243dvla,
  author = {Zhen, Haoyu and Qiu, Xiaowen and Chen, Peihao and Yang, Jincheng and Yan, Xin and Du, Yilun and Hong, Yining and Gan, Chuang},
  title = {3D-VLA: 3D Vision-Language-Action Generative World Model},
  journal = {arXiv preprint arXiv:2403.09631},
  year = {2024},
}

Acknowledgement

Here we would like to thank the following resources for their great work: