This repository is the official implementation of the paper "A Simple Approach For Visual Room Rearrangement: 3D Mapping & Semantic Search" at ICLR 2023. Our method won the 2022 Rearrangement Challenge at the Embodied AI Workshop, and serves as a baseline for the 2023 challenge.
You can install MaSS by installing required packages with pip install -r requirements.txt
, following the installation instructions for the ai2thor-rearrangement package on GitHub, and then installing detectron2.
Once these are installed, MaSS can be installed via pip install -e .
.
Our method was developed using PyTorch 1.10.2. Newer versions may be compatible, but are untested.
Two model checkpoints are required to run MaSS:
detectron2
available here.policy.pth
in the same directory as this README.Following the instructions here first export the challenge package to your PYTHONPATH
.
export PYTHONPATH=$PYTHONPATH::/path/to/ai2thor-rearrangement
Then you can run the agent by calling agent.py with your python environment.
python -u agent.py \
--logdir ./testing-the-agent --stage val \
--semantic-search-walkthrough \
--semantic-search-unshuffle \
--use-feature-matching \
--start-task 0 --total-tasks 20
The above command runs MaSS using our Semantic Search policy to select navigation goals during the walkthrough phase and the unshuffle phase. In addition, the --use-feature-matching
option uses image features to match instances of objects between the unshuffle phase and walkthrough phase, and should be used as it improves %FixedStrict
by 7.03 points in our experiments.
If you find our work helpful in your research, consider citing our paper at ICLR 2023:
@inproceedings{
trabucco2023a,
title={A Simple Approach for Visual Room Rearrangement: 3D Mapping and Semantic Search},
author={Brandon Trabucco and Gunnar A Sigurdsson and Robinson Piramuthu and Gaurav S. Sukhatme and Ruslan Salakhutdinov},
booktitle={The Eleventh International Conference on Learning Representations },
year={2023},
url={https://openreview.net/forum?id=1C6nCCaRe6p}
}
In addition, consider citing the Rearrangement Challenge benchmark:
@InProceedings{RoomR,
author = {Luca Weihs and Matt Deitke and Aniruddha Kembhavi and Roozbeh Mottaghi},
title = {Visual Room Rearrangement},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2021}
}