Aiming at interpretable and flexible cooperative perception, we propose the concept of query cooperation in this paper, which enables instance-level feature interaction among agents via the query stream. To specifically describe the query cooperation, a representative cooperative perception framework (QUEST) is proposed. It performs cross-agent query interaction by fusion and complementation, which are designed for co-aware objects and unaware objects respectively. Taking camera-based vehicle-infrastructure cooperative perception as a typical scenario, we generate the camera-centric cooperation labels of DAIR-V2X-Seq and evaluate the proposed framework on it. The experimental results not only demonstrate the effectiveness but also show the advantages of transmission flexibility and robustness to packet dropout. In addition, we discuss the pros and cons of query cooperation paradigm from the possible extensions and foreseeable limitations.
For technical details, please refer to:
Really glad that our work is valuable for you. Actually, we are not planning to open-source the QUEST because of the requirement of the enterprise partners, but it is applied to a unified framework for End-to-End Cooperative Autonomous Driving, called UniV2X.
The official code of UniV2X is open-sourced at UniV2X-Github, and your can find more implementation details of query cooperation in it.
For technical details of UniV2X, please refer to UniV2X-Paper
If you find our work useful in your research, please consider citing:
@InProceedings{fan2023quest,
author = {Fan, Siqi and Yu, Haibao and Yang, Wenxian and Yuan, Jirui and Nie, Zaiqing},
title = {QUEST: Query Stream for Practical Cooperative Perception},
booktitle = {IEEE International Conference on Robotics and Automation (ICRA)},
month = {May},
year = {2024},
pages = {}
}