HuangZhe885 / Collaborative-Perception

AEV2V: Accurate and Efficient Vehicle-to-Vehicle Collaborative Perception
4 stars 0 forks source link
collaborative-perception object-detection v2v-communication

Collaborative 3D Object Detection by Intelligent Vehicles Considering Semantic Information and Agent Heterogeneity

image

Collaborative perception can significantly enhance perception performance through information sharing among multiple smart vehicles and roadside perception systems. Existing methods often require sharing and aggregating all information from other collaborators.

WechatIMG206

However, in practice, most information shared by collaborators is largely irrelevant to the final perception, making it challenging to identify distant and occluded objects. To address this issue, we propose a novel collaborative framework, named Semantic Aware Heterogeneous Network (SAHNet), which shares and fuses perceptually crucial and useful information among heterogeneous collaborators to improve the performance of 3D object detection. Specifically, we firstly design Foreground and Boundary Feature Selection (FBFS) to enhance meaningful feature extraction. A Heterogeneous Feature Transfer module (HFF) is then proposed to consider collaborators' heterogeneity to better transfer perception-critical features. Finally, we put forward a Semantic Feature Fusion module (SFF) that effectively aggregates features using semantic information.

https://github.com/HuangZhe885/Collaborative-Perception/assets/44192081/2a860911-ee6f-44b6-ab13-17cf3e4aacad

METHOD

image

RESULTS

We evaluate the effectiveness of the proposed AEV2V on the V2V perception dataset V2XSet.

image

Visualization of collaboration

image image

CONCLUSION

In this paper, we have developed a novel deep learning based framework for collaborative object detection to improve the perception capabilities of autonomous vehicles. It consists of multiple modules designed to learn critical information from collaborating vehicles and address data heterogeneity. This framework can integrate the learned critical features into the final information fusion, assisting single vehicle in being informed about distant occluded objects and enhancing decision-making capabilities. Extensive experiments conducted on three public datasets demonstrate the effectiveness and robustness of the proposed method, showcasing better performance than the state-of-the-art methods. We believe that our research results contribute to improving the accuracy of collaborative perception in autonomous driving. Future work will investigate methods to compensate for delay in feature transmission and adjust object pose and continue to deepen our understanding of collaborative 3D object detection.