Loading [a11y]/accessibility-menu.js
SGDet3D: Semantics and Geometry Fusion for 3D Object Detection Using 4D Radar and Camera | IEEE Journals & Magazine | IEEE Xplore

SGDet3D: Semantics and Geometry Fusion for 3D Object Detection Using 4D Radar and Camera


Abstract:

4D millimeter-wave radar has gained attention as an emerging sensor for autonomous driving in recent years. However, existing 4D radar and camera fusion models often fail...Show More

Abstract:

4D millimeter-wave radar has gained attention as an emerging sensor for autonomous driving in recent years. However, existing 4D radar and camera fusion models often fail to fully exploit complementary information within each modality and lack deep cross-modal interactions. To address these issues, we propose a novel 4D radar and camera fusion method, named SGDet3D, for 3D object detection. Specifically, we first introduce a dual-branch fusion module that employs geometric depth completion and semantic radar PillarNet to comprehensively leverage geometric and semantic information within each modality. Then we introduce an object-oriented attention module that employs localization-aware cross-attention to facilitate deep interactions across modalites by allowing queries in bird's-eye view (BEV) to attend to interested image tokens. We validate our SGDet3D on the TJ4DRadSet and View-of-Delft (VoD) datasets. Experimental results demonstrate that SGDet3D effectively fuses 4D radar data and camera image and achieves state-of-the-art performance.
Published in: IEEE Robotics and Automation Letters ( Volume: 10, Issue: 1, January 2025)
Page(s): 828 - 835
Date of Publication: 09 December 2024

ISSN Information:

Funding Agency:


Contact IEEE to Subscribe

References

References is not available for this document.