ObjectFusion: Multi-modal 3D object detection with object-centric fusion
Recent progress on multi-modal 3D object detection has featured BEV (Bird-Eye-View) based fusion, which effectively unifies both LiDAR point clouds and camera images in a shared BEV space. Nevertheless, it is not trivial to perform camera-to-BEV transformation due to the inherently ambiguous depth e...
Saved in:
Main Authors: | , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2023
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/8306 https://ink.library.smu.edu.sg/context/sis_research/article/9309/viewcontent/Cai_ObjectFusion_Multi_modal_3D_Object_Detection_with_Object_Centric_Fusion_ICCV_2023_paper.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-9309 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-93092023-12-05T03:19:20Z ObjectFusion: Multi-modal 3D object detection with object-centric fusion CAI, Q. PAN, Y. YAO, T. NGO, Chong-wah MEI, T. Recent progress on multi-modal 3D object detection has featured BEV (Bird-Eye-View) based fusion, which effectively unifies both LiDAR point clouds and camera images in a shared BEV space. Nevertheless, it is not trivial to perform camera-to-BEV transformation due to the inherently ambiguous depth estimation of each pixel, resulting in spatial misalignment between these two multi-modal features. Moreover, such transformation also inevitably leads to projection distortion of camera image features in BEV space. In this paper, we propose a novel Object-centric Fusion (ObjectFusion) paradigm, which completely gets rid of camera-to-BEV transformation during fusion to align object-centric features across different modalities for 3D object detection. ObjectFusion first learns three kinds of modality-specific feature maps (i.e., voxel, BEV, and image features) from LiDAR point clouds and its BEV projections, camera images. Then a set of 3D object proposals are produced from the BEV features via a heatmap-based proposal generator. Next, the 3D object proposals are reprojected back to voxel, BEV, and image spaces. We leverage voxel and RoI pooling to generate spatially aligned object-centric features for each modality. All the object-centric features of three modalities are further fused at object level, which is finally fed into the detection heads. Extensive experiments on nuScenes dataset demonstrate the superiority of our ObjectFusion, by achieving 69.8% mAP on nuScenes validation set and improving BEVFusion by 1.3%. 2023-10-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8306 https://ink.library.smu.edu.sg/context/sis_research/article/9309/viewcontent/Cai_ObjectFusion_Multi_modal_3D_Object_Detection_with_Object_Centric_Fusion_ICCV_2023_paper.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University 3D object detection Multi-modal Fusion-based approach Artificial Intelligence and Robotics Robotics |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
3D object detection Multi-modal Fusion-based approach Artificial Intelligence and Robotics Robotics |
spellingShingle |
3D object detection Multi-modal Fusion-based approach Artificial Intelligence and Robotics Robotics CAI, Q. PAN, Y. YAO, T. NGO, Chong-wah MEI, T. ObjectFusion: Multi-modal 3D object detection with object-centric fusion |
description |
Recent progress on multi-modal 3D object detection has featured BEV (Bird-Eye-View) based fusion, which effectively unifies both LiDAR point clouds and camera images in a shared BEV space. Nevertheless, it is not trivial to perform camera-to-BEV transformation due to the inherently ambiguous depth estimation of each pixel, resulting in spatial misalignment between these two multi-modal features. Moreover, such transformation also inevitably leads to projection distortion of camera image features in BEV space. In this paper, we propose a novel Object-centric Fusion (ObjectFusion) paradigm, which completely gets rid of camera-to-BEV transformation during fusion to align object-centric features across different modalities for 3D object detection. ObjectFusion first learns three kinds of modality-specific feature maps (i.e., voxel, BEV, and image features) from LiDAR point clouds and its BEV projections, camera images. Then a set of 3D object proposals are produced from the BEV features via a heatmap-based proposal generator. Next, the 3D object proposals are reprojected back to voxel, BEV, and image spaces. We leverage voxel and RoI pooling to generate spatially aligned object-centric features for each modality. All the object-centric features of three modalities are further fused at object level, which is finally fed into the detection heads. Extensive experiments on nuScenes dataset demonstrate the superiority of our ObjectFusion, by achieving 69.8% mAP on nuScenes validation set and improving BEVFusion by 1.3%. |
format |
text |
author |
CAI, Q. PAN, Y. YAO, T. NGO, Chong-wah MEI, T. |
author_facet |
CAI, Q. PAN, Y. YAO, T. NGO, Chong-wah MEI, T. |
author_sort |
CAI, Q. |
title |
ObjectFusion: Multi-modal 3D object detection with object-centric fusion |
title_short |
ObjectFusion: Multi-modal 3D object detection with object-centric fusion |
title_full |
ObjectFusion: Multi-modal 3D object detection with object-centric fusion |
title_fullStr |
ObjectFusion: Multi-modal 3D object detection with object-centric fusion |
title_full_unstemmed |
ObjectFusion: Multi-modal 3D object detection with object-centric fusion |
title_sort |
objectfusion: multi-modal 3d object detection with object-centric fusion |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2023 |
url |
https://ink.library.smu.edu.sg/sis_research/8306 https://ink.library.smu.edu.sg/context/sis_research/article/9309/viewcontent/Cai_ObjectFusion_Multi_modal_3D_Object_Detection_with_Object_Centric_Fusion_ICCV_2023_paper.pdf |
_version_ |
1784855627891212288 |