CoFiI2P: coarse-to-fine correspondences-based image to point cloud registration
Image-to-point cloud (I2P) registration is a fundamental task for robots and autonomous vehicles to achieve cross-modality data fusion and localization. Current I2P registration methods primarily focus on estimating correspondences at the point or pixel level, often neglecting global alignment. As a...
Saved in:
Main Authors: | , , , , , , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2025
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/182367 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-182367 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1823672025-01-27T01:42:02Z CoFiI2P: coarse-to-fine correspondences-based image to point cloud registration Kang, Shuhao Liao, Youqi Li, Jianping Liang, Fuxun Li, Yuhao Zou, Xianghong Li, Fangning Chen, Xieyuanli Dong, Zhen Yang, Bisheng School of Electrical and Electronic Engineering Delta-NTU Corporate Laboratory Engineering Coarse-to-fine correspondences Transformer network Image-to-point cloud (I2P) registration is a fundamental task for robots and autonomous vehicles to achieve cross-modality data fusion and localization. Current I2P registration methods primarily focus on estimating correspondences at the point or pixel level, often neglecting global alignment. As a result, I2P matching can easily converge to a local optimum if it lacks high-level guidance from global constraints. To improve the success rate and general robustness, this letter introduces CoFiI2P, a novel I2P registration network that extracts correspondences in a coarse-to-fine manner. First, the image and point cloud data are processed through a two-stream encoder-decoder network for hierarchical feature extraction. Second, a coarse-to-fine matching module is designed to leverage these features and establish robust feature correspondences. Specifically, in the coarse matching phase, a novel I2P transformer module is employed to capture both homogeneous and heterogeneous global information from the image and point cloud data. This enables the estimation of coarse super-point/super-pixel matching pairs with discriminative descriptors. In the fine matching module, point/pixel pairs are established with the guidance of super-point/super-pixel correspondences. Finally, based on matching pairs, the transformation matrix is estimated with the EPnP-RANSAC algorithm. Experiments conducted on the KITTI Odometry dataset demonstrate that CoFiI2P achieves impressive results, with a relative rotation error (RRE) of 1.14 degrees and a relative translation error (RTE) of 0.29 meters, while maintaining real-time speed. These results represent a significant improvement of 84% in RRE and 89% in RTE compared to the current state-of-the-art (SOTA) method. Additional experiments on the Nuscenes dataset confirm our method's generalizability. This work was supported in part by the National Natural Science Foundation under Project 42201477 and Project 42130105, in part by the Open Fund of Hubei Luojia Laboratory under Grant 2201000054, and in part by Open Fund of Key Laboratory of Urban Spatial Information, Ministry of Natural Resources under Grant 2023ZD001. 2025-01-27T01:42:02Z 2025-01-27T01:42:02Z 2024 Journal Article Kang, S., Liao, Y., Li, J., Liang, F., Li, Y., Zou, X., Li, F., Chen, X., Dong, Z. & Yang, B. (2024). CoFiI2P: coarse-to-fine correspondences-based image to point cloud registration. IEEE Robotics and Automation Letters, 9(11), 10264-10271. https://dx.doi.org/10.1109/LRA.2024.3466068 2377-3766 https://hdl.handle.net/10356/182367 10.1109/LRA.2024.3466068 2-s2.0-85205519040 11 9 10264 10271 en IEEE Robotics and Automation Letters © 2024 IEEE. All rights reserved. |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering Coarse-to-fine correspondences Transformer network |
spellingShingle |
Engineering Coarse-to-fine correspondences Transformer network Kang, Shuhao Liao, Youqi Li, Jianping Liang, Fuxun Li, Yuhao Zou, Xianghong Li, Fangning Chen, Xieyuanli Dong, Zhen Yang, Bisheng CoFiI2P: coarse-to-fine correspondences-based image to point cloud registration |
description |
Image-to-point cloud (I2P) registration is a fundamental task for robots and autonomous vehicles to achieve cross-modality data fusion and localization. Current I2P registration methods primarily focus on estimating correspondences at the point or pixel level, often neglecting global alignment. As a result, I2P matching can easily converge to a local optimum if it lacks high-level guidance from global constraints. To improve the success rate and general robustness, this letter introduces CoFiI2P, a novel I2P registration network that extracts correspondences in a coarse-to-fine manner. First, the image and point cloud data are processed through a two-stream encoder-decoder network for hierarchical feature extraction. Second, a coarse-to-fine matching module is designed to leverage these features and establish robust feature correspondences. Specifically, in the coarse matching phase, a novel I2P transformer module is employed to capture both homogeneous and heterogeneous global information from the image and point cloud data. This enables the estimation of coarse super-point/super-pixel matching pairs with discriminative descriptors. In the fine matching module, point/pixel pairs are established with the guidance of super-point/super-pixel correspondences. Finally, based on matching pairs, the transformation matrix is estimated with the EPnP-RANSAC algorithm. Experiments conducted on the KITTI Odometry dataset demonstrate that CoFiI2P achieves impressive results, with a relative rotation error (RRE) of 1.14 degrees and a relative translation error (RTE) of 0.29 meters, while maintaining real-time speed. These results represent a significant improvement of 84% in RRE and 89% in RTE compared to the current state-of-the-art (SOTA) method. Additional experiments on the Nuscenes dataset confirm our method's generalizability. |
author2 |
School of Electrical and Electronic Engineering |
author_facet |
School of Electrical and Electronic Engineering Kang, Shuhao Liao, Youqi Li, Jianping Liang, Fuxun Li, Yuhao Zou, Xianghong Li, Fangning Chen, Xieyuanli Dong, Zhen Yang, Bisheng |
format |
Article |
author |
Kang, Shuhao Liao, Youqi Li, Jianping Liang, Fuxun Li, Yuhao Zou, Xianghong Li, Fangning Chen, Xieyuanli Dong, Zhen Yang, Bisheng |
author_sort |
Kang, Shuhao |
title |
CoFiI2P: coarse-to-fine correspondences-based image to point cloud registration |
title_short |
CoFiI2P: coarse-to-fine correspondences-based image to point cloud registration |
title_full |
CoFiI2P: coarse-to-fine correspondences-based image to point cloud registration |
title_fullStr |
CoFiI2P: coarse-to-fine correspondences-based image to point cloud registration |
title_full_unstemmed |
CoFiI2P: coarse-to-fine correspondences-based image to point cloud registration |
title_sort |
cofii2p: coarse-to-fine correspondences-based image to point cloud registration |
publishDate |
2025 |
url |
https://hdl.handle.net/10356/182367 |
_version_ |
1823108706844278784 |