Sensor fusion for autonomous mobile robot
The sparsity of point clouds and lack of sufficient semantic information present significant challenges for existing LiDAR-only 3D detection methods, particularly in robotic applications that demand high accuracy and efficiency. To address point cloud sparsity, recent approaches have explored the co...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Master by Coursework |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/181815 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-181815 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1818152024-12-20T15:47:42Z Sensor fusion for autonomous mobile robot Yu, Zhuochen Andy Khong W H School of Electrical and Electronic Engineering AndyKhong@ntu.edu.sg Engineering Autonomous mobile robot The sparsity of point clouds and lack of sufficient semantic information present significant challenges for existing LiDAR-only 3D detection methods, particularly in robotic applications that demand high accuracy and efficiency. To address point cloud sparsity, recent approaches have explored the conversion of RGB images into virtual points through depth completion, enabling fusion with LiDAR data. While these methods improve point cloud density, they often introduce substantial computational overhead due to the high density of generated virtual points and do not fully exploit the rich semantic information from images. In this work, VKIFNet is introduced as an efficient multi-modal feature fusion framework designed to enhance 3D perception for robotic systems by integrating virtual key instances with LiDAR points across multiple stages. VKIFNet incor porates three core modules. First, SKIS (Semantic Key Instance Selection) is presented, which filters and preserves only essential virtual key instances while leveraging semantic information from virtual points. This approach significantly reduces computational demands and allows critical image-derived features to be retained in 3D space, crucial for efficient robotic operation. The second module is a new fusion technique called VIFF (Virtual Instance Focused Fusion), which performs multi-level fusion of Virtual Key Instances and LiDAR data in both BEV (Bird’s-Eye View) and 3D space. This fusion method enhances spatial awareness and ensures that both LiDAR and image-derived features contribute to a more robust understanding of the environment. Lastly, VIRA (Virtual-Instance-to-Real Attention) is introduced as a lightweight attention mechanism that utilizes the features of relevant LiDAR points to refine the virtual key instances with minimal computational overhead, optimizing the model for real-time robotic applications. VKIFNet demonstrates substantial improvements in detection performance on the KITTI and JRDB datasets, showcasing its potential for high-precision 3D perception in robotics. Master's degree 2024-12-20T12:20:34Z 2024-12-20T12:20:34Z 2024 Thesis-Master by Coursework Yu, Z. (2024). Sensor fusion for autonomous mobile robot. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/181815 https://hdl.handle.net/10356/181815 en application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering Autonomous mobile robot |
spellingShingle |
Engineering Autonomous mobile robot Yu, Zhuochen Sensor fusion for autonomous mobile robot |
description |
The sparsity of point clouds and lack of sufficient semantic information present significant challenges for existing LiDAR-only 3D detection methods, particularly in robotic applications that demand high accuracy and efficiency. To address point cloud sparsity, recent approaches have explored the conversion of RGB images into virtual points through depth completion, enabling fusion with LiDAR data. While these methods improve point cloud density, they often introduce substantial computational overhead due to the high density of generated virtual points and do not fully exploit the rich semantic information from images.
In this work, VKIFNet is introduced as an efficient multi-modal feature fusion framework designed to enhance 3D perception for robotic systems by integrating virtual key instances with LiDAR points across multiple stages. VKIFNet incor porates three core modules. First, SKIS (Semantic Key Instance Selection) is presented, which filters and preserves only essential virtual key instances while leveraging semantic information from virtual points. This approach significantly reduces computational demands and allows critical image-derived features to be retained in 3D space, crucial for efficient robotic operation. The second module is a new fusion technique called VIFF (Virtual Instance Focused Fusion), which performs multi-level fusion of Virtual Key Instances and LiDAR data in both BEV (Bird’s-Eye View) and 3D space. This fusion method enhances spatial awareness and ensures that both LiDAR and image-derived features contribute to a more robust understanding of the environment. Lastly, VIRA (Virtual-Instance-to-Real Attention) is introduced as a lightweight attention mechanism that utilizes the features of relevant LiDAR points to refine the virtual key instances with minimal computational overhead, optimizing the model for real-time robotic applications. VKIFNet demonstrates substantial improvements in detection performance on the KITTI and JRDB datasets, showcasing its potential for high-precision 3D perception in robotics. |
author2 |
Andy Khong W H |
author_facet |
Andy Khong W H Yu, Zhuochen |
format |
Thesis-Master by Coursework |
author |
Yu, Zhuochen |
author_sort |
Yu, Zhuochen |
title |
Sensor fusion for autonomous mobile robot |
title_short |
Sensor fusion for autonomous mobile robot |
title_full |
Sensor fusion for autonomous mobile robot |
title_fullStr |
Sensor fusion for autonomous mobile robot |
title_full_unstemmed |
Sensor fusion for autonomous mobile robot |
title_sort |
sensor fusion for autonomous mobile robot |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/181815 |
_version_ |
1819113053083402240 |