3D spatial perception for underwater robots using point cloud data from orthogonal multibeam sonars fusion
Enhancing 3D spatial perception for underwater robots is crucial in improving their capability to carry out complex operations. However, this is a challenging problem due to the severe visibility for optical and sparse spatial data for acoustic imaging sensors underwater. To address these problems,...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Doctor of Philosophy |
Language: | English |
Published: |
Nanyang Technological University
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/172750 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-172750 |
---|---|
record_format |
dspace |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Mechanical engineering::Robots |
spellingShingle |
Engineering::Mechanical engineering::Robots Nicholas Sadjoli 3D spatial perception for underwater robots using point cloud data from orthogonal multibeam sonars fusion |
description |
Enhancing 3D spatial perception for underwater robots is crucial in improving their capability to carry out complex operations. However, this is a challenging problem due to the severe visibility for optical and sparse spatial data for acoustic imaging sensors underwater. To address these problems, Orthogonal Multibeam Sonar Fusion (OMSF) was previously developed in the literature, producing 3D point cloud data (PCD) using fusion from a pair of multibeam forward-looking sonars (MFLS) orthogonal to each other. Orthogonal orientation refers to a configuration where the image planes between a ‘horizontal’ and ‘vertical’ sensor pair are oriented perpendicularly (90 degrees) to each other. Forward-oriented MFLS creates 2D images of the environment based on intensities of multiple beamformed acoustic signals reflected from objects. While there exists research on OMSF for environment mapping, no prior controlled testing to determine OMSF accuracy and sensitivity to operational and environmental factors, was published. Methodology 3D data for perception beyond mapping and object scanning has also not been explored. This thesis presents several works to address these research gaps. First, controlled simulation and pool tests are performed to determine OMSF sensitivity and accuracy for factors of sensor frequency, object scale, object shape, and relative sensor rotation. Test results show the method has on average 38.09% higher accuracy using high frequency sensors on larger scaled objects and is up to 43% more accurate for objects with solid surfaces than hollow frames. This work also shows that with proper compensation, OMSF accuracy is robust against relative sensor rotations as long as the line-of-sight of the target is maintained. To the best of our knowledge, these results present the first known controlled documentation of OMSF sensitivity, and confirm that this method carries over and preserves known properties of singular MFLS sonars. This work will be useful for applications with known target objects, such as automated garage docking, allowing assessment of method suitability and result optimization.
Next, an application scenario of automated garage docking for an Autonomous Underwater Vehicle (AUVs) was devised to assess the feasibility of OMSF for classification and pose estimation, using simulation and pool tests. First, a PCD-based classification technique was developed by integrating OMSF PCDs as input to a binary classifier based on PointNet++, and trained on low cost offline object PCD scans. To address data sparsity, an OMSF-based volumetric filtering method was developed to re-include dense points from raw sonar features into the input. Pool test results show higher efficiency for low granularity sampling, and the volumetric filtering enables classification to achieve 25% and 37% respectively higher success rate and confidence, compared to using inputs from raw 3D projection of sonar features. Furthermore, an MFLS-based pose estimation technique was proposed, consisting of object width normalization to address limited training data for pose regression, a deterministic bounding box regression using Orthogonal Feature Matching (OFM), and an image-based relative pose regression. Pool test results show that OFM bounding box regression produces 4.28% higher mean Intersection over Union (IoU) and a 10% increase for the (> 25%) IoU metric compared to methods based on standard MFLS filtering, and that pose regression using Convolutional Neural Network achieves the highest success rate compared to other tested methods. Results also show the proposed end-to-end technique having < 10^o error, comparable to those of existing optical-based methods. In conclusion, these works present novel use cases of OMSF for perception-based applications, resulting in
PCD-based classification with low training cost and achieving better performance compared to using naive 3D sonar feature projection, along with MFLS-based pose estimation that has comparable performance to existing optical-based methods and inherently more robust in turbid waters. |
author2 |
Cai Yiyu |
author_facet |
Cai Yiyu Nicholas Sadjoli |
format |
Thesis-Doctor of Philosophy |
author |
Nicholas Sadjoli |
author_sort |
Nicholas Sadjoli |
title |
3D spatial perception for underwater robots using point cloud data from orthogonal multibeam sonars fusion |
title_short |
3D spatial perception for underwater robots using point cloud data from orthogonal multibeam sonars fusion |
title_full |
3D spatial perception for underwater robots using point cloud data from orthogonal multibeam sonars fusion |
title_fullStr |
3D spatial perception for underwater robots using point cloud data from orthogonal multibeam sonars fusion |
title_full_unstemmed |
3D spatial perception for underwater robots using point cloud data from orthogonal multibeam sonars fusion |
title_sort |
3d spatial perception for underwater robots using point cloud data from orthogonal multibeam sonars fusion |
publisher |
Nanyang Technological University |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/172750 |
_version_ |
1787590723891101696 |
spelling |
sg-ntu-dr.10356-1727502024-01-04T06:32:51Z 3D spatial perception for underwater robots using point cloud data from orthogonal multibeam sonars fusion Nicholas Sadjoli Cai Yiyu School of Mechanical and Aerospace Engineering Saab Singapore Pte. Ltd. MYYCai@ntu.edu.sg Engineering::Mechanical engineering::Robots Enhancing 3D spatial perception for underwater robots is crucial in improving their capability to carry out complex operations. However, this is a challenging problem due to the severe visibility for optical and sparse spatial data for acoustic imaging sensors underwater. To address these problems, Orthogonal Multibeam Sonar Fusion (OMSF) was previously developed in the literature, producing 3D point cloud data (PCD) using fusion from a pair of multibeam forward-looking sonars (MFLS) orthogonal to each other. Orthogonal orientation refers to a configuration where the image planes between a ‘horizontal’ and ‘vertical’ sensor pair are oriented perpendicularly (90 degrees) to each other. Forward-oriented MFLS creates 2D images of the environment based on intensities of multiple beamformed acoustic signals reflected from objects. While there exists research on OMSF for environment mapping, no prior controlled testing to determine OMSF accuracy and sensitivity to operational and environmental factors, was published. Methodology 3D data for perception beyond mapping and object scanning has also not been explored. This thesis presents several works to address these research gaps. First, controlled simulation and pool tests are performed to determine OMSF sensitivity and accuracy for factors of sensor frequency, object scale, object shape, and relative sensor rotation. Test results show the method has on average 38.09% higher accuracy using high frequency sensors on larger scaled objects and is up to 43% more accurate for objects with solid surfaces than hollow frames. This work also shows that with proper compensation, OMSF accuracy is robust against relative sensor rotations as long as the line-of-sight of the target is maintained. To the best of our knowledge, these results present the first known controlled documentation of OMSF sensitivity, and confirm that this method carries over and preserves known properties of singular MFLS sonars. This work will be useful for applications with known target objects, such as automated garage docking, allowing assessment of method suitability and result optimization. Next, an application scenario of automated garage docking for an Autonomous Underwater Vehicle (AUVs) was devised to assess the feasibility of OMSF for classification and pose estimation, using simulation and pool tests. First, a PCD-based classification technique was developed by integrating OMSF PCDs as input to a binary classifier based on PointNet++, and trained on low cost offline object PCD scans. To address data sparsity, an OMSF-based volumetric filtering method was developed to re-include dense points from raw sonar features into the input. Pool test results show higher efficiency for low granularity sampling, and the volumetric filtering enables classification to achieve 25% and 37% respectively higher success rate and confidence, compared to using inputs from raw 3D projection of sonar features. Furthermore, an MFLS-based pose estimation technique was proposed, consisting of object width normalization to address limited training data for pose regression, a deterministic bounding box regression using Orthogonal Feature Matching (OFM), and an image-based relative pose regression. Pool test results show that OFM bounding box regression produces 4.28% higher mean Intersection over Union (IoU) and a 10% increase for the (> 25%) IoU metric compared to methods based on standard MFLS filtering, and that pose regression using Convolutional Neural Network achieves the highest success rate compared to other tested methods. Results also show the proposed end-to-end technique having < 10^o error, comparable to those of existing optical-based methods. In conclusion, these works present novel use cases of OMSF for perception-based applications, resulting in PCD-based classification with low training cost and achieving better performance compared to using naive 3D sonar feature projection, along with MFLS-based pose estimation that has comparable performance to existing optical-based methods and inherently more robust in turbid waters. Doctor of Philosophy 2023-12-19T07:50:36Z 2023-12-19T07:50:36Z 2023 Thesis-Doctor of Philosophy Nicholas Sadjoli (2023). 3D spatial perception for underwater robots using point cloud data from orthogonal multibeam sonars fusion. Doctoral thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/172750 https://hdl.handle.net/10356/172750 10.32657/10356/172750 en This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0). application/pdf Nanyang Technological University |