3D localization for unmanned vehicles using visual inputs

This thesis describes our efforts to tackle the 3D localization problem for unmanned vehicles using only visual data. Given the video stream of a camera, we wish to estimate the location of the camera accurately in real-time. We propose an indoor visual odometry system with a RGB-D camera pointing...

Full description

Saved in:
Bibliographic Details
Main Author: Mou, Wei
Other Authors: Wang Han
Format: Theses and Dissertations
Language:English
Published: 2017
Subjects:
Online Access:http://hdl.handle.net/10356/70516
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-70516
record_format dspace
spelling sg-ntu-dr.10356-705162023-07-04T17:24:53Z 3D localization for unmanned vehicles using visual inputs Mou, Wei Wang Han School of Electrical and Electronic Engineering DRNTU::Engineering::Electrical and electronic engineering This thesis describes our efforts to tackle the 3D localization problem for unmanned vehicles using only visual data. Given the video stream of a camera, we wish to estimate the location of the camera accurately in real-time. We propose an indoor visual odometry system with a RGB-D camera pointing at the ceiling. The term visual odometry is chosen for its functional similarity with wheel odometry which incrementally estimates the position of a robot by counting the number of turns of its wheels over time. Similarly, visual odometry estimates the position of the robot by integrating its motion changes inferred from the images that captured by the on-board cameras. The main contribution of this algorithm is the introduction of principal direction detection that can greatly reduce error accumulation problem in most visual odometry estimation approaches. The proposed approach can be operated in real-time and it performs well even with cameras disturbance. For robots working in outdoor environments, an efficient visual odometry system is developed. Keypoints are detected using the FAST detector. The proposed feature descriptor is designed in such a way that it is not only invariant to rotation and llumination changes but also the difference between two descriptors can be computed very efficiently using Intel Streaming SIMD Extensions (SSE) instruction. The feature matching process is accelerated using prior statistical analysis of maximum and minimum feature displacements. Experimental results show that the proposed system can perform accurate visual odometry very efficiently in outdoor environments. In order to localize distant objects on the sea surface, an automatic self-calibration approach for wide baseline stereo cameras using sea surface images is introduced. Compared to the traditional stereo calibration method using calibration pattern, the proposed self-calibration approach automatically estimate cameras’ rotation matrices for stereo rig using the sea horizon and a point at infinite distance. Doctor of Philosophy (EEE) 2017-04-26T06:02:49Z 2017-04-26T06:02:49Z 2017 Thesis Mou, W. (2017). 3D localization for unmanned vehicles using visual inputs. Doctoral thesis, Nanyang Technological University, Singapore. http://hdl.handle.net/10356/70516 10.32657/10356/70516 en 174 p. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic DRNTU::Engineering::Electrical and electronic engineering
spellingShingle DRNTU::Engineering::Electrical and electronic engineering
Mou, Wei
3D localization for unmanned vehicles using visual inputs
description This thesis describes our efforts to tackle the 3D localization problem for unmanned vehicles using only visual data. Given the video stream of a camera, we wish to estimate the location of the camera accurately in real-time. We propose an indoor visual odometry system with a RGB-D camera pointing at the ceiling. The term visual odometry is chosen for its functional similarity with wheel odometry which incrementally estimates the position of a robot by counting the number of turns of its wheels over time. Similarly, visual odometry estimates the position of the robot by integrating its motion changes inferred from the images that captured by the on-board cameras. The main contribution of this algorithm is the introduction of principal direction detection that can greatly reduce error accumulation problem in most visual odometry estimation approaches. The proposed approach can be operated in real-time and it performs well even with cameras disturbance. For robots working in outdoor environments, an efficient visual odometry system is developed. Keypoints are detected using the FAST detector. The proposed feature descriptor is designed in such a way that it is not only invariant to rotation and llumination changes but also the difference between two descriptors can be computed very efficiently using Intel Streaming SIMD Extensions (SSE) instruction. The feature matching process is accelerated using prior statistical analysis of maximum and minimum feature displacements. Experimental results show that the proposed system can perform accurate visual odometry very efficiently in outdoor environments. In order to localize distant objects on the sea surface, an automatic self-calibration approach for wide baseline stereo cameras using sea surface images is introduced. Compared to the traditional stereo calibration method using calibration pattern, the proposed self-calibration approach automatically estimate cameras’ rotation matrices for stereo rig using the sea horizon and a point at infinite distance.
author2 Wang Han
author_facet Wang Han
Mou, Wei
format Theses and Dissertations
author Mou, Wei
author_sort Mou, Wei
title 3D localization for unmanned vehicles using visual inputs
title_short 3D localization for unmanned vehicles using visual inputs
title_full 3D localization for unmanned vehicles using visual inputs
title_fullStr 3D localization for unmanned vehicles using visual inputs
title_full_unstemmed 3D localization for unmanned vehicles using visual inputs
title_sort 3d localization for unmanned vehicles using visual inputs
publishDate 2017
url http://hdl.handle.net/10356/70516
_version_ 1772828840224096256