ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments

With the rapid development of artificial intelligence, robots, and autonomous driving technologies, visual SLAM technology has received extensive attention from research communities. However, the current research of visual SLAM systems is mainly based on static and simple environments, and the syste...

Full description

Saved in:
Bibliographic Details
Main Author: Chen, Peiyu
Other Authors: Xie Lihua
Format: Thesis-Master by Coursework
Language:English
Published: Nanyang Technological University 2022
Subjects:
Online Access:https://hdl.handle.net/10356/154873
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-154873
record_format dspace
spelling sg-ntu-dr.10356-1548732023-07-04T17:42:03Z ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments Chen, Peiyu Xie Lihua School of Electrical and Electronic Engineering ELHXIE@ntu.edu.sg Engineering::Electrical and electronic engineering With the rapid development of artificial intelligence, robots, and autonomous driving technologies, visual SLAM technology has received extensive attention from research communities. However, the current research of visual SLAM systems is mainly based on static and simple environments, and the system performance could be severely degraded in complex environments. Navigation and mapping in dynamic environment is a very challenging problem for autonomous robots. In this dissertation, we develop semantic SLAM by combining ORB-SLAM3 with YOLOv3 neural network. Our proposed system includes five parallel threads: semantic segmentation, tracking, local mapping, loop and map merging and ATLAS. ORB-SLAM3-YOLOv3 uses YOLOv3 to preprocess the image and segment the prior dynamic objects in frames. Then we use black mask to cover the dynamic objects to reduce the impact of the dynamic objects. Finally, we test the accuracy of the proposed system under Ubuntu 16.04. Experimental results show that our proposed method can effectively reduce the influence of dynamic objects on the TUM and KITTI dataset. The absolute trajectory accuracy in ORB-SLAM3-YOLOv3 can be improved compared with ORB-SLAM3. The computational time of our SLAM system can achieve 120ms per frame with CPU. Master of Science (Computer Control and Automation) 2022-01-13T23:20:54Z 2022-01-13T23:20:54Z 2021 Thesis-Master by Coursework Chen, P. (2021). ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/154873 https://hdl.handle.net/10356/154873 en application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering
spellingShingle Engineering::Electrical and electronic engineering
Chen, Peiyu
ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments
description With the rapid development of artificial intelligence, robots, and autonomous driving technologies, visual SLAM technology has received extensive attention from research communities. However, the current research of visual SLAM systems is mainly based on static and simple environments, and the system performance could be severely degraded in complex environments. Navigation and mapping in dynamic environment is a very challenging problem for autonomous robots. In this dissertation, we develop semantic SLAM by combining ORB-SLAM3 with YOLOv3 neural network. Our proposed system includes five parallel threads: semantic segmentation, tracking, local mapping, loop and map merging and ATLAS. ORB-SLAM3-YOLOv3 uses YOLOv3 to preprocess the image and segment the prior dynamic objects in frames. Then we use black mask to cover the dynamic objects to reduce the impact of the dynamic objects. Finally, we test the accuracy of the proposed system under Ubuntu 16.04. Experimental results show that our proposed method can effectively reduce the influence of dynamic objects on the TUM and KITTI dataset. The absolute trajectory accuracy in ORB-SLAM3-YOLOv3 can be improved compared with ORB-SLAM3. The computational time of our SLAM system can achieve 120ms per frame with CPU.
author2 Xie Lihua
author_facet Xie Lihua
Chen, Peiyu
format Thesis-Master by Coursework
author Chen, Peiyu
author_sort Chen, Peiyu
title ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments
title_short ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments
title_full ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments
title_fullStr ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments
title_full_unstemmed ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments
title_sort orb-slam3-yolov3 : a visual slam based on deep learning for dynamic environments
publisher Nanyang Technological University
publishDate 2022
url https://hdl.handle.net/10356/154873
_version_ 1772827788733054976