ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments
With the rapid development of artificial intelligence, robots, and autonomous driving technologies, visual SLAM technology has received extensive attention from research communities. However, the current research of visual SLAM systems is mainly based on static and simple environments, and the syste...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Master by Coursework |
Language: | English |
Published: |
Nanyang Technological University
2022
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/154873 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-154873 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1548732023-07-04T17:42:03Z ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments Chen, Peiyu Xie Lihua School of Electrical and Electronic Engineering ELHXIE@ntu.edu.sg Engineering::Electrical and electronic engineering With the rapid development of artificial intelligence, robots, and autonomous driving technologies, visual SLAM technology has received extensive attention from research communities. However, the current research of visual SLAM systems is mainly based on static and simple environments, and the system performance could be severely degraded in complex environments. Navigation and mapping in dynamic environment is a very challenging problem for autonomous robots. In this dissertation, we develop semantic SLAM by combining ORB-SLAM3 with YOLOv3 neural network. Our proposed system includes five parallel threads: semantic segmentation, tracking, local mapping, loop and map merging and ATLAS. ORB-SLAM3-YOLOv3 uses YOLOv3 to preprocess the image and segment the prior dynamic objects in frames. Then we use black mask to cover the dynamic objects to reduce the impact of the dynamic objects. Finally, we test the accuracy of the proposed system under Ubuntu 16.04. Experimental results show that our proposed method can effectively reduce the influence of dynamic objects on the TUM and KITTI dataset. The absolute trajectory accuracy in ORB-SLAM3-YOLOv3 can be improved compared with ORB-SLAM3. The computational time of our SLAM system can achieve 120ms per frame with CPU. Master of Science (Computer Control and Automation) 2022-01-13T23:20:54Z 2022-01-13T23:20:54Z 2021 Thesis-Master by Coursework Chen, P. (2021). ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/154873 https://hdl.handle.net/10356/154873 en application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Electrical and electronic engineering |
spellingShingle |
Engineering::Electrical and electronic engineering Chen, Peiyu ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments |
description |
With the rapid development of artificial intelligence, robots, and autonomous driving technologies, visual SLAM technology has received extensive attention from research communities. However, the current research of visual SLAM systems is mainly based on static and simple environments, and the system performance could be severely degraded in complex environments. Navigation and mapping in dynamic environment is a very challenging problem for autonomous robots. In this dissertation, we develop semantic SLAM by combining ORB-SLAM3 with YOLOv3 neural network. Our proposed system includes five parallel threads: semantic segmentation, tracking, local mapping, loop and map merging and ATLAS. ORB-SLAM3-YOLOv3 uses YOLOv3 to preprocess the image and segment the prior dynamic objects in frames. Then we use black mask to cover the dynamic objects to reduce the impact of the dynamic objects. Finally, we test the accuracy of the proposed system under Ubuntu 16.04. Experimental results show that our proposed method can effectively reduce the influence of dynamic objects on the TUM and KITTI dataset. The absolute trajectory accuracy in ORB-SLAM3-YOLOv3 can be improved compared with ORB-SLAM3. The computational time of our SLAM system can achieve 120ms per frame with CPU. |
author2 |
Xie Lihua |
author_facet |
Xie Lihua Chen, Peiyu |
format |
Thesis-Master by Coursework |
author |
Chen, Peiyu |
author_sort |
Chen, Peiyu |
title |
ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments |
title_short |
ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments |
title_full |
ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments |
title_fullStr |
ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments |
title_full_unstemmed |
ORB-SLAM3-YOLOv3 : a visual SLAM based on deep learning for dynamic environments |
title_sort |
orb-slam3-yolov3 : a visual slam based on deep learning for dynamic environments |
publisher |
Nanyang Technological University |
publishDate |
2022 |
url |
https://hdl.handle.net/10356/154873 |
_version_ |
1772827788733054976 |