Deep learning feature-based visual SLAM

Visual SLAM is often used in autonomous agents for localization and mapping. A key element of visual SLAM is the feature detector-descriptor used to extract tracking keypoints from images of the environment. Classical visual SLAM algorithms rely on hand-crafted methods, which have low computation...

全面介紹

Saved in:
書目詳細資料
主要作者: Yong, Duan Kai
其他作者: Lam Siew Kei
格式: Final Year Project
語言:English
出版: Nanyang Technological University 2023
主題:
在線閱讀:https://hdl.handle.net/10356/166008
標簽: 添加標簽
沒有標簽, 成為第一個標記此記錄!
實物特徵
總結:Visual SLAM is often used in autonomous agents for localization and mapping. A key element of visual SLAM is the feature detector-descriptor used to extract tracking keypoints from images of the environment. Classical visual SLAM algorithms rely on hand-crafted methods, which have low computational complexity but suffer from poor accuracy. Deep-learned feature detector descriptors have the potential to improve feature accuracy and robustness and thus tracking performance. This paper aims to compare the advantages and disadvantages of classical and deeplearning feature detector-descriptors for use in mono-SLAM through a review of literature and comparing the experimental performance of ORB-SLAM2 and SuperPoint-SLAM. A broad set of experiments covering multiple conditions such as dynamic scenes, illumination changes, and fast motion have been performed to determine baseline performance as well as reveal which conditions favour each detector-descriptor type.