A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation
This paper proposes a solution to the problem of mobile robotic localization using visual indoor image sequences with a biologically inspired spatio-temporal neural network approach. The system contains three major subsystems: a feature extraction module, a scene quantization module and a spatio-tem...
Saved in:
Main Authors: | , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2013
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/107260 http://hdl.handle.net/10220/17664 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-107260 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1072602020-05-28T07:19:13Z A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation Nguyen, Vu Anh Goh, Wooi Boon Starzyk, Janusz A. School of Computer Engineering DRNTU::Engineering::Computer science and engineering This paper proposes a solution to the problem of mobile robotic localization using visual indoor image sequences with a biologically inspired spatio-temporal neural network approach. The system contains three major subsystems: a feature extraction module, a scene quantization module and a spatio-temporal long-term memory (LTM) module. During learning, the scene quantization module clusters the visual images set into scene tokens. A K-Iteration Fast Learning Artificial Neural Network (KFLANN) is employed as the core unit of the quantization module. The KFLANN network is driven by intrinsic statistics of the data stream and therefore does not require the number of clusters to be predefined. In addition, the KFLANN performance is less sensitive to data presentation ordering compared to popular clustering methods such as k-means, and can therefore produce a consistent number of stable centroids. Using scene tokens, the topological structure of the environment can be composed into sequences of tokens. These sequences are then learnt and stored in memory units in an LTM architecture, which is able to continuously and robustly recognize the visual input stream. The design of memory units addresses two critical problems in spatio-temporal learning, namely error tolerance and memory forgetting. The primary objective of this work is to explore the synergy between the strength of KFLANN and LTM models to address the visual topological localization problem. We demonstrate the efficiency and efficacy of the proposed framework on the challenging COsy Localization Dataset. 2013-11-15T05:07:02Z 2019-12-06T22:27:31Z 2013-11-15T05:07:02Z 2019-12-06T22:27:31Z 2013 2013 Journal Article Nguyen, V. A., Starzyk, J. A., & Goh, W. B. (2013). A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation. Robotics and autonomous systems, 61(2), 1744-1758. 0921-8890 https://hdl.handle.net/10356/107260 http://hdl.handle.net/10220/17664 10.1016/j.robot.2012.12.004 en Robotics and autonomous systems |
institution |
Nanyang Technological University |
building |
NTU Library |
country |
Singapore |
collection |
DR-NTU |
language |
English |
topic |
DRNTU::Engineering::Computer science and engineering |
spellingShingle |
DRNTU::Engineering::Computer science and engineering Nguyen, Vu Anh Goh, Wooi Boon Starzyk, Janusz A. A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation |
description |
This paper proposes a solution to the problem of mobile robotic localization using visual indoor image sequences with a biologically inspired spatio-temporal neural network approach. The system contains three major subsystems: a feature extraction module, a scene quantization module and a spatio-temporal long-term memory (LTM) module. During learning, the scene quantization module clusters the visual images set into scene tokens. A K-Iteration Fast Learning Artificial Neural Network (KFLANN) is employed as the core unit of the quantization module. The KFLANN network is driven by intrinsic statistics of the data stream and therefore does not require the number of clusters to be predefined. In addition, the KFLANN performance is less sensitive to data presentation ordering compared to popular clustering methods such as k-means, and can therefore produce a consistent number of stable centroids. Using scene tokens, the topological structure of the environment can be composed into sequences of tokens. These sequences are then learnt and stored in memory units in an LTM architecture, which is able to continuously and robustly recognize the visual input stream. The design of memory units addresses two critical problems in spatio-temporal learning, namely error tolerance and memory forgetting. The primary objective of this work is to explore the synergy between the strength of KFLANN and LTM models to address the visual topological localization problem. We demonstrate the efficiency and efficacy of the proposed framework on the challenging COsy Localization Dataset. |
author2 |
School of Computer Engineering |
author_facet |
School of Computer Engineering Nguyen, Vu Anh Goh, Wooi Boon Starzyk, Janusz A. |
format |
Article |
author |
Nguyen, Vu Anh Goh, Wooi Boon Starzyk, Janusz A. |
author_sort |
Nguyen, Vu Anh |
title |
A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation |
title_short |
A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation |
title_full |
A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation |
title_fullStr |
A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation |
title_full_unstemmed |
A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation |
title_sort |
spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation |
publishDate |
2013 |
url |
https://hdl.handle.net/10356/107260 http://hdl.handle.net/10220/17664 |
_version_ |
1681058082246885376 |