A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation

This paper proposes a solution to the problem of mobile robotic localization using visual indoor image sequences with a biologically inspired spatio-temporal neural network approach. The system contains three major subsystems: a feature extraction module, a scene quantization module and a spatio-tem...

Full description

Saved in:
Bibliographic Details
Main Authors: Nguyen, Vu Anh, Goh, Wooi Boon, Starzyk, Janusz A.
Other Authors: School of Computer Engineering
Format: Article
Language:English
Published: 2013
Subjects:
Online Access:https://hdl.handle.net/10356/107260
http://hdl.handle.net/10220/17664
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-107260
record_format dspace
spelling sg-ntu-dr.10356-1072602020-05-28T07:19:13Z A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation Nguyen, Vu Anh Goh, Wooi Boon Starzyk, Janusz A. School of Computer Engineering DRNTU::Engineering::Computer science and engineering This paper proposes a solution to the problem of mobile robotic localization using visual indoor image sequences with a biologically inspired spatio-temporal neural network approach. The system contains three major subsystems: a feature extraction module, a scene quantization module and a spatio-temporal long-term memory (LTM) module. During learning, the scene quantization module clusters the visual images set into scene tokens. A K-Iteration Fast Learning Artificial Neural Network (KFLANN) is employed as the core unit of the quantization module. The KFLANN network is driven by intrinsic statistics of the data stream and therefore does not require the number of clusters to be predefined. In addition, the KFLANN performance is less sensitive to data presentation ordering compared to popular clustering methods such as k-means, and can therefore produce a consistent number of stable centroids. Using scene tokens, the topological structure of the environment can be composed into sequences of tokens. These sequences are then learnt and stored in memory units in an LTM architecture, which is able to continuously and robustly recognize the visual input stream. The design of memory units addresses two critical problems in spatio-temporal learning, namely error tolerance and memory forgetting. The primary objective of this work is to explore the synergy between the strength of KFLANN and LTM models to address the visual topological localization problem. We demonstrate the efficiency and efficacy of the proposed framework on the challenging COsy Localization Dataset. 2013-11-15T05:07:02Z 2019-12-06T22:27:31Z 2013-11-15T05:07:02Z 2019-12-06T22:27:31Z 2013 2013 Journal Article Nguyen, V. A., Starzyk, J. A., & Goh, W. B. (2013). A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation. Robotics and autonomous systems, 61(2), 1744-1758. 0921-8890 https://hdl.handle.net/10356/107260 http://hdl.handle.net/10220/17664 10.1016/j.robot.2012.12.004 en Robotics and autonomous systems
institution Nanyang Technological University
building NTU Library
country Singapore
collection DR-NTU
language English
topic DRNTU::Engineering::Computer science and engineering
spellingShingle DRNTU::Engineering::Computer science and engineering
Nguyen, Vu Anh
Goh, Wooi Boon
Starzyk, Janusz A.
A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation
description This paper proposes a solution to the problem of mobile robotic localization using visual indoor image sequences with a biologically inspired spatio-temporal neural network approach. The system contains three major subsystems: a feature extraction module, a scene quantization module and a spatio-temporal long-term memory (LTM) module. During learning, the scene quantization module clusters the visual images set into scene tokens. A K-Iteration Fast Learning Artificial Neural Network (KFLANN) is employed as the core unit of the quantization module. The KFLANN network is driven by intrinsic statistics of the data stream and therefore does not require the number of clusters to be predefined. In addition, the KFLANN performance is less sensitive to data presentation ordering compared to popular clustering methods such as k-means, and can therefore produce a consistent number of stable centroids. Using scene tokens, the topological structure of the environment can be composed into sequences of tokens. These sequences are then learnt and stored in memory units in an LTM architecture, which is able to continuously and robustly recognize the visual input stream. The design of memory units addresses two critical problems in spatio-temporal learning, namely error tolerance and memory forgetting. The primary objective of this work is to explore the synergy between the strength of KFLANN and LTM models to address the visual topological localization problem. We demonstrate the efficiency and efficacy of the proposed framework on the challenging COsy Localization Dataset.
author2 School of Computer Engineering
author_facet School of Computer Engineering
Nguyen, Vu Anh
Goh, Wooi Boon
Starzyk, Janusz A.
format Article
author Nguyen, Vu Anh
Goh, Wooi Boon
Starzyk, Janusz A.
author_sort Nguyen, Vu Anh
title A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation
title_short A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation
title_full A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation
title_fullStr A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation
title_full_unstemmed A spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation
title_sort spatio-temporal long-term memory approach for visual place recognition in mobile robotic navigation
publishDate 2013
url https://hdl.handle.net/10356/107260
http://hdl.handle.net/10220/17664
_version_ 1681058082246885376