Robust loop closure by textual cues in challenging environments
Loop closure is an important task in robot navigation. However, existing methods mostly rely on some implicit or heuristic features of the environment, which can still fail to work in common environments such as corridors, tunnels, and warehouses. Indeed, navigating in such featureless, degenerative...
Saved in:
Main Authors: | , , , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2025
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/182125 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-182125 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1821252025-01-10T15:44:00Z Robust loop closure by textual cues in challenging environments Jin, Tongxing Nguyen,Thien-Minh Xu, Xinhang Yang, Yizhuo Yuan, Shenghai Li, Jianping Xie, Lihua School of Electrical and Electronic Engineering Center for Advanced Robotics Technology Innovation (CARTIN) Computer and Information Science Loop closure Localization Loop closure is an important task in robot navigation. However, existing methods mostly rely on some implicit or heuristic features of the environment, which can still fail to work in common environments such as corridors, tunnels, and warehouses. Indeed, navigating in such featureless, degenerative, and repetitive (FDR) environments would also pose a significant challenge even for humans, but explicit text cues in the surroundings often provide the best assistance. This inspires us to propose a multi-modal loop closure method based on explicit human-readable textual cues in FDR environments. Specifically, our approach first extracts scene text entities based on Optical Character Recognition (OCR), then creates a \textit{local} map of text cues based on accurate LiDAR odometry and finally identifies loop closure events by a graph-theoretic scheme. Experiment results demonstrate that this approach has superior performance over existing methods that rely solely on visual and LiDAR sensors. To benefit the community, we release the source code and datasets at https://github.com/TongxingJin/TXTLCD. National Research Foundation (NRF) Submitted/Accepted version This work was supported by National Research Foundation, Singapore, under its Medium-Sized Center for Advanced Robotics Technology Innovation (CARTIN). 2025-01-09T05:41:09Z 2025-01-09T05:41:09Z 2024 Journal Article Jin, T., Nguyen, T., Xu, X., Yang, Y., Yuan, S., Li, J. & Xie, L. (2024). Robust loop closure by textual cues in challenging environments. IEEE Robotics and Automation Letters, 10(1), 812 --819. https://dx.doi.org/10.1109/LRA.2024.3511397 2377-3766 https://hdl.handle.net/10356/182125 10.1109/LRA.2024.3511397 1 10 812 - 819 en IEEE Robotics and Automation Letters © 2025 IEEE. All rights reserved. This article may be downloaded for personal use only. Any other use requires prior permission of the copyright holder. The Version of Record is available online at http://doi.org/10.1109/LRA.2024.3511397. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science Loop closure Localization |
spellingShingle |
Computer and Information Science Loop closure Localization Jin, Tongxing Nguyen,Thien-Minh Xu, Xinhang Yang, Yizhuo Yuan, Shenghai Li, Jianping Xie, Lihua Robust loop closure by textual cues in challenging environments |
description |
Loop closure is an important task in robot navigation. However, existing methods mostly rely on some implicit or heuristic features of the environment, which can still fail to work in common environments such as corridors, tunnels, and warehouses. Indeed, navigating in such featureless, degenerative, and repetitive (FDR) environments would also pose a significant challenge even for humans, but explicit text cues in the surroundings often provide the best assistance.
This inspires us to propose a multi-modal loop closure method based on explicit human-readable textual cues in FDR environments. Specifically, our approach first extracts scene text entities based on Optical Character Recognition (OCR), then creates a \textit{local} map of text cues based on accurate LiDAR odometry and finally identifies loop closure events by a graph-theoretic scheme.
Experiment results demonstrate that this approach has superior performance over existing methods that rely solely on visual and LiDAR sensors.
To benefit the community, we release the source code and datasets at https://github.com/TongxingJin/TXTLCD. |
author2 |
School of Electrical and Electronic Engineering |
author_facet |
School of Electrical and Electronic Engineering Jin, Tongxing Nguyen,Thien-Minh Xu, Xinhang Yang, Yizhuo Yuan, Shenghai Li, Jianping Xie, Lihua |
format |
Article |
author |
Jin, Tongxing Nguyen,Thien-Minh Xu, Xinhang Yang, Yizhuo Yuan, Shenghai Li, Jianping Xie, Lihua |
author_sort |
Jin, Tongxing |
title |
Robust loop closure by textual cues in challenging environments |
title_short |
Robust loop closure by textual cues in challenging environments |
title_full |
Robust loop closure by textual cues in challenging environments |
title_fullStr |
Robust loop closure by textual cues in challenging environments |
title_full_unstemmed |
Robust loop closure by textual cues in challenging environments |
title_sort |
robust loop closure by textual cues in challenging environments |
publishDate |
2025 |
url |
https://hdl.handle.net/10356/182125 |
_version_ |
1821237190996262912 |