iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning

This paper proposes iTD3-CLN, a Deep Reinforcement Learning (DRL) based low-level motion controller, to achieve map-less autonomous navigation in dynamic scene. We consider three enhancements to the Twin Delayed DDPG (TD3) for the navigation task: N-step returns, Priority Experience Replay, and a ch...

Full description

Saved in:
Bibliographic Details
Main Authors: Jiang, Haoge, Esfahani, Mahdi Abolfazli, Wu, Keyu, Wan, Kong-wah, Heng, Kuan-kian, Wang, Han, Jiang, Xudong
Other Authors: School of Electrical and Electronic Engineering
Format: Article
Language:English
Published: 2022
Subjects:
Online Access:https://hdl.handle.net/10356/163356
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-163356
record_format dspace
spelling sg-ntu-dr.10356-1633562022-12-05T01:58:19Z iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning Jiang, Haoge Esfahani, Mahdi Abolfazli Wu, Keyu Wan, Kong-wah Heng, Kuan-kian Wang, Han Jiang, Xudong School of Electrical and Electronic Engineering Engineering::Electrical and electronic engineering Deep Reinforcement Learning Collision Avoidance This paper proposes iTD3-CLN, a Deep Reinforcement Learning (DRL) based low-level motion controller, to achieve map-less autonomous navigation in dynamic scene. We consider three enhancements to the Twin Delayed DDPG (TD3) for the navigation task: N-step returns, Priority Experience Replay, and a channel-based Convolutional Laser Network (CLN) architecture. In contrast to the conventional methods such as the DWA, our approach is found superior in the following ways: no need for prior knowledge of the environment and metric map, lower reliance on an accurate sensor, learning emergent behavior in dynamic scene that is intuitive, and more remarkably, able to transfer to the real robot without further fine-tuning. Our extensive studies show that in comparison to the original TD3, the proposed approach can obtain approximately 50% reduction in training to get same performance, 50% higher accumulated reward, and 30–50% increase in generalization performance when tested in unseen environments. Videos of our experiments are available at https://youtu.be/BRN0Gk5oLOc (Simulation) and https://youtu.be/yIxGH9TPQCc (Real experiment). Agency for Science, Technology and Research (A*STAR) This research is partly supported by A*STAR Grant No. 192 2500049 from the National Robotics Programme (NRP) , Singapore. 2022-12-05T01:58:19Z 2022-12-05T01:58:19Z 2022 Journal Article Jiang, H., Esfahani, M. A., Wu, K., Wan, K., Heng, K., Wang, H. & Jiang, X. (2022). iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning. Neurocomputing, 503, 118-128. https://dx.doi.org/10.1016/j.neucom.2022.06.102 0925-2312 https://hdl.handle.net/10356/163356 10.1016/j.neucom.2022.06.102 2-s2.0-85133902711 503 118 128 en 192 2500049 Neurocomputing © 2022 Elsevier B.V. All rights reserved.
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering
Deep Reinforcement Learning
Collision Avoidance
spellingShingle Engineering::Electrical and electronic engineering
Deep Reinforcement Learning
Collision Avoidance
Jiang, Haoge
Esfahani, Mahdi Abolfazli
Wu, Keyu
Wan, Kong-wah
Heng, Kuan-kian
Wang, Han
Jiang, Xudong
iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning
description This paper proposes iTD3-CLN, a Deep Reinforcement Learning (DRL) based low-level motion controller, to achieve map-less autonomous navigation in dynamic scene. We consider three enhancements to the Twin Delayed DDPG (TD3) for the navigation task: N-step returns, Priority Experience Replay, and a channel-based Convolutional Laser Network (CLN) architecture. In contrast to the conventional methods such as the DWA, our approach is found superior in the following ways: no need for prior knowledge of the environment and metric map, lower reliance on an accurate sensor, learning emergent behavior in dynamic scene that is intuitive, and more remarkably, able to transfer to the real robot without further fine-tuning. Our extensive studies show that in comparison to the original TD3, the proposed approach can obtain approximately 50% reduction in training to get same performance, 50% higher accumulated reward, and 30–50% increase in generalization performance when tested in unseen environments. Videos of our experiments are available at https://youtu.be/BRN0Gk5oLOc (Simulation) and https://youtu.be/yIxGH9TPQCc (Real experiment).
author2 School of Electrical and Electronic Engineering
author_facet School of Electrical and Electronic Engineering
Jiang, Haoge
Esfahani, Mahdi Abolfazli
Wu, Keyu
Wan, Kong-wah
Heng, Kuan-kian
Wang, Han
Jiang, Xudong
format Article
author Jiang, Haoge
Esfahani, Mahdi Abolfazli
Wu, Keyu
Wan, Kong-wah
Heng, Kuan-kian
Wang, Han
Jiang, Xudong
author_sort Jiang, Haoge
title iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning
title_short iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning
title_full iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning
title_fullStr iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning
title_full_unstemmed iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning
title_sort itd3-cln: learn to navigate in dynamic scene through deep reinforcement learning
publishDate 2022
url https://hdl.handle.net/10356/163356
_version_ 1751548590654750720