iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning
This paper proposes iTD3-CLN, a Deep Reinforcement Learning (DRL) based low-level motion controller, to achieve map-less autonomous navigation in dynamic scene. We consider three enhancements to the Twin Delayed DDPG (TD3) for the navigation task: N-step returns, Priority Experience Replay, and a ch...
Saved in:
Main Authors: | , , , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2022
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/163356 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-163356 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1633562022-12-05T01:58:19Z iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning Jiang, Haoge Esfahani, Mahdi Abolfazli Wu, Keyu Wan, Kong-wah Heng, Kuan-kian Wang, Han Jiang, Xudong School of Electrical and Electronic Engineering Engineering::Electrical and electronic engineering Deep Reinforcement Learning Collision Avoidance This paper proposes iTD3-CLN, a Deep Reinforcement Learning (DRL) based low-level motion controller, to achieve map-less autonomous navigation in dynamic scene. We consider three enhancements to the Twin Delayed DDPG (TD3) for the navigation task: N-step returns, Priority Experience Replay, and a channel-based Convolutional Laser Network (CLN) architecture. In contrast to the conventional methods such as the DWA, our approach is found superior in the following ways: no need for prior knowledge of the environment and metric map, lower reliance on an accurate sensor, learning emergent behavior in dynamic scene that is intuitive, and more remarkably, able to transfer to the real robot without further fine-tuning. Our extensive studies show that in comparison to the original TD3, the proposed approach can obtain approximately 50% reduction in training to get same performance, 50% higher accumulated reward, and 30–50% increase in generalization performance when tested in unseen environments. Videos of our experiments are available at https://youtu.be/BRN0Gk5oLOc (Simulation) and https://youtu.be/yIxGH9TPQCc (Real experiment). Agency for Science, Technology and Research (A*STAR) This research is partly supported by A*STAR Grant No. 192 2500049 from the National Robotics Programme (NRP) , Singapore. 2022-12-05T01:58:19Z 2022-12-05T01:58:19Z 2022 Journal Article Jiang, H., Esfahani, M. A., Wu, K., Wan, K., Heng, K., Wang, H. & Jiang, X. (2022). iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning. Neurocomputing, 503, 118-128. https://dx.doi.org/10.1016/j.neucom.2022.06.102 0925-2312 https://hdl.handle.net/10356/163356 10.1016/j.neucom.2022.06.102 2-s2.0-85133902711 503 118 128 en 192 2500049 Neurocomputing © 2022 Elsevier B.V. All rights reserved. |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Electrical and electronic engineering Deep Reinforcement Learning Collision Avoidance |
spellingShingle |
Engineering::Electrical and electronic engineering Deep Reinforcement Learning Collision Avoidance Jiang, Haoge Esfahani, Mahdi Abolfazli Wu, Keyu Wan, Kong-wah Heng, Kuan-kian Wang, Han Jiang, Xudong iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning |
description |
This paper proposes iTD3-CLN, a Deep Reinforcement Learning (DRL) based low-level motion controller, to achieve map-less autonomous navigation in dynamic scene. We consider three enhancements to the Twin Delayed DDPG (TD3) for the navigation task: N-step returns, Priority Experience Replay, and a channel-based Convolutional Laser Network (CLN) architecture. In contrast to the conventional methods such as the DWA, our approach is found superior in the following ways: no need for prior knowledge of the environment and metric map, lower reliance on an accurate sensor, learning emergent behavior in dynamic scene that is intuitive, and more remarkably, able to transfer to the real robot without further fine-tuning. Our extensive studies show that in comparison to the original TD3, the proposed approach can obtain approximately 50% reduction in training to get same performance, 50% higher accumulated reward, and 30–50% increase in generalization performance when tested in unseen environments. Videos of our experiments are available at https://youtu.be/BRN0Gk5oLOc (Simulation) and https://youtu.be/yIxGH9TPQCc (Real experiment). |
author2 |
School of Electrical and Electronic Engineering |
author_facet |
School of Electrical and Electronic Engineering Jiang, Haoge Esfahani, Mahdi Abolfazli Wu, Keyu Wan, Kong-wah Heng, Kuan-kian Wang, Han Jiang, Xudong |
format |
Article |
author |
Jiang, Haoge Esfahani, Mahdi Abolfazli Wu, Keyu Wan, Kong-wah Heng, Kuan-kian Wang, Han Jiang, Xudong |
author_sort |
Jiang, Haoge |
title |
iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning |
title_short |
iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning |
title_full |
iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning |
title_fullStr |
iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning |
title_full_unstemmed |
iTD3-CLN: learn to navigate in dynamic scene through Deep Reinforcement Learning |
title_sort |
itd3-cln: learn to navigate in dynamic scene through deep reinforcement learning |
publishDate |
2022 |
url |
https://hdl.handle.net/10356/163356 |
_version_ |
1751548590654750720 |