BND*-DDQN: learn to steer autonomously through deep reinforcement learning

It is vital for mobile robots to achieve safe autonomous steering in various changing environments. In this paper, a novel end-to-end network architecture is proposed for mobile robots to learn steering autonomously through deep reinforcement learning. Specifically, two sets of feature representatio...

Full description

Saved in:
Bibliographic Details
Main Authors: Wu, Keyu, Wang, Han, Esfahani, Mahdi Abolfazli, Yuan, Shenghai
Other Authors: School of Electrical and Electronic Engineering
Format: Article
Language:English
Published: 2022
Subjects:
Online Access:https://hdl.handle.net/10356/159818
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-159818
record_format dspace
spelling sg-ntu-dr.10356-1598182022-07-04T03:25:00Z BND*-DDQN: learn to steer autonomously through deep reinforcement learning Wu, Keyu Wang, Han Esfahani, Mahdi Abolfazli Yuan, Shenghai School of Electrical and Electronic Engineering Engineering::Electrical and electronic engineering Feature Extraction Training It is vital for mobile robots to achieve safe autonomous steering in various changing environments. In this paper, a novel end-to-end network architecture is proposed for mobile robots to learn steering autonomously through deep reinforcement learning. Specifically, two sets of feature representations are first extracted from the depth inputs through two different input streams. The acquired features are then merged together to derive both linear and angular actions simultaneously. Moreover, a new action selection strategy is also introduced to achieve motion filtering by taking the consistency in angular velocity into account. Besides, in addition to the extrinsic rewards, the intrinsic bonuses are also adopted during training to improve the exploration capability. Furthermore, it is worth noting the proposed model is readily transferable from the simple virtual training environment to much more complicated real-world scenarios so that no further fine-tuning is required for real deployment. Compared to the existing methods, the proposed method demonstrates significant superiority in terms of average reward, convergence speed, success rate, and generalization capability. In addition, it exhibits outstanding performance in various cluttered real-world environments containing both static and dynamic obstacles. A video of our experiments can be found at https://youtu.be/19jrQGG1oCU. Nanyang Technological University National Research Foundation (NRF) This work was supported in part by ST Engineering-NTU Corporate Laboratory and in part by NRF. 2022-07-04T03:25:00Z 2022-07-04T03:25:00Z 2019 Journal Article Wu, K., Wang, H., Esfahani, M. A. & Yuan, S. (2019). BND*-DDQN: learn to steer autonomously through deep reinforcement learning. IEEE Transactions On Cognitive and Developmental Systems, 13(2), 249-261. https://dx.doi.org/10.1109/TCDS.2019.2928820 2379-8920 https://hdl.handle.net/10356/159818 10.1109/TCDS.2019.2928820 2-s2.0-85069924995 2 13 249 261 en IEEE Transactions on Cognitive and Developmental Systems © 2019 IEEE. All rights reserved.
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering
Feature Extraction
Training
spellingShingle Engineering::Electrical and electronic engineering
Feature Extraction
Training
Wu, Keyu
Wang, Han
Esfahani, Mahdi Abolfazli
Yuan, Shenghai
BND*-DDQN: learn to steer autonomously through deep reinforcement learning
description It is vital for mobile robots to achieve safe autonomous steering in various changing environments. In this paper, a novel end-to-end network architecture is proposed for mobile robots to learn steering autonomously through deep reinforcement learning. Specifically, two sets of feature representations are first extracted from the depth inputs through two different input streams. The acquired features are then merged together to derive both linear and angular actions simultaneously. Moreover, a new action selection strategy is also introduced to achieve motion filtering by taking the consistency in angular velocity into account. Besides, in addition to the extrinsic rewards, the intrinsic bonuses are also adopted during training to improve the exploration capability. Furthermore, it is worth noting the proposed model is readily transferable from the simple virtual training environment to much more complicated real-world scenarios so that no further fine-tuning is required for real deployment. Compared to the existing methods, the proposed method demonstrates significant superiority in terms of average reward, convergence speed, success rate, and generalization capability. In addition, it exhibits outstanding performance in various cluttered real-world environments containing both static and dynamic obstacles. A video of our experiments can be found at https://youtu.be/19jrQGG1oCU.
author2 School of Electrical and Electronic Engineering
author_facet School of Electrical and Electronic Engineering
Wu, Keyu
Wang, Han
Esfahani, Mahdi Abolfazli
Yuan, Shenghai
format Article
author Wu, Keyu
Wang, Han
Esfahani, Mahdi Abolfazli
Yuan, Shenghai
author_sort Wu, Keyu
title BND*-DDQN: learn to steer autonomously through deep reinforcement learning
title_short BND*-DDQN: learn to steer autonomously through deep reinforcement learning
title_full BND*-DDQN: learn to steer autonomously through deep reinforcement learning
title_fullStr BND*-DDQN: learn to steer autonomously through deep reinforcement learning
title_full_unstemmed BND*-DDQN: learn to steer autonomously through deep reinforcement learning
title_sort bnd*-ddqn: learn to steer autonomously through deep reinforcement learning
publishDate 2022
url https://hdl.handle.net/10356/159818
_version_ 1738844826811498496