Reinforcement learning based algorithm design for mobile robot dynamic obstacle avoidance

A good path planning strategy is an essential part of the autonomous navigation system when moving in a new environment. In the real case, there are not only static obstacles but also dynamic obstacles in the surrounding environment. In the future, drone transportation may become the main method of...

Full description

Saved in:
Bibliographic Details
Main Author: Gu, Bochen
Other Authors: Hu Guoqiang
Format: Thesis-Master by Coursework
Language:English
Published: Nanyang Technological University 2021
Subjects:
Online Access:https://hdl.handle.net/10356/152376
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-152376
record_format dspace
spelling sg-ntu-dr.10356-1523762023-07-04T17:41:14Z Reinforcement learning based algorithm design for mobile robot dynamic obstacle avoidance Gu, Bochen Hu Guoqiang School of Electrical and Electronic Engineering GQHu@ntu.edu.sg Engineering::Electrical and electronic engineering::Control and instrumentation::Robotics A good path planning strategy is an essential part of the autonomous navigation system when moving in a new environment. In the real case, there are not only static obstacles but also dynamic obstacles in the surrounding environment. In the future, drone transportation may become the main method of logistics, which, commonly, multiple robots work together in a limited space. In some traditional path planning strategies, the routes of the robots are pre-defined since the environment is already known. In the contrast, however, in most cases, the robot cannot get global information about the environment but only can detect the surrounding environment. It is necessary for robots to learn the changes in the environment. In this dissertation, a grid world environment with static and dynamic obstacles is set, while reinforcement learning and two kinds of deep Q networks are used to do the training. After the training, the agent will know how to avoid obstacles when going forward to the terminal. Firstly, the basic theory of Q learning and deep Q network is introduced. In Q learning, the position of the agent is taken as the states. And set the epsilon greedy algorithm as the exploration strategy and updating the reward dictionary every step since the environment is always changing. After that, double DQN and dueling DQN are used to improve the performance of it in dynamic environments. At last, the networks are improved by the PER policy, and their performances are compared. Master of Science (Computer Control and Automation) 2021-08-06T03:23:15Z 2021-08-06T03:23:15Z 2021 Thesis-Master by Coursework Gu, B. (2021). Reinforcement learning based algorithm design for mobile robot dynamic obstacle avoidance. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/152376 https://hdl.handle.net/10356/152376 en application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering::Control and instrumentation::Robotics
spellingShingle Engineering::Electrical and electronic engineering::Control and instrumentation::Robotics
Gu, Bochen
Reinforcement learning based algorithm design for mobile robot dynamic obstacle avoidance
description A good path planning strategy is an essential part of the autonomous navigation system when moving in a new environment. In the real case, there are not only static obstacles but also dynamic obstacles in the surrounding environment. In the future, drone transportation may become the main method of logistics, which, commonly, multiple robots work together in a limited space. In some traditional path planning strategies, the routes of the robots are pre-defined since the environment is already known. In the contrast, however, in most cases, the robot cannot get global information about the environment but only can detect the surrounding environment. It is necessary for robots to learn the changes in the environment. In this dissertation, a grid world environment with static and dynamic obstacles is set, while reinforcement learning and two kinds of deep Q networks are used to do the training. After the training, the agent will know how to avoid obstacles when going forward to the terminal. Firstly, the basic theory of Q learning and deep Q network is introduced. In Q learning, the position of the agent is taken as the states. And set the epsilon greedy algorithm as the exploration strategy and updating the reward dictionary every step since the environment is always changing. After that, double DQN and dueling DQN are used to improve the performance of it in dynamic environments. At last, the networks are improved by the PER policy, and their performances are compared.
author2 Hu Guoqiang
author_facet Hu Guoqiang
Gu, Bochen
format Thesis-Master by Coursework
author Gu, Bochen
author_sort Gu, Bochen
title Reinforcement learning based algorithm design for mobile robot dynamic obstacle avoidance
title_short Reinforcement learning based algorithm design for mobile robot dynamic obstacle avoidance
title_full Reinforcement learning based algorithm design for mobile robot dynamic obstacle avoidance
title_fullStr Reinforcement learning based algorithm design for mobile robot dynamic obstacle avoidance
title_full_unstemmed Reinforcement learning based algorithm design for mobile robot dynamic obstacle avoidance
title_sort reinforcement learning based algorithm design for mobile robot dynamic obstacle avoidance
publisher Nanyang Technological University
publishDate 2021
url https://hdl.handle.net/10356/152376
_version_ 1772826915676094464