UR robot manipulator path planning via reinforcement learning

With the increasing demand of people's intelligent life, there have been new developments in robotic field. Combined with neural network, traditional reinforcement learning algorithms is also improved to adapt to the situation of high-dimensional continuous space. As the most basic step of mani...

Full description

Saved in:
Bibliographic Details
Main Author: Wang, Xiaoye
Other Authors: Hu Guoqiang
Format: Thesis-Master by Coursework
Language:English
Published: Nanyang Technological University 2021
Subjects:
Online Access:https://hdl.handle.net/10356/152896
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-152896
record_format dspace
spelling sg-ntu-dr.10356-1528962023-07-04T17:40:20Z UR robot manipulator path planning via reinforcement learning Wang, Xiaoye Hu Guoqiang School of Electrical and Electronic Engineering GQHu@ntu.edu.sg Engineering::Electrical and electronic engineering::Control and instrumentation::Robotics With the increasing demand of people's intelligent life, there have been new developments in robotic field. Combined with neural network, traditional reinforcement learning algorithms is also improved to adapt to the situation of high-dimensional continuous space. As the most basic step of manipulator motion control, path planning has also become more intelligent with the development of reinforcement learning. First of all, the path planned in the Cartesian space needs to be converted into a path of joint angles, which will play an important role in subsequent motion control. We use the D-H parameter method to model the kinematics of the manipulator, find out its forward and inverse kinematics and verify its correctness. Secondly, we discuss the obstacle avoidance effect and its limitations of the traditional obstacle avoidance method artificial potential field method in a known environment, and use the RRT algorithm to improve it when it falls into a local optimum. Thirdly, we use the DDPG algorithm in reinforcement learning to plan the trajectory, which is suitable for continuous state and action situation. Through the training of the model, the simple path planning between 2 given points of the robot arm is completed by training the RL model. We analyze the problems of sparse rewards and propose some possible solutions. Master of Science (Computer Control and Automation) 2021-10-14T04:46:50Z 2021-10-14T04:46:50Z 2021 Thesis-Master by Coursework Wang, X. (2021). UR robot manipulator path planning via reinforcement learning. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/152896 https://hdl.handle.net/10356/152896 en ISM-DISS-02242 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering::Control and instrumentation::Robotics
spellingShingle Engineering::Electrical and electronic engineering::Control and instrumentation::Robotics
Wang, Xiaoye
UR robot manipulator path planning via reinforcement learning
description With the increasing demand of people's intelligent life, there have been new developments in robotic field. Combined with neural network, traditional reinforcement learning algorithms is also improved to adapt to the situation of high-dimensional continuous space. As the most basic step of manipulator motion control, path planning has also become more intelligent with the development of reinforcement learning. First of all, the path planned in the Cartesian space needs to be converted into a path of joint angles, which will play an important role in subsequent motion control. We use the D-H parameter method to model the kinematics of the manipulator, find out its forward and inverse kinematics and verify its correctness. Secondly, we discuss the obstacle avoidance effect and its limitations of the traditional obstacle avoidance method artificial potential field method in a known environment, and use the RRT algorithm to improve it when it falls into a local optimum. Thirdly, we use the DDPG algorithm in reinforcement learning to plan the trajectory, which is suitable for continuous state and action situation. Through the training of the model, the simple path planning between 2 given points of the robot arm is completed by training the RL model. We analyze the problems of sparse rewards and propose some possible solutions.
author2 Hu Guoqiang
author_facet Hu Guoqiang
Wang, Xiaoye
format Thesis-Master by Coursework
author Wang, Xiaoye
author_sort Wang, Xiaoye
title UR robot manipulator path planning via reinforcement learning
title_short UR robot manipulator path planning via reinforcement learning
title_full UR robot manipulator path planning via reinforcement learning
title_fullStr UR robot manipulator path planning via reinforcement learning
title_full_unstemmed UR robot manipulator path planning via reinforcement learning
title_sort ur robot manipulator path planning via reinforcement learning
publisher Nanyang Technological University
publishDate 2021
url https://hdl.handle.net/10356/152896
_version_ 1772825510010683392