From reinforcement learning to classical path planning: motion planning with obstacle avoidance

This project investigates the comparative performance of Reinforcement Learning (RL) and sampling-based motion planning methods in robotics, focusing on obstacle avoidance, illustrated in a 3D and 2D environment respectively with a singular agent and obstacle present. This is broken down into two ph...

Full description

Saved in:
Bibliographic Details
Main Author: Ng, Tze Minh
Other Authors: Yeo Chai Kiat
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2024
Subjects:
Online Access:https://hdl.handle.net/10356/181149
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-181149
record_format dspace
spelling sg-ntu-dr.10356-1811492024-11-18T00:46:50Z From reinforcement learning to classical path planning: motion planning with obstacle avoidance Ng, Tze Minh Yeo Chai Kiat College of Computing and Data Science ASCKYEO@ntu.edu.sg Computer and Information Science Reinforcement learning Motion planning This project investigates the comparative performance of Reinforcement Learning (RL) and sampling-based motion planning methods in robotics, focusing on obstacle avoidance, illustrated in a 3D and 2D environment respectively with a singular agent and obstacle present. This is broken down into two phases. The approach involves first replicating the results of a chosen research paper on Soft Actor Critic with Prioritised Experience Replay (SACPER) and running it on a simulation software. Then, a comparative analysis of different sampling-based motion planning algorithms is generated. Through this process, insights into how differing scenarios and tasks call for different methods for optimal performance will be uncovered. Phase 1 involving the implementation of SACPER was unable to learn due to a stagnant reward curve, which necessitated the need for increased time and computing resources. Phase 2 investigated how sampling-based methods performed in a 2D environment based on slight changes in the environment. Overall, this project contributes to the understanding of motion planning for robotics, emphasizing the strengths and limitations of learning and sampling-based strategies. Future developments in considering a hybrid approach between learning and sampling-based strategies could be pioneered. Bachelor's degree 2024-11-18T00:46:50Z 2024-11-18T00:46:50Z 2024 Final Year Project (FYP) Ng, T. M. (2024). From reinforcement learning to classical path planning: motion planning with obstacle avoidance. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/181149 https://hdl.handle.net/10356/181149 en SCSE23-1186 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Computer and Information Science
Reinforcement learning
Motion planning
spellingShingle Computer and Information Science
Reinforcement learning
Motion planning
Ng, Tze Minh
From reinforcement learning to classical path planning: motion planning with obstacle avoidance
description This project investigates the comparative performance of Reinforcement Learning (RL) and sampling-based motion planning methods in robotics, focusing on obstacle avoidance, illustrated in a 3D and 2D environment respectively with a singular agent and obstacle present. This is broken down into two phases. The approach involves first replicating the results of a chosen research paper on Soft Actor Critic with Prioritised Experience Replay (SACPER) and running it on a simulation software. Then, a comparative analysis of different sampling-based motion planning algorithms is generated. Through this process, insights into how differing scenarios and tasks call for different methods for optimal performance will be uncovered. Phase 1 involving the implementation of SACPER was unable to learn due to a stagnant reward curve, which necessitated the need for increased time and computing resources. Phase 2 investigated how sampling-based methods performed in a 2D environment based on slight changes in the environment. Overall, this project contributes to the understanding of motion planning for robotics, emphasizing the strengths and limitations of learning and sampling-based strategies. Future developments in considering a hybrid approach between learning and sampling-based strategies could be pioneered.
author2 Yeo Chai Kiat
author_facet Yeo Chai Kiat
Ng, Tze Minh
format Final Year Project
author Ng, Tze Minh
author_sort Ng, Tze Minh
title From reinforcement learning to classical path planning: motion planning with obstacle avoidance
title_short From reinforcement learning to classical path planning: motion planning with obstacle avoidance
title_full From reinforcement learning to classical path planning: motion planning with obstacle avoidance
title_fullStr From reinforcement learning to classical path planning: motion planning with obstacle avoidance
title_full_unstemmed From reinforcement learning to classical path planning: motion planning with obstacle avoidance
title_sort from reinforcement learning to classical path planning: motion planning with obstacle avoidance
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/181149
_version_ 1816858935782539264