From reinforcement learning to classical path planning: motion planning with obstacle avoidance
This project investigates the comparative performance of Reinforcement Learning (RL) and sampling-based motion planning methods in robotics, focusing on obstacle avoidance, illustrated in a 3D and 2D environment respectively with a singular agent and obstacle present. This is broken down into two ph...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/181149 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-181149 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1811492024-11-18T00:46:50Z From reinforcement learning to classical path planning: motion planning with obstacle avoidance Ng, Tze Minh Yeo Chai Kiat College of Computing and Data Science ASCKYEO@ntu.edu.sg Computer and Information Science Reinforcement learning Motion planning This project investigates the comparative performance of Reinforcement Learning (RL) and sampling-based motion planning methods in robotics, focusing on obstacle avoidance, illustrated in a 3D and 2D environment respectively with a singular agent and obstacle present. This is broken down into two phases. The approach involves first replicating the results of a chosen research paper on Soft Actor Critic with Prioritised Experience Replay (SACPER) and running it on a simulation software. Then, a comparative analysis of different sampling-based motion planning algorithms is generated. Through this process, insights into how differing scenarios and tasks call for different methods for optimal performance will be uncovered. Phase 1 involving the implementation of SACPER was unable to learn due to a stagnant reward curve, which necessitated the need for increased time and computing resources. Phase 2 investigated how sampling-based methods performed in a 2D environment based on slight changes in the environment. Overall, this project contributes to the understanding of motion planning for robotics, emphasizing the strengths and limitations of learning and sampling-based strategies. Future developments in considering a hybrid approach between learning and sampling-based strategies could be pioneered. Bachelor's degree 2024-11-18T00:46:50Z 2024-11-18T00:46:50Z 2024 Final Year Project (FYP) Ng, T. M. (2024). From reinforcement learning to classical path planning: motion planning with obstacle avoidance. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/181149 https://hdl.handle.net/10356/181149 en SCSE23-1186 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science Reinforcement learning Motion planning |
spellingShingle |
Computer and Information Science Reinforcement learning Motion planning Ng, Tze Minh From reinforcement learning to classical path planning: motion planning with obstacle avoidance |
description |
This project investigates the comparative performance of Reinforcement Learning (RL) and sampling-based motion planning methods in robotics, focusing on obstacle avoidance, illustrated in a 3D and 2D environment respectively with a singular agent and obstacle present. This is broken down into two phases. The approach involves first replicating the results of a chosen research paper on Soft Actor Critic with Prioritised Experience Replay (SACPER) and running it on a simulation software. Then, a comparative analysis of different sampling-based motion planning algorithms is generated. Through this process, insights into how differing scenarios and tasks call for different methods for optimal performance will be uncovered.
Phase 1 involving the implementation of SACPER was unable to learn due to a stagnant reward curve, which necessitated the need for increased time and computing resources. Phase 2 investigated how sampling-based methods performed in a 2D environment based on slight changes in the environment.
Overall, this project contributes to the understanding of motion planning for robotics, emphasizing the strengths and limitations of learning and sampling-based strategies. Future developments in considering a hybrid approach between learning and sampling-based strategies could be pioneered. |
author2 |
Yeo Chai Kiat |
author_facet |
Yeo Chai Kiat Ng, Tze Minh |
format |
Final Year Project |
author |
Ng, Tze Minh |
author_sort |
Ng, Tze Minh |
title |
From reinforcement learning to classical path planning: motion planning with obstacle avoidance |
title_short |
From reinforcement learning to classical path planning: motion planning with obstacle avoidance |
title_full |
From reinforcement learning to classical path planning: motion planning with obstacle avoidance |
title_fullStr |
From reinforcement learning to classical path planning: motion planning with obstacle avoidance |
title_full_unstemmed |
From reinforcement learning to classical path planning: motion planning with obstacle avoidance |
title_sort |
from reinforcement learning to classical path planning: motion planning with obstacle avoidance |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/181149 |
_version_ |
1816858935782539264 |