End to end autonomous driving simulation based on reinforcement learning
This paper presents a comprehensive study that explores the application of reinforcement learning (RL) algorithms, specifically Deep Q-Network (DQN) and Soft Actor Critic (SAC), in the context of end-to-end autonomous driving. The research project utilizes the SMARTS Simulator, an open-source softwa...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/177154 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
Summary: | This paper presents a comprehensive study that explores the application of reinforcement learning (RL) algorithms, specifically Deep Q-Network (DQN) and Soft Actor Critic (SAC), in the context of end-to-end autonomous driving. The research project utilizes the SMARTS Simulator, an open-source software tailored for RL applications in autonomous driving scenarios. Employing an end-to-end approach, the research project utilizes RGB image inputs into an Artificial Neural Network, with image recognition facilitated by the Vision Transformer model.
The study begins with a review of RL theory, covering key concepts such as the Markov Decision Process, exploration-exploitation strategies, policy, rewards, value functions, and the taxonomy of RL methods. Following this, the report introduces the DQN and SAC algorithms, providing insights into their applications in autonomous driving scenarios. Additionally, the study explores the Vision Transformer model for image recognition tasks within the autonomous driving domain.
Subsequently, the paper outlines the experimental setup, detailing the environment, scenario, action space, reward structure, and termination conditions specific to the autonomous driving context. The design and implementation section clarifies the architectural framework and the specific implementation strategies within the SMARTS Simulator environment. Finally, the paper presents and discusses empirical results of applying both DQN and SAC algorithm from the experiments conducted. |
---|