Leveraging deep generative models for non-parametric distributions in reinforcement learning
This thesis explores the use of deep generative models to enhance distribution representations in reinforcement learning (RL), leading to improved exploration, stability, and performance. It focuses on two roles of distributions in RL: policy distributions and action distributions. For policy distri...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Doctor of Philosophy |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/173455 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-173455 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1734552024-03-07T08:52:05Z Leveraging deep generative models for non-parametric distributions in reinforcement learning Tang, Shi Yuan Zhang Jie School of Computer Science and Engineering ZhangJ@ntu.edu.sg Computer and Information Science This thesis explores the use of deep generative models to enhance distribution representations in reinforcement learning (RL), leading to improved exploration, stability, and performance. It focuses on two roles of distributions in RL: policy distributions and action distributions. For policy distributions, an adversarial hypernetwork (AH) architecture enables multi-policy learning, allowing algorithms to converge to diverse local optima. The AH framework is generalized to single-policy RL algorithms, with a self-distillation mechanism for better learning efficiency. In the second part, the thesis investigates the benefits of using deep generative Fully-parameterized Quantile Function (FQF) in the actor of Soft Actor-Critic (SAC) to overcome uni-modality assumptions in stochastic policy implementations. The thesis demonstrates the theoretical boundedness of the entropy regularization term in FQF. Overall, this work proposes leveraging deep generative models to address performance inconsistencies and limitations of traditional modality assumptions in RL distributions. Doctor of Philosophy 2024-02-05T08:53:43Z 2024-02-05T08:53:43Z 2023 Thesis-Doctor of Philosophy Tang, S. Y. (2023). Leveraging deep generative models for non-parametric distributions in reinforcement learning. Doctoral thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/173455 https://hdl.handle.net/10356/173455 10.32657/10356/173455 en This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0). application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science |
spellingShingle |
Computer and Information Science Tang, Shi Yuan Leveraging deep generative models for non-parametric distributions in reinforcement learning |
description |
This thesis explores the use of deep generative models to enhance distribution representations in reinforcement learning (RL), leading to improved exploration, stability, and performance. It focuses on two roles of distributions in RL: policy distributions and action distributions. For policy distributions, an adversarial hypernetwork (AH) architecture enables multi-policy learning, allowing algorithms to converge to diverse local optima. The AH framework is generalized to single-policy RL algorithms, with a self-distillation mechanism for better learning efficiency. In the second part, the thesis investigates the benefits of using deep generative Fully-parameterized Quantile Function (FQF) in the actor of Soft Actor-Critic (SAC) to overcome uni-modality assumptions in stochastic policy implementations. The thesis demonstrates the theoretical boundedness of the entropy regularization term in FQF. Overall, this work proposes leveraging deep generative models to address performance inconsistencies and limitations of traditional modality assumptions in RL distributions. |
author2 |
Zhang Jie |
author_facet |
Zhang Jie Tang, Shi Yuan |
format |
Thesis-Doctor of Philosophy |
author |
Tang, Shi Yuan |
author_sort |
Tang, Shi Yuan |
title |
Leveraging deep generative models for non-parametric distributions in reinforcement learning |
title_short |
Leveraging deep generative models for non-parametric distributions in reinforcement learning |
title_full |
Leveraging deep generative models for non-parametric distributions in reinforcement learning |
title_fullStr |
Leveraging deep generative models for non-parametric distributions in reinforcement learning |
title_full_unstemmed |
Leveraging deep generative models for non-parametric distributions in reinforcement learning |
title_sort |
leveraging deep generative models for non-parametric distributions in reinforcement learning |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/173455 |
_version_ |
1794549355297898496 |