Making reinforcement learning more data-efficient

Recently, researchers showed that applying deep neural networks to Reinforcement Learning (RL) can improve the agent’s ability to learn action policies directly from high dimensional input states. However, there are still several challenges present when applying these algorithms to real world...

Full description

Saved in:
Bibliographic Details
Main Author: Cao, Jeffery Siming
Other Authors: Tan Yap Peng
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2021
Subjects:
Online Access:https://hdl.handle.net/10356/149413
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-149413
record_format dspace
spelling sg-ntu-dr.10356-1494132023-07-07T18:31:39Z Making reinforcement learning more data-efficient Cao, Jeffery Siming Tan Yap Peng School of Electrical and Electronic Engineering A*STAR Bai Feng Jun Zhu Hai Yue EYPTan@ntu.edu.sg Engineering::Electrical and electronic engineering Recently, researchers showed that applying deep neural networks to Reinforcement Learning (RL) can improve the agent’s ability to learn action policies directly from high dimensional input states. However, there are still several challenges present when applying these algorithms to real world problems. One huge challenge is that they require a lot of interactions with the environment to learn a suitable action policy and have poor data-efficiency. In supervised learning tasks, data augmentation has been shown to improve the data-efficiency of deep learning models. This paper investigates the effects of data augmentation and other semi-supervised methods on a state-of-the-art RL algorithm, namely the Contrastive Unsupervised Representation for RL (CURL) algorithm. We demonstrate that data augmentation can improve the data-efficiency of the CURL algorithm when applied to the DeepMind Control Suite benchmark running on OpenAI Gym. Furthermore, we investigate possible applications of the Mean Teacher semi-supervised learning method to CURL Bachelor of Engineering (Electrical and Electronic Engineering) 2021-05-31T04:48:38Z 2021-05-31T04:48:38Z 2021 Final Year Project (FYP) Cao, J. S. (2021). Making reinforcement learning more data-efficient. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/149413 https://hdl.handle.net/10356/149413 en B3311-201 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering
spellingShingle Engineering::Electrical and electronic engineering
Cao, Jeffery Siming
Making reinforcement learning more data-efficient
description Recently, researchers showed that applying deep neural networks to Reinforcement Learning (RL) can improve the agent’s ability to learn action policies directly from high dimensional input states. However, there are still several challenges present when applying these algorithms to real world problems. One huge challenge is that they require a lot of interactions with the environment to learn a suitable action policy and have poor data-efficiency. In supervised learning tasks, data augmentation has been shown to improve the data-efficiency of deep learning models. This paper investigates the effects of data augmentation and other semi-supervised methods on a state-of-the-art RL algorithm, namely the Contrastive Unsupervised Representation for RL (CURL) algorithm. We demonstrate that data augmentation can improve the data-efficiency of the CURL algorithm when applied to the DeepMind Control Suite benchmark running on OpenAI Gym. Furthermore, we investigate possible applications of the Mean Teacher semi-supervised learning method to CURL
author2 Tan Yap Peng
author_facet Tan Yap Peng
Cao, Jeffery Siming
format Final Year Project
author Cao, Jeffery Siming
author_sort Cao, Jeffery Siming
title Making reinforcement learning more data-efficient
title_short Making reinforcement learning more data-efficient
title_full Making reinforcement learning more data-efficient
title_fullStr Making reinforcement learning more data-efficient
title_full_unstemmed Making reinforcement learning more data-efficient
title_sort making reinforcement learning more data-efficient
publisher Nanyang Technological University
publishDate 2021
url https://hdl.handle.net/10356/149413
_version_ 1772825165856505856