Making reinforcement learning more data-efficient
Recently, researchers showed that applying deep neural networks to Reinforcement Learning (RL) can improve the agent’s ability to learn action policies directly from high dimensional input states. However, there are still several challenges present when applying these algorithms to real world...
Saved in:
主要作者: | |
---|---|
其他作者: | |
格式: | Final Year Project |
語言: | English |
出版: |
Nanyang Technological University
2021
|
主題: | |
在線閱讀: | https://hdl.handle.net/10356/149413 |
標簽: |
添加標簽
沒有標簽, 成為第一個標記此記錄!
|
總結: | Recently, researchers showed that applying deep neural networks to Reinforcement
Learning (RL) can improve the agent’s ability to learn action policies directly from high
dimensional input states. However, there are still several challenges present when
applying these algorithms to real world problems. One huge challenge is that they require
a lot of interactions with the environment to learn a suitable action policy and have poor
data-efficiency.
In supervised learning tasks, data augmentation has been shown to improve the
data-efficiency of deep learning models. This paper investigates the effects of data
augmentation and other semi-supervised methods on a state-of-the-art RL algorithm,
namely the Contrastive Unsupervised Representation for RL (CURL) algorithm. We
demonstrate that data augmentation can improve the data-efficiency of the CURL
algorithm when applied to the DeepMind Control Suite benchmark running on
OpenAI Gym. Furthermore, we investigate possible applications of the Mean Teacher
semi-supervised learning method to CURL |
---|