Knowledge Transfer for Deep Reinforcement Learning with Hierarchical Experience Replay
The process for transferring knowledge of multiple reinforcement learning policies into a single multi-task policy via distillation technique is known as policy distillation. When policy distillation is under a deep reinforcement learning setting, due to the giant parameter size and the huge state s...
Saved in:
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Conference or Workshop Item |
Language: | English |
Published: |
2017
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/83043 http://hdl.handle.net/10220/42453 https://aaai.org/ocs/index.php/AAAI/AAAI17/paper/view/14478 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-83043 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-830432019-12-06T15:10:46Z Knowledge Transfer for Deep Reinforcement Learning with Hierarchical Experience Replay Yin, Haiyan Pan, Sinno Jialin School of Computer Science and Engineering Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17) Deep reinforcement learning Policy distillation The process for transferring knowledge of multiple reinforcement learning policies into a single multi-task policy via distillation technique is known as policy distillation. When policy distillation is under a deep reinforcement learning setting, due to the giant parameter size and the huge state space for each task domain, it requires extensive computational efforts to train the multi-task policy network. In this paper, we propose a new policy distillation architecture for deep reinforcement learning, where we assume that each task uses its task specific high-level convolutional features as the inputs to the multi-task policy network. Furthermore, we propose a new sampling framework termed hierarchical prioritized experience replay to selectively choose experiences from the replay memories of each task domain to perform learning on the network. With the above two attempts, we aim to accelerate the learning of the multi-task policy network while guaranteeing a good performance. We use Atari 2600 games as testing environment to demonstrate the efficiency and effectiveness of our proposed solution for policy distillation. Accepted version 2017-05-18T07:42:58Z 2019-12-06T15:10:46Z 2017-05-18T07:42:58Z 2019-12-06T15:10:46Z 2017-01-01 2017 Conference Paper Yin, H., & Pan, S. J. (2017). Knowledge Transfer for Deep Reinforcement Learning with Hierarchical Experience Replay. Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17), 1640-1646. https://hdl.handle.net/10356/83043 http://hdl.handle.net/10220/42453 https://aaai.org/ocs/index.php/AAAI/AAAI17/paper/view/14478 199546 en © 2017 Association for the Advancement of Artificial Intelligence. This is the author created version of a work that has been peer reviewed and accepted for publication by Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17), Association for the Advancement of Artificial Intelligence. It incorporates referee’s comments but changes resulting from the publishing process, such as copyediting, structural formatting, may not be reflected in this document. The published version is available at: [https://aaai.org/ocs/index.php/AAAI/AAAI17/paper/view/14478]. 7 p. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
country |
Singapore |
collection |
DR-NTU |
language |
English |
topic |
Deep reinforcement learning Policy distillation |
spellingShingle |
Deep reinforcement learning Policy distillation Yin, Haiyan Pan, Sinno Jialin Knowledge Transfer for Deep Reinforcement Learning with Hierarchical Experience Replay |
description |
The process for transferring knowledge of multiple reinforcement learning policies into a single multi-task policy via distillation technique is known as policy distillation. When policy distillation is under a deep reinforcement learning setting, due to the giant parameter size and the huge state space for each task domain, it requires extensive computational efforts to train the multi-task policy network. In this paper, we propose a new policy distillation architecture for deep reinforcement learning, where we assume that each task uses its task specific high-level convolutional features as the inputs to the multi-task policy network. Furthermore, we propose a new sampling framework termed hierarchical prioritized experience replay to selectively choose experiences from the replay memories of each task domain to perform learning on the network. With the above two attempts, we aim to accelerate the learning of the multi-task policy network while guaranteeing a good performance. We use Atari 2600 games as testing environment to demonstrate the efficiency and effectiveness of our proposed solution for policy distillation. |
author2 |
School of Computer Science and Engineering |
author_facet |
School of Computer Science and Engineering Yin, Haiyan Pan, Sinno Jialin |
format |
Conference or Workshop Item |
author |
Yin, Haiyan Pan, Sinno Jialin |
author_sort |
Yin, Haiyan |
title |
Knowledge Transfer for Deep Reinforcement Learning with Hierarchical Experience Replay |
title_short |
Knowledge Transfer for Deep Reinforcement Learning with Hierarchical Experience Replay |
title_full |
Knowledge Transfer for Deep Reinforcement Learning with Hierarchical Experience Replay |
title_fullStr |
Knowledge Transfer for Deep Reinforcement Learning with Hierarchical Experience Replay |
title_full_unstemmed |
Knowledge Transfer for Deep Reinforcement Learning with Hierarchical Experience Replay |
title_sort |
knowledge transfer for deep reinforcement learning with hierarchical experience replay |
publishDate |
2017 |
url |
https://hdl.handle.net/10356/83043 http://hdl.handle.net/10220/42453 https://aaai.org/ocs/index.php/AAAI/AAAI17/paper/view/14478 |
_version_ |
1681042700578586624 |