Handling long and richly constrained tasks through constrained hierarchical reinforcement learning

Safety in goal directed Reinforcement Learning (RL) settings has typically been handled through constraints over trajectories and have demonstrated good performance in primarily short horizon tasks. In this paper, we are specifically interested in the problem of solving temporally extended decision...

Full description

Saved in:
Bibliographic Details
Main Authors: LU, Yuxiao, SINHA, Arunesh, VARAKANTHAM, Pradeep
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2024
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8595
https://ink.library.smu.edu.sg/context/sis_research/article/9598/viewcontent/handling_long.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9598
record_format dspace
spelling sg-smu-ink.sis_research-95982024-01-25T08:38:10Z Handling long and richly constrained tasks through constrained hierarchical reinforcement learning LU, Yuxiao SINHA, Arunesh VARAKANTHAM, Pradeep Safety in goal directed Reinforcement Learning (RL) settings has typically been handled through constraints over trajectories and have demonstrated good performance in primarily short horizon tasks. In this paper, we are specifically interested in the problem of solving temporally extended decision making problems such as robots cleaning different areas in a house while avoiding slippery and unsafe areas (e.g., stairs) and retaining enough charge to move to a charging dock; in the presence of complex safety constraints. Our key contribution is a (safety) Constrained Search with Hierarchical Reinforcement Learning (CoSHRL) mechanism that combines an upper level constrained search agent (which computes a reward maximizing policy from a given start to a far away goal state while satisfying cost constraints) with a low-level goal conditioned RL agent (which estimates cost and reward values to move between nearby states). A major advantage of CoSHRL is that it can handle constraints on the cost value distribution (e.g., on Conditional Value at Risk, CVaR) and can adjust to flexible constraint thresholds without retraining. We perform extensive experiments with different types of safety constraints to demonstrate the utility of our approach over leading approaches in constrained and hierarchical RL. 2024-02-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8595 https://ink.library.smu.edu.sg/context/sis_research/article/9598/viewcontent/handling_long.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Artificial Intelligence and Robotics Databases and Information Systems
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Artificial Intelligence and Robotics
Databases and Information Systems
spellingShingle Artificial Intelligence and Robotics
Databases and Information Systems
LU, Yuxiao
SINHA, Arunesh
VARAKANTHAM, Pradeep
Handling long and richly constrained tasks through constrained hierarchical reinforcement learning
description Safety in goal directed Reinforcement Learning (RL) settings has typically been handled through constraints over trajectories and have demonstrated good performance in primarily short horizon tasks. In this paper, we are specifically interested in the problem of solving temporally extended decision making problems such as robots cleaning different areas in a house while avoiding slippery and unsafe areas (e.g., stairs) and retaining enough charge to move to a charging dock; in the presence of complex safety constraints. Our key contribution is a (safety) Constrained Search with Hierarchical Reinforcement Learning (CoSHRL) mechanism that combines an upper level constrained search agent (which computes a reward maximizing policy from a given start to a far away goal state while satisfying cost constraints) with a low-level goal conditioned RL agent (which estimates cost and reward values to move between nearby states). A major advantage of CoSHRL is that it can handle constraints on the cost value distribution (e.g., on Conditional Value at Risk, CVaR) and can adjust to flexible constraint thresholds without retraining. We perform extensive experiments with different types of safety constraints to demonstrate the utility of our approach over leading approaches in constrained and hierarchical RL.
format text
author LU, Yuxiao
SINHA, Arunesh
VARAKANTHAM, Pradeep
author_facet LU, Yuxiao
SINHA, Arunesh
VARAKANTHAM, Pradeep
author_sort LU, Yuxiao
title Handling long and richly constrained tasks through constrained hierarchical reinforcement learning
title_short Handling long and richly constrained tasks through constrained hierarchical reinforcement learning
title_full Handling long and richly constrained tasks through constrained hierarchical reinforcement learning
title_fullStr Handling long and richly constrained tasks through constrained hierarchical reinforcement learning
title_full_unstemmed Handling long and richly constrained tasks through constrained hierarchical reinforcement learning
title_sort handling long and richly constrained tasks through constrained hierarchical reinforcement learning
publisher Institutional Knowledge at Singapore Management University
publishDate 2024
url https://ink.library.smu.edu.sg/sis_research/8595
https://ink.library.smu.edu.sg/context/sis_research/article/9598/viewcontent/handling_long.pdf
_version_ 1789483283382272000