Value-based subgoal discovery and path planning for reaching long-horizon goals

Learning to reach long-horizon goals in spatial traversal tasks is a significant challenge for autonomous agents. Recent subgoal graph-based planning methods address this challenge by decomposing a goal into a sequence of shorter-horizon subgoals. These methods, however, use arbitrary heuristics for...

Full description

Saved in:
Bibliographic Details
Main Authors: PATERIA, Shubham, SUBAGDJA, Budhitama, TAN, Ah-hwee, QUEK, Chai
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2023
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8114
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9117
record_format dspace
spelling sg-smu-ink.sis_research-91172023-09-06T10:06:03Z Value-based subgoal discovery and path planning for reaching long-horizon goals PATERIA, Shubham SUBAGDJA, Budhitama TAN, Ah-hwee QUEK, Chai Learning to reach long-horizon goals in spatial traversal tasks is a significant challenge for autonomous agents. Recent subgoal graph-based planning methods address this challenge by decomposing a goal into a sequence of shorter-horizon subgoals. These methods, however, use arbitrary heuristics for sampling or discovering subgoals, which may not conform to the cumulative reward distribution. Moreover, they are prone to learning erroneous connections (edges) between subgoals, especially those lying across obstacles. To address these issues, this article proposes a novel subgoal graph-based planning method called learning subgoal graph using value-based subgoal discovery and automatic pruning (LSGVP). The proposed method uses a subgoal discovery heuristic that is based on a cumulative reward (value) measure and yields sparse subgoals, including those lying on the higher cumulative reward paths. Moreover, LSGVP guides the agent to automatically prune the learned subgoal graph to remove the erroneous edges. The combination of these novel features helps the LSGVP agent to achieve higher cumulative positive rewards than other subgoal sampling or discovery heuristics, as well as higher goal-reaching success rates than other state-of-the-art subgoal graph-based planning methods. 2023-02-07T08:00:00Z text https://ink.library.smu.edu.sg/sis_research/8114 info:doi/10.1109/TNNLS.2023.3240004 Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Long-horizon goal-reaching motion planning path planning reinforcement learning (RL) subgoal discovery subgoal graph Databases and Information Systems OS and Networks
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Long-horizon goal-reaching
motion planning
path planning
reinforcement learning (RL)
subgoal discovery
subgoal graph
Databases and Information Systems
OS and Networks
spellingShingle Long-horizon goal-reaching
motion planning
path planning
reinforcement learning (RL)
subgoal discovery
subgoal graph
Databases and Information Systems
OS and Networks
PATERIA, Shubham
SUBAGDJA, Budhitama
TAN, Ah-hwee
QUEK, Chai
Value-based subgoal discovery and path planning for reaching long-horizon goals
description Learning to reach long-horizon goals in spatial traversal tasks is a significant challenge for autonomous agents. Recent subgoal graph-based planning methods address this challenge by decomposing a goal into a sequence of shorter-horizon subgoals. These methods, however, use arbitrary heuristics for sampling or discovering subgoals, which may not conform to the cumulative reward distribution. Moreover, they are prone to learning erroneous connections (edges) between subgoals, especially those lying across obstacles. To address these issues, this article proposes a novel subgoal graph-based planning method called learning subgoal graph using value-based subgoal discovery and automatic pruning (LSGVP). The proposed method uses a subgoal discovery heuristic that is based on a cumulative reward (value) measure and yields sparse subgoals, including those lying on the higher cumulative reward paths. Moreover, LSGVP guides the agent to automatically prune the learned subgoal graph to remove the erroneous edges. The combination of these novel features helps the LSGVP agent to achieve higher cumulative positive rewards than other subgoal sampling or discovery heuristics, as well as higher goal-reaching success rates than other state-of-the-art subgoal graph-based planning methods.
format text
author PATERIA, Shubham
SUBAGDJA, Budhitama
TAN, Ah-hwee
QUEK, Chai
author_facet PATERIA, Shubham
SUBAGDJA, Budhitama
TAN, Ah-hwee
QUEK, Chai
author_sort PATERIA, Shubham
title Value-based subgoal discovery and path planning for reaching long-horizon goals
title_short Value-based subgoal discovery and path planning for reaching long-horizon goals
title_full Value-based subgoal discovery and path planning for reaching long-horizon goals
title_fullStr Value-based subgoal discovery and path planning for reaching long-horizon goals
title_full_unstemmed Value-based subgoal discovery and path planning for reaching long-horizon goals
title_sort value-based subgoal discovery and path planning for reaching long-horizon goals
publisher Institutional Knowledge at Singapore Management University
publishDate 2023
url https://ink.library.smu.edu.sg/sis_research/8114
_version_ 1779157159485898752