Towards explaining sequences of actions in multi-agent deep reinforcement learning models

Although Multi-agent Deep Reinforcement Learning (MADRL) has shown promising results in solving complex real-world problems, the applicability and reliability of MADRL models are often limited by a lack of understanding of their inner workings for explaining the decisions made. To address this issue...

Full description

Saved in:
Bibliographic Details
Main Authors: KHAING, Phyo Wai, GENG, Minghong, SUBAGDJA, Budhitama, PATERIA, Shubham, TAN, Ah-hwee
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2023
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8076
https://ink.library.smu.edu.sg/context/sis_research/article/9079/viewcontent/p2325.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9079
record_format dspace
spelling sg-smu-ink.sis_research-90792023-09-07T07:55:53Z Towards explaining sequences of actions in multi-agent deep reinforcement learning models KHAING, Phyo Wai GENG, Minghong SUBAGDJA, Budhitama PATERIA, Shubham TAN, Ah-hwee Although Multi-agent Deep Reinforcement Learning (MADRL) has shown promising results in solving complex real-world problems, the applicability and reliability of MADRL models are often limited by a lack of understanding of their inner workings for explaining the decisions made. To address this issue, this paper proposes a novel method for explaining MADRL by generalizing the sequences of action events performed by agents into high-level abstract strategies using a spatio-temporal neural network model. Specifically, an interval-based memory retrieval procedure is developed to generalize the encoded sequences of action events over time into short sequential patterns. In addition, two abstraction algorithms are introduced, one for abstracting action events across multiple agents and the other for further abstracting the episodes over time into short sequential patterns, which can then be translated into symbolic form for interpretation. We evaluate the proposed method using the StarCraft Multi Agent Challenge (SMAC) benchmark task, which shows that the method is able to derive high-level explanations of MADRL models at various levels of granularity. 2023-06-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8076 info:doi/10.5555/3545946.3598922 https://ink.library.smu.edu.sg/context/sis_research/article/9079/viewcontent/p2325.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Multi Agent Deep Reinforcement Learning Explainable Artificial Intelligence Explainable Deep Reinforcement Learning Databases and Information Systems
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Multi Agent Deep Reinforcement Learning
Explainable Artificial Intelligence
Explainable Deep Reinforcement Learning
Databases and Information Systems
spellingShingle Multi Agent Deep Reinforcement Learning
Explainable Artificial Intelligence
Explainable Deep Reinforcement Learning
Databases and Information Systems
KHAING, Phyo Wai
GENG, Minghong
SUBAGDJA, Budhitama
PATERIA, Shubham
TAN, Ah-hwee
Towards explaining sequences of actions in multi-agent deep reinforcement learning models
description Although Multi-agent Deep Reinforcement Learning (MADRL) has shown promising results in solving complex real-world problems, the applicability and reliability of MADRL models are often limited by a lack of understanding of their inner workings for explaining the decisions made. To address this issue, this paper proposes a novel method for explaining MADRL by generalizing the sequences of action events performed by agents into high-level abstract strategies using a spatio-temporal neural network model. Specifically, an interval-based memory retrieval procedure is developed to generalize the encoded sequences of action events over time into short sequential patterns. In addition, two abstraction algorithms are introduced, one for abstracting action events across multiple agents and the other for further abstracting the episodes over time into short sequential patterns, which can then be translated into symbolic form for interpretation. We evaluate the proposed method using the StarCraft Multi Agent Challenge (SMAC) benchmark task, which shows that the method is able to derive high-level explanations of MADRL models at various levels of granularity.
format text
author KHAING, Phyo Wai
GENG, Minghong
SUBAGDJA, Budhitama
PATERIA, Shubham
TAN, Ah-hwee
author_facet KHAING, Phyo Wai
GENG, Minghong
SUBAGDJA, Budhitama
PATERIA, Shubham
TAN, Ah-hwee
author_sort KHAING, Phyo Wai
title Towards explaining sequences of actions in multi-agent deep reinforcement learning models
title_short Towards explaining sequences of actions in multi-agent deep reinforcement learning models
title_full Towards explaining sequences of actions in multi-agent deep reinforcement learning models
title_fullStr Towards explaining sequences of actions in multi-agent deep reinforcement learning models
title_full_unstemmed Towards explaining sequences of actions in multi-agent deep reinforcement learning models
title_sort towards explaining sequences of actions in multi-agent deep reinforcement learning models
publisher Institutional Knowledge at Singapore Management University
publishDate 2023
url https://ink.library.smu.edu.sg/sis_research/8076
https://ink.library.smu.edu.sg/context/sis_research/article/9079/viewcontent/p2325.pdf
_version_ 1779157141148401664