Constrained multiagent reinforcement learning for large agent population

Learning control policies for a large number of agents in a decentralized setting is challenging due to partial observability, uncertainty in the environment, and scalability challenges. While several scalable multiagent RL (MARL) methods have been proposed, relatively few approaches exist for large...

Full description

Saved in:
Bibliographic Details
Main Authors: LING, Jiajing, SINGH, Arambam James, NGUYEN, Duc Thien, KUMAR, Akshat
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2023
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8091
https://ink.library.smu.edu.sg/context/sis_research/article/9094/viewcontent/978_3_031_26412_2_12_pv.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9094
record_format dspace
spelling sg-smu-ink.sis_research-90942023-09-07T07:26:33Z Constrained multiagent reinforcement learning for large agent population LING, Jiajing SINGH, Arambam James NGUYEN, Duc Thien KUMAR, Akshat Learning control policies for a large number of agents in a decentralized setting is challenging due to partial observability, uncertainty in the environment, and scalability challenges. While several scalable multiagent RL (MARL) methods have been proposed, relatively few approaches exist for large scale constrained MARL settings. To address this, we first formulate the constrained MARL problem in a collective multiagent setting where interactions among agents are governed by the aggregate count and types of agents, and do not depend on agents’ specific identities. Second, we show that standard Lagrangian relaxation methods, which are popular for single agent RL, do not perform well in constrained MARL settings due to the problem of credit assignment—how to identify and modify behavior of agents that contribute most to constraint violations (and also optimize primary objective alongside)? We develop a fictitious MARL method that addresses this key challenge. Finally, we evaluate our approach on two large-scale real-world applications: maritime traffic management and vehicular network routing. Empirical results show that our approach is highly scalable, can optimize the cumulative global reward and effectively minimize constraint violations, while also being significantly more sample efficient than previous best methods. 2023-09-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8091 info:doi/10.1007/978-3-031-26412-2_12 https://ink.library.smu.edu.sg/context/sis_research/article/9094/viewcontent/978_3_031_26412_2_12_pv.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Constraint optimization Multi-agent systems Multiagent reinforcement learning Artificial Intelligence and Robotics Operations Research, Systems Engineering and Industrial Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Constraint optimization
Multi-agent systems
Multiagent reinforcement learning
Artificial Intelligence and Robotics
Operations Research, Systems Engineering and Industrial Engineering
spellingShingle Constraint optimization
Multi-agent systems
Multiagent reinforcement learning
Artificial Intelligence and Robotics
Operations Research, Systems Engineering and Industrial Engineering
LING, Jiajing
SINGH, Arambam James
NGUYEN, Duc Thien
KUMAR, Akshat
Constrained multiagent reinforcement learning for large agent population
description Learning control policies for a large number of agents in a decentralized setting is challenging due to partial observability, uncertainty in the environment, and scalability challenges. While several scalable multiagent RL (MARL) methods have been proposed, relatively few approaches exist for large scale constrained MARL settings. To address this, we first formulate the constrained MARL problem in a collective multiagent setting where interactions among agents are governed by the aggregate count and types of agents, and do not depend on agents’ specific identities. Second, we show that standard Lagrangian relaxation methods, which are popular for single agent RL, do not perform well in constrained MARL settings due to the problem of credit assignment—how to identify and modify behavior of agents that contribute most to constraint violations (and also optimize primary objective alongside)? We develop a fictitious MARL method that addresses this key challenge. Finally, we evaluate our approach on two large-scale real-world applications: maritime traffic management and vehicular network routing. Empirical results show that our approach is highly scalable, can optimize the cumulative global reward and effectively minimize constraint violations, while also being significantly more sample efficient than previous best methods.
format text
author LING, Jiajing
SINGH, Arambam James
NGUYEN, Duc Thien
KUMAR, Akshat
author_facet LING, Jiajing
SINGH, Arambam James
NGUYEN, Duc Thien
KUMAR, Akshat
author_sort LING, Jiajing
title Constrained multiagent reinforcement learning for large agent population
title_short Constrained multiagent reinforcement learning for large agent population
title_full Constrained multiagent reinforcement learning for large agent population
title_fullStr Constrained multiagent reinforcement learning for large agent population
title_full_unstemmed Constrained multiagent reinforcement learning for large agent population
title_sort constrained multiagent reinforcement learning for large agent population
publisher Institutional Knowledge at Singapore Management University
publishDate 2023
url https://ink.library.smu.edu.sg/sis_research/8091
https://ink.library.smu.edu.sg/context/sis_research/article/9094/viewcontent/978_3_031_26412_2_12_pv.pdf
_version_ 1779157152103923712