Constrained multiagent reinforcement learning for large agent population
Learning control policies for a large number of agents in a decentralized setting is challenging due to partial observability, uncertainty in the environment, and scalability challenges. While several scalable multiagent RL (MARL) methods have been proposed, relatively few approaches exist for large...
Saved in:
Main Authors: | , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2023
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/8091 https://ink.library.smu.edu.sg/context/sis_research/article/9094/viewcontent/978_3_031_26412_2_12_pv.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-9094 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-90942023-09-07T07:26:33Z Constrained multiagent reinforcement learning for large agent population LING, Jiajing SINGH, Arambam James NGUYEN, Duc Thien KUMAR, Akshat Learning control policies for a large number of agents in a decentralized setting is challenging due to partial observability, uncertainty in the environment, and scalability challenges. While several scalable multiagent RL (MARL) methods have been proposed, relatively few approaches exist for large scale constrained MARL settings. To address this, we first formulate the constrained MARL problem in a collective multiagent setting where interactions among agents are governed by the aggregate count and types of agents, and do not depend on agents’ specific identities. Second, we show that standard Lagrangian relaxation methods, which are popular for single agent RL, do not perform well in constrained MARL settings due to the problem of credit assignment—how to identify and modify behavior of agents that contribute most to constraint violations (and also optimize primary objective alongside)? We develop a fictitious MARL method that addresses this key challenge. Finally, we evaluate our approach on two large-scale real-world applications: maritime traffic management and vehicular network routing. Empirical results show that our approach is highly scalable, can optimize the cumulative global reward and effectively minimize constraint violations, while also being significantly more sample efficient than previous best methods. 2023-09-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8091 info:doi/10.1007/978-3-031-26412-2_12 https://ink.library.smu.edu.sg/context/sis_research/article/9094/viewcontent/978_3_031_26412_2_12_pv.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Constraint optimization Multi-agent systems Multiagent reinforcement learning Artificial Intelligence and Robotics Operations Research, Systems Engineering and Industrial Engineering |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Constraint optimization Multi-agent systems Multiagent reinforcement learning Artificial Intelligence and Robotics Operations Research, Systems Engineering and Industrial Engineering |
spellingShingle |
Constraint optimization Multi-agent systems Multiagent reinforcement learning Artificial Intelligence and Robotics Operations Research, Systems Engineering and Industrial Engineering LING, Jiajing SINGH, Arambam James NGUYEN, Duc Thien KUMAR, Akshat Constrained multiagent reinforcement learning for large agent population |
description |
Learning control policies for a large number of agents in a decentralized setting is challenging due to partial observability, uncertainty in the environment, and scalability challenges. While several scalable multiagent RL (MARL) methods have been proposed, relatively few approaches exist for large scale constrained MARL settings. To address this, we first formulate the constrained MARL problem in a collective multiagent setting where interactions among agents are governed by the aggregate count and types of agents, and do not depend on agents’ specific identities. Second, we show that standard Lagrangian relaxation methods, which are popular for single agent RL, do not perform well in constrained MARL settings due to the problem of credit assignment—how to identify and modify behavior of agents that contribute most to constraint violations (and also optimize primary objective alongside)? We develop a fictitious MARL method that addresses this key challenge. Finally, we evaluate our approach on two large-scale real-world applications: maritime traffic management and vehicular network routing. Empirical results show that our approach is highly scalable, can optimize the cumulative global reward and effectively minimize constraint violations, while also being significantly more sample efficient than previous best methods. |
format |
text |
author |
LING, Jiajing SINGH, Arambam James NGUYEN, Duc Thien KUMAR, Akshat |
author_facet |
LING, Jiajing SINGH, Arambam James NGUYEN, Duc Thien KUMAR, Akshat |
author_sort |
LING, Jiajing |
title |
Constrained multiagent reinforcement learning for large agent population |
title_short |
Constrained multiagent reinforcement learning for large agent population |
title_full |
Constrained multiagent reinforcement learning for large agent population |
title_fullStr |
Constrained multiagent reinforcement learning for large agent population |
title_full_unstemmed |
Constrained multiagent reinforcement learning for large agent population |
title_sort |
constrained multiagent reinforcement learning for large agent population |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2023 |
url |
https://ink.library.smu.edu.sg/sis_research/8091 https://ink.library.smu.edu.sg/context/sis_research/article/9094/viewcontent/978_3_031_26412_2_12_pv.pdf |
_version_ |
1779157152103923712 |