HiSOMA: A hierarchical multi-agent model integrating self-organizing neural networks with multi-agent deep reinforcement learning

被引:4
作者
Geng, Minghong [1 ]
Pateria, Shubham [1 ]
Subagdja, Budhitama [1 ]
Tan, Ah-Hwee [1 ]
机构
[1] Singapore Management Univ, Sch Comp & Informat Syst, 80 Stamford Rd, Singapore 178902, Singapore
关键词
Multi-agent deep reinforcement learning; Hierarchical control; Self-organizing neural networks; LEVEL;
D O I
10.1016/j.eswa.2024.124117
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-agent deep reinforcement learning (MADRL) has shown remarkable advancements in the past decade. However, most current MADRL models focus on task-specific short-horizon problems involving a small number of agents, limiting their applicability to long-horizon planning in complex environments. Hierarchical multiagent models offer a promising solution by organizing agents into different levels, effectively addressing tasks with varying planning horizons. However, these models often face constraints related to the number of agents or levels of hierarchies. This paper introduces HiSOMA, a novel hierarchical multi-agent model designed to handle long-horizon, multi-agent, multi -task decision-making problems. The top-level controller, FALCON, is modeled as a class of self-organizing neural networks (SONN), designed to learn high-level decision rules as internal cognitive codes to modulate middle-level controllers in a fast and incremental manner. The middle-level controllers, MADRL models, in turn receive modulatory signals from the higher level and regulate bottom-level controllers, which learn individual action policies generating primitive actions and interacting directly with the environment. Extensive experiments across different levels of the hierarchical model demonstrate HiSOMA's efficiency in tackling challenging long-horizon problems, surpassing a number of non-hierarchical MADRL approaches. Moreover, its modular design allows for extension into deeper hierarchies and application to more complex tasks with heterogeneous controllers. Demonstration videos and codes can be found on our project web page: https://smu-ncc.github.io.
引用
收藏
页数:11
相关论文
共 51 条
[1]  
Ahilan S, 2019, Arxiv, DOI arXiv:1901.08492
[2]  
[Anonymous], 2004, Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems-Volume
[3]  
Bin Peng X, 2019, Arxiv, DOI arXiv:1905.09808
[4]   Synergies in coordination: a comprehensive overview of neural, computational, and behavioral approaches [J].
Bruton, Michaela ;
O'Dwyer, Nicholas .
JOURNAL OF NEUROPHYSIOLOGY, 2018, 120 (06) :2761-2774
[5]   A comprehensive survey of multiagent reinforcement learning [J].
Busoniu, Lucian ;
Babuska, Robert ;
De Schutter, Bart .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2008, 38 (02) :156-172
[6]   FUZZY ART - FAST STABLE LEARNING AND CATEGORIZATION OF ANALOG PATTERNS BY AN ADAPTIVE RESONANCE SYSTEM [J].
CARPENTER, GA ;
GROSSBERG, S ;
ROSEN, DB .
NEURAL NETWORKS, 1991, 4 (06) :759-771
[7]   ART-2 - SELF-ORGANIZATION OF STABLE CATEGORY RECOGNITION CODES FOR ANALOG INPUT PATTERNS [J].
CARPENTER, GA ;
GROSSBERG, S .
APPLIED OPTICS, 1987, 26 (23) :4919-4930
[8]  
Carroll M, 2019, ADV NEUR IN, V32
[9]  
Christianos F, 2021, PR MACH LEARN RES, V139
[10]   Multi-Agent Deep Reinforcement Learning for Large-Scale Traffic Signal Control [J].
Chu, Tianshu ;
Wang, Jie ;
Codeca, Lara ;
Li, Zhaojian .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2020, 21 (03) :1086-1095