The field of multi-agent reinforcement learning (MARL) is rapidly advancing, with a focus on developing more efficient and effective algorithms for cooperative and competitive environments. Recent research has explored the use of centralized training with decentralized execution, permutation equivariant architectures, and autoregressive action world models to improve performance in complex multi-agent settings. Additionally, there is a growing interest in developing more realistic and challenging benchmarks, such as those that incorporate continuous action spaces and uncertain environments. Noteworthy papers in this area include the proposal of Centralized Permutation Equivariant (CPE) learning, which has been shown to substantially improve performance in cooperative benchmarks, and the introduction of MAPF-World, an autoregressive action world model that enables more informed and coordinated decision-making in multi-agent path finding scenarios. Other notable contributions include the development of novel communication mechanisms, such as self-attention-based modules, and the investigation of payoff manipulation in repeated multi-objective Stackelberg games.
Advances in Multi-Agent Reinforcement Learning
Sources
Nominal Evaluation Of Automatic Multi-Sections Control Potential In Comparison To A Simpler One- Or Two-Sections Alternative With Predictive Spray Switching
DCT-MARL: A Dynamic Communication Topology-Based MARL Algorithm for Connected Vehicle Platoon Control
MACTAS: Self-Attention-Based Module for Inter-Agent Communication in Multi-Agent Reinforcement Learning