Prior Coordination
Prior coordination, the ability of multiple agents to achieve a shared goal without pre-arranged plans, is a crucial area of research spanning diverse fields like economics, robotics, and online social dynamics. Current research focuses on developing algorithms and models, including reinforcement learning (particularly multi-agent variants like QMIX and DDPG), graph neural networks, and game-theoretic approaches, to enable effective coordination in complex, often partially observable environments. These advancements have significant implications for improving traffic management, optimizing multi-robot systems, enhancing online community interactions, and creating more robust and efficient decentralized systems in general.
Papers
Stateful active facilitator: Coordination and Environmental Heterogeneity in Cooperative Multi-Agent Reinforcement Learning
Dianbo Liu, Vedant Shah, Oussama Boussif, Cristian Meo, Anirudh Goyal, Tianmin Shu, Michael Mozer, Nicolas Heess, Yoshua Bengio
Meta Navigation Functions: Adaptive Associations for Coordination of Multi-Agent Systems
Matin Macktoobian, Guillaume Ferdinand Duc