Multi Agent Cooperation
Multi-agent cooperation research focuses on designing algorithms enabling multiple independent agents to collaboratively achieve shared goals, often in complex, partially observable environments. Current research emphasizes developing robust and efficient algorithms, including those based on large language models (LLMs), reinforcement learning (RL), and novel architectures like normative modules, to improve coordination and communication among agents. These advancements have implications for diverse fields, such as robotics, autonomous systems, and human-computer interaction, by enabling more effective collaboration in both simulated and real-world scenarios. The development of task-agnostic communication strategies and methods for aligning individual and collective objectives are particularly active areas of investigation.
Papers
Building Cooperative Embodied Agents Modularly with Large Language Models
Hongxin Zhang, Weihua Du, Jiaming Shan, Qinhong Zhou, Yilun Du, Joshua B. Tenenbaum, Tianmin Shu, Chuang Gan
Multi-Agent Cooperation via Unsupervised Learning of Joint Intentions
Shanqi Liu, Weiwei Liu, Wenzhou Chen, Guanzhong Tian, Yong Liu