ProAgent: Building Proactive Cooperative Agents with Large Language Models

Ceyao Zhang1,2, Kaijie Yang3, Siyi Hu4, Zihao Wang2,5, Guanghe Li2, Yihang Sun2, Cheng Zhang2, Zhaowei Zhang2,5, Anji Liu2, Song-Chun Zhu5, Xiaojun Chang4, Junge Zhang3, Feng Yin1, Yitao Liang2, Yaodong Yang2,
1CUHK-Shenzhen, 2Insitute of AI, PKU, 3Insitute of Automation, CAS, 4RELER, AAII, UTS, 5BIGAI

Abstract

Building agents with adaptive behavior in cooperative tasks stands as a paramount goal in the realm of multi-agent systems. Current approaches to developing cooperative agents rely primarily on learning-based methods, whose policy generalization depends heavily on the diversity of teammates they interact with during the training phase. Such reliance, however, constrains the agents' capacity for strategic adaptation when cooperating with unfamiliar teammates, which becomes a significant challenge in zero-shot coordination scenarios. To address this challenge, we propose ProAgent, a novel framework that harnesses large language models (LLMs) to create proactive agents capable of dynamically adapting their behavior to enhance cooperation with teammates. ProAgent can analyze the present state, and infer the intentions of teammates from observations. It then updates its beliefs in alignment with the teammates' subsequent actual behaviors. Moreover, ProAgent exhibits a high degree of modularity and interpretability, making it easily integrated into various of coordination scenarios. Experimental evaluations conducted within the Overcooked-AI environment unveil the remarkable performance superiority of ProAgent, outperforming five methods based on self-play and population-based training when cooperating with AI agents. Furthermore, in partnered with human proxy models, its performance exhibits an average improvement exceeding 10% compared to the current state-of-the-art method.






ProAgent cooperates with AI agent in Overcooked-AI.

BibTeX


        @inproceedings{zhang2024proagent,
          title={Pro{A}gent: Building Proactive Cooperative Agents with Large Language Models},
          author={Zhang, Ceyao and Yang, Kaijie and Hu, Siyi and Wang, Zihao and Li, Guanghe and Sun, Yihang and Zhang, Cheng and Zhang, Zhaowei and Liu, Anji and Zhu, Song-Chun and Chang, Xiaojun and Zhang, Junge and Yin, Feng and Liang, Yitao  and Yang, Yaodong},
          booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
          volume={38},
          number={16},
          pages={17591--17599},
          year={2024}
        }