Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Multi-Agent Systems in the Era of Foundation Models: Opportunities, Challenges and Futures

MARL-GPT: Foundation Model for Multi-Agent Reinforcement Learning

Maria Nesterova · Mikhail Kolosov · Anton Andreychuk · Egor Cherepanov · Alexey Kovalev · Konstantin Yakovlev · Aleksandr Panov · Alexey Skrynnik


Abstract:

Recent advances in multi-agent reinforcement learning (MARL) have demonstrated success in numerous challenging domains and environments, but typically require specialized models for each task. In this work, we propose a coherent methodology that makes it possible for a single GPT-based model to learn and perform well across diverse MARL environments and tasks, including StarCraft Multi-Agent Challenge, Google Research Football and POGEMA. Our approach, which we dub MARL-GPT, combines data-driven imitation learning with Q-value learning that leverages learning (at scale) on the expert trajectories (400M for SMAC, 10M for GRF, 1B for POGEMA).Experiments show that MARL-GPT achieves competitive performance compared to specialized baselines in all tested environments. Thus, our findings suggest that it is, indeed, possible to build a multi-task transformer-based model for a wide variety of (significantly different) multi-agent problems paving the way to the fundamental MARL model (akin to ChatGPT, Llama, Mistral etc. in natural language modeling).

Chat is not available.