Papers
arxiv:2501.13200

SRMT: Shared Memory for Multi-agent Lifelong Pathfinding

Published on Jan 22
· Submitted by alsu-sagirova on Jan 24
#1 Paper of the day
Authors:
,

Abstract

Multi-agent reinforcement learning (MARL) demonstrates significant progress in solving cooperative and competitive multi-agent problems in various environments. One of the principal challenges in MARL is the need for explicit prediction of the agents' behavior to achieve cooperation. To resolve this issue, we propose the Shared Recurrent Memory Transformer (SRMT) which extends memory transformers to multi-agent settings by pooling and globally broadcasting individual working memories, enabling agents to exchange information implicitly and coordinate their actions. We evaluate SRMT on the Partially Observable Multi-Agent Pathfinding problem in a toy Bottleneck navigation task that requires agents to pass through a narrow corridor and on a POGEMA benchmark set of tasks. In the Bottleneck task, SRMT consistently outperforms a variety of reinforcement learning baselines, especially under sparse rewards, and generalizes effectively to longer corridors than those seen during training. On POGEMA maps, including Mazes, Random, and MovingAI, SRMT is competitive with recent MARL, hybrid, and planning-based algorithms. These results suggest that incorporating shared recurrent memory into the transformer-based architectures can enhance coordination in decentralized multi-agent systems. The source code for training and evaluation is available on GitHub: https://github.com/Aloriosa/srmt.

Community

Paper author Paper submitter

We augmented the multi-agent reinforcement learning (MARL) policy network with a shared memory workspace where agents can broadcast their experiences to cooperate and coordinate their behavior while maintaining decentralization during training and execution.

SRMT2 (2).jpg

Paper author Paper submitter

We tested our approach on multi-agent path finding tasks of various levels of difficulty and demonstrated that SRMT consistently outperforms MARL communication baselines and SRMT is competitive with recent MARL, hybrid, and planning-based algorithms.

The source code is available on GitHub:
https://github.com/Aloriosa/srmt

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2501.13200 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2501.13200 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2501.13200 in a Space README.md to link it from this page.

Collections including this paper 2