Scaling Teams or Scaling Time? Memory Enabled Lifelong Learning in LLM Multi-Agent Systems
Abstract
LLM multi-agent systems exhibit non-monotonic scaling behavior where memory design significantly impacts long-term performance, with smaller teams sometimes outperforming larger ones when experience reuse is optimized.
Large language model (LLM) multi-agent systems can scale along two distinct dimensions: by increasing the number of agents and by improving through accumulated experience over time. Although prior work has studied these dimensions separately, their interaction under realistic cost constraints remains unclear. In this paper, we introduce a conceptual scaling view of multi-agent systems that jointly considers team size and lifelong learning ability, and we study how memory design shares this landscape. To this end, we propose LLMA-Mem, a lifelong memory framework for LLM multi-agent systems under flexible memory topologies. We evaluate LLMA-Mem on MultiAgentBench across coding, research, and database environments. Empirically, LLMA-Mem consistently improves long-horizon performance over baselines while reducing cost. Our analysis further reveals a non-monotonic scaling landscape: larger teams do not always produce better long-term performance, and smaller teams can outperform larger ones when memory better supports the reuse of experience. These findings position memory design as a practical path for scaling multi-agent systems more effectively and more efficiently over time.
Community
Excited to share our work on the scaling of multi-agent systems! We move beyond just "adding more agents" to jointly study accumulated experience as a second dimension.
The Framework: We introduce LLMA-Mem, a lifelong memory system with flexible topologies (individual vs. shared) for MAS.
Key Insight: We find a non-monotonic scaling landscape—smaller teams with superior memory design often outperform larger, "memory-poor" teams while significantly reducing inference costs.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Structurally Aligned Subtask-Level Memory for Software Engineering Agents (2026)
- Experience as a Compass: Multi-agent RAG with Evolving Orchestration and Agent Prompts (2026)
- AdaMem: Adaptive User-Centric Memory for Long-Horizon Dialogue Agents (2026)
- Collaborative Multi-Agent Optimization for Personalized Memory System (2026)
- Choosing How to Remember: Adaptive Memory Structures for LLM Agents (2026)
- Agent Q-Mix: Selecting the Right Action for LLM Multi-Agent Systems through Reinforcement Learning (2026)
- Memex(RL): Scaling Long-Horizon LLM Agents via Indexed Experience Memory (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Get this paper in your agent:
hf papers read 2604.03295 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper