Learning to Share: Selective Memory for Efficient Parallel Agentic Systems
AI 摘要
提出了Learning to Share (LTS),一种用于并行Agentic系统的学习型共享内存机制,提升效率。
主要贡献
- 提出了LTS:一个学习型共享内存机制。
- 设计了轻量级的内存控制器,决定信息是否加入内存。
- 使用步进式强化学习训练控制器,识别全局有用信息。
方法论
引入全局内存库,通过强化学习训练控制器,选择性地将中间步骤添加到内存中,实现跨团队的信息复用。
原文摘要
Agentic systems solve complex tasks by coordinating multiple agents that iteratively reason, invoke tools, and exchange intermediate results. To improve robustness and solution quality, recent approaches deploy multiple agent teams running in parallel to explore diverse reasoning trajectories. However, parallel execution comes at a significant computational cost: when different teams independently reason about similar sub-problems or execute analogous steps, they repeatedly perform substantial overlapping computation. To address these limitations, in this paper, we propose Learning to Share (LTS), a learned shared-memory mechanism for parallel agentic frameworks that enables selective cross-team information reuse while controlling context growth. LTS introduces a global memory bank accessible to all teams and a lightweight controller that decides whether intermediate agent steps should be added to memory or not. The controller is trained using stepwise reinforcement learning with usage-aware credit assignment, allowing it to identify information that is globally useful across parallel executions. Experiments on the AssistantBench and GAIA benchmarks show that LTS significantly reduces overall runtime while matching or improving task performance compared to memory-free parallel baselines, demonstrating that learned memory admission is an effective strategy for improving the efficiency of parallel agentic systems. Project page: https://joefioresi718.github.io/LTS_webpage/