Agent Tuning & Optimization 相关度: 9/10

UMEM: Unified Memory Extraction and Management Framework for Generalizable Memory

Yongshi Ye, Hui Jiang, Feihu Jiang, Tian Lan, Yichao Du, Biao Fu, Xiaodong Shi, Qianghuai Jia, Longyue Wang, Weihua Luo
arXiv: 2602.10652v1 发布: 2026-02-11 更新: 2026-02-11

AI 摘要

UMEM联合优化LLM的记忆提取与管理,通过语义邻域建模提高记忆泛化能力,在交互任务中表现出色。

主要贡献

  • 提出UMEM框架,联合优化记忆提取和管理
  • 引入语义邻域建模和GRPO优化,提升记忆泛化性
  • 实验证明UMEM在多轮交互任务中显著优于基线

方法论

使用大型语言模型,通过语义邻域建模和GRPO优化,联合训练记忆提取和管理模块,提升记忆的泛化能力。

原文摘要

Self-evolving memory serves as the trainable parameters for Large Language Models (LLMs)-based agents, where extraction (distilling insights from experience) and management (updating the memory bank) must be tightly coordinated. Existing methods predominately optimize memory management while treating memory extraction as a static process, resulting in poor generalization, where agents accumulate instance-specific noise rather than robust memories. To address this, we propose Unified Memory Extraction and Management (UMEM), a self-evolving agent framework that jointly optimizes a Large Language Model to simultaneous extract and manage memories. To mitigate overfitting to specific instances, we introduce Semantic Neighborhood Modeling and optimize the model with a neighborhood-level marginal utility reward via GRPO. This approach ensures memory generalizability by evaluating memory utility across clusters of semantically related queries. Extensive experiments across five benchmarks demonstrate that UMEM significantly outperforms highly competitive baselines, achieving up to a 10.67% improvement in multi-turn interactive tasks. Futhermore, UMEM maintains a monotonic growth curve during continuous evolution. Codes and models will be publicly released.

标签

LLM Memory Agent Generalization

arXiv 分类

cs.CL