菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-11
📄 Abstract - UMEM: Unified Memory Extraction and Management Framework for Generalizable Memory

Self-evolving memory serves as the trainable parameters for Large Language Models (LLMs)-based agents, where extraction (distilling insights from experience) and management (updating the memory bank) must be tightly coordinated. Existing methods predominately optimize memory management while treating memory extraction as a static process, resulting in poor generalization, where agents accumulate instance-specific noise rather than robust memories. To address this, we propose Unified Memory Extraction and Management (UMEM), a self-evolving agent framework that jointly optimizes a Large Language Model to simultaneous extract and manage memories. To mitigate overfitting to specific instances, we introduce Semantic Neighborhood Modeling and optimize the model with a neighborhood-level marginal utility reward via GRPO. This approach ensures memory generalizability by evaluating memory utility across clusters of semantically related queries. Extensive experiments across five benchmarks demonstrate that UMEM significantly outperforms highly competitive baselines, achieving up to a 10.67% improvement in multi-turn interactive tasks. Futhermore, UMEM maintains a monotonic growth curve during continuous evolution. Codes and models will be publicly released.

顶级标签: llm agents model training
详细标签: memory extraction memory management self-evolving agents generalization reinforcement learning 或 搜索:

UMEM:面向可泛化记忆的统一记忆提取与管理框架 / UMEM: Unified Memory Extraction and Management Framework for Generalizable Memory


1️⃣ 一句话总结

这篇论文提出了一个名为UMEM的统一框架,通过联合优化记忆提取和管理过程,并引入语义邻域建模来防止模型过拟合到具体实例,从而让基于大语言模型的智能体能够积累更具泛化性和鲁棒性的记忆,在多项任务中显著提升了性能。

源自 arXiv: 2602.10652