AMA-Bench:评估智能体应用中的长程记忆能力 / AMA-Bench: Evaluating Long-Horizon Memory for Agentic Applications
1️⃣ 一句话总结
这篇论文提出了一个名为AMA-Bench的新基准测试,专门用于评估大语言模型在真实、长期运行的智能体应用中的记忆能力,并针对现有记忆系统的不足,设计了一个包含因果图和工具增强检索的新型记忆系统AMA-Agent,显著提升了性能。
Large Language Models (LLMs) are deployed as autonomous agents in increasingly complex applications, where enabling long-horizon memory is critical for achieving strong performance. However, a significant gap exists between practical applications and current evaluation standards for agent memory: existing benchmarks primarily focus on dialogue-centric, human-agent interactions. In reality, agent memory consists of a continuous stream of agent-environment interactions that are primarily composed of machine-generated representations. To bridge this gap, we introduce AMA-Bench (Agent Memory with Any length), which evaluates long-horizon memory for LLMs in real agentic applications. It features two key components: (1) a set of real-world agentic trajectories across representative agentic applications, paired with expert-curated QA, and (2) a set of synthetic agentic trajectories that scale to arbitrary horizons, paired with rule-based QA. Our comprehensive study shows that existing memory systems underperform on AMA-Bench primarily because they lack causality and objective information and are constrained by the lossy nature of similarity-based retrieval employed by many memory systems. To address these limitations, we propose AMA-Agent, an effective memory system featuring a causality graph and tool-augmented retrieval. Our results demonstrate that AMA-Agent achieves 57.22% average accuracy on AMA-Bench, surpassing the strongest memory system baselines by 11.16%.
AMA-Bench:评估智能体应用中的长程记忆能力 / AMA-Bench: Evaluating Long-Horizon Memory for Agentic Applications
这篇论文提出了一个名为AMA-Bench的新基准测试,专门用于评估大语言模型在真实、长期运行的智能体应用中的记忆能力,并针对现有记忆系统的不足,设计了一个包含因果图和工具增强检索的新型记忆系统AMA-Agent,显著提升了性能。
源自 arXiv: 2602.22769