菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-05-11
📄 Abstract - MAGE: Multi-Agent Self-Evolution with Co-Evolutionary Knowledge Graphs

Self-evolving language-model agents must decide what to learn next and how to preserve what they have learned across iterations. Existing systems typically carry this cross-iteration knowledge as natural-language feedback, flat episodic memory, or implicit reinforcement signals, none of which cleanly supports a frozen weak backbone at inference time. This paper introduces MAGE (Multi-Agent Graph-guided Evolution), a framework that externalizes self-knowledge into a four-subgraph co-evolutionary knowledge graph. Its experience subgraph stores both teacher-written failure corrections and the learner's own past correct reasoning traces, which are retrieved as task-conditioned guidance for a frozen execution model. During evolution, the graph, a task-level search bandit, and a skill-level routing bandit are updated from the same reward stream, while the learner's backbone remains unchanged. We further provide structural analysis showing how append-only memory growth, bounded curriculum coverage, and task-filtered retrieval together support stable improvement of the retrieval substrate for frozen-learner evolution. Across nine benchmarks spanning mathematical reasoning, multi-hop and open-domain question answering, spatio-temporal analysis, financial numerical reasoning, medical multiple-choice, an open-world survival game, and web navigation, MAGE achieves strong performance against prompt-based frozen-backbone baselines. Ablations show that self-harvested success traces and teacher-written corrections are complementary, with success memories contributing most on reasoning-template-heavy tasks and corrective memories supporting harder composition and interaction settings.

顶级标签: llm agents
详细标签: knowledge graph self-evolution multi-agent frozen backbone benchmark 或 搜索:

MAGE:基于协同进化知识图谱的多智能体自我进化框架 / MAGE: Multi-Agent Self-Evolution with Co-Evolutionary Knowledge Graphs


1️⃣ 一句话总结

MAGE提出了一种让语言模型在自我进化中不改变自身参数的方法,通过构建包含经验、任务和技能等四个子图的协同进化知识图谱来存储和检索历史知识,从而在推理时让冻结的弱模型利用这些外部知识持续提升性能,在数学推理、问答、游戏导航等多个任务上取得了优于传统提示方法的成果。

源自 arXiv: 2605.10064