菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-15
📄 Abstract - PACEvolve: Enabling Long-Horizon Progress-Aware Consistent Evolution

Large Language Models (LLMs) have emerged as powerful operators for evolutionary search, yet the design of efficient search scaffolds remains ad hoc. While promising, current LLM-in-the-loop systems lack a systematic approach to managing the evolutionary process. We identify three distinct failure modes: Context Pollution, where experiment history biases future candidate generation; Mode Collapse, where agents stagnate in local minima due to poor exploration-exploitation balance; and Weak Collaboration, where rigid crossover strategies fail to leverage parallel search trajectories effectively. We introduce Progress-Aware Consistent Evolution (PACEvolve), a framework designed to robustly govern the agent's context and search dynamics, to address these challenges. PACEvolve combines hierarchical context management (HCM) with pruning to address context pollution; momentum-based backtracking (MBB) to escape local minima; and a self-adaptive sampling policy that unifies backtracking and crossover for dynamic search coordination (CE), allowing agents to balance internal refinement with cross-trajectory collaboration. We demonstrate that PACEvolve provides a systematic path to consistent, long-horizon self-improvement, achieving state-of-the-art results on LLM-SR and KernelBench, while discovering solutions surpassing the record on Modded NanoGPT.

顶级标签: llm agents model training
详细标签: evolutionary search long-horizon planning context management search dynamics self-improvement 或 搜索:

PACEvolve:实现长周期、进度感知的一致性进化 / PACEvolve: Enabling Long-Horizon Progress-Aware Consistent Evolution


1️⃣ 一句话总结

这篇论文提出了一个名为PACEvolve的新框架,它通过智能管理搜索过程中的信息和动态协调不同探索路径,解决了大语言模型在长期进化搜索中容易出现的三大问题,从而让AI系统能更稳定、高效地自我改进并发现更好的解决方案。

源自 arXiv: 2601.10657