菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-16
📄 Abstract - Discovering Novel LLM Experts via Task-Capability Coevolution

Frontier model developers aim to train models continually to possess emergent, diverse capabilities. To extend capabilities, the current pre-training and post-training paradigm requires manually starting training runs with static datasets or reward functions every time. Addressing this limitation, our work pursues the insight that open-endedness (via the coevolution of models and tasks) can discover models with increasingly novel skills in a single run. We introduce a new model development framework that extends coevolution to large language model (LLM) discovery, open-ended \textit{Assessment Coevolving with Diverse Capabilities} (AC/DC). AC/DC evolves both LLMs via model merging and natural language tasks via synthetic data generation. AC/DC discovers growing archives of LLMs that surpass the capabilities of larger LLMs while taking up less GPU memory. In particular, our LLM populations achieve a broader Coverage of expertise than other curated models or baselines on downstream benchmarks, without \textit{any} explicit benchmark optimization. Furthermore, AC/DC improves Coverage over time, continually innovates on tasks and models, and improves performance in multi-agent best-of-N selection. Our findings highlight the potential of coevolution as a means of discovering broader sets of capabilities from base LLMs. Overall, AC/DC brings us one step closer to a profoundly new paradigm of LLM development, where continual improvements to the diversity of model capabilities can be accelerated by leveraging existing models as stepping stones to increasingly powerful models.

顶级标签: llm model training agents
详细标签: coevolution model merging open-ended learning capability discovery synthetic data generation 或 搜索:

通过任务与能力协同演化发现新型大语言模型专家 / Discovering Novel LLM Experts via Task-Capability Coevolution


1️⃣ 一句话总结

这篇论文提出了一种名为AC/DC的新框架,它通过让大语言模型和自然语言任务在同一个系统中协同进化,自动发现并培养出具备多样化、新颖能力且效率更高的模型,为持续提升AI能力开辟了一条自动化新路径。

源自 arXiv: 2604.14969