菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-14
📄 Abstract - MAXS: Meta-Adaptive Exploration with LLM Agents

Large Language Model (LLM) Agents exhibit inherent reasoning abilities through the collaboration of multiple tools. However, during agent inference, existing methods often suffer from (i) locally myopic generation, due to the absence of lookahead, and (ii) trajectory instability, where minor early errors can escalate into divergent reasoning paths. These issues make it difficult to balance global effectiveness and computational efficiency. To address these two issues, we propose meta-adaptive exploration with LLM agents this https URL, a meta-adaptive reasoning framework based on LLM Agents that flexibly integrates tool execution and reasoning planning. MAXS employs a lookahead strategy to extend reasoning paths a few steps ahead, estimating the advantage value of tool usage, and combines step consistency variance and inter-step trend slopes to jointly select stable, consistent, and high-value reasoning steps. Additionally, we introduce a trajectory convergence mechanism that controls computational cost by halting further rollouts once path consistency is achieved, enabling a balance between resource efficiency and global effectiveness in multi-tool reasoning. We conduct extensive empirical studies across three base models (MiMo-VL-7B, Qwen2.5-VL-7B, Qwen2.5-VL-32B) and five datasets, demonstrating that MAXS consistently outperforms existing methods in both performance and inference efficiency. Further analysis confirms the effectiveness of our lookahead strategy and tool usage.

顶级标签: llm agents systems
详细标签: reasoning framework lookahead strategy tool usage trajectory stability meta-adaptive exploration 或 搜索:

MAXS:基于大语言模型智能体的元自适应探索 / MAXS: Meta-Adaptive Exploration with LLM Agents


1️⃣ 一句话总结

本文提出了一种名为MAXS的智能推理框架,它通过前瞻性策略和轨迹稳定性评估,有效解决了大语言模型智能体在工具调用时目光短浅和推理路径不稳定的问题,从而在保证推理质量的同时提升了计算效率。

源自 arXiv: 2601.09259