菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-11
📄 Abstract - ET-Agent: Incentivizing Effective Tool-Integrated Reasoning Agent via Behavior Calibration

Large Language Models (LLMs) can extend their parameter knowledge limits by adopting the Tool-Integrated Reasoning (TIR) paradigm. However, existing LLM-based agent training framework often focuses on answers' accuracy, overlooking specific alignment for behavior patterns. Consequently, agent often exhibits ineffective actions during TIR tasks, such as redundant and insufficient tool calls. How to calibrate erroneous behavioral patterns when executing TIR tasks, thereby exploring effective trajectories, remains an open-ended problem. In this paper, we propose ET-Agent, a training framework for calibrating agent's tool-use behavior through two synergistic perspectives: Self-evolving Data Flywheel and Behavior Calibration Training. Specifically, we introduce a self-evolutionary data flywheel to generate enhanced data, used to fine-tune LLM to improve its exploration ability. Based on this, we implement an two-phases behavior-calibration training framework. It is designed to progressively calibrate erroneous behavioral patterns to optimal behaviors. Further in-depth experiments confirm the superiority of \ourmodel{} across multiple dimensions, including correctness, efficiency, reasoning conciseness, and tool execution accuracy. Our ET-Agent framework provides practical insights for research in the TIR field. Codes can be found in this https URL

顶级标签: llm agents model training
详细标签: tool use behavior calibration agent training reasoning agents fine-tuning 或 搜索:

ET-Agent:通过行为校准激励有效的工具集成推理智能体 / ET-Agent: Incentivizing Effective Tool-Integrated Reasoning Agent via Behavior Calibration


1️⃣ 一句话总结

这篇论文提出了一个名为ET-Agent的训练框架,它通过自我进化的数据循环和两阶段行为校准训练,来纠正大语言模型智能体在使用外部工具时出现的无效行为(如工具调用冗余或不足),从而提升其任务执行的正确性和效率。

源自 arXiv: 2601.06860