菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-21
📄 Abstract - A Self-Evolving Framework for Efficient Terminal Agents via Observational Context Compression

As model capabilities advance, research has increasingly shifted toward long-horizon, multi-turn terminal-centric agentic tasks, where raw environment feedback is often preserved in the interaction history to support future decisions. However, repeatedly retaining such feedback introduces substantial redundancy and causes cumulative token cost to grow quadratically with the number of steps, hindering long-horizon reasoning. Although observation compression can mitigate this issue, the heterogeneity of terminal environments makes heuristic-based or fixed-prompt methods difficult to generalize. We propose TACO, a plug-and-play, self-evolving Terminal Agent Compression framework that automatically discovers and refines compression rules from interaction trajectories for existing terminal agents. Experiments on TerminalBench (TB 1.0 and TB 2.0) and four additional terminal-related benchmarks (i.e., SWE-Bench Lite, CompileBench, DevEval, and CRUST-Bench) show that TACO consistently improves performance across mainstream agent frameworks and strong backbone models. With MiniMax-2.5, it improves performance on most benchmarks while reducing token overhead by around 10%. On TerminalBench, it brings consistent gains of 1%-4% across strong agentic models, and further improves accuracy by around 2%-3% under the same token budget. These results demonstrate the effectiveness and generalization of self-evolving, task-aware compression for terminal agents.

顶级标签: agents llm systems
详细标签: compression terminal agents token efficiency self-evolving long-horizon 或 搜索:

一种通过观察上下文压缩实现高效终端智能体的自演化框架 / A Self-Evolving Framework for Efficient Terminal Agents via Observational Context Compression


1️⃣ 一句话总结

本文提出了一种名为TACO的自适应框架,能让终端操作型AI智能体在长任务交互过程中,自动学习和优化压缩历史观察信息的方法,从而在不降低甚至提升任务表现的前提下,显著减少计算开销和令牌消耗。

源自 arXiv: 2604.19572