菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-14
📄 Abstract - Local-Splitter: A Measurement Study of Seven Tactics for Reducing Cloud LLM Token Usage on Coding-Agent Workloads

We present a systematic measurement study of seven tactics for reducing cloud LLM token usage when a small local model can act as a triage layer in front of a frontier cloud model. The tactics are: (1) local routing, (2) prompt compression, (3) semantic caching, (4) local drafting with cloud review, (5) minimal-diff edits, (6) structured intent extraction, and (7) batching with vendor prompt caching. We implement all seven in an open-source shim that speaks both MCP and the OpenAI-compatible HTTP surface, supporting any local model via Ollama and any cloud model via an OpenAI-compatible endpoint. We evaluate each tactic individually, in pairs, and in a greedy-additive subset across four coding-agent workload classes (edit-heavy, explanation-heavy, general chat, RAG-heavy). We measure tokens saved, dollar cost, latency, and routing accuracy. Our headline finding is that T1 (local routing) combined with T2 (prompt compression) achieves 45-79% cloud token savings on edit-heavy and explanation-heavy workloads, while on RAG-heavy workloads the full tactic set including T4 (draft-review) achieves 51% savings. We observe that the optimal tactic subset is workload-dependent, which we believe is the most actionable finding for practitioners deploying coding agents today.

顶级标签: llm agents systems
详细标签: token reduction coding agents cost optimization model cascading efficiency tactics 或 搜索:

Local-Splitter:针对编码智能体工作负载,减少云端大语言模型令牌使用的七种策略的测量研究 / Local-Splitter: A Measurement Study of Seven Tactics for Reducing Cloud LLM Token Usage on Coding-Agent Workloads


1️⃣ 一句话总结

这项研究系统地测量了七种策略,通过让一个小型本地模型充当‘把关人’来筛选任务,从而有效减少调用昂贵云端大语言模型的令牌消耗,最高可节省79%的成本,并且发现最佳策略组合取决于具体的工作负载类型。

源自 arXiv: 2604.12301