📄
Abstract - AgenticCache: Cache-Driven Asynchronous Planning for Embodied AI Agents
Embodied AI agents increasingly rely on large language models (LLMs) for planning, yet per-step LLM calls impose severe latency and cost. In this paper, we show that embodied tasks exhibit strong plan locality, where the next plan is largely predictable from the current one. Building on this, we introduce AgenticCache, a planning framework that reuses cached plans to avoid per-step LLM calls. In AgenticCache, each agent queries a runtime cache of frequent plan transitions, while a background Cache Updater asynchronously calls the LLM to validate and refine cached entries. Across four multi-agent embodied benchmarks, AgenticCache improves task success rate by 22% on average across 12 configurations (4 benchmarks x 3 models), reduces simulation latency by 65%, and lowers token usage by 50%. Cache-based plan reuse thus offers a practical path to low-latency, low-cost embodied agents. Code is available at this https URL.
AgenticCache:面向具身AI智能体的缓存驱动异步规划框架 /
AgenticCache: Cache-Driven Asynchronous Planning for Embodied AI Agents
1️⃣ 一句话总结
为解决具身AI智能体每次行动都需要调用大语言模型导致的高延迟和高成本问题,本文提出了一种名为AgenticCache的规划框架,它利用缓存已生成的行动序列来替代频繁的模型调用,并通过后台异步更新保证缓存内容的准确性,在多个标准测试中将任务成功率提升22%,同时将延迟和计算成本分别降低65%和50%。