菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-04
📄 Abstract - Textual Planning with Explicit Latent Transitions

Planning with LLMs is bottlenecked by token-by-token generation and repeated full forward passes, making multi-step lookahead and rollout-based search expensive in latency and compute. We propose EmbedPlan, which replaces autoregressive next-state generation with a lightweight transition model operating in a frozen language embedding space. EmbedPlan encodes natural language state and action descriptions into vectors, predicts the next-state embedding, and retrieves the next state by nearest-neighbor similarity, enabling fast planning computation without fine-tuning the encoder. We evaluate next-state prediction across nine classical planning domains using six evaluation protocols of increasing difficulty: interpolation, plan-variant, extrapolation, multi-domain, cross-domain, and leave-one-out. Results show near-perfect interpolation performance but a sharp degradation when generalization requires transfer to unseen problems or unseen domains; plan-variant evaluation indicates generalization to alternative plans rather than memorizing seen trajectories. Overall, frozen embeddings support within-domain dynamics learning after observing a domain's transitions, while transfer across domain boundaries remains a bottleneck.

顶级标签: llm agents systems
详细标签: planning latent transitions embedding space next-state prediction generalization 或 搜索:

基于显式潜在状态转移的文本规划 / Textual Planning with Explicit Latent Transitions


1️⃣ 一句话总结

这篇论文提出了一种名为EmbedPlan的新方法,它通过让模型在固定的文本语义向量空间中预测和检索下一个状态,从而绕过了传统大语言模型逐字生成和重复计算的瓶颈,实现了更快速、高效的文本规划,但其跨领域泛化能力仍有待提升。

源自 arXiv: 2602.04557