基于显式潜在状态转移的文本规划 / Textual Planning with Explicit Latent Transitions
1️⃣ 一句话总结
这篇论文提出了一种名为EmbedPlan的新方法,它通过让模型在固定的文本语义向量空间中预测和检索下一个状态,从而绕过了传统大语言模型逐字生成和重复计算的瓶颈,实现了更快速、高效的文本规划,但其跨领域泛化能力仍有待提升。
Planning with LLMs is bottlenecked by token-by-token generation and repeated full forward passes, making multi-step lookahead and rollout-based search expensive in latency and compute. We propose EmbedPlan, which replaces autoregressive next-state generation with a lightweight transition model operating in a frozen language embedding space. EmbedPlan encodes natural language state and action descriptions into vectors, predicts the next-state embedding, and retrieves the next state by nearest-neighbor similarity, enabling fast planning computation without fine-tuning the encoder. We evaluate next-state prediction across nine classical planning domains using six evaluation protocols of increasing difficulty: interpolation, plan-variant, extrapolation, multi-domain, cross-domain, and leave-one-out. Results show near-perfect interpolation performance but a sharp degradation when generalization requires transfer to unseen problems or unseen domains; plan-variant evaluation indicates generalization to alternative plans rather than memorizing seen trajectories. Overall, frozen embeddings support within-domain dynamics learning after observing a domain's transitions, while transfer across domain boundaries remains a bottleneck.
基于显式潜在状态转移的文本规划 / Textual Planning with Explicit Latent Transitions
这篇论文提出了一种名为EmbedPlan的新方法,它通过让模型在固定的文本语义向量空间中预测和检索下一个状态,从而绕过了传统大语言模型逐字生成和重复计算的瓶颈,实现了更快速、高效的文本规划,但其跨领域泛化能力仍有待提升。
源自 arXiv: 2602.04557