Textual Planning with Explicit Latent Transitions
AI 摘要
EmbedPlan通过在冻结语言嵌入空间中进行状态转移预测,加速LLM规划过程。
主要贡献
- 提出EmbedPlan,一种基于嵌入空间的规划方法
- 使用轻量级的状态转移模型代替自回归生成
- 无需微调编码器即可实现快速规划
方法论
将状态和动作描述编码为向量,预测下一状态嵌入,通过最近邻搜索检索下一状态。
原文摘要
Planning with LLMs is bottlenecked by token-by-token generation and repeated full forward passes, making multi-step lookahead and rollout-based search expensive in latency and compute. We propose EmbedPlan, which replaces autoregressive next-state generation with a lightweight transition model operating in a frozen language embedding space. EmbedPlan encodes natural language state and action descriptions into vectors, predicts the next-state embedding, and retrieves the next state by nearest-neighbor similarity, enabling fast planning computation without fine-tuning the encoder. We evaluate next-state prediction across nine classical planning domains using six evaluation protocols of increasing difficulty: interpolation, plan-variant, extrapolation, multi-domain, cross-domain, and leave-one-out. Results show near-perfect interpolation performance but a sharp degradation when generalization requires transfer to unseen problems or unseen domains; plan-variant evaluation indicates generalization to alternative plans rather than memorizing seen trajectories. Overall, frozen embeddings support within-domain dynamics learning after observing a domain's transitions, while transfer across domain boundaries remains a bottleneck.