AI Agents 相关度: 9/10

Textual Planning with Explicit Latent Transitions

Eliezer Shlomi, Ido Levy, Eilam Shapira, Michael Katz, Guy Uziel, Segev Shlomov, Nir Mashkif, Roi Reichart, Sarah Keren
arXiv: 2602.04557v1 发布: 2026-02-04 更新: 2026-02-04

AI 摘要

EmbedPlan通过在冻结语言嵌入空间中进行状态转移预测,加速LLM规划过程。

主要贡献

  • 提出EmbedPlan,一种基于嵌入空间的规划方法
  • 使用轻量级的状态转移模型代替自回归生成
  • 无需微调编码器即可实现快速规划

方法论

将状态和动作描述编码为向量,预测下一状态嵌入,通过最近邻搜索检索下一状态。

原文摘要

Planning with LLMs is bottlenecked by token-by-token generation and repeated full forward passes, making multi-step lookahead and rollout-based search expensive in latency and compute. We propose EmbedPlan, which replaces autoregressive next-state generation with a lightweight transition model operating in a frozen language embedding space. EmbedPlan encodes natural language state and action descriptions into vectors, predicts the next-state embedding, and retrieves the next state by nearest-neighbor similarity, enabling fast planning computation without fine-tuning the encoder. We evaluate next-state prediction across nine classical planning domains using six evaluation protocols of increasing difficulty: interpolation, plan-variant, extrapolation, multi-domain, cross-domain, and leave-one-out. Results show near-perfect interpolation performance but a sharp degradation when generalization requires transfer to unseen problems or unseen domains; plan-variant evaluation indicates generalization to alternative plans rather than memorizing seen trajectories. Overall, frozen embeddings support within-domain dynamics learning after observing a domain's transitions, while transfer across domain boundaries remains a bottleneck.

标签

LLM Planning Embedding State Transition

arXiv 分类

cs.CL