Chameleon: Episodic Memory for Long-Horizon Robotic Manipulation
AI 摘要
Chameleon提出了一种几何感知的多模态记忆模型,用于解决机器人操作中的长程依赖问题。
主要贡献
- 提出Chameleon记忆模型,利用几何信息进行精确回忆
- 引入Camo-Dataset,一个真实机器人数据集,用于评估记忆能力
- 实验证明Chameleon在感知混淆环境中提升了决策可靠性和长程控制
方法论
通过几何grounded的多模态tokens保存上下文,使用可微的记忆栈实现目标导向的回忆。
原文摘要
Robotic manipulation often requires memory: occlusion and state changes can make decision-time observations perceptually aliased, making action selection non-Markovian at the observation level because the same observation may arise from different interaction histories. Most embodied agents implement memory via semantically compressed traces and similarity-based retrieval, which discards disambiguating fine-grained perceptual cues and can return perceptually similar but decision-irrelevant episodes. Inspired by human episodic memory, we propose Chameleon, which writes geometry-grounded multimodal tokens to preserve disambiguating context and produces goal-directed recall through a differentiable memory stack. We also introduce Camo-Dataset, a real-robot UR5e dataset spanning episodic recall, spatial tracking, and sequential manipulation under perceptual aliasing. Across tasks, Chameleon consistently improves decision reliability and long-horizon control over strong baselines in perceptually confusable settings.