Multimodal Learning 相关度: 8/10

Lifelong Imitation Learning with Multimodal Latent Replay and Incremental Adjustment

Fanqi Yu, Matteo Tiezzi, Tommaso Apicella, Cigdem Beyan, Vittorio Murino
arXiv: 2603.10929v1 发布: 2026-03-11 更新: 2026-03-11

AI 摘要

提出一种终身模仿学习框架,利用多模态潜在空间和增量调整实现策略持续优化。

主要贡献

  • 提出基于多模态潜在空间的终身模仿学习框架
  • 引入增量特征调整机制,稳定任务嵌入
  • 在LIBERO基准测试中取得了SOTA结果

方法论

利用多模态潜在空间存储经验,通过增量特征调整机制约束任务嵌入,实现策略持续优化。

原文摘要

We introduce a lifelong imitation learning framework that enables continual policy refinement across sequential tasks under realistic memory and data constraints. Our approach departs from conventional experience replay by operating entirely in a multimodal latent space, where compact representations of visual, linguistic, and robot's state information are stored and reused to support future learning. To further stabilize adaptation, we introduce an incremental feature adjustment mechanism that regularizes the evolution of task embeddings through an angular margin constraint, preserving inter-task distinctiveness. Our method establishes a new state of the art in the LIBERO benchmarks, achieving 10-17 point gains in AUC and up to 65% less forgetting compared to previous leading methods. Ablation studies confirm the effectiveness of each component, showing consistent gains over alternative strategies. The code is available at: https://github.com/yfqi/lifelong_mlr_ifa.

标签

终身学习 模仿学习 多模态学习 持续学习

arXiv 分类

cs.CV cs.RO