Multimodal Learning 相关度: 9/10

Gaze-Regularized Vision-Language-Action Models for Robotic Manipulation

Anupam Pani, Yanchao Yang
arXiv: 2603.23202v1 发布: 2026-03-24 更新: 2026-03-24

AI 摘要

论文提出了一种基于人类视觉注意力的机器人操作学习框架,提升了机器人操作的性能和可解释性。

主要贡献

  • 提出基于人类注视的VLA模型正则化训练方法
  • 无需额外硬件即可提升机器人操作性能
  • 增强模型的可解释性,学习人类操作策略

方法论

利用时序聚集的注视热图,通过KL散度正则化Transformer的注意力,引导模型关注任务相关特征。

原文摘要

Despite advances in Vision-Language-Action (VLA) models, robotic manipulation struggles with fine-grained tasks because current models lack mechanisms for active visual attention allocation. Human gaze naturally encodes intent, planning, and execution patterns -- offering a powerful supervisory signal for guiding robot perception. We introduce a gaze-regularized training framework that aligns VLA models' internal attention with human visual patterns without architectural modifications or inference-time overhead. Our method transforms temporally aggregated gaze heatmaps into patch-level distributions and regularizes the transformer's attention through KL divergence, creating an inductive bias toward task-relevant features while preserving deployment efficiency. When integrated into existing VLA architectures, our approach yields 4-12% improvements across manipulation benchmarks. The gaze-regularized models reach equivalent performance with fewer training steps and maintain robustness under lighting variations and sensor noise. Beyond performance metrics, the learned attention patterns produce interpretable visualizations that mirror human strategies, enhancing trust in robotic systems. Moreover, our framework requires no eye-tracking equipment and applies directly to existing datasets. These results demonstrate that human perceptual priors can significantly accelerate robot learning while improving both task performance and system interpretability.

标签

机器人操作 视觉语言行动模型 人类注视 注意力机制 可解释性

arXiv 分类

cs.CV