Multimodal Learning 相关度: 9/10

TTA-Vid: Generalized Test-Time Adaptation for Video Reasoning

Soumya Shamarao Jahagirdar, Edson Araujo, Anna Kukleva, M. Jehanzeb Mirza, Saurabhchand Bhati, Samuel Thomas, Brian Kingsbury, Rogerio Feris, James R. Glass, Hilde Kuehne
arXiv: 2604.00696v1 发布: 2026-04-01 更新: 2026-04-01

AI 摘要

TTA-Vid利用测试时强化学习,无需标注数据即可使视频理解模型适应新领域。

主要贡献

  • 提出TTA-Vid,一种测试时视频理解自适应方法
  • 使用批量感知频率奖励作为伪标签更新模型
  • 提出多臂老虎机策略进行自适应帧选择

方法论

利用测试时强化学习,通过批量感知频率奖励和多臂老虎机策略,使模型在推理时自适应视频数据。

原文摘要

Recent video reasoning models have shown strong results on temporal and multimodal understanding, yet they depend on large-scale supervised data and multi-stage training pipelines, making them costly to train and difficult to adapt to new domains. In this work, we leverage the paradigm of Test-Time Reinforcement Learning on video-language data to allow for adapting a pretrained model to incoming video samples at test-time without explicit labels. The proposed test-time adaptation for video approach (TTA-Vid) combines two components that work simultaneously: (1) a test-time adaptation that performs step-by-step reasoning at inference time on multiple frame subsets. We then use a batch-aware frequency-based reward computed across different frame subsets as pseudo ground truth to update the model. It shows that the resulting model trained on a single batch or even a single sample from a dataset, is able to generalize at test-time to the whole dataset and even across datasets. Because the adaptation occurs entirely at test time, our method requires no ground-truth annotations or dedicated training splits. Additionally, we propose a multi-armed bandit strategy for adaptive frame selection that learns to prioritize informative frames, guided by the same reward formulation. Our evaluation shows that TTA-Vid yields consistent improvements across various video reasoning tasks and is able to outperform current state-of-the-art methods trained on large-scale data. This highlights the potential of test-time reinforcement learning for temporal multimodal understanding.

标签

测试时自适应 强化学习 视频理解 多模态学习

arXiv 分类

cs.CV