TTA-Vid: Generalized Test-Time Adaptation for Video Reasoning
AI 摘要
TTA-Vid利用测试时强化学习,无需标注数据即可使视频理解模型适应新领域。
主要贡献
- 提出TTA-Vid,一种测试时视频理解自适应方法
- 使用批量感知频率奖励作为伪标签更新模型
- 提出多臂老虎机策略进行自适应帧选择
方法论
利用测试时强化学习,通过批量感知频率奖励和多臂老虎机策略,使模型在推理时自适应视频数据。
原文摘要
Recent video reasoning models have shown strong results on temporal and multimodal understanding, yet they depend on large-scale supervised data and multi-stage training pipelines, making them costly to train and difficult to adapt to new domains. In this work, we leverage the paradigm of Test-Time Reinforcement Learning on video-language data to allow for adapting a pretrained model to incoming video samples at test-time without explicit labels. The proposed test-time adaptation for video approach (TTA-Vid) combines two components that work simultaneously: (1) a test-time adaptation that performs step-by-step reasoning at inference time on multiple frame subsets. We then use a batch-aware frequency-based reward computed across different frame subsets as pseudo ground truth to update the model. It shows that the resulting model trained on a single batch or even a single sample from a dataset, is able to generalize at test-time to the whole dataset and even across datasets. Because the adaptation occurs entirely at test time, our method requires no ground-truth annotations or dedicated training splits. Additionally, we propose a multi-armed bandit strategy for adaptive frame selection that learns to prioritize informative frames, guided by the same reward formulation. Our evaluation shows that TTA-Vid yields consistent improvements across various video reasoning tasks and is able to outperform current state-of-the-art methods trained on large-scale data. This highlights the potential of test-time reinforcement learning for temporal multimodal understanding.