Multimodal Learning 相关度: 9/10

ViKey: Enhancing Temporal Understanding in Videos via Visual Prompting

Yeonkyung Lee, Dayun Ju, Youngmin Kim, Seil Kang, Seong Jae Hwang
arXiv: 2603.23186v1 发布: 2026-03-24 更新: 2026-03-24

AI 摘要

ViKey通过视觉提示和关键词帧映射,提升视频LLM在稀疏帧下的时间推理能力。

主要贡献

  • 提出ViKey框架,结合视觉提示和关键词帧映射
  • 利用帧索引作为字典键,连接文本提示和相关帧
  • 在稀疏帧下显著提升时间推理能力,接近稠密帧性能

方法论

提出训练无关的ViKey框架,结合视觉提示(帧序号标注)和关键词帧映射模块,增强LLM的时间推理能力。

原文摘要

Recent advancements in Video Large Language Models (VideoLLMs) have enabled strong performance across diverse multimodal video tasks. To reduce the high computational cost of processing dense video frames, efficiency-oriented methods such as frame selection have been widely adopted. While effective at minimizing redundancy, these methods often cause notable performance drops on tasks requiring temporal reasoning. Unlike humans, who can infer event progression from sparse visual cues, VideoLLMs frequently misinterpret temporal relations when intermediate frames are omitted. To address this limitation, we explore visual prompting (VP) as a lightweight yet effective way to enhance temporal understanding in VideoLLMs. Our analysis reveals that simply annotating each frame with explicit ordinal information helps the model perceive temporal continuity. This visual cue also supports frame-level referencing and mitigates positional ambiguity within a sparsely sampled sequence. Building on these insights, we introduce ViKey, a training-free framework that combines VP with a lightweight Keyword-Frame Mapping (KFM) module. KFM leverages frame indices as dictionary-like keys to link textual cues to the most relevant frames, providing explicit temporal anchors during inference. Despite its simplicity, our approach substantially improves temporal reasoning and, on some datasets, preserves dense-frame baseline performance with as few as 20% of frames.

标签

视频LLM 时间推理 视觉提示 多模态学习

arXiv 分类

cs.CV