Multimodal Learning 相关度: 9/10

An Approach to Enriching Surgical Video Datasets for Fine-Grained Spatial-Temporal Understanding of Vision-Language Models

Lennart Maack, Alexander Schlaefer
arXiv: 2604.00784v1 发布: 2026-04-01 更新: 2026-04-01

AI 摘要

提出SurgSTU-Pipeline自动生成手术视频数据集,提升VLM对手术视频时空理解能力。

主要贡献

  • 提出了SurgSTU-Pipeline,一个用于生成手术视频数据集的确定性流程
  • 构建了包含15万个细粒度时空问答样本的SurgSTU数据集
  • 验证了SurgSTU数据集在提升VLM手术视频时空理解方面的有效性

方法论

提出了SurgSTU-Pipeline,通过时序和空间连续性过滤,自动生成高质量手术视频问答数据集,用于训练VLM。

原文摘要

Surgical video understanding is a crucial prerequisite for advancing Computer-Assisted Surgery. While vision-language models (VLMs) have recently been applied to the surgical domain, existing surgical vision-language datasets lack in capturing and evaluating complex, interleaved spatial-temporal dynamics. Creating large scale datasets that accurately represent fine-grained spatial-temporal relationships in surgical videos is challenging due to costly manual annotations or error-prone generation using large language models. To address this gap, we introduce the SurgSTU-Pipeline, a deterministic generation pipeline featuring temporal and spatial continuity filtering to reliably create surgical datasets for fine-grained spatial-temporal multimodal understanding. Applying this pipeline to publicly available surgical datasets, we create the SurgSTU dataset, comprising 7515 video clips densely extended with 150k fine-grained spatial-temporal question-answer samples. Our comprehensive evaluation shows that while state-of-the-art generalist VLMs struggle in zero-shot settings, their spatial-temporal capabilities can be improved through in-context learning. A fine-tuned VLM on the SurgSTU training dataset achieves highest performance among all spatial-temporal tasks, validating the dataset's efficacy to improve spatial-temporal understanding of VLMs in surgical videos. Code will be made publicly available.

标签

VLM Surgical Video Understanding Dataset Generation

arXiv 分类

cs.CV