Multimodal Learning 相关度: 9/10

SAM3-LiteText: An Anatomical Study of the SAM3 Text Encoder for Efficient Vision-Language Segmentation

Chengxi Zeng, Yuxuan Jiang, Ge Gao, Shuai Wang, Duolikun Danier, Bin Zhu, Stevan Rudinac, David Bull, Fan Zhang
arXiv: 2602.12173v1 发布: 2026-02-12 更新: 2026-02-12

AI 摘要

SAM3-LiteText通过知识蒸馏,大幅减少SAM3文本编码器参数,提升视觉语言分割效率。

主要贡献

  • 分析了视觉语言分割中文本提示的冗余性
  • 提出了轻量级文本编码框架SAM3-LiteText
  • 通过实验证明了参数减少的同时保持性能

方法论

通过大规模分析发现冗余,使用知识蒸馏将MobileCLIP作为学生模型优化SAM3文本编码器。

原文摘要

Vision-language segmentation models such as SAM3 enable flexible, prompt-driven visual grounding, but inherit large, general-purpose text encoders originally designed for open-ended language understanding. In practice, segmentation prompts are short, structured, and semantically constrained, leading to substantial over-provisioning in text encoder capacity and persistent computational and memory overhead. In this paper, we perform a large-scale anatomical analysis of text prompting in vision-language segmentation, covering 404,796 real prompts across multiple benchmarks. Our analysis reveals severe redundancy: most context windows are underutilized, vocabulary usage is highly sparse, and text embeddings lie on low-dimensional manifold despite high-dimensional representations. Motivated by these findings, we propose SAM3-LiteText, a lightweight text encoding framework that replaces the original SAM3 text encoder with a compact MobileCLIP student that is optimized by knowledge distillation. Extensive experiments on image and video segmentation benchmarks show that SAM3-LiteText reduces text encoder parameters by up to 88%, substantially reducing static memory footprint, while maintaining segmentation performance comparable to the original model. Code: https://github.com/SimonZeng7108/efficientsam3/tree/sam3_litetext.

标签

视觉语言分割 模型压缩 知识蒸馏 轻量级模型

arXiv 分类

cs.AI