Multimodal Learning 相关度: 9/10

SARE: Sample-wise Adaptive Reasoning for Training-free Fine-grained Visual Recognition

Jingxiao Yang, DaLin He, Miao Pan, Ge Su, Wenqi Zhang, Yifeng Hu, Tangwei Li, Yuke Li, Xuhong Zhang
arXiv: 2603.17729v1 发布: 2026-03-18 更新: 2026-03-18

AI 摘要

SARE提出了一种样本自适应的推理框架,用于无需训练的细粒度视觉识别。

主要贡献

  • 提出样本自适应推理框架SARE
  • 结合快速检索和精细推理的级联设计
  • 自反思经验机制,利用历史错误进行推理引导

方法论

SARE通过级联设计结合检索和推理,并引入自反思机制,针对不同样本自适应地进行推理。

原文摘要

Recent advances in Large Vision-Language Models (LVLMs) have enabled training-free Fine-Grained Visual Recognition (FGVR). However, effectively exploiting LVLMs for FGVR remains challenging due to the inherent visual ambiguity of subordinate-level categories. Existing methods predominantly adopt either retrieval-oriented or reasoning-oriented paradigms to tackle this challenge, but both are constrained by two fundamental limitations:(1) They apply the same inference pipeline to all samples without accounting for uneven recognition difficulty, thereby leading to suboptimal accuracy and efficiency; (2) The lack of mechanisms to consolidate and reuse error-specific experience causes repeated failures on similar challenging cases. To address these limitations, we propose SARE, a Sample-wise Adaptive textbfREasoning framework for training-free FGVR. Specifically, SARE adopts a cascaded design that combines fast candidate retrieval with fine-grained reasoning, invoking the latter only when necessary. In the reasoning process, SARE incorporates a self-reflective experience mechanism that leverages past failures to provide transferable discriminative guidance during inference, without any parameter updates. Extensive experiments across 14 datasets substantiate that SARE achieves state-of-the-art performance while substantially reducing computational overhead.

标签

细粒度视觉识别 大型视觉语言模型 自适应推理

arXiv 分类

cs.CV cs.AI