SARE: Sample-wise Adaptive Reasoning for Training-free Fine-grained Visual Recognition
AI 摘要
SARE提出了一种样本自适应的推理框架,用于无需训练的细粒度视觉识别。
主要贡献
- 提出样本自适应推理框架SARE
- 结合快速检索和精细推理的级联设计
- 自反思经验机制,利用历史错误进行推理引导
方法论
SARE通过级联设计结合检索和推理,并引入自反思机制,针对不同样本自适应地进行推理。
原文摘要
Recent advances in Large Vision-Language Models (LVLMs) have enabled training-free Fine-Grained Visual Recognition (FGVR). However, effectively exploiting LVLMs for FGVR remains challenging due to the inherent visual ambiguity of subordinate-level categories. Existing methods predominantly adopt either retrieval-oriented or reasoning-oriented paradigms to tackle this challenge, but both are constrained by two fundamental limitations:(1) They apply the same inference pipeline to all samples without accounting for uneven recognition difficulty, thereby leading to suboptimal accuracy and efficiency; (2) The lack of mechanisms to consolidate and reuse error-specific experience causes repeated failures on similar challenging cases. To address these limitations, we propose SARE, a Sample-wise Adaptive textbfREasoning framework for training-free FGVR. Specifically, SARE adopts a cascaded design that combines fast candidate retrieval with fine-grained reasoning, invoking the latter only when necessary. In the reasoning process, SARE incorporates a self-reflective experience mechanism that leverages past failures to provide transferable discriminative guidance during inference, without any parameter updates. Extensive experiments across 14 datasets substantiate that SARE achieves state-of-the-art performance while substantially reducing computational overhead.