Delving into Spectral Clustering with Vision-Language Representations
AI 摘要
该论文提出一种基于视觉-语言表征的谱聚类方法,显著提升了聚类性能。
主要贡献
- 提出基于视觉-语言模型中跨模态对齐的谱聚类方法
- 引入神经正切核并使用积极名词进行锚定
- 提出正则化的亲和力扩散机制
方法论
利用预训练视觉-语言模型的跨模态对齐,通过神经正切核和亲和力扩散机制构建亲和力矩阵,进行谱聚类。
原文摘要
Spectral clustering is known as a powerful technique in unsupervised data analysis. The vast majority of approaches to spectral clustering are driven by a single modality, leaving the rich information in multi-modal representations untapped. Inspired by the recent success of vision-language pre-training, this paper enriches the landscape of spectral clustering from a single-modal to a multi-modal regime. Particularly, we propose Neural Tangent Kernel Spectral Clustering that leverages cross-modal alignment in pre-trained vision-language models. By anchoring the neural tangent kernel with positive nouns, i.e., those semantically close to the images of interest, we arrive at formulating the affinity between images as a coupling of their visual proximity and semantic overlap. We show that this formulation amplifies within-cluster connections while suppressing spurious ones across clusters, hence encouraging block-diagonal structures. In addition, we present a regularized affinity diffusion mechanism that adaptively ensembles affinity matrices induced by different prompts. Extensive experiments on \textbf{16} benchmarks -- including classical, large-scale, fine-grained and domain-shifted datasets -- manifest that our method consistently outperforms the state-of-the-art by a large margin.