PCA-Seg: Revisiting Cost Aggregation for Open-Vocabulary Semantic and Part Segmentation
AI 摘要
PCA-Seg提出并行成本聚合方法,解决开放词汇语义分割中语义和空间信息的干扰问题。
主要贡献
- 提出并行成本聚合(PCA)范式
- 设计专家驱动的感知学习(EPL)模块
- 提出特征正交解耦(FOD)策略
方法论
通过并行处理语义和上下文信息,并利用专家驱动的感知学习和特征正交解耦,增强模型对视觉-语言对齐信息的理解。
原文摘要
Recent advances in vision-language models (VLMs) have garnered substantial attention in open-vocabulary semantic and part segmentation (OSPS). However, existing methods extract image-text alignment cues from cost volumes through a serial structure of spatial and class aggregations, leading to knowledge interference between class-level semantics and spatial context. Therefore, this paper proposes a simple yet effective parallel cost aggregation (PCA-Seg) paradigm to alleviate the above challenge, enabling the model to capture richer vision-language alignment information from cost volumes. Specifically, we design an expert-driven perceptual learning (EPL) module that efficiently integrates semantic and contextual streams. It incorporates a multi-expert parser to extract complementary features from multiple perspectives. In addition, a coefficient mapper is designed to adaptively learn pixel-specific weights for each feature, enabling the integration of complementary knowledge into a unified and robust feature embedding. Furthermore, we propose a feature orthogonalization decoupling (FOD) strategy to mitigate redundancy between the semantic and contextual streams, which allows the EPL module to learn diverse knowledge from orthogonalized features. Extensive experiments on eight benchmarks show that each parallel block in PCA-Seg adds merely 0.35M parameters while achieving state-of-the-art OSPS performance.