Multimodal Learning 相关度: 7/10

IConE: Batch Independent Collapse Prevention for Self-Supervised Representation Learning

Konstantinos Almpanakis, Anna Kreshuk
arXiv: 2603.15263v1 发布: 2026-03-16 更新: 2026-03-16

AI 摘要

IConE提出了一种不依赖batch size的自监督学习方法,通过全局可学习实例嵌入防止表征坍塌。

主要贡献

  • 提出了IConE框架,解耦了坍塌预防和batch size
  • 引入了可学习的辅助实例嵌入,通过多样性目标正则化
  • 在小batch size和类别不平衡数据集上表现优于现有方法

方法论

IConE维护一个全局可学习实例嵌入集合,并通过显式的多样性目标进行正则化,防止表征坍塌。

原文摘要

Self-supervised learning (SSL) has revolutionized representation learning, with Joint-Embedding Architectures (JEAs) emerging as an effective approach for capturing semantic features. Existing JEAs rely on implicit or explicit batch interaction -- via negative sampling or statistical regularization -- to prevent representation collapse. This reliance becomes problematic in regimes where batch sizes must be small, such as high-dimensional scientific data, where memory constraints and class imbalance make large, well-balanced batches infeasible. We introduce IConE (Instance-Contrasted Embeddings), a framework that decouples collapse prevention from the training batch size. Rather than enforcing diversity through batch statistics, IConE maintains a global set of learnable auxiliary instance embeddings regularized by an explicit diversity objective. This transfers the anti-collapse mechanism from the transient batch to a dataset-level embedding space, allowing stable training even when batch statistics are unreliable, down to batch size 1. Across diverse 2D and 3D biomedical modalities, IConE outperforms strong contrastive and non-contrastive baselines throughout the small-batch regime (from B=1 to B=64) and demonstrates marked robustness to severe class imbalance. Geometric analysis shows that IConE preserves high intrinsic dimensionality in the learned representations, preventing the collapse observed in existing JEAs as batch sizes shrink.

标签

自监督学习 对比学习 表征学习 小批量 类别不平衡

arXiv 分类

cs.CV cs.LG