Multimodal Learning 相关度: 8/10

Universal Algorithm-Implicit Learning

Stefano Woerner, Seong Joon Oh, Christian F. Baumgartner
arXiv: 2602.14761v1 发布: 2026-02-16 更新: 2026-02-16

AI 摘要

提出通用元学习框架和算法隐式学习概念,Transformer实现跨域、跨模态和高类别任务的元学习。

主要贡献

  • 提出算法隐式学习框架
  • 设计TAIL元学习模型
  • 实现跨域、跨模态和高类别泛化

方法论

提出TAIL,一种基于Transformer的算法隐式元学习器,通过随机投影和嵌入处理不同模态和类别。

原文摘要

Current meta-learning methods are constrained to narrow task distributions with fixed feature and label spaces, limiting applicability. Moreover, the current meta-learning literature uses key terms like "universal" and "general-purpose" inconsistently and lacks precise definitions, hindering comparability. We introduce a theoretical framework for meta-learning which formally defines practical universality and introduces a distinction between algorithm-explicit and algorithm-implicit learning, providing a principled vocabulary for reasoning about universal meta-learning methods. Guided by this framework, we present TAIL, a transformer-based algorithm-implicit meta-learner that functions across tasks with varying domains, modalities, and label configurations. TAIL features three innovations over prior transformer-based meta-learners: random projections for cross-modal feature encoding, random injection label embeddings that extrapolate to larger label spaces, and efficient inline query processing. TAIL achieves state-of-the-art performance on standard few-shot benchmarks while generalizing to unseen domains. Unlike other meta-learning methods, it also generalizes to unseen modalities, solving text classification tasks despite training exclusively on images, handles tasks with up to 20$\times$ more classes than seen during training, and provides orders-of-magnitude computational savings over prior transformer-based approaches.

标签

Meta-learning Transformer Multimodal Learning Few-shot Learning Generalization

arXiv 分类

cs.LG cs.AI cs.CV