Layer-Specific Lipschitz Modulation for Fault-Tolerant Multimodal Representation Learning
AI 摘要
提出一种用于容错多模态表示学习的层特异性Lipschitz调制框架。
主要贡献
- 提出基于Lipschitz和Jacobian的故障敏感性度量
- 设计了两阶段自监督训练方案,增强异常检测和纠正能力
- 引入层特异性Lipschitz调制和梯度裁剪来控制模块敏感性
方法论
基于扰动传播理论,提出Lipschitz调制,构建多模态卷积自编码器,通过自监督学习进行故障检测和校正。
原文摘要
Modern multimodal systems deployed in industrial and safety-critical environments must remain reliable under partial sensor failures, signal degradation, or cross-modal inconsistencies. This work introduces a mathematically grounded framework for fault-tolerant multimodal representation learning that unifies self-supervised anomaly detection and error correction within a single architecture. Building upon a theoretical analysis of perturbation propagation, we derive Lipschitz- and Jacobian-based criteria that determine whether a neural operator amplifies or attenuates localized faults. Guided by this theory, we propose a two-stage self-supervised training scheme: pre-training a multimodal convolutional autoencoder on clean data to preserve localized anomaly signals in the latent space, and expanding it with a learnable compute block composed of dense layers for correction and contrastive objectives for anomaly identification. Furthermore, we introduce layer-specific Lipschitz modulation and gradient clipping as principled mechanisms to control sensitivity across detection and correction modules. Experimental results on multimodal fault datasets demonstrate that the proposed approach improves both anomaly detection accuracy and reconstruction under sensor corruption. Overall, this framework bridges the gap between analytical robustness guarantees and practical fault-tolerant multimodal learning.