Multimodal Learning 相关度: 8/10

Learning Transferable Sensor Models via Language-Informed Pretraining

Yuliang Chen, Arvind Pillai, Yu Yvonne Wu, Tess Z. Griffin, Lisa Marsch, Michael V. Heinz, Nicholas C. Jacobson, Andrew Campbell
arXiv: 2603.11950v1 发布: 2026-03-12 更新: 2026-03-12

AI 摘要

SLIP通过语言信息预训练传感器模型,提升跨领域零样本迁移能力,实现语义理解和生成推理。

主要贡献

  • 提出SLIP框架,用于学习语言对齐的传感器表示。
  • 结合对比对齐和传感器条件描述,提升判别理解和生成推理能力。
  • 支持不同时间分辨率和可变长度输入,无需额外训练。

方法论

利用预训练语言模型,通过交叉注意力机制和灵活的patch-embedder,实现传感器数据和语言之间的对齐。

原文摘要

Modern sensing systems generate large volumes of unlabeled multivariate time-series data. This abundance of unlabeled data makes self-supervised learning (SSL) a natural approach for learning transferable representations. However, most existing approaches are optimized for reconstruction or forecasting objectives and often fail to capture the semantic structure required for downstream classification and reasoning tasks. While recent sensor-language alignment methods improve semantic generalization through captioning and zero-shot transfer, they are limited to fixed sensor configurations, such as predefined channel sets, signal lengths, or temporal resolutions, which hinders cross-domain applicability. To address these gaps, we introduce \textbf{SLIP} (\textbf{S}ensor \textbf{L}anguage-\textbf{I}nformed \textbf{P}retraining), an open-source framework for learning language-aligned representations that generalize across diverse sensor setups. SLIP integrates contrastive alignment with sensor-conditioned captioning, facilitating both discriminative understanding and generative reasoning. By repurposing a pretrained decoder-only language model via cross-attention and introducing an elegant, flexible patch-embedder, SLIP supports different temporal resolutions and variable-length input at inference time without additional retraining. Across 11 datasets, SLIP demonstrates superior performance in zero-shot transfer, signal captioning, and question answering. It achieves a 77.14% average linear-probing accuracy, a 5.93% relative improvement over strong baselines, and reaches 64.83% accuracy in sensor-based question answering.

标签

传感器 自监督学习 语言模型 迁移学习

arXiv 分类

cs.AI cs.LG