Unified Restoration-Perception Learning: Maritime Infrared-Visible Image Fusion and Segmentation
AI 摘要
提出用于海事场景红外-可见光图像融合和分割的统一复原-感知学习框架。
主要贡献
- 构建红外-可见光海事船舶数据集(IVMSD)
- 提出多任务互补学习框架(MCLF)
- 设计频率-空间增强互补(FSEC)模块和语义-视觉一致性注意(SVCA)模块
方法论
构建IVMSD数据集,提出MCLF框架,结合FSEC和SVCA模块,实现图像复原、融合和语义分割的多任务协同学习。
原文摘要
Marine scene understanding and segmentation plays a vital role in maritime monitoring and navigation safety. However, prevalent factors like fog and strong reflections in maritime environments cause severe image degradation, significantly compromising the stability of semantic perception. Existing restoration and enhancement methods typically target specific degradations or focus solely on visual quality, lacking end-to-end collaborative mechanisms that simultaneously improve structural recovery and semantic effectiveness. Moreover, publicly available infrared-visible datasets are predominantly collected from urban scenes, failing to capture the authentic characteristics of coupled degradations in marine environments. To address these challenges, the Infrared-Visible Maritime Ship Dataset (IVMSD) is proposed to cover various maritime scenarios under diverse weather and illumination conditions. Building upon this dataset, a Multi-task Complementary Learning Framework (MCLF) is proposed to collaboratively perform image restoration, multimodal fusion, and semantic segmentation within a unified architecture. The framework includes a Frequency-Spatial Enhancement Complementary (FSEC) module for degradation suppression and structural enhancement, a Semantic-Visual Consistency Attention (SVCA) module for semantic-consistent guidance, and a cross-modality guided attention mechanism for selective fusion. Experimental results on IVMSD demonstrate that the proposed method achieves state-of-the-art segmentation performance, significantly enhancing robustness and perceptual quality under complex maritime conditions.