PatchDenoiser: Parameter-efficient multi-scale patch learning and fusion denoiser for medical images
AI 摘要
PatchDenoiser通过多尺度patch学习和融合,高效降噪医学图像,保留细节。
主要贡献
- 提出了一种轻量级的医学图像降噪框架PatchDenoiser
- 采用多尺度patch学习和空间感知融合策略
- 在实际医疗数据上验证了方法的有效性和高效性
方法论
PatchDenoiser将降噪分解为局部纹理提取和全局上下文聚合,并通过空间感知patch融合策略结合。
原文摘要
Medical images are essential for diagnosis, treatment planning, and research, but their quality is often degraded by noise from low-dose acquisition, patient motion, or scanner limitations, affecting both clinical interpretation and downstream analysis. Traditional filtering approaches often over-smooth and lose fine anatomical details, while deep learning methods, including CNNs, GANs, and transformers, may struggle to preserve such details or require large, computationally expensive models, limiting clinical practicality. We propose PatchDenoiser, a lightweight, energy-efficient multi-scale patch-based denoising framework. It decomposes denoising into local texture extraction and global context aggregation, fused via a spatially aware patch fusion strategy. This design enables effective noise suppression while preserving fine structural and anatomical details. PatchDenoiser is ultra-lightweight, with far fewer parameters and lower computational complexity than CNN-, GAN-, and transformer-based denoisers. On the 2016 Mayo Low-Dose CT dataset, PatchDenoiser consistently outperforms state-of-the-art CNN- and GAN-based methods in PSNR and SSIM. It is robust to variations in slice thickness, reconstruction kernels, and HU windows, generalizes across scanners without fine-tuning, and reduces parameters by ~9x and energy consumption per inference by ~27x compared with conventional CNN denoisers. PatchDenoiser thus provides a practical, scalable, and computationally efficient solution for medical image denoising, balancing performance, robustness, and clinical deployability.