Abstract In recent years, significant progress has been made in image denoising, driven by advancements in modern deep neural networks such as Convolutional Neural Networks
(CNNs) and Transformers. However, existing denoising models often face limitations due to inherent local inductive biases or the quadratic computational complexity of their architectures. Recently, Selective Structured State Space Models, exemplified by Mamba, have shown considerable potential in modeling long-range dependencies with linear complexity. Despite this promise, their application in low-level computer vision tasks remains underexplored. In this study, we introduce a straightforward yet effective model, the Dual-Stream State-Space Module (DSSM), specifically designed for image denoising. The DSSM serves as the core component of our model, combining convolutional operations with the
Mamba model to enhance the capability of capturing both local features and global context.
This dual-stream approach leverages the inherent structure of local image patches while effectively modeling long-range dependencies, resulting in feature representations tailored for denoising tasks. Extensive experiments demonstrate the effectiveness of our method. For instance, on the SIDD dataset, our model achieves performance on par with state-of-the-art methods, maintaining a global receptive field while operating with similar computational costs.
Read full abstract