Search for a command to run...
The methods employed for image demosaicing and denoising play a pivotal role in image acquisition and restoration, and have been extensively studied over the past few decades. Traditionally, these tasks are performed sequentially, with demosaicing followed by denoising, or vice versa, treating each process independently. While this approach can enhance image quality, it often leads to issues such as color inaccuracies and information loss, as the outcome of the first task influences the second. Consequently, the integration of joint demosaicing and denoising (JDD) has become a focal point in recent research. Deep convolutional neural networks have shown promising results in addressing JDD challenges. This study introduces an end-to-end network, termed the Frequency-domain Features learning Network (FFNet), designed to tackle the JDD problem. Unlike conventional methods that focus on spatial domain features, FFNet utilizes frequency-domain (FD) characteristics to capture both global and local image details. Based on the vision Transformer architecture, FFNet consists of two key components: a global Fourier block (GFB), which uses global attention to determine the weights of FD parameters, and an MLP-based local Fourier block (LFB), which improves local feature extraction. These blocks are integrated with a channel attention mechanism to form the frequency-domain attention block (FAB), the core element of FFNet. Extensive experimental results on benchmark datasets demonstrate that FFNet achieves superior performance in terms of both quantitative metrics (PSNR/SSIM) and visual quality compared to existing state-of-the-art JDD methods. Furthermore, we provide a comprehensive analysis of its computational efficiency, including parameter count, FLOPs, and inference time, showing a competitive trade-off between performance and complexity.