Next Article in Journal
Thermal Image Processing for Respiratory Estimation from Cubical Data with Expandable Depth
Previous Article in Journal
Noninvasive Prediction of Sperm Retrieval Using Diffusion Tensor Imaging in Patients with Nonobstructive Azoospermia
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Dehazing with Recursive Gated Convolution in U-Net: A Novel Approach for Image Dehazing

1
Graduate School of Artificial Intelligence, Jeonju University, Jeonju-si 55069, Republic of Korea
2
Artificial Intelligence Research Center, Jeonju University, Jeonju-si 55069, Republic of Korea
*
Author to whom correspondence should be addressed.
J. Imaging 2023, 9(9), 183; https://doi.org/10.3390/jimaging9090183
Submission received: 16 August 2023 / Revised: 6 September 2023 / Accepted: 7 September 2023 / Published: 11 September 2023
(This article belongs to the Topic Computer Vision and Image Processing)

Abstract

:
Image dehazing, a fundamental problem in computer vision, involves the recovery of clear visual cues from images marred by haze. Over recent years, deploying deep learning paradigms has spurred significant strides in image dehazing tasks. However, many dehazing networks aim to enhance performance by adopting intricate network architectures, complicating training, inference, and deployment procedures. This study proposes an end-to-end U-Net dehazing network model with recursive gated convolution and attention mechanisms to improve performance while maintaining a lean network structure. In our approach, we leverage an improved recursive gated convolution mechanism to substitute the original U-Net’s convolution blocks with residual blocks and apply the SK fusion module to revamp the skip connection method. We designate this novel U-Net variant as the Dehaze Recursive Gated U-Net (DRGNet). Comprehensive testing across public datasets demonstrates the DRGNet’s superior performance in dehazing quality, detail retrieval, and objective evaluation metrics. Ablation studies further confirm the effectiveness of the key design elements.

1. Introduction

Image dehazing represents a key research theme in computer vision. Its primary goal is to retrieve the crispness and detail from images compromised by haze. Its relevance extends across various sectors, including remote sensing image processing, autonomous driving, video surveillance, and image restoration. Turbid substances such as haze, smoke, and water droplets in the atmosphere lead to light refraction, scattering, and absorption during transmission, causing image degradation, obscured details, contrast reduction, and color distortion. This degradation process is typically captured by the atmospheric scattering model [1,2,3]:
I ( x ) = J ( x ) t ( x ) + A ( 1 t ( x ) )
Here, I is the observed hazy image, J is the latent haze-free image, t is the medium transmission map, A is the global atmospheric light, and x represents the brightness or pixel value of a location in the haze image. In more detail, J ( x ) t ( x ) is called direct attenuation and describes the scene’s radiance and decay in the medium. A ( 1 t ( x ) ) is called airlight, results from previously scattered light, and leads to the shift of the scene color. If the atmosphere is homogeneous, then the transmission t can be represented by t x = e β d ( x ) , where β is the scattering coefficient of the atmosphere. It indicates that the scene radiance is attenuated exponentially with the scene depth d . Image dehazing methods aspire to yield the latent haze-free image J , using the input hazy image I directly or indirectly. It is important to note that Formula (1) represents a simplified expression of the Koschmieder model, suitable for small scattering approximation. This approximation applies to a low haze concentration, short light propagation distances, and isotropic scattering. While the simplified scattering model might overlook certain complexities in real-world scenarios, it can suffice for experimental and algorithmic validation purposes. Therefore, it finds widespread application in the field of image dehazing.
Traditional image dehazing methods, largely dependent on physical models and manually designed priors, often fall short in complex scenes. However, with the advent of deep learning, image dehazing methods rooted in deep neural networks have emerged as a focal research area. Early dehazing networks, primarily based on the atmospheric scattering model, calculated J by estimating A and t values separately. These methods, however, lacked effectiveness in restoring details and textures in intricate scenes. Current advanced dehazing networks deviate from the atmospheric scattering model, employing pairs of hazy and haze-free images for model training and restoring high-quality haze-free images by learning the mapping relationship between hazy and haze-free images. The significant advantage of this approach lies in its capacity to generate haze-free images directly from input hazy images, lessening the dependence on prior knowledge and strengthening model robustness. Despite the remarkable progress in dehazing networks, their growing complexity poses challenges as performance improves. To enhance the dehazing performance while maintaining network simplicity, we integrated minor modifications to the classic U-Net [4] architecture, resulting in a high-performance, lean, compact dehazing model: the Dehaze Recursive Gated U-Net (DRGNet). In particular, we used the classic U-Net as our base architecture, which possesses local residuals [5] and global residuals [6], thereby enabling the extraction of multi-scale image information. We established residual blocks using an improved recursive gated convolution mechanism [7], effectively infusing dynamic weights and high-order spatial interactions into the model, thereby improving feature extraction. We also improved the SK module [8] to embed a channel attention mechanism into the model, enabling the dynamic fusion of low- and high-level features. To validate the efficacy of the DRGNet, we engineered three variants of differing depths and executed trials using numerous prevalent image dehazing datasets. As depicted in Figure 1, the DRGNet delivers superior performance and comparatively lower computational overhead while preserving a lean structure. Additionally, we performed an array of ablation studies to explore the influence of specific design elements within DRGNet on the model’s dehazing outcome.
In summary, our work introduces a lean dehazing network: the DRGNet. Compared to current mainstream dehazing networks, the DRGNet offers an outstanding dehazing performance while requiring fewer parameters and incurring lower computational overhead. A simplified network can be easier to train, infer, and deploy. It also shows that we can still achieve better and better performance through careful design when not using complex architectures.

2. Related Works

We can categorize mainstream image dehazing methods into four groups.: methods based on image enhancement, methods based on prior knowledge, methods based on fusion, and methods based on deep learning. In recent years, many research works have used multiple methods simultaneously to enhance performance.
Image enhancement-based methods usually do not use the physical model of haze generation and do not pay much attention to image quality. They use image enhancement technology to increase the visual effect of foggy images, thereby highlighting specific details. Standard image enhancement methods include histogram equalization [9,10], Gamma correction [11,12,13,14], multi-scale retinex [15], white balance method [16], median filtering [17,18], and others. However, dehazing methods based only on image enhancement usually face problems such as pixel oversaturation.
Image fusion-based methods aim to select the best regions from multiple images to synthesize high-quality images. Among these methods, the method proposed in [19] is representative. This method first uses gamma correction coefficients to generate a set of front-exposed images, and then pixel-wise weight maps are constructed by analyzing both global and local exposedness to guide the fusion process. This method is very computationally efficient and generates haze-free images. The color quality is good and the details are precise. In addition, classic image fusion-based dehazing methods also include [20,21], but they have high computational complexity.
Methods based on prior knowledge generally employ physical laws or manually crafted rules to tackle image dehazing. Among them, the Dark Channel Prior (DCP) method [22] stands out as the epitome. The DCP method, premised on physical assumptions, mainly targets outdoor hazy images. Guided by the atmospheric scattering model, it exploits the dark channels within an image to estimate atmospheric light and transmittance, subsequently restoring the haze-free image. Other notable prior knowledge-based methods encompass the Color Attenuation Prior method [23] and the Edge-Preserving Decomposition-Based method [24]. These methods are lauded for their robust interpretability, efficiency, and simplicity, offering promising dehazing results in specific settings. However, they need more adaptability, resulting in a subpar performance in complex environments. These classic prior knowledge-based methodologies have played a crucial role in early image dehazing research, setting the foundation and inspiration for ensuing studies.
On the other hand, learning-based methods utilize deep learning techniques for dehazing tasks, emerging as a research hotspot in recent years. Based on their dependence, these methods can be further divided into two subcategories: those reliant on the atmospheric scattering model and those that are not. Early dehazing networks [25,26,27] typically rested on the atmospheric scattering model, accepting hazy images as inputs, outputting medium transmission maps or global atmospheric light, and restoring latent haze-free images per the atmospheric scattering model. However, in contrast to prior knowledge-based methods, these early dehazing networks needed to exhibit substantial advancements in principles or performance.
As deep learning has matured and our understanding of image dehazing has become more nuanced, the most advanced dehazing networks have veered away from rigid physical assumptions and manually designed rules. Instead, these networks now employ a data-driven strategy, generating haze-free images directly from their hazy counterparts [28,29,30,31,32,33,34,35,36,37]. Depending on the architecture, commonly used dehazing networks, which function independently of the atmospheric scattering model, encompass encoder–decoder networks, GAN-based networks, attention-based networks, and Transformer-based networks. These advanced networks demonstrate superior generalization capabilities and robustness compared to their early predecessors. As a result, they deliver a superior performance in complex scenarios, such as those involving dense and uneven haze, thus asserting themselves as the prevailing approach. Despite their advancements, intricate network architectures pose their challenges. They enhance network performance and significantly escalate the network’s complexity, making training, inference, and deployment more challenging. Balancing performance enhancement and complexity is an ongoing area of investigation in dehazing networks, and this is also the problem we intend to solve in this paper. In addition to the methods based on supervised learning introduced above, some researchers have proposed defogging methods based on unsupervised and semi-supervised methods. Ref. [38] proposed a method to continue unsupervised learning, solely using real-world outdoor images and tuning the network’s parameters by directly minimizing the DCP. Ref. [39] proposes a semi-supervised learning network: SAD-Net. SAD-Net utilizes both synthetic datasets and natural hazy images for training and uses an attention mechanism to increase dehazing performance. These methods can reduce the dependence of the dehazing network on the data set and have a more robust generalization.

3. Methods

The Dehaze Recursive Gated U-Net (DRGNet) is a 7-stage adaptation of the U-Net, depicted in Figure 2. Each stage of the DRGNet comprises one or more stacked Recursive Gated Convolution Blocks (RG-Conv Blocks), which employ the recursive gated convolution. We recognized the limitations of the original recursive gated convolution for image dehazing tasks. As such, we implemented several enhancements. We also introduced SK fusion modules, derived from the SK module, and can dynamically fuse features from the encoder and decoder while introducing an attention mechanism to the model.

3.1. Motivation

With the development of deep learning technology and the deepening of people’s understanding of image dehazing tasks, dehazing models based on deep learning are performing better and better in image dehazing tasks. However, as illustrated in Figure 1, we have observed that these high-performing dehazing models often come with complex architectures, numerous parameters, and significant computational overhead. Such complexity in models is detrimental to training, inference, and deployment. Therefore, a lean model with excellent dehazing performance is necessary and holds significant research value.
Owing to the haze effect precipitated by atmospheric scattering, the transmission of light is disrupted by airborne particulates, causing specific pixel values in the image to darken. Channels vary in their response to light scattering and absorption, resulting in differential darkening effects. A typical method in the Prior-based approach is the DCP method, positing differences in pixel value distribution across channels in a hazy image window. To some extent, the DCP method’s success verifies the effectiveness of channel-by-channel feature extraction. Recursive gated convolution operates by executing gated convolutions recursively on the input information’s channel dimension to extract features. Intuitively, superior to traditional convolution, recursive gated convolution more efficiently merges features from different channels, thereby enhancing image dehazing efficiency. The direct application of original recursive gated convolution to image dehazing is challenging, leading to our adaptation of it to cater to image dehazing task specifics. U-Net is an encoder–decoder convolutional neural network design characterized by simplicity, fewer parameters, fast inference speed, and proficiency in image detail information. Combining recursive gated convolution with U-Net architecture offers a lean, high-performance image dehazing network. Attention mechanisms have gained significant attention in recent years. Our study has revealed that by making subtle modifications to the SK module, it is possible to introduce an attention mechanism to the model while dynamically fusing feature maps from different branches.

3.2. Recursive Gated Convolution Block

We propose that the Recursive Gated Convolution Block (RG-Conv Block) is fundamentally rooted in the recursive gated convolution mechanism. Unlike conventional convolution blocks, the RG-Conv Block improves model performance by explicitly modeling high-order interaction mechanisms. Figure 3 depicts the structure of an RG-Conv block with a second-order interaction. Depending on the circumstances, we can easily adjust the value of the order, thereby improving the model’s capability.
Let’s define X R H W × C as the input feature map and C represents the channel number of X . Initially, we leverage a projection layer to effect a dimensional transformation on X . Subsequently, the projection layer’s output is divided by channel dimension, leading to the generation of a series of features:
P 0 H W × C 0 , D 0 H W × C 0 , D 1 H W × C 1 , , D n 1 H W × C n 1 = P r o j X R H W × 2 C
In Equation (2), n symbolizes the value of the order while the channel number C k of each feature adheres to the following conditions: C 0 + k = 0 n 1 C k = 2 C and C k = C / 2 n k 1 .
We undertake gated convolution [40] in a recursive style. For an RG-Conv block featuring n -order interaction, the execution of gated convolution is necessary n times. We assume that the output of k -th gated convolution is P k . The computation process of P k + 1 is in line with the Equation (3):
P k + 1 = ϕ k ( S i g m o i d ( g k ( P k ) ) f k ( D k ) ) R H W × C k + 1
In this context, f k stands for Depth-wise Convolution [41] and ⊙ represents element-wise multiplication, with g k and ϕ k being piecewise functions. In Equation (4), L i n e a r C k 1 , C k denotes the mapping of the channel dimension of the feature vector from C k 1 to C k . Equation (5) is the same.
g k = I d e n t i t y ,   k = 0 L i n e a r C k 1 , C k ,   k > 0
ϕ k = I d e n t i t y ,   k + 1 < n L i n e a r C k , C ,   k = n 1
It is worth mentioning that we performed pertinent ablation experiments in the experimental section. The findings demonstrated that bounded functions such as the Sigmoid function [42] and the Hard-Sigmoid function [43] could effectively circumvent gradient explosion. On the contrary, eliminating the activation function or utilizing unbounded functions such as the ReLU function [44] leads to gradient explosion, while employing the Tanh activation function could also trigger gradient explosion. This likelihood escalates with increasing order n .

3.3. SK Fusion Module

The SK module is an attention module that can significantly improve the performance of the model when applied to basic computer vision tasks [45,46]. The SK fusion module we used is a variation of the SK module.
In DRG-Net, as depicted in Figure 2, there are skip connections between the encoding and decoding stages, and the features of the encoding and decoding stages are fused to help the network better restore the detailed information. We use the SK fusion module to fuse features from different stages dynamically. For the feature map f 1 R c 1 × h w from the encoding stage and the feature map f 2 R c 2 × h w from the decoding stage, their process of dynamic fusion by the SK fusion module is shown in Figure 4.
For feature map f 1 , it undergoes a linear function g f 1 = L i n e a r c 1 , c 2 , which matches its channel dimensions with f 2 . Subsequently, g f 1 and f 2 are fused through element-wise addition to generate a feature map denoted as s . Following this, global average pooling [47] is employed to generate channel-wise statistics, multi-layer perceptrons and the softmax function are then used to obtain channel attention for the feature map s , and the channel dimension separates the result to procure the fusion weights of separate branches. The above process can be described by Equation (6).
w 1 , w 2 = S p l i t S o f t m a x M L P G A P g f 1 + f 2
Lastly, feature maps from distinct branches are added to the fusion weights in a weighted manner to produce the final output of the SK fusion module:
O u t = w 1 g f 1 + w 2 f 2

3.4. Training Loss

In the training process of DRGNet, we elected to use the L1 loss function. The principal distinction between the L1 and L2 loss (MSE loss) functions lies in their handling of errors. The L2 loss function penalizes more significant errors more severely (as errors are squared), making the model inclined to minimize these substantial errors. However, this may also make the model excessively sensitive to noise or outliers. In image dehazing, such sensitivity could lead the model to overcorrect hazy effects, resulting in the loss or over-sharpening of image details. In contrast, the L1 loss function yields superior results in image dehazing tasks. In the ablation experiments in the subsequent chapters, we also verified this point of view.

3.5. Network Architecture Details

DRGNet is a 7-stage derivative of the U-Net blueprint. To simplify the model and ensure performance and stability during training, DRGNet’s architecture details are as follows: each stage accommodates RG-Conv blocks at a quantity ratio of {1:1:1:2:1:1:1}. RG-Blocks in the same stage should have the same interaction order, and the interaction orders of RG-Block in different stages are: {1, 2, 2, 3, 2, 2, 1}, while the kernel size of DW-Conv within each RG-Block is 5. To rigorously evaluate the dehazing capabilities of DRGNet, we engineered three depth-differentiated variants: DRGNet-T, DRGNet-B, and DRGNET-L. Table 1 delineates the architectural details of each variant. In subsequent ablation experiments we will verify the rationality of this design.

4. Experimental

4.1. Data Set and Experimental Setup

Experiments were conducted on the RESIDE [48] and Haze-4K [49] datasets to validate our method. RESIDE, with its diverse data sources and image content, is subdivided into ITS (Indoor Training Set), OTS (Outdoor Training Set), and SOTS (Synthetic Objective Testing Set). We utilized ITS (13,990 image pairs) and OTS (313,950 image pairs) to train the dehazing network and the indoor and outdoor scenes from the SOTS dataset (500 image pairs each) for testing. The Haze4K dataset, a synthetic dataset, includes 4000 paired images from both indoor and outdoor scenes. A total of 3000 pairs were employed for training and the remaining 1000 for testing.
We employed the PyTorch framework for coding and ran the training and testing on an A100 (80 G) graphics card. During training, the input image dimension was set to 256 × 256. The mini-batch size for variants of DRGNet at varying depths was 32, with 1000 epochs of training. We found that the size of the learning rate significantly influences training stability; hence, we opted for smaller learning rates for deeper variants. For the T, B, and L variants, according to the linear scaling rule [50], we established learning rates of {12 × 10−4, 8 × 10−4, 4 × 10−4}, respectively. We incrementally elevated the learning rate using a warmup strategy for 50 epochs. Subsequently, the cosine decay strategy [51] is employed to gradually reduce the learning rate to 1/100 of the initial value. In the training process, we utilized the AdamW optimizer [52].

4.2. Quantitative Comparison and Qualitative Analysis

To corroborate the efficacy of our proposed method, we executed a quantitative performance comparison between the DRGNet and the baseline methods. The results are in Table 2. In this quantitative comparison, the DRGNet performed well on the extensive RESIDE and more concise Haze4K datasets. Compared to existing methods, the DRGNet strikes a balance between performance and model complexity. For example, when compared with the classic dehazing network FFA-Net, the DRGNet-T only used about 1% of the MACs and around 20% of the parameters, and improved PSNR [53] and SSIM [53] by 2.47 and 0.005, respectively. Furthermore, when contrasted with the more advanced MixDehazeNet-S, the DRGNet-T demonstrated a comparable dehazing performance while reducing the MACs and parameters by approximately 76% and 47%, respectively.
Figure 5 showcases the visual dehazing outcomes of the DRGNet-T compared with other methods. Our method outshines existing methods in terms of dehazing quality. The DRGNet mitigates haze influence, resulting in amplified clarity, enhanced contrast, and restored color images. It proficiently retains details and textures camouflaged by the haze, ensuring aesthetically pleasing outputs. Moreover, our network performs robustly when processing hazy images from various scenes, consistently delivering reliable dehazing results. It efficiently mitigates haze while suppressing artifacts, representing an optimal compromise between these two aspects of image restoration.

4.3. Ablation Study

To analyze the model’s key components, we conducted ablation studies. Table 3 shows the results of ablation experiments, with Figure 6 showing the training process of the DRGNet-T in selected ablation experiments.
In the DRGNet, we used a series of recursive gated convolution blocks with different orders to facilitate feature extraction. When the order was 1, the recursive gated convolution block was degraded to a standard gated convolution block. Our ablation studies showed performance differences between these two mechanisms, suggesting that the recursive gated convolution mechanism significantly improves the model’s performance. Furthermore, we examined the activation function design within the recursive gated convolution blocks. We incorporated a sigmoid function into the traditional recursive gated convolution mechanism to prevent gradient explosion. Subsequent ablation studies confirmed the effectiveness of our approach: on removing the bounded function or resorting to unbounded functions such as ReLU, NaN values predictably emerged during the training phase. Lastly, we explored the influence of order. The experimental results show that the order can affect the model’s performance and efficiency. If the order is too small, it will affect the model’s performance. If the order is too large, it will become difficult to train and infer the model. Therefore, deploying blocks in different orders across stages at different depths is a feasible strategy, which we adopt.
It is crucial to highlight that the RG-Conv block introduces a notably significant increase in latency. We conjecture that this could stem from the recursive gating mechanism, which engages DW-Conv at each recursion step. Conventional deep learning frameworks often enhance computational efficiency through optimization libraries (e.g., cuDNN) and hardware acceleration (e.g., GPUs). Nonetheless, these optimizations might not seamlessly support DW-Conv in specific scenarios, resulting in suboptimal efficiency.
The original SK Module boasts a lightweight nature, seamlessly integrating without imposing substantial additional computational costs. Conventional U-Net models commonly resort to concatenation fusion for amalgamating features sourced from diverse branches. However, within the DRGNet, we utilized the SK Fusion module, resulting in a conspicuous elevation in the model’s dehazing prowess. The improvement may be attributed to the following factors: (1) The SK fusion module introduces an attention mechanism to the model. (2) Unlike concatenation fusion, the SK fusion module can dynamically fuse features from different branches.
The kernel size refers to the convolution kernel size in the DW-Conv of the recursive gated convolution block. Our experiments revealed that kernel size 5 is an optimal choice for the DRGNet-T. However, any increase or decrease in kernel size led to a significant drop in performance.
The experimental results show that compared with the L2 loss function, the improvement brought about by using the L1 loss function is significant. Simultaneously, the simplicity of the L1 loss function also reduces the computational cost and latency of the model.
Irrespective of whether the depth or width of the DRGNet-T is augmented, a significant enhancement in the model’s performance ensues. This observation underscores the robust scalability inherent in the DRGNet. It is noteworthy that, when compared to broader networks, deeper networks exhibit an enhanced performance. However, this advantage comes at the expense of increased parameters and computational demands. Simultaneously, as the network depth expands, the count of RG-Conv blocks also rises, resulting in a substantial escalation of latency in deeper networks.

5. Discussion

This paper introduces an efficient dehazing network: the DRGNet. We use the U-Net with local and global residuals as the basic architecture to extract multi-scale information of features. To enhance the extraction of features from various channels in hazy images, we have improved the recursive gated convolution mechanism and created RG-Conv blocks using it.
Simultaneously, we have leveraged SK fusion to supersede the conventional cascade fusion. This innovation empowers the model’s fusion layer to dynamically integrate feature maps from distinct branches while incorporating an attention mechanism into the model’s framework.
We have presented three distinct variants of the DRGNet, varying in depth, and evaluated their performance across multiple publicly available datasets. The test results unequivocally showcase the exceptional capabilities of the DRGNet. We conducted extensive tests to confirm the effectiveness of our proposed designs. This will provide valuable guidance for future researchers to analyze the key designs that enhance the performance in dehazing networks.

Author Contributions

Conceptualization, Z.W.; Data curation, P.L.; Formal analysis, J.M.; Funding acquisition, J.M.; Investigation, J.J.; Methodology, Z.W.; Project administration, Z.W.; Resources, J.J.; Software, Z.W.; Supervision, J.M.; Validation, J.M.; Visualization, J.J.; Writing—original draft, Z.W.; Writing—review and editing, J.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Experimental data can be found in the following link: https://github.com/AquaLaker/Efficient-Dehazing-with-Recursive-Gated-Convolution-in-U-Net-A-Novel-Approach-for-Image-Dehazing (accessed on 6 September 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. McCartney, E.J. Optics of the Atmosphere: Scattering by Molecules and Particles; John Wiley and Sons: New York, NY, USA, 1976. [Google Scholar]
  2. Nayar, S.K.; Narasimhan, S.G. Vision in Bad Weather. In Proceedings of the 7th IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; IEEE: Piscataway, NJ, USA, 1999; Volume 2, pp. 820–827. [Google Scholar]
  3. Narasimhan, S.G.; Nayar, S.K. Vision and the Atmosphere. Int. J. Comput. Vis. 2002, 48, 233–254. [Google Scholar] [CrossRef]
  4. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  5. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  6. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian Denoiser: Residual Learning of Deep Cnn for Image Denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [PubMed]
  7. Rao, Y.; Zhao, W.; Tang, Y.; Zhou, J.; Lim, S.N.; Lu, J. Hornet: Efficient High-Order Spatial Interactions with Recursive Gated Convolutions. Adv. Neural Inf. Process. Syst. 2022, 35, 10353–10366. [Google Scholar]
  8. Li, X.; Wang, W.; Hu, X.; Yang, J. Selective Kernel Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 510–519. [Google Scholar]
  9. Galdran, A. Image Dehazing by Artificial Multiple-Exposure Image Fusion. Signal Process. 2018, 149, 135–147. [Google Scholar] [CrossRef]
  10. Fu, X.; Wang, J.; Zeng, D.; Huang, Y.; Ding, X. Remote Sensing Image Enhancement Using Regularized-Histogram Equalization and DCT. IEEE Geosci. Remote Sens. Lett. 2015, 12, 2301–2305. [Google Scholar] [CrossRef]
  11. Mi, Z.; Zhou, H.; Zheng, Y.; Wang, M. Single Image Dehazing via Multi-scale Gradient Domain Contrast Enhancement. IET Image Process. 2016, 10, 206–214. [Google Scholar] [CrossRef]
  12. Zheng, L.; Shi, H.; Gu, M. Infrared Traffic Image Enhancement Algorithm Based on Dark Channel Prior and Gamma Correction. Mod. Phys. Lett. B 2017, 31, 1740044. [Google Scholar] [CrossRef]
  13. Gao, Y.; Chen, H.; Li, H.; Zhang, W. Single Image Dehazing Using Local Linear Fusion. IET Image Process. 2018, 12, 637–643. [Google Scholar] [CrossRef]
  14. Ju, M.; Ding, C.; Zhang, D.; Guo, Y.J. Gamma-Correction-Based Visibility Restoration for Single Hazy Images. IEEE Signal Process. Lett. 2018, 25, 1084–1088. [Google Scholar] [CrossRef]
  15. Wang, J.; Lu, K.; Xue, J.; He, N.; Shao, L. Single Image Dehazing Based on the Physical Model and MSRCR Algorithm. IEEE Trans. Circuits Syst. Video Technol. 2017, 28, 2190–2199. [Google Scholar] [CrossRef]
  16. Liu, C.; Zhao, J.; Shen, Y.; Zhou, Y.; Wang, X.; Ouyang, Y. Texture Filtering Based Physically Plausible Image Dehazing. Vis. Comput. 2016, 32, 911–920. [Google Scholar] [CrossRef]
  17. Tarel, J.-P.; Hautiere, N.; Cord, A.; Gruyer, D.; Halmaoui, H. Improved Visibility of Road Scene Images under Heterogeneous Fog. In Proceedings of the 2010 IEEE Intelligent Vehicles Symposium, La Jolla, CA, USA, 21–24 June 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 478–485. [Google Scholar]
  18. Tarel, J.-P.; Hautiere, N. Fast Visibility Restoration from a Single Color or Gray Level Image. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 27 September–4 October 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 2201–2208. [Google Scholar]
  19. Zhu, Z.; Wei, H.; Hu, G.; Li, Y.; Qi, G.; Mazur, N. A Novel Fast Single Image Dehazing Algorithm Based on Artificial Multiexposure Image Fusion. IEEE Trans. Instrum. Meas. 2020, 70, 1–23. [Google Scholar] [CrossRef]
  20. Yuan, F.; Zhou, Y.; Xia, X.; Shi, J.; Fang, Y.; Qian, X. Image Dehazing Based on a Transmission Fusion Strategy by Automatic Image Matting. Comput. Vis. Image Underst. 2020, 194, 102933. [Google Scholar] [CrossRef]
  21. Gao, Y.; Su, Y.; Li, Q.; Li, H.; Li, J. Single Image Dehazing via Self-Constructing Image Fusion. Signal Process. 2020, 167, 107284. [Google Scholar] [CrossRef]
  22. He, K.; Sun, J.; Tang, X. Single Image Haze Removal Using Dark Channel Prior. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 2341–2353. [Google Scholar]
  23. Zhu, Q.; Mai, J.; Shao, L. A Fast Single Image Haze Removal Algorithm Using Color Attenuation Prior. IEEE Trans. Image Process. 2015, 24, 3522–3533. [Google Scholar]
  24. Li, Z.; Zheng, J. Edge-Preserving Decomposition-Based Single Image Haze Removal. IEEE Trans. Image Process. 2015, 24, 5432–5441. [Google Scholar] [CrossRef]
  25. Cai, B.; Xu, X.; Jia, K.; Qing, C.; Tao, D. Dehazenet: An End-to-End System for Single Image Haze Removal. IEEE Trans. Image Process. 2016, 25, 5187–5198. [Google Scholar] [CrossRef]
  26. Ren, W.; Liu, S.; Zhang, H.; Pan, J.; Cao, X.; Yang, M.-H. Single Image Dehazing via Multi-Scale Convolutional Neural Networks. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part II 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 154–169. [Google Scholar]
  27. Li, B.; Peng, X.; Wang, Z.; Xu, J.; Feng, D. Aod-Net: All-in-One Dehazing Network. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4770–4778. [Google Scholar]
  28. Chen, D.; He, M.; Fan, Q.; Liao, J.; Zhang, L.; Hou, D.; Yuan, L.; Hua, G. Gated Context Aggregation Network for Image Dehazing and Deraining. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa Village, HI, USA, 7–11 January 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1375–1383. [Google Scholar]
  29. Liu, X.; Ma, Y.; Shi, Z.; Chen, J. Griddehazenet: Attention-Based Multi-Scale Network for Image Dehazing. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7314–7323. [Google Scholar]
  30. Dong, H.; Pan, J.; Xiang, L.; Hu, Z.; Zhang, X.; Wang, F.; Yang, M.-H. Multi-Scale Boosted Dehazing Network with Dense Feature Fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2157–2167. [Google Scholar]
  31. Dong, J.; Pan, J. Physics-Based Feature Dehazing Networks. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XXX 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 188–204. [Google Scholar]
  32. Qin, X.; Wang, Z.; Bai, Y.; Xie, X.; Jia, H. FFA-Net: Feature Fusion Attention Network for Single Image Dehazing. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 11908–11915. [Google Scholar]
  33. Ye, T.; Jiang, M.; Zhang, Y.; Chen, L.; Chen, E.; Chen, P.; Lu, Z. Perceiving and Modeling Density Is All You Need for Image Dehazing. arXiv 2021, arXiv:2111.09733. [Google Scholar]
  34. Hong, M.; Liu, J.; Li, C.; Qu, Y. Uncertainty-Driven Dehazing Network. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 22 February–1 March 2022; Volume 36, pp. 906–913. [Google Scholar]
  35. Song, Y.; Zhou, Y.; Qian, H.; Du, X. Rethinking Performance Gains in Image Dehazing Networks. arXiv 2022, arXiv:2209.11448. [Google Scholar]
  36. Lu, L.; Xiong, Q.; Chu, D.; Xu, B. MixDehazeNet: Mix Structure Block for Image Dehazing Network. arXiv 2023, arXiv:2305.176542023. [Google Scholar]
  37. Song, Y.; He, Z.; Qian, H.; Du, X. Vision Transformers for Single Image Dehazing. IEEE Trans. Image Process. 2023, 32, 1927–1941. [Google Scholar] [CrossRef] [PubMed]
  38. Golts, A.; Freedman, D.; Elad, M. Unsupervised Single Image Dehazing Using Dark Channel Prior Loss. IEEE Trans. Image Process. 2019, 29, 2692–2701. [Google Scholar] [CrossRef] [PubMed]
  39. Sun, Z.; Zhang, Y.; Bao, F.; Wang, P.; Yao, X.; Zhang, C. Sadnet: Semi-Supervised Single Image Dehazing Method Based on an Attention Mechanism. ACM Trans. Multimed. Comput. Commun. Appl. 2022, 18, 1–23. [Google Scholar] [CrossRef]
  40. Dauphin, Y.N.; Fan, A.; Auli, M.; Grangier, D. Language Modeling with Gated Convolutional Networks. In Proceedings of the International Conference on Machine Learning; PMLR, Sydney, Australia, 6–11 August 2017; pp. 933–941. [Google Scholar]
  41. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  42. Cybenko, G. Approximation by Superpositions of a Sigmoidal Function. Math. Control Signals Syst. 1989, 2, 303–314. [Google Scholar] [CrossRef]
  43. Nwankpa, C.; Ijomah, W.; Gachagan, A.; Marshall, S. Activation Functions: Comparison of Trends in Practice and Research for Deep Learning. arXiv 2018, arXiv:1811.03378. [Google Scholar]
  44. Nair, V.; Hinton, G.E. Rectified Linear Units Improve Restricted Boltzmann Machines. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), Haifa, Israel, 21–24 June 2010; pp. 807–814. [Google Scholar]
  45. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.-H.; Shao, L. Learning Enriched Features for Fast Image Restoration and Enhancement. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 1934–1948. [Google Scholar] [CrossRef]
  46. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.-H.; Shao, L. Learning Enriched Features for Real Image Restoration and Enhancement. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XXV 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 492–511. [Google Scholar]
  47. Lin, M.; Chen, Q.; Yan, S. Network in Network. arXiv 2013, arXiv:1312.4400. [Google Scholar]
  48. Li, B.; Ren, W.; Fu, D.; Tao, D.; Feng, D.; Zeng, W.; Wang, Z. Benchmarking Single-Image Dehazing and Beyond. IEEE Trans. Image Process. 2018, 28, 492–505. [Google Scholar] [CrossRef]
  49. Liu, Y.; Zhu, L.; Pei, S.; Fu, H.; Qin, J.; Zhang, Q.; Wan, L.; Feng, W. From Synthetic to Real: Image Dehazing Collaborating with Unlabeled Real Data. In Proceedings of the 29th ACM International Conference on Multimedia, Online, 20–24 October 2021; pp. 50–58. [Google Scholar]
  50. Goyal, P.; Dollár, P.; Girshick, R.; Noordhuis, P.; Wesolowski, L.; Kyrola, A.; Tulloch, A.; Jia, Y.; He, K. Accurate, Large Minibatch Sgd: Training Imagenet in 1 Hour. arXiv 2017, arXiv:1706.02677. [Google Scholar]
  51. Loshchilov, I.; Hutter, F. Sgdr: Stochastic Gradient Descent with Warm Restarts. arXiv 2016, arXiv:1608.03983. [Google Scholar]
  52. Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar]
  53. Hore, A.; Ziou, D. Image Quality Metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 2366–2369. [Google Scholar]
  54. Tu, Z.; Talebi, H.; Zhang, H.; Yang, F.; Milanfar, P.; Bovik, A.; Li, Y. Maxim: Multi-Axis Mlp for Image Processing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5769–5780. [Google Scholar]
  55. Nie, J.; Pang, Y.; Xie, J.; Han, J.; Li, X. Binocular Image Dehazing via a Plain Network without Disparity Estimation. IEEE Trans. Multimed. 2022; in press. [Google Scholar]
  56. Luo, P.; Xiao, G.; Gao, X.; Wu, S. LKD-Net: Large Kernel Convolution Network for Single Image Dehazing. In Proceedings of the 2023 IEEE International Conference on Multimedia and Expo (ICME), Brisbane, Australia, 10–14 July 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1601–1606. [Google Scholar]
  57. Chen, Z.; He, Z.; Lu, Z.-M. DEA-Net: Single Image Dehazing Based on Detail-Enhanced Convolution and Content-Guided Attention. arXiv 2023, arXiv:2301.04805. [Google Scholar]
Figure 1. Comparison of DRGNet with Alternative Approaches on the SOTS Indoor Dataset. In this graph, we employ Multiply–Accumulate Operations (MACs) as a metric for quantifying computational load and efficiency across different models. Additionally, we utilize the Peak Signal-to-Noise Ratio (PSNR) as a benchmark to assess the dehazing performance of each model. The size of the data points in our visual representation corresponds to the parameter count of the respective models.
Figure 1. Comparison of DRGNet with Alternative Approaches on the SOTS Indoor Dataset. In this graph, we employ Multiply–Accumulate Operations (MACs) as a metric for quantifying computational load and efficiency across different models. Additionally, we utilize the Peak Signal-to-Noise Ratio (PSNR) as a benchmark to assess the dehazing performance of each model. The size of the data points in our visual representation corresponds to the parameter count of the respective models.
Jimaging 09 00183 g001
Figure 2. The overall architecture of DRGNet.
Figure 2. The overall architecture of DRGNet.
Jimaging 09 00183 g002
Figure 3. Recursive Gated Convolution Block with second-order interactions.
Figure 3. Recursive Gated Convolution Block with second-order interactions.
Jimaging 09 00183 g003
Figure 4. The process of SK Fusion Module fuses various branches. R represents the dimension information of different feature vectors.
Figure 4. The process of SK Fusion Module fuses various branches. R represents the dimension information of different feature vectors.
Jimaging 09 00183 g004
Figure 5. Qualitative comparisons on RESIDE-IN and RESODE-OUT datasets. The first two rows are indoor images, the last two rows are outdoor images, and the last column represents the corresponding ground truth.
Figure 5. Qualitative comparisons on RESIDE-IN and RESODE-OUT datasets. The first two rows are indoor images, the last two rows are outdoor images, and the last column represents the corresponding ground truth.
Jimaging 09 00183 g005
Figure 6. Ablation Study: Training DRGNet-T on SOTS Indoor Dataset. The X-axis represents the number of training iterations, and the Y-axis represents the dehazing performance of the model (we use the Peak Signal-to-Noise Ratio (PSNR) to measure if PSNR = 0 indicates that Nan appeared during the training process). For better visualization, we calculate the average every twenty points and plot it.
Figure 6. Ablation Study: Training DRGNet-T on SOTS Indoor Dataset. The X-axis represents the number of training iterations, and the Y-axis represents the dehazing performance of the model (we use the Peak Signal-to-Noise Ratio (PSNR) to measure if PSNR = 0 indicates that Nan appeared during the training process). For better visualization, we calculate the average every twenty points and plot it.
Jimaging 09 00183 g006
Table 1. Architecture details of DRGNet with varying depths. “Depth” indicates the number of blocks per stage, “Order” indicates the interaction order of blocks per stage.
Table 1. Architecture details of DRGNet with varying depths. “Depth” indicates the number of blocks per stage, “Order” indicates the interaction order of blocks per stage.
NameDepthOrder
DRGNet-T{2, 2, 2, 4, 2, 2, 2}{1, 2, 2, 3, 2, 2, 1}
DRGNet-B{4, 4, 4, 8, 4, 4, 4}{1, 2, 2, 3, 2, 2, 1}
DRGNet-L{8, 8, 8, 16, 8, 8, 8}{1, 2, 2, 3, 2, 2, 1}
Table 2. Benchmarking Dehazing Techniques on Three Datasets. We use a single A100 (80 G) graphics card to train our dehazing network. For a fair comparison, we used a single RTX3090 graphics card to test the overhead of DRGNet. Data for other methods in the table are taken from their respective papers. ‘-’ indicates that there are no such data in the original paper.
Table 2. Benchmarking Dehazing Techniques on Three Datasets. We use a single A100 (80 G) graphics card to train our dehazing network. For a fair comparison, we used a single RTX3090 graphics card to test the overhead of DRGNet. Data for other methods in the table are taken from their respective papers. ‘-’ indicates that there are no such data in the original paper.
ModelRESIDE-INRESIDE-OUTHaze4KOverhead
PSNRSSIMPSNRSSIMPSNRSSIMMACsParamLatency
DCP [22]16.620.81819.130.81514.010.760---
DehazeNet [25]19.820.82124.750.92719.120.8400.581 G0.009 M0.919 ms
MSCNN [26]19.840.83322.060.90814.010.5100.525 G0.008 M0.619 ms
AOD-Net [27]20.510.81624.140.92017.150.8300.115 G0.002 M0.390 ms
GCANet [28]30.230.980--25.090.92318.41 G0.702 M3.695 ms
GridDehazeNet [29]32.160.98430.860.982--21.49 G0.956 M9.905 ms
MSBDN [30]33.670.98533.480.98222.990.85041.54 G31.35 M13.250 ms
PFDN [31]32.680.976----50.46 G11.27 M4.809 ms
FFA-Net [32]36.390.98933.570.98426.960.950287.8 G4.456 M55.91 ms
PMNet [33]38.410.99034.740.985--81.13 G18.90 M28.08 ms
UDN [34]38.620.99134.920.987---4.25 M-
gUNet-T [35]37.990.99334.520.98331.600.9842.595 G0.805 M3.391 ms
MixDehazeNet-S [36]39.470.99535.090.985--22.06 G3.16 M14.56 ms
DehazeFormer-T [37]35.150.98933.710.982--6.658 G0.686 M10.59 ms
MAXIM [54]38.110.99134.190.985--216 G14.10 M-
SGID-PFF [55]38.520.99130.200.975--152.80 G13.87 M20.92 ms
LKD-B [56]38.570.99334.810.983--12.20 G1.22 M-
DEA-Net [57]40.200.99336.030.98933.190.9932.23 G3.653 M7.093 ms
DRGNet-T38.860.99434.810.98332.420.9862.907 G0.939 M7.57 ms
DRGNet-B39.820.99535.320.98432.890.9875.207 G1.675 M13.70 ms
DRGNet-L40.760.99636.330.98633.210.9889.803 G3.146 M25.77 ms
Table 3. Ablation study of network architectures.
Table 3. Ablation study of network architectures.
MethodsRESIDE-INOverhead
PSNRSSIMMACsParamLatency
Baseline38.860.9942.907 G0.939 M7.57 ms
RGC GC38.34 0.993 2.607 G 0.806 M 6.21 ms
Sigmoid Hard-Sigmoid39.02 0.995 2.909 G 0.939 M7.67 ms
     ReLUNaNNaN2.907 G0.939 M7.81 ms
     GeLUNaNNaN2.907 G0.939 M7.43 ms
Order [1, 1, 1, 1, 1, 1, 1]38.34 0.993 2.607 G 0.806 M 6.21 ms
    [2, 2, 2, 2, 2, 2, 2]38.76 0.9943.036 G 0.918 M 7.54 ms
    [3, 3, 3, 3, 3, 3, 3]38.07 0.991 3.181 G 0.951 M 9.26 ms
SK Fusion Cat37.59 0.990   3.115 G 0.958 M 7.31 ms
      Sum37.32 0.989   2.904 G 0.934 M 7.10 ms
Kernel Size = 5 337.64 0.989 2.69 G 0.904 M 7.11 ms
        738.00 0.991 3.239 G 0.991 M 7.67 ms
L1 LOSS L2 LOSS38.49 0.993 2.909 G 0.939 M7.81 ms
Depth × 239.82 0.995 5.207 G 1.675 M 13.70 ms
Width ×   2 39.70 0.995 4.965 G 1.638 M 7.35 ms
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Z.; Jia, J.; Lyu, P.; Min, J. Efficient Dehazing with Recursive Gated Convolution in U-Net: A Novel Approach for Image Dehazing. J. Imaging 2023, 9, 183. https://doi.org/10.3390/jimaging9090183

AMA Style

Wang Z, Jia J, Lyu P, Min J. Efficient Dehazing with Recursive Gated Convolution in U-Net: A Novel Approach for Image Dehazing. Journal of Imaging. 2023; 9(9):183. https://doi.org/10.3390/jimaging9090183

Chicago/Turabian Style

Wang, Zhibo, Jia Jia, Peng Lyu, and Jeongik Min. 2023. "Efficient Dehazing with Recursive Gated Convolution in U-Net: A Novel Approach for Image Dehazing" Journal of Imaging 9, no. 9: 183. https://doi.org/10.3390/jimaging9090183

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop