Next Article in Journal
Scale Effect of Land Cover Classification from Multi-Resolution Satellite Remote Sensing Data
Previous Article in Journal
ECKN: An Integrated Approach for Position Estimation, Packet Routing, and Sleep Scheduling in Wireless Sensor Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Focus Image Fusion via Distance-Weighted Regional Energy and Structure Tensor in NSCT Domain

1
College of Information Science and Engineering, Xinjiang University, Urumqi 830046, China
2
School of Information and Electronics, Beijing Institute of Technology, Beijing 100081, China
3
School of Mechanical Engineering, Guangxi University, Nanning 530004, China
4
State Key Laboratory of Remote Sensing Science, Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100101, China
5
Department of Electronic Engineering, Tsinghua University, Beijing 100084, China
*
Authors to whom correspondence should be addressed.
Sensors 2023, 23(13), 6135; https://doi.org/10.3390/s23136135
Submission received: 15 May 2023 / Revised: 21 June 2023 / Accepted: 26 June 2023 / Published: 4 July 2023
(This article belongs to the Section Sensing and Imaging)

Abstract

:
In this paper, a multi-focus image fusion algorithm via the distance-weighted regional energy and structure tensor in non-subsampled contourlet transform domain is introduced. The distance-weighted regional energy-based fusion rule was used to deal with low-frequency components, and the structure tensor-based fusion rule was used to process high-frequency components; fused sub-bands were integrated with the inverse non-subsampled contourlet transform, and a fused multi-focus image was generated. We conducted a series of simulations and experiments on the multi-focus image public dataset Lytro; the experimental results of 20 sets of data show that our algorithm has significant advantages compared to advanced algorithms and that it can produce clearer and more informative multi-focus fusion images.

1. Introduction

In order to obtain richer and more useful information and to improve the completeness and accuracy of scene description, multi-focus images obtained from video sensors are usually fused [1,2]. Multi-focus image fusion uses effective information processing methods to fuse clear and focused information from different images, resulting in a high-definition fully focused image [3,4,5]. Figure 1 depicts an example of multi-focus image fusion.
After many years of extensive and in-depth research, the image fusion method has made significant progress. Image fusion approaches can be divided into three categories: transform domain-based, edge-preserving filtering-based, and deep learning-based algorithms [6,7,8,9].
The most widely used transform domain-based methods are the curvelet transform [10], contourlet transform [11,12], shearlet transform [13,14], etc. Kumar et al. [15] constructed an intelligent multi-modal image fusion technique utilizing the fast discrete curvelet transform and type-2 fuzzy entropy, and the fusion results demonstrate the efficiency of this model in terms of subjective and objective assessment. Kumar et al. [16] constructed an image fusion technique via an improved multi-objective meta-heuristic algorithm with fuzzy entropy in fast discrete curvelet transform domain; a comparison of the developed methodology over the state-of-the-art models observes enhanced performance with respect to visual quality assessment. Li et al. [17] integrated the curvelet and discrete wavelet transform for multi-focus image fusion, and it can obtain advanced fusion performance. Zhang et al. [18] constructed an image fusion model using the contourlet transform. The average-based fusion rule and region energy-based fusion rule are performed on the low- and high-frequency sub-bands, respectively; the fused image is analyzed using fast non-local clustering for multi-temporal synthetic aperture radar image change detection; and this technique generates state-of-the-art change detection performance on both small- and large-scale datasets. Li et al. [19] constructed an image fusion approach using sparse representation and local energy in shearlet transform domain; this method can obtain good fusion performance. Hao et al. [20] introduced a multi-scale decomposition optimization-based image fusion method via gradient-weighted local energy; the non-subsampled shearlet transform is utilized to decompose the source images into low- and high-frequency sub-images, and then the low-frequency components are divided into base layers and texture layers, the base layers are merged according to the intrinsic attribute-based energy fusion rule, and the structure tensor-based gradient-weighted local energy operator fusion rule is utilized to merge the texture layers and high-frequency sub-bands. This fusion method can achieve a superior fusion capability in both qualitative and quantitative assessments.
The methods based on edge-preserving filtering also hold an important position in the field of image fusion, such as guided image filtering [21], cross bilateral filtering [22], Gaussian curvature filtering [23], sub-window variance filtering [24], etc. Feng et al. [24] constructed the multi-channel dynamic threshold neural P systems-based image fusion method via a visual saliency map in the sub-window variance filter domain; this algorithm can obtain effective fusion performance both on visual quality and quantitative evaluations. Zhang et al. [25] introduced a double joint edge preservation filter-based image fusion technique via a non-globally saliency gradient operator; it can obtain excellent subjective and objective performance. Jiang et al. [26] introduced the image fusion approach utilizing entropy measure between intuitionistic fuzzy set joint Gaussian curvature filtering.
The image fusion approaches based on deep learning have generated good performance, and these deep learning-based fusion models can be divided into supervised- and unsupervised-based approaches [27,28]. In terms of supervised algorithms, some effective fusion methods have been generated. The deep convolutional neural network introduced by Liu et al. [29] is used for multi-focus image fusion, and it generates good fusion results. The multi-scale visual attention deep convolutional neural network introduced by Lai et al. [30] is utilized to fuse multi-focus images. Wang et al. [31] introduced weakly supervised image fusion with modal synthesis and enhancement. The self-supervised residual feature learning model constructed by Wang et al. [32] is used to fuse multi-focus images. The attention mechanism-based image fusion approach via supervised learning was constructed by Jiang et al. [33]. In terms of unsupervised algorithms, Jin et al. [34] introduced the Transformer and U-Net for image fusion. Zhang et al. [35] presented another unsupervised generative adversarial network with adaptive and gradient joint constraints for image fusion. Liu et al. [36] constructed a generative adversarial network-based unsupervised back project dense network for image fusion. These deep learning-based fusion methods also achieve good performance.
In order to improve the fusion performance of multi-focus images, a distance-weighted regional energy-based multi-focus image fusion algorithm via the structure tensor in non-subsampled contourlet transform domain was constructed. The distance-weighted regional energy-based fusion rule and the structure tensor-based fusion rule were used to merge low- and high-frequency sub-images, respectively. The experimental results of the Lytro dataset show that the proposed fusion method can produce fusion effects that are superior to other traditional and deep learning methods in terms of subjective visual and objective evaluation.

2. Backgrounds

2.1. NSCT

The non-subsampled contourlet transform (NSCT) is a transformation with multi-directional, multi-scale, and translation invariance [12]. Its basic framework structure is divided into two parts: the non-subsampled pyramid (NSP) decomposition mechanism and the non-subsampled directional filter bank (NSDFB) decomposition mechanism. The NSCT first utilizes NSP decomposition to perform multi-scale decomposition on the source image and then uses NSDFB decomposition to further decompose the high-frequency components in the direction, ultimately obtaining sub-band images of the source image at different scales and in different directions [12]. Figure 2 shows the NSCT basic framework structure diagram and the NSCT frequency domain partition diagram.

2.2. Structure Tensor

For a multi-channel image, f x , y = f 1 x , y ,   f 2 x , y , ,   f 3 x , y , the grayscale images and color images are two special cases when m = 1 and m = 3, respectively. The square of variation of f x , y at position x , y in direction θ for any ε 0 + can be given by the following [37]:
d f 2 = f x + ε cos θ , y + ε sin θ f x , y 2 2   i = 1 m f i x ε cos θ + f i y ε sin θ 2
The rate of change C θ of image f x , y at position x , y can be expressed as follows:
C θ = i = 1 m f i x cos θ + f i y sin θ 2 = cos θ , sin θ i = 1 m f i x 2 i = 1 m f i x f i y i = 1 m f i y f i x i = 1 m f i y 2 cos θ , sin θ T = cos θ , sin θ i = 1 m f i f i T cos θ , sin θ T
where f i = f i x , f i y T , the following second-moment positive semi-definite matrix is
P = i = 1 m f i f i T = E F F G
where P is called the structure tensor and where E, F, and G are defined as
E = i = 1 m f i x 2
F = i = 1 m f i x f i y
G = i = 1 m f i y 2
The structure tensor-based focus detection operator (STO) is given by
S = λ 1 + λ 2 2 + 0.5 λ 1 λ 2 2
where λ 1 and λ 2 are the eigenvalues of the structure tensor and can be computed by the following:
λ 1 , 2 = 1 2 E + G ± G E 2 + 4 F 2

3. The Proposed Multi-Focus Image Fusion Method

The multi-focus image fusion method based on the distance-weighted regional energy and structure tensor in NSCT domain is proposed in this section, and the schematic of the proposed algorithm is depicted in Figure 3. This fusion method can be divided into four parts: NSCT decomposition, low-frequency components fusion, high-frequency components fusion, and the inverse NSCT. More details can be seen in the follows.

3.1. NSCT Decomposition

The multi-focus images A and B were decomposed into low-frequency components and high-frequency components through the NSCT, and the corresponding coefficients are defined as L A , H A l , k and L B , H B l , k , respectively.

3.2. Low-Frequency Components Fusion

Low-frequency components contain more energy and information in the images. In this section, the distance-weighted regional energy (DWRE)-based rule was used to merge the low-pass sub-bands, and it is defined as follows [38]:
D W R E L X i , j = p = 1 1   q = 1 1 W p + 2 , q + 2 L X i + p , j + q 2
where W shows a 3 × 3 matrix that allocates weights to the neighboring coefficients and where W is defined as
W = 1 1 + 2 1 2 1 1 + 2 1 2 1 1 2 1 1 + 2 1 2 1 1 + 2
Here, each entry of W is achieved by reciprocating ‘1 + distance of the respective position from its center, i.e., W 2 , 2 ’, which is depicted in the following:
W i , j = 1 1 + i 2 2 + j 2 2
The fused low-frequency component L F i , j is constructed by the following:
L F i , j = L A i , j   if   D W R E L A i , j D W R E L B i , j L B i , j   else
where D W R E L X i , j | X A , B is the DWRE of L X i , j estimated over a 3 × 3 neighborhood centered at the i , j th coefficient.

3.3. High-Frequency Components Fusion

High-frequency components contain more details and noise in the images. In this section, the structure tensor-based focus detection operator (STO) fusion rule was used to process the high-frequency components, and the fused coefficients H F l , k i , j can be computed by
H F l , k i , j = H X l , k i , j × M ˜ X l , k i , j X A , B
where
M ˜ X l , k i , j = 1   if   a , b Ω 1   M X l , k i + a , j + b > 0.5 × s i z e Ω 1 0   else
M X l , k i , j = 1   if   arg max S A l , k i , j , S B l , k i , j 0   else
where H X l , k i , j shows the high-frequency coefficient of image X A , B at the l -th scale and k -th direction at location i , j and where S X l , k i , j shows the structure salient image of H X l , k i , j computed by STO. Ω 1 is a local region centered at i , j with the size of W 1 × V 1 . Equation (14) is the classical operator called the consistency verification, which can improve the robustness and rectify some wrong focusing detection of M X l , k i , j .

3.4. Inverse NSCT

The final fused image F was generated by the inverse NSCT, which was performed on fused low- and high-frequency components L F , H F l , k , and F is defined as follows:
F i , j = L A i , j + H F l , k i , j
The main steps of the proposed multi-focus image fusion approach are summarized in Algorithm 1.
Algorithm 1 Proposed multi-focus image fusion method
Input: the source images: A and B
Parameters: The number of NSCT decomposition levels: L, the number of directions at each decomposition level: K ( l ) ,   l [ 1 , L ]
Main step:
Step 1: NSCT decomposition
For each source image X { A , B }
  Perform NSCT decomposition on X to generate { L X , H X l , k } ,   l [ 1 , L ] ,   k [ 1 , K ( l ) ] ;
End
Step 2: Low-frequency components fusion
For each source image X { A , B }
  Calculate the DWRE for LX utilizing Equations (9)–(11);
End
Merge LA and LB utilizing Equations (12) to generate LF;
Step 3: High-frequency components fusion
For each Level l = 1 : L
  For each direction k = 1 : K ( l )
   For each source image X A , B
    Calculate the structure salient image S X l , k i , j of H X l , k i , j computed by STO using Equation (7);
    Calculate the consistency verification using Equations (14) and (15);
   End
   Merge H A l , k and H B l , k using Equation (13)
  End
End
Step 4. Inverse NSCT
Perform inverse NSCT on L F , H F l , k to generate F;
Output: the fused image F.

4. Experimental Results and Discussions

In this section, the Lytro dataset with twenty pair images (Figure 4) are utilized to experiment, and eight state-of-the-art fusion algorithms are used to be compared with our method, namely, multi-focus image fusion using a bilateral-gradient-based sharpness criterion (BGSC) [39], the non-subsampled contourlet transform and fuzzy-adaptive reduced pulse-coupled neural network (NSCT) [40], multi-focus image fusion in gradient domain (GD) [41], the unified densely connected network for image fusion (FusionDN) [42], the fast unified image fusion network based on the proportional maintenance of gradient and intensity (PMGI) [43], image fusion based on target-enhanced multi-scale transform decomposition (TEMST) [44], the unified unsupervised image fusion network (U2Fusion) [45], and zero-shot multi-focus image fusion (ZMFF) [3]. Subjective and objective assessments are also used; in terms of objective assessment, eight metrics, including edge-based similarity measurement Q A B / F [46], the feature mutual information metric Q F M I [47], the gradient-based metric Q G [48], the structural similarity-based metric Q E [48], the phase congruency-based metric Q P [48], the structural similarity-based metric Q Y introduced by Yang et al. [48], the average gradient metric Q A G [22], and the average pixel intensity metric Q A P I [22], are used. The larger the values of these indicators, the better the fusion effect. In the proposed method, the size of Ω 1 is set as 5 × 5; the ‘vk’ and ‘pyrexc’ are used as the pyramid filter and directional filter, respectively; and the four decomposition levels with 2, 2, 2, and 2 directions from a coarser scale to a finer scale are used.

4.1. Qualitative Comparisons

In this section, we selected five sets of data from the Lytro dataset for result display; the qualitative comparisons of different methods are shown in Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9. We have enlarged some areas for easy observation and comparison. In Figure 5, we can see that the fusion result generated by the BGSC method has a weak far-focus fusion effect and that black spots appear around the building; the fusion results computed by the NSCT, GD, FusionDN, PMGI, and TEMST methods are somewhat blurry; the U2Fusion method generates a dark fusion image which limits the observation of some areas in the image; and the ZMFF method produces fully focused image results. However, compared to our method, our result produces moderate brightness and clear edges, especially the brightness of the red lock in the image, which is better than ZMFF.
In Figure 6, we can see that the fusion results generated by the BGSC and PMGI methods are blurry, making it difficult to observe the detailed information in the fused images; the fusion images computed by NSCT, FusionDN, TEMST, and ZMFF are similar, these images all yielding fully focused image results; the GD method generates a high-brightness fusion image, with the color information in the image being partially lost; the U2Fusion method generates a sharpened fused image, with some areas of the image having lost information, such as the branches of the tree on the right side and the branches of the small tree that is far away in the image, all of which are darker in the image, which is not conducive to obtaining information in a fully focused image; and the fusion image calculated by the proposed method is an all-focused image, with the details and edge information of the image being well preserved and with moderate brightness and ease of observation.
In Figure 7, we can denote that the fusion images calculated by the BGSC, GD, FusionDN, PMGI, and TEMST methods appear blurry, the images generated by BGSC and TEMST especially exhibiting severe distortion; the image computed by the NSCT method shows a basic full-focus result with some artifacts in the image, making it appear to have texture overlays; the U2Fusion method generates a dark image; and the fusion image calculated by the ZMFF method is very close to the result obtained by our algorithm. However, our fusion effect is better, with a clearer image and the almost seamless integration of all the information into the fully focused image.
In Figure 8, we can see that the fused images computed by the BGSC, FusionDN, and PMGI methods appear blurry, especially the forest information in the far distance of the images, making it difficult to observe the details; the images generated by the NSCT, TEMST, and ZMFF methods are similar; the fused image obtained by the GD method has high brightness and clarity; the fused image computed by the U2Fusion method has some dark regions, especially in the mouth area of the horse, with the loss of information being severe; and the fused image achieved by the proposed method has moderate brightness and retains more image information.
In Figure 9, we can see that the fusion images obtained by the BGSC, FusionDN, and PMGI approaches exhibit a certain degree of the blurring phenomenon, which makes it difficult to observe the information in the images, the image generated by the BGSC method especially not achieving full focus and the information on the left side of the image being severely lost; the NSCT, TEMST, and ZMFF methods all obtain a basic fully focused image from which information about the entire scene can be observed; the GD method gives a high-brightness fusion image but can also cause some details in the image to be lost; the fused image achieved by the U2Fusion method has many darker areas, causing severe information loss (for example, the information about the girl’s clothing and headscarf in the image cannot be captured correctly); the fusion image obtained by our algorithm fully preserves the information of the two source images. Additionally, it not only has moderate brightness, but the detailed information of the entire scene can also be fully observed and obtained.

4.2. Quantitative Comparisons

In this section, the quantitative comparisons of different methods are shown in Table 1, Table 2, Table 3, Table 4, Table 5 and Table 6 and Figure 10. In Table 1, we can see that the metrics data Q A B / F , Q F M I , Q G , Q E , Q P , Q Y , and Q A G computed by the proposed method are the best; the metric data Q A P I generated by the GD method is the best, but the corresponding indicator result of our algorithm still ranks third. In Table 2, we can see that the metrics data QAB/F, QFMI, QG, QE, QP, and QY generated by our method are the best. The metric data Q A G generated by the U2Fusion method is the best, while our method ranks second; the metric data Q A P I generated by the GD method is the best, and our method ranks fourth. In Table 3, we denoted that the metrics data QAB/F, QFMI, QG, QE, QP, QY and QAG computed by our method are the best; the metric data Q A P I generated by the FusionDN method is the best, and our method ranks fourth. In Table 4, we can see that the metrics data QAB/F, QFMI, QG, QE, QP, and QY generated by our method are the best; the metrics data Q A G and Q A P I generated by the GD method are the best. In Table 5, we denoted that the metrics data Q A B / F , Q F M I , Q G , Q E , Q P , Q Y , and Q A G computed by our method are the best; the metric data Q A P I computed by the FusionDN method is the best.
Figure 10 shows the objective experimental results of different algorithms with respect to 20 sets of image pairs, and the data results of the same method with respect to different image pairs are connected into a curve, with the average indicator value on the right side of the indicator graph. Table 6 shows the average quantitative assessment of the different methods in Figure 10. In Figure 10 and Table 6, we can see that the average metrics data Q A B / F , Q F M I , Q G , Q E , Q P , and Q Y generated by our method are the best; the average metrics date Q A G and Q A P I generated by the U2Fusion and GD methods are the best, respectively, and the two corresponding average metrics data computed by our method rank second.
Through rigorous qualitative and quantitative evaluation and analysis, the results show that our algorithm stands out in the fusion effect of multi-focus images. Compared with state-of-the-art algorithms, we have achieved the best fusion effect and have the advantages of rich information and clear images.

5. Conclusions

In this paper, a novel multi-focus image fusion method based on the distance-weighted regional energy and structure tensor in NSCT domain was proposed. The structure tensor-based fusion rule was utilized to fuse low-frequency sub-bands, and the distance-weighted regional energy-based fusion rule was utilized to fused high-frequency sub-bands. The proposed method was experimented on the Lytro dataset with 20 paired images, and the fusion results demonstrate that our method can generate state-of-the-art fusion performance in terms of image information, definition, and brightness, making the seamless fusion of multi-focus images possible. In future work, we will improve and expand the application of this algorithm in the field of multi-modal image fusion so that it has better universality in image fusion.

Author Contributions

The experimental measurements and data collection were carried out by M.L., L.L., Z.J. and H.M. The manuscript was written by M.L. with the assistance of L.L., Q.J., Z.J., L.C. and H.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Science Foundation of China under Grant No. 62261053; the Shanghai Aerospace Science and Technology Innovation Fund under Grant No. SAST2019-048; and the Cross-Media Intelligent Technology Project of the Beijing National Research Center for Information Science and Technology (BNRist) under Grant No. BNR2019TD01022.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, H.; Xu, H. Image fusion meets deep learning: A survey and perspective. Inf. Fusion 2021, 76, 323–336. [Google Scholar] [CrossRef]
  2. Karim, S.; Tong, G. Current advances and future perspectives of image fusion: A comprehensive review. Inf. Fusion 2023, 90, 185–217. [Google Scholar] [CrossRef]
  3. Hu, X.; Jiang, J.; Liu, X.; Ma, J. ZMFF: Zero-shot multi-focus image fusion. Inf. Fusion 2023, 92, 127–138. [Google Scholar] [CrossRef]
  4. Zafar, R.; Farid, M.; Khan, M. Multi-focus image fusion: Algorithms, evaluation, and a library. J. Imaging 2020, 6, 60. [Google Scholar] [CrossRef]
  5. Dong, Y.; Chen, Z.; Li, Z.; Gao, F. A multi-branch multi-scale deep learning image fusion algorithm based on DenseNet. Appl. Sci. 2022, 12, 10989. [Google Scholar] [CrossRef]
  6. Singh, S.; Singh, H. A review of image fusion: Methods, applications and performance metrics. Digit. Signal Process. 2023, 137, 104020. [Google Scholar] [CrossRef]
  7. Li, L.; Ma, H.; Jia, Z.; Si, Y. A novel multiscale transform decomposition based multi-focus image fusion framework. Multimed. Tools Appl. 2021, 80, 12389–12409. [Google Scholar] [CrossRef]
  8. Li, L.; Si, Y.; Wang, L.; Jia, Z.; Ma, H. A novel approach for multi-focus image fusion based on SF-PAPCNN and ISML in NSST domain. Multimed. Tools Appl. 2020, 79, 24303–24328. [Google Scholar] [CrossRef]
  9. Wu, P.; Hua, Z.; Li, J. Multi-scale siamese networks for multi-focus image fusion. Multimed. Tools Appl. 2023, 82, 15651–15672. [Google Scholar] [CrossRef]
  10. Candes, E.; Demanet, L. Fast discrete curvelet transforms. Multiscale Model. Simul. 2006, 5, 861–899. [Google Scholar] [CrossRef] [Green Version]
  11. Do, M.N.; Vetterli, M. The contourlet transform: An efficient directional multiresolution image representation. IEEE Trans. Image Process. 2005, 14, 2091–2106. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Da, A.; Zhou, J.; Do, M. The nonsubsampled contourlet transform: Theory, design, and applications. IEEE Trans. Image Process. 2006, 15, 3089–3101. [Google Scholar]
  13. Guo, K.; Labate, D. Optimally sparse multidimensional representation using shearlets. SIAM J. Math. Anal. 2007, 39, 298–318. [Google Scholar] [CrossRef] [Green Version]
  14. Easley, G.; Labate, D.; Lim, W.Q. Sparse directional image representations using the discrete shearlet transform. Appl. Comput. Harmon. Anal. 2008, 25, 25–46. [Google Scholar] [CrossRef] [Green Version]
  15. Kumar, N.; Prasad, T.; Prasad, K. An intelligent multimodal medical image fusion model based on improved fast discrete curvelet transform and type-2 fuzzy entropy. Int. J. Fuzzy Syst. 2023, 25, 96–117. [Google Scholar] [CrossRef]
  16. Kumar, N.; Prasad, T.; Prasad, K. Multimodal medical image fusion with improved multi-objective meta-heuristic algorithm with fuzzy entropy. J. Inf. Knowl. Manag. 2023, 22, 2250063. [Google Scholar] [CrossRef]
  17. Li, S.; Yang, B. Multifocus image fusion by combining curvelet and wavelet transform. Pattern Recognit. Lett. 2008, 29, 1295–1301. [Google Scholar] [CrossRef]
  18. Zhang, W.; Jiao, L.; Liu, F. Adaptive contourlet fusion clustering for SAR image change detection. IEEE Trans. Image Process. 2022, 31, 2295–2308. [Google Scholar] [CrossRef]
  19. Li, L.; Lv, M.; Jia, Z.; Ma, H. Sparse representation-based multi-focus image fusion method via local energy in shearlet domain. Sensors 2023, 23, 2888. [Google Scholar] [CrossRef] [PubMed]
  20. Hao, H.; Zhang, B.; Wang, K. MGFuse: An infrared and visible image fusion algorithm based on multiscale decomposition optimization and gradient-weighted local energy. IEEE Access 2023, 11, 33248–33260. [Google Scholar] [CrossRef]
  21. Li, S.; Kang, X.; Hu, J. Image fusion with guided filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar] [PubMed]
  22. Shreyamsha Kumar, B. Image fusion based on pixel significance using cross bilateral filter. Signal Image Video Process. 2015, 9, 1193–1204. [Google Scholar] [CrossRef]
  23. Tan, W.; Zhou, H. Fusion of multi-focus images via a Gaussian curvature filter and synthetic focusing degree criterion. Appl. Opt. 2018, 57, 10092–10101. [Google Scholar] [CrossRef] [PubMed]
  24. Feng, X.; Fang, C.; Qiu, G. Multimodal medical image fusion based on visual saliency map and multichannel dynamic threshold neural P systems in sub-window variance filter domain. Biomed. Signal Process. Control 2023, 84, 104794. [Google Scholar] [CrossRef]
  25. Zhang, Y.; Lee, H. Multi-sensor infrared and visible image fusion via double joint edge preservation filter and non-globally saliency gradient operator. IEEE Sens. J. 2023, 23, 10252–10267. [Google Scholar] [CrossRef]
  26. Jiang, Q.; Huang, J. Medical image fusion using a new entropy measure between intuitionistic fuzzy sets joint Gaussian curvature filter. IEEE Trans. Radiat. Plasma Med. Sci. 2023, 7, 494–508. [Google Scholar] [CrossRef]
  27. Zhang, C. Multifocus image fusion using a convolutional elastic network. Multimed. Tools Appl. 2022, 81, 1395–1418. [Google Scholar] [CrossRef]
  28. Ma, W.; Wang, K.; Li, J. Infrared and visible image fusion technology and application: A review. Sensors 2023, 23, 599. [Google Scholar] [CrossRef]
  29. Liu, Y.; Chen, X.; Peng, H. Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 2017, 36, 191–207. [Google Scholar] [CrossRef]
  30. Lai, R.; Li, Y.; Guan, J. Multi-scale visual attention deep convolutional neural network for multi-focus image fusion. IEEE Access 2019, 7, 114385–114399. [Google Scholar] [CrossRef]
  31. Wang, L.; Liu, Y. MSE-Fusion: Weakly supervised medical image fusion with modal synthesis and enhancement. Eng. Appl. Artif. Intell. 2023, 119, 105744. [Google Scholar] [CrossRef]
  32. Wang, Z.; Li, X. A self-supervised residual feature learning model for multifocus image fusion. IEEE Trans. Image Process. 2022, 31, 4527–4542. [Google Scholar] [CrossRef] [PubMed]
  33. Jiang, L.; Fan, H.; Li, J. A multi-focus image fusion method based on attention mechanism and supervised learning. Appl. Intell. 2022, 52, 339–357. [Google Scholar] [CrossRef]
  34. Jin, X.; Xi, X.; Zhou, D. An unsupervised multi-focus image fusion method based on Transformer and U-Net. IET Image Process. 2023, 17, 733–746. [Google Scholar] [CrossRef]
  35. Zhang, H.; Le, Z.; Shao, Z. MFF-GAN: An unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion. Inf. Fusion 2021, 66, 40–53. [Google Scholar] [CrossRef]
  36. Liu, S.; Yang, L. BPDGAN: A GAN-based unsupervised back project dense network for multi-modal medical image fusion. Entropy 2022, 24, 1823. [Google Scholar] [CrossRef]
  37. Li, X.; Zhou, F.; Tan, H. Multi-focus image fusion based on nonsubsampled contourlet transform and residual removal. Signal Process. 2021, 184, 108062. [Google Scholar] [CrossRef]
  38. Panigrahy, C.; Seal, A. Parameter adaptive unit-linking pulse coupled neural network based MRI-PET/SPECT image fusion. Biomed. Signal Process. Control 2023, 83, 104659. [Google Scholar] [CrossRef]
  39. Tian, J.; Chen, L.; Ma, L. Multi-focus image fusion using a bilateral gradient-based sharpness criterion. Opt. Commun. 2011, 284, 80–87. [Google Scholar] [CrossRef]
  40. Das, S.; Kundu, M.K. A neuro-fuzzy approach for medical image fusion. IEEE Trans. Biomed. Eng. 2013, 60, 3347–3353. [Google Scholar] [CrossRef]
  41. Paul, S.; Sevcenco, I.; Agathoklis, P. Multi-exposure and multi-focus image fusion in gradient domain. J. Circuits Syst. Comput. 2016, 25, 1650123. [Google Scholar] [CrossRef] [Green Version]
  42. Xu, H.; Ma, J.; Le, Z. FusionDN: A unified densely connected network for image fusion. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI), New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12484–12491. [Google Scholar]
  43. Zhang, H.; Xu, H.; Xiao, Y. Rethinking the image fusion: A fast unified image fusion network based on proportional maintenance of gradient and intensity. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI), New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12797–12804. [Google Scholar]
  44. Chen, J.; Li, X.; Luo, L. Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf. Sci. 2020, 508, 64–78. [Google Scholar] [CrossRef]
  45. Xu, H.; Ma, J.; Jiang, J. U2Fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 502–518. [Google Scholar] [CrossRef] [PubMed]
  46. Qu, X.; Yan, J.; Xiao, H. Image fusion algorithm based on spatial frequency-motivated pulse coupled neural networks in nonsubsampled contourlet transform domain. Acta Autom. Sin. 2008, 34, 1508–1514. [Google Scholar] [CrossRef]
  47. Haghighat, M.; Razian, M. Fast-FMI: Non-reference image fusion metric. In Proceedings of the IEEE 8th International Conference on Application of Information and Communication Technologies, Astana, Kazakhstan, 15–17 October 2014; pp. 424–426. [Google Scholar]
  48. Liu, Z.; Blasch, E.; Xue, Z. Objective assessment of multiresolution image fusion algorithms for context enhancement in night vision: A comparative study. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 94–109. [Google Scholar] [CrossRef]
Figure 1. Multi-focus image fusion. (a) Near-focused image; (b) far-focused image; and (c) fused image.
Figure 1. Multi-focus image fusion. (a) Near-focused image; (b) far-focused image; and (c) fused image.
Sensors 23 06135 g001
Figure 2. (a) NSCT basic framework structure diagram; (b) NSCT frequency domain partition diagram.
Figure 2. (a) NSCT basic framework structure diagram; (b) NSCT frequency domain partition diagram.
Sensors 23 06135 g002
Figure 3. The schematic of the proposed fusion algorithm.
Figure 3. The schematic of the proposed fusion algorithm.
Sensors 23 06135 g003
Figure 4. Lytro dataset.
Figure 4. Lytro dataset.
Sensors 23 06135 g004
Figure 5. Lytro-06 fusion results. (a) BGSC; (b) NSCT; (c) GD; (d) FusionDN; (e) PMGI; (f) TEMST; (g) U2Fusion; (h) ZMFF; and (i) proposed.
Figure 5. Lytro-06 fusion results. (a) BGSC; (b) NSCT; (c) GD; (d) FusionDN; (e) PMGI; (f) TEMST; (g) U2Fusion; (h) ZMFF; and (i) proposed.
Sensors 23 06135 g005aSensors 23 06135 g005b
Figure 6. Lytro-10 fusion results. (a) BGSC; (b) NSCT; (c) GD; (d) FusionDN; (e) PMGI; (f) TEMST; (g) U2Fusion; (h) ZMFF; and (i) proposed.
Figure 6. Lytro-10 fusion results. (a) BGSC; (b) NSCT; (c) GD; (d) FusionDN; (e) PMGI; (f) TEMST; (g) U2Fusion; (h) ZMFF; and (i) proposed.
Sensors 23 06135 g006aSensors 23 06135 g006b
Figure 7. Lytro-13 fusion results. (a) BGSC; (b) NSCT; (c) GD; (d) FusionDN; (e) PMGI; (f) TEMST; (g) U2Fusion; (h) ZMFF; and (i) proposed.
Figure 7. Lytro-13 fusion results. (a) BGSC; (b) NSCT; (c) GD; (d) FusionDN; (e) PMGI; (f) TEMST; (g) U2Fusion; (h) ZMFF; and (i) proposed.
Sensors 23 06135 g007aSensors 23 06135 g007b
Figure 8. Lytro-15 fusion results. (a) BGSC; (b) NSCT; (c) GD; (d) FusionDN; (e) PMGI; (f) TEMST; (g) U2Fusion; (h) ZMFF; and (i) proposed.
Figure 8. Lytro-15 fusion results. (a) BGSC; (b) NSCT; (c) GD; (d) FusionDN; (e) PMGI; (f) TEMST; (g) U2Fusion; (h) ZMFF; and (i) proposed.
Sensors 23 06135 g008aSensors 23 06135 g008b
Figure 9. Lytro-18 fusion results. (a) BGSC; (b) NSCT; (c) GD; (d) FusionDN; (e) PMGI; (f) TEMST; (g) U2Fusion; (h) ZMFF; and (i) proposed.
Figure 9. Lytro-18 fusion results. (a) BGSC; (b) NSCT; (c) GD; (d) FusionDN; (e) PMGI; (f) TEMST; (g) U2Fusion; (h) ZMFF; and (i) proposed.
Sensors 23 06135 g009aSensors 23 06135 g009b
Figure 10. The line chart of metrics data.
Figure 10. The line chart of metrics data.
Sensors 23 06135 g010aSensors 23 06135 g010b
Table 1. The quantitative assessment of different methods in Figure 5.
Table 1. The quantitative assessment of different methods in Figure 5.
QAB/FQMFIQGQEQPQYQAGQAPI
BGSC0.50830.89100.51040.46080.49020.87148.2490103.9694
NSCT0.69310.89630.69340.83220.78180.920510.1363104.1275
GD0.68430.89140.68740.73120.77180.81869.6352120.8782
FusionDN0.53980.88720.54590.72740.57690.77897.8779110.6457
PMGI0.28190.88350.28290.34290.40290.60424.3240105.6984
TEMST0.66320.89240.66330.77130.77020.9323 9.8586104.3450
U2Fusion0.59190.88630.59570.76460.69070.687210.079097.9331
ZMFF0.67020.88980.67360.82900.77630.91469.8506101.9007
Proposed0.71900.89680.71930.84190.82450.946110.3594105.8613
Table 2. The quantitative assessment of different methods in Figure 6.
Table 2. The quantitative assessment of different methods in Figure 6.
QAB/FQMFIQGQEQPQYQAGQAPI
BGSC0.60420.8409 0.60590.62890.61810.942421.4182100.9177
NSCT0.72920.84310.73210.83450.74980.986124.1701100.8183
GD0.71720.83970.72110.80100.74030.931324.0096118.4325
FusionDN0.57620.83210.58000.74860.63800.9205 24.6566109.2992
PMGI0.47430.83520.4806 0.6013 0.58410.8205 13.6582105.7407
TEMST0.71590.84040.71930.8274 0.74210.981523.7422101.4723
U2Fusion0.56670.83200.57120.72190.61860.866625.578293.3662
ZMFF0.70890.84290.71200.83160.74430.977224.068098.6039
Proposed0.73810.84530.74120.83980.76490.989324.7951102.7896
Table 3. The quantitative assessment of different methods in Figure 7.
Table 3. The quantitative assessment of different methods in Figure 7.
QAB/FQMFIQGQEQPQYQAGQAPI
BGSC0.35100.81380.34630.27340.24380.715014.575482.8519
NSCT0.76090.85280.75620.88740.74510.955620.282484.1720
GD0.75970.84770.75430.79060.77200.866819.6094107.0923
FusionDN0.64140.83980.63120.77180.64400.835617.8845108.0876
PMGI0.23690.83910.23090.33420.47730.62858.8447101.0823
TEMST0.65630.84160.64910.75980.65440.916518.414283.8760
U2Fusion0.62390.83540.61570.77890.61560.798218.115070.1179
ZMFF0.75850.84900.75130.89630.76520.967119.252979.9660
Proposed0.79090.85350.78810.90410.79190.972020.773385.8043
Table 4. The quantitative assessment of different methods in Figure 8.
Table 4. The quantitative assessment of different methods in Figure 8.
QAB/FQMFIQGQEQPQYQAGQAPI
BGSC0.68080.94520.67660.83620.63600.88738.0699112.4417
NSCT0.78900.94600.78500.93350.85850.94129.9235112.0279
GD0.73600.92470.73280.84630.79660.839711.8692124.0953
FusionDN0.65700.93020.65290.85120.64660.83679.5025110.2910
PMGI0.48460.93260.48170.59480.49540.68244.6840107.4757
TEMST0.78500.94460.78300.93360.84640.947710.1213112.5434
U2Fusion0.63080.92050.62950.78280.67920.786911.6540105.0806
ZMFF0.76950.93900.76730.92700.82930.92899.7681111.2188
Proposed0.81060.94630.81020.93670.89030.972310.3106112.7676
Table 5. The quantitative assessment of different methods in Figure 9.
Table 5. The quantitative assessment of different methods in Figure 9.
QAB/FQMFIQGQEQPQYQAGQAPI
BGSC0.61000.93080.61170.53000.64410.93394.655282.9920
NSCT0.68040.94100.68080.83080.71050.90205.100382.5546
GD0.67100.93010.66780.71180.70760.79645.3423104.4747
FusionDN0.56410.93280.56110.71920.47080.72524.7077105.7243
PMGI0.41560.92720.41470.49730.26990.61392.737397.1696
TEMST0.68990.93750.69020.80640.77990.94335.142882.9629
U2Fusion0.54940.93160.54690.71900.53620.68465.345872.9192
ZMFF0.68310.93900.68070.83680.76910.91005.144181.7770
Proposed0.72090.94440.72100.84400.82380.95875.433883.2692
Table 6. The average quantitative assessment of different methods in Figure 10.
Table 6. The average quantitative assessment of different methods in Figure 10.
QAB/FQMFIQGQEQPQYQAGQAPI
BGSC0.52420.88260.52100.51730.47420.85129.3365111.7289
NSCT0.71030.89720.70580.86440.76160.920811.3477111.8662
GD0.70340.88870.69870.78740.74660.860811.6844126.7408
FusionDN0.60180.88330.59520.76630.62210.822410.7865110.9212
PMGI0.39010.88150.38570.47360.46200.67385.8684108.0326
TEMST0.67220.89150.66640.80080.72630.915310.8599112.2319
U2Fusion0.61430.88440.60930.78350.66570.791212.0343108.1656
ZMFF0.70870.89250.70300.86870.78530.931311.2183109.8769
Proposed0.73880.89860.73660.87450.81000.953011.7371113.3531
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lv, M.; Li, L.; Jin, Q.; Jia, Z.; Chen, L.; Ma, H. Multi-Focus Image Fusion via Distance-Weighted Regional Energy and Structure Tensor in NSCT Domain. Sensors 2023, 23, 6135. https://doi.org/10.3390/s23136135

AMA Style

Lv M, Li L, Jin Q, Jia Z, Chen L, Ma H. Multi-Focus Image Fusion via Distance-Weighted Regional Energy and Structure Tensor in NSCT Domain. Sensors. 2023; 23(13):6135. https://doi.org/10.3390/s23136135

Chicago/Turabian Style

Lv, Ming, Liangliang Li, Qingxin Jin, Zhenhong Jia, Liangfu Chen, and Hongbing Ma. 2023. "Multi-Focus Image Fusion via Distance-Weighted Regional Energy and Structure Tensor in NSCT Domain" Sensors 23, no. 13: 6135. https://doi.org/10.3390/s23136135

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop