Next Article in Journal
Ergodic Capacity of NOMA-Based Multi-Antenna LMS Systems with Imperfect Limitations
Previous Article in Journal
Research on Multiple Spectral Ranges with Deep Learning for SpO2 Measurement
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Image Super-Resolution via Self-Calibrated Feature Fuse

1
College of Information Science and Engineering, Xinjiang University, Urumqi 830046, China
2
College of Mathematics and System Science, Xinjiang University, Urumqi 830046, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(1), 329; https://doi.org/10.3390/s22010329
Submission received: 2 December 2021 / Revised: 29 December 2021 / Accepted: 31 December 2021 / Published: 2 January 2022
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Recently, many super-resolution reconstruction (SR) feedforward networks based on deep learning have been proposed. These networks enable the reconstructed images to achieve convincing results. However, due to a large amount of computation and parameters, SR technology is greatly limited in devices with limited computing power. To trade-off the network performance and network parameters. In this paper, we propose the efficient image super-resolution network via Self-Calibrated Feature Fuse, named SCFFN, by constructing the self-calibrated feature fuse block (SCFFB). Specifically, to recover the high-frequency detail information of the image as much as possible, we propose SCFFB by self-transformation and self-fusion of features. In addition, to accelerate the network training while reducing the computational complexity of the network, we employ an attention mechanism to elaborate the reconstruction part of the network, called U-SCA. Compared with the existing transposed convolution, it can greatly reduce the computation burden of the network without reducing the reconstruction effect. We have conducted full quantitative and qualitative experiments on public datasets, and the experimental results show that the network achieves comparable performance to other networks, while we only need fewer parameters and computational resources.

1. Introduction

The essence of the SR task is to reconstruct the high-resolution (HR) image from a low-resolution (LR) image. The SR task is a hot and challenging point in low-level computer vision, which is mainly based on the fact that the reconstruction of a LR image into a HR image has different reconstruction directions with different environments, so it is inherently ill-posed. The problem is essentially a one-to-many relationship, which is difficult to solve with a specific mathematical relation formula. To solve this problem, many traditional methods have been proposed (e.g., interpolation-based methods [1] and degenerate model-based methods [2]), but their reconstruction results are not satisfactory. With the booming rise of deep learning (DL) techniques, convolutional neural networks (CNNs) have been attempted to constrain the solution space from LR to HR and have shown excellent performance.
Dong et al. [3] first applied CNN to the SR task and constructed the SRCNN model by establishing a direct relationship between LR and HR, which is obviously superior to the traditional non-DL method. Then, inspired by traditional sparse coding, Wang et al. [4] proposed a progressive upsampling method to achieve better HR generation at large upsampling factors (e.g., ×4). Due to the outstanding performance of the VGG [5] network on the ImageNet classification task, Kim et al. [6,7] increased the depth of the SR network to 20 layers to achieve better performance and showed that it greatly outperformed the SRCNN. We found that a deeper network model had a significant impact on improving the reconstruction performance of the network. But deeper networks are not conducive to their training, so some researchers have introduced higher learning rate and residual learning to alleviate this problem. Meanwhile, in order to effectively reduce parameters in the network, recursive learning was introduced in the DRCN [7] proposed by Kim et al. to save network parameters significantly. Similarly, DRRN [8] proposed by Tai et al. not only uses recursive learning to save network parameters but also introduces global and local residuals to promote the network training. Meanwhile, the MemNet [9] proposed by Tai et al. solves the CNN long-term dependency problem by applying recursive units and gate units. However, both algorithms require a long time and substantial graphical memory consumption during the training and testing phases. The main reason is that the first stage of these two models is to upsample LR, and the whole network process is to train the upsampled images without any downsampling operation, which introduces great computational consumption.
To address the above issues, we further explore the lightweight and reconstruction performance of single image SR network. In this paper, we propose a simple and efficient SR network via Self-Calibrated Feature Fuse (SCFFN) to achieve better balance between network performance and applicability. In the network proposed in this paper, the two crucial parts of the proposed network are the high-level semantic information learning part and the reconstruction part. In the deep feature extraction part, we propose the SCFFB according to the recent self-calibrated convolution [10]. The SCFFB has no complicated connection structure and up/down sampling operation as these are unfriendly to hardware acceleration. We summarize some previous work and find that other work is to use transpose convolution and sub-pixel convolution layer to implement image upsampling, and there is little work to study the impact of the reconstruction component on network performance and efficiency, but this structure is less efficient. In the network proposed in this paper, we use traditional NN and convolutional layers to improve the efficiency of the network and reduce parameters, while our introduced SCA improves the network performance at a small parameters cost. Therefore, the reconstruction part contains NN, SCA and convolutional layers. In general, because our network has no complicated connection mode, SCFFN is simpler and more efficient than the previous methods. As can be seen in Figure 1, our network achieves relatively optimal results in terms of parameters, model complexity and network reconstruction performance compared to the previous network.
The main contributions of this work are as follows:
  • We have carefully designed a simple and effective lightweight SR network (SCFFN), and extensive experimental results demonstrate the superiority of our network over other networks.
  • Inspired by self-calibrated convolution [10], we constructed a novel high-level feature learning block (SCFFB) for the SR task, which enables extract high-level information by its own feature fusion.
  • To further improve the efficiency of the network, we used the traditional nearest neighbor interpolation method and the simple channel attention mechanism we designed in the reconstruction part, named U-SCA block.

2. Related Work

In recent years, supervised deep learning methods have been introduced to various computer vision tasks [11] and have achieved remarkable success. Garcia et al. [12] summarized the application and achievements of deep learning in semantic segmentation in recent years. At the same time, to compensate for the disadvantage that supervised learning requires a large number of labels, researchers have proposed a semi-supervised learning method, where there is only a small amount of label data and a large amount of unlabel data. The mean teachers method proposed by Vainen et al. [13] effectively improves the network performance of semi-supervised learning by averaging model weights. Doulamis et al. [14] proposed a semi-supervised learning method for object classification/tracking, which allowed the unsupervised data to initially configure the network, and then triggered the gradient descent optimization scheme to fine-tune the data. In addition, an adaptive method was proposed, which allowed the model to be dynamically modified according to the current visual conditions. Baur et al. [15] embedded the concept of auxiliary manifold of semi-supervised learning into FCNs to solve the segmentation of MS lesions. For SR tasks, numerous CNN-based methods have been proposed [16,17,18,19,20] to improve the reconstruction performance of the network, however, the network parameters and computational consumption limit their application in industry [21].
Dong et al. [3] constructed the first super-resolution reconstruction method based on deep learning through three convolution layers, named SRCNN, whose reconstruction performance is obviously better than that of traditional methods, but the input of SRCNN is bicubic interpolated image, which greatly increases the computational cost and training time. To improve the reconstruction performance, the VDSR proposed by Kim et al. [6] used global residual learning to expand the network to 20 layers, and their experimental results demonstrated that increasing the depth of the network could improve the performance of the network. Later, more and more researchers designed very deep and complex networks to improve network performance, but the consequence was that the network parameters, complexity and training cost were greatly increased. For example, EDSR [19] achieved an unprecedented breakthrough in image reconstruction performance and won the 2017 NTIRE competition, but the network parameters and depth were 43 M and 69 layers respectively. Zhang et al. proposed RDN [16] by introducing dense connection in the residual block of EDSR, which includes 22 M parameters and 128 layers. Meanwhile, Zhang et al. proposed RCAN [17] by applying the channel attention mechanism to the residual structure for the first time, which greatly improved the network performance. The network depth reached 400 layers but only needed 15.59 M network parameters. Although these methods achieve good performance, they are not suitable for use in devices with limited resources. For mobile devices, in the case of limited available memory and inference time, we should aim at the highest possible SR performance. Moreover, many situations (such as video applications, smartphones, edge devices, etc.) require good performance and faster reconstruction speeds. It is therefore essential to design a lightweight and efficient model to meet these requirements. However, most of the algorithms mentioned above have a large number of parameters and require much more memory consumption, so these networks are unaffordable for mobile devices with limited resources in practical applications.
Many fast, lightweight networks have been designed to address these issues. To accelerate network training and reduce computational costs, FSRCNN [22] implements network upsampling in the final stage of the network. This structure enables the whole network to learn high-level information in the low-dimensional space. Meanwhile, the ESPCN [23] proposes an effective sub-pixel convolutional layer to implement image upsampling. In order to reduce network parameters, some researchers adopt parameter sharing and recursive structure. For example, the DRCN [7] proposed by Kim et al. is the first to apply recursive structures to the SR task, while introducing residual connection to alleviate the gradient disappearance caused by too many recurrences. The DRRN [8] proposed by Tai et al. introduced the recursive structure deeper through gradient cropping and achieved good results. Meanwhile, Tai et al. adopted recursive units and gate units to solve the long-term dependency problem of the network and proposed MemNet [9] for multi-tasks (including image denoising, SR and JPEG deblocking). The abovementioned methods all adopt recursive structure and parameter sharing to reduce the complexity of the model. Ahuja et al. introduced the Laplace pyramid structure to the SR task to address the speed and reconstruction performance of SR, proposing LapSRN [24] and MS-LapSRN [25], both of which use LR images as input to progressively reconstruct multi-scale HR images. Similar to this work, the ProSR [26] proposed by Perazzi et al. took this structure while reconstructing higher quality images at large scaling factors. The CARN [16] proposed by Ahn et al. uses cascade connection to efficiently transmit information, thus realizing the lightweight of the network. SRResNet [27] improves the performance by removing unnecessary blocks. Later, Hui et al. proposed lightweight IDN [18] and IMDN [28] based on information distillation. In a word, it is of great significance to design a lightweight and efficient SR network.
In this paper, we further explored the lightweight and reconstruction performance of the SR network. Inspired by self-calibration convolution [10], we designed a simple and efficient SR network, namely SCFFN. Consistent with the learning-based reconstruction method (e.g., pixel shuffling [29]) that realize upsampling at the last stage of the network. However, the reconstruction module of most networks consists of upsampling (sub-pixel convolution or transpose convolution) and convolution layers. The reconstruction part in SCFFN adopts NN, the SCA of our design and two convolutional layers. We conclude from previous work that attention mechanisms [30,31] can improve network performance, but there is little work on the impact of the reconstruction stage on network performance. Therefore, in this work, we employ an attention mechanism-based U-SCA block in the reconstruction phase to better reconstruct images. Sufficient experiments have shown that our network is superior to most methods in parameters and complexity.

3. Method

In this part, the proposed network structure is described in detail. Section 3.1 introduces the overall framework and implementation process of our network. Section 3.2 describes in detail SCFFB, which is the core of our algorithm. Section 3.3 shows the reconstruction module of our network. Section 3.4 presents the loss function we need to train the network.

3.1. Network Architecture

Figure 2 shows the general architecture of SCFFN, which consists of three modules, the feature extraction block (FEB), the high-level semantic information learning module (i.e., a stacked series of SCFFB) and the reconstruction module (U-SCA).
The LR image is first fed to the FEB for shallow feature extraction, which consists of two convolution layers. The FEB can be expressed as:
f 0   = C o n v ( I L R )
f 1 = C o n v ( f 0 )
where C o n v indicates a convolutional layer with kernel 3 for shallow feature extraction and f 1 is the extracted feature.
We then use a series of stacked SCFFB as the nonlinear mapping module for the network in this work to generate a powerful representation of the LR image high-level features f n . We denote the SCFFB as H S C F F B ( · ) , the shallow features f 1 flow through the nonlinear feature mapping module to obtain the high-level features f n , which can be expressed as:
f n = H S C F F B n ( H S C F F B n 1 ( H S C F F B 1 ( f 1 )   )   )
where f n is the output feature map of the n-th SCFFB.
The skip connection is used to supplement the details of the original information to the obtained f n , which can also effectively prevent the gradient from vanishing:
F n = f n + f 0
where, + is element-wise addition, F n is the final feature map of low resolution.
Finally, the NN, the SCA we designed and two convolutional layers as the reconstructed part of SCFFN. We first upsample the feature map F n to the target size through NN and then performed feature enhancement by modeling the feature map channels (SCA). In addition, we employ a global skip connection f U P , and the high-level semantic information learned by the network is summed element-wise to obtain a detail-rich super-resolution image. As a result, we obtain:
I S R = H U P ( F n ) + f U P ( I L R )
where H U P ( · ) is the reconstruction unit in our network, I S R is the high-resolution image after the final reconstruction of our network, and f U P indicates the perform bilinear interpolation operation.
Table 1 shows the parameter settings for our network. “Input” and “Output” denote the input and output flowing through the corresponding network layers, respectively. “Layers” represents the number of corresponding layers. SCFFB is the high-level information extraction block with the number of 12 (discussed in Setion 4 of the article), NN is the nearest neighbour interpolation upsampling method, and “s” is the scaling factor. It is worth mentioning that when s = 4, in order to reduce the serious mosaic and sawtooth phenomenon of the image caused by too large an upsampling factor, we split it into two × 2 upsampling.

3.2. Self-Calibrated Feature Fuse Block

As the core part of the SCFFN, the nonlinear mapping module consists of a series of stacked SCFFBs. SCFFB performs feature crossing to refine features. It first feeds the input feature map into the two branches and then strengthens the common part by element-wise product while increasing the nonlinear capability of the network. By multiple SCFFBs, the boundaries of the high-level features are sharpened. At the same time, we add a local skip connection to SCFFB, which can avoid the gradient disappearance caused by multiple products and compensate for the low-frequency information.
Here, we use f n 1 and f n to denote the input and output of the n-th SCFFB respectively. Similar to SCNet [10], the SCFFB contains two parts. To reduce the complexity of the proposed network, we first reduce the dimension in the upper part by 1 × 1 convolution, FFB then performs feature refinement to enhance the common parts among features, producing a feature map with clear boundaries. The under part is a local residual connection to retain the original details. The SCFFB structure is shown in Figure 3. For the input feature f n 1 , we have:
f n 1 = C o n v ( f n 1 )
f n 1 = C o n v ( f n 1 )
where f n 1 and f n 1 are only half of the number of channels of f n 1 , C o n v means that 1 × 1 convolution layer is used for dimension reduction.
The structure of FFB is shown in Figure 3. The upper part of FFB is about up-down symmetry. The feature f n 1 and f n 1 first pass through the 3 × 3 convolution layer to adapt to the subsequent changes while increasing the nonlinear capability of the network. Then perform an element-by-element multiplication to highlight the high frequency information of the image, to achieve the transformation and fusion of the features. The fused features have clear boundaries and rich semantics. Finally, we use the shortcut to retain the original information of LR to generate the final output feature map f n . The whole process is expressed as:
f f u s e = C o n v ( f n 1 )     C o n v ( f n 1 )
H 1 = C o n v ( f f u s e ) ,   H 2 = C o n v ( f f u s e )
f f u s e = C o n c a t ( [ H 1 , H 2 ] )
f n = f n 1 + C o n v ( f f u s e )
where means the element-wise product, C o n c a t ( [ H 1 ,   H 2 ] ) is the concatenation operation of feature maps H 1 and H 2 , f f u s e and f f u s e are the feature maps generated in the middle.

3.3. U-SCA Block

In the previous work, the reconstruction part of the network is often easily overlooked, because processing the up-sampled image will introduce a lot of parameters and computation. Therefore, the reconstruction module of SR network consists of an upsampling layer (sub-pixel convolution or transposed convolution) and a convolution layer. But the reconstruction part also has a significant impact on the reconstruction performance of the network.
In the reconstruction stage of the network, we choose the simple and fast nearest neighbor (NN) interpolation method, which will also introduce mosaic and sawtooth effects. To alleviate this problem, this paper introduces the attention mechanism. Because convolution layer treats each channel-wise feature equally, it is unfriendly to the feature image after up-sampling by NN. For example, the network should pay attention to areas (edges, contours, etc.) with rich high-frequency information. Therefore, we modelled the interdependence among feature channels, introduce a simple channel attention mechanism (SCA) in the up-sampling phase of the network (the structure is as follows). We expect the learning of high-level features to be enhanced by explicitly modelling channel interdependencies, so that the network is able to alleviate the mosaic and jagged introduced by NN. The network structure of SCA is shown in Figure 4. The experimental results show that the SCA we introduce has a positive effect on the performance of the network, while only a few parameters need to be introduced.
In U-SCA, we first upsample the fine feature map of the nonlinear mapping learning unit to the target size by traditional NN. In order to make the network more effective and have fewer parameters, we then reduce the dimension of the upsampled feature map (i.e., reduce the number of channels) and input it into SCA for information enhancement. Finally, the number of channels is reduced three channels (i.e., RGB) by a convolution layer. The mathematical expression is:
h 1 = C o n v ( H N N ( f n ) )
where H N N   ( · ) represents the nearest neighbour interpolation function, C o n v is a 3 × 3 convolutional layer, while reducing the image dimension from 40 to 24 to ensure the efficiency of the network. h 1 is the feature map after upsampling and dimensionality reduction.
The obtained feature map h 1 is sent to SCA block to enhance the features. Specifically, firstly, through a global average pooling, then through an attention activation (Sigmoid) function layer, we get an attention weight vector α R 1 × 1 × C , where C denotes the number of channels of the weight vector, here C = 24 , and finally apply the weight vector to the feature map by element-wise multiplication and addition to generate a residual map with abundant details. Its expression formula is given by:
α = S i g m i o d ( C o n v ( A v g ( h 1 ) ) )
o u t = α   h 1 + h 1
where A v g ( · ) is the global average pooling function, S i g m i o d ( · ) is the Sigmoid function, and is the element-wise product operation between two feature maps.

3.4. Loss Functions

The loss function is one of the most important parts of deep neural network, which determines the direction of our network optimization. In the SR task, because L1 [32] loss function punishes the relative error of abnormal samples less than the MSE (L2) loss function. Numerous experiments prove that the MSE loss function can produce blurred images, so in our network, the L1 loss function is used to optimize our network. The network loss function formula can be expressed as:
L ( θ ) = 1 N i = 1 N | | H S C F F N ( I L R i ) I H R i | | 1
where H S C F F N ( · ) denotes the network function the proposed in the paper, θ is a learnable parameter in SCFFN network, and | | · | | 1 is the l 1 norm. { I L R i , I H R i } is the training dataset pair, I L R i and I H R i indicate the input LR images and the corresponding ground-truth images respectively, and N represents the batch-size of training datasets.

4. Experiments

In this section, we verify the effectiveness of our method through sufficient experiments and the final results show its superiority. Section 4.1 introduces the proposed network training settings (such as datasets, evaluation indicators and training settings). Section 4.2 analyses each part of our network in detail, such as the impact of the number of SCFFB on performance, the effectiveness of the U-SCA, etc. Section 4.3 compares the proposed method with other algorithms in terms of objective metrics and visualization results.

4.1. Settings

We are using DIV2K [33] containing 800 high quality images as the training dataset. Due to the relative simplicity of the proposed network, we did not perform any data augmentation on the training dataset, but in the image preprocessing, we cut the HR images in the training dataset into small-size images as input to enhance the data. Also, the small size of the image better allows the network to learn local information. We evaluated the SR results of five standard benchmark test datasets under the peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) [29]: Set5 [34], Set14 [35], BSD100 [36], Urban100 [37] and Manga109 [38]. At the same time, our model is also evaluated in the quantitative indicators of cost calculation (Multi-Adds). For a fair comparison, the results of the experimental quantitative analysis of our network, like other networks, were evaluated on the luminance (Y) channel in YCbCr channel.
Consistent with the existing network, we downsampled the ground-truth images in MATLAB using bicubic downsampling to generate LR, where the downsampling factors include ( × 2, × 3, × 4), and the final training dataset pair is formed.
We set the input batchsize to 32 to train our network. Also, to ensure that our network can fully learn the information in LR, we set the patch-size of LR input to 64. Meanwhile, we use Adam [25] and MultiStepLR learning scheme to optimize the network parameters. In Adam, we set β 1 = 0.9 , β 2 = 0.99 , and ϵ = 10 8 , and the initial learning rate is 7 × 10 4 . For every 250,000 iterations, the learning rate was cropped by 0.5, and the total of 1,000,000 iterations are trained. We use Pytorch deep learning framework to implement our algorithm and train it in TITAN RTX.

4.2. Model Analysis

We first explored the number of SCFFB, then verified the effectiveness of U-SCA and compared the effects of other upsampling (such as transpose convolution, etc.). Finally, we qualitatively compare and visualize the proposed methods.

4.2.1. Number of SCFFB Studies

In this section, we analyze the influence of the number of SCFFB on the performance of network through experiments. As the core component of our network, the number of SCFFB affects the final performance of our network to some extent. We should not only consider the performance of the network, but also pay attention to the parameters and computation of the network. As shown in Figure 5 and Table 2, the results show that when n = 12, 13, 14, the reconstruction results of the network are close. We know that the larger the n, the deeper the network, and the more network parameters and computation. The purpose of this paper is to explore the lightweight of SR network. Therefore, when the results are similar, we choose a model with relatively few parameters and calculations, so in this paper, n = 12 is chosen as our final network. It should be noted that SCFFN + (n = i), Parameters, Multi-Adds, PSNR and SSIM in Table 2 represent the corresponding network parameters, complexity and the average PSNR/SSIM of five common datasets on × 4 when the number of SCFFB is i.

4.2.2. Ablation Study

As can be seen from Figure 2 of the network architecture proposed in this paper, our network adds a local residual connection (LRC) (as seen in Equation (4)) to supplement the original rough information of LR and effectively conduct gradient transmission. We can see from Table 3 and Figure 6 that the LRC is critical to the performance improvement of our network. Due to the network is deeper, the weights of the shallow network may not be updated in time during the training process of the network, causing a significant drop in the performance of the network. Therefore, the LRC is essential in our network. SCFFN-LRC indicates removal of the LRC from the SCFFN network (see Equation (4)).
We also made a detailed experimental comparison on the reconstruction part of the network. Firstly, we replace the reconstruction part of SCFFN with NN and two convolutional layers, named Base + NN. Then, like other networks, we use transposed convolution to perform upsampling, where kernel_size = 6/7/8, padding = 2/2/2 and stride = 2/3/4 to achieve ×2/×3/×4 perceptibly, denoted Base + Deconv. At the same time, we also made an experimental comparison of U-SCA, we removed the addition branch in SCA, denoted as Base + NN + (B-add). The results of the ablation experiment are presented in Table 4 and Figure 6, where the deconvolution layer dramatically increases the parameters of the network without increasing the performance of the network, while we find that the Multi-Adds for just one transposed convolution is 94.4 G, thus demonstrating that the reconstruction part of our design improves the performance of the network at a small cost. We also find from Table 4 that the designed SCA also has a positive effect on the network performance improvement. Overall, our well-designed reconstruction part is crucial to our network. It is worth mentioning that the “Base” in Table 4 refers to the network after the reconstruction part is removed by SCFFN, the PSNR/SSIM in the results of Table 3 and Table 4 is the average value evaluated on five common test datasets (×4), the experimental results in Figure 6 are tested in Set5 (×4).

4.2.3. Loss Analysis

In this part, we explore the influence of L1 and L2 loss functions on network performance. The experimental results are shown in Figure 7, the results show that the network performance optimized by L1 loss function is better than that optimized by L2 loss function, so L1 loss function is more suitable for our network. The result is evaluated on Set5 (×2).

4.2.4. Visual Analysis

We have visualized the intermediate feature map of the network. As shown in Figure 8. The first column represents the input image of the network, the second column represents the feature map of the image after shallow feature extraction, the third column shows the feature map after deep feature learning module, and the fourth column represents the features of NN upsampling. The last column shows the features after SCA. From these feature maps, we can find that the shallow feature map contains abundant low-frequency information. After the deep learning module, the extracted feature image retains a large amount of high-frequency information. After NN upsampling, the feature map has obvious mosaic and jaggedness phenomenon. Finally, after our proposed SCA, the high-frequency information of the image is clearly displayed, and at the same time, it can effectively alleviate the shortage of NN upsampling. Experiments show that the method we designed is very effective for lightweight SR.

4.3. Comparison with State-of-the-Arts

4.3.1. Network Parameters

After sufficient training, comparison of our model with state-of-the-art methods on the five public test datasets (see Table 5), including SRCNN [3], FSRCNN [23], VDSR [6], DRCN [7], DRRN [8], MemNet [9], CARN [21], LapSRN [24], SRResNet [27], IDMN [28], MAFFSRN [39], MADNet [40] and SMSR [41]. For a fair comparison, we only consider the models with equivalent Multi-Adds for comparison, and therefore models that were too deep and too large, such as RDN [16] and RCAN [17], were excluded here. According to the convention, we choose PSNR and SSIM [33] as metrics. The comparison results in network parameters, reconstruction effect (PSNR) and Multi-Adds (G) are shown in Figure 1 and Figure 9. Figure 1 shows that our method can balance the parameters, reconstruction performance and Multi-Adds well. It can also be seen from Figure 9a that although the Multi-Adds (27 G vs. 19.3 G) of our network is slightly higher than MAFFSRN in × 4 upscaling factors and can achieve similar performance (in Table 5 shows that we have fewer Multi-Adds on × 2). It can obtain from Figure 9b that the parameters of our network are only half that of MAFFSRN (267 K vs. 441 K). Therefore, compared with other methods, our network is lighter and more efficient. It is worth noting that Multi-Adds are estimated on 720p (1280 × 720) HR image, and Figure 1 shows our method on Set5 ( × 2) compared to other methods, and Figure 9 compare on Set5 ( × 4).

4.3.2. Comparison of Reconstruction Performance and Visual Effects of the Network

In this subsection, we show the quantitative and qualitative results of SCFFN compared with state-of-the-art models (including SRCNN [3], FSRCNN [23], VDSR [6], DRCN [7], DRRN [8], MemNet [9], CARN [21], LapSRN [24], SRResNet [27], IMDN [28], MAFFSRN [39], MADNet [40] and SMSR [41]) on performance comparison on the three upscaling factors × 2, × 3 and × 4. The quantitative results of our network are presented in Table 5, which includes Multi-Adds that show the complexity of the model and parameters. Specifically, CARN has achieved comparable performance to SCFFN, but its parameters are close to 1592 K, about six times that of the proposed method. The parameters of the proposed network in this work are only 37% of IMDN, but comparable results can be achieved. Complete experimental results demonstrate that the proposed lightweight network SCFFN achieves comparable performance to other state-of-the-art methods on multiple datasets and scale factors, but we only need fewer parameters and Multi-Adds. It is worth noting that MAFFSRN is the work from the AIM 2020 Efficient SR Challenge, which ranked the network third and fourth in terms of Multi-Adds and parameters, respectively, but its code is not publicly available.
We selected an image from the Set5, Set14 and Urban100 test datasets respectively for comparison of the visual reconstruction details (shown in Figure 10), we can see that our method is superior to other methods in details, such as stripes. For the image “ppt3” and “Baby”, we observe that most comparison methods will produce obvious artifacts and blurring effects, while our method produce more accurate lines. On the structural details in “img008”, the proposed network in this paper can realize reconstruction with less artifacts.

4.4. Discussion

Through the above ablation research and comparative experiments, we found that the image super-resolution reconstruction has great challenges in terms of trade-offs among network parameters, reconstruction performance and computational complexity, but the SR lightweight network we designed has achieved good results. However, there is room for optimization in our approach. Similar to most SR networks, it is difficult to minimize network parameters, performance and computational complexity at the same time. However, compared with other comparison networks, we only need fewer parameters and computational complexity (Multi-Adds) to achieve considerable performance.

5. Conclusions

In this work, we propose a lightweight network SCFFN for the SR task, in which SCFFB is the basic building block. SCFFB performs feature crossing to refine features. Specifically, the input features are first fed into two branches, and then the common part is strengthened by element-wise multiplication while increasing the nonlinear ability of the network, so that the fused features have the characteristics of clear boundary, etc. At the same time, we add the local skip connection, which not only avoids gradient dispersion caused by multiple multiplications but also supplements low-frequency information. In the reconstruction part of the network, we adopt the traditional nearest neighbor interpolation upsampling and introduce SCA to model the features channel to alleviate the mosaic and sawtooth phenomenon caused by NN. Comprehensive experiments show that the proposed method achieves comparable performance with other advanced methods, but we only need less network parameters and computational complexity.
In the future work, we will continue to explore the lightweight of SR network and try to introduce non-parametric attention mechanism or dynamic convolution layer to enhance information extraction in the high-level information learning stage of the network. In order to design a more effective up-sampling operation for the reconstruction part of the network, we can try to combine the depth separable convolution or group convolution into the transposed convolution layer to reduce the network parameters. At the same time, in the future work, we will apply this work to video SR or introduce it into the real world for real-time broadcasting.

Author Contributions

Conceptualization, C.T.; methodology, L.W.; software, C.T.; validation, S.C. and C.T.; formal analysis, S.C.; investigation, L.W.; resource, L.W.; writing—original draft preparation, C.T.; writing—review and editing, S.C.; visualization, S.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Xinjiang Uygur Autonomous Region grant number 2019D01C033, Tianshan Innovation Team of Xinjiang Uygur Autonomous Region grant number 2020D14044, National Science Foundation of China under Grant U1903213, 61771416 and 62041110, the National Key R&D Program of China under Grant 2018YFB1403202.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. Our training set DIV2k can be obtained from Available online: https://data.vision.ee.ethz.ch/cvl/DIV2K/ (accessed on 18 October 2021). The URLs of test sets Set5, Set14, BSD100, Urban 100 and Manga109 are [Low-Complexity Single-Image Super-Resolution (inria.fr)], Available online: https://sites.google.com/site/romanzeyde/research-interests (accessed on 18 October 2021), Available online: https://www.eecs.berkeley.edu/Research/Projects/CS/vision/bsds/ (accessed on 18 October 2021), Available online: https://sites.google.com/site/jbhuang0604/publications/struct_sr (accessed on 18 October 2021) and Available online: http://www.manga109.org/en/ respectively (accessed on 18 October 2021).

Conflicts of Interest

We declare no conflict of interest.

References

  1. Zhang, L.; Wu, X.L. An edge-guided image interpolation algorithm via directional filtering and data fusion. IEEE Trans. Image Process. 2006, 15, 2226–2238. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Zhang, K.B.; Gao, X.B.; Tao, D.C.; Li, X.L. Single Image Super-Resolution with Non-Local Means and Steering Kernel Regression. IEEE Trans. Image Process. 2012, 21, 4544–4556. [Google Scholar] [CrossRef] [PubMed]
  3. Dong, C.; Loym, C.C.; He, K.M.; Tang, X.O. Learning a Deep Convolutional Network for Image Super-Resolution. In Proceedings of the 13th European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 184–199. [Google Scholar]
  4. Wang, Z.W.; Liu, D.; Yang, J.C.; Han, W.; Thomas, S.H. Deep Networks for Image Super-Resolution with Sparse Prior. In Proceedings of the 2015 IEEE/CVF International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 370–378. [Google Scholar]
  5. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  6. Kim, J.; Lee, J.K.; Lee, K.M. Accurate Image Super-Resolution Using Very Deep Convolutional Networks. In Proceedings of the 2016 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26–30 June 2016; pp. 1646–1654. [Google Scholar]
  7. Kim, J.; Lee, J.K.; Lee, K.M. Deeply-Recursive Convolutional Network for Image Super-Resolution. In Proceedings of the 2016 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26–30 June 2016; pp. 1637–1645. [Google Scholar]
  8. Tai, Y.; Yang, J.; Liu, X.M. Image Super-Resolution via Deep Recursive Residual Network. In Proceedings of the 2017 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2790–2798. [Google Scholar]
  9. Tai, Y.; Yang, J.; Liu, X.M.; Xu, C.Y. MemNet: A Persistent Memory Network for Image Restoration. In Proceedings of the 2017 IEEE/CVF International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4549–4557. [Google Scholar]
  10. Liu, J.J.; Hou, Q.B.; Cheng, M.M.; Wang, C.H.; Feng, J.S. Improving Convolutional Networks with Self-Calibrated Convolutions. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 10093–10102. [Google Scholar]
  11. Voulodimos, A.; Doulamis, N.; Doulamis, A.; Protopapadakis, E. Deep learning for computer vision: A brief review. Comput. Intell. Neurosci. 2018, 2018, 7068349:1–7068349:13. [Google Scholar] [CrossRef] [PubMed]
  12. Garcia-Garcia, A.; Orts-Escolano, S.; Oprea, S.; Villena-Martinez, V.; Garcia-Rodriguez, J. A review on deep learning techniques applied to semantic segmentation. arXiv 2017, arXiv:1704.06857. [Google Scholar]
  13. Tarvainen, A.; Valpola, H. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. arXiv 2017, arXiv:1703.01780. [Google Scholar]
  14. Doulamis, N.; Doulamis, A. Semi-supervised deep learning for object tracking and classification. In Proceedings of the 2014 IEEE International Conference on Image Processing, Paris, France, 27–30 October 2014; pp. 848–852. [Google Scholar]
  15. Baur, C.; Albarqouni, S.; Navab, N. Semi-supervised deep learning for fully convolutional networks. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Quebec City, QC, Canada, 10–14 September 2017; pp. 311–319. [Google Scholar]
  16. Zhang, Y.L.; Tian, Y.P.; Kong, Y.; Zhong, B.N.; Fu, Y. Residual Dense Network for Image Super-Resolution. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2472–2481. [Google Scholar]
  17. Zhang, Y.L.; Li, K.P.; Li, K.; Wang, L.C.; Zhong, B.N.; Fu, Y. Image Super-Resolution Using Very Deep Residual Channel Attention Networks. In Proceedings of the 15th European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 294–310. [Google Scholar]
  18. Hui, Z.; Wang, X.; Gao, X. Fast and Accurate Single Image Super-Resolution via Information Distillation Network. In Proceedings of the 15th European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 723–731. [Google Scholar]
  19. Dai, T.; Cai, J.R.; Zhang, Y.B.; Xia, S.T.; Zhang, L. Second-Order Attention Network for Single Image Super-Resolution. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 11065–11074. [Google Scholar]
  20. Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced Deep Residual Networks for Single Image Super-Resolution. In Proceedings of the 2017 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1132–1140. [Google Scholar]
  21. Ahn, N.; Kang, B.; Sohn, K.A. Fast, Accurate, and Lightweight Super-Resolution with Cascading Residual Network. In Proceedings of the 15th European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 256–272. [Google Scholar]
  22. Dong, C.; Loy, C.C.; Tang, X.O. Accelerating the Super-Resolution Convolutional Neural Network. In Proceedings of the 14th European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 391–407. [Google Scholar]
  23. Shi, W.Z.; Caballero, J.; Huszar, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z.H. Real-Time Single Image and Video Super-Resolution Using an Efficient Sub-Pixel Convolutional Neural Network. In Proceedings of the 2016 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26–30 June 2016; pp. 1874–1883. [Google Scholar]
  24. Ahuja, N.; Lai, W.S.; Huang, J.B.; Yang, M.H. Deep Laplacian Pyramid Networks for Fast and Accurate Super-Resolution. In Proceedings of the 2017 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5835–5843. [Google Scholar]
  25. Ahuja, N.; Lai, W.S.; Huang, J.B.; Yang, M.H. Fast and Accurate Image Super-Resolution with Deep Laplacian Pyramid Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 2599–2613. [Google Scholar]
  26. Perazzi, F.; Wang, Y.F.; McWilliams, B.; Hornung, B.; Hornung, A.S.; Schroers, C. A Fully Progressive Approach to Single-Image Super-Resolution. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Munich, Germany, 8–14 September 2018; pp. 864–873. [Google Scholar]
  27. Ledig, C.; Theis, L.; Huszar, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.P.; Tejani, A.; Totz, J.; Wang, Z.H.; et al. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. In Proceedings of the 2017 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 105–114. [Google Scholar]
  28. Hui, Z.; Gao, X.B.; Yang, Y.C.; Wang, X.M. Lightweight Image Super-Resolution with Information Multi-distillation Network. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 2024–2032. [Google Scholar]
  29. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Jiang, X.; Zhao, C.; Zhu, M.; Hao, Z.; Gao, W. Residual Spatial and Channel Attention Networks for Single Image Dehazing. Sensors 2021, 21, 7922. [Google Scholar] [CrossRef] [PubMed]
  31. Zhu, T.; Luo, W.; Yu, F. Multi-Branch Convolutional Neural Network for Automatic Sleep Stage Classification with Embedded Stage Refinement and Residual Attention Channel Fusion. Sensors 2020, 20, 6592. [Google Scholar] [CrossRef] [PubMed]
  32. Dong, C.; Loy, C.C.; He, K.M.; Tang, X.O. Image Super-Resolution Using Deep Convolutional Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 295–307. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Agustsson, E.; Timofte, R. NTIRE 2017 Challenge on Single Image Super-Resolution: Dataset and Study. In Proceedings of the 2017 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1122–1131. [Google Scholar]
  34. Bevilacqua, M.; Roumy, A.; Guillemot, C.; Alberi-Morel, M.L. Low-Complexity Single-Image Super-Resolution based on Nonnegative Neighbor Embedding. In Proceedings of the Electronic British Machine Vision Conference 2012 (BMVC), Guildford, UK, 3–7 September 2012; pp. 1–10. [Google Scholar]
  35. Yang, J.C.; Wright, J.; Huang, T.S.; Ma, Y. Image Super-Resolution Via Sparse Representation. IEEE Trans. Image Process. 2010, 19, 2861–2873. [Google Scholar] [CrossRef] [PubMed]
  36. Martin, D.R.; Fowlkes, C.C.; Tal, D.; Malik, J. A Database of Human Segmented Natural Images and its Application to Evaluating Segmentation Algorithms and Measuring Ecological Statistics. In Proceedings of the 2001 IEEE/CVF International Conference on Computer Vision, Vancouver, BC, Canada, 7–14 July 2001; pp. 416–425. [Google Scholar]
  37. Huang, J.B.; Singh, A.; Ahuja, N. Single image super-resolution from transformed self-exemplars. In Proceedings of the 2015 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 8–10 June 2015; pp. 5197–5206. [Google Scholar]
  38. Matsui, Y.; Ito, K.; Aramaki, Y.; Fujimoto, A.; Ogawa, T.; Yamasaki, T.; Aizawa, K. Sketch-based manga retrieval using manga109 dataset. Multimed. Tools Appl. 2017, 76, 21811–21838. [Google Scholar] [CrossRef] [Green Version]
  39. Muqeet, A.; Hwang, J.; Yang, S.B.; Kang, J.H.; Kim, K.W.; Bae, S.H. Ultra Lightweight Image Super-Resolution with Multi-Attention Layers. arXiv 2020, arXiv:2008.12912. [Google Scholar]
  40. Lan, R.; Sun, L.; Liu, Z.B.; Lu, H.; Pang, C.; Luo, X. MADNet: A Fast and Lightweight Network for Single-Image Super Resolution. IEEE Trans. Cybern. 2021, 51, 1443–1453. [Google Scholar] [CrossRef] [PubMed]
  41. Wang, L.; Dong, X.; Wang, Y.; Ying, X.; Lin, Z.; An, W.; Guo, Y. Exploring Sparsity in Image Super-Resolution for Efficient Inference. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 19–25 June 2021; pp. 4917–4926. [Google Scholar]
Figure 1. Performance comparison of our algorithm with other lightweight algorithms on Set5 ( × 2), where the horizontal coordinates represent the computational complexity of the network (Multi-Adds), the vertical coordinates represent the PSNR of the network and the size of the circles in the figure indicates the network parameters. Multi-Adds are computed on 720p HR images.
Figure 1. Performance comparison of our algorithm with other lightweight algorithms on Set5 ( × 2), where the horizontal coordinates represent the computational complexity of the network (Multi-Adds), the vertical coordinates represent the PSNR of the network and the size of the circles in the figure indicates the network parameters. Multi-Adds are computed on 720p HR images.
Sensors 22 00329 g001
Figure 2. The proposed network structure of SCFFN.
Figure 2. The proposed network structure of SCFFN.
Sensors 22 00329 g002
Figure 3. The proposed SCFFB, as the critical part of our network, for high-level semantic information extraction of LR images.
Figure 3. The proposed SCFFB, as the critical part of our network, for high-level semantic information extraction of LR images.
Sensors 22 00329 g003
Figure 4. The simple channel attention (SCA) mechanism block of the network reconstruction part.
Figure 4. The simple channel attention (SCA) mechanism block of the network reconstruction part.
Sensors 22 00329 g004
Figure 5. Number n of SCFFB vs. PSNR analysis, the result is evaluated on Set5 ( × 4).
Figure 5. Number n of SCFFB vs. PSNR analysis, the result is evaluated on Set5 ( × 4).
Sensors 22 00329 g005
Figure 6. Ablation experiment analysis, the result was evaluated in Set5 ( × 4).
Figure 6. Ablation experiment analysis, the result was evaluated in Set5 ( × 4).
Sensors 22 00329 g006
Figure 7. Analysis of L1 and L2 Loss Functions. (a) is the PSNR value evaluated on Set5 (×2), (b) is the SSIM value evaluated on Set5 ( × 2).
Figure 7. Analysis of L1 and L2 Loss Functions. (a) is the PSNR value evaluated on Set5 (×2), (b) is the SSIM value evaluated on Set5 ( × 2).
Sensors 22 00329 g007
Figure 8. Visual feature maps.
Figure 8. Visual feature maps.
Sensors 22 00329 g008
Figure 9. Comparison of our method with other methods in terms of Multi-Adds and Parameters. (a) PSNR vs. Multi-Adds. (b) PSNR vs. Parameters. The results shown are for experiments conducted on Set5 ( × 4).
Figure 9. Comparison of our method with other methods in terms of Multi-Adds and Parameters. (a) PSNR vs. Multi-Adds. (b) PSNR vs. Parameters. The results shown are for experiments conducted on Set5 ( × 4).
Sensors 22 00329 g009
Figure 10. Comparison of the visual effect of the method in this paper with other methods on × 4.
Figure 10. Comparison of the visual effect of the method in this paper with other methods on × 4.
Sensors 22 00329 g010
Table 1. Setting of network structure parameters of our algorithm.
Table 1. Setting of network structure parameters of our algorithm.
ModuleLayer NameLayersKernel SizeInputOutput
Shallow feature
extraction
Conv1 3 × 3 3 ×   64   × 64 40 ×   64 × 64
Conv1 1 × 1 40 ×   64 × 64 40 ×   64 × 64
Deep feature extractionSCFFB12 1 × 1 40 ×   64 × 64 20 ×   64 × 64
1 × 1 40 ×   64 × 64 20 ×   64 × 64
3 × 3 20 ×   64 × 64 20 ×   64 × 64
3 × 3 20 ×   64 × 64 20   ×   64 × 64
3 × 3 20 ×   64 × 64 20 ×   64 × 64
3 × 3 20 ×   64 × 64 20   ×   64 × 64
3 × 3 40 ×   64 × 64 40 ×   64 × 64
Conv1 3 × 3 40 ×   64 × 64 40 ×   64 × 64
Network reconstructionNN1 or 2 40 ×   64 × 64 40 ×   ( s · 64 ) ×   ( s · 64 )
Conv1 or 2 3 × 3 40 ×   64 × 64 24 ×   ( s · 64 ) ×   ( s · 64 )
SCA1 or 2 3 × 3 24 ×   64 × 64 24 ×   ( s · 64 ) ×   ( s · 64 )
Conv1 or 2 3 × 3 24 ×   64 × 64 3 ×   ( s · 64 ) ×   ( s · 64 )
Table 2. Performance analysis of the number of SCFFB.
Table 2. Performance analysis of the number of SCFFB.
MethodParametersMulti-AddsPSNR (dB)SSIM
SCFFN + (n = 9)214 K24.2 G28.860.8190
SCFFN + (n = 10)232 K25.2 G28.870.8188
SCFFN + (n = 11)249 K26.2 G28.850.8190
SCFFN + (n = 12)267 K27.2 G28.930.8203
SCFFN + (n = 13)285 K28.2 G28.910.8199
SCFFN + (n = 14)302 K29.2 G28.920.8202
Table 3. Regarding the effect of LRC in Equation (4) on performance results, we test results on × 4 and PSNR, SSIM is the average of the test results on the five public datasets.
Table 3. Regarding the effect of LRC in Equation (4) on performance results, we test results on × 4 and PSNR, SSIM is the average of the test results on the five public datasets.
MethodParametersMulti-AddsPSNR (dB)SSIM
SCFFN-LRC267 K27.2 G27.900.8023
SCFFN267 K27.2 G28.930.8203
Table 4. Experimental analysis on the reconstruction part of the network. We test results on × 4 and PSNR, SSIM is the average of the test results on the five public datasets.
Table 4. Experimental analysis on the reconstruction part of the network. We test results on × 4 and PSNR, SSIM is the average of the test results on the five public datasets.
MethodParametersMulti-AddsPSNR (dB)SSIM
Base + NN266 K27.2 G28.860.8188
Base + Deconv353 K116.8 G28.860.8191
Base + NN + (SCA-add)267 K27.2 G28.890.8194
SCFFN (ours)267 K27.2 G28.930.8203
Table 5. Comparison of the results of the proposed algorithm with the state-of-the-art models on × 2, × 3, × 4.
Table 5. Comparison of the results of the proposed algorithm with the state-of-the-art models on × 2, × 3, × 4.
ScaleMethodParams (K)Multi-Adds (G)Set5
PSNR/SSIM
Set14
PSNR/SSIM
BSD100
PSNR/SSIM
Urban100
PSNR/SSIM
Manga109
PSNR/SSIM
2SRCNN5752.736.66/0.954232.45/0.906731.36/0.887929.50/0.894635.60/0.9663
FSRCNN136.037.00/0.955832.63/0.908831.53/0.892029.88/0.902036.67/0.9710
VDSR66661237.53/0.958733.03/0.912431.90/0.896030.76/0.914037.22/0.9750
DRCN177417,97437.63/0.958833.04/0.911831.85/0.894230.75/0.913337.55/0.9732
LapSRN25129.937.52/0.959132.99/0.912431.80/0.895230.41/0.910337.27/0.9740
DRRN2986796.937.74/0.959133.23/0.913632.05/0.897331.23/0.918837.88/0.9749
MemNet6782662.437.78/0.959733.28/0.914232.08/0.897831.31/0.919537.72/0.9740
CARN1592222.837.76/0.959033.52/0.916632.09/0.897831.92/0.925638.36/0.9765
SRResNet1518146.138.05/0.960733.64/0.917832.22/0.900232.23/0.929538.05/0.9607
IMDN694159.637.91/0.959433.59/0.916932.15/0.898732.14/0.927438.79/0.9764
MAFFSRN40277.237.97/0.960333.49/0.917032.14/0.899431.96/0.9268-\-
MADNet878178.137.94/0.960433.46/0.916732.10/0.898831.74/0.9246-\-
SMSR985131.638.00/0.960133.64/0.917932.17/0.899032.19/0.928438.76/0.9771
SCFFN(ours)2566838.01/0.960433.52/0.916932.12/0.899031.93/0.926138.51/0.9768
3SRCNN5752.732.75/0.909029.30/0.821528.41/0.786326.24/0.798930.48/0.9117
FSRCNN13K5.033.18/0.914029.37/0.824028.53/0.791026.43/0.808031.10/0.9210
VDSR66661233.66/0.921329.77/0.831428.82/0.797627.14/0.827932.01/0.9340
DRCN177417,974.333.82/0.922629.76/0.831128.80/0.796327.15/0.827632.24/0.9343
DRRN2986796.934.03/0.924429.96/0.834928.95/0.800427.53/0.837832.71/0.9179
MemNet6782662.434.09/0.924830.00/0.835028.96/0.800127.56/0.837632.51/0.9369
CARN1592118.834.29/0.925530.29/0.840729.06/0.803428.06/0.849333.50/0.9440
SRResNet1554190.234.41/0.927430.36/0.842729.11/0.805528.20/0.853533.54/0.9448
IMDN70371.734.32/0.925930.31/0.840929.07/0.803628.15/0.851033.58/0.9434
MAFFSRN41834.234.32/0.926930.35/0.842929.09/0.805228.13/0.8521-\-
MADNet93088.434.26/0.926230.29/0.841029.04/0.803327.91/0.8464-\-
SMSR993100.534.40/0.927030.33/0.841229.10/0.805028.25/0.853633.68/0.9445
SCFFN(ours)2563734.29/0.926330.27/0.840929.04/0.803427.98/0.848133.30/0.9427
4SRCNN5752.730.48/0.862827.49/0.750326.90/0.710124.52/0.722127.66/0.8505
FSRCNN134.630.71/0.865727.59/0.753526.98/0.710524.62/0.720827.90/0.8517
VDSR665612.631.35/0.883828.01/0.767427.29/0.725125.18/0.752428.83/0.8809
DRCN177417,976.331.53/0.885428.02/0.767027.23/0.723325.14/0.751028.98/0.8816
LapSRN813149.431.54/0.885029.19/0.772027.32/0.728025.21/0.756029.09/0.8845
DRRN2976796.931.68/0.888828.21/0.772027.38/0.728425.44/0.763829.46/0.8960
MemNet6772662.431.74/0.889328.26/0.772327.40/0.728125.50/0.763029.42/0.8942
CARN1592222.832.13/0.893728.60/0.780627.58/0.734926.07/0.783730.47/0.9084
SRResNet1518146.132.17/0.895128.61/0.782327.59/0.736526.12/0.787130.48/0.9087
IMDN71541.132.19/0.893628.57/0.780327.54/0.734226.03/0.782930.44/0.9065
MAFFSRN44119.332.18/0.894828.58/0.781227.57/0.736126.04/0.7848-\-
MADNet100254.132.11/0.893928.52/0.779927.52/0.734025.89/0.7782-\-
SMSR100657.232.12/0.893228.55/0.780827.55/0.735126.11/0.786830.54/0.9085
SCFFN(ours)2672732.18/0.895028.56/0.780927.54/0.735226.01/0.783230.36/0.9070
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tan, C.; Cheng, S.; Wang, L. Efficient Image Super-Resolution via Self-Calibrated Feature Fuse. Sensors 2022, 22, 329. https://doi.org/10.3390/s22010329

AMA Style

Tan C, Cheng S, Wang L. Efficient Image Super-Resolution via Self-Calibrated Feature Fuse. Sensors. 2022; 22(1):329. https://doi.org/10.3390/s22010329

Chicago/Turabian Style

Tan, Congming, Shuli Cheng, and Liejun Wang. 2022. "Efficient Image Super-Resolution via Self-Calibrated Feature Fuse" Sensors 22, no. 1: 329. https://doi.org/10.3390/s22010329

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop