1. Introduction
Ghost imaging (GI) [
1] is an imaging technique that produces the object image via the intensity correlation of two beams: one beam interacts with the target but is collected by a bucket detector without spatial resolution (only recording the total intensity); the other beam records the spatial distribution of the light field of the source but never interacts with the target. The detector in either beam cannot “see” the object by itself, but the object image can be recovered by calculating the intensity correlation between these two arms’ signals [
2,
3,
4]. Later, it was found that the reference arm containing the spatially resolved array detector can be removed by using a programmable spatial light modulator (SLM) encoded with preset modulated patterns [
5,
6].
Over the last two decades, GI has attracted a lot of attention [
7,
8,
9,
10,
11] and has been used in many fields, such as microscopic imaging [
12], optical encryption [
13], and cryptographic key distribution [
14]. However, the signal-to-noise ratios (SNRs) of conventional GI methods are extremely low, even under a large number of measurements, and increasing the number of measurements only provides a limited improvement in image quality. To improve the image quality and imaging efficiency of the GI, many reconstruction algorithms have been proposed, such as background-removal GI (BGI) [
15], high-order GI (HGI) [
16,
17], differential GI (DGI) [
18], pseudo-inverse GI (PGI) [
19], compressive GI [
20,
21,
22], adaptive compressive GI [
23], iterative GI [
24,
25,
26], Gerchberg–Saxton-like GI [
27], iterative eigenmode GI [
28], joint iteration GI [
29], singular value decomposition GI [
30], etc. However, these methods have their limitations. The BGI finds it hard to deal with quasitransparent complex objects; the HGI relies on the increase in exponential power to improve visibility; the DGI can provide a considerable but relatively limited SNR improvement, which is mainly dependent on the object [
27,
31]; the PGI generally takes a long time to reconstruct large-scale object images and is sensitive to noise. The compressive GI algorithms require the sparse priors of objects and take huge matrix calculations; the adaptive compressive GI aims to reduce the sampling rate from the perspective of the object’s hierarchical structure, but has high requirements for modulated pattern design. Differing from the compressive GI method and its variants, the iterative GI, Gerchberg–Saxton-like GI, iterative eigenmode GI, joint iteration GI methods do not need to rely on the sparse priors of images, and can finish the image reconstruction task only using iterations based on statistical averages, without large-scale matrix multiplication, which undoubtedly greatly reduces computational consumption. However, these iterative variants of GI either have strict constraints or make some loose approximations of the expression of noise. In addition to these, researchers have made other attempts, such as singular value decomposition GI, but still need time-consuming calculations. Therefore, it is very necessary to construct a high-efficiency, high-quality image reconstruction method.
In this paper, inspired by the idea of classic gradient descent algorithms [
32,
33,
34,
35], we propose a new image reconstruction method, called gradient-descent-like ghost imaging (GGI). It is worth noting that the aim of the gradient descent method is to find the solution, which minimizes the objective function by updating the parameters in the opposite direction of the gradient of the objective function [
32]. In addition, in view of the previous analysis, the correlation function (statistical average function) has a natural advantage as an iterative carrier, and we also know that, in GI, any form of intensity correlation functions can be regarded as a transformation function performed on the original image. Based on the above ideas, we gradually search for the optimal solution of the rewritten objective function to acquire high-quality image reconstruction. Both simulation and optical experiments will be performed to verify the performance of this proposed method against the noise without any sparse prior knowledge of the object images. The performance of this method in the recovery of complex object images will also be investigated.
2. Principle of Gradient-Descent-like Ghost Imaging
In this section, we will first recall the theory of gradient descent, then briefly review the common intensity correlation functions in GI and derive their matrix expression forms, and finally introduce the principle of our GGI.
2.1. Gradient Descent Theory
To our knowledge, gradient descent (also known as steepest descent) is a popular strategy, which is widely used in machine learning and deep learning to solve both convex and non-convex problems. Its idea is to gradually minimize an objective function, parameterized by a model’s parameters through iterations along the opposite direction of the gradient of objective function [
32].
Let J be a function with respect to an independent vector , denoted as , and the gradient at can be written as , where ∇ denotes the gradient operator. The opposite direction of the gradient direction is also called the gradient flow of the variable . Assume , where is a positive scalar, is the unit vector of , defined as . Then we can obtain , which is the core iterative expression of gradient descent. To obtain the minimum value of , the process of gradient descent can be performed via the following steps:
- (1)
Compute the current gradient (partial derivative) with respect to ;
- (2)
Multiply the current gradient by a step size (i.e., learning rate) , i.e., , and update the variable via ;
- (3)
Repeat Steps (1–2) until the difference between the values obtained from two adjacent iterations is small enough (less than the preset termination threshold ); at this moment, the objective function reaches its minimum;
- (4)
Output the current independent variable , which is exactly the value that minimizes the function .
For a linear measurement model, there is a hypothesis function
where
(
) is the model parameter to be evaluated,
is the coefficient or weight. To evaluate the fitting of the algorithm, a loss function can generally be used. Minimizing the loss function will help us acquire the best fitting, and the corresponding model parameters are the optimal solution. In linear regression, the loss function is usually the square of the difference between the hypothesis function and the sample output. For the convenience of derivation, we use a loss function defined as half of the mean square error (MSE):
where
denotes the
jth actual measured value. Then, we calculate the partial derivative of
J with respect to
:
Thus, the iterative expression can be rewritten as
which is also called batch gradient descent [
32] because it uses the gradient data of all samples when calculating the gradient.
Since each
(
) can also be a column vector, the hypothesis function can be written as
, where
is a column vector of
consisting of expected measurements,
is a column vector of
to be reconstructed, and
A is a matrix of
, which can be written as
. Next, we will introduce the matrix representation of gradient descent. The foregoing loss function can be rewritten in matrix form as
where
is a column vector of
, which consists of actual measurements (sample outputs), and
T denotes the transposition operator. Then, the partial derivative of
with respect to
can be computed via
Thus, the updated expression of
can be rewritten as
Figure 1a shows the schematic diagram of gradient descent. The minimum value of
can be obtained by iterating along the direction of gradient descent. First, it is worth mentioning that the value
represents the length of each step in the gradient direction during iterations. If the value
is too large, we cannot guarantee that the gradient will be decreased in each iteration, nor can we guarantee the convergence. If the value
is too small, it will lead to a painfully slow convergence and long calculation time, but the update value can achieve almost an optimal solution to the objective function, as it is unlikely that the stepping will miss any useful gradient position. Thus, this value
determines the convergence speed of the iterations and whether the iterations can reach the optimal solution. Second, if the function
is convex, the iterative result starting from one initial value will be the optimal solution by a large probability; if
is non-convex, it is necessary to perform the gradient descent strategy multiple times, each with different initial values, to get rid of the local optimal solution, and then the solution with the smallest functional value should be selected from these iterative results. Generally, different initial values may lead to different minimum values of the objective function. Therefore, we should carefully choose the initial value of iterations, as well as the iteration step size.
2.2. Intensity Correlation Functions in GI
In GI, the object
is illuminated by several modulated patterns
, where the superscript
denotes the
jth modulation,
x and
y stand for the spatial coordinates on
x-axis and
y-axis. The total intensity collected by the bucket detector can be written as
. The object image can be retrieved by calculating the intensity correlation between the patterns
and bucket values
. In the following, we will briefly introduce BGI, HGI, DGI, logarithmic GI (LGI), trigonometric ghost imaging (TGI) [
36].
First of all, the most classic second-order correlation in GI can be written as
where
stands for the ensemble average of the signal
u. It is worth mentioning that this classic second-order correlation function is a prototype algorithm of GI. Obviously, this function is too simple to obtain a satisfactory image quality, even with a large number of measurements, and the reconstructed image contains non-negligible background noise (which can be treated as the direct component). Therefore, this formula has basically withdrawn from history, but its derivatives have gradually become the mainstream algorithms.
For example, based on the above formula, by making
and
separately subtract the average terms of
and
, we can acquire the functional form of the BGI [
15]:
Thus, this formula uses
to describe the background noise. By subtracting this product term from
, the BGI can generate a good-quality image for simple objects with high transmittance, but fails to work in the case of unstable light sources (e.g., with temperature drift) or quasitransparent complex objects.
By calculating high-order correlation [
16], we can obtain
where
p and
q are the power indices of the bucket and reference signals, respectively. The HGI can improve the visibility and contrast by selecting appropriate
p and
q values, but it cannot effectively remove background noise.
By replacing
in BGI with a new differential term
(
), which reflects the relative fluctuations in the bucket values, we will obtain the functional form of the DGI [
18]:
Compared with the BGI, the term
in the DGI can better describe the background noise. Therefore, the DGI can more accurately remove background noise, and significantly improve the image quality, even in harsh or noisy measurement environments.
The expression of the LGI is defined [
36] as follows:
where
C is the base of the logarithm function. According to the logarithmic operational rule, the above formula can also be rewritten as
; thus, some part of background noise is removed.
The formula of the TGI [
36] can be written as
where
,
and
are the maximum and minimum of
, respectively. Here,
d takes even integers to generate positive images [
36]. Since the TGI only performs a triangular transformation on
on the basis of the classic second-order correlation function, the reconstruction quality obtained via TGI is not very different from the recovered result of the classic second-order correlation function.
We find that Equations (
8)–(
13) can be divided into two categories: functions without subtractive background terms and functions with subtractive background terms. The first category, containing the classic second-order correlation function, HGI and TGI, mainly focuses on enhancing the image visibility by optimizing bucket values, modulated patterns or both; the second category, including the BGI, DGI and LGI, tries to characterize the background noise and subtracts it from the reconstructed image of the classic second-order correlation function to improve the image quality. Among them, the DGI is recognized as the best statistical correlation algorithm due to its excellent and stable imaging performance.
To simplify the derivation, we will also derive the matrix forms of these functions. Each modulated pattern can be reshaped into a row vector of
,
m row vectors reshaped from
m patterns will form a measurement matrix
A of
. In the same way, the object image
can be flattened into a column vector
O of
. Thus, the ideal bucket values can be written as
, and the actual bucket values will be denoted by
. In the ideal measurement environment,
. Then, we will have
,
,
,
, where
E denotes a matrix consisting of all ones and its subscripts stand for the dimensions of this matrix. In the following, we can rewrite the above intensity correlation functions as
where
and
denote the powers that are performed on each element of the matrix,
characterizes the average light field,
, and
. Although the above derivations are carried out under noise-free conditions, the actual recorded bucket values are generally accompanied by the scaling of light intensities as well as measurement noise
e, i.e.,
, where
denotes the proportional coefficient determined by the light attenuation, collection efficiency, photoelectric conversion efficiency, etc. However, this does not affect the above derivations because, for
, where
, we only need to replace original
O with
.
2.3. Gradient-Descent-like Ghost Imaging
Next, let us first review the composition of the original image O. Generally, the image O can be considered to be composed of the object part and background part (a direct component), i.e., (the values in both and are greater than 0); thus, we will have . According to the previous classification, it is easy to notice that , , are in forms of . In terms of GI principle, we expect to make the element values of this subtractive term as close to those of as possible. The closer these two sets of values are, the higher the reconstruction quality of the intensity correlation function is. We find that the form of looks very similar to the gradient form of the traditional gradient descent algorithm. If we regard the above three intensity correlation functions as the image gradients , then their primitive functions can be written as the transform forms of . happens to be the gradient of the loss function without any transformation. Different from the loss function (e.g., ) in traditional gradient descent algorithm, here, is to make as close to a direct component (rather than ) as possible so that will reach its minimum near the optimal solution (while in classic gradient descent algorithm, is to make Y as close to as possible so that the residual error can be minimized; thus, both and its gradient will vanish to 0 near the optimal solution). However, it does not affect the correlation functions being regarded here as the gradients, because, no matter whether or is used, the ultimate goal is to minimize the objective function by reducing the gradient value. Following this idea, tries to characterize the background noise, so intends to extract the object part, and the values of are expected to be continuously reduced to make the reconstructed object part closer to the optimal solution. and can be expressed as the transform forms of (), namely, to make the subtractive term in equal to zero, but with relatively poor imaging performance. Therefore, most of the intensity correlation functions (especially the ones with subtractive background terms) can be regarded as the gradients that are expected to be minimized.
According to the above theoretical analysis and proof, we propose a GGI method. It has been proven that, no matter what correlation function we use to recover the object image, the probability of recovered pixel values located in the pixel region of the same original gray value of the object follows a Gaussian distribution, whose average value has a linear relationship with this given original gray value [
31,
37,
38]. For this reason, all intensity correlation algorithms can be regarded as the functions
F of the object image
, denoted as
. Here, we let
(with respect to the current estimate image
O) represent the above intensity correlation functions.
can be regarded as the gradient of
, i.e., the partial derivative
. Since the original object image to be recovered is generally unknown in advance,
will fail to work because its change cannot be observed intuitively. Fortunately, there are many equivalent loss functions available as alternatives to assert the convergence of the iterations in a direct way, such as
,
,
,
and no-reference image quality assessment, where TV is short for total variation. Here, we choose to use Tenengrad (TNG) function [
39,
40] (a similar method to the TV norm) as the algorithm’s termination criterion, which calculates the sum of the squares of all pixels’ gradient values in a current estimate that are greater than a certain threshold, and is defined [
39] as
where
is the preset threshold. In this function,
is the gradient, which can be written as
where
and
are the gradients along
x and
y axes, respectively, which can be obtained by the well-known Sobel (discrete differentiation) operator. Generally, the sharper the edge of the natural image, the higher the gradient value (before computing the value of TNG, the column vector
O needs to be reshaped into a matrix form). In its implementation, we only need to compare the difference between the values of the TNG function obtained by two adjacent iterations to determine whether the algorithm has converged.
Figure 1b gives the algorithm flow chart of the proposed GGI, and its algorithm steps can be described as follows:
- (1)
Compute the current with respect to the current estimate O;
- (2)
Let
O minus
times the current estimate obtained in Step (1) to produce an updated estimate
:
- (3)
Judge whether the difference between and is greater than a predefined threshold (which is set to in this work): if not, the algorithm will be terminated, the current estimate O is the optimal solution, which will then be reshaped to an image matrix U; if so, replace O with and repeat Steps (1–2);
- (4)
Output the reconstructed image U.
It should be noted that the main iterative formula as given in Equation (
22) is only related to the current estimate
O (a possible solution to the recovered ghost image), predefined threshold
and newly computed function
, while intensity correlation functions depend on the current bucket values and modulated patterns. This means that the function
in Equation (
22) can be replaced with intensity correlation functions, with the inputs serving as the modulated patterns and the new bucket values calculated with the current estimate
O. In other words, in each iteration, the modulated patterns remain unchanged, while the bucket values are updated according to the current estimate
O. Therefore, the bucket values used in the first iteration are obtained from actual measurements, while the ones used in the subsequent iterations are numerically calculated. In addition, since the intensity correlation function
, loss function
and its equivalent TNG function are all convex, by following the direction of the slope of the convex surface created by the loss function downhill, the iterations that start from some initial values tend to reach a valley, and the (local) optimal solution will be acquired after multiple iterations by a very large probability according to the theory of gradient descent. The iterative formulas that use intensity correlation functions with subtractive background terms will definitely achieve better convergence than the cases using intensity correlation functions without subtractive background terms, because the gradient functions in the former cases will tend towards the minimum near the optimal solution and are more conducive to the convergence of iterations.
Next, for the sake of simplification, we abbreviate the gradient-descent-like BGI, gradient-descent-like HGI, gradient-descent-like DGI, gradient-descent-like LGI and gradient-descent-like TGI as the GBGI, GHGI, GDGI, GLGI, GTGI, respectively.
3. Simulation Results and Analysis
To demonstrate the performance of the proposed method, we used random binary patterns for numerical simulations. Here, we set and for the HGI function, as a large p will increase the image visibility and a small q will dramatically suppress the noise. For LGI, the base of the logarithmic function is set to , while, in TGI, we set .
Figure 2a shows the binary image labelled “01” consisting of
pixels. The DGI, GBGI, GHGI, GDGI, GLGI and GTGI are all computed with 20,480 measurements and the step size
is set at 0.1 (which will be explained below). It should be noted that all these reconstruction algorithms are based on statistical intensity correlation, which means that oversampling is generally required in classic GI functions, but the large-scale matrix multiplication and prior knowledge of image sparsity are no longer needed, so the computational complexity is very low. Although oversampling is not as efficient as a simple scanning, GI, as an indirect computational imaging technique, only needs to record the total light intensity, which can effectively avoid the average distribution of luminous flux across dimensions, as in point scanning. Especially under ultra-low-light illumination, this advantage will become more prominent. In addition, GI has a better anti-turbulence and a better imaging sensitivity than scanning. Since the focus of this manuscript is more on how to improve the imaging quality of the intensity correlation functions through only a few iterations, the number of measurements used here (20,480) is just an example to illustrate the performance improvement, and the subsampling cases will be discussed later. The reconstruction results of the above algorithms are given in
Figure 2b–g, with their SNRs and running time
t marked right below the figures. The algorithms are running under the operating system Windows 10, with the memory of 16.0 GB and the central processor AMD Ryzen 7 4800H of 2.90 GHz integrated with Radeon Graphics. It can clearly be seen from the simulation results that the SNRs of the GBGI, GDGI, and GLGI are higher than those of the GHGI and GTGI (consistent with our previous theoretical analysis), which means that they gradually approach the position of the optimal solution after a few iterations. We note that the SNR of the GHGI is slightly lower than that of DGI, because the high-order operation of the bucket values amplifies the noise while enhancing the signal. The GTGI also fails to acquire the performance gain, mainly due to the absence of subtractive background term. It is worth noting that, among these, the SNR of the GDGI can reach 9.60, which is several times that of the DGI. The main reason the GDGI can outperform other variants is that the DGI function expression it uses happens to be
, whose primitive function is exactly the loss function
without any transformation. Besides this, the calculation time of the GDGI is 3.57 s, with only a small increase in computing time compared to the DGI, which is acceptable in most cases. Therefore, the GDGI can achieve a much better reconstruction performance without much increase in computing time compared with the DGI.
Figure 3a shows a
grayscale image labelled “02”, whose gray value range is [0, 1]. With the same number of measurements, the above algorithms are executed again. Here, we set
. The results are presented in
Figure 3b–g. The SNR and calculation time of the DGI is 3.86 and 2.59 s, respectively. In the recovered image of the DGI, the contour of the object is recognizable, but the background noise is messy and the detailed information is missing. As we can see, the GDGI acquires the best performance, with an SNR value of up to 32.70. The image retrieved by the GDGI contains only a small amount of noise and shows a great ability to resolve the details. In addition, the SNRs of the GBGI and GLGI are more than three times that of the DGI, while the GHGI’s performance was the worst (as it hardly acquires any useful object information) among these GGI variants due to the aforementioned HGI restrictions. The calculation time of the above methods is very close, with all less than 4 s, which further indicates that the GDGI can obtain high-quality reconstructed images, with the noise being well-suppressed at the expense of a negligible increase in computing time.
As mentioned above, the gradient descent results are greatly influenced by the step size
. Given that the GDGI has the best performance among the above variant algorithms, without loss of generality, we will mainly focus on the GDGI algorithm and discuss the relationship between the SNR and
value in depth. Take the binary image “01” and grayscale image “02” ranging from 0 to 1, as shown in
Figure 2a and
Figure 3a for examples, we drew the SNR curves of the GDGI as functions of the step size
as well as the calculation time
t for the original images “01” and “02”, respectively, as shown in
Figure 4a–d. The step size
was set from 0.005 to 0.5 with a 0.005 stepping increase, and, for each step size, both the SNR and calculation time were recorded.
As can be seen from
Figure 4a,b, with the increase in the step size
, the SNRs of the reconstructed images of “01” present an oscillating downward trend, and the calculation time
t decreases sharply and tends to be stable when
. As we know, when the step size
is small, the gradient descend stepping distance of the each iteration is too small, so the convergence speed of iterations is slow and the calculation time is long. However, in this case, the algorithm is unlikely to miss the optimal solution in its iterative process, leading to a relatively high SNR with a high probability. When the step size
becomes large, the iterative convergence speed increases rapidly, but the iteration precision will drop, and there is a high probability that the optimal solution will be missed, causing a relatively low SNR. Furthermore, for some large
values, the current step of the iterations will cross the position of the optimal solution, which will lead to fluctuations, as shown in
Figure 4a,c. Taking the DGI results of the binary image “01” (i.e., SNR = 1.48 and
t = 2.62 s, see
Figure 2b as a reference (which is drawn as straight lines in
Figure 4a,b), it can be seen from the curves that no matter how the value of
changes in the range from 0 to 0.5, the SNRs of the GDGI are always higher than those of the DGI, and when
is greater than 0.1, the calculation time of GDGI is almost the same as that of the DGI. The above results demonstrate that our proposed method has a more efficient imaging performance than the DGI in the binary case.
For the grayscale image “02”, from
Figure 4c,d we can see that the SNRs of reconstructed images present an overall downward trend and the calculation time
t also drops sharply with the increase in step size
. When
, the SNRs of the GDGI are almost constant. When
further increases, the SNR gradually changes from a small-amplitude fluctuation to a large-amplitude fluctuation, as shown in
Figure 4c. This is because the increase in the step size
causes an increase in the gradient descent stepping distance (when the iteration is close to the optimal solution but the preset termination conditions are not met, the jump will continue; the larger the step size, the more likely it is to jump out of the region where the optimal solution is located and enter a new reciprocating iteration, which may cause reciprocating oscillations in SNR values). Similarly, we took the DGI results of the grayscale image “02” (i.e., SNR = 3.86 and
t = 2.59 s, see
Figure 3b) as a reference (which is also drawn as straight lines in
Figure 4c,d). We can see from
Figure 4c,d that the SNRs of the GDGI never fall below those of the DGI, and when
is greater than 0.1, the calculation times of the GDGI and DGI are nearly the same. This proves that our scheme is also suitable for complex grayscale objects. According to the above results, to form a trade-off between the imaging quality and running time, the appropriate value range of the step size
should range from 0.05 to 0.15. For simplicity, in the following, we set
. In addition, since the result of
is always positive, the iterative updating direction is usually opposite to the gradient direction, but we believe that
, in some special cases, can also take a negative value; for example, when the step size is too large, the algorithm misses the optimal solution and
should be changed to a negative value. This appears complicated and needs further study in the future, so it will not be discussed in this article.
Next, to verify the universality and effectiveness of our algorithm, we tested another binary image and a grayscale image whose gray value range is [0, 1], marked as “03” and “04”, respectively.
Figure 5 shows the original images and the corresponding imaging results of the DGI, PGI and GDGI with different sampling rates
r (defined as the ratio of the number of measurements to the total number of image pixels). As can be seen from
Figure 5b–m,o–z, when
r = 0.5, the image quality of the PGI is slightly better than those of the DGI and GDGI for the image “03”, while the image quality of the GDGI is 0.59 and 0.82 times higher than those of the DGI and PGI for the image “04”. When
r = 1, the image quality of the GDGI is 0.08 and 0.14 times higher than those of the DGI and PGI for the image “03”, and 1.99 and 2.57 times higher than those of the DGI and PGI for the image “04”. When
r = 3 and 5, the imaging performance of the GDGI is much better than those of the DGI and PGI. Therefore, it can be inferred that, with the increase in sampling rate
r, the improvement in image quality will be more significant. In addition, since our algorithm does not require preprocessing or optimizing of the measurement data, the use of a gradient-descent-like method can also fully improve the utilization of data, thereby greatly saving additional computational overhead. Here, the average running time of the GDGI is only 1.19 and 1.18 times those of the DGI or PGI, respectively, which means that a significant improvement in imaging quality can be achieved without much running time increment, i.e., our scheme has a relatively high return-to-investment ratio. Besides, through this simulation, it also demonstrates that our algorithm works well for different types of images.
In the following, we performed additional numerical simulations to test the robustness of the GDGI against noise. Taking the image “02” as an example, we further added white Gaussian noise to the bucket value
to simulate the noisy measurements. Then, we calculated the SNRs of reconstructed images under different detection signal-to-noise ratios (DSNRs), i.e., the power ratio of the signal to the measurement noise, which can be expressed as
[
29], where
and
denote the variances of the bucket values and measurement noise, respectively. As shown in
Figure 6a, when the Gaussian random noise with different variances is added to the bucket signal, the SNRs of images recovered by the DGI, PGI and GDGI decrease with the decrease in the DSNR. It can be clearly seen from the curves that when the DSNR is greater than 5 dB, the imaging performance of the GDGI is significantly better than those of the DGI and PGI; when the DSNR is less than 5 dB, the SNRs of these three methods become closer, but the SNRs of the GDGI are still slightly higher than those of the DGI and PGI. This is because the statistical averaging of the intensity correlation is accompanied by reconstruction noise. The lower the DSNR, the harsher the measurement condition and the more difficult it is to distinguish the useful signal from the noise, which leads to a larger reconstruction noise and makes it more difficult to recognize the object part in the image. In ultra-low DSNR conditions, all algorithms including convex optimization will tend to fail. As shown in
Figure 6b–j, we made some comparisons among the DGI, PGI and GDGI, with the DSNR changing from 35 dB to 15 dB. When DSNR = 35 dB, 25 dB and 15 dB, the SNRs of the GDGI are 8.87, 6.71 and 3.13 times those of the DGI, and are 16.91, 12.79, 5.87 times those of the PGI, showing that our algorithm has a good anti-noise ability compared with the traditional GI algorithm under noisy measurements.