Next Article in Journal
Prediction of Extracellular Matrix Proteins by Fusing Multiple Feature Information, Elastic Net, and Random Forest Algorithm
Next Article in Special Issue
Multipoint Fractional Iterative Methods with (2α + 1)th-Order of Convergence for Solving Nonlinear Problems
Previous Article in Journal
Diffeological Statistical Models, the Fisher Metric and Probabilistic Mappings
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Modified Hestenes-Stiefel-Type Derivative-Free Method for Large-Scale Nonlinear Monotone Equations

College of Mathematics and Computational Science, Changsha University of Science and Technology, Changsha 410114, China
*
Author to whom correspondence should be addressed.
Mathematics 2020, 8(2), 168; https://doi.org/10.3390/math8020168
Submission received: 26 December 2019 / Revised: 16 January 2020 / Accepted: 21 January 2020 / Published: 30 January 2020
(This article belongs to the Special Issue Iterative Methods for Solving Nonlinear Equations and Systems 2020)

Abstract

:
The goal of this paper is to extend the modified Hestenes-Stiefel method to solve large-scale nonlinear monotone equations. The method is presented by combining the hyperplane projection method (Solodov, M.V.; Svaiter, B.F. A globally convergent inexact Newton method for systems of monotone equations, in: M. Fukushima, L. Qi (Eds.) Reformulation: Nonsmooth, Piecewise Smooth, Semismooth and Smoothing Methods, Kluwer Academic Publishers. 1998, 355–369) and the modified Hestenes-Stiefel method in Dai and Wen (Dai, Z.; Wen, F. Global convergence of a modified Hestenes-Stiefel nonlinear conjugate gradient method with Armijo line search. Numer Algor. 2012, 59, 79–93). In addition, we propose a new line search for the derivative-free method. Global convergence of the proposed method is established if the system of nonlinear equations are Lipschitz continuous and monotone. Preliminary numerical results are given to test the effectiveness of the proposed method.

1. Introduction

In this paper, we consider the problem of finding numerical solutions of the following large-scale nonlinear equations
F ( x ) = 0 ,
where the function F : R n R n is monotone and continuous. If F ( x ) is monotone, it implies that the following inequality holds
F ( x ) F ( y ) , x y 0 , x , y R n .
Nonlinear monotone equations can be applied in different fields, for example, they are used as subproblems in the generalized proximal algorithms with Bregman distances [1]. Some monotone variational inequality problems can be converted into nonlinear monotone equations [2]. Monotone systems of equations can also be applied in L 1 -norm regularization sparse optimization problems (see [3,4]) and discrete mathematics such as graph theory (see [5,6]).
Being aware of the important applications of nonlinear monotone equations, in recent years, many scholars have paid attention to propose efficient algorithms for solving problem (1). These algorithms are mainly divided into the following categories.
Each of the Newton-type method, Levenberg-Marquardt method, and quasi-Newton method enjoy fast local convergence property, and are attractive (see [7,8,9,10,11,12,13]). However, for large-scale problems, a drawback of these methods is that at each iteration these algorithms require computing a large-scale linear system of equations by using approximate systems of equations or a Jacobian matrix. The large demand of storage for matrix results in improper handling of large-scale nonlinear monotone systems.
In recent years, gradient-type algorithms have attracted the attention of many scholars. The main reasons are low storage requirements, easy implementation, and global convergence under mild conditions. For example, the spectral gradient method [14] only needs to use gradient information which is easy but effective for an optimization problem. From a different perspective, the spectral gradient method [14] is extended to solve nonlinear monotone equations by combining it with the projection method (see [15,16]).
In addition, for large-scale unconstrained optimization problems, the conjugate gradient method (CG) is another easy but effective method, due to the two attractive features: one is the low memory requirement, the other is strong global convergence properties. In recent years, the conjugate gradient method has achieved rich results (see [17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36]) from the perspective of the sufficient descent property, qausi-Newton direction and conjugacy condition. Inspired by the extension of spectral gradient method to nonlinear monotone equations, CG methods have been applied in solving the nonlinear monotone equations problem. For example, PRP-type method ([37,38,39,40,41,42,43]); Perry conjugate gradient method ([44]); Liu-Storey type method [45], among others.
In this paper, we will focus on extending the Hestenes-Stiefel (HS) CG method to solve large-scale nonlinear monotone equations. To the best of our knowledge, the HS CG method [46] is generally considered as the most efficient CG method for computing performance. However, the HS CG method does not enjoy sufficient descent property. Based on the modified secant equation [10], Dai and Wen [47] propose a modified HS conjugate gradient method that can generate sufficient descent directions (i.e., c > 0 exists such that g k T d k < c g k 2 ). Global convergence results under the Armijo line search are obtained in Dai and Wen [47]. Hence, we aim to present a derivative-free method to solve the nonlinear monotone Equations (1). The proposed method can be seem as a further study of the modified HS CG method in Dai and Wen [47] for unconstrained optimization problems.
Our paper makes two contributions to large-scale nonlinear monotone equations. Firstly, a new line search is proposed for the derivative-free method. A significant advantage of this line search is that it is easier to obtain the search stepsize. Secondly, we propose a derivative-free method for solving large-scale nonlinear monotone equations which combines the modified Hestenes-Stiefel method in Dai and Wen [47] for unconstrained optimization problems and the hyperplane projection method [13]. A good property of the proposed method is that it is suitable to solve large-scale nonlinear monotone equations due to its lower storage requirement.
The rest of the article is organized as follows. In Section 2, we give the algorithm and prove the sufficient descent property. The global convergence is proved in Section 3. We report the numerical results In Section 4. The last Section gives the conclusion.

2. Algorithm and the Sufficient Descent Property

In this section, we will present the derivative-free method for solving problem (1), that is a combination of the modified Hestenes-Stiefel method [47] and the hyperplane projection method [13]. Different from the traditional conjugate gradient method, the iteration sequence { x k + 1 } is obtained in two steps at each iteration.
In the first step, the algorithm produces an iterative sequence { z k = x k + α k d k } , where d k is the search direction, and α k > 0 is the steplength obtained by a suitable line search. For most iterative algorithms of optimization problems, the line search plays an important role in convergence analysis and numerical calculation. Zhang and Zhou [15] obtained the steplength α k > 0 by the following Armijio-type line search: calculating the search steplength α k = m a x { β ρ i : i = 0 , 1 , , } such that
F ( x k + α k d k ) T d k σ α k d k 2 ,
where β is some initial attempt for α k , β > 0 and ρ ( 0 , 1 ) .
In addition, Li and Li [39] introduced an alternative line search, that is, computing the search steplength α k = m a x { β ρ i : i = 0 , 1 , , } such that
F ( x k + α k d k ) T d k σ α k F ( x k + α k d k ) · d k 2 ,
From the above introduction, we can see that the steplength α k is obtained by calculating α k = m a x { β ρ i : i = 0 , 1 , , } such that (3) or (4) is satisfied. If the point x k is far from the solution, the obtained steplength α k may be very small. Taking this into account, we present the following line search rule where the steplength α k is obtained by computing α k = m a x { β ρ i : i = 0 , 1 , , } such that
F ( x k + α k d k ) T d k σ α k min { d k 2 , F ( x k + α k d k ) d k 2 , F ( x k ) T d k } .
In the second step, { x k + 1 } can be determined by using x k , z k , F ( z k ) via the hyperplane projection method [13]. Now, let’s introduce how to generate { x k + 1 } via the hyperplane projection method [13]. Along the search direction d k > 0 , we can generate a point z k = x k + α k d k by a suitable line search such that
F ( z k ) T ( x k z k ) > 0 .
On the other hand, the monotonicity of F implies that for any solution x ( F ( x ) = 0 ) , the following inequality holds
F ( z k ) T ( x z k ) = ( F ( z k ) F ( x ) ) T ( z k x ) 0 .
From (6) and (7), we can see that { F ( z k ) T ( x k z k ) > 0 } holds for any x k and { F ( z k ) T ( x z k ) 0 } holds for the solution x . Therefore, from (6) and (7), there is a hyperplane
H k = { x R n | F ( z k ) T ( x z k ) = 0 } ,
which can strictly separate the current point x k from the x (zero point) of equation in (1).
Following Solodov and Svaiter [13] and Zhang and Zhou [15], taking the projection of x k onto the hyperplane (8) as the next iterate x k + 1 is a reasonable choice. In detail, the next iterate point x k + 1 can be computed by
x k + 1 = x k F ( z k ) T ( x k z k ) F ( z k ) 2 F ( z k ) .
In what following, we pay our attention to the search direction which plays a crucial role in an iterative algorithm. Our main starting point is to extend the search direction of Dai and Wen [47] to nonlinear monotone equations problem (1). Similar to Dai and Wen [47] for unconstrained optimization, we give the search direction as
d k = F 0 , if k = 0 , F k + β k NHZ d k 1 , if k 1 ,
where
β k NHZ = F k T y k 1 d k 1 T w k 1 μ y k 1 2 ( d k 1 T w k 1 ) 2 F k T d k 1 , μ > 1 4 ,
w k 1 = y k 1 + γ s ¯ k 1 , γ > 0 , y k 1 = F ( x k ) F ( x k 1 ) , s ¯ k 1 = z k 1 x k 1 = α k 1 d k 1 .
For simplicity, we refer to (10) and (11) as NHZ method hereafter.
Further, in this paper, the function F is assumed to satisfy the following assumptions witch are often utilized in convergence analysis for nonlinear monotone equations (see, [37,38,39,40,41,42,43,44,45,48]).
Assumption A1.
( A 1 )
The F is a monotone function:
( F ( x ) F ( y ) ) T ( x y ) 0 , x , y R n .
( A 2 )
The F is Lipschitz continuous function, namely, there exists a L > 0 such that
F ( x ) F ( y ) L x y , x , y R n .
In what following, we will describe the proposed Algorithm 1.
Algorithm 1: NHZ derivative-free method.
Step 0: Given x 0 R n as an initial point, and the constants ε > 0 , β > 0 , σ > 0 , ρ ( 0 , 1 ) . Then set k := 0.
Step 1: Calculate F ( x k ) . If  F ( x k ) ε , stop the algorithm. Otherwise, go to step 2.
Step 2: Determine the search direction d k by (10), (11) and (12).
Step 3: Calculate the search steplength α k by (5). Let z k = x k + α k d k .
Step 4: Calculate F ( z k ) . If  F ( z k ) ε , stop the algorithm. Otherwise, calculate x k + 1 by using the projection (9). Set k := k + 1 and go to Step 1.
In what follow, we will show that the proposed NHZ derivative-free method enjoys the sufficient descent property which plays an important role in proof of convergence. From now on, we use F k to denote F ( x k ) .
Theorem 1.
The search direction d k generated by (10), (11) and (12) is sufficient descent direction. That is, if d k T w k 0 , then we have
F k T d k ( 1 1 4 μ ) F k 2 , μ > 1 4 .
Proof. 
When k = 0 , we have
F 0 T d 0 = F 0 2 ( 1 1 4 μ ) F 0 2 .
It is obvious that (15) is satisfied for k = 0 . □
Now we will show that the sufficient descent condition (15) holds for k 1 . We can obtain from (10) and (11) that
F k T d k = F k 2 + β k F k T d k 1 = F k 2 + F k T y k 1 d k 1 T w k 1 μ y k 1 2 ( d k 1 T w k 1 ) 2 F k T d k 1 F k T d k 1 = F k T y k 1 ( d k 1 T w k 1 ) ( F k T d k 1 ) F k 2 ( d k 1 T w k 1 ) 2 μ y k 1 2 ( F k T d k 1 ) 2 ( d k 1 T w k 1 ) 2 .
Define
u k = 1 2 μ ( d k 1 T w k 1 ) F k , v k = 2 μ ( F k T d k 1 ) y k 1 .
By using the Equation (16) and the inequality u k T v k 1 / 2 ( u k 2 + v k 2 ) , we have
F k T d k = u k T v k 1 / 2 ( u k 2 + v k 2 ) ( d k 1 T w k 1 ) 2 ( 1 1 4 μ ) ( d k 1 T w k 1 ) 2 ( d k 1 T w k 1 ) 2 F k 2 ( 1 1 4 μ ) F k 2 .
Thus (15) holds for k 1 .

3. Global Convergence Analysis

Now, we will investigate the global convergence of Algorithm 1. Firstly, we give the following Lemma which shows the line search strategy (5) is well-defined if the search directions { d k } satisfy the sufficient descent property.
Lemma 1.
If the iterative sequences { x k } and { z k } are generated by the Algorithm 1. Then, there always exists a steplength α k satisfying the line search (5).
Proof. 
Assume that, for any nonnegative integer i for β ρ i , the line search strategy (5) does not hold in the k 0 -th iterate, then we have
F ( x k 0 + β ρ i d k 0 ) T d k 0 < σ β ρ i min { d k 0 2 , F ( x k 0 + β ρ i d k 0 ) d k 0 2 , F ( x k 0 ) T d k 0 } .
Letting i , we have from the continuity of F and ρ ( 0 , 1 ) that
F ( x k 0 + β ρ i d k 0 ) T d k 0 < 0 ,
which contradicts (15). The proof is completed. □
The next Lemma indicates that the line search strategy (5) provides a lower bound for steplength α k .
Lemma 2.
If the iterative sequences { x k } and { z k } are generated by the Algorithm 1. Then, we can obtain that
α k min β , δ ρ ( L + σ ) F k 2 d k 2 ,
where δ = 1 1 4 μ .
Proof. 
Tf α k = β , it is obviously that (19) holds. Suppose that α k β , then we can obtain that α k = ρ 1 α k does not satisfy the line search process (5). That is,
F ( z k ) T d k < σ α k min { d k 2 , F ( z k ) d k 2 , F ( x k ) T d k } σ α k d k 2 ,
where z k = x k + α k d k .  □
From the sufficient descent condition (15), we have that
( 1 1 4 μ ) F k 2 δ F k 2 F k T d k .
From the Lipschitz continuity of F (14), (20) and (21), we can obtain that
F k T d k = ( F ( z k ) F ( x k ) ) T d k F ( z k ) T d k F ( z k ) F ( x k ) d k + σ α k d k 2 L z k x k d k + σ α k d k 2 = L α k d k 2 + σ α k d k 2 = ρ 1 α k ( L + σ ) d k 2 .
Therefore, the above inequalities and (21) imply
α k δ ρ ( L + σ ) F k 2 d k 2 .
This shows that Lemma about the search steplength α k holds.
The next Lemma is proved by Solodov and Svaiter (see Lemma 2.1 in [13]), which can also hold for Algorithm 1. Now, we give this lemma without proof, because its proof is similar in Solodov & Svaiter [13].
Lemma 3.
Assume the function F is monotone and the Lipschitz continuous condition (14) holds. If the iterative sequences { x k } is generated by the Algorithm 1, then for any x , such that F ( x ) = 0 , we can obtain that
x k + 1 x 2 x k x 2 x k + 1 x k 2 .
In particular, the iterative sequence { x k } is bounded and
k = 0 x k + 1 x k 2 < .
Remark 1.
The above Lemma 2 confirms that the sequence { x k x } decreases with k. In addition, (22) implies that
lim k x k + 1 x k = 0 .
Theorem 2.
If the iterative sequences { x k } is generated by the Algorithm 1, then, we can have that
lim k α k d k = 0 .
Proof. 
We can obtain from (5) and (9) that, for any k,
x k + 1 x k = | F ( z k ) T ( x k z k ) | F ( z k ) = α k F ( z k ) T d k F ( z k ) σ α k 2 d k 2 .
In particular, it follows from (23) and (25) that
lim k α k d k = 0 .
 □
Lemma 4.
If the iterative sequences { x k } is generated by the Algorithm 1, and x satisfies F ( x ) = 0 , z k = x k + α k d k , α k = ρ 1 α k . Then, { F ( z k ) } and { F k } are bounded, i.e, there is a constant M 0 , such that
F ( z k ) M , F k M .
Proof. 
By the Lemma 2, we have
x k x x 0 x .
From (26), we have that there is a constant M 1 > 0 such that α k d k M 1 . Hence
z k x x k x + α k d k x 0 x + ρ 1 α k d k x 0 x + M 1 .
Since the function F ( x ) is Lipschitz continuous, we can easily obtain the following two inequalities
F ( z k ) F ( z k ) F ( x ) L z k x L ( x 0 x + ρ 1 M 1 ,
and
F k F ( x k ) F ( x ) L x k x L x 0 x .
Let M = m a x { L x 0 x , L ( x 0 x + ρ 1 M 1 } . We can obtain (27). □
In what following, we will give the global convergence theorem for our proposed method.
Theorem 3.
If the iterative sequences { x k } is generated by the Algorithm 1. We can obtain that
lim inf k F k = 0 .
Proof. 
We will prove this Theorem by contradiction. Assume that (29) is not true. Then it implies there is a constant ε > 0 , s.t. F k > ε .
Since F k 0 , we have from (15) that d k 0 . Hence, the monotonicity of F together with (10) implies
s ¯ k 1 T w k 1 = F ( z k 1 ) F ( x k 1 ) , z k 1 x k 1 + γ s ¯ k 1 T s ¯ k 1 γ s ¯ k 1 T s ¯ k 1 .
This together with the definition of s ¯ k 1 implies
d k 1 T w k 1 γ α k 1 d k 1 2 .
We have from (11), (12) and (30) that
| β k NHZ | = | F k T y k 1 d k 1 T w k 1 μ y k 1 2 ( d k 1 T w k 1 ) 2 F k T d k 1 | L α k 1 d k 1 F k γ α k 1 d k 1 2 + μ L 2 α k 1 2 d k 1 2 γ 2 α k 1 2 d k 1 4 F k d k 1 ( L γ + μ L 2 γ 2 ) F k d k 1 .
Therefore, from (10) and (30), we can obtain
d k F k + | β k NHZ | d k 1 F k + ( L γ + μ L 2 γ ) F k ( 1 + L γ + μ L 2 γ 2 ) M .
Define C = ( 1 + L γ + μ L 2 γ 2 ) M . Then, we can have d k C . □
It follows from Lemmas 2, Lemmas 3, F k ε and d k ε that for all k sufficiently large,
α k d k min β , δ ρ ( L + σ ) F k 2 d k 2 d k min β ε , δ ρ ε 2 ( L + σ ) C > 0 .
It is obvious that the above inequality contradicts with (24). That is, (29) holds. And we complete this proof.

4. Numerical Experiments

Now, we will give some numerical experiments to test the numerical performance of our proposed method. We try to test the NHZ Algorithm 1 and compare it’s performance with the spectral gradient (SG) method [15] and the MPRP method in [39]. In the testing experiments, all codes were written in Matlab R2018a, and run on a Lenovo PC with 4 GB RAM memory.
To obtain better numerical performance, we select the following initial steplength as in [39] and [43]
β = | F ( x k ) T d k d k T ( F ( x k + ϵ d k ) F ( x k ) ) / ϵ | | F ( x k ) T d k d k T F ( x k ) d k | , ϵ = 10 8 .
We set ρ = 0 . 5 , σ = 2 . Further, let β = 1 if β < 10 4 .
Following the MPRP method in [39], we terminate the iterative process if the following condition
min { F ( x k ) , F ( z k ) } a t o l + r t o l F ( x 0 )
is satisfied, and r t o l = a t o l = 10 4 .
The numerical performance of SG, MPRP, and NHZ methods are tested by using the following five nonlinear monotone equations problem with different various sizes and initial points.
Problem 1
([49]). The specific expression of the function F ( x ) is defined as
F i ( x ) = 2 x i s i n ( | x i | ) , i = 1 , , n .
Problem 2
([49]). The specific expression of the function F ( x ) is defined as
F 1 ( x ) = 2 x 1 + s i n ( x 1 ) 1 , F i ( x ) = 2 x i 1 + 2 x i + s i n ( x i ) 1 , i = 2 , , n 1 , F n ( x ) = 2 x n + s i n ( x n ) 1 .
Problem 3
([50]). The specific expression of the function F ( x ) is defined as
F i ( x ) = x i s i n ( x i ) , i = 1 , , n .
Problem 4
([50]). The specific expression of the function F ( x ) is defined as
F ( x ) = A x + g ( x )
where g ( x ) = ( e x 1 1 , e x 2 1 , , e x n 1 ) T
A = 2 1 1 2 1 1 1 2
Problem 5.
The specific expression of the function F ( x ) is defined as
F ( x ) = A x + | X | B
where | X | = ( | x 1 | , | x 2 | , , | x n | ) T , B = ( 1 , 1 , , 1 ) T , and
A = 2 1 1 2 1 1 1 2 .
The numerical results for 5 tested problems are reported in Table 1, Table 2, Table 3, Table 4 and Table 5 respectively, where the given initial points are x 1 = ( 0.1 , , 0.1 ) T , x 2 = ( 1 , , 1 ) T , x 3 = ( 1 , 1 2 , , 1 n ) T , x 4 = ( 10 , , 10 ) T , x 5 = ( 0.1 , , 0.1 ) T , x 6 = ( 1 , , 1 ) T . Meanwhile, the numerical results are listed in the Table 1, Table 2, Table 3, Table 4 and Table 5 where (Time) is the CPU time (in seconds), (Iter) is the number of iterations and (Feval) is the number of function evaluations.
Table 1, Table 2, Table 3, Table 4 and Table 5 report the numerical results of the proposed algorithm, the spectral gradient (SG) method [15] and the MPRP method in [39] with five tested problems where the test indicators of Time, Iter and Feval are used to evaluate numerical performance.
Comparing the CPU time within the tested algorithms, we note that our proposed algorithm needs lower time consuming than both the spectral gradient (SG) method [15] and the MPRP method in [39] for all tested problems, and the difference is substantial and significant especially for large-scale problems. In addition, the MPRP method in [39] needs lower CPU time than the spectral gradient (SG) method [15]. Assessing the number of iterations within the tested algorithms, we find that the NHZ method requires fewer iterations than the spectral gradient (SG) method [15] and the MPRP method in [39] for all tested problems. We also note that our proposed algorithm requires fewer number of function evaluations for all tested problems, and the difference is substantial and significant.
In sum, from the numerical results in Table 1, Table 2, Table 3, Table 4 and Table 5, it isn’t difficult to see that the proposed algorithm performs better than the spectral gradient (SG) method [15] and the MPRP method in [39] for three test indicators which implies that the modified Hestenes-Stiefel-based derivative-free method is computationally efficient for nonlinear monotone equations.

5. Conclusions

This paper aims to present a modified Hestenes-Stiefel method to solve the nonlinear monotone equations which combines the hyperplane projection method [13] and the modified Hestenes-Stiefel method in Dai and Wen [47]. In the proposed method, the search direction satisfies sufficient descent conditions. A new line search is proposed for the derivative-free method. Under appropriate conditions, the proposed method converges globally. The given numerical results show the presented method is more efficient compared to the methods proposed by the spectral gradient method Zhang and Zhou [15] and the MPRP method in Li and Li [39].
In addition, we also expect that our proposed method and its further modifications could produce new applications for problems in relevant areas of symmetric equations [51], image processing [52], and finance [53,54,55].

Author Contributions

Conceptualization, Z.D.; methodology, Z.D. and H.Z.; software, Z.D. and H.Z.; validation, Z.D.; formal analysis, Z.D.; investigation, Z.D.; resources, Z.D. and H.Z.; data curation, Z.D.; writingoriginal draft preparation, Z.D. and H.Z.; writing-review and editing Z.D. and H.Z.; visualization, Z.D. and H.Z.; supervision, Z.D. and H.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the NSF of China granted 71771030, 11301041, and fund of Hunan Provincial Education Department granted 19A007.

Acknowledgments

This work was supported by the National Natural Science Foundation of China, grants 71771030, and 11301041; and the Scientific Research Fund of Hunan Provincial Education Department, grant number 19A007.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Iusem, A.N.; Solodov, M.V. Newton-type methods with generalized distances for constrained optimization. Optimization 1997, 41, 257–278. [Google Scholar] [CrossRef]
  2. Zhao, Y.B.; Li, D.H. Monotonicity of fixed point and normal mapping associated with variational inequality and its application. SIAM J. Optim. 2001, 4, 962–973. [Google Scholar] [CrossRef]
  3. Figueiredo, M.; Nowak, R.; Wright, S.J. Gradient projection for sparse reconstruction, application to compressed sensing and other inverse problems. IEEE J-STSP 2007, 1, 586–597. [Google Scholar] [CrossRef] [Green Version]
  4. Xiao, Y.; Zhu, H. A conjugate gradient method to solve convex constrained monotone equations with applications in compressive sensing. J. Math. Anal. Appl. 2013, 405, 310–319. [Google Scholar] [CrossRef]
  5. Shang, Y. Vulnerability of networks: Fractional percolation on random graphs. Phys. Rev. E 2014, 89, 012813. [Google Scholar] [CrossRef]
  6. Shang, Y. Super Connectivity of Erdos-Renyi Graphs. Mathematics 2019, 7, 267. [Google Scholar] [CrossRef] [Green Version]
  7. Brown, P.N.; Saad, Y. Convergence theory of nonlinear Newton-Krylov algorithms. SIAM J. Optim. 1994, 4, 297–330. [Google Scholar]
  8. Gasparo, M. A nonmonotone hybrid method for nonlinear systems. Optim. Methods Softw. 2000, 13, 79–94. [Google Scholar] [CrossRef]
  9. Griewank, A. The global convergence of Broyden-like methods with suitable line search. J. Austral. Math. Soc. Ser. B 1996, 28, 75–92. [Google Scholar] [CrossRef] [Green Version]
  10. Li, D.H.; Fukushima, M. A modified BFGS method and its global convergence in nonconvex minimization. J. Comput. Appl. Math. 2001, 129, 15–35. [Google Scholar] [CrossRef] [Green Version]
  11. Li, D.H.; Fukushima, M. A derivative-free line search and global convergence of Broyden-like method for nonlinear equations. Optim. Methods Softw. 2000, 13, 583–599. [Google Scholar] [CrossRef]
  12. Martínez, J.M. A family of quasi-Newton methods for nonlinear equations with direct secant updates of matrix factorizations. SIAM J. Numer. Anal. 1990, 27, 1034–1049. [Google Scholar] [CrossRef]
  13. Solodov, M.V.; Svaiter, B.F. A globally convergent inexact Newton method for systems of monotone equations. In Reformulation: Nonsmooth, Piecewise Smooth, Semismooth and Smoothing Methods; Fukushima, M., Qi, L., Eds.; Kluwer Academic Publishers: Dordrecht, The Netherlands, 1998; pp. 355–369. [Google Scholar]
  14. Barzilai, J.; Borwein, J.M. Two-point step size gradient methods. IMA J. Numer. Anal. 1998, 8, 141–148. [Google Scholar] [CrossRef]
  15. Zhang, L.; Zhou, W.J. Spectral gradient projection method for solving nonlinear monotone equations. J. Comput. Appl. Math. 2006, 196, 478–484. [Google Scholar] [CrossRef] [Green Version]
  16. Yu, Z.S.; Ji, L.N.; Sun, J.; Xiao, Y.H. Spectral gradient projection method for monotone nonlinear equations with convex constraints. Appl. Numer. Math. 2009, 59, 2416–2423. [Google Scholar] [CrossRef]
  17. Hager, W.W.; Zhang, H. A new conjugate gradient method with guaranteed descent and an efficient line search. SIAM J. Optim. 2005, 16, 170–192. [Google Scholar] [CrossRef] [Green Version]
  18. Zhang, L.; Zhou, W.J.; Li, D.H. A descent modified Polak-Ribière-Polyak conjugate gradient method and its global convergence. IMA J. Numer. Anal. 2006, 26, 629–640. [Google Scholar] [CrossRef]
  19. Cheng, W.Y. A two-term PRP-based descent method. Numer. Funct. Anal. Optim. 2007, 28, 1217–1230. [Google Scholar] [CrossRef]
  20. Zhang, L.; Zhou, W.J.; Li, D.H. Global convergence of a modified Fletcher-Reeves conjugate gradient method with Armijo-type line search. Numer. Math. 2006, 104, 561–572. [Google Scholar] [CrossRef]
  21. Dai, Z.; Zhu, H. Stock return predictability from a mixed model perspective. Pac-Basin. Finac. J. 2020, 60, 101267. [Google Scholar] [CrossRef]
  22. Narushima, Y.; Yabe, H.; Ford, J.A. A three-term conjugate gradient method with sufficient descent property for unconstrained optimization. SIAM J. Optim. 2011, 21, 212–230. [Google Scholar] [CrossRef] [Green Version]
  23. Andrei, N. A simple three-term conjugate gradient algorithm for unconstrained optimization. J. Comp. Appl. Math. 2013, 241, 19–29. [Google Scholar] [CrossRef]
  24. Andrei, N. On three-term conjugate gradient algorithms for unconstrained optimization. Appl. Math. Comput. 2013, 219, 6316–6327. [Google Scholar] [CrossRef]
  25. Liu, Z.X.; Liu, H.W.; Dong, X.L. A new adaptive Barzilai and Borwein method for unconstrained optimization. Optim. Lett. 2018, 12, 845–873. [Google Scholar] [CrossRef]
  26. Babaie-Kafaki, S.; Reza, G. The Dai-Liao nonlinear conjugate gradient method with optimal parameter choices. Eur. J. Oper. Res. 2014, 234, 625–630. [Google Scholar] [CrossRef]
  27. Babaie-Kafaki, S. On optimality of the parameters of self-scaling memoryless quasi-Newton updating formulae. J. Optim. Theory Appl. 2015, 167, 91–101. [Google Scholar] [CrossRef]
  28. Yuan, G.; Zhang, M. A three-terms Polak-Ribiére-Polyak conjugate gradient algorithm for large-scale nonlinear equations. J. Comput. Appl. Math. 2015, 286, 186–195. [Google Scholar] [CrossRef]
  29. Yuan, G.; Meng, Z.H.; Li, Y. A modified Hestenes and Stiefel conjugate gradient algorithm for large-scale nonsmooth minimizations and nonlinear equations. J. Optim. Theory. Appl. 2016, 168, 129–152. [Google Scholar] [CrossRef]
  30. Dong, X.; Han, D.; Reza, G.; Li, X.; Dai, Z. Some new three-term Hestenes-Stiefel conjugate gradient methods with affine combination. Optimization 2017, 66, 759–776. [Google Scholar] [CrossRef]
  31. Dong, X.; Han, D.; Dai, Z.; Li, L.; Zhu, J. An accelerated three-term conjugate gradient method with sufficient descent condition and conjugacy condition. J. Optim. Theory Appl. 2018, 179, 944–961. [Google Scholar] [CrossRef]
  32. Li, M.; Feng, H. A sufficient descent Liu-Storey conjugate gradient method for unconstrained optimization problems. Appl Math Comput. 2011, 218, 1577–1586. [Google Scholar]
  33. Dai, Z.; Wen, F. Another improved Wei-Yao-Liu nonlinear conjugate gradient method with sufficient descent property. Appl Math Comput. 2012, 218, 4721–4730. [Google Scholar] [CrossRef]
  34. Dai, Z.; Tian, B. Global convergence of some modified PRP nonlinear conjugate gradient methods. Optim. Lett. 2011, 5, 615–630. [Google Scholar] [CrossRef]
  35. Dai, Z. Comments on a new class of nonlinear conjugate gradient coefficients with global convergence properties. Appl. Math. Computat. 2016, 276, 297–300. [Google Scholar] [CrossRef] [Green Version]
  36. Dai, Z.; Zhou, H.; Wen, F.; He, S. Efficient predictability of stock return volatility: The role of stock market implied volatility. N. Am. J. Econ. Finance. 2020. forthcoming. [Google Scholar] [CrossRef]
  37. Cheng, W.Y. A PRP type method for systems of monotone equations. Math. Comput. Model. 2009, 50, 15–20. [Google Scholar] [CrossRef]
  38. Yu, G. A derivative-free method for solving large-scale nonlinear systems of equations. J. Ind. Manag. Optim. 2010, 6, 149–160. [Google Scholar] [CrossRef]
  39. Li, Q.; Li, D.H. A class of derivative-free methods for large-scale nonlinear monotone equations. IMA J. Numer. Anal. 2011, 31, 1625–1635. [Google Scholar] [CrossRef]
  40. Yu, G. Nonmonotone spectral gradient-type methods for large-scale unconstrained optimization and nonlinear systems of equations. Pac. J. Optim. 2011, 7, 387–404. [Google Scholar]
  41. Zhou, W.; Shen, D. An inexact PRP conjugate gradient method for symmetric nonlinear equations. Numer. Funct. Anal. Optim. 2014, 35, 370–388. [Google Scholar] [CrossRef]
  42. Sun, M.; Wang, X.; Feng, D. A family of conjugate gradient methods for large-scale nonlinear equations. J. Inequal. Appl. 2017, 236, 1–8. [Google Scholar]
  43. Zhou, W.; Wang, F. A PRP-based residual method for large-scale monotone nonlinear equations. Appl. Math. Comput. 2015, 261, 1–7. [Google Scholar] [CrossRef]
  44. Dai, Z.; Chen, X.; Wen, F. A modified Perry’s conjugate gradient method-based derivative-free method for solving large-scale nonlinear monotone equation. Appl. Math. Comput. 2015, 270, 378–386. [Google Scholar] [CrossRef] [Green Version]
  45. Li, M. A Liu-Storey type method for solving large-scale nonlinear monotone equations. Numer. Funct. Anal. Optim. 2014, 35, 310–322. [Google Scholar] [CrossRef]
  46. Hestenes, M.R.; Stiefel, E. Methods of conjugate gradients for solving linear systems. J. Res. Natl. Bur. Stand. 1952, 49, 409–436. [Google Scholar] [CrossRef]
  47. Dai, Z.; Wen, F. Global convergence of a modified Hestenes-Stiefel nonlinear conjugate gradient method with Armijo line search. Numer Algor. 2012, 59, 79–93. [Google Scholar] [CrossRef]
  48. Yan, Q.R.; Peng, X.Z.; Li, D.H. A globally convergent derivative-free method for solving large-scale nonlinear monotone equations. J. Comput. Appl. Math. 2010, 234, 649–657. [Google Scholar] [CrossRef] [Green Version]
  49. Zhou, W.J.; Li, D.H. Limited memory BFGS method for nonlinear monotone equations. J. Comput. Math. 2007, 25, 89–96. [Google Scholar]
  50. Zhou, W.J.; Li, D.H. A globally convergent BFGS method for nonlinear monotone equations without any merit functions. Math. Comput. 2008, 77, 2231–2240. [Google Scholar] [CrossRef]
  51. Zhou, W.J. A modified BFGS type quasi-Newton method with line search for symmetric nonlinear equations problems. J. Comput. Appl. Math. 2020, 357, 454. [Google Scholar] [CrossRef]
  52. Gao, P.T.; He, C.; Liu, Y. An adaptive family of projection methods for constrained monotone nonlinear equations with applications. Appl. Math. Comput. 2019, 359, 1–16. [Google Scholar] [CrossRef]
  53. Dai, Z.; Zhu, H. Forecasting stock market returns by combining sum-of-the-parts and ensemble empirical mode decomposition. Appl. Econ. 2019. [Google Scholar] [CrossRef]
  54. Dai, Z.; Zhu, H.; Wen, F. Two nonparametric approaches to mean absolute deviation portfolio selection model. J. Ind. Manag. Optim. 2019. [Google Scholar] [CrossRef]
  55. Dai, Z.; Zhou, H. Prediction of stock returns: Sum-of-the-parts method and economic constraint method. Sustainability 2020, 12, 541. [Google Scholar] [CrossRef] [Green Version]
Table 1. Numerical results for the tested Problem 1 with various sizes and given initial points.
Table 1. Numerical results for the tested Problem 1 with various sizes and given initial points.
InitialDim. SG MPRP NHZ
TimeIterFevalTimeIterFevalTimeIterFeval
x 1 10001.1616370.8112390.651029
x 2 10001.1616370.8112390.651029
x 3 10000.7713260.558270.48722
x 4 10001.2415430.859420.78735
x 5 10001.1514440.536280.48524
x 6 10001.1514440.536280.48524
x 1 50006.9618375.3213424.551132
x 2 50006.9617375.0112393.851129
x 3 50004.1611223.017242.43620
x 4 500010.4518627.3812605.991148
x 5 50006.8814443.406282.84524
x 6 50006.9615453.406282.84524
x 1 10,00027.77183821.15134215.721132
x 2 10,00027.62173619.65123914.361127
x 3 10,00014.90102011.927248.96620
x 4 10,00044.65206935.15147230.981268
x 5 10,00029.55154513.8662811.22524
x 6 10,00029.55154513.8662811.22524
Table 2. Numerical results for the tested Problem 2 with various sizes and given initial points.
Table 2. Numerical results for the tested Problem 2 with various sizes and given initial points.
InitialDim. SG MPRP NHZ
TimeIterFevalTimeIterFevalTimeIterFeval
x 1 10000.673266460.171965910.12155568
x 2 10000.643647140.112036120.10180584
x 3 10000.643436910.191955880.17174555
x 4 10001.184689410.152437410.14220709
x 5 10000.674799880.171945850.15162549
x 6 10000.654318570.161895710.14165529
x 1 500010.3872320478.1948714658.104681448
x 2 500016.87753262212.85769231012.637442205
x 3 50009.0182428497.9846914117.654421324
x 4 500018.741023326115.96929284015.519012781
x 5 500010.05122630537.5945313637.504421320
x 6 500011.3283624766.3237111226.203381103
x 1 10,00047.90960200234.59539162210.014601508
x 2 10,00082.621334406665.061023307260.7910013003
x 3 10,00050.99833246934.30516155332.325011502
x 4 10,00056.822042666839.651668507536.3116025003
x 5 10,00049.63832226831.57497149730.254361405
x 6 10,00045.70850170625.36396120222.243751106
Table 3. Numerical results for the tested Problem 3 with various sizes and given initial points.
Table 3. Numerical results for the tested Problem 3 with various sizes and given initial points.
InitialDim. SG MPRP NHZ
TimeIterFevalTimeIterFevalTimeIterFeval
x 1 10001.1616370.8112390.621028
x 2 10001.1717360.8312390.711128
x 3 10000.7711240.578270.49728
x 4 10001.2514440.889420.75732
x 5 10001.1613420.566280.48522
x 6 10001.1613420.576280.48522
x 1 50006.9817365.4213424.631132
x 2 50006.9817365.1112393.951130
x 3 50004.2910223.127242.34620
x 4 500010.5719647.4612606.251152
x 5 50006.9913423.526283.92524
x 6 50006.9913423.526283.92524
x 1 10,00027.78173621.35134215.971132
x 2 10,00027.79173619.75123915.861130
x 3 10,00015.6592611.997249.98619
x 4 10,00044.85206935.36147229.981260
x 5 10,00029.89144513.9862812.56524
x 6 10,00029.89144513.9862813.59624
Table 4. Numerical results for the tested Problem 4 with various sizes and given initial points.
Table 4. Numerical results for the tested Problem 4 with various sizes and given initial points.
InitialDim. SG MPRP NHZ
TimeIterFevalTimeIterFevalTimeIterFeval
x 1 10000.202194310.06502160.0538168
x 2 10000.282614630.06562520.0546185
x 3 10000.282243290.05341520.0332137
x 4 10000.222635290.071004210.0696399
x 5 10000.281834030.06421870.0540177
x 6 10000.282124240.06602610.0548218
x 1 50002.152634561.11482091.0547183
x 2 50002.452253781.19462240.9238169
x 3 50001.651222670.62271170.6529128
x 4 50003.412655582.591094832.49104455
x 5 50002.862904671.21532311.0744189
x 6 50002.972314771.20542341.1648213
x 1 10,0005.302785023.96451953.8342185
x 2 10,0006.262375744.27412103.8438158
x 3 10,0005.622755851.9342962.6235142
x 4 10,00018.1533359610.861175339.45109498
x 5 10,00013.523415954.34492123.8544186
x 6 10,00013.553365534.89562463.7848195
Table 5. Numerical results for the tested Problem 5 with various sizes and given initial points.
Table 5. Numerical results for the tested Problem 5 with various sizes and given initial points.
InitialDim. SG MPRP NHZ
TimeIterFevalTimeIterFevalTimeIterFeval
x 1 10000.891192890.66471990.4438168
x 2 10000.781222630.45221050.442498
x 3 10000.691302350.35482090.2838120
x 4 10000.851902490.47371650.343598
x 5 10000.751942480.55942370.4566192
x 6 10001.222254620.791743960.75142372
x 1 50002.321132601.22512210.9842172
x 2 50002.921282700.56221050.582896
x 3 50003.802284121.11472000.7944144
x 4 50003.502164241.20482060.7944142
x 5 50003.002264431.17472060.8144122
x 6 50006.574618815.063087074.25262628
x 1 10,0005.92662093.90441913.4238184
x 2 10,0006.866821851.12210549.12198
x 3 10,0005.76601814.24472043.2338132
x 4 10,00011.846922710.5482098.5244148
x 5 10,00010.55682214.02451963.8242168
x 6 10,00012.468932610.1742627.8368232

Share and Cite

MDPI and ACS Style

Dai, Z.; Zhu, H. A Modified Hestenes-Stiefel-Type Derivative-Free Method for Large-Scale Nonlinear Monotone Equations. Mathematics 2020, 8, 168. https://doi.org/10.3390/math8020168

AMA Style

Dai Z, Zhu H. A Modified Hestenes-Stiefel-Type Derivative-Free Method for Large-Scale Nonlinear Monotone Equations. Mathematics. 2020; 8(2):168. https://doi.org/10.3390/math8020168

Chicago/Turabian Style

Dai, Zhifeng, and Huan Zhu. 2020. "A Modified Hestenes-Stiefel-Type Derivative-Free Method for Large-Scale Nonlinear Monotone Equations" Mathematics 8, no. 2: 168. https://doi.org/10.3390/math8020168

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop