Solving nonlinear equation systems via clustering-based adaptive speciation di ﬀ erential evolution

: In numerical computation, locating multiple roots of nonlinear equations (NESs) in a single run is a challenging work. In order to solve the problem of population grouping and parameters settings during the evolutionary, a clustering-based adaptive speciation di ﬀ erential evolution, referred to as CASDE, is presented to deal with NESs. CASDE o ﬀ ers three advantages: 1) the clustering with dynamic clustering sizes is used to set clustering sizes for di ﬀ erent problems; 2) adaptive parameter control at the niche level is proposed to enhance the search ability and e ﬃ ciency; 3) re-initialization mechanism motivates the algorithm to search new roots and saves computing resources. To evaluate the performance of CASDE, we select 30 problems with di ﬀ erent features as test suite. Experimental results indicate that the speciation clustering with dynamic clustering sizes, niche adaptive parameter control, and re-initialization mechanism when combined together in a synergistic manner can improve the ability to ﬁnd multiple roots in a single run. Additionally, our method is also compared with other state-of-the-art methods, which is capable of obtaining better results in terms of peak ratio and success rate. Finally, two practical mechanical problems are used to verify the performance of CASDE


Introduction
Many real-world applications can be transformed into nonlinear equations (NESs), such as physics [1], engineerings [2,3], economics [4], and so on. Generally, a NES contains multiple roots. Each root is equally important because it can provide multiple selections to the decision makers so that they can make a better decision [5]. In recent years, solving NESs have received widespread attention from researchers. However, it may cause the great challenge in mathematical field, especially to locate multiple roots in a single run.
Evolutionary algorithms (EAs) draw lessons from the evolution of biological operations in nature [6]. It includes the basic operations of population initialization, crossover mutation operator, retention mechanism, and so on. Among them, differential evolution (DE) [7] is a famous optimization technique and a versatile function optimizer. Due to its features of easy-to-implement and robust adaptability, DE has been widely applied for many optimization problems [8,9]. More specifically, Li et al. presented an enhanced adaptive DE algorithm to extract the parameters of photovoltaic models [10]. Mohanmed et al. designed a novel mutation strategy to enhance SHADE and LSHADE algorithm to solve the global numerical optimization [11]. Pierezan et al. proposed a modified self-adaptive differential evolution to deal with the static force capability optimization of humanoids robots [12]. Santos Coelho et al. presented a self-adaptive chaotic differential evolution algorithm using gamma distribution to solve unconstrained global optimization problem [13]. Li et al. hybrid differential evolution algorithm with modified CoDE and JADE to improve the performance of solving the global optimization problem [14]. Civicioglu et al. adopted bezier search differential evolution algorithm to solve the numerical function optimization [15]. Zhao et al. proposed a collaborative LSHADE algorithm with comprehensive learning mechanism to slove the non-separable optimization problem and obtained the competitive results [16].
In recent years, DE is often used to solve NESs because it is insensitive to the characteristics of NESs, such as non-convexity and discontinuity, it has been applied to solve NESs [17]. However, DE encounters following two dilemma: 1) due to the lack of diversity preserving mechanism, it hardly locates multiple roots in a single run; 2) although some DE variants can obtain multiple roots, they have the problems of parameter settings, such as cluster size [18], repulsion radius [19].
To effectively solve NESs, it is important to maintain the population diversity during the evolutionary process. Clustering is perceived to be an effective methods, which can pratition the whole population into different species [20]. Thus, some clustering-based methods have been developed to find the roots of NESs [18,21]. It requires to give the number of cluster in advance. Too small the number of clusters may lose some roots. In contrast, too large cluster number may obtain the roots with low accuracy since the algorithm cannot make full use of computational resources to exploit in each cluster. However, it is difficult to specify the appropriate number of clusters for different NESs.
Parameter control in evolutionary computation plays an important role to the robustness of the algorithm [22]. Recently, some adaptive or self-adaptive parameter control methods [23][24][25] have been proposed to dynamically control the parameters according to different fitness landscapes of optimization problems. These methods can significantly improve the search ability of DE, but most of them focus on global optimum. Thus, how to develop a self-adaptive strategy suited for NESs is still a problem to be solved.
After clustering, each subpopulation might gradually converge to a narrow range of space during the evolution process. If an individual fitness less than the threshold, it is considered as a candidate solution. By this stage, continuous optimization in this subpoulation prefers the exploitation, which is able to improve the quality of root. However, it is not conducive to the population diversity and wastes more computational resource to seek the same root. Therefore, it is an urgent issue to deal with the species contained the found root.
Based on the above considerations, we combine the species clustering with dynamic cluster sizes, niche adaptive parameter control, and re-initialization mechanism to locate multiple roots of NESs. The proposed method is referred as a clustering-based adaptive speciation differential evolution (CASDE). In CASDE, species clustering with dynamic cluster sizes can alleviate the trivial task to set the cluster number. Moreover, the adaptive parameter control is employed to dynamically adjust at the niche level, thereby improving the search efficiency of the algorithm. To verify the performance of our method, we select 30 NESs from the literature [26] as test suite. Experimental results demonstrate that our method obtains highly competitive results compared with other state-of-the-art methods.
The major contributions of this paper are summarized as follows: • The species clustering with dynamic cluster sizes, niche adaptive parameter control, and re-initialization technique are combined together in a synergistic manner can greatly enhance the problem-solving capability. Among them, the species clustering with dynamic cluster sizes-maintaining the population diversity; niche adaptive parameter control-enhancing the exploitation performance in each species and avoiding trivial task to set parameters; and re-initiation technique-seeking the new roots with an reasonable using computational resource reasonably. • The niche parameter adaptive method is proposed to ensure that the parameters can be changed according to different problem landscapes, so as to improve the efficiency of the algorithm. • The effectiveness of the three different parts of components has been experimentally verified. The results show that these three parts combined together in a synergistic manner can greatly enhance the performance for solving NESs. • To further evaluate the performance of CASDE, two cases of motor system are selected to measure the performance of the proposed approach. The results have been confirmed the effectiveness of CASDE on real-world problems.
The rest of this paper is summarized as follows. Section 2 introduces background knowledge about the transformed optimization problem, different evolution. Section 3 reviews the related work for solving NESs. In Section 4, the proposed CASDE is described in detail. The experimental results and discussion are respectively given in Section 5, follow by the discussion in Section 6. In Section 7, two practical problem are used to test the performance of the algorithm. Finally, Section 8 concludes this paper.

Problem statement
Generally, a NES can be expressed as follows: where n is the number of equations, x = (x 1 , x 2 , . . ., x D ) denotes a decision vector, x ∈ S, and S ⊆ R D is the search space. Generally, where i = 1, · · · , D, x i and x i are the lower and upper bound of x i , respectively. Before solving a NES with optimization algorithm, it is commonly transformed into a single-objective optimization problem: Subsequently, solving NES is equivalent to find the global optimal of the transformed optimization problem in Eq (2.2).

Differential evolution
Differential evolution (DE) adopts three operators, including mutation, crossover, and selection, to deal with the population during the search process. Generally, a population contains NP real-valued vectors: X = {x 1 , x 2 , ..., x NP }. NP is the population size.

Mutation
Mutation operator is used to generate a mutant vector v i according to the parent population. Two well-known mutation strategies are shown below: r 1 , r 2 and r 3 respectively represent different random indices selected from the population, and they are distinct from the base index i. F is a scale factor that controls the size of difference vector. x best is the best individual in current population.

Selection
In original DE algorithm, greedy selection operator is employed to select the individual with better fitness value for a minimization problem. If the trial vector x i is better than x i , x i is set to x i ; otherwise, x i keeps unchanged.
where f (·) is the objective function to be minimized.

Algorithms for NES problems
Some stochastic approaches were proposed to solve NESs, which consist of three categories: multiobjective-optimization-based ones, single-objective-optimization-based ones, and constraint-optimization-based ones.

Multiobjective optimization based-methods
Multiobjective optimization can obtain a group of Pareto optimal solutions [27,28], which is similar to locate different roots of NESs. Therefore, more attention has been paid to the use of multiobjective optimization to solve NESs. In [29], a NES was first transformed into a multiobjective optimization problem, and then solve the optimization problem via a evolution algorithm. This kind of transformation technique ensures that multiple roots can be found in a single run. However, it may cause the curse of dimensionality if the number of equation is too large. To solve this problem, In [30], a bi-objective transformation technique was proposed to transform a NES into multiobjective optimization problems with two objectives. However, since only one decision variable is utilized to design the location function, it may lose several roots. For this purpose, Gong et al. [31] presented a weighted bi-objective transformation technique (A-WeB) for NESs. In [32], Naidu and Ojha employed a hybrid cooperative multiobjective optimization IWO to solve NESs, where multiple populations are used to deal with multiple objectives.

Single-objective optimization based-methods
In general, a NES is usually transformed into a single-objective optimization problem, as shown in Eq (2.2). Two common methods are used to solve this problem: clustering-based methods and repulsion-based methods.

Clustering-based methods
Clustering [20] can divide the population into different subpopulations, thus enabling the algorithm to search for more than one promising area. Clustering can maintain the diversity of population, so several researchers put forward clustering-based methods to locate multiple roots of NESs. In [18], the clustering technique was employed to separate the estimated locations of solutions, and then invasive weed optimization was used to calculate the exact solution in each cluster. In [33], Fuzzy Clustering Means combines Luus-Jaakola random search and the Nelder-Mead simplex method to solve NESs. In [21], Multistart and Minfinder methods based on clustering were applied for locating multiple roots of NESs.

Repulsion-based methods
The repulsion techniques can generate a repulsive regions around the obtained roots, which can increase the variety of population. Based on the repulsion techniques, several methods have been developed to solve NESs. In [34], repulsion technique combined with simulated annealing (SA) to compute critical points in binary systems. Henderson et al. [35] designed a combination of continuous SA and repulsion technique to locate multiple roots of double retrograde vaporization. In [18], a two-phase root-finder was developed to find the roots of NESs, in which invasive weed optimization located the exact roots while repulsion technique was used to preserve the population diversity. In [36], a biased random-key genetic algorithm (BRKGA) was carried out many times to detect the roots. In [19], the improved harmony search algorithm combined with the repulsion methods to solve NESs. In [26], a new approach consisting of the repulsion technique, adaptive parameter control, and diversity preserving mechanism, named RADE, was designed. Further, In [37], a general framework based on the dynamic repulsion technique and evolutionary algorithms (DR-JADE) is presented.

Constraint-optimization-based methods
The third category is the constrained optimization transformation approach. It translates a NES into a constrained optimization problem: Based on this translated optimization problem, Kuri-Morales [38] developed a penalty function to handle constraints and used a genetic algorithm (GA) to locate the root. Pourrajabian et al. [39] combined augmented lagrangian function with GA to find the optimal solution.

Motivation
Generally, the number of clustering should be given before the algorithm runs. However, it is difficult to set appropriate clustering number for NESs. Besides, after clustering, there are different subpopulations in the search space. Different subpopulations represent multiple promising regions that may have roots. Thus, how to improve the search ability in each subpopulation also needs further research. Moreover, as the search proceeds, each subpopulation will converge to a narrow range as the Algorithm 1: Speciation clustering with dynamic cluster sizes Input: population P, cluster size set C Output: a set of species Sort P in ascending order according to fitness value; while P is not empty do Random select a cluster size M from C; Select the best individual in P as a new seed; Find M − 1 individuals closest to the species seed and combine them as a species; Remove these M individuals from P end number of iterations increases. If a root is found in the subpopulation, continuous optimization of the subpopulation will result in loss of diversity and waste of computing resources. In addition, several researchers have adopted improved DE algorithm to solve NESs, which has gained extensive attention and improved the performance of problem solving [26,37]. For this reason, this paper still focuses on enhancing the DE algorithm by integrating other techniques in order to obtain satisfactory results.
Based on the above considerations, a re-initialization clustering-based adaptive differential evolution, named CASDE, is presented to solve NESs. In CASDE, a dynamical cluster sizing technique is employed to solve the problem that the number of clusters is difficult to set. Meanwhile, niche adaptive parameter setting is applied to improve the search ability in each subpopulation and avoid the trivial task of parameter settings. Moreover, the re-initialization mechanism will be triggered if a root has found in a subpopulation. For one thing, it can preserve the population diversity; for another, it is a benefit for exploration ability of the search algorithm and increases the probability to find new roots in other promising regions.

Dynamic clustering size
In [40], a speciation clustering was proposed to divide the population into different subpopulations. To reduce the sensitivity of the cluster size, dynamic cluster sizing technique [41] was introduced into speciation clustering. The effectiveness of this simple scheme was verified by experiments.
The process of dynamic clustering size (DCS) is outlined in Algorithm 1. First, the population is sorted according to fitness value in ascending order. Second, a random integer M is selected from the cluster size set C. Based on such integer, M − 1 individuals that is close to species seed are combined with the species seed to form a species. Finally, the algorithm divides the whole population into a number of species with M individuals. It is noted that M is a random integer selected from C.

Niche adaptive parameter control
After clustering, different subpopulations contain different promising regions, which have different landscape characteristics in the search space. To improve the search ability and avoid the trivial task of parameter settings in the subpopulations , a niche adaptive differential evolution is proposed in this section.
At each generation, the crossover rate CR i, j of each individual x i in cluster j is independently generated according to a normal distribution of mean CR j and standard deviation 0.1 and truncated to [0,1], µ CR is updated as follows: where c is constant number between 0 and 1; S CR j is the set of the crossover rates CR i, j in each species. mean L (.) is the Lehmer mean Similarly, the mutation factor F i, j of each individual x i in species j is independently generated according to a Cauchy distribution.
The parameter µ F updates as follow: where mean A (.) is the arithmetic mean; and S F j is the set of mutation factor F i, j at each species. Comparison with JADE [24], we modify the parameter adaptive method to some extent. S F j does update via arithmetic mean whereas S CR j is used Lehmer mean to revise. The reasons are two-fold: i) the adaptation S F j put a greater emphasis on normal mutation factor by using the arithmetic mean instead of a Lehmer mean. The arithmetic mean is helpful to propagate average mutation factors, which improve the exploitation ability in each species; ii) to improve population diversity in the species and avoid trapping in local optima, Lehmer mean is used to update S CR j .

Reinitialization mechanism
As the algorithm proceeding, the candidate solution can be found in the subpopulation. Continuous optimization in such subpopulation may pays more attention to exploitation rather than exploration. It may lead to loss of population diversity and waste of computational resource. Therefore, the reinitialization mechanism is used to solve this problem.
If a subpopulation locates a root, it is considered as convergence during the run. The root is stored into a archive A. Subsequently, all of the individuals in the subpopulation will reinitialize for maintaining population diversity. Additionally, F and CR of each individual are respectively set to 0.5 and 0.9.
It is worth mentioning that our approach may find the same root during the run. Thus, a method to update the archive is employed to avoid encountering this dilemma. Algorithm 2 outlines the process of updating the archive. In Algorithm 2, x * is one of the root in A, and is a small fixed value to avoid storing the same root in A. In this paper, is set to 0.01.
In reinitialization mechanism, we initialize the entire subpopulation instead of the found root, which has two advantages. On the one hand, it can use computational resource efficiently and benefit population diversity. On the other hand, to some extent, this mechanism prevents the remaining individuals in the subpopulation from continuing to locate the same root.

Algorithm 2: Archive updating
Input: Solution x and > 0 Output:

The proposed framework: CASDE
The framework of CASDE is outlined in Algorithm 3. One iteration in CASDE includes following steps: Step 1: Dividing population (in line 7). Algorithm 1 partitions the population into multiple subpopulations.
Step 4: Updating the archive (in line 24). If a root is located, it will be stored in the archive.
Step 5: Reinitializating the subpopulation (in lines 25,26). All individuals in such subpopulation are reinitialized; F i, j and CR i, j of each individual are also respectively set to 0.5 and 0.9.
The above iteration is repeated until the termination criterion was met.
It is worth mentioning that a hybrid mutation strategy ( "DE/rand/1" and "DE/best/1" ) is used to generate the mutation vectors (in lines [12][13][14][15]. The reason is that the hybrid strategy can balance the ability of exploration and exploitation to some extent and improve the search ability of the algorithm. Respectively update F i, j and CR i, j in S F j and S CR j Update µ F and µ CR via Eqs (4.5) and (4.2); end Find the minimal fitness value f (min) in j-th species if f (min) < τ then Record the corresponding individual in the archive via Algorithm 2 Re-initialize the individuals in j-th species Reset the F i, j and CR i, j to 0.5 and 0.9 in j-th species end NFE = NFE + NP; end

Experiment results and analysis
In this section, we mainly focuses on experimental results and analysis, including the impact of different parts in CASDE, comparing CASDE with state-of-the-art methods.

Test problems
To evaluate the performance of different methods, frequently-used NESs (F01-F30) benchmark problems are selected as a test suite. They have different characteristics, and some derive from real-world applications, such as multiple steady states problem (F08) [42], robot kinematics problem (F17) [43], and molecular conformation (F23) [44]. The test problems are briefly introduced in Table 1.

Performance metrics for NESs
To evaluate the effectiveness of algorithms, two performance Metrics in [26,37] : (Root ratio (RR) and Success Rate (SR)), are employed in this paper.
• Root ratio (RR): It is the ratio of found roots number to total roots number over multiple runs within NFEs max : where N r is the number of runs; N f,i is the found roots number in the i-th run; NoR is the total roots number of a NES. • Success rate (S R): It is the ratio of successful runs * to total runs : where N r,s is the number of successful runs.

Influence of different parts in CASDE
CASDE contains three parts, i.e., dynamic clustering sizes, niche adaptive parameter control, and the re-initialization mechanism. This subsection mainly dedicates to discuss the impact of different parts of CASDE on performance of solving NESs.
1) CASDE. The dynamic clustering sizes, niche adaptive parameter control, and the re-initialization mechanism were combined.
2) DCS-DE. It is the algorithm that DE combines with dynamic clustering sizes.
3) CASDE/DA. The re-initialization mechanism was removed from CASDE; 4) CASDE/DR. The niche adaptive parameter control was removed from CASDE. F and CR were respectively set to 0.5 and 0.9; 5) CASDE/AR. The dynamic clustering sizes were not used in CASDE. The clustering size were set to 5.
The detailed experiment result in terms of RR and S R is shown in Table A1. It is obvious that CASDE obtained the best average RR value, i.e., 0.9951 and the best average S R value, i.e., 0.9556. Additionally, CASDE successfully solves 26 out of 30 NESs over 30 independent runs. In contrast, CASDE/AR, CASDE/DR, CASDE/DA and DCS-DE. successfully solve 23, 20, 15, 13 NESs over 30 independent runs, respectively.
The statistical test results acquired by the multiple-problem Wilcoxon's test are showed in Table 3. In addition, the ranking results obtained from the Friedman's test are shown in Table 2. From Table 3, CASDE consistently offers significantly better results than CASDE/DR, CASDE/DA, and DCS-DE due to the fact that p-values are less than 0.05 in all the cases. Additionally, CASDE also obtains the best ranking as shown in Table 2. In what follows, we try to analyze the influence of different parts of CASDE on the performance of solving NESs.
• DCS-DE: from Table A1, DCS-DE can successfully solve 13 out of 30 NESs, i.e., F01, F05, F06, F09-F11, F18, F20, F21, F26-F30. However, one feature of them is that they contain a small number of the known roots. For example, the known roots of this kind of NESs are no more than 8. Thus, DCS-DE is more suitable for solving these NESs. * A successful run is considered as a run where all known optima of a NES are found.
• CASDE/DA: The combination of dynamic clustering size and niche adaptive parameter control can improve the performance of the algorithm to some extent. As shown in Table A1, RR and S R values obtained by CASDE/DA were higher than DCS-DE. • CASDE/DR: From Table A1, the re-initialization method can improve the performance of the algorithm. For example, RR and S R values obtained by CASDE/DR were 0.9365 and 0.8156, respectively, significantly higher than those obtained by DCS-DE. • CASDE/AR: Similarly, the combination of niche adaptive parameter control and re-initialization mechanism can also improve the performance of the algorithm. • CASDE: It obtains the best RR and S R values. The reasons are as follows: 1) dynamic clustering size can reduce the sensitivity of clustering number and maintain population diversity; 2) niche adaptive parameter control reduces the tedious task of parameter settings and improves the search ability of the algorithm; 3) the re-initialization mechanism can improve the population diversity and detect the new roots.

Comparison with other state-of-the-art methods
CASDE is compared with the following state-of-the-art algorithms: • Decomposition-based differential evolution with reinitialization: DDE/R. • A weighted biobjective transformation technique for NESs: A-WeB [45].
The parameter settings are given in Table 4. All experiments are run 30 times for fair comparison. In addition, the advantage of CMA-ES, JADE and COA are to find the global optimal. In order to improve the performance of solving NESs, we combine niche techniques with these algorithms.  Tables 5 and 6, respectively. It is obvious that CASDE obtains the highest average values in both RR (0.9951) and S R (0.9556). Moreover, compared with other methods, it also achieves the highest ranking in Table 5. Meanwhile, from Table 6 obtained by the Wilcoxon test, CASDE significantly exceeds other methods in terms of RR and S R except DDE/R and FONDE since all p−values are less than 0.05.
Moreover, the results † obtained by Nemenyi test are listed in Table 7. We can observe the results of the best classified algorithms in the competition, like CASDE, we see that there are significant differences with algorithms like A-WeB, RADE, EMO-MMO,N-CMA-ES, N-JADE and N-COA but this difference is not significant for DDE/R, RADE, DREA, FONDE, and ANDE. More specific, CASDE has better performance than DDE/R, RADE, DREA, FONDE, and ANDE according to the results achieved by the Friedman test and the Wilcoxon test. † In order to save space, we will only list the comparison between CASDE and other methods

Discussion
As shown in Section 5.2, the superior performance of CASDE has been evaluated. In this section, the influence of fixed clustering size, the impact of different parameter settings, and different mutation operators are studied.

Influence of fixed clustering size
An important aspect concerning CASDE is the clustering size. Therefore, this section mainly discusses the impact of different clustering sizes on CASDE.
To make it simple, we respectively set the fixed sizes as 5, 10, 20, which is for the precise division. They are employed to replace the dynamic clustering size previously used in CASDE. Therefore, three CASDE variants, i.e., CASDE-5, CASDE-10 and CASDE-20, are developed ‡ .
The detailed experiment results are shown in Table A4 for both RR and S R in the supplementary file. From Table A4, compared with CASDE-5, CASDE-10, and CASDE-20, CASDE also obtains the best average RR and S R values. Furthermore, the case where the fixed clustering size is equal to 20 is worth discussing. This large clustering size leads to a small number of species. Therefore, it is not conductive to solving the NESs contained many roots. From Table A4 , it can be seen that CASDE-20 successfully solve 16 out of 30 NESs over 30 independent runs. One characteristic of these NESs is that they contain few roots. In contrast, if the NESs has many roots, such as F03, F04, F13, F16, F17 , F23, and F24, CASDE-20 is difficult in locating all the roots in a single run. Tables 8 and 9. We can see clearly that CASDE achieves the best ranking value from Table 8. Meanwhile, from Table 9, CASDE significantly outperforms CASDE-20 at α = 0.05 for RR and S R.

The statical results from the Friedman test and Wilcoxon test are shown in
Based on the above analysis, comparison with the fixed clustering size, the dynamic clustering size is able to improve the performance of algorithm. Besides, dynamic clustering sizes alleviates the trivial task to give the proper clustering size for different NESs problems.

Study on different parameter settings
To improve the search ability and avoid the trivial tasks of parameter setting, niche adaptive parameter setting is used in CASDE. In this subsection, we verify the effect of the static parameter settings (F and CR) on the performance of CASDE. Therefore, the adaptive parameter setting is removed from the CASDE, and nine diverse sets of parameters are used in CASDE, i.e., The detailed results based on RR and S R values are respectively shown in Tables A5 and A6 in the supplementary material. Additionally, the average ranking obtained by the Friedman test are reported in Table 10 and the results derived from the Wilcoxon test are shown in Tabel 11.
From Tables A5 and A6, we can see that (F = 0.1, CR = 0.9) get the higher average RR and S R values over 30 independent runs. Besides, different parameters can influence the performance of RCDE for solving NESs. For example, three sets of parameters, such as (F = 0.1, CR = 0.1), (F = 0.5, CR = 0.1), and (F = 0.9, CR = 0.1) obtain poor results due to a small CR value. Therefore, we can conclude that the parameter settings has significant impact on the CASDE. Its optimal value is difficult to give in advance and problem-dependent. From Table 10, CASDE get the highest ranking for RR and S R via the Friedman test. Moreover, Table 11 shows that CASDE is better than the static parameter settings, because it can obtain higher R + than R − values in all cases. In general, the proposed niche adaptive parameter setting is effective. More importantly, it can avoid setting the proper values of F and CR for different NESs. Table 11. Results obtained by the willcoxon test for algorithm casde in terms of RR and S R compared with casde-5, casde-10, and casde-20.

. On different mutation operators
In CASDE, "DE/rand/1" and "DE/best/1" as shown in Eqs (2.3) and (2.4) are selected randomly during the search process. The hybrid mutation strategy can contribute to the balance of exploration and exploitation. In this subsection, we replace the hybrid mutation strategy with single mutation operator to evaluate the effectiveness of CASDE . To this end, we focus on two CASDE variants: 1) CASDE-1, i.e., CASDE with "DE/rand/1"; 2) CASDE-2, i.e., CASDE with "DE/best/1". The parameter settings are remained the same for two CASDE variants.
The detailed experiment results are shown in Table A7 in the supplement file. Compared with CASDE-1 and CASDE-2, CASDE obtains the highest average RR and S R values. In addition, several interesting phenomena appear in the experiment results. For F04, F12, and F23, CASDE obtains better RR and S R values than CASDE-1; for F13, CASDE-1 gets the best result; whereas CASDE-2 shows the best performance for F16. Therefore, different test instances require reasonable use of mutation operators to improve the effectiveness of algorithm. Hence, this hybrid strategy is used to remedy the drawback to some extent.
The statistical results derived from Friedman and Wilcoxon test are reported in Tables 12 and 13. We can observe that CASDE achieves the best ranking from Table 12. Meanwhile, from Table 13, CASDE is better than CASDE-1 and CASDE-2. settings, because it can obtain higher R + than R − values in all cases.

Solving real-world NES problems
In the previous experimental results, the superiority of CASDE is verified. In this section, the effectiveness of CASDE on real-world problems is studied. Two cases of motor system [50] are selected to measure the performance of CASDE.

Test cases
1) Case-1. This case mainly focuses on the solution of internal variables in steady-state operation of synchronous generator with magnetic saturation. For example, saturation characteristics of a synchronous generator, s = 0.047 6E 12 q , synchronous reactance x d = x q = 2.264, reactance of armature reaction x ad = x aq = 2.104, field winding self-inductance x F = 2.209, field winding resistance r F = 0.0008, stator resistance r = 0.02, steady state operation active power P = 0.8, whereas reactive power Q = 0.4, terminal voltage U = 1.0∠0. It needs to calculate the internal variables of the unit.
Firstly, the following equation can be derived from the relationship between the internal variables during the steady-state operation of the synchronous generator and the saturation factor of the magnetic circuit: x ad(sa) (1 + s) = x ad(un) x ad(sa) I = E q where δ is power angle; ϕ denotes power factor angle; E q is the air gap voltage; x ad(sa) and x ad(un) respectively illustrate saturated direct axis armature reactive reactance and Unsaturated direct axis armature reactive reactance; I is resultant current, and s is saturation factor. Let: Thus, the motor problem can be transformed into a NES:  Table 14. Table 14. Roots of Case 1. 2) Case-2. It is used to solve the circuit model parameters of synchronous generator d-axis equivalent circuit. Several known parameters of a synchronous generator are described as follows: 497s, T * d = 0.035s, r F = 0.000856, x e = 0.232. The work is to calculate the parameters of D -axis equivalent circuit model (Canay model). Follow this, the equations can be obtained according to equivalent circuit model: Let: x 1 = x de , x 2 = T de , x 3 = T * de , x 4 = x e , x 5 = x de , x 6 = x * de, x 7 = b, x 8 = r Fce , x 9 = x Fce . Application Case-2 can be transformed into a NES as follow:  Table 15. Table 15. Roots of Case-2.

Results
In this subsection, CASDE is compared with the above ten methods, the parameter settings shown in Table 4. To make a fair comparison, each method is executed over 30 runs. For the two problems, NFEs max = 200, 000.
The detailed results are reported in Tables 16 and 17, where the best results are listed in bold. It can be observed that CASDE obtains the highest average RR values, i.e., 0.9750 and the highest average S R values, i.e., 0.9000. Thus, comparison with other ten methods, our proposed approach, CASDE, is effective and efficient for solving the two motor systems problems.

Conclusions and future work
Solving NESs is a very challenging task due to the fact that it needs to locate multiple roots of NESs in a single run. To address this issue, we propose a re-initialization clustering-based adaptive differential evolution, named CASDE, in which the dynamic clustering sizes, niche adaptive parameter control, and re-initialization mechanism were combined together to solve NESs effectively. The performance of CASDE was verified by 30 NESs selected from the literature. Experimental results demonstrated that CASDE is able to locate multiple roots in a single run. In addition, comparison with other state-of-the-art methods, our approach also obtains significant performance. Moreover, we execute extensive experiments to analyze the performance of our approach, as well as the effectiveness of different parts of CASDEand the influence of different parameter settings. From the experiments, we can testify that CASDE can be considered as an effective alternative to solve NESs.
In the near future, we plan to use the fusion of reinforcement learning and evolutionary algorithm to solve NESs. By evaluating the actions using different strategies and parameters in a certain environment, the algorithm will have the characteristic of autonomous decision. Moreover, we will employ CASDE to solve other complex re-world NESs.   Table A5. Influence of different parameter setting in CASDE with respect to the peak ratio.