A Novel Selection Approach for Genetic Algorithms for Global Optimization of Multimodal Continuous Functions

Genetic algorithms (GAs) are stochastic-based heuristic search techniques that incorporate three primary operators: selection, crossover, and mutation. These operators are supportive in obtaining the optimal solution for constrained optimization problems. Each operator has its own benefits, but selection of chromosomes is one of the most essential operators for optimal performance of the algorithms. In this paper, an improved genetic algorithm-based novel selection scheme, i.e., stairwise selection (SWS) is presented to handle the problems of exploration (population diversity) and exploitation (selection pressure). For its global performance, we compared with several other selection schemes by using ten well-known benchmark functions under various dimensions. For a close comparison, we also examined the significance of SWS based on the statistical results. Chi-square goodness of fit test is also used to evaluate the overall performance of the selection process, i.e., mean difference between observed and expected number of offspring. Hence, the overall empirical results along with graphical representation endorse that the SWS outperformed in terms of robustness, stability, and effectiveness other competitors through authentication of performance index (PI).


Introduction
e basic idea of genetic algorithms (GAs) was originated by John Holland in 1960s and was further developed in his book "Adaptation in Natural and Artificial Systems" published in 1975 [1]. GAs are the most efficient procedure to understand and solve problems for which have limited information. ese algorithms are able to effectively handle both unconstrained and constrained optimization problems depending on a process of natural selection through biological evolution. e working mechanism of GAs is linked with a search space that contains all possible solutions. Each part of the search space represents one sufficient solution, and its fitness values will be marked by these sufficient solutions, and a set of these solutions is called a population. A set of sufficient solutions will be carried on to the next generation, but weak solutions will be dead based on "survival of fittest" by Darwin's theory of evolution [2].
ere are two significant points in the GA process: one is starting point initialization in search space and other is assigning of fitness function [3]. GA starts with the initialization of a population or potential solutions of the problems.
is initialization is represented by the chromosomes (individuals), which are a set of genes, with each gene carrying the features of dataset. ese chromosomes have their own fitness values depending on the objectives function, so it is very important to determine the solvable objective function.
GA works with the set of solutions and not the decision variables like the other statistical techniques [4][5][6]. After creating the solutions which are represented by the chromosomes, each of these chromosomes will be evaluated for their fitness depending on the fitness function. Chromosomes that have the fittest value will survive to the next generation. e fitness function depends on the objective of the problem statement. Most of the fitness will be made equal to the objective function value. If the problem statement is to have a minimum cost of some product, then the optimization function here is to find the lowest of the fitness values [7]. Specification in the fitness function is one of the crucial problems in GA because it will determine which chromosomes can survive to the next generation and which will be eliminated from the population.
During the GA process, the feasible solutions in the search space cannot be obtained without reproduction and recombination. Reproduction phase of GA is initiated by the selection of better individuals that will produce new offspring for the next generation with the hope that the next generation will be improved. e core idea of the selection procedure is to enhance the quality of solutions by giving preference to the most suitable individuals and avoiding bad individuals. By combining the current population's solutions, the new population will hopefully contain better solutions and avoiding loss of genetic material. Furthermore, to make the process more reliable, some of the features in the solutions will be mutated or changed with minor probability. e purpose of crossover and mutation will definitely support to generate better a population than the old one [2,3].
As we can notice from Figure 1, GA is a stochastic-based heuristic search procedure which is used to set problembased parameters and making decisions about the following: (i) Generate initial population (ii) e process of parents' selection for reproduction of offspring (iii) Crossover and mutation of individuals (iv) Predefine stopping criteria e function of elitism is to make sure that the good and strong chromosomes can be carried to the next generation by storing them outside the current population. Elitism is helpful in presenting the best solution during the process of crossover and mutation [8,9]. is can be applied in many ways; one way is to combine both parents and child to produce a new population with all competing to survive to the next generation. e use of elitism can help to converge at a global optimal solution [10].
In order to converge at global optima and avoid the local stagnation, a systematic tradeoff mechanism between exploration and exploitation is compulsory. Most of the stochastic-based heuristic search algorithms try to create a balance between two contradictory measures of their performance: exploration (population diversity) and exploitation (selection pressure). Exploration means the capability of an algorithm to search or explore every region of the possible search space and exploitation means to converge at the optimum solution as soon as possible.
e suitable adjustment between exploration and exploitation increases the performance of the GA. In this paper, we will handle this problem with help of the proposed selection procedure. e aim of the selection procedure is to exploit the suitable features of fitted individuals in the context of improved solutions, which technically guide the GA for the convergence to a feasible solution for optimization problem [2]. e GA is widely used in various fields of human endeavor including machine learning [11], scheduling [12], signal processing [13], energy [14], robotics [15], manufacturing [16], mathematics [17], routing [18], and many more. e rest of the paper is organized as follows: concise detail about some conventional selection schemes is discussed in Section 2. Mathematical derivation along with proposed selection scheme is presented in Section 3. Detailed description about benchmark functions are defined in Section 4, while simulated results by using well known benchmark functions along with evaluation tools are revealed and discussed in Section 5. Conclusions of the study are presented in the last section of this paper.

Review of Genetic Algorithm Selection Process
ere are no specific criteria or theoretical justification to choose an appropriate selection scheme for various problems. is can be an alarming situation due to the application of an inappropriate selection technique on numerical data which can lead to poor performance of the GA regarding reliability of the results. In this section, we will review the reproduction process of individuals and also will present the performance comparison regarding shortcomings and advantages of different selection schemes. Hence, there are several schemes for the selection of individuals from the population. So, for the purpose of conducting comparative performance evaluation studies, numerous GA selection techniques exist in the literature: roulette wheel selection/fitness proportional selection (RWS), linear rank selection (LRS), tournament selection (TS), stochastic remainder selection (SRS), etc.
Roulette wheel selection is another name of fitness proportional selection.
is selection technique uses the proportion of the solutions which will affect the area in the wheel. e higher proportions will have a larger area in the wheel and vice versa. In RWS, the wheel will be partitioned according to the probability where the higher probability will have a bigger area and the lower probability will have a smaller area. In this selection technique, a circular wheel connected with a fixed pointer is used for choosing different individuals, which is on the border of the circular wheel [2]. e first individual is selected when the area of the circular wheel comes in front of the fixed pointer. e second individual is selected through the same procedure, and this procedure will be replicated till the selection of last individual. It is very obvious that the individual with highest fitness value will acquire the greater portion on circular wheel and will have a higher possibility of arriving in front of the wheel's fixed pointer when the wheel is spun. erefore, the probability p i of selecting individuals is directly proportional on its fitness value [19]: where f i is a fitness value of i th individual and W denotes the size of population.

Computational Intelligence and Neuroscience
RWS is a biased selection because the chance of the small area being selected is very low [2]. is selection scheme still has an advantage where the weaker solutions have a limited chance to be selected and may survive in the next generation [20,21].
In the literature, there are some other selection techniques to overcome the above shortcomings. Hence, LRS is one of the most popular selection techniques, which is more beneficial to handle premature convergence issue as compared to RWS. is selection scheme is focused on rankbased selection procedure, which provides a better opportunity to weaker individuals in the context of uniform scaling. e chromosomes are selected with the probability p i that is linearly proportional to the rank of chromosomes.
where i is the rank of individual according to its fitness value and W is the size of population. Furthermore, φ + and φ − are parameters representing the best and worst selection of individuals linked with their ranks, respectively. For the estimation of the above function in equation (2), the constraints are φ + � 2 − φ − and φ − ≥ 0. e limitation of this scheme is slower convergence to optimal solution because difference between the best fitted chromosome and other chromosome is not significant due to closeness of values. So, LRS is more beneficial than other techniques due to standardized scaling procedure and also useful to overcome the problem of premature convergence [22]. TS is an extensively used selection technique in GAs. It is also applicable in most of the applied research problems.
is selection scheme can be implemented competently and is amenable to parallelization [21]. e simplest form of TS is based on randomized selection of two individuals and conducting a competition to decide which chromosome will win and get selected for the mating pool, and then comparing it to a predetermined selection probability p i . Hence, the predetermined selection probability for individual p i for (t − 1) tournament is given by where W is defined as the population size and t is size of the tournament. For the binary tournament, t � 2, and for large tournament, t > 2. e probability of parameters provides a suitable procedure for adjusting the selection pressure. e TS can also be further extended to involve more than two individuals if desired [22]. e basic idea of the SRS technique is based on the deterministic sampling technique [20]. Each chromosome (individual) in the population has the selection probability based on its comparative fitness value. e SRS uses a concept of removing or copying the strings based on the values of the reproduction counts. e process is done by computing the reproduction count associated with each string. At first, the probability of selection p i , where f i is a fitness value of i th individual. Hence, the expected number of individuals in the mating pool is calculated as population size W:  Computational Intelligence and Neuroscience Integer portion of p i is used to choose as an individual deterministically and then uses RWS or flipping a coin to deal the remaining fractional portion and to fill the rest of portion in mating pool. For example, if the value of p i � 3.8 as described in Figure 2, which means that three copies of chromosomes are directly placed in the mating pool because of integer portion, then the fractional portion of the parents are chosen stochastically.
ere are two methods to deal with remainder portion of p i ; the first is SRS with replacement and other is SRS without replacement. In SRS with replacement, the remainder part of p i is used to size the portion of RWS process. e resultant probability is proportionate of fractional portion of its scaled value.
is selection mechanism provides maximum opportunity of selecting best-fitted individuals of the population. In SRS without replacement technique, flipping a coin determines whether the fractional portion of scale value receives another copy or not.

Defining Problem.
In the above context, most of the operators follow one extreme, i.e., exploitation or exploration. erefore, for achieving the optimal solution, it is more beneficial to adjust selection pressure which maintained population diversity during the selection process. More illustratively, we considered RWS and LRS which are both extremes in selection of individuals [22]. Generally, LRS mainly focuses on maintaining population diversity (more technically known as exploration) by compromising selection pressure resulting delayed convergence and RWS emphasizes on selection pressure (known as exploitation) with shortcomings of premature convergence.

Proposed Scheme (Proportionate Selection).
To overcome the shortcomings of conventional selection schemes, we proposed a balanced selection approach associated with suitable tradeoff between exploitation and exploration, which basically decreases the effect of selection pressure and assure some genetic diversity within population. In other words, it will be a fine adjustment between selection pressure and loss of population diversity.
Here, the newly proposed selection scheme will be helpful in improving the search space through proportionate probabilistic approach. e initiation of probabilistic weights to individuals will definitely introduce greater diversity in the population, thus offering better solutions with sustainable convergence speed. us, the new selection scheme creates a sustainable adjustment between exploitation and exploration. Hence, a modified selection scheme is going to be proposed, named stairwise selection (SWS). Its objective is to overcome the disadvantages of other selection schemes by providing a comparatively better opportunity to the weak individuals for maintaining population diversity.
is newly selection mechanism is designed in such a way that the resulting generation has a limited chance of deterioration. e working phenomenon of SWS proceeds by assigning ranks to all individuals from worst to best criterion according to their fitness values. e ranked population of size W is given below: First, we divided the whole population into five equal portions as Hence, the selection probability of each individual "i" is according to the following function: where q 1 + q 2 + q 3 + q 4 + q 5 � 1 and the suitable probabilities weights are revealed in e pseudocode of SWS is given in Algorithm 1.

4
Computational Intelligence and Neuroscience e performance of the GA is usually examined through the optimum value and number of generations required to get the optimum solution. For visual understanding and close comparison of different selection schemes, we considered a population of ten individuals. Figure 3(a) shows that the individuals "1" to "3" have a limited chance to get selected because of the small portion in the roulette wheel instead of "7" to "10" with higher portion. Hence, current distribution of individuals in RWS increases selection pressure and reduce population diversity. Conversely, the distribution of LRS for individuals will delay the convergence due to uniform scaling. Figure 3(c) shows that TS is giving more weight to individuals "1" to "3" as compared to RWS, which means that TS is somehow managing selection  Generate the individuals of size W Sort in ascending order after fitness Create a table t t ⟵ 1: ALGORITHM 1: e pseudocode of stairwise selection scheme.
Computational Intelligence and Neuroscience pressure and population diversity. Now, the newly proposed selection scheme, i.e., (SWS) has a better control over the above two extremes, i.e., selection pressure and population diversity. Because individuals "1" to "3" have a sufficient chance to be selected and "7" to "10" also have an adequate representation, there is an adequate balance between exploitation and exploration. For more realistic visual comparison, we considered a population of hundred individuals. Figure 4 clearly visualizes that the graphical line of SWS occurs in between conventional selection schemes, which reflects that this novel selection scheme seems to have a better control over selection pressure, and it is more beneficial to maintain population diversity. In other words, it would be a perfect tradeoff between exploration and exploitation.

e Sampling
Methodology. An efficient sampling procedure is required to select individuals for mating process through the mechanism of two-step selection. is   Computational Intelligence and Neuroscience sampling procedure fills the mating pool with copies of individuals of the given population, while respecting the selection probabilities p i , such that the observed and expected number of individuals are equal. Among the widely used sampling procedures, we commonly used the roulette wheel sampling technique (or Monte Carlo sampling) for evaluating the efficiency of the newly proposed SWS operator.

Chi-Square Goodness-of-Fit
Measure. χ 2 is used as a tool to measure the mean difference between observed and expected number of offspring. is measure was first time introduced by Schell and Wegenkittl [23] for average accuracy. Initially, there are k mutually exclusive classes Let ε j � i∈C j e i denote the cumulative expectation and O j � i∈C j o i represent the observed/actual copies of individuals in the mating pool followed by the sampling process. Preferably, the order of ε j should be W/kfor 1 ≤ j ≤ k. So, on average, each class contains equal number of individuals, and there should be at least 10 number of classes to attain the required accuracy. Schell and Wegenkittl [23] suggested the Chi-square test as a measure to evaluate the efficiency of the sampling procedure as follows: In the context of the roulette wheel sampling scenario, the abovementioned constraint, i.e.,ε j ≥ 10, χ should follow Chi-square distribution with k − 1 degree of freedom. is distribution is asymptotic of χ under multinomial distributed o i when W ⟶ ∞. According to the present research study, the concern-fixed parameters are the population size W � 100, number of classes � 10, and total number of tests s � 100. e results in Table 1 reveal the probability distribution of SWS along with corresponding cumulative expectation, which are close to W/k � 100/10. We used χ SW,R to evaluate the results of χ. In χ SW,R , SW denotes the proposed operator that assigns selection probabilities to the individuals and R represents a technique of sampling algorithm. Mainly, this test is used to estimate the expectation and its variance. e population generated randomly with predefined specific individuals and used the probability distribution R to assign them probabilities for the process of selection followed by sampling procedure R is applied to obtain instance of O j and χ SW,R , respectively. e sample mean and variance can be obtained through sequence (χ SW,R ) with 1 ≤ w ≤ s as given below: For the purpose of evaluation, this technique is compared with theoretical distribution χ 2 k− 1 at 99% confidence level. e mean and variance of χ 2 distribution are k − 1 � 9 and 2(k − 1) � 18 for 10 classes. Hence, the corresponding estimates of e and σ 2 are 9.1025 and 19.8583, respectively. e above estimates are almost similar and comparatively more accurate in terms of symbolic representation between assigning probabilities to the individuals and the number of copies related to their respective probabilities coming in the mating pool. e simulated results authenticate the overall performance of the sampling procedure with respect to probability distribution of SWS. Hence, the roulette wheel sampling technique provides the empirical distribution function that cannot be significantly different from theoretical distribution χ 2 k− 1 regarding e and σ 2 estimates.

Benchmark Functions
ere is not a rule of thumb for the evaluating the performance of the GA by choosing an appropriate optimization function. erefore, the performance of the algorithm is based on the nature of the problem regarding variation rate in objective function, the number of local optima, etc. [24]. A multimodal function has at least two local optima. e efficient search procedure must be proficient of eliminating the region around local optimum in context of the search for global optima. e scenario becomes more complex in situation of random distribution of local optima in the search space. e dimensionality of the search space is another significant factor which makes the problem more complicated. A comprehensive study regarding dimensionality problem and its characteristics was carried out by Friedman [25]. During the search process, value regarding global optimum needs to be obtained efficiently. Hence, the areas close to local minima must be avoided as much as possible. If the local optima are randomly distributed in the search area, then it is considered to be a most difficult problem. e Computational Intelligence and Neuroscience optimization process focuses on obtaining the global optimum point; consequently, the regions nearby local optima should be circumvented because the optimization process might be stuck at local optima and then local optima are considered to be as global optima. To evaluate the performance and sustainability of the proposed selection operators, we used ten unimodal, multimodal, separable or nonseparable, convex, and continuous benchmark functions. Table 2 presents the list of benchmark functions [16,[26][27][28][29][30][31][32][33][34][35][36][37][38][39][40][41][42] utilized to appraise the efficiency of the suggested evolutionary methods. Hence, the benchmark function's name, limit, properties, and fitness function are presented in Table 2.
ese benchmark factions have varying complexities that are most commonly applied in many comparative studies. e necessary details regarding these benchmarks are given below:

Experimental Setup.
In this section, we focused on the experimental results of four conventional and one proposed GA selection schemes. e overall efficiency of these selection schemes can be influenced by the use of fixed parameters with additional experimental conditions. Hence, the suitable values for fixed parameters such as population size, crossover and mutation probability, number of generation, and scaling function. Table 3 shows the value of fixed parameters that are used for optimization problems. e performance of these selection schemes is evaluated on ten benchmark functions using MATLAB version R2015a. e simulated results of these runs are obtained in terms of mean and standard deviation (S.D). An independent t-test is also executed to examine the significant difference between different selection schemes. e p value along with mean and S.D of thirty runs are reported in subsequent tables. e sign of " * " indicates the significant difference with the proposed technique and "a" defines the significance difference with reference technique.

Experimental Results.
In this experimental study, the optimum values regarding GA were obtained through screening experimentation and trial run. e algorithms were executed thirty times, and the mean value and standard deviation are taken as final results. All experiments are terminated in this study when number of generations achieved the maximum numbers of generation. e basic objective of this study is to make a comparison between different conventional selection schemes with the proposed one in the context of optimal solution by using benchmark functions. e overall statistical results of Table 4 clearly show that SWS obtained a minimum mean value and low S.D compared to other selection techniques from 10 to 100 dimensions. But there is a nonsignificant difference between SWS and TS at some benchmark functions. As we can notice for Axis Parallel Hyper Ellipsoid function, when dimensions of the study increase from 10 to 100, the average rate of change is in between 706 and 3052 because of function complexity. Hence, the minimum average rate of change is 706 in SWS and maximum is 3052 under RW at lower dimensions. e p value of t-tests further reduced with increase in the dimensions of experiment that actually tends toward significance of the results. About Colville function, SWS is the best-performing selection technique with the mean value of 1.39 at 10 dimensions with highly significant differences. When we increase the dimensions up to 100, the optimum value increases up to 5940 in the Colville function. Hence, the average rate of change is much high due to complexity. According to Table 4, the results of Ellipsoidal family function reveal that the proposed selection scheme (SWS) is the best-performing approach with minimum mean value of 0.0000 at lower dimensions, but at higher dimensions, the average rate of change is 187286 which is at the higher side. Another unimodal function is Rosenbrock; its statistical results about SWS are close to the theoretical optimum value which means that the proposed selection technique is efficiently handle complex problems at higher dimensions.
e average rate of change in the Schaffer function is considerably low which shows that SWS efficiently performs at higher dimensions. e optimum value of SWS is ranging from 4.14 to 45.61 in the Schaffer function for 10-100 dimensions.
According to the results of the Beale function, Table 5 shows that the optimum value is obtained through TS. Moreover, SWS has significant difference with LRS but not with RWS, TS, and SRS at lower dimensions. When we increase the dimension, p value will also reduce from 0.9807 to 0.0000, and the average rate of change of TS is 583 which is considered close to the theoretical optimum value as compared to other selection techniques including SWS. e SWS also achieves the minimum average rate on the Bohachevsky function, i.e., 98. Furthermore, the average rate of change is 84 for SWS which is the lowest in all other schemes from low to high dimensions. Moreover, the results of Bohachevsky benchmark function in Table 5 reveal that SWS distinctly performs better than all other selection schemes in terms of least empirical values. Moreover, by increasing the dimensions of experiment, SWS significantly differs at higher dimensions and show nonsignificance difference at lower dimensions with TS and RS.
According to the results in Table 5, SWS is considerably close to the theoretical optimum value under Drop-wave and Egg-holder benchmark functions, but the average rate of change in Egg-holder function is much higher when we increase the dimensions as compare to Drop-wave function. Hence, SWS efficiently handles selection pressure and makes In the context of above discussion, it is demonstrated the substantial amount of effectiveness of the newly proposed selection technique over the standard GA techniques. Additionally, SWS selection technique ensured a broader and comprehensive search and avoided premature convergence to the optimum solutions in unimodal and multimodal benchmark functions. e newly proposed technique efficiently handles the problem of selection pressure and extends the diversity by intensifying the scope of the search process. is scheme is also reducing the possibility of less favorable solutions at higher as well as lower dimensions. In addition, the proportionate selection strategy ensures that best solutions are always carried forward to the next generation. In fact, SWS enhances the exploration of future generations and reduces the chance of premature convergence at local minima.

Overall Performance.
e empirical results of conventional selection schemes (RWS, TS, LRS, and SRS) along with proposed SWS are evaluated on ten benchmark functions. e statistical results of Table 6 reveal that SWS outperforms in almost all benchmark functions regarding robustness, stability, and effectiveness of the solutions.
TS is the second best selection scheme because its optimum values are considerably close to SWS and sometimes have nonsignificant difference between these two. SWS equally efficient for unimodal and multimodal functions but the average rate of change is comparatively high in multimodal functions. Furthermore, SWS also performs efficiently when increasing the dimensions of experiment from 10 to 100 and also establish a suitable adjustment between exploitation and exploration. e influence of results in Table 6 confirms that SWS has a firm grip on controlling selection pressure and population diversity.

Performance Index (PI).
After descriptively evaluating the performance of stairwise selection operator with others, our next goal is to make a comparison between GAs' selection schemes based on relative performance index (PI) defined by Bharti [43].
is performance index was specifically used to analyze the behavior of some controlled stochastic search techniques. e PI is a widely used mechanism for comparing population-based heuristic algorithms [44,45]. e PI can be mathematically derived in following way:  where where M i � mean value of objective function for i th optimization problem, LM i � least mean value of objective function obtained by all algorithms for i th optimization problem, S i � standard deviation of objective function for i th optimization problem, LS i � least standard deviation value of objective function obtained by all algorithms for i th optimization problem, MAE i � the value of mean absolute error of objective function for i th optimization problem, LMAE i � least mean absolute error value of objective function obtained by all algorithms for i th optimization problem, W p � the total population to be analyzed.   θ 1 , θ 2 , and θ 3 (θ 1 + θ 2 + θ 3 � 1 and 0 ≤ θ 1 , θ 2 , θ 3 ≤ 1) are weights assigned to three statistics that were considered, respectively.
In the context of the above definition, it is revealed that PI is a function of θ 1 , θ 2 , and θ 3 , respectively. Since θ 1 + θ 2 + θ 3 � 1, one of θ i , i � 1, 2, 3 could be eliminated to reduce the number of dependent variables from the expression of PI (equation (12)). However, it is still difficult to graphically examine the behavior of all GAs' selection techniques due to overlapping of the surface plot of PI. So, we adopt the modified mechanism is the subsequent section by assigning same weights to any two terms in PI (equation (12)). Hence, the PI becomes a function of single variable. e resultant cases are given below: e graphical representation for cases (1-3) in Figures 5-7 reveal that the horizontal axis define weights (wt) and performance index (PI) scaled on the vertical axis. e PI of proposed SWS is superimposed in Figures 5 and 7 as compared to other selection schemes which show a substantial enhancement towards perfection. Moreover, SWS shows considerable improvement at lower weights in terms of PI in Figure 6. More specifically, the graphical representation of PI endorses the improved performance of SWS.

Conclusions
In current study, we focused on the relative performance among various selection techniques to obtain the optimal solution for given test problems. A set of selection techniques including roulette wheel selection (RWS), linear rank selection (LRS), tournament selection (TS), stochastic remainder selection (SRS), and stairwise selection (SWS) were considered, and their performance was evaluated through ten well-known benchmark functions with 10 to 100 dimensions. ese benchmark functions cover various characteristics including convex, separable, nonseparable, unimodal, and multimodal. Additionally, the results of Chisquare goodness of fit test show improvements regarding proposed selection technique, and there is also an   insignificant difference between expected and actual number of offsprings. e statistical results of this study also show that the proposed selection technique (SWS) performed best in nine out of ten benchmark functions because of proportionate selection methodology. Furthermore, the simulated results reveal that the performance of SWS is significantly improved for unimodal and multimodal benchmark functions. When increasing the dimensions of experiments, SWS also performed efficiently under complex circumstances of dimensionality. e variability of results reveals that the proposed scheme has a better control over selection pressure and loss of population diversity. erefore, SWS found a suitable adjustment between exploitation and exploration due to split ranked ideology. According to the results, TS is the second best selection technique after SWS, and sometimes there is insignificance difference between these two. Finally, the numerical outcomes of proposed technique are very close to theoretical optimum value which is an evidence of the best-performing selection technique with authentication of performance index (PI).

Data Availability
e data used to support the findings of this manuscript are taken from the website (https://www.sfu.ca/ssurjano/ optimization.html).

Conflicts of Interest
e authors declare that they have no conflicts of interest.