Next Article in Journal
Enhanced Example Diffusion Model via Style Perturbation
Previous Article in Journal
Positive Periodic Solutions for a First-Order Nonlinear Neutral Differential Equation with Impulses on Time Scales
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Parallel Binary Rafflesia Optimization Algorithm and Its Application in Feature Selection Problem

1
College of Computer Science and Engineering, Shandong University of Science and Technology, Qingdao 266590, China
2
Department of Information Management, Chaoyang University of Technology, Taichung 41349, Taiwan
3
Departments of Computer Information System and Computer Sciences, Faculty of Computer Science and Informatics, Amman Arab University, Amman 11953, Jordan
*
Author to whom correspondence should be addressed.
Symmetry 2023, 15(5), 1073; https://doi.org/10.3390/sym15051073
Submission received: 25 April 2023 / Revised: 9 May 2023 / Accepted: 10 May 2023 / Published: 12 May 2023

Abstract

:
The Rafflesia Optimization Algorithm (ROA) is a new swarm intelligence optimization algorithm inspired by Rafflesia’s biological laws. It has the advantages of high efficiency and fast convergence speed, and it effectively avoids falling into local optimum. It has been used in logistics distribution center location problems, and its superiority has been demonstrated. It is applied to solve the problem of continuity, but there are many binary problems to be solved in the actual situation. Thus, we designed a binary version of ROA. We used transfer functions to change continuous values into binary values, and binary values are used to symmetrically represent the meaning of physical problems. In this paper, four transfer functions are implemented to binarize ROA so as to improve the original transfer function for the overall performance of the algorithm. In addition, on the basis of the algorithm, we further improve the algorithm by adopting a parallel strategy, which improves the convergence speed and global exploration ability of the algorithm. The algorithm is verified on 23 benchmark functions, and the parallel binary ROA has a better performance than some other existing algorithms. In the aspect of the application, this paper adopts the datasets on UCI for feature selection. The improved algorithm has higher accuracy and selects fewer features.

1. Introduction

With the rise of the Internet industry, the scale of data available to us is getting larger and larger, so this requires more and more of our ability to perform information filtering. Data mining has received increasing attention in recent years [1,2]. The acquisition of data requires not only comprehensive but also efficient access [3]. Before formally processing data, we need to perform data pre-processing. Feature selection is a good way of data pre-processing [4]. In a large dataset, there are many features that may be unnecessary or redundant and not relevant to our ultimate goal of classification, so it is necessary to reduce the dimensionality of the data and focus on exploring those features that are more representative to achieve a more efficient classification and processing problem [5,6]. However, in our inexperience, it is difficult to estimate which features are useful and which ones are redundant.
Davies has shown that the feature selection problem is an NP problem, which means that we can only find the optimal subset of features by exhaustive enumeration. However, the time cost of using the exhaustive method is excessively high. This cost is unacceptable in most cases. Therefore it is expected that a more optimal method can be used to solve this problem. A heuristic algorithm is a method that expects an optimal or suboptimal solution at an acceptable cost and has become a mainstream approach to solve complex problems [7,8,9,10,11,12]. Particle swarm optimization is a very classical heuristic algorithm developed based on the behavior of biological swarms in nature [13,14,15,16]. Whale optimization algorithm is inspired by the social behavior of humpback whales, which is more competitive than traditional algorithms and has been applied to various fields [17,18,19,20]. In recent years, heuristic algorithms have received more and more attention from scholars, and many excellent algorithms have been proposed [21,22,23,24,25,26,27]. According to the shortcomings of algorithms, many heuristic algorithms for further optimization have also been proposed by scholars [28,29,30,31]. ROA is an algorithm based on the living habits of Rafflesia, and has the advantages of fast convergence and not easily falling into local optimum [32].
Traditional optimization algorithms are generally applied to solve continuous problems. Many optimization algorithms are applied to various problems [33,34,35,36,37]. However, in reality, there are many discrete problems that need to be solved, such as feature selection. To date, many scholars have proposed many binary algorithms, and all of them have achieved good results [38,39,40,41]. The binary PSO algorithm has been applied to the problem of feature selection with good results, and this is the first time that a binary algorithm has been proposed [42]. The improved binary Grey wolf optimization algorithm is proposed by Hu [43]. In the binary algorithm, we will use 1 or 0 to symmetrically indicate whether the corresponding feature in the dataset is to be used for classification. In the paper, a new parallel strategy is proposed and new transfer functions are introduced to achieve better results in the application of the feature selection problem.
Parallel strategies have been widely used in recent years to improve heuristic algorithms and have obtained good results [44,45]. The parallel strategy is to group the original populations, explore them independently of each other, and communicate with them every certain number of iterations through a certain communication strategy to achieve faster convergence and prevent premature maturity [46,47,48]. Schutte has demonstrated the effectiveness of parallel strategies through numerous experiments [49].
The binarization of algorithms is often implemented with transfer functions [50,51]. The S-type transfer function is applied in the binary PSO algorithm to complete the binarization of the continuity values [42]. Zhuang applied a four-family transfer function for the binarization of the quasi-affine transformation evolution algorithm and demonstrated the superiority of the algorithm by 23 benchmark test functions [52]. This paper focuses on the binarization of the ROA algorithm using transfer functions and the optimization of the algorithm by a parallel strategy. The improved algorithm was tested on the UCI dataset for feature selection and compared with the binary GWO and binary PSO algorithms, and the experimental results showed that the algorithm outperformed the above two algorithms in some cases.
The general organization of this paper is as follows. Section 2 describes related works. Section 3 proposes new transfer functions based on mathematical analysis and proposes a parallel strategy to optimize the algorithm. Section 4 demonstrates the performance of the algorithm by testing it on benchmark functions. Section 5 presents the application of the algorithm to feature selection. Section 6 summarizes and discusses this paper.

2. Related Works

2.1. ROA

Rafflesia is a saprophytic organism. Its main axis is extremely short, and there are no leaves and underground stems. When it blooms, it emits a peculiar smell, thus attracting certain insects to pollinate it. In this process, because of its unique structure, some attracted insects may be trapped during pollination and die. After the flowering period has passed, the petals will wither, and the fruit will be ripe at this time. There will be many tiny seeds in the fruit. After the fruit falls to the ground, the seeds will be randomly taken to various places in various ways to find a suitable germination place. Based on the above characteristics, the ROA algorithm is divided into three stages to implement.
The stage of attracting insects is divided into two strategies. The first strategy seeks to replace the inferior individual. In this strategy, 1/3 of the individual with poor fitness is replaced by a new one. Each dimension of the newly added individual is abstracted into a three-dimensional space for calculation, and the dimension model is illustrated in Figure 1. The equations for calculating the individual X i position are (1)–(3) to update its position.
X i k = X b e s t k + d · s i n β k c o s γ k
d = k = 1 D ( X R k X b e s t k ) 2
X w o r s t i = X i
where d is the distance between X i and X b e s t . X R is a random individual in the population. X b e s t is the best individual in the population. X w o r s t represents the poor individual in the population. β k is a random value between [0, π /2]; γ k is random value between [0, π ].
The second strategy is to update 2/3 of individuals with better fitness. The individual velocity update equation is derived from the insect flapping flight model. The individual speed update equations are as shown in Equations (4)–(7). The individual update Equation is (8):
v 1 = ω 0 2 A 2 s i n 2 ( ω 0 t + θ ) + B 2 c o s 2 ( ω 1 t + θ )
v 2 = v 2 ω 0 c o s ( ω 0 t + θ + ϕ )
v = v 1 + v 2
l e = C · v · t + ( X b e s t X ( t ) ) · ( 1 C ) · r a n d
X ( t ) = X ( t ) + l e
where v 1 and v 2 represent the translation speed and rotation speed, respectively, and ω 0 and ω 1 represent the frequency periods of flapping and lateral flapping wings, respectively, both with values of 0.025. A is the amplitude of the wing during movement, with the value of 2.5. B is the lateral offset, with a value of 0.1. ϕ representing the phase difference between translation and rotation, with the value of −0.78545. The value range of θ is (0, 2pi). The initial value of v 2 is within the range (0, 2pi). C is a random number in (−1, 1).
According to the principle of “natural selection by nature, survival of the fittest”, the stage of swallowing insects will eliminate the individual with the worst fitness every certain number of iterations, thus ensuring the quality of the solution and speeding up the efficiency of the algorithm.
At the stage of spreading seeds, the position of the Rafflesia is represented by the initial optimal individual at this stage. At the same time, other individuals will randomly search around for an environment suitable for growth. At this stage, individuals update the equation as follows:
X ( t ) k = X b e s t k + r d · e x p ( i t e m M a x _ i t e r 1 ) · s i g n ( r a n d 0.5 )
where k ( k = 1 , , D ) is the population dimension. i t e r and M a x _ i t e r represent the current number of iterations and the maximum number of iterations, respectively. Where s i g n ( r a n d 0.5 ) is a symbolic function with a value of 1 or −1. r d represents the range of individual values in the population, and its expression equation is as follows:
r d = r a n d · ( u b l b ) + l b
u b is the upper limit of the search space and l b is the lower limit of the search space. Algorithm 1 is the pseudo code for ROA.
Algorithm 1:ROA
Input: f(x): fitness function; N: individual number; d: function dimension; lb: maximum boundary; ub: minimum boundary; M a x _ i t e r : maximum number of iteration
Output: global optimal value.
Initialize the related parameters of ROA
Randomly generate the positions of the insects
while iter< M a x _ i t e r doSymmetry 15 01073 i001

2.2. Transfer Function

The transfer function is a common method used in metaheuristic algorithms to convert from continuous variables into binary variables. We can use transfer functions to map a continuous variable to [0, 1] and then convert them into binary values by subsequent processing. It is reasonable and convenient to use transfer functions for binary operations.
In [43], Hu uses transfer functions to binarize the algorithm. As shown in Equation (11), where A denotes a vector of coefficients and D denotes a vector of distances between the current individual and the superior individual.
F ( A D ) = 1 1 + e 10 ( A D 0.5 )
After mapping through the transfer function in the above equation, we can use Equation (12) to iterate over the positions of individuals. r a n d denotes a random number that is uniformly distributed between 0 and 1.
X ( t + 1 ) i = 0 , F ( A D ) r a n d 1 , F ( A D ) > r a n d
In this paper, we use four transfer function families to map continuous values. Herein, there are S-shaped, U-shaped, and V-shaped families of transfer functions are proposed by Mirjalili [53,54]. The expressions are given in the following Table 1, Table 2 and Table 3. Guo proposed a family of Z-shaped transfer functions in [55] and used them in binarization. The expressions are shown in Table 4.
The position update equation of the S-shaped transfer functions is (12), which means that the particle is more likely to become 0 when its velocity is small and 1 when its velocity is large. However, because the nature of the U-shaped, V-shaped, and Z-shaped transfer functions is different from that of S-shaped transfer functions, it is not possible to use the same position update method as the S-shaped transfer functions. Zhuang proposed the location update Equation (13). In this updated formula, when the individual’s speed is higher, there is a higher possibility of changing to a complementary position.
X ( t + 1 ) = 1 X ( t ) , F ( A D ) > r a n d X ( t ) , F ( A D ) r a n d

3. Analysis and Proposed Parallel Binary ROA

In this section, a mathematical analysis of the search space of ROA is carried out. The initial transfer function is changed on the basis of mathematical analysis, and the algorithm is binarized. The algorithm is optimized using a parallel strategy, and two inter-group communication methods are proposed to improve the convergence speed of the algorithm and reduce the risk of premature convergence.

3.1. Mathematical Analysis

In the ROA algorithm, the location of the insect can be at any point in the search space, but in the binary ROA, its location can only be chosen between 0 and 1. Therefore, the search space of the original algorithm needs to be analyzed to obtain its specific range so that the corresponding transfer function can be better constructed to binarize the algorithm.
In order to make it easier to understand, we take one dimension of the individual to analyze. From (7) and (8), we know that X i = X i + c v t + ( X b e s t X i ) · ( 1 c ) · r a n d , where v = v 1 + v 2 , v 1 = ω 0 2 A 2 s i n 2 ( ω 0 t + θ ) + B 2 c o s 2 ( ω 1 t + θ ) , v 2 = v 2 ω 0 c o s ( ω 0 t + θ + ϕ ) . Furthermore, it is known that the values of X i and X b e s t can only be chosen between 0 or 1, so the final value of X has the four following cases.
(1)
if X i = 0 and X b e s t = 0
X i = 0 + c v t + ( 0 0 ) · ( 1 c ) · r a n d = c v
(2)
if X i = 0 and X b e s t = 1
X i = 0 + c v t + ( 1 0 ) · ( 1 c ) · r a n d = c v + ( 1 c ) · r a n d
(3)
if X i = 1 and X b e s t = 0
X i = 1 + c v t + ( 0 1 ) · ( 1 c ) · r a n d = 1 + c v ( 1 c ) · r a n d
(4)
if X i = 1 and X b e s t = 1
X i = 1 + c v t + ( 1 1 ) · ( 1 c ) · r a n d = 1 + c v
We can know that c is a random number in (−1, 1) and r a n d is a random number between (0, 1). After the above analysis, we derive v s . ( 0.156 , 0.188 ) , and conclude that the range of values of X i ( 1.188 , 2.156 ) . The conclusions we analyze here can be used for the improvement of the transfer function later.

3.2. New Transfer Functions

The transfer function is extremely important for the binary of the algorithm. In Section 2.2, we refer to four families of transfer functions to binarize the algorithm. This section will improve the transfer functions based on the mathematical analysis of the algorithm in Section 3.1.
The original transfer function is not fully applicable in this algorithm’s binary process because the search space to which the original transfer function is adapted does not match this algorithm. An acceptable transfer function allows individuals to converge to the optimum more quickly, while an unsuitable transfer function may cause the algorithm to converge slowly. We take S-shaped transfer functions as an example, and we hope that the greater the value of the individual in the process of position update, the greater the probability that the individual position will take the value of 1, and vice versa for the value of 0. From Section 3.1, we know that the search space of the ROA is [−1.188, 2.156], the center of the search space is 0.484, and the range of values of the S 1 transfer function in the search space is [0.0850, 0.9868]. Therefore, we deform the transfer functions so that we can achieve better results, and the deformed S-shaped transfer functions are shown in Table 5.
Similar deformation operations are performed for other transfer function families, which enable the values that can be obtained in the original search space to be more evenly distributed between [0, 1], resulting in better results. The deformed transfer functions are shown in Table 6, Table 7 and Table 8. A comparison of the original and improved transfer function series is shown in Figure 2.

3.3. Parallel Strategy

Parallel strategy is often used in the optimization of algorithms. Through parallels, we can obtain better quality solutions and faster convergence. The concrete practice of a parallel strategy is to explore the optimal solution through grouping, and each group is independent of the other. Then, when certain conditions are met, communication between groups is carried out, thus completing the information exchange between groups.
In the parallel strategy used in this algorithm, we group the original population into a main population and two subpopulations. Subpopulation 1 selects the better individuals of the main population to search because optimal individual may appear in the vicinity of the better ones. Subpopulation 2 randomly selects individuals in the main population to search so as to prevent the population from falling into a local optimum.
The communication strategy is extremely important for the algorithm. A good communication method can make the algorithm obtain faster convergence and make the algorithm less likely to fall into the local optimum. In order to achieve faster convergence and improve the quality of the solution, two intergroup communication strategies are proposed. Strategy 1 compares the best individuals in the subpopulation with the best in the main population after every certain number of iterations, and if the individuals in the subpopulation are better, the worst individuals in the main population are replaced. This strategy can make the algorithm converge faster. Strategy 2 is to monitor the optimal individuals in the main population, and if no update is made for a long time, the algorithm is considered to be possibly stuck in a local optimum, and then, some individuals that are optimal in the main population are updated. The update is performed by combining the different dimensions of previously good individuals to form new individuals, and then replacing some of the best individuals in the main population with new ones. Figure 3 depicts the communication strategy for a parallel strategy.

4. Experiments, Results, and Analysis

In this section, we will use 23 benchmark functions to examine the exploration and development capabilities of the algorithm. The benchmark test function used, although specifically for evaluating continuous optimization algorithms, still has the original properties of the function when each dimension can only take binary values, so the benchmark test function can be used to detect binary optimization. In Table 9, f 1 f 7 represents unimodal benchmark functions; in Table 10, f 8 f 13 represents common multimodal benchmark functions; and in Table 11, f 14 f 23 represents multimodal benchmark functions in low dimension. D i m in the table is the dimension of the function, f m i n denotes the minimum value that the function can obtain, and s p a c e denotes the search space of the function.
There is no optimal local solution in the unimodal function, but only one global optimal solution and the convergence speed of the algorithm can be measured by the single-peaked function. The common multimodal function is relatively complex, with multiple locally optimal solutions, so it can test whether the algorithm can dispose of the local optimal and thus reach the optimal global solution. Besides having multiple locally optimal solutions, the multimodal functions in low dimension with its too low dimensionality makes the algorithm more prone to prematureness, so it can strictly verify the convergence results of the algorithm.

Experimental Results

In experiments, we examine the impact of the improved transfer functions and the addition of the parallel strategy on the algorithm. Due to space limitation, we chose four transfer functions, namely S 3 , V 2 , U 3 , Z 3 , in four families of transfer functions as examples for our experiments. We also compared the improved algorithm with BPSO and BGWO to perform the experiments [42,43].
In our experiments, we will judge the algorithm’s merit by the quality of the solution and its stability. In the experiments, the algorithms were run 20 times on each benchmark function, and 150 iterations were passed during each run, with a total of 30 individuals in the population. Table 12 shows the experimental results when the algorithm uses the original transfer function, and Table 13 shows the experimental results when the algorithm uses the improved strategy. For the convenience of reading, the experimental data are retained to four decimal places using rounding.
As can be seen from Table 12, BROA_U3, BROA_V2, and BROA_Z3 have better results on the unimodal benchmark functions. This proves the good convergence performance of the ROA algorithm using the original V-shaped, U-shaped, and Z-shaped transfer function binarization. BGWO and BROA_Z3 achieved the best results on f8 and f13. BROA_U3, BROA_V2, and BROA_Z3 achieved the best performance on f9–f12, which indicates their good global exploration capability. In contrast, the ROA algorithm was optimized with the S-shaped transfer function and performs poorly. In the multimodal benchmark functions in low dimensions, all algorithms achieve better results.
Table 13 shows the experimental results of the algorithm after applying the improved strategy. For ease of reading, the improved algorithm is indicated in blue font if it is better than the original algorithm, and in red font if the improved algorithm is inferior to the original algorithm. The improved BROA_S3 achieves better results on both f1–f3 and f5–f13, indicating that the improved strategy is very effective on it, allowing it to obtain better convergence performance and more easily escape the local optimum. The improved BROA_V2 has better results on f5, f7, f8, f13; the improved BROA_U3 has better results on f7, f8, f13; and the improved BROA_V2 has better results on f5, f7. The differences between the fitness values of the improved algorithm and the original algorithm on the test functions are shown in Figure 4. On the figure, we can more clearly see the performance improvement due to the algorithm improvement. The above results show that the performance of the algorithm is significantly improved after applying the improved strategy. The parallel strategy of the algorithm allows it to converge to the optimum faster and effectively prevents the algorithm from falling into a local optimum. Among the algorithms that use different transfer functions for binarization, the improved strategy is the most effective in optimizing the algorithm for the binarization of the S-shaped transfer function.
In order to demonstrate that the improved algorithm is significantly different from the original algorithm, the Wilcoxon rank-sum test was performed on the results of the algorithm before and after improvement at the 5% significance level. We assume that there is no significant difference between the improved algorithm and the original algorithm. The obtained results are shown in Table 14. The original hypothesis is rejected if the value p in the table is less than 0.05, proving that the improved algorithm is significantly different from the one before improvement. From the table, we can see that PBROA_S3 has p-values less than 0.05 on f1, f2, f4, f5, f7; PBROA_V2 has p-values less than 0.05 on f7; PBROA_U3 has p-values less than 0.05 on f4, f7, f8; and PBROA_Z3 has p-values lower than 0.05 on f5, f7. From the results analyzed above, we know that there is a significant difference between the improved algorithm and the original algorithm. This further proves that the improved strategy enhances the performance of the algorithm.

5. Application of Feature Selection

Because the original data have a lot of redundancy, it makes the data processing difficult and time consuming. Feature selection is a very important data pre-processing approach, where the raw data are initially processed by feature selection, and the processed data become more accurate and streamlined. In this section, we apply the improved algorithm to feature selection.

5.1. Dataset

The experiments test the performance of the algorithm using 13 datasets. These datasets are all from the UCI machine learning repository and vary in number and dimensionality [56]. The specific parameters of the datasets are shown in Table 15.

5.2. Experimental Results and Analysis

KNN is a simple method for data classification in data mining, which is based on the principle that each individual type can be represented by its K-nearest neighbors [57]. The calculation of distance in KNN generally uses the Manhattan distance or the Euclidean distance. The specific calculation is as in Equation (14)
D p ( x , x t e s t ) = ( i = 1 n | x i x t e s t i | p ) 1 p
where p can be taken as 1 or 2; when p = 1 , this equation indicates the calculation of Manhattan distance, and when p = 2 , this equation indicates the calculation of Euclidean distance. Where x denotes training data, x t e s t denotes the test data, and n denotes the dimensionality of the data.
In this experiment, the K-fold cross-validation is adopted, the original dataset is divided into K parts, and K-1 parts are used as training data each time, leaving one datum as the test datum. After repeating K experiments, the final result is taken as the average of K experiments. In the experiment, we should balance the classification accuracy and the number of features selected. The purpose of feature selection is to streamline the experimental data, so we want to obtain a lower classification error and, at the same time, select as few features as possible. The experimental error is calculated as shown in Equation (15)
f i t n e s s = a · e r r o r + ( 1 a ) f l a g d i m
where e r r o r denotes the classification error, a denotes the proportion of the classification error in the fitness, and in this experiment, a takes the value of 0.99. f l a g denotes the number of selected features, and d i m denotes the total number of features in this dataset.
Euclidean and 5 are the values of the parameters of Distance and NumNeighbors in KNN, and the value of K-fold parameter in cross validation is set to 2 in this experiment. In the experiments, the number of all populations was set to 30, the number of iterations was set to 150, and it was run 20 times on each dataset. In the algorithm used in the experiments, when the individual takes the value of 0, it means that the current top feature is not selected, and when the individual takes the value of 1, this means that the current top feature is selected.
The data in Table 16 are the fitness values, where the best data are indicated in bold for ease of reading. In addition, we also use Figure 5 to show the experimental results, which makes the comparison of the experimental results clearer. From the data in the table and figure, we can see that BROA_Z3 achieves the best results for the datasets Breast Cancer, Breast Cancer Wisconsin, and Glass. For the dataset Iris, several algorithms achieve better results due to its low dimensionality. For BGWO, it performs well in the datasets South German Credit, Flags, and Image Segmentation. BROA_S3 has good performance in the datasets Dermatology, Credit Approval, Chess(kr-vs-kp), Hepatitis, Statlog (Australian Credit Approval), and Wall-Following Robot Navigation datasets. These results show the superiority of the parallel binary ROA algorithm, which can outperform BGWO and BPSO algorithms in most cases, and BROA_S3 has higher accuracy among several improved algorithms. Compared with the traditional algorithm, the new algorithm proposed in this paper has a better convergence performance, it does not easily fall into local optimum, and tends to find better solutions. The above results also confirm the superiority of the improved parallel optimization algorithm in practical applications, where the parallel strategy can explore more cases and obtain better results.
The data in Table 17 are the number of selected features, where the best data are indicated in bold for the ease of reading. To make the experimental effect more understandable, we also show the experimental results with Figure 6. From the data in the table and figure, we can see that, for the dataset Iris, several algorithms give the best results. BPSO gives the best results for the dataset Breast Cancer, Hepatitis. BROA_S3 gives the best results for the dataset Dermatology. BROA_V2 selected fewer features on the datasets South German Credit, Flags, Credit Approval, Image Segmentation, and Statlog (Australian Credit Approval). BROA_U3 has better results on the datasets Wine and Wall-Following Robot Navigation. BROA_Z3 selected fewer features on the dataset Breast Cancer Wisconsin, Glass, Chess (kr-vs-kp). From the analysis of the above results, it is known that BROA_S3 is more accurate but biased in the selection of more features, and BROA_V2 will be more biased to select fewer features to achieve classification.

6. Conclusions

The parallel binary Rafflesia Optimization Algorithm is able to solve discrete problems, and in this paper, it is applied to the problem of feature selection with good results. In this paper, the transfer function is crucial in the binary algorithm. We analyze the search space of the original algorithm and propose new transfer functions based on it, thus binarizing the original algorithm. To further improve the performance of the algorithm, a parallel strategy is used to optimize the algorithm. The remarkable performance of the parallel binary Rafflesia Optimization Algorithm is demonstrated by running on the benchmark functions. Finally, we successfully applied the algorithm to the problem of feature selection and achieved the classification of features by KNN and cross-validation methods. Parallel binary ROA has been shown to possess acceptable classification accuracy in experiments. In this paper, we only used KNN to implement feature selection, which can be combined with a neural network for classification in the future and other ways to reduce the classification error. The binary conversion of the algorithm also increases the computation time, and future work will also include reducing the running time of the algorithm.

Author Contributions

Conceptualization, J.-S.P. and S.-C.C.; Data curation, P.H.; Formal analysis, J.-S.P. and S.-C.C.; Investigation, J.-S.P. and H.-J.S.; Methodology, J.-S.P., H.-J.S., S.-C.C. and P.H.; Resources, H.-J.S. and P.H.; Software, H.-J.S. and S.-C.C.; Validation, J.-S.P., S.-C.C. and H.A.S.; Writing—original draft, H.-J.S.; Writing—review and editing, J.-S.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lin, J.C.W.; Wu, T.Y.; Fournier-Viger, P.; Lin, G.; Zhan, J.; Voznak, M. Fast algorithms for hiding sensitive high-utility itemsets in privacy-preserving utility mining. Eng. Appl. Artif. Intell. 2016, 55, 269–284. [Google Scholar] [CrossRef]
  2. Romero, C.; Ventura, S. Data mining in education. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2013, 3, 12–27. [Google Scholar] [CrossRef]
  3. Chen, C.M.; Chen, L.; Gan, W.; Qiu, L.; Ding, W. Discovering high utility-occupancy patterns from uncertain data. Inf. Sci. 2021, 546, 1208–1229. [Google Scholar] [CrossRef]
  4. Karabulut, E.M.; Özel, S.A.; Ibrikci, T. A comparative study on the effect of feature selection on classification accuracy. Procedia Technol. 2012, 1, 323–327. [Google Scholar] [CrossRef]
  5. Kumar, V.; Minz, S. Feature selection: A literature review. SmartCR 2014, 4, 211–229. [Google Scholar] [CrossRef]
  6. Chandrashekar, G.; Sahin, F. A survey on feature selection methods. Comput. Electr. Eng. 2014, 40, 16–28. [Google Scholar] [CrossRef]
  7. Meng, Z.; Pan, J.S.; Tseng, K.K. PaDE: An enhanced Differential Evolution algorithm with novel control parameter adaptation schemes for numerical optimization. Knowl.-Based Syst. 2019, 168, 80–99. [Google Scholar] [CrossRef]
  8. Xue, X.; Chen, J.; Yao, X. Efficient user involvement in semiautomatic ontology matching. IEEE Trans. Emerg. Top. Comput. Intell. 2018, 5, 214–224. [Google Scholar] [CrossRef]
  9. Sun, C.; Jin, Y.; Cheng, R.; Ding, J.; Zeng, J. Surrogate-assisted cooperative swarm optimization of high-dimensional expensive problems. IEEE Trans. Evol. Comput. 2017, 21, 644–660. [Google Scholar] [CrossRef]
  10. Wang, H.; Rahnamayan, S.; Sun, H.; Omran, M.G. Gaussian bare-bones differential evolution. IEEE Trans. Cybern. 2013, 43, 634–647. [Google Scholar] [CrossRef]
  11. Salehi, S.; Selamat, A.; Mashinchi, M.R.; Fujita, H. The synergistic combination of particle swarm optimization and fuzzy sets to design granular classifier. Knowl.-Based Syst. 2015, 76, 200–218. [Google Scholar] [CrossRef]
  12. Wang, J.; Liu, Y.; Chen, J.; Yang, X. An Ensemble Framework to Forest Optimization Based Reduct Searching. Symmetry 2022, 14, 1277. [Google Scholar] [CrossRef]
  13. Jain, M.; Saihjpal, V.; Singh, N.; Singh, S.B. An Overview of Variants and Advancements of PSO Algorithm. Appl. Sci. 2022, 12, 8392. [Google Scholar] [CrossRef]
  14. Meng, Z.; Zhong, Y.; Mao, G.; Liang, Y. PSO-sono: A novel PSO variant for single-objective numerical optimization. Inf. Sci. 2022, 586, 176–191. [Google Scholar] [CrossRef]
  15. Pervaiz, S.; Bangyal, W.H.; Ashraf, A.; Nisar, K.; Haque, M.R.; Ibrahim, A.; Ag, A.B.; Chowdhry, B.; Rasheed, W.; Rodrigues, J.J.; et al. Comparative Research Directions of Population Initialization Techniques using PSO Algorithm. Intell. Autom. Soft Comput. 2022, 32, 1427–1444. [Google Scholar] [CrossRef]
  16. Feng, H.M. Self-generation RBFNs using evolutional PSO learning. Neurocomputing 2006, 70, 241–251. [Google Scholar] [CrossRef]
  17. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  18. Guo, F.; Zhang, H.; Xu, Y.; Xiong, G.; Zeng, C. Isokinetic Rehabilitation Trajectory Planning of an Upper Extremity Exoskeleton Rehabilitation Robot Based on a Multistrategy Improved Whale Optimization Algorithm. Symmetry 2023, 15, 232. [Google Scholar] [CrossRef]
  19. Li, Y.; Li, X.; Liu, J.; Tu, X. Gaussian perturbation whale optimization algorithm based on nonlinear strategy. Int. J. Perform. Eng. 2019, 15, 1829. [Google Scholar] [CrossRef]
  20. Aljarah, I.; Faris, H.; Mirjalili, S. Optimizing connection weights in neural networks using the whale optimization algorithm. Soft Comput. 2018, 22, 1–15. [Google Scholar] [CrossRef]
  21. Wu, T.Y.; Lin, J.C.W.; Zhang, Y.; Chen, C.H. A grid-based swarm intelligence algorithm for privacy-preserving data mining. Appl. Sci. 2019, 9, 774. [Google Scholar] [CrossRef]
  22. Chu, S.C.; Tsai, P.W.; Pan, J.S. Cat swarm optimization. In Proceedings of the Pacific Rim International Conference on Artificial Intelligence, Guilin, China, 7–11 August 2006; pp. 854–858. [Google Scholar]
  23. Meng, Z.; Yang, C. Two-stage differential evolution with novel parameter control. Inf. Sci. 2022, 596, 321–342. [Google Scholar] [CrossRef]
  24. Song, P.C.; Chu, S.C.; Pan, J.S.; Yang, H. Simplified Phasmatodea population evolution algorithm for optimization. Complex Intell. Syst. 2022, 8, 2749–2767. [Google Scholar] [CrossRef]
  25. Pan, J.S.; Zhang, L.G.; Wang, R.B.; Snášel, V.; Chu, S.C. Gannet optimization algorithm: A new metaheuristic algorithm for solving engineering optimization problems. Math. Comput. Simul. 2022, 202, 343–373. [Google Scholar] [CrossRef]
  26. Nguyen, T.-T.; Dong-Nguyen, T.; Ngo, T.-G.; Nguyen, V.-T. An Optimal Thresholds for Segmenting Medical Images Using Improved Swarm Algorithm. J. Inf. Hiding Multimed. Signal Process. 2022, 13, 12–21. [Google Scholar]
  27. Xue, X.; Jiang, C. Matching sensor ontologies with multi-context similarity measure and parallel compact differential evolution algorithm. IEEE Sens. J. 2021, 21, 24570–24578. [Google Scholar] [CrossRef]
  28. Wang, G.; Guo, L.; Wang, H.; Duan, H.; Liu, L.; Li, J. Incorporating mutation scheme into krill herd algorithm for global numerical optimization. Neural Comput. Appl. 2014, 24, 853–871. [Google Scholar] [CrossRef]
  29. Wang, G.G.; Hao, G.S.; Cheng, S.; Cui, Z. An improved monarch butterfly optimization with equal partition and f/t mutation. In Proceedings of the Advances in Swarm Intelligence: 8th International Conference, ICSI 2017, Fukuoka, Japan, 27 July–1 August 2017; pp. 106–115. [Google Scholar]
  30. Deng, W.; Xu, J.; Zhao, H. An improved ant colony optimization algorithm based on hybrid strategies for scheduling problem. IEEE Access 2019, 7, 20281–20292. [Google Scholar] [CrossRef]
  31. Arasteh, B.; Seyyedabbasi, A.; Rasheed, J.; Abu-Mahfouz, A.M. Program Source-Code Re-Modularization Using a Discretized and Modified Sand Cat Swarm Optimization Algorithm. Symmetry 2023, 15, 401. [Google Scholar] [CrossRef]
  32. Pan, J.S.; Fu, Z.; Hu, C.C.; Tsai, P.W.; Chu, S.C. Rafflesia Optimization Algorithm Applied in the Logistics Distribution Centers Location Problem. J. Internet Technol. 2022, 23, 1541–1555. [Google Scholar]
  33. Trong-The Nguyen, T.D.N.; Nguyen, V.T. An Optimizing Pulse Coupled Neural Network based on Golden Eagle Optimizer for Automatic Image Segmentation. J. Inf. Hiding Multimed. Signal Process. 2022, 13, 155–164. [Google Scholar]
  34. Xue, X. A compact firefly algorithm for matching biomedical ontologies. Knowl. Inf. Syst. 2020, 62, 2855–2871. [Google Scholar] [CrossRef]
  35. Chen, X.; Cheng, L.; Liu, C.; Liu, Q.; Liu, J.; Mao, Y.; Murphy, J. A WOA-based optimization approach for task scheduling in cloud computing systems. IEEE Syst. J. 2020, 14, 3117–3128. [Google Scholar] [CrossRef]
  36. Ramachandran, M.; Ganesh, E. Energy Optimized Joint Channel Assignment and Routing using Cat Swarm Optimization (CSO) Algorithm in CRAHN. J. Green Eng. 2020, 202, 3434–3449. [Google Scholar]
  37. Gao, D.; Wang, G.G.; Pedrycz, W. Solving fuzzy job-shop scheduling problem using DE algorithm improved by a selection mechanism. IEEE Trans. Fuzzy Syst. 2020, 28, 3265–3275. [Google Scholar] [CrossRef]
  38. Mirjalili, S.; Mirjalili, S.M.; Yang, X.S. Binary bat algorithm. Neural Comput. Appl. 2014, 25, 663–681. [Google Scholar] [CrossRef]
  39. Mirjalili, S.; Hashim, S.Z.M. BMOA: Binary magnetic optimization algorithm. Int. J. Mach. Learn. Comput. 2012, 2, 204. [Google Scholar] [CrossRef]
  40. Hussien, A.G.; Oliva, D.; Houssein, E.H.; Juan, A.A.; Yu, X. Binary whale optimization algorithm for dimensionality reduction. Mathematics 2020, 8, 1821. [Google Scholar] [CrossRef]
  41. Akan, T.; Agahian, S.; Dehkharghani, R. Binbro: Binary battle royale optimizer algorithm. Expert Syst. Appl. 2022, 195, 116599. [Google Scholar]
  42. Kennedy, J.; Eberhart, R.C. A discrete binary version of the particle swarm algorithm. In Proceedings of the 1997 IEEE International Conference on Systems, Man, and Cybernetics, Computational Cybernetics and Simulation, Orlando, FL, USA, 12–15 October 1997; Volume 5, pp. 4104–4108. [Google Scholar]
  43. Hu, P.; Pan, J.S.; Chu, S.C. Improved binary grey wolf optimizer and its application for feature selection. Knowl.-Based Syst. 2020, 195, 105746. [Google Scholar] [CrossRef]
  44. Song, Y.; Wu, D.; Deng, W.; Gao, X.Z.; Li, T.; Zhang, B.; Li, Y. MPPCEDE: Multi-population parallel co-evolutionary differential evolution for parameter optimization. Energy Convers. Manag. 2021, 228, 113661. [Google Scholar] [CrossRef]
  45. Biscani, F.; Izzo, D. A parallel global multiobjective framework for optimization: Pagmo. J. Open Source Softw. 2020, 5, 2338. [Google Scholar] [CrossRef]
  46. Pan, J.S.; Sun, B.; Chu, S.C.; Zhu, M.; Shieh, C.S. A Parallel Compact Gannet Optimization Algorithm for Solving Engineering Optimization Problems. Mathematics 2023, 11, 439. [Google Scholar] [CrossRef]
  47. Ding, S.; Du, W.; Zhao, X.; Wang, L.; Jia, W. A new asynchronous reinforcement learning algorithm based on improved parallel PSO. Appl. Intell. 2019, 49, 4211–4222. [Google Scholar] [CrossRef]
  48. Khankhour, H.; Abdoun, O.; Abouchabaka, J. Parallel genetic approach for routing optimization in large ad hoc networks. Int. J. Electr. Comput. Eng. (IJECE) 2022, 12, 748–755. [Google Scholar] [CrossRef]
  49. Schutte, J.F.; Reinbolt, J.A.; Fregly, B.J.; Haftka, R.T.; George, A.D. Parallel global optimization with the particle swarm algorithm. Int. J. Numer. Methods Eng. 2004, 61, 2296–2315. [Google Scholar] [CrossRef]
  50. Rashedi, E.; Nezamabadi-Pour, H.; Saryazdi, S. BGSA: Binary gravitational search algorithm. Nat. Comput. 2010, 9, 727–745. [Google Scholar] [CrossRef]
  51. Abdel-Basset, M.; Mohamed, R.; Mirjalili, S. A binary equilibrium optimization algorithm for 0–1 knapsack problems. Comput. Ind. Eng. 2021, 151, 106946. [Google Scholar] [CrossRef]
  52. Chu, S.C.; Zhuang, Z.; Li, J.; Pan, J.S. A novel binary QUasi-affine transformation evolutionary (QUATRE) algorithm. Appl. Sci. 2021, 11, 2251. [Google Scholar] [CrossRef]
  53. Mirjalili, S.; Lewis, A. S-shaped versus V-shaped transfer functions for binary particle swarm optimization. Swarm Evol. Comput. 2013, 9, 1–14. [Google Scholar] [CrossRef]
  54. Mirjalili, S.; Zhang, H.; Mirjalili, S.; Chalup, S.; Noman, N. A novel U-shaped transfer function for binary particle swarm optimisation. In Soft Computing for Problem Solving 2019; Springer: Cham, Switzerland, 2020; pp. 241–259. [Google Scholar]
  55. Guo, S.; Wang, J.; Guo, M. Z-shaped transfer functions for binary particle swarm optimization algorithm. Comput. Intell. Neurosci. 2020, 2020, 6502807. [Google Scholar] [CrossRef] [PubMed]
  56. Asuncion, A.; Newman, D. UCI machine learning repository, 2007.
  57. Zhang, S.; Li, X.; Zong, M.; Zhu, X.; Wang, R. Efficient kNN classification with different numbers of nearest neighbors. IEEE Trans. Neural Netw. Learn. Syst. 2017, 29, 1774–1785. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The model of the calculated dimensions.
Figure 1. The model of the calculated dimensions.
Symmetry 15 01073 g001
Figure 2. The original (solid lines) and improved (dotted lines) families of transfer functions.
Figure 2. The original (solid lines) and improved (dotted lines) families of transfer functions.
Symmetry 15 01073 g002
Figure 3. Communication strategy for parallel strategy.
Figure 3. Communication strategy for parallel strategy.
Symmetry 15 01073 g003
Figure 4. The difference between algorithm fitness before and after applying the improved strategy.
Figure 4. The difference between algorithm fitness before and after applying the improved strategy.
Symmetry 15 01073 g004
Figure 5. The result of fitness value.
Figure 5. The result of fitness value.
Symmetry 15 01073 g005
Figure 6. The number of selected features.
Figure 6. The number of selected features.
Symmetry 15 01073 g006
Table 1. S-shaped families of transfer functions.
Table 1. S-shaped families of transfer functions.
Function NameFormula
S 1 ( x ) 1 1 + e 2 x
S 2 ( x ) 1 1 + e x
S 3 ( x ) 1 1 + e x / 2
S 4 ( x ) 1 1 + e x / 3
Table 2. V-shaped families of transfer functions.
Table 2. V-shaped families of transfer functions.
Function NameFormula
V 1 ( x ) | e r f ( π 2 x ) |
V 2 ( x ) | t a n h ( x ) |
V 3 ( x ) | x 1 + x 2 |
V 4 ( x ) | 2 π a r c t a n ( 2 π x ) |
Table 3. U-shaped families of transfer functions.
Table 3. U-shaped families of transfer functions.
Function NameFormula
U 1 ( x ) m i n ( | x 1.5 | , 1 )
U 2 ( x ) m i n ( | x 2 | , 1 )
U 3 ( x ) m i n ( | x 3 | , 1 )
U 4 ( x ) m i n ( | x 4 | , 1 )
Table 4. Z-shaped families of transfer functions.
Table 4. Z-shaped families of transfer functions.
Function NameFormula
Z 1 ( x ) 1 2 x
Z 2 ( x ) 1 5 x
Z 3 ( x ) 1 8 x
Z 4 ( x ) 1 20 x
Table 5. Improved S-shaped families of transfer functions.
Table 5. Improved S-shaped families of transfer functions.
Function NameFormula
S 1 ( x ) 1 1 + e 2 ( x 0.484 )
S 2 ( x ) 1 1 + e 5 ( x 0.484 )
S 3 ( x ) 1 1 + e 10 ( x 0.484 )
S 3 ( x ) 1 1 + e 15 ( x 0.484 )
Table 6. Improved V-shaped families of transfer functions.
Table 6. Improved V-shaped families of transfer functions.
Function NameFormula
V 1 ( x ) | e r f ( π 2 ( 4 · ( x 0.484 ) ) ) |
V 2 ( x ) | t a n h ( 3 · ( x 0.484 ) ) |
V 3 ( x ) | 1 + 1 . 672 2 1.672 · ( x 0.484 ) 1 + ( x 0.484 ) 2 |
V 4 ( x ) | 1 2 π a r c t a n ( π 2 · 5.016 ) · 2 π a r c t a n ( 6 π ( x 0.484 ) ) |
Table 7. Improved U-shaped families of transfer functions.
Table 7. Improved U-shaped families of transfer functions.
Function NameFormula
U 1 ( x ) m i n ( | ( x 0.484 ) 1.5 0 . 516 1.5 | , 1 )
U 2 ( x ) m i n ( | ( x 0.484 ) 2 0 . 516 2 | , 1 )
U 3 ( x ) m i n ( | ( x 0.484 ) 3 0 . 516 3 | , 1 )
U 4 ( x ) m i n ( | ( x 0.484 ) 4 0 . 516 4 | , 1 )
Table 8. Improved Z-shaped families of transfer functions.
Table 8. Improved Z-shaped families of transfer functions.
Function NameFormula
Z 1 ( x ) 1 2 x 0.484 1 2 1.602
Z 2 ( x ) 1 5 x 0.484 1 5 1.602
Z 3 ( x ) 1 8 x 0.484 1 8 1.602
Z 4 ( x ) 1 20 x 0.484 1 20 1.602
Table 9. Unimodal benchmark functions.
Table 9. Unimodal benchmark functions.
NameFunctionSpace D im f min
Sphere f 1 ( x ) = i = 1 n x i 2 [−100, 100]300
Schwefel’s function 2.21 f 2 ( x ) = i = 1 n | x i | + i = 1 n | x i | [−10, 10]300
Schwefel’s function 1.2 f 3 ( x ) = i = 1 n ( j 1 i x j ) 2 [−100, 100]300
Schwefel’s function 2.22 f 4 ( x ) = m a x i { | x i | , 1 i n } [−100, 100]300
Rosenbrock f 5 ( x ) = i = 1 n 1 [ 100 ( x i + 1 x i 2 ) + ( x i 1 ) 2 ] [−30, 30]300
Step f 6 ( x ) = i = 1 n ( [ x i + 0.5 ] ) 2 [−100, 100]300
Dejong’s noisy f 7 ( x ) = i = 1 n i x i 4 + r a n d o m ( 0 , 1 ) [−1.28, 1.28]300
Table 10. Common multimodal benchmark functions.
Table 10. Common multimodal benchmark functions.
NameFunctionSpace D im f min
Schwefel f 8 ( x ) = i = 1 n x i s i n ( | x i | ) [−500, 500]30−12,569
Rastringin f 9 ( x ) = i = 1 n [ x i 2 10 c o s ( 2 π x i ) + 10 ] [−5.12, 5.12]300
Ackley f 10 ( x ) = 20 e x p ( 0.2 1 n i = 1 n x i 2 ) [−32, 32]300
e x p ( 1 n i = 1 n c o s ( 2 π x i ) ) + 20 + e
Griewank f 11 ( x ) = 1 4000 i = 1 n x i 2 i = 1 n c o s ( x i i ) + 1 [−600, 600]300
Generalized penalized 1 f 12 ( x ) = π n { 10 s i n ( π y 1 ) + i = 1 n 1 ( y i 1 ) 2 [ 1 + [−50, 50]300
10 s i n 2 ( π y i + 1 ) ] + ( y n 1 ) 2 } + i = 1 n u ( x i , 10 , 100 , 4 )
y i = 1 + x i + 1 4 u ( x i , a , k , m ) =
k ( x i a ) m x i > a 0 , a < x i < a k ( x i a ) m x i < a
Generalized penalized 2 f 13 ( x ) = 0.1 { s i n 2 ( 3 π x 1 ) + [−50, 50]300
i = 1 n ( x i 1 ) 2 [ 1 + s i n 2 ( 3 π x i + 1 ) ] +
( x n 1 ) 2 [ 1 + s i n 2 ( 2 π x n ) ] } +
i = 1 n u ( x i , 10 , 100 , 4 )
Table 11. Multimodal benchmark functions in low dimension.
Table 11. Multimodal benchmark functions in low dimension.
NameFunctionSpace D im f min
Fifth of Dejong f 14 ( x ) = ( 1 500 j = 1 25 1 j + i = 1 2 ( x i a i j ) 6 ) [−65, 65]21
Kowalik f 15 ( x ) = i = 1 1 1 [ a i x 1 ( b i 2 + b i x 2 ) b i 2 + b i x 3 + x 4 ] 2 [−5, 5]40.00030
Six-hump camel back f 16 ( x ) = 4 x 1 2 2.1 x 1 4 + 1 3 x 1 6 + x 1 x 2 4 x 2 2 + 4 x 2 4 [−5, 5]2−1.0316
Branins f 17 ( x ) = ( x 2 5.1 4 π 2 x 1 2 + 5 π x 1 6 ) 2 + 10 ( 1 1 8 π ) c o s x 1 + 10 [-5, 5]20.398
Goldstein–Price f 18 ( x ) = [ 1 + ( x 1 + x 2 + 1 ) 2 ( 19 14 x 1 + 3 x 1 2 14 x 2 + 6 x 1 x 2 + 3 x 2 2 ) ] [−2, 2]23
× [ 30 + ( 2 x 1 3 x 2 ) 2 × ( 18 32 x 1 + 12 x 1 2 + 48 x 2 36 x 1 x 2 + 27 x 2 2 ) ]
Hartman 1 f 19 ( x ) = i = 1 4 c i e x p ( j = 1 3 a i j ( x j p i j ) 2 ) [1, 3]3−3.86
Hartman 2 f 20 ( x ) = i = 1 4 c i e x p ( j = 1 6 a i j ( x j p i j ) 2 ) [0, 1]6−3.32
Shekel 1 f 21 ( x ) = i = 1 5 [ ( X a i ) ( X a i ) T + c i ] 1 [0, 10]4−10.1532
Shekel 2 f 22 ( x ) = i = 1 7 [ ( X a i ) ( X a i ) T + c i ] 1 [0, 10]4−10.4028
Shekel 3 f 23 ( x ) = i = 1 10 [ ( X a i ) ( X a i ) T + c i ] 1 [0, 10]4−10.5363
Table 12. The statistical results of the original transfer function.
Table 12. The statistical results of the original transfer function.
FunctionBPSOBGWOBROA_S3BROA_V2BROA_U3BROA_Z3
advstdadvstdadvstdadvstdadvstdadvstd
f10.93330.63973.10001.29590.06670.25370.00000.00000.00000.00000.00000.0000
f20.73330.58332.80001.21490.16670.37900.00000.00000.00000.00000.00000.0000
f38.03336.6978117.4667103.40040.46671.22430.00000.00000.00000.00000.00000.0000
f41.00000.00001.00000.00001.00000.00000.80000.40680.00000.00001.00000.0000
f5259.8333112.419216.966738.592031.033348.343230.033318.166829.00000.000034.766741.6270
f68.70001.126513.23332.14858.16671.09337.50000.00007.50000.00007.50000.0000
f74.72103.006535.400219.31170.72111.55940.00060.00060.00050.00040.00100.0011
f8−24.51490.3653−25.24410.0000−25.18800.2135−25.13190.2909−25.07580.3423−25.24410.0000
f91.00000.64332.56671.52410.23330.43020.00000.00000.00000.00000.00000.0000
f100.55480.35111.33550.26140.02390.13090.00000.00000.00000.00000.00000.0000
f110.02230.01300.13580.06010.00250.00660.00000.00000.00000.00000.00000.0000
f121.77460.07712.05880.17691.67900.03821.66900.00001.66900.00001.66900.0000
f130.10000.06430.00000.00000.00670.02540.02000.04070.00670.02540.00000.0000
f1412.67050.000012.67050.000012.67050.000012.67050.000012.67050.000012.67050.0000
f150.14840.00000.14840.00000.14840.00000.14840.00000.14840.00000.14840.0000
f160.00000.00000.00000.00000.00000.00000.00000.00000.00000.00000.00000.0000
f1727.70290.000027.70290.000027.70290.000027.70290.000027.70290.000027.70290.0000
f18600.00000.0000600.00000.0000600.00000.0000600.00000.0000600.00000.0000600.00000.0000
f19−0.33480.0000−0.33370.0063−0.33480.0000−0.33480.0000−0.33480.0000−0.33480.0000
f20−0.16570.0000−0.13340.0554−0.16570.0000−0.16570.0000−0.16570.0000−0.16570.0000
f21−5.05520.0000−5.05520.0000−5.05520.0000−5.05520.0000−5.05520.0000−5.05520.0000
f22−5.08770.0000−5.08770.0000−5.08770.0000−5.08770.0000−5.08770.0000−5.08770.0000
f23−5.12850.0000−5.12850.0000−5.12850.0000−5.12850.0000−5.12850.0000−5.12850.0000
Table 13. Statistical results of applying improvement strategies.
Table 13. Statistical results of applying improvement strategies.
FunctionBPSOBGWOBROA_S3BROA_V2BROA_U3BROA_Z3
advstdadvstdadvstdadvstdadvstdadvstd
f10.93330.63973.10001.29590.0000 0.00000.00000.00000.00000.00000.00000.0000
f20.73330.58332.80001.21490.00000.00000.00000.00000.00000.00000.00000.0000
f38.03336.6978117.4667103.40040.00000.00000.00000.00000.00000.00000.00000.0000
f41.00000.00001.00000.00001.00000.00000.86670.34570.00000.00001.00000.0000
f5259.8333112.419216.966738.59203.500019.170328.03335.294729.00000.000029.466724.1400
f68.70001.126513.23332.14857.50000.00007.50000.00007.50000.00007.50000.0000
f74.72103.006535.400219.31170.00040.00030.00050.00050.00030.00040.00080.0008
f8−24.51490.3653−25.24410.0000−25.24410.0000−25.24410.0000−25.21610.1536−25.24410.0000
f91.00000.64332.56671.52410.00000.00000.00000.00000.00000.00000.00000.0000
f100.55480.35111.33550.26140.00000.00000.00000.00000.00000.00000.00000.0000
f110.02230.01300.13580.06010.00000.00000.00000.00000.00000.00000.00000.0000
f121.77460.07712.05880.17691.66900.00001.66900.00001.66900.00001.66900.0000
f130.10000.06430.00000.00000.00000.00000.00330.01830.00000.00000.00000.0000
f1412.67050.000012.67050.000012.67050.000012.67050.000012.67050.000012.67050.0000
f150.14840.00000.14840.00000.14840.00000.14840.00000.14840.00000.14840.0000
f160.00000.00000.00000.00000.00000.00000.00000.00000.00000.00000.00000.0000
f1727.70290.000027.70290.000027.70290.000027.70290.000027.70290.000027.70290.0000
f18600.00000.0000600.00000.0000600.00000.0000600.00000.0000600.00000.0000600.00000.0000
f19−0.33480.0000−0.33370.0063−0.33480.0000−0.33480.0000−0.33480.0000−0.33480.0000
f20−0.16570.0000−0.13340.0554−0.16570.0000−0.16570.0000−0.16570.0000−0.16570.0000
f21−5.05520.0000−5.05520.0000−5.05520.0000−5.05520.0000−5.05520.0000−5.05520.0000
f22−5.08770.0000−5.08770.0000−5.08770.0000−5.08770.0000−5.08770.0000−5.08770.0000
f23−5.12850.0000−5.12850.0000−5.12850.0000−5.12850.0000−5.12850.0000−5.12850.0000
Table 14. p-value of the Wilcoxon rank-sum test.
Table 14. p-value of the Wilcoxon rank-sum test.
FunctionPBROA_S3PBROA_V2PBROA_U3PBROA_Z3
f10.0156111
f20.0425111
f30.2500111
f40.00010.25000.00001
f50.00800.86590.74650.0021
f60.0625111
f70.00130.00230.00340.0033
f810.25000.04251
f91111
f100.2500111
f110.1250110.2500
f120.1250111
f1310.12500.03351
f141111
f151111
f161111
f171111
f181111
f191111
f201111
f211111
f221111
f231111
Table 15. Information about the testing datasets.
Table 15. Information about the testing datasets.
DatasetInstancesNumber of FeaturesNumber of CategoriesAttribute Types
Breast Cancer28492Categorical
Breast Cancer Wisconsin69996Integer
Glass21496Real
Iris15043Real
South German Credit1000204Integer, real
Wine178133Integer, real
Flags194288Categorical, integer
Dermatology366336Categorical, integer
Credit Approval690152Categorical, integer, real
Chess(kr-vs-kp)3196362Categorical
Hepatitis155192Categorical, integer, real
Image Segmentation210197Real
Statlog (Australian Credit Approval)690142Categorical, integer, real
Wall-Following Robot Navigation Data5456254Real
Table 16. The result of fitness value.
Table 16. The result of fitness value.
DatasetBPSOBGWOPBROA_S3PBROA_V2PBROA_U3PBROA_Z3
Breast Cancer0.10940.10300.10200.10390.10790.1010
Breast Cancer Wisconsin0.00420.00460.00370.00400.00410.0036
Glass0.12730.12580.12790.12140.12830.1208
Iris0.00250.00290.00250.00250.00250.0025
South German Credit0.46470.44210.45440.45690.45800.4594
Wine0.00200.00340.00180.00190.00170.0018
Flags0.22390.18770.20680.21970.22010.2120
Dermatology0.00350.00460.00310.00320.00320.0035
Credit Approval0.08170.08170.07040.07840.07800.0748
Chess(kr-vs-kp)0.02870.03230.02250.03140.03820.0257
Hepatitis0.02710.02260.00980.01790.01760.0195
Image Segmentation0.00680.00460.00670.00980.00660.0087
Statlog (Australian Credit Approval)0.06680.07640.06000.06600.06630.0649
Wall-Following Robot Navigation Data0.04470.05000.03590.03950.04060.0408
Table 17. The number of selected features.
Table 17. The number of selected features.
DatasetBPSOBGWOPBROA_S3PBROA_V2PBROA_U3PBROA_Z3
Breast Cancer3.13.63.53.653.33.4
Breast Cancer Wisconsin3.754.153.33.63.653.25
Glass5.355.054.86.46.24.75
Iris11.151111
South German Credit9.6598.458.49.29
Wine2.654.452.42.452.252.3
Flags10.611.25109.710.759.9
Dermatology11.4515.1510.1510.4510.4511.4
Credit Approval5.26.355.54.555.155.15
Chess (kr-vs-kp)19.519.9519.4522.629.1518.95
Hepatitis6.056.056.556.656.16.7
Image Segmentation8.58.658.27.4587.6
Statlog (Australian Credit Approval)4.24.554.153.553.94.45
Wall-Following Robot Navigation4.856.654.64.74.454.6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pan, J.-S.; Shi, H.-J.; Chu, S.-C.; Hu, P.; Shehadeh, H.A. Parallel Binary Rafflesia Optimization Algorithm and Its Application in Feature Selection Problem. Symmetry 2023, 15, 1073. https://doi.org/10.3390/sym15051073

AMA Style

Pan J-S, Shi H-J, Chu S-C, Hu P, Shehadeh HA. Parallel Binary Rafflesia Optimization Algorithm and Its Application in Feature Selection Problem. Symmetry. 2023; 15(5):1073. https://doi.org/10.3390/sym15051073

Chicago/Turabian Style

Pan, Jeng-Shyang, Hao-Jie Shi, Shu-Chuan Chu, Pei Hu, and Hisham A. Shehadeh. 2023. "Parallel Binary Rafflesia Optimization Algorithm and Its Application in Feature Selection Problem" Symmetry 15, no. 5: 1073. https://doi.org/10.3390/sym15051073

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop