Next Article in Journal
Biomimetic Action of Zinc Hydroxyapatite on Remineralization of Enamel and Dentin: A Review
Next Article in Special Issue
On the Mechanical Behaviour of Biomimetic Cornstalk-Inspired Lightweight Structures
Previous Article in Journal
Study on Performance Simulation of Vascular-like Flow Channel Model Based on TPMS Structure
Previous Article in Special Issue
Development of a Bionic Tube with High Bending-Stiffness Properties Based on Human Tibiofibular Shapes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Coverage Optimization of Heterogeneous Wireless Sensor Network Based on Improved Wild Horse Optimizer

1
Electrical Engineering College, Guizhou University, Guiyang 550025, China
2
College of Forestry, Guizhou University, Guiyang 550025, China
3
College of Computer Science and Technology, Guizhou University, Guiyang 550025, China
4
Power China Guizhou Engineering Co., Ltd., Guiyang 550001, China
*
Authors to whom correspondence should be addressed.
Biomimetics 2023, 8(1), 70; https://doi.org/10.3390/biomimetics8010070
Submission received: 11 January 2023 / Revised: 29 January 2023 / Accepted: 2 February 2023 / Published: 6 February 2023
(This article belongs to the Special Issue Bio-Inspired Design and Optimisation of Engineering Systems)

Abstract

:
One of the most important challenges for heterogeneous wireless sensor networks (HWSNs) is adequate network coverage and connectivity. Aiming at this problem, this paper proposes an improved wild horse optimizer algorithm (IWHO). Firstly, the population’s variety is increased by using the SPM chaotic mapping at initialization; secondly, the WHO and Golden Sine Algorithm (Golden-SA) are hybridized to improve the WHO’s accuracy and arrive at faster convergence; Thirdly, the IWHO can escape from a local optimum and broaden the search space by using opposition-based learning and the Cauchy variation strategy. The results indicate that the IWHO has the best capacity for optimization by contrasting the simulation tests with seven algorithms on 23 test functions. Finally, three sets of coverage optimization experiments in different simulated environments are designed to test the effectiveness of this algorithm. The validation results demonstrate that the IWHO can achieve better and more effective sensor connectivity and coverage ratio compared to that of several algorithms. After optimization, the HWSN’s coverage and connectivity ratio attained 98.51% and 20.04%, and after adding obstacles, 97.79% and 17.44%, respectively.

1. Introduction

Over the past decade, wireless sensor networks (WSNs) have been used in different fields, such as urban management, environmental monitoring, disaster prevention, and military applications, etc. [1,2,3,4]. A large number of tiny sensors make up the self-organizing distributed network system known as the WSN, and the sensors are typically heterogeneous. In WSN applications, coverage and connectivity are important indicators for determining whether real-time data can be provided to users through the inter-collaboration of sensors. However, the traditional WSN coverage approach deploys sensors at random. This approach will result in insufficient coverage, causing communication conflicts [5,6]. In existing research, scholars usually consider coverage when optimizing HWSN coverage, but connectivity is frequently overlooked. Therefore, this paper studies how to improve the coverage and connectivity of HWSNs.
The swarm intelligence (SI) optimization algorithm is a biologically inspired method that is one of the most successful strategies for solving optimization problems [7,8]. It is characterized by a fast search speed and strong search capability, avoiding complex theoretical derivation. Examples include particle swarm optimization (PSO) [9], bald eagle search optimization algorithm (BES) [10], cuckoo search (CS) [11], sparrow search algorithm (SSA) [12], northern goshawk optimization (NGO) [13], mayfly optimization algorithm (MA) [14], gray wolf optimization algorithm (GWO) [15], Harris hawks optimization (HHO) [16], coot optimization algorithm (COOT) [17], wild horse optimizer (WHO) [18] and other algorithms.
The WHO was put forth by Naruei et al. in 2021 as a method of solving algebraic optimization issues. Its optimization performance has significant advantages over the majority of classical algorithms, and it has been widely used to solve various engineering problems. In 2022, Milovanović et al. applied the WHO to multi-objective energy management in microgrids [19]. Ali et al. applied the WHO to the frequency regulation of a hybrid multi-area power system with a new type of combined fuzzy fractional order PI and TID controllers [20]. Furthermore, many researchers have improved WHO to improve its optimization capability. In 2022, Li et al. proposed a hybrid multi-strategy improved wild horse optimizer, which can improve the algorithm’s convergence speed, accuracy, and stability [21]. Ali et al. proposed an improved wild horse optimization algorithm for reliability-based optimal DG planning of radial distribution networks. This algorithm is a high-performance optimization method in terms of exploration–exploitation balance and convergence speed [22].
In 2017, Tanyildizi et al. proposed the Gold-SA algorithm [23], which is based on the sine trigonometric function. This algorithm uses a golden sine operator to condense the solution space, efficiently avoiding the local optimal outcome and quickly approaching the global optimum. Additionally, the algorithm contains few parameters and algorithm-dependent operators, which can be well integrated with the other algorithms. In 2022, Wang et al. proposed an improved crystal structure algorithm for engineering optimization problems. This algorithm makes good use of the relationship between the golden sine operator and the unit circle to make the algorithm exploration space more comprehensive, which can effectively speed up the convergence rate of the algorithm [24]. Yuan et al. proposed a hybrid golden jackal optimization and golden sine algorithm with dynamic lens imaging learning for global optimization problems; the golden sine algorithm is integrated to improve the ability and efficiency of golden jackal optimization [25]. In 2023, Jia et al. proposed the fusion swarm-intelligence-based decision optimization for energy-efficient train-stopping schemes. Their algorithm incorporates the golden sine strategy to improve the performance of the algorithm [26].
In recent years, SI optimization algorithms have been used by many scholars for the study of WSN coverage optimization, and fruitful results have been achieved with the continuous development of SI. In 2013, Huang et al. proposed an AFSA-based coverage optimization method for WSN. Simulation results show that AFSA increases the sensors’ coverage in WSN [27]. In 2015, Zhang proposed a hybrid algorithm of particle swarm and firefly, with particle swarm as the main body and firefly for local search, thus improving the sensor coverage [28]. In 2016, Wu et al. suggested an improved adaptive PSO-based coverage optimization. This approach first increases the evolution factor and aggregation factor to improve the inertia weights, and then, in order to ensure that the particle population is diverse, it introduces a collision resilience strategy during each iteration of the algorithm [29]. In 2018, Lu et al. proposed an FA-based WSN coverage optimization technique that involves switching out two sensors’ placements at once to increase network coverage [30]. In 2019, Nguyen et al. suggested a powerful genetic algorithm based on coverage optimization, effectively addressing various drawbacks of the current metaheuristic algorithms [31].
Although these SI optimization algorithms have produced many positive results, there is still room for further research into the algorithm’s performance and the optimization of WSN coverage. This research suggests an IWHO to optimize sensor coverage and connectivity. The main contributions are the following:
  • We improve the WHO algorithm in order to achieve better optimization. The SPM chaotic map is used to improve the population’s quality. The WHO and Golden-SA are hybridized to improve the WHO’s accuracy and arrive at faster convergence. The Cauchy variation and opposition-based learning strategies are also used to avoid falling into a local optimum and broaden the search space.
  • We test 23 test functions and compare the results to the performance of the IWHO and seven other algorithms. The findings reveal that IWHO has a stronger optimization performance than the others. We use the IWHO to optimize the coverage of a homogeneous WSN and compare the performance with five other algorithms and four improved algorithms proposed in References. The experimental data demonstrate that IWHO can optimize WSN coverage more effectively.
  • The HWSN coverage problem is optimized using the IWHO, which significantly increases coverage and the connectivity ratio. With the situation of barriers, the same high level of coverage and connectivity of sensors is attained.

2. WSN Coverage Model

Suppose that the monitoring area is a two-dimensional region with an area of M × N. The n sensors are randomly arranged in this area and they can be expressed as U = {u1, u2un}. Assume that the sensors are heterogeneous with different sensing radii Rs and communication radii Rc, and Rc 2Rs. Every sensor can move, and their position can be updated instantly. The sensor is centered on itself and has a sensing radius Rs as its radius, covering a circular area. If the coordinates of the detected arbitrary sensor Ui are (xi,yi), the coordinates of the target detected sensor Oj are (xj,yj). The Euclidean distance from the detected arbitrary sensor Ui to the target detected sensor Oj is expressed as:
d ( U i , O j ) = ( x i x j ) 2 + ( y i y j ) 2
The probability of sensor Oi being perceived by sensor Ui is denoted by p(Ui,Oi). It signifies that the goal is covered and the probability is 1 when the distance between sensors is smaller than Rs. It is 0 if it is not covered. The expression is as follows:
p U i , O j = 0 R s > d U i , O j 1 R s d U i , O j
Joint sensors’ perception probabilities are defined as follows:
P U , O j = 1 i = 1 n 1 p U i , O j
The coverage ratio is an important indicator of the HWSN deployment problem. The coverage ratio is calculated as follows:
f 1 = C o v = j = 1 M × N P U , O j M × N
The utilization of sensor coverage is evaluated using coverage efficiency. Higher coverage efficiency means achieving the same coverage area with fewer sensors. It is determined by dividing the region’s effective coverage range by the sum of all of the sensors’ coverage ranges. The coverage efficiency CE is calculated as shown in the following equation, where Ai denotes the area covered by the i-th sensor:
C E = i = 1 n A i i = 1 n A i
In the coverage problem, the connectivity ratio is equally as important as the coverage among sensors. To ensure the reliability of network connectivity, each sensor should be able to connect with at least two or more sensors. If the separation between two sensors is within the Rc, it is known that the two sensors can be connected to each other. The connectivity ratio between sensor Oi and target detection sensor Oj (i ≠ j) can be defined as:
p O i , O j = 0 R c > d O i , O j 1 R c d O i , O j
As shown in Figure 1, a network is connected if there is a path between any two sensors. The connectivity ratio is the proportion of connected paths to the maximum connected paths between sensors. The expression is as follows:
f 2 = i = 1 n j = 1 n P O i , O j n ( n 1 ) / 2 ( i j )
The amount of paths between any two nodes is n(n − 1)/2. Therefore, according to sensor coverage and connection, the objective function is:
max F ( f 1 , f 2 ) = w 1 f 1 + w 2 f 2 s t . w 1 + w 2 = 1 j = 1 M × N P ( U , O j ) M × N p ( U i , O j ) 0 p ( O i , O j ) 0
After several experiments, the values of w1 and w2, respectively were 0.9 and 0.1.

3. Wild Horse Optimizer

The social behavior of wild horses served as a model for the WHO. In the population construction of wild horses, there exist stallions and the rest of the horse herd. The WHO is designed and optimized for various problems based on group behavior, grazing, mating, dominance, and leadership among the stallions and herds in the wild horse population.
  • Establishing an initial population and choosing leaders
The population members are first distributed at random throughout the search ranges. In the beginning, we group this population. If there are N members overall, the number of stallions is G = (N × PS). PS represents the percentage of stallions in the herd, and it serves as the algorithm’s control parameter. The algorithm begins with the group leaders being chosen at random, and as the algorithm progresses, the leaders are chosen based on which group member has the best fitness function.
  • Grazing and mating of horses
The stallion is regarded as the center of the grazing area, and the group members move about the center to promote grazing behavior. We propose Equation (9) to model the grazing behavior. Members of the group move and conduct searches with varying radii around the leader:
X i , G j = 2 Z cos ( 2 π R Z ) ( S t a l l i o n j X i , G j ) + S t a l l i o n j
where Stallion is the leader’s position and R is a random number within [−2, 2]; Z is calculated as:
P = R 1 < T D R ;   I D X = ( P = = 0 ) ;   Z = R 2 Θ I D X + R 3 Θ ( ~ I D X )
where P is a vector consisting of 0 and 1. The random numbers R 1 , R 2 , and R 3 have a uniform distribution and fall between [0, 1]. Returns for the IDX indices of the random vector R 1 that satisfy the condition (P == 0). During algorithm execution, TDR declines, starting at 1 and eventually reaching 0. The expression is as follows:
T D R = 1 i t e r 1 max i t e r
where iter indicates how many iterations are being performed right now, and maxiter indicates the maximum number of iterations.
In order to simulate the behavior of horses leaving and mating, Equation (12) has proposed the same Crossover operator as the mean value type:
X G , K p = C r o s s o v e r ( X G , i q , X G , j z ) , i j k , p = q = e n d ; C r o s s o v e r = M e a n
  • Group is led by a leader
In nature, leaders mostly guide groups to appropriate living environments. If another population dominates the habitat, then that population must leave it. Equation (13) allows calculation of the location of the next habitat searched by the leader in each population:
S t a l l i o n G i = 2 Z cos ( 2 π R Z ) ( W H S t a l l i o n G i ) + W H , i f R 3 > 0.5   2 Z cos ( 2 π R Z ) ( W H S t a l l i o n G i ) W H , e l s e
where WH is the current location of the most suitable habitat, and R, RZ, and Z are defined as before.
  • Exchange of leaders
As mentioned in the population initialization phase, the two positions are switched if a group member has a greater fitness value than the leader:
S t a l l i o n G i = X G i , i f cos t ( X G i ) < cos t ( S t a l l i o n G i )     S t a l l i o n G i , e l s e

4. Improved Wild Horse Optimizer

To address the issues of the original algorithm, including the difficulty of escaping local optima, lack of accuracy during convergence, and slow speed, three methods are introduced in this study as ways to enhance the WHO algorithm. They improve the starting population, boost optimization speed and accuracy, increase the optimal solution’s disruption, and aid the algorithm’s exit from the local optimum.

4.1. SPM Chaotic Mapping Initialization Population

The optimization ability can benefit from a diversified initial population. In the original WHO algorithm, the rand function is used to randomly initialize the population, which results in uneven distribution and overlapping individuals, and the population diversity decreases rapidly in the later iterations. Chaos is a unique and widespread form of acyclic motion in nonlinear systems, which is widely used in population intelligence algorithms for optimizing the diversity of populations because of its ergodic and stochastic nature. In this paper, we introduce the SPM chaotic mapping model, which has superior chaotic and ergodic properties [32]. The expression is shown as follows:
X t + 1 = mod ( x ( t ) η + μ sin ( π x ( t ) ) + r , 1 ) , 0 x ( t ) < η mod ( x ( t ) / n 0.5 η + μ sin ( π x ( t ) ) + r , 1 ) , 0 x ( t ) < 0.5 mod ( 1 x ( t ) / n 0.5 η + μ sin ( π ( 1 x ( t ) ) ) + r , 1 ) , 0.5 x ( t ) < 1 η mod ( 1 x ( t ) η + μ sin ( π ( 1 x ( t ) ) ) + r , 1 ) , 1 η x ( t ) < 1
Scholars usually choose different chaotic mapping models for optimization of population initialization of population intelligence algorithms. This paper selects the Logistic mapping and Sine mapping with high usage rate, and compares them with SPM mapping under the condition of setting the same initial value and iterating 2000 times. Figure 2 shows the histograms of the three chaotic mappings, where the horizontal coordinate is the chaotic value and the vertical coordinate is the frequency of that chaotic value. The results prove that SPM mapping has better chaos performance and traversal. SPM mapping was therefore chosen to enhance population variety and make the population distribution more uniform. Figure 3 shows the population distribution of different chaotic mappings when the population size is 2000. The Logistic and Sine mappings show many individuals overlapping at the boundary, whereas the SPM mapping has a more uniform distribution.

4.2. Golden Sine Algorithm

To fix the disadvantages of the stallion position update method in the WHO algorithm, the Golden-SA algorithm is used in this research. In this paper, the stallion position is updated using the golden sine operator to condense the algorithm’s solution space and enhance the capability of the optimal search. The expression is shown as follows:
S t a l l i o n G i = S t a l l i o n G i sin ( r 1 ) r 2 sin ( r 1 ) x 1 W H x 2 S t a l l i o n G i
where r1, r2 are random numbers between [0, 2π] and [0, π]. In the following iteration, r1 determines the distance that individual i travels and r2 determines the direction in which individual i travels. x1 and x2 are the golden partition coefficients. During the iteration, they are utilized to condense the search space and direct the solution to the place that is globally optimal. Its partitioning implementation is shown below:
x 1 = a 1 τ + b τ x 2 = a τ + b 1 τ τ = 5 1 / 2
where a and b are the initial golden mean, and τ is the golden mean ratio.

4.3. Cauchy Variation and Opposition-Based Learning

The WHO algorithm does not perturb the optimal solution after each iteration, which can keep the solution in a locally optimal state. To address this issue, we employ Cauchy variation and opposition-based learning strategies to perturb the solution. This was inspired by MAO et al., who proposed a chaotic squirrel search algorithm with a mixture of stochastic opposition-based learning and Cauchy variation in 2021 [33].

4.3.1. Cauchy Variation

Gaussian and Cauchy distributions are two similar classical probability density distribution functions, and Gaussian variation has also been used by scholars in algorithm improvements. Figure 4 displays the probability density function curves for both.
In comparison to the Gaussian distribution, as seen in Figure 4, the Cauchy distribution is longer and flatter at both ends. It flattens out as it approaches 0, moves more slowly, and has a smaller peak close to the origin. As a result, Cauchy variance has better perturbation properties than Gaussian variance. As a result, introducing the Cauchy variation strategy can expand the search space and improve the perturbation ability. The expression is as follows:
W H = W H ( 1 + ( 1 / M ax _ i t e r ) tan ( π ( r a n d 0.2 ) ) )

4.3.2. Opposition-Based Learning

By building the reverse solution from the existing solution, the opposition-based learning strategy can broaden the solution space of algorithms. To determine which solution is preferable, existing solution is contrasted with the reverse solution. The expression for incorporating the opposition-based learning strategy into the WHO algorithm is as follows:
W H b a c k = u b + r a n d ( l b W H ) W H = W H b a c k + b 1 W H W H b a c k b 1 = ( max _ i t e r l / max _ i t e r ) l
where WHback is the reverse solution to the stallion’s optimal position in the lth generation, rand is a random matrix of dimension obeying the standard uniform distribution of (0, 1), upper and lower boundaries are denoted by ub and lb, respectively, and b1 denotes the information exchange control parameter.

4.3.3. A Dynamic Selection Probability

This paper sets a dynamic selection probability Pz to choose the strategy to update the stallion position more appropriately. The Pz is shown as follows:
P z = exp ( 1 l / max _ i t e r ) 20 + 0.05
Pz will be compared to a number chosen at random between (0,1). If Pz > rand then a opposition-based learning strategy starts to work. If Pz < rand then the Cauchy variation strategy is utilized to disturb at the present stallion.

4.4. The Pseudo Code of IWHO

  • Initialize the first population of horses using the new SPM chaotic sequence
  • Input IWHO parameters, PC = 0.13, PS = 0.1, a = π, b = −π
  • Calculate the fitness
  • Create foal groups and select stallions
  • Find the best horse
  • While the end criterion is not satisfied
  • Calculate TDR using Equation (11)
  • For number of stallions
  • Calculate Z using Equation (10)
  • For number of foals of any group
  • If rand > PC
  • Update the position of the foal using Equation (9)
  • Else
  • Update the position of the foal using Equation (12)
  • End
  • End
  • Update the position of the StallionGi using Equation (16)
  • If cost (StallionGi) < cost (Stallion)
  • Stallion = StallionGi
  • End
  • Sort foals of group by cost
  • Select foal with minimum cost
  • If cost (foal) < cost (Stallion)
  • Exchange foal and Stallion position using Equation (14)
  • End
  • Calculate Pz using Equation (20)
  • If Pz < rand
  • Update the position of Stallion using Equation (19)
  • Else
  • Update the position of Stallion using Equation (18)
  • End
  • End
  • Update optimum
  • End

4.5. Time Complexity Analysis

Time complexity is a significant factor in determining an algorithm’s quality and shows how effectively it operates. The time complexity of the WHO algorithm can be represented as O(N×D×L), where N is the entire population, D is the search space’s dimension, and L is the maximum number of iterations. The following is a depiction of the IWHO algorithm’s time complexity analysis:
  • The population is initialized using the SPM chaotic mapping model, and the time complexity is as follows:
T 1 = O N 1 + N 2 × D = O N × D
  • The position update formula of the original stallion is replaced by the golden sine strategy without the addition of any extra strategies. The time complexity is as follows:
T 2 = O N × D × L
  • After iterations, incorporating the Cauchy variance and opposition-based learning, the time complexity is as follows:
T 3 = O N × D × L + O N × D × L = O N × D × L
  • Thus, the IWHO time complexity is as follows:
T 1 + T 2 + T 3 = O N × D + O N × D × L + O N × D × L = O N × D × L
In conclusion, IWHO has the same time complexity as WHO, and the three improvement techniques do not make the algorithm’s time complexity any more difficult.

5. IWHO Algorithm-Based Coverage Optimization Design

The process of finding a suitable habitat for a horse herd is analogous to the process of obtaining the optimal coverage of sensors, and the position of the stallion represents the coordinates covered by the sensors. Using the same number of sensors to cover a bigger area while maintaining effective communication is the aim of WSN optimization coverage based on the IWHO. These are the steps:
Step 1: Enter the size of the area to be detected by the WSN, the number of sensors, sensing radius, communication radius, and the IWHO algorithm’s settings;
Step 2: The population is initialized according to Equation (15), where each individual represents a coverage scheme. At this step, the sensors are dispersed randomly around the monitoring region, and Equation (8) is used to determine the initial coverage and connectivity;
Step 3: Update the location information of stallions and foals, and calculate the corresponding adaptation degree. Update the coverage ratio and connectivity ratio according to Equation (8). Find the optimal sensor location;
Step 4: Create a new solution by perturbing at the optimal solution position through dynamic probabilistic selection of a Cauchy variation or a opposition-based learning strategy;
Step 5: Immediately exit the loop if the condition is met. Output the sensor’s best coverage scheme.

6. Simulation Experiments and Analysis

6.1. IWHO Algorithm Performance Test Analysis

6.1.1. Simulation Test Environment

The environment for this simulation test was: Windows 10 Professional, 64-bit OS, Intel(R) Core (TM) i5-4210H CPU @2.90 GHz, 8GB. The simulation software was MATLAB 2016a.

6.1.2. Comparison Objects and Parameter Settings

In this paper, the WHO, SSA, NGO, MA, PSO, COOT, GWO and the IWHO algorithms were selected for comparison. In order to make a fair comparison between each algorithm, we have unified the number of consumed fitness evaluations in the experiment, and the number of consumed fitness evaluations by each algorithm is 30,000. The parameters ware set as shown in Table 1.
In GWO, α represents the control parameter of a grey wolf when hunting prey. In SSA, ST represents the alarm value, PD represents the number of producers, SD represents the number of sparrows who perceive the danger. In MA, g represents the inertia weight, a1 represents the personal learning coefficient, and a2 and a3 represent the global learning coefficient. In PSO, c1 and c2 represent the learning coefficient, wmin and wmax represent the upper and lower limits of inertia weight. In COOT, R, R1 and R2 are random vectors along the dimensions of the problem.

6.1.3. Benchmark Functions Test

To verify the IWHO algorithm’s capacity for optimization, 23 benchmark functions were chosen for simulation in this study. There are three categories of test function, among which F1–F7 listed in Table 2 are single-peak test functions, F8–F13 are multi-peak test functions listed in Table 3 and F14–F23 are fixed-dimension test functions listed in Table 4. The single-peak test functions are characterized by a single extreme value and are used to test the convergence speed and convergence accuracy of the IWHO. The multi-peak test functions are characterized by multiple local extremes, which can be applied to evaluate the IWHO algorithm’s local and global search capabilities.

6.1.4. Simulation Test Results

After 30 independent runs, Table 5 displays the average value and standard deviation for eight algorithms. The IWHO outperforms others. For the single-peak test functions F1–F4 and F7, the IWHO has the best performance and the fastest convergence speed, with a standard deviation of 0, demonstrating that it is robust and stable. IWHO has the second-best performance for F5 and F6. For F9 and F11, although most of the algorithms converge to the ideal optimum, as illustrated in Figure 5i,k, the IWHO converges faster. Both the IWHO and SSA algorithms have the best results for F10, however, the IWHO goes through fewer iterations, as can be seen in Figure 5j. For F13, both the PSO algorithm and the SSA algorithm outperform the IWHO, which is ranked third among the eight optimization algorithms. The results in Table 5 and Figure 5 for the fixed-dimension test functions show that the IWHO converges to the theoretical optimum with a standard deviation of 0.
In conclusion, the results of 30 independent experiments with eight different algorithms and 23 benchmark test functions show that the IWHO can converge to the ideal optimal values for 14 test functions, F1, F3, F9, F11, F14–F23. This means that the convergence accuracy of IWHO has been improved a lot, but when optimizing some test functions, although the optimal value can be obtained, the convergence speed can still be improved. Additionally, some test functions still fall into local optimization. Among the 23 test results, IWHO ranked first 20 times, second twice, and third once. It can be concluded that the IWHO performs better than the WHO algorithm and effectively improves some of its flaws. Similarly, the IWHO has significant advantages over other algorithms.

6.1.5. Wilcoxon Rank Sum Test

Such data analysis lacks integrity and scientific validity if it only compares and analyzes the mean and standard deviation of the different algorithms. We therefore chose a non-parametric statistical test, the Wilcoxon rank sum test, to further validate the algorithm’s performance. We ran each algorithm independently for 30 times in 23 test functions. For Wilcoxon rank sum test and p calculation, we compared the experimental results of the other algorithms with those of the IWHO. When p < 5%, it was marked it as “+”, indicating that IWHO was better than the comparison algorithm. When p > 5%, it was marked as “−”, indicating that IWHO is inferior to the comparison algorithm. When p was equal to 1, it indicates that it is not suitable for judgment.
The comparison results are shown in Table 6. In the comparison of various algorithms, most of the rank sum test p values are less than 0.05, which indicates that the IWHO has significant differences with other algorithms; that is, the IWHO algorithm has better optimization performance.

6.2. Coverage Performance Simulation Test Analysis

To verify the performance of IWHO in optimizing coverage in HWSNs, three experiments were set up by simulating different scenarios and setting different parameters.
  • Experiment 1 used homogenous sensors, and the IWHO was used to improve the coverage ratio of WSN. In order to demonstrate the IWHO algorithm’s efficacy, five algorithms and four4 improved algorithms were chosen for comparison.
  • The results of Experiment 1 show how effective the IWHO is in solving the sensor coverage optimization issue. In Experiment 2, the IWHO was used in an HWSN to increase sensor coverage and ensure sensor connectivity.
  • In Experiment 3, an obstacle was added to the monitoring area to simulate a more realistic scenario. The sensors must avoid the obstacle for coverage, and the IWHO was used to improve the coverage and connectivity ratio of the sensors in the monitoring area.

6.2.1. Simulation Experiment 1 Comparison Results Analysis

In Experiment 1, we only considered the improvement in coverage ratio using Equation (4) as the objective function. To avoid the possibility of the algorithm, the algorithm was run thirty times independently. Table 7 displays the settings for the sensor parameters.
Figure 6a shows how the sensors were initially covered in the monitoring area at random. With the optimization of the IWHO, the overlapping sensors started to decrease. Finally, as shown in Figure 6b, they were evenly distributed in the monitoring area. Table 8 demonstrates that the initial coverage reached 79.13%, while after the optimization of the IWHO the coverage reached 97.58%, which is an improvement of 18.45%. At the beginning, there were more redundant sensors in the region, and the region had more obvious energy voids and appeared to be cluttered. After the optimization of the algorithm, the sensor distribution became obviously uniform and the coverage ratio was significantly improved. As a result, the IWHO is effective for WSN coverage optimization.
WHO, PSO, WOA, BES, HHO algorithms were selected for further comparative experiments. All algorithms were run independently thirty times. Table 6 displays the settings for the sensor parameters.
Table 9 demonstrates that the IWHO, with a coverage ratio of 97.58%, has the best optimization result. The results of the experiments can prove that the IWHO shows better results than the WHO algorithm in solving the WSN coverage optimization problem, with an average coverage improvement of 5.52% and a coverage efficiency improvement of 3.93% after thirty independent runs. It also outperforms the other four algorithms. The coverage ratio after IWHO optimization is 14.56% higher than the worst PSO algorithm and 1.1% higher than the best HHO algorithm. Figure 7 shows that the IWHO maintains a better coverage ratio for almost all iterations. Additionally, the IWHO algorithm’s coverage efficiency is the highest, coming in at 69.08%, surpassing the PSO algorithm’s coverage efficiency by 10.3%. It demonstrates that the IWHO is optimized by reducing redundancy in sensor coverage.
To further verify the superiority of IWHO in optimizing sensors coverage, the IWHO was compared with the four improved algorithms. To ensure fairness, the parameter settings are kept consistent with those in reference.
As Table 10, Table 11, Table 12 and Table 13 demonstrate, with the same parameter values, the IWHO improves WSN coverage.

6.2.2. Simulation Experiment 2 Comparison Results Analysis

In a complex sensor coverage environment, it is difficult to unify the types of sensors. Therefore, more of the HWSN is often covered in real environments. In Experiment 2, two different types of sensors were set up, and the sensors were dispersed at random throughout the monitoring area. The IWHO was used to optimize the HWSN coverage. Table 14 displays the settings for the sensor parameters. N1,2 indicates the number of two types of sensors.
As shown in Figure 8a,b and Table 15, IWHO improved coverage equally well when two kinds of different sensors are deployed. The optimized coverage achieved 98.51%, a 17.08% improvement over the initial state. In contrast to Experiment 1, IWHO improved connectivity while improving coverage. As illustrated in Figure 8c,d and Table 16, some of the sensors were not connected in the initial state, and the connectivity ratio was only 16.03%. However, after optimization, the connectivity ratio rose to 20.04%, and the overall network connectivity improved.

6.2.3. Simulation Experiment 3 Comparison Results Analysis

In the actual coverage area, there are some non-deployable units. In order to simulate a more realistic simulation scenario, a 25 m × 25 m obstacle was added to the monitoring area. Experiment 3 evaluated the IWHO algorithm’s performance in optimizing HWNs coverage with the existence of obstacles. Table 13 displays the settings for the sensor parameters.
Similarly, as shown in Figure 9a,b and Table 17, IWHO can also effectively improve the coverage ratio after adding obstacles to the monitoring area, reaching 99.25%, which is 16.96% higher than the initial state. As shown in Figure 9c,d and Table 18, some sensors were not connected at the start. Following optimization, the network connectivity ratio increased to 17.44%.

7. Conclusions

This paper proposes an IWHO to increase the coverage and connectivity of HWSNs. The WHO algorithm’s flaws include a slow rate of convergence, poor convergence accuracy, and a propensity for falling into a local optimum. How can the algorithm be made to perform better as a result? The population is initialized using the SPM chaotic mapping model, followed by the integration of the Golden-SA to enhance the algorithm’s search for optimization, and finally, the best solution is perturbed using the opposition-based learning and Cauchy variation strategy to avoid a local optimum being reached. This study evaluates the optimization performance of the IWHO, the WHO algorithm, and another seven algorithms using 23 benchmark test functions to confirm the performance of the IWHO. According to the results of the simulation, the IWHO outperforms other algorithms in terms of convergence accuracy and speed, as well as the ability to escape local optimization. Although the IWHO can find the ideal optimal value for the majority of test functions, the convergence rate can still be improved. For individual test functions, IWHO still falls into local optimization. To further verify IWHO’s superiority in optimizing sensor coverage, five other algorithms and four improved algorithms were used. The outcomes demonstrate that the IWHO achieves optimum sensor coverage and is superior to other algorithms. The coverage issue of HWSNs is finally resolved via IWHO, which optimizes sensor coverage and enhances sensor connectivity. After adding an obstacle, it can also be optimized to obtain good results.
However, the current research is still insufficient. We can continue enhancing the IWHO to increase its performance and researching the IWHO for multi-objective optimization so that it can handle more complex HWSN coverage and engineering issues.

Author Contributions

Conceptualization, C.Z.; methodology, C.Z.; software, C.Z.; validation, C.Z.; investigation, C.Z. and J.Y.; resources, C.Z. and J.Y.; data curation, C.Z.; writing—original draft preparation, C.Z.; writing—review and editing, J.Y. and C.Z.; visualization, C.Z.; supervision, C.L. and S.Y.; project administration, T.Q. and W.T.; funding acquisition, Z.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by NNSF of China (No. 61640014, No. 61963009), Guizhou provincial Science and Technology Projects (No. Qiankehe Zhicheng[2022]Yiban017, No. Qiankehe Zhicheng[2019]2152), Innovation group of Guizhou Education Department (No. Qianjiaohe KY[2021]012), Engineering Research Center of Guizhou Education Department (No. Qianjiaoji[2022]043, No. Qianjiaoji[2022]040), Science and Technology Fund of Guizhou Province (No. Qiankehejichu[2020]1Y266, No. Qiankehejichu [ZK [2022]Yiban103]), Science and Technology Project of Power Construction Corporation of China, Ltd. (No. DJ-ZDXM-2020-19), CASE Library of IOT (KCALK201708).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kavitha, M.; Geetha, B.G. An efficient city energy management system with secure routing communication using WSN. Clust. Comput. 2019, 22, 13131–13142. [Google Scholar] [CrossRef]
  2. Laiqa, B.I.; Rana, M.A.L.; Muhammad, F.; Hamza, A. Smart city based autonomous water quality monitoring system using WSN. Wirel. Pers. Commun. 2020, 115, 1805–1820. [Google Scholar]
  3. Ditipriya, S.; Rina, K.; Sudhakar, T. Semisupervised classification based clustering approach in WSN for forest fire detection. Wirel. Pers. Commun. 2019, 109, 2561–2605. [Google Scholar]
  4. Sunny, K.; Jan, I.; Bhawani, S.C. WSN-based monitoring and fault detection over a medium-voltage power line using two-end synchronized method. Electr. Eng. 2018, 100, 83–90. [Google Scholar]
  5. Zhang, Q.; Fok, M.P. A two-phase coverage-enhancing algorithm for hybrid wireless sensor networks. Sensors 2017, 17, 117. [Google Scholar] [CrossRef]
  6. Adulyasas, A.; Sun, Z.; Wang, N. Connected coverage optimization for sensor scheduling in wireless sensor networks. IEEE Sens. J. 2015, 15, 3877–3892. [Google Scholar] [CrossRef]
  7. Al-Betar, M.A.; Awadallah, M.A.; Faris, H. Natural selection methods for grey wolf optimizer. Expert Syst. Appl. 2018, 113, 481–498. [Google Scholar] [CrossRef]
  8. Gao, W.; Guirao, J.G.; Basavanagoud, B.; Wu, J.Z. Partial multi-dividing ontology learning algorithm. Inform. Sci. 2018, 467, 35–58. [Google Scholar] [CrossRef]
  9. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95—International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995; pp. 1942–1948. [Google Scholar]
  10. Alsattar, H.A.; Zaidan, A.A.; Zaidan, B.B. Novel meta-heuristic bald eagle search optimization algorithm. Artif. Intell. Rev. 2020, 53, 2237–2264. [Google Scholar] [CrossRef]
  11. Yang, X.S.; Deb, S. Cuckoo search: Recent advances and applications. Neural Comput. Appl. 2014, 24, 169–174. [Google Scholar] [CrossRef]
  12. Xue, J.K.; Shen, B. A novel swarm intelligence optimization approach: Sparrow search algorithm. Syst. Sci. Control Eng. Open Access J. 2020, 8, 22–34. [Google Scholar] [CrossRef]
  13. Dehghani, M.; Hubálovský, Š.; Trojovský, P. Northern goshawk optimization: A new swarm-based algorithm for solving optimization problems. IEEE Access 2021, 9, 162059–162080. [Google Scholar] [CrossRef]
  14. Konstantinos, Z.; Stelios, T. A mayfly optimization algorithm. Comput. Ind. Eng. 2020, 145, 106559. [Google Scholar]
  15. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef] [Green Version]
  16. Heidari, A.A.; Mirjalili, S.; Faris, H.; Ibrahim, A.; Majdi, M.; Huiling, C. Harris hawks optimization: Algorithm and applications. Future Gener. Comput. Syst. 2019, 97, 849–872. [Google Scholar] [CrossRef]
  17. Naruei, I.; Keynia, F. A new optimization method based on COOT bird natural life model. Expert Syst. Appl. 2021, 183, 115352. [Google Scholar] [CrossRef]
  18. Naruei, I.; Keynia, F. Wild horse optimizer: A new meta-heuristic algorithm for solving engineering optimization problems. Eng. Comput. -Ger. 2021, 17, 3025–3056. [Google Scholar] [CrossRef]
  19. Milovanović, M.; Klimenta, D.; Panić, M.; Klimenta, J.; Perović, B. An application of Wild Horse Optimizer to multi-objective energy management in a micro-grid. Electr. Eng. 2022, 104, 4521–4541. [Google Scholar] [CrossRef]
  20. Ali, M.; Kotb, H.; AboRas, M.K.; Abbasy, H.N. Frequency regulation of hybrid multi-area power system using wild horse optimizer based new combined Fuzzy Fractional-Order PI and TID controllers. Alex. Eng. J. 2022, 61, 12187–12210. [Google Scholar] [CrossRef]
  21. Li, Y.; Yuan, Q.; Han, M.; Cui, R.; Cui, R. Hybrid Multi-Strategy Improved Wild Horse Optimizer. Adv. Intell. Syst. 2022, 4, 2200097. [Google Scholar] [CrossRef]
  22. Ali, M.H.; Kamel, S.; Hassan, H.M.; Véliz, M.T.; Zawbaa, H.M. An improved wild horse optimization algorithm for reliability based optimal DG planning of radial distribution networks. Energy Rep. 2022, 8, 582–604. [Google Scholar] [CrossRef]
  23. Tanyildizi, E.; Demir, G. Golden sine algorithm: A novel math-inspired algorithm. Adv. Electr. Comput. Eng. 2017, 17, 71–78. [Google Scholar] [CrossRef]
  24. Wang, W.T.; Tian, J.; Wu, D. An Improved Crystal Structure Algorithm for Engineering Optimization Problems. Electronics 2022, 11, 4109. [Google Scholar] [CrossRef]
  25. Yuan, P.L.; Zhang, T.H.; Yao, L.G.; Lu, Y.; Zhuang, W.B. A Hybrid Golden Jackal Optimization and Golden Sine Algorithm with Dynamic Lens-Imaging Learning for Global Optimization Problems. Appl. Sci. 2022, 12, 9709. [Google Scholar] [CrossRef]
  26. Jia, X.G.; Zhou, X.B.; Bao, J.; Zhai, J.Y.; Yan, R. Fusion Swarm-Intelligence-Based Decision Optimization for Energy-Efficient Train-Stopping Schemes. Appl. Sci. 2023, 13, 1497. [Google Scholar] [CrossRef]
  27. Huang, Y.Y.; Li, K.Q. Coverage optimization of wireless sensor networks based on artificial fish swarm algorithm. Appl. Res. Comput. 2013, 30, 554–556. [Google Scholar]
  28. Zhang, Q. Research on Coverage Optimization of Wireless Sensor Networks Based on Swarms Intelligence Algorithm. Master’s thesis, Hunan University, Changsha, China, 2015. [Google Scholar]
  29. Wu, Y.; He, Q.; Xu, T. Application of improved adaptive particle swarm optimization algorithm in WSN coverage optimization. Chin. J. Sens. Actuators 2016, 2016, 559–565. [Google Scholar]
  30. Lu, X.L.; Cheng, W.; He, Q.; Yang, J.H.; Xie, X.L. Coverage optimization based on improved firefly algorithm for mobile wireless sensor networks. In Proceedings of the 2018 IEEE 4th International Conference on Computer and Communications, Chengdu, China, 7–10 December 2018; pp. 899–903. [Google Scholar]
  31. Nguyen, T.; Hanh, T.; Thanh, B.; Marimuthu, S.P. An efficient genetic algorithm for maximizing area coverage in wireless sensor networks. Inform. Sci. 2019, 488, 58–75. [Google Scholar]
  32. Ban, D.H.; Lv, X.; Wang, X. Efficient image encryption algorithm based on 1D chaotic map. Coll. Comput. Inf. 2020, 47, 278–284. [Google Scholar]
  33. Mao, Q.; Zhang, Q. Improved sparrow algorithm combining Cauchy mutation and opposition-based learning. J. Front. Comput. Sci. Technol. 2021, 15, 1155–1164. [Google Scholar]
  34. Lu, C.; Li, X.B.; Yu, W.J.; Zeng, Z.; Yan, M.M.; Li, X. Sensor network sensing coverage optimization with improved artificial bee colony algorithm using teaching strategy. Computing 2021, 103, 1439–1460. [Google Scholar] [CrossRef]
  35. Huang, Y.H.; Zhang, J.; Wei, W.; Qin, T.; Fan, Y.C.; Luo, X.M.; Yang, J. Research on coverage optimization in a WSN based on an improved coot bird algorithm. Sensors 2022, 22, 3383. [Google Scholar] [CrossRef] [PubMed]
  36. Miao, Z.M.; Yuan, X.F.; Zhou, F.Y.; Qiu, X.J.; Song, Y.; Chen, K. Grey wolf optimizer with an enhanced hierarchy and its application to the wireless sensor network coverage optimization problem. Appl. Soft Comput. 2020, 96, 106602. [Google Scholar] [CrossRef]
  37. Wang, S.P.; Yang, X.P.; Wang, X.Q.; Qian, Z.H. A virtual force algorithm-lévy-embedded grey wolf optimization algorithm for wireless sensor network coverage optimization. Sensors 2019, 19, 2735. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. HWSN coverage model.
Figure 1. HWSN coverage model.
Biomimetics 08 00070 g001
Figure 2. Chaotic Mapping Histogram. (a) Logistic mapping; (b) Sine mapping; (c) SPM mapping.
Figure 2. Chaotic Mapping Histogram. (a) Logistic mapping; (b) Sine mapping; (c) SPM mapping.
Biomimetics 08 00070 g002
Figure 3. Chaotic Mapping Scatter Diagram. (a) Logistic mapping; (b) Sine mapping; (c) SPM mapping.
Figure 3. Chaotic Mapping Scatter Diagram. (a) Logistic mapping; (b) Sine mapping; (c) SPM mapping.
Biomimetics 08 00070 g003
Figure 4. Curves of the probability density functions.
Figure 4. Curves of the probability density functions.
Biomimetics 08 00070 g004
Figure 5. Comparison of convergence curves. (a) F1; (b) F2; (c) F3; (d) F4; (e) F5; (f) F6; (g) F7; (h) F8; (i) F9; (j) F10; (k) F11; (l) F12; (m) F13; (n) F14; (o) F15; (p) F16; (q) F17; (r) F18; (s) F19; (t) F20; (u) F21; (v) F22; (w) F23.
Figure 5. Comparison of convergence curves. (a) F1; (b) F2; (c) F3; (d) F4; (e) F5; (f) F6; (g) F7; (h) F8; (i) F9; (j) F10; (k) F11; (l) F12; (m) F13; (n) F14; (o) F15; (p) F16; (q) F17; (r) F18; (s) F19; (t) F20; (u) F21; (v) F22; (w) F23.
Biomimetics 08 00070 g005
Figure 6. Coverage maps of sensors. (a) Initial coverage map of sensors; (b) Optimized coverage map of sensors.
Figure 6. Coverage maps of sensors. (a) Initial coverage map of sensors; (b) Optimized coverage map of sensors.
Biomimetics 08 00070 g006
Figure 7. Comparison of coverage convergence curves.
Figure 7. Comparison of coverage convergence curves.
Biomimetics 08 00070 g007
Figure 8. Coverage and connectivity map of sensors. (a) Initial coverage map of sensors; (b) Optimized coverage map of sensors; (c) Initial connectivity map of sensors; (d) Optimized connectivity map of sensors.
Figure 8. Coverage and connectivity map of sensors. (a) Initial coverage map of sensors; (b) Optimized coverage map of sensors; (c) Initial connectivity map of sensors; (d) Optimized connectivity map of sensors.
Biomimetics 08 00070 g008
Figure 9. Obstacle coverage and connectivity map of sensors. (a) Initial obstacle coverage map of sensors; (b) Optimized obstacle coverage map of sensors; (c) Initial obstacle connectivity map of sensors; (d) Optimized obstacle connectivity map of sensors.
Figure 9. Obstacle coverage and connectivity map of sensors. (a) Initial obstacle coverage map of sensors; (b) Optimized obstacle coverage map of sensors; (c) Initial obstacle connectivity map of sensors; (d) Optimized obstacle connectivity map of sensors.
Biomimetics 08 00070 g009
Table 1. Parameter settings of the algorithm.
Table 1. Parameter settings of the algorithm.
AlgorithmParameters
GWOα = [0, 2]
SSAST = 0.6, PD = 0.7, SD = 0.2
MAg = 0.8, a1 = 1, a2 = a3 = 1.5
PSOc1,c2 = 2, wmin = 0.2, wmax = 0.9
COOTR = [−1, 1], R1 = R2 = [0, 1]
WHOPS = 0.1, PC = 0.13
IWHOPS = 0.1, PC = 0.13, a = π, b = −π
Table 2. Single-peak test functions.
Table 2. Single-peak test functions.
FFunctionRangeDimfmin
F1 f 1 x = i = 1 n x i 2 [−100, 100]300
F2 f 2 x = i = 1 n x i + i = 1 n x i [−10, 10]300
F3 f 3 x = i = 1 n j = 1 n x j 2 [−100, 100]300
F4 f 4 x = max x i , 1 i n [−100, 100]300
F5 f 5 x = i = 1 n 1 100 x i + 1 x i 2 2 + x i 1 2 [−30, 30]300
F6 f 6 x = i = 1 n 1 x i + 0.5 2 [−100, 100]300
F7 f 7 x = max x i , 1 i n [−1.28, 1.28]300
Table 3. Multi-peak test functions.
Table 3. Multi-peak test functions.
FFunctionRangeDimfmin
F8 f 8 x = i = 1 n x i sin ( x i ) [−500, 500]30−418.9829 × n
F9 f 9 x = i = 1 n x i 2 10   cos ( 2 π x i ) + 10 [−5.12, 5.12]300
F10 f 10 x = 20 exp ( 0.2 1 n i = 1 n x i 2 ) exp ( 1 n i = 1 n cos ( 2 π x i ) ) + 20 + e [−32, 32]300
F11 f 11 x = i = 1 n x i 2 4000 i = 1 n cos ( x i i ) 1 [−600, 600]300
F12 f 12 x = π n 10 sin 2 ( π y i ) + i = 1 n 1 ( y i 1 ) 2 [ 1 + 10 sin 2 ( π y i + 1 ) ] + ( y n 1 ) 2 + i = 1 n u ( x i , 10 , 100 , 4 ) [−50, 50]300
F13 f 13 x = 0.1 sin 2 ( 3 π x i ) + i = 1 n ( x i 1 ) 2 1 + sin 2 ( 3 π x i + 1 ) + x n 1 2 1 + sin 2 ( 2 π x n ) + i = 1 n u ( x i , 5 , 100 , 4 ) [−50, 50]300
Table 4. Fixed-dimension test functions.
Table 4. Fixed-dimension test functions.
FFunctionRangeDimfmin
F14 f 14 x = ( 1 500 + j = 1 25 1 j + i = 1 2 ( x i a i j ) 6 ) 1 [−65, 65]21
F15 f 15 x = i = 1 11 a i x 1 ( b i 2 + b i x 2 ) b i 2 + b i x 3 + x 4 2 [−5, 5]40.00030
F16 f 16 x = 4 x 1 2 2.1 x 1 4 + 1 3 x 1 6 + x 1 x 2 4 x 2 2 + 4 x 2 4 [−5, 5]2−1.0316
F17 f 17 x = ( x 2 5.1 4 π 2 x 1 2 + 5 π x 1 6 ) 2 + 10 ( 1 1 8 π ) cos x 1 + 10 [−5, 5]20.398
F18 f 18 x = [ 1 + ( x 1 + x 2 + 1 ) 2 ( 19 14 x 1 + 3 x 1 2 14 x 2 + 6 x 1 x 2 + 3 x 2 2 ) ] [ 30 + ( 2 x 1 3 x 2 ) 2 ( 18 32 x 1 + 12 x 1 2 + 48 x 2 36 x 1 x 2 + 27 x 2 2 ) ] [−2, 2]23
F19 f 19 x = i = 1 4 c i exp ( j = 1 3 a i j ( x i p i j ) 2 ) [1, 3]3−3.86
F20 f 20 x = i = 1 4 c i exp ( j = 1 6 a i j ( x i p i j ) 2 ) [0, 1]6−3.32
F21 f 21 x = i = 1 5 [ ( X a i ) ( X a i ) T + c i ] 1 [0, 10]4−10.1532
F22 f 22 x = i = 1 7 [ ( X a i ) ( X a i ) T + c i ] 1 [0, 10]4−10.4028
F23 f 23 x = i = 1 10 [ ( X a i ) ( X a i ) T + c i ] 1 [0, 10]4−10.5363
Table 5. Results of single-peak and multi-peak benchmark functions.
Table 5. Results of single-peak and multi-peak benchmark functions.
FunctionCriteriaIWHOWHOMAPSONGOSSAGWOCOOT
F1avg06.629 × 10−914.714 × 10−122.737 × 10−103.979 × 10−2111.122 × 10−434.529 × 10−591.321 × 10−39
std03.567 × 10−902.421 × 10−111.086 × 10−904.747 × 10−439.959 × 10−127.114 × 10−39
F2avg1.025 × 10−2912.506 × 10−509.438 × 10−84.021 × 10−44.77 × 10−1061.461 × 10−261.109 × 10−343.414 × 10−26
std01.344 × 10−491.563 × 10−73.329 × 10−32.553 × 10−1056.542 × 10−261.475 × 10−341.834 × 10−25
F3avg01.572 × 10−561.859 × 1031.0857.325 × 10−2077.957 × 10−383.953 × 10−155.870 × 10−44
std07.437 × 10−569.828 × 1023.62 × 10−104.285 × 10−371.486 × 10−143.109 × 10−43
F4avg2.494 × 10−826.154 × 10−373.457 × 1011.102 × 1014.512 × 10−1031.243 × 10−371.642 × 10−143.053 × 10−7
std01.232 × 10−368.8073.057 × 1022.311 × 10−1026.698 × 10−372.238 × 10−141.644 × 10−6
F5avg2.485 × 10−15.292 × 1015.338 × 1012.987 × 1012.810 × 1018.877 × 10−32.686 × 1012.971 × 101
std2.283 × 10−110.059 × 1024.787 × 1011.551 × 1019.521 × 10−17.998 × 10−39.602 × 10−17.874
F6avg4.562 × 10−101.595 × 10−86.424 × 10−148.325 × 10−102.5532.641 × 10−56.713 × 10−18.996 × 10−3
std1.356 × 10−94.095 × 10−81.867 × 10−131.826 × 10−95.301 × 10−11.352 × 10−53.432 × 10−15.114 × 10−3
F7avg3.703 × 10−53.498 × 10−41.234 × 10−21.982 × 10−21.079 × 10−41.077 × 10−39.155 × 10−41.098 × 10−3
std3.150 × 10−51.879 × 10−42.861 × 10−33.175 × 10−32.009 × 10−51.331 × 10−41.368 × 10−49.571 × 10−4
F8avg−1.232 × 104−9.025 × 103−1.053 × 104−2.752 × 103−7.968 × 103−8.195 × 103−6.279 × 103−7.543 × 103
std4.721 × 1027.426 × 1023.316 × 1023.766 × 1025.581 × 1022.483 × 1037.877 × 1026.622 × 102
F9avg007.6174.905 × 101001.518 × 10−14.661 × 10−13
std004.2521.301 × 101008.176 × 10−12.334 × 10−12
F10avg8.881 × 10−163.611 × 10−153.965 × 10−15.111 × 10−63.73 × 10−158.881 × 10−161.581 × 10−142.836 × 10−11
std01.760 × 10−155.369 × 10−17.288 × 10−61.421 × 10−1502.495 × 10−151.525 × 10−10
F11avg001.709 × 10−21.397 × 101003.712 × 10−33.33 × 10−17
std001.993 × 10−23.651008.541 × 10−39.132 × 10−17
F12avg2.310 × 10−111.727 × 10−22.839 × 10−22.901 × 10−11.524 × 10−13.643 × 10−63.721 × 10−22.676 × 10−2
std6.685 × 10−113.863 × 10−25.284 × 10−23.890 × 10−16.048 × 10−21.858 × 10−61.853 × 10−27.688 × 10−2
F13avg2.080 × 10−26.233 × 10−21.054 × 10−21.831 × 10−32.5795.259 × 10−36.249 × 10−15.709 × 10−2
std3.187 × 10−23.187 × 10−22.092 × 10−24.094 × 10−34.391 × 10−16.224 × 10−32.481 × 10−14.912 × 10−2
F14avg9.98 × 10−19.98 × 10−19.98 × 10−19.98 × 10−19.98 × 10−11.171 × 1011.999.98 × 10−1
std0003.047 × 10−1209.536 × 10−19.920 × 10−10
F15avg3.075 × 10−47.653 × 10−43.075 × 10−46.441 × 10−43.075 × 10−43.163 × 10−42.036 × 10−27.271 × 10−4
std1.038 × 10−82.375 × 10−43.833 × 10−43.369 × 10−44.578 × 10−46.587 × 10−42.491 × 10−85.714 × 10−5
F16avg−1.031−1.031−1.031−1.031−1.031−1.031−1.031−1.031
std01.570 × 10−16001.57 × 10−161.087 × 10−94.158 × 10−101.776 × 10−15
F17avg3.978 × 10−13.978 × 10−13.978 × 10−13.978 × 10−13.978 × 10−13.978 × 10−13.978 × 10−13.978 × 10−1
std001.153 × 10−7001.180 × 10−91.498 × 10−71.667 × 10−10
F18avg33333333
std003.140 × 10−167.021 × 10−1605.518 × 10−81.017 × 10−58.038 × 10−14
F19avg−3.862−3.862−3.862−3.862−3.862−3.862−3.862−3.862
std00004.440 × 10−163.859 × 10−73.810 × 10−34.440 × 10−16
F20avg−3.322−3.626−3.322−3.203−3.322−3.322−3.261−3.322
std05.944 × 10−2005.916 × 10−71.097 × 10−66.038 × 10−21.416 × 10−10
F21avg−1.015 × 1016.418−6.391−1.015 × 101−1.015 × 101−1.015 × 101−7.604−1.015 × 101
std03.7353.76106.21 × 10−56.568 × 10−62.5481.433 × 10−10
F22avg−1.040 × 101−7.765−1.040 × 101−7.063−1.040 × 101−1.040 × 101−1.040 × 101−1.040 × 101
std02.63703.3392.660 × 10−51.666 × 10−53.949 × 10−41.810 × 10−11
F23avg−1.053 × 101−1.053 × 101−1.053 × 101−6.671−1.053 × 101−1.053 × 101−1.053 × 101−1.053 × 101
std0003.8642.882 × 10−86.094 × 10−61.722 × 10−56.436 × 10−12
Table 6. The result of Wilcoxon rank sum test.
Table 6. The result of Wilcoxon rank sum test.
FunctionWHOMAPSONGOSSAGWOCOOT
F11.734 × 10−61.734 × 10−61.821 × 10−61.734 × 10−61.734 × 10−61.734 × 10−61.734 × 10−6
F21.734 × 10−61.734 × 10−67.691 × 10−61.734 × 10−61.734 × 10−61.734 × 10−61.734 × 10−6
F31.734 × 10−61.734 × 10−64.01 × 10−51.734 × 10−61.734 × 10−61.734 × 10−61.734 × 10−6
F41.734 × 10−61.734 × 10−61.360 × 10−51.734 × 10−61.734 × 10−61.734 × 10−61.734 × 10−6
F53.882 × 10−66.564 × 10−62.224 × 10−61.734 × 10−61.734 × 10−61.734 × 10−65.171 × 10−1
F61.639 × 10−51.356 × 10−11.734 × 10−61.734 × 10−61.734 × 10−61.734 × 10−61.734 × 10−6
F71.734 × 10−61.734 × 10−61.734 × 10−61.734 × 10−61.734 × 10−61.734 × 10−61.734 × 10−6
F81.734 × 10−62.353 × 10−61.734 × 10−61.734 × 10−64.729 × 10−61.734 × 10−61.734 × 10−6
F911.734 × 10−61.734 × 10−6112.441 × 10−41.25 × 10−2
F107.744 × 10−61.734 × 10−61.734 × 10−62.727 × 10−616.932 × 10−71.789 × 10−5
F1111.734 × 10−61.734 × 10−6112.5 × 10−12.5 × 10−1
F128.466 × 10−62.182 × 10−24.729 × 10−61.734 × 10−61.734 × 10−61.734 × 10−61.734 × 10−6
F137.035 × 10−12.623 × 10−11.025 × 10−51.734 × 10−62.585 × 10−31.734 × 10−61.020 × 10−2
F148.134 × 10−11.562 × 10−27.744 × 10−11.562 × 10−27.723 × 10−61.592 × 10−35.712 × 10−3
F151.036 × 10−34.405 × 10−15.446 × 10−28.188 × 10−52.623 × 10−11.915 × 10−12.411 × 10−4
F1611111.734 × 10−61.734 × 10−66.25 × 10−2
F171.734 × 10−61.734 × 10−61.734 × 10−62.437 × 10−47.821 × 10−61.734 × 10−61.734 × 10−6
F187.626 × 10−12.186 × 10−113.75 × 10−11.734 × 10−61.734 × 10−61.47 × 10−6
F1911111.734 × 10−61.734 × 10−65 × 10−1
F202.492 × 10−26.875 × 10−12.148 × 10−29.271 × 10−31.494 × 10−51.891 × 10−46.088 × 10−3
F211.856 × 10−21.530 × 10−41.058 × 10−34.508 × 10−25.791 × 10−24.193 × 10−25.689 × 10−2
F221.114 × 10−28.596 × 10−21.397 × 10−28.365 × 10−21.975 × 10−26.732 × 10−21.556 × 10−2
F237.275 × 10−31.879 × 10−22.456 × 10−31.504 × 10−11.359 × 10−24.4913 × 10−27.139 × 10−1
+/=/−16/4/315/2/618/3/216/4/317/3/319/0/418/0/5
Table 7. Experimental 1 parameter configurations.
Table 7. Experimental 1 parameter configurations.
ParametersValues
Area of monitoring (S)100 m × 100 m
Sensing radius (Rs)10 m
Number of sensors (N)45
Number of iterations (iteration)150
Table 8. Comparison of initial coverage and optimize coverage experimental results.
Table 8. Comparison of initial coverage and optimize coverage experimental results.
SensorsInitial Coverage RatioOptimize Coverage Ratio
N = 4579.13%97.58%
Table 9. Comparison of experimental results from various algorithms.
Table 9. Comparison of experimental results from various algorithms.
AlgorithmCoverage RatioCoverage Efficiency
PSO83.02%58.75%
BES91.32%64.63%
WOA94.62%66.96%
HHO96.48%68.28%
WHO92.06%65.15%
IWHO97.58%69.08%
Table 10. Comparison of TABC and IWHO experimental results.
Table 10. Comparison of TABC and IWHO experimental results.
MethodCoverage Ratio
TABC [34]96.07%
IWHO97.58%
Table 11. Comparison of COOTCLCO and IWHO experimental results.
Table 11. Comparison of COOTCLCO and IWHO experimental results.
MethodCoverage Ratio
COOTCLCO [35]96.99%
IWHO99.17%
Table 12. Comparison of GWO-EH and IWHO experimental results.
Table 12. Comparison of GWO-EH and IWHO experimental results.
MethodCoverage Ratio
GWO-EH [36]83.64%
IWHO97.78%
Table 13. Comparison of VFLGWO and IWHO experimental results.
Table 13. Comparison of VFLGWO and IWHO experimental results.
MethodCoverage Ratio
VFLGWO [37]94.52%
IWHO99.48%
Table 14. Experimental 2 parameter configurations.
Table 14. Experimental 2 parameter configurations.
ParametersValues
Area of monitoring (S)100 m × 100 m
Sensing radius (Rs1)12 m
Sensing radius (Rs2)10 m
Communication radius (Rs1)24 m
Communication radius (Rs2)20 m
Number of sensor sensors (N1,2)20
Number of iterations (iteration)150
Table 15. Comparison of HWSN coverage optimization experimental results.
Table 15. Comparison of HWSN coverage optimization experimental results.
SensorsInitial Coverage RatioOptimize Coverage Ratio
N1,2 = 2081.43%98.51%
Table 16. Comparison of HWSN connectivity optimization experimental results.
Table 16. Comparison of HWSN connectivity optimization experimental results.
SensorsInitial Connectivity RatioOptimize Connectivity Ratio
N1,2 = 2016.03%20.04%
Table 17. Comparison of HWSN obstacle coverage optimization experimental results.
Table 17. Comparison of HWSN obstacle coverage optimization experimental results.
SensorsInitial Coverage RatioOptimize Coverage Ratio
N1,2 = 2086.61%97.79%
Table 18. Comparison of HWSN obstacle connectivity optimization experimental results.
Table 18. Comparison of HWSN obstacle connectivity optimization experimental results.
SensorsInitial Connectivity RatioOptimize Connectivity Ratio
N1,2 = 2015.77%17.44%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zeng, C.; Qin, T.; Tan, W.; Lin, C.; Zhu, Z.; Yang, J.; Yuan, S. Coverage Optimization of Heterogeneous Wireless Sensor Network Based on Improved Wild Horse Optimizer. Biomimetics 2023, 8, 70. https://doi.org/10.3390/biomimetics8010070

AMA Style

Zeng C, Qin T, Tan W, Lin C, Zhu Z, Yang J, Yuan S. Coverage Optimization of Heterogeneous Wireless Sensor Network Based on Improved Wild Horse Optimizer. Biomimetics. 2023; 8(1):70. https://doi.org/10.3390/biomimetics8010070

Chicago/Turabian Style

Zeng, Chuijie, Tao Qin, Wei Tan, Chuan Lin, Zhaoqiang Zhu, Jing Yang, and Shangwei Yuan. 2023. "Coverage Optimization of Heterogeneous Wireless Sensor Network Based on Improved Wild Horse Optimizer" Biomimetics 8, no. 1: 70. https://doi.org/10.3390/biomimetics8010070

Article Metrics

Back to TopTop