Next Article in Journal
Negative Group Delay Metamaterials Based on Split-Ring Resonators and Their Application
Next Article in Special Issue
Diagnosis of Stator Winding and Permanent Magnet Faults of PMSM Drive Using Shallow Neural Networks
Previous Article in Journal
Multi-Objective Optimization Algorithms for a Hybrid AC/DC Microgrid Using RES: A Comprehensive Review
Previous Article in Special Issue
A Prognosis Technique Based on Improved GWO-NMPC to Improve the Trajectory Tracking Control System Reliability of Unmanned Underwater Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Performance of Electronic Current Transformer Fault Diagnosis Model: Using an Improved Whale Optimization Algorithm and RBF Neural Network

1
Jinshan Power Supply Company, State Grid Shanghai Electric Power, Shanghai 200001, China
2
College of Electrical Engineering and Control Science, Nanjing Tech University, Nanjing 210037, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(4), 1066; https://doi.org/10.3390/electronics12041066
Submission received: 19 December 2022 / Revised: 13 January 2023 / Accepted: 21 January 2023 / Published: 20 February 2023

Abstract

:
With the widely application of electronic transformers in smart grids, transformer faults have become a pressing problem. However, reliable fault diagnosis of electronic current transformers (ECT) is still an open problem due to the complexity and diversity of fault types. In order to solve this problem, this paper proposes an ECT fault diagnosis model based on radial basis function neural network (RBFNN) and optimizes the model parameters and the network size of RBFNN simultaneously via an improved whale optimization algorithm (WOA) to improve the classification accuracy and robustness of RBFNN. Since the classical WOA is easy to fall into a locally optimal performance, a hybrid multi-strategies WOA algorithm (CASAWOA) is proposed for further improvement in optimization performance. Firstly, we introduced the tent chaotic map strategy to improve the population diversity of WOA. Secondly, we introduced nonlinear convergence factor and adaptive inertia weight to enhance the exploitation ability of the WOA. Finally, on the premise of ensuring the convergence speed of the algorithm, we modified the simulated annealing mechanism in order to prevent premature convergence. The benchmark function tests show that the CASAWOA outperforms other state-of-the-art WOA algorithms in terms of convergence speed and exploration ability. Furthermore, to validate the performance of ECT fault diagnosis model based on CASAWOA-RBFNN, a comprehensive analysis of eight fault diagnosis methods is conducted based on the ECT fault samples collected from the detection circuit. The experimental results show that the CASAWOA-RBFNN achieves an accuracy of 97.77% in ECT fault diagnosis, which is 9.8% better than WOA-RBF and which shows promising engineering practicality.

1. Introduction

In recent years, electronic current transformers (ECTs) have become an important part of the smart grid,. These devices provide indispensable measurement information for intelligent substations and ensure safe, reliable, economically viable, and efficient operation of the power grid. [1]. Since transformers usually operate in harsh environments (such as substations) for long periods of time, ECTs have a high potential for failure during operation. Transformer failures may cause limited implementation of auxiliary functions of the device, which will threaten the normal operation of the substation system and even cause power supply interruption and profit loss [2]. Effective fault diagnosis can help engineers troubleshoot early faults and improve the reliability of the power system [3]. Therefore, accurate and timely ECT fault diagnosis techniques have widely attracted attention in the past decades.
At present, numerous studies have been carried out for fault diagnosis of ECTs, and these methods can be mainly divided into three categories: fault diagnosis methods based on analytical mathematical models, fault diagnosis methods based on signal processing, and knowledge-based fault diagnosis methods [4,5,6]. The methods based on analytical mathematical models compare measurement information by establishing mathematical models. However, these methods are highly dependent on how high the accuracy of the established model is and therefore have a low robustness. Signal processing-based fault diagnosis methods decompose the time-frequency features of the collected signals and utilize the feature vectors to locate the faults. Chen et al. [7] extracted the descending feature point by using Hilbert transform and used the cutoff points of the sampled data for detection and feature estimation. Xiong et al. [8] analyzed the features of electric quantities in the primary system and used wavelet transform for multi-scale modulus maxima processing. In this processing, the abrupt change moment of the signal is determined and compared to the abrupt change moments of multiple transformers to determine ECT faults. However, in the case of noisy signals in abnormal operating conditions, fault diagnosis with single logic variable combinations suffers from insufficient adaptability and low fault diagnosis accuracy. Compared with the above-mentioned methods, knowledge-based fault diagnosis methods have numerous advantages in terms of diagnostic accuracy and robustness [9]. Some knowledge-based methods (such as expert system, fuzzy theory, and fault tree analysis) have played an important part in fault diagnosis models for ECTs.
In the era of Industry 4.0, there is widespread interest in diagnosing equipment failure issues in an intelligent manner. With the increase in computer computing capacity, machine learning techniques have gradually emerged, and some well-known artificial intelligence methods (such as the support vector machine (SVM), artificial neural network (ANN) [10], adaptive encoders [11], extreme learning machines [12], the back propagation neural network (BPNN) [13], and many other algorithms) are broadly employed in transformer fault diagnosis. Li et al. [14] proposed a genetic algorithm optimized support vector machine to improve the classification performance of SVM. L. Qu et al. [15] introduced radial basis function to the SVM (RBF-SVM), which significantly improves the diagnosis accuracy and generalization ability. However, the SVM algorithm is essentially a binary classification algorithm, which is hard to effectively use to solve multi-classification problems. Moreover, the SVM based diagnosis method is still less accurate than the ANN-based model. As one of the most well-known structures of ANN, BPNNs have the advantages of strong generalization ability and parallel processing, but the parameters of BPNN are extensive, such as connection weights, the threshold, and the topology of the network. The process used to determine the optimal parameters of BPNN is time consuming, and it is easy to fall into local optima, resulting in an unsatisfied classification performance. In [16], the improved distributed parallel firefly algorithm is proposed to optimize the parameters of BPNN, which improves the diagnosis accuracy of BPNN. However, the problems of low convergence speed in the BPNN training process and low generalization ability have not been well addressed in the proposed model. Radial basis function neural network (RBFNN) is a network structure that uses a special transfer function in the hidden layer. Being different from BPNNs, RBFNNs do not need to train the global connection weights, but they do adjust only certain important weights that affect the outputs [17]. Compared with other machine learning methods, RBFNN has the advantages of strong nonlinear approximation ability, a fast convergence rate, and good generalization performance, which have potential for application in ECT fault diagnosis.
However, similar to BPNN, the network parameters (i.e., the coordinates of centers, the widths of neurons, and the output weights) of RBFNN largely affect the network performance. Determining the optimal network parameters according to the training data can improve the classification performance of RBFNNs, and this task can be expressed as an optimization problem [18]. Metaheuristic algorithms have shown its strong capacity when dealing with numerical optimization problems. Naturally inspired metaheuristic algorithms can be generally divided into two categories: evolutionary algorithms (EAs) and swarm intelligence (SI) algorithms [19]. EAs imitate the process of individual evolution, while SI algorithms imitate some behaviors of social animals (such as flying, seeking food, or collecting resources). One of the most well-known SI algorithms is particle swarm optimization (PSO). Han et al. [20] proposed an adaptive PSO to optimize the RBFNN size and parameters. The experimental results showed that the PSO-optimized RBFNN outperformed other RBFNNs in terms of solving nonlinear problems. In addition to PSO optimized neural networks, genetic algorithms (GA) [21], cuckoo searches [22], bat algorithms [23] and grey wolf algorithms [24] have shown enhanced performance over traditional methods in optimization searching, but they have also exposed that most of the algorithms still have the problems of falling into local optima caused by the stochastic nature of metaheuristic algorithms.
Heuristic-based RBFNN formulate the parameter-seeking problem as an optimization task and establish the objective function. Then, metaheuristic algorithms are used to solve the optimal network parameters. In order to find the most appropriate RBFNN parameters as well as improve ECT fault diagnosis accuracy, we proposed a ECT fault diagnosis model based on chaos adaptive simulated annealing based whale optimization algorithm (CASAWOA) optimized RBFNN (CASAWOA-RBF). The following are the primary contributions of this article:
  • We designed a detection circuit for electronic current transformers. Based on this design, we collected data through the detection points in the circuit, which can provide samples for training the RBFNN;
  • We introduced the tent chaotic map strategy to enhance the population diversity of WOA, which helps accelerate the convergence speed of the algorithm;
  • We introduced nonlinear convergence factor and adaptive inertia weight to enhance the local exploitation ability and global searching abilities of the WOA;
  • We adjusted the annealing function of the SA algorithm, which makes the annealing speed vary according to the fitness value of the accepted worse solution. This will not only help the algorithm to avoid premature convergence but also improve the convergent speed in the late evolution;
  • We proposed the CASAWOA-RBF as a tool to solve the ECT fault diagnosis problem.
The rest parts of this article are organized as follows: relevant previous studies are introduced in Section 2. We elaborate the proposed algorithm in Section 3, and we compare the optimization and fault diagnosis performances of CASAWOA with those of other WOA algorithms in Section 4. Finally, a conclusion is summarized in Section 5.

2. Materials and Methods

2.1. Introduction of Detection Circuit for Electronic Current Transformers

In order to effectively collect various types of fault samples from ECTs, we designed a detection circuit for ECT fault diagnosis by combining the fault types and the structural characteristics of ECTs. Some of the faults in electronic current transformers can be identified with a single signal, while some faults require the identification of multiple signals [7]. By installing several key detection points to detect the current and voltage parameters on the primary and secondary sides of the electronic current transformer, the system parameters of the ECT can be obtained to determine whether the transformer is operating under normal operating conditions. Figure 1 shows the principle diagram of the detection circuit of ECT.
In Figure 1, CT represents the electronic current transformer. I A , I B , I C represent the A, B, and C three phases current, respectively. Voltage U a , U c and current I a , I c represent the measured voltage and current of the metering unit “1”and “2”, respectively. The secondary side voltage of CT 1 and CT 2 is u a and u c , respectively. When the ECT is short-circuited, the impedance change is detected by applying 1 KHz signal to accumulate multiple signals. In case of the secondary short circuit in CT, the network impedance varies according to the load and the fault is determined through several detected parameters.
For the faults on the primary and secondary sides, the fault diagnosis is conducted by analyzing the data of seven detection points, including I A , I B , U a , U c , I a , I c , u a , and u c . The fault types and corresponding detection point variation ranges proposed in this paper are shown below:
(1)
CT 1 Primary side short circuit: fluctuation of u a exceeds 10%;
(2)
CT 2 Primary side short circuit: fluctuation of u c exceeds 10%;
(3)
CT 1 Short circuit (in front of the secondary side): fluctuation of I a exceeds 10%;
(4)
CT 2 Short circuit (in front of the secondary side): fluctuation of I b exceeds 10%;
(5)
CT 1 Short circuit (at the back of the secondary side): fluctuation of I a and u a exceed 10% at the same time;
(6)
CT 2 Short circuit (at the back of the secondary side): fluctuation of I b and u c exceed 10% at the same time;
(7)
CT Phase short circuit (secondary side): fluctuation of I A , I B , u a and u c more than 10% at the same time.

2.2. Introduction of RBF Neural Network

RBFNN is a three-layer feedforward network, which usually consists of an input layer, a hidden layer, and an output layer [25,26]. As an example, Figure 2 shows the structure of a multiple-input and multiple-output RBFNN. The RBFNN employs radial basis functions as the activation function of the hidden layer neurons, and the output layer is a linear combination of the hidden layer outputs, which can be expressed as y n = i = 1 j ω i n φ i ,   n [ 1 , p ] . Here, ω i n is the output weight between the i-th hidden neuron and the n-th output neuron. φ i denotes the output of the i-th hidden neuron, which is usually defined by a Gaussian radial basis function:
φ i = e X μ k / δ k 2 ,   i [ 1 ,   j ] ,   X = [ x 1 , x 2 , , x m ] T
where δ k is the neuron width of k-th hidden neuron and μ k denotes the center vector of the k-th hidden neuron.
As shown in Figure 2, in the RBFNN, the hidden layers can map the input vector X from the low dimensions to the high dimensions. Therefore, it can convert the low-dimensional linearly inseparable problem to the high-dimensional linearly separable problem, which can accelerate the learning rate and avoid being trapped into local optima.

2.3. Introduction of Whale Optimization Algorithm (WOA)

The WOA is a SI algorithm for numerical optimization with the advantages of easy implementation, few adjustment parameters, and excellent stability [27]. WOA stimulates the special predation mechanism of humpback whales, and it accomplishes the task of finding the optimal solution through three main stages: searching for prey, encircling prey, and the bubble-net attacking strategy.

2.3.1. Searching the Prey

In the searching stage, humpback whales perform a random search according to Equation (2):
X ( t + 1 ) = X r a n d A × D
D = | C × X r a n d X ( t ) |
where t is the current iteration time, A and C are coefficient vectors, and X r a n d is the random position corresponding to the whale. The coefficient vectors A and C can be obtained according to Equations (4) and (5).
A = 2 a r a
C = 2 r
where a is the linearly decreasing momentum from 2 to 0 during the iteration and r is a random vector between 0 and 1.

2.3.2. Encircling the Prey

In the encircling prey stage, the humpback whales surround the prey by selecting the optimal prey location according to Equation (6) when | A | 1 . After the algorithm determines the current optimal solution, other search individuals will keep approaching the current optimal solution and updating the next generation of candidates.
X ( t + 1 ) = X b e s t ( t ) A × D
D = | C × X b e s t ( t ) X ( t ) |
where X b e s t ( t ) is the current optimal solution.

2.3.3. Bubble-Net Attacking Strategy

The bubble-net attacking strategy simulates the special movements of whales in bubble-net foraging and establishes two mechanisms as follows:
(1)
Shrinking Encircling Mechanism:
This mechanism is accomplished by the value of the parameter a in Equation (4). Since the trend of A depends on the change of a , A is a random value between a and a . Setting an arbitrary value in [−1, 1], the new position of the whale can be redefined by the physical space distance between the original selected position and the current optimal selected position.
(2)
Logarithmic Spiral Updating Position:
This technique first needs to calculate the distance between the whale and the prey. Then, the humpback whale moves with a conical logarithmic spiral motion toward prey. This movement can be expressed as follows:
X ( t + 1 ) = D × e b l × cos ( 2 π l ) + X b e s t ( t )
where D = | X b e s t ( t ) X ( t ) | is the distance between the whale and the prey. b is a constant for defining the logarithmic spiral shapes. l is a random value in [−1, 1].
Humpback whales perform shrinking, encircling, and spiral motions in space simultaneously. As a result, the possibility of choosing two mechanisms is the same, which can be mathematically expressed according to Equation (9).
X ( t + 1 ) = { X b e s t ( t ) A × D p < 0.5 D × e b l × cos ( 2 π l ) + X b e s t ( t )    p 0.5
where p is a random number in the range [1].

3. Proposed Methods

In this section, we detail the CASAWOA method, which is used to train RBFNN. Then, the proposed CASAWOA-RBFNN is introduced with the aim of determining the optimal parameters of RBFNN to achieve better ECT fault diagnosis performance.

3.1. Chaos Adaptive Simulated Annealing Based Whale Optimization Algorithm (CASAWOA)

The WOA has shown to obtain satisfied accuracy when dealing with low-dimensional unimodal optimization tasks. However, the problems with the whale optimization algorithm (such as relying on the initial solution, not having purposeful performance in finding the optimal solution, and being likely to fall into local optima) make the algorithm ineffective in dealing with high-dimensional problems. In this section, we propose an improved WOA algorithm to improve the performance of the algorithm by chaotic mapping, nonlinear convergence factor, adaptive inertia weight, and simulated annealing mechanism.

3.1.1. Tent Chaotic Map Strategy

As demonstrated in Section 2.3, the conventional WOA is completely random when initializing populations, which will affect the convergence speed and the performance of the algorithm [28]. However, theoretical studies have shown that convergence can be greatly accelerated if the diversity of the initial solutions can be guaranteed. Therefore, we initialize the population using the tent chaotic map strategy. The mathematical expression is as follows:
x n + 1 = { x n β , 0 x n < t h r e s h o l d 1 x n 1 β , t h r e s h o l d x n 1
When t h r e s h o l d = 0.5 , the tent map achieves the most uniform distribution characteristics [28]. It should be noted that, as shown in Figure 3, there are cycles of 0.2, 0.4, 0.6, and 0.8 in the tent mapping, and there will also be unstable cycle points of 0.25, 0.5, and 0.75; all of these points will iterate to 0. Therefore, we added a mechanism to the tent chaotic mapping to avoid iterating into these unstable periodic points. The main steps of the tent chaotic map are shown in Table 1.

3.1.2. Nonlinear Convergence Factor

According to Equations (1) and (2), it is clear that the convergence factor depends mainly on the parameter r . These parameters determine whether the algorithm performs the global search or the spiral search. As a result, the convergence factor has a direct impact on the performance of the algorithm. The convergence factor of the basic WOA algorithm linearly decreases and cannot be accurately adjusted and traded off in the case of complex and nonlinear optimization problems. Thus, we propose a nonlinear convergence factor with the following updated equation:
a = 1 cos [ ( 1 t / M I t e r ) m × π ]
where t is the current iteration, M I t e r is the maximum number of cycles, and m is the nonlinear adjustment factor that controls the decreasing degree of the convergence factor a .

3.1.3. Adaptive Inertia Weight

In order to improve the global searching ability, the idea of inertia weights in the PSO algorithm is introduced in this paper. It was found in the literature [29] that the reasonable setting of inertia weights can maximize the performance of the algorithm in the optimization search. In this paper, we improve the global and local exploration performance of WOA by introducing a new nonlinear inertia weight, which is formulated as follows:
w = 2 × [ sin ( π × t 2 × M a x _ i t e r + π ) + 1 ] n
where n is a random number between 0 and 1.
Certain perturbations in the late iteration of the algorithm can make it easier for the algorithm to jump out of the local optima [29]. Equation (12) shows that at the beginning of the iteration, a large w can make it easier for global exploration, while at the later stage of the iteration of the algorithm, a smaller w can make the algorithm have better convergence and find the optimal solution. Therefore, Equation (9) is substituted with Equation (13):
X ( t + 1 ) = { X ( t + 1 ) = w X best   ( t ) A × D in   case   of   p < 0.5 X ( t + 1 ) = D × e b l × cos ( 2 π l ) + w X b e s t ( t )    in   case   of   p 0.5

3.1.4. The Improved Simulated Annealing Mechanism

The simulated annealing (SA) algorithm [30] is used to solve combinatorial optimization problems. The SA algorithm provides an effective approximate solution algorithm for NP-hard problems and overcomes the defects of trapping in local optima and the dependence on initial value in other optimization processes. In this paper, the main reason for employing the SA algorithm is to reduce the risk by offsetting the selection pressure.
One of the drawbacks of the original SA algorithms, which use fixed values of the annealing parameter and initial temperature, is that it may decrease the convergence speed. To overcome these drawbacks, we propose a modified probability of accepting worse solution which is shown in Equation (14):
P a = exp ( f ( v i ) f ( x i ) f ( v i ) T m )
where f ( v i ) , f ( x i ) are the objective function and T m is the value of current temperature. When the T m is high, P a provides the worst solution with a high probability of acceptance; this ensures that the algorithm will not be trapped in the local optima. In the later stage of the algorithm, the probability of accepting worse solutions decreases, which means the algorithm becomes greedier for finding the potentially optimal solution.
Enlighted by the Cauchy annealing schedule [31], CASAWOA decreases the temperature after accepting a worse solution as Equation (15) shows.
T m = T s 1 + λ
where T s is the initial temperature and T m is the following temperature in the next iteration. The definition of the annealing factor λ is shown in Equation (16):
λ = η f i t n e w f i t n e w + f i t c u r r e n t
where η is a positive factor between 0 and 1 and f i t n e w and f i t c u r r e n t are the fitness values of new and current solutions. Because the SA mechanism is based on the premise that the fitness of the new solution is lower than that of the current solution, we have f i t n e w < f i t c u r r e n t . Therefore, this annealing parameter λ is always between 0 and 0.5. Equation (16) makes the annealing rate vary with the quality of the solutions: when the algorithm accepts a worse solution with high fitness, its temperature attenuation is tiny, ensuring that the algorithm can accept subsequent worse solutions; if the fitness of the accepted worse solution is much lower than that of the current solution, Equation (15) will help CASAWOA to decrease the temperature sharper to achieve an accelerating convergence speed. The pseudocode for the SA process in the CASAWOA is shown in Algorithm 1.
Algorithm 1 The Simulated Annealing Search Process in the CASAWOA
1: x i j = current solution; f i t c u r r e n t = current solution fitness value
2: v i j = new solution generated; f i t n e w = new solution fitness value
3: Initialize the temperature T s .
4: Calculate the fitness value of two solutions
5: if  f i t n e w > fitness of f i t c u r r e n t  then
6:    Accept v i j
7: else
8:    Calculate the probability P a by Equation (14)
9: end if
10: if rand (0, 1) > P a  then
11:    Accept v i j
12:    Upgrade the temperature by Equation (15)
13: else
14:    Accept x i j
15: end if

3.1.5. Specific Steps of the CASAWOA

In a nutshell, the main improvement of WOA focuses on population diversity, local exploitation ability, and global optimization ability. The pseudocode of the CASAWOA is given as Algorithm 2 shows:
Algorithm 2 The Pseudocode of the CASAWOA
1: Initialize CASAWOA parameters
2: Initialize population X i ( i = 1 , 2 , , n ) according to Equation (10)
3: Calculate the fitness value of X i marked as f i t c u r r e n t
4: X b e s t = the best search agent
5: for Iteration = 1 to M I t e r  do
6:   for  X i , i = 1 , 2 , 3 , , n
7:       Update a , A , C , w , l and p
8:       if  p < 0.5  then
9:          if  | A | < 1  then
10:             Update the current solution by Equation (13) in the case of p < 0.5
11:          else ( | A | 1 )
12:             Select a random search agent
13:             Update the current solution by Equation (2)
14:          end if
15:       if  p 0.5  then
16:          Update the current solution by Equation (13) in the case of p 0.5
17:        end if
18:   end for
19:  Check if there are solutions that exceed the search space and revise them.
20:  Calculate the fitness of the new solution marked as f i t n e w
21:  Update the solution according to the SA process in Algorithm 1
22: end for
23: Output X b e s t

3.2. CASAWOA Optimized RBF Neural Network (CASAWOA-RBF)

In this section, we demonstrate the training process for optimizing the parameters and the network size in CASAWOA-RBFNN. In the initialization, CASAWOA generates the initial solution according to Equation (10). As shown in Equation (17), let x i denote the initial position of the i-th whale.
x i = [ μ i , 1 T , δ i , 1 , ω i , 1 , μ i , 2 T , δ i , 2 , ω i , 2 , , μ i , H i T , δ i , H i , ω i , H i ]
where μ i , 1 T , δ i , 1 , ω i , 1 are the center weights, the hidden layer neuron width, and the output weights, respectively. H i is the number of the hidden layer neuron (RBF neural network size) and d i m i is the dimension of the i-th whale satisfying dim i = ( 2 + i n p u t n ) H i , where i n p u t n is the number of the input variables.
In RBFNN, the fitness function of the whale represents the accuracy of the training network. Based on the comprehensive consideration of network accuracy and network size, the fitness function based on root mean square error is defined as Equation (18) shows.
f i t i ( x i ) = 1 L ( | x i | ) i = 1 L ( | x i | ) ( y i O i ) 2 + γ H i
where L ( | x i | ) represents the number of data pairs in x i . y i and O i denote the actual and desired output values of RBFNN, respectively. γ is the adjustment factor, which is generally taken as 0.03 according to experience [32].
CASAWOA approaches the position of the optimal whale according to Equation (13). The dimensions of the other whales are updated by changing the number of hidden layer neurons, as shown in Equation (19)
H i = { H i + 1     if   H b e s t H i H i 1     if   H b e s t < H i
where H b e s t is the network size (number of the hidden layer neuron) corresponding to the optimal whale individuals. Considering the problem of uniform dimensionality of whale populations, we adopt the maximum dimension criterion in Ref. [33], i.e., all whale individuals share a common virtual space marked as dim v , and if the virtual space dim v is larger than the actual space, the remaining positions of the virtual space will be initialized randomly. After the update process, the virtual space will be emptied.
From the discussion above, the main steps of the CASAWOA-RBF neural network (CASAWOA-RBF) are summarized as follows:
Step 1: Initialize the CASAWOA parameters, the maximum iteration number M I t e r , dimensionality, population size N , initial temperature T s , and nonlinear adjustment factor m .
Step 2: Generate the RBFNN as required by the ECT fault diagnosis and initialize the position of whales according to Equation (10) marked as X = [ x 1 , x 2 , , x N ] , where x i , i [ 1 , N ] consists of the center weights μ i , the hidden layer neuron width δ i and output weights ω i .
Step 3: Calculate the fitness value of each whale. The fitness function is given in Equation (17).
Step 4: Iterate through all the whale individuals in the initial solution X and update the solutions via the bubble-net attacking strategy according to Equation (13).
Step 5: Calculate the fitness of the new solution and mark them as f i t n e w .
Step 6: Update the solutions according to the SA process in Algorithm 1.
Step 7: If the number of iterations has reached the M I t e r , continue to the next step; otherwise, return to Step 3.
Step 8: Input the optimized number of hidden layer neuron H into the RBF neural network and determine the network structure.
Step 9: Feed the optimized center weights μ i , the hidden layer neuron width δ i and output weights ω i into the RBF neural network.
Step 10: Feed the test samples into the trained RBF network to classify the fault types of ECT.

4. Simulation and Results

In order to evaluate the proposed CASAWOA and the performance of the CASAWOA optimized RBF neural network for CT fault diagnosis, the simulations are conducted with MATLAB 2020b. Section “Results for benchmark functions” compares the optimization performance of the proposed CASAWOA with other WOA in five test functions and section “Performance of CT fault diagnosis based on CASAWOA optimized RBF network” compares the performance of CT fault diagnosis based on CASAWOA-RBF network with eight other improved RBF networks.

4.1. Results for Benchmark Functions

As illustrated in Section 3.1, we took several measures to improve the WOA, so we intend to test our CASAWOA with the existing WOA [27] and SAWOA [32] through numerical benchmark functions. According to the best parameters of SAWOA elaborated in [32], the parameters of WOA, SAWOA, and CASAWOA are set as shown in Table 2. All the WOA-based algorithms were tested on five well-known benchmark functions as shown in Table 3. Further, 30 independent runs are carried out in each case in our simulations.
The comparative results between WOA, SAWOA, and CASAWOA are given in Table 4, and the best results of all the algorithms are bolded. The ‘Mean’ column shows the average best values, and the ‘SD’ column contains the standard deviation of the best values. It can be seen that in all experiments, CASAWOA achieved better results than the other WOA algorithms, and the standard deviation of the results obtained by CASAWOA is also the best. This indicates that the optimization performance of CASAWOA is more stable than that of the other methods.
In order to show the convergence speed performance, we make comparisons with SAWOA and WOA in the case of D = 30 . Figure 4 shows the details of convergence speed in different functions through three WOA algorithms.
From the figures presented, the CASAWOA not only achieves better results in all the test functions, it also outperforms SAWOA and WOA in terms of convergence speed in the Griewank, Rastrigin, and Rosenbrock functions. For the Ackley function, CASAWOA has a better best-mean-value than WOA and SAWOA but has a lower convergent speed than WOA.
From the simulation results, for the five test benchmark functions, CASAWOA has a better ability to optimize functions than WOA and SAWOA, especially in the case of the Rosenbrock and Griewank functions. Since the network training problem is considered an optimization task, having a high similarity with the presented benchmark functions is important. We hope that the CASAWOA will also perform better when tackling the neural network training problem.

4.2. Performance of ECT Fault Diagnosis Based on CASAWOA Optimized RBF Network

4.2.1. Data Acquisition and Preprocessing

Faults in the primary side can be identified with a single signal, while some faults in the secondary side require identification with multiple signals. For the faults on the primary and secondary sides of the ECT, the test environment platform was established to collect seven crucial parameters as the sample data. We collected 140 groups of data under normal operation state and 140 groups of data in each of the seven fault states. Then, we randomly selected 80% of the samples in each category to train the RBF neural network and the remaining 20% of the dataset functioned as the test set. Table 5 shows some of the collected sample data.
The segmented datasets are imported into the input layer of the RBFNN for training. It should be noted that, since the proposed ECT fault diagnosis model is a neural network-based classifier, the fault types of ECT need to be recoded and imported into the network, the fault types and corresponding code are shown in Table 6.

4.2.2. Compared Methods

To verify the performance of CASAWOA optimized RBF neural network, we first compared CASAWOA-RBF with SAWOA-RBF, WOA-RBF, and RBF neural network vertically; second, a cross-sectional comparison was conducted by choosing several metaheuristics combined with RBFNN to constitute different ECT fault diagnosis models. Finally, we compared the accuracy of ECT fault diagnosis by CASAWOA-RBF with existing neural network structures.
Firstly, we conducted a longitudinal comparison of CASAWOA-RBF, SAWOA-RBF, WOA-RBF and RBF-NN. Secondly, we selected four state-of-the-art metaheuristic algorithms combined with RBFNN: the gray wolf optimization (GWO)-RBF [34], the artificial bee colony (ABC)-RBF [35], the salp swarm algorithm (SSA)-RBF [36], and the seagull optimization (SOA)-RBF [37]. We then compared these combined neural networks with our algorithm. Finally, we selected BP neural networks [13], extreme learning machines (ELM) [12], and probabilistic neural networks (PNN) [10] for further comparison with CASAWOA-RBF.

4.2.3. Results of Model Comparison

The calculation platform for the simulation is Ryzen R5 [email protected] with 16 G memory. We took the segmented dataset as the training set for RBFNN training. Since the sample data collected from the test platform are seven-dimensional, the number of input layers is 7 and the number of output layers is four. The specific parameters of each metaheuristic in the simulations are shown in Table 7, where M I t e r is the maximum iteration, N is the population size N , T 0 is the initial temperature, and L i m i t is a special threshold used to abandon the solution in ABC algorithm.
Figure 5 shows the diagnosis results of the first set of comparison experiments (the WOA-RBF, the SAWOA-RBF, and the CASAWOA-RBF). The overall accuracy of the four models is presented in Table 8. It is clear that the RBF-NN achieves the lowest classification accuracy. Compared with other WOA models based on RBF networks, CASAWOA-RBF achieves the highest accuracy of classification with 97.78%. which is higher than WOA-RBF and CASAWOA-RBF. From Figure 5 and Table 8, it can be demonstrated that the optimization of the model parameters and the network size of RBFNN by CASAWOA can substantially improve the accuracy of RBF network as a classifier.
For further validation of the effectiveness of CASAWOA-RBF on various aspects of the training set and test set, mean square error (MSE) is adopted to quantitatively evaluate three models in the training and the test process. Figure 6 shows the MSE of four models in comparison experiments. It is obvious that the dispersion of CASAWOA-RBF is the smallest compared to SAWOA-RBF, WOA-RBF, and RBF-NN. This also implies that the improved SA mechanism can improve the accuracy of finding global optimal values.
Due to the rapid development of the metaheuristic methods in the past decades, we intend to test our CASAWOA with other well-known latecomers. Based on the same datasets, we compared the CASAWOA-RBF with other RBF-NN fault diagnosis models based on intelligent algorithms in several aspects.
The visualization results of the metaheuristic-based diagnosis models and the accuracy of each diagnostic model are given in Figure 7 and Table 9, respectively. Compared with the results of the first simulation, these network models based on the novel metaheuristic algorithms perform better than the WOA-RBF fault diagnosis model in general. However, as shown in Table 10, it can still be found that, although the accuracy of WOA-RBF is only 91.52% (which is lower than ABC-RBF, GWO-RBF, SSA-RBF, and ABC-RBF), the improved CASAWOA-RBF achieves 97.78% in terms of accuracy, which is better than all of the above algorithms. This indicates that the proposed CASAWOA still outperforms the state-of-the-art SIs such as SSA, GWO, SOA, and so forth. In addition, as shown in Figure 8, the MSEs of the other four metaheuristic-based RBF networks are all larger than the CASAWOA-RBF model, which indicates that CASAWOA-RBF achieves better ECT fault diagnosis performance.
In the third set of experiments, we used existing neural network models and classical neural network models (BP, ELM, and PNN) to compare with the CASAWOA-RBF model. The average accuracy for each model and the corresponding diagnosis results are given in Table 10 and Figure 9, respectively.
From the comparison with existing neural networks for ECT fault diagnosis, BP-NN shows the lowest overall accuracy of 86.61%; PNN has a significant advantage over traditional BP-NN and ELM in the accuracy of fault diagnosis, reaching 93.3% (even higher than WOA-RBF and SOA-RBF in the first two sets of experiments). Compared with PNN, BP-NN, and ELM, the CASAWOA-RBF model still has the highest diagnosis accuracy.
In addition to the accuracy, the recall and precision of all the neural network models are shown in Table 11 and Table 12, respectively. It is clear that the CASAWOA-RBF neural network has the highest recall for six fault types and achieves 100% recall rate in the fault “ CT 1 Short circuit” and “ CT Phase short circuit”. From Table 12, the CASAWOA-RBF model still has the highest precision in most of the fault type classifications. This indicates that the fault diagnosis model based on CASAWOA-RBF can provide more accurate classification results compared with other models, which is crucial in practical applications in substations.

5. Conclusions

In this article, we proposed an improved hybrid optimization algorithm based on the SA algorithm and WOA algorithm (CASAWOA) to obtain further improvement in solving the problem of ECT fault diagnosis. The chaos tent map is used to enhance the convergence speed at the early stage of iteration. In order to further improve the convergence speed of the algorithm in the medium term of iteration, nonlinear convergence factor and adaptive inertia weight are introduced in the bubble-net attacking stage. Furthermore, we introduced the modified SA mechanism in order to prevent premature convergence.
Compared with WOA and SAWOA, CASAWOA achieves higher accuracy and convergence speed. Further simulations were conducted for evaluating the performance of ECT fault diagnosis by the CASAWOA optimized RBF neural network. The results showed an enhanced accuracy and promising engineering practicability.
However, CASAWOA has a slow convergence rate when dealing with unimodal optimization problems, which needs further improvement in the further study. In addition, our test samples are collected based on detection circuits. In practical application scenarios, we still need to consider the effect of noise on data collection. As a part of future research in this area, we will investigate the characteristics of photoelectric current transformers and improve CASAWOA to combine with probabilistic neural networks for photoelectric current transformers fault diagnosis.

Author Contributions

Conceptualization, P.Y. and T.W.; methodology, P.Y. and T.W.; writing—original draft preparation, P.Y. and L.C.; funding acquisition, P.Y. and L.C.; data curation and writing—review and editing, H.Y., C.M. and H.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by State Grid Science and Technology Project under grant No. SGSHJS00HBJS210087.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yang, S.; Bryant, A.; Mawby, P.; Xiang, D.; Ran, L.; Tavner, P. An Industry-Based Survey of Reliability in Power Electronic Converters. IEEE Trans. Ind. Appl. 2011, 47, 1441–1451. [Google Scholar] [CrossRef]
  2. Wang, X.; Li, Q.; Li, C.; Yang, R.; Su, Q. Reliability assessment of the fault diagnosis methodologies for transformers and a new diagnostic scheme based on fault info integration. IEEE Trans. Dielectr. Electr. Insul. 2013, 20, 2292–2298. [Google Scholar] [CrossRef]
  3. Li, J.; Li, G.; Hai, C.; Guo, M. Transformer Fault Diagnosis Based on Multi-Class AdaBoost Algorithm. IEEE Access 2022, 10, 1522–1532. [Google Scholar] [CrossRef]
  4. Dai, J.; Song, H.; Sheng, G.; Jiang, X. Dissolved gas analysis of insulating oil for power transformer fault diagnosis with deep belief network. IEEE Trans. Dielectr. Electr. Insul. 2017, 24, 2828–2835. [Google Scholar] [CrossRef]
  5. Liu, Y.; Han, W.; Liu, F.; Li, Q.; Zhou, N. A collaborative diagnosis of abrupt-changing fault of electronic instrument transformer based on wavelet transform and WVD. Power Syst. Prot. Control 2019, 47, 163–170. [Google Scholar]
  6. Wang, H.; Tang, K.; Xu, R.; Zhu, X.; Li, X. Research on digital substation electronic transformer gradual variability fault diagnosis method. Power Syst. Prot. Control 2012, 40, 53–58. [Google Scholar]
  7. Chen, T.; Tian, G.; Sophian, A. Feature extraction and selection for defect classification of pulsed eddy current NDT. NDT E Int. 2008, 41, 467–476. [Google Scholar] [CrossRef]
  8. Xiong, X.; He, N.; Yu, J. Diagnosis of abrupt-changing fault of electronic instrument transformer in digital substation based on wavelet transform. Power Syst. Technol. 2010, 7, 181–185. [Google Scholar]
  9. Yu, T.; Wang, H.; Chen, X.; Tan, K. Fault Diagnosis and Fault Tolerant Method of Single Stage Matrix Type Power Electronic Transformer. In Proceedings of the 2021 IEEE 1st International Power Electronics and Application Symposium (PEAS), Beijing, China, 13–15 November 2021. [Google Scholar]
  10. Mian, T.; Choudhary, A.; Fatima, F. Vibration and infrared thermography based multiple fault diagnosis of bearing using deep learning. Nondestruct. Test. Eval. 2022, 1–22. [Google Scholar] [CrossRef]
  11. Chine, W.; Mellit, A.; Lughi, V.; Malek, A.; Sulligoi, G.; Pavan, A.M. A novel fault diagnosis technique for photovoltaic systems based on artificial neural networks. Renew Energy 2016, 90, 501–512. [Google Scholar] [CrossRef]
  12. Li, S.; Wu, G.; Gao, B.; Hao, C.; Xin, D.; Yin, X. Interpretation of DGA for transformer fault diagnosis with complementary SaE-ELM and arctangent transform. IEEE Trans. Dielectr. Electr. Insul. 2016, 23, 586–595. [Google Scholar] [CrossRef]
  13. Duan, H.; Xie, H.; Lu, Y. Transformer On-line Monitoring and Fault Diagnosis System Based on DRNN and PAS. In Proceedings of the 2019 IEEE 9th International Conference on Electronics Information and Emergency Communication (ICEIEC), Beijing, China, 12–14 July 2019. [Google Scholar]
  14. Li, J.; Zhang, Q.; Wang, K.; Wang, J.; Zhou, T.; Zhang, Y. Optimal dissolved gas ratios selected by genetic algorithm for power transformer fault diagnosis based on support vector machine. IEEE Trans. Dielectr. Electr. Insul. 2016, 23, 1198–1206. [Google Scholar] [CrossRef]
  15. Qu, L.; Zhou, H.; Liu, C.; Lu, Z. Study on Multi-RBF-SVM for Transformer Fault Diagnosis. In Proceedings of the 2018 17th International Symposium on Distributed Computing and Applications for Business Engineering and Science (DCABES), Wuxi, China, 19–23 October 2018. [Google Scholar]
  16. Li, Z.; Chen, W.; Shan, J.; Yang, Z.; Cao, L. Enhanced Distributed Parallel Firefly Algorithm Based on the Taguchi Method for Transformer Fault Diagnosis. Energies 2022, 15, 3017. [Google Scholar] [CrossRef]
  17. Meng, K.; Dong, Z.; Wang, D.; Wong, K. A Self-Adaptive RBF Neural Network Classifier for Transformer Fault Analysis. IEEE Trans. Power Syst. 2010, 25, 1350–1360. [Google Scholar] [CrossRef]
  18. Singla, P.; Subbarao, K.; Junkins, J. Direction-dependent learning approach for radial basis function networks. IEEE Trans. Neural Netw. 2007, 18, 203–222. [Google Scholar] [CrossRef]
  19. Chen, H.; Xu, Y.; Wang, M. A balanced whale optimization algorithm for constrained engineering design problems. Appl. Math. Model. 2019, 71, 45–59. [Google Scholar] [CrossRef]
  20. Lee, J.H.; Kim, J.W.; Song, J.Y.; Kim, Y.J.; Jung, S.Y. A novel memetic algorithm using modified particle swarm optimization and mesh adaptive direct search for PMSM design. IEEE Trans. Magn. 2015, 52, 1–4. [Google Scholar] [CrossRef]
  21. Choi, K.; Jang, D.H.; Kang, S.I.; Lee, J.H.; Chung, T.K.; Kim, H.S. Hybrid algorithm combing genetic algorithm with evolution strategy for antenna design. IEEE Trans. Magn. 2015, 52, 1–4. [Google Scholar] [CrossRef]
  22. Li, X.; Ma, S.; Yang, G. Synthesis of difference patterns for monopulse antennas by an improved cuckoo search algorithm. IEEE Antennas Wirel. Propag. Lett. 2016, 16, 141–144. [Google Scholar] [CrossRef]
  23. Senthilnath, J.; Kulkarni, S.; Benediktsson, J.A.; Yang, X.S. A novel approach for multispectral satellite image classification based on the bat algorithm. IEEE Geosci. Remote Sens. Lett. 2016, 13, 599–603. [Google Scholar] [CrossRef] [Green Version]
  24. Zhou, Y.; Yang, X.; Tao, L.; Yang, L. Transformer Fault Diagnosis Model Based on Improved Gray Wolf Optimizer and Probabilistic Neural Network. Energies 2021, 14, 3029. [Google Scholar] [CrossRef]
  25. Liu, X.; He, H. Fault Diagnosis for TE Process Using RBF Neural Network. IEEE Access 2021, 9, 118453–118460. [Google Scholar] [CrossRef]
  26. Li, Y.; Qiang, S.; Zhuang, X.; Kaynak, O. Robust and adaptive backstepping control for nonlinear systems using RBF neural networks. IEEE Trans. Neural Netw. 2004, 15, 693–701. [Google Scholar] [CrossRef] [PubMed]
  27. Mirjalili, S.; Lewis, A. The Whale Optimization Algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  28. Yang, D.; Li, G.; Cheng, G. On the efficiency of chaos optimization algorithms for global optimization. Chaos Solitons Fractals 2007, 34, 1366–1375. [Google Scholar] [CrossRef]
  29. Sun, H.; Xiao, H.; Hu, Q. Analysis for load distribution of tandem cold rolling based on constrained multi-objective evolutionary algorithm. China Mech. Eng. 2017, 28, 93–100. [Google Scholar]
  30. Kirpatrick, S.; Gelatt, C.D.; Vecchi, M.P. Optimization by simulated annealing. Science 1987, 220, 606–615. [Google Scholar]
  31. Lee, C.Y. Fast simulated annealing with a multivariate Cauchy distribution and the configuration’s initial temperature. J. Korean Phys. Soc. 2015, 66, 1457–1466. [Google Scholar] [CrossRef]
  32. Mafarja, M.M.; Mirjalili, S. Hybrid Whale Optimization Algorithm with simulated annealing for feature selection. Neurocomputing 2017, 260, 302–312. [Google Scholar] [CrossRef]
  33. Han, H.; Zhang, L.; Hou, Y.; Qiao, F. Nonlinear Model Predictive Control Based on a Self-Organizing Recurrent Neural Network. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 402–415. [Google Scholar] [CrossRef]
  34. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef] [Green Version]
  35. Kay, B.; Karaboga, D. A modified Artificial Bee Colony algorithm for real-parameter optimization. Inf. Sci. 2012, 192, 120–142. [Google Scholar]
  36. Mirjalili, S.; Gandomi, A.H.; Mirjalili, S.Z.; Saremi, S.; Faris, H.; Mirjalili, S.M. Salp Swarm Algorithm: A bio-inspired optimizer for engineering design problems. Adv. Eng. Softw. 2017, 114, 163–191. [Google Scholar] [CrossRef]
  37. Dhiman, G.; Kumar, V. Seagull optimization algorithm: Theory and its applications for large-scale industrial engineering problems. Knowl. Based Syst. 2019, 165, 169–196. [Google Scholar] [CrossRef]
Figure 1. Electronic current transformer fault detection principle diagram.
Figure 1. Electronic current transformer fault detection principle diagram.
Electronics 12 01066 g001
Figure 2. The structure of RBF neural network.
Figure 2. The structure of RBF neural network.
Electronics 12 01066 g002
Figure 3. Tent chaotic map distribution.
Figure 3. Tent chaotic map distribution.
Electronics 12 01066 g003
Figure 4. CASAWOA convergence speed for benchmarks (a) comparative convergence speed of the Sphere function; (b) comparative convergence speed of the Rosenbrock function; (c) comparative convergence speed of the Rastrigin function; (d) comparative convergence speed of the Ackley function; and (e) comparative convergence speed of the Griewank function.
Figure 4. CASAWOA convergence speed for benchmarks (a) comparative convergence speed of the Sphere function; (b) comparative convergence speed of the Rosenbrock function; (c) comparative convergence speed of the Rastrigin function; (d) comparative convergence speed of the Ackley function; and (e) comparative convergence speed of the Griewank function.
Electronics 12 01066 g004
Figure 5. The results of fault diagnosis of the three WOA based RBF network: (a) RBF-NN; (b) WOA-RBF; (c) SAWOA-RBF; (d) CASAWOA-RBF.
Figure 5. The results of fault diagnosis of the three WOA based RBF network: (a) RBF-NN; (b) WOA-RBF; (c) SAWOA-RBF; (d) CASAWOA-RBF.
Electronics 12 01066 g005
Figure 6. The MSE of the four RBF neural networks in comparison experiments.
Figure 6. The MSE of the four RBF neural networks in comparison experiments.
Electronics 12 01066 g006
Figure 7. The results of fault diagnosis of the three WOA based RBF networks: (a) GWO-RBF; (b)ABC-RBF; (c) SSA-RBF; (d) SOA-RBF.
Figure 7. The results of fault diagnosis of the three WOA based RBF networks: (a) GWO-RBF; (b)ABC-RBF; (c) SSA-RBF; (d) SOA-RBF.
Electronics 12 01066 g007
Figure 8. The MSE of the metaheuristic-based RBF network in comparison experiments.
Figure 8. The MSE of the metaheuristic-based RBF network in comparison experiments.
Electronics 12 01066 g008
Figure 9. The results of fault diagnosis of the existing neural network: (a) BP-NN; (b) ELM; (c) PNN; (d) CAWAWOA-RBF.
Figure 9. The results of fault diagnosis of the existing neural network: (a) BP-NN; (b) ELM; (c) PNN; (d) CAWAWOA-RBF.
Electronics 12 01066 g009
Table 1. Main steps of tent chaotic map.
Table 1. Main steps of tent chaotic map.
Main StepsDescription
Step 1Randomly generate the initial values x 0 , and record them in the array y ( 1 ) = x 0 ,   i = j = 1
Step 2Generating the sequences of x iteratively according to Equation (10).
Step 3Check if the termination condition is reached. If yes, go to Step 5; otherwise: if x i = { 0 , 0.25 , 0.5 , 0.75 } or x i = x i k ,   k = { 0 , 1 , 2 , 3 , 4 } , go to Step 4, otherwise go back to Step 2.
Step 4Change the initial value x ( i ) = y ( j + 1 ) = y ( j ) + m , j = j + 1, where m is the random value.
Step 5Stop.
Table 2. Main parameters of WOA-based algorithms.
Table 2. Main parameters of WOA-based algorithms.
AlgorithmsPopulation SizeMaximum Iteration NumberInitial TemperatureηNonlinear Adjustment FactorAnnealing Factor
WOA302000---
SAWOA302000100--0.93
CASAWOA3020001000.11.6
WOA: Whale optimization algorithm; SAWOA: Hybrid whale optimization algorithm with simulated annealing; CASAWOA: Chaos adaptive whale optimization algorithm with simulated annealing.
Table 3. Benchmark functions.
Table 3. Benchmark functions.
IDIDDimensionRangeOptimum
F1 f 1 ( x ) = i = 1 n x i 2 30[−100, 100]0
F2 f 3 ( x ) = i = 1 D 1 100 ( x i + 1 x i 2 ) 2 + ( x i 1 ) 2 30[−15, 15]0
F3 f 4 ( x ) = i = 1 n [ x i 2 10 cos ( 2 π x i ) + 10 ] 30[−5.1, 5.1]0
F4 f 5 ( x ) = 20 exp ( 0.2 1 n i = 1 n x i 2 ) exp ( 1 n i = 1 n cos ( 2 π x i ) ) + 20 + e 30[−32, 32]0
F5 f 6 ( x ) = 1 4000 i = 1 n x i 2 i = 1 n cos ( x i i ) + 1 30[−600, 600]0
F1: Sphere; F2: Rosenbrock; F3: Rastrigin; F4: Ackley; F5: Griewank.
Table 4. The results of WOA, SAWOA, and CASAWOA in the benchmark functions.
Table 4. The results of WOA, SAWOA, and CASAWOA in the benchmark functions.
IDWOASAWOACASAWOA
MeanSDMeanSDMeanSD
F11.62 × 10−179.66 × 10−189.29 × 10−185.27 × 10−1800 × 10+0000 × 10+00
F20.03558780.03813290.02578020.02886295.21 × 10−61.30 × 10−5
F300 × 10+0000 × 10+0000 × 10+0000 × 10+0000 × 10+0000 × 10+00
F43.84 × 10−151.34 × 10−151.24 × 10−151.09 × 10−158.88 × 10−1600 × 10+00
F51.1 × 10−136.7 × 10−138.88 × 10−171.72 × 10−1600 × 10+0000 × 10+00
Table 5. Selected sample data from the test platform.
Table 5. Selected sample data from the test platform.
Index I A /A I a /mA I c /mA U a /V U c /V u a /v u c /V Fault Type
11.06498.2399.98173.1173.11.4141.413 CT 1 Primary side short circuit
21.031121.71100.03173.1173.10.4261.414 CT 2 Primary side short circuit
31.05498.35124.34173.1173.11.4130.332 CT 1 Short circuit (in front of the secondary side)
41.09262.32100.0173.2173.21.4111.411 CT 2 Short circuit (in front of the secondary side)
51.00299.3151.01173.1173.11.4131.413 CT 1 Short circuit (at the back of the secondary side)
61.03212.1699.86173.2173.10.0181.414 CT 2 Short circuit (at the back of the secondary side)
71.066100.134.316173.2173.11.4130.054 CT Phase short circuit (Secondary side)
81.12650.0650.4173.1173.20.7070.708Normal
Table 6. The types and corresponding code format of the electronic current transformer status.
Table 6. The types and corresponding code format of the electronic current transformer status.
IndexFault TypeFault Code
F1Normal0000
F2 CT 1 Primary side short circuit0001
F3 CT 2 Primary side short circuit0010
F4 CT 1 Short circuit (in front of secondary side)0011
F5 CT 2 Short circuit (in front of secondary side)0100
F6 CT 1 Short circuit (at the back of the secondary side)0101
F7 CT 2 Short circuit (at the back of the secondary side)0110
F8 CT Phase short circuit (Secondary side)0111
Table 7. Detailed parameters of each model.
Table 7. Detailed parameters of each model.
MethodsParameters Settings
WOA-RBF N = 10 ,   M I t e r = 100
SAWOA-RBF N = 10 ,   M I t e r = 100 ,   T 0 = 100 , annealing factor = 0.93
CASAWOA-RBF N = 10 ,   M I t e r = 100 ,   T 0 = 100 , factor η = 0.1
nonlinear adjustment factor = 1.6
GWO-RBF N = 10 ,   M I t e r = 100
ABC-RBF N = 10 ,   M I t e r = 100 ,   L i m i t = 5
SSA-RBF N = 10 ,   M I t e r = 100
SOA-RBF N = 10 ,   M I t e r = 100
ELMNumber of hidden neuron = 20
PNNSmoothing factor = 0.06
Table 8. The accuracy of WOA-based models and RBF-NN model.
Table 8. The accuracy of WOA-based models and RBF-NN model.
MethodsAccuracy (%)
RBF-NN89.29%
WOA-RBF91.52%
SAWOA-RBF95.09%
CASAWOA-RBF97.78%
Table 9. The accuracy of each metaheuristic-based model.
Table 9. The accuracy of each metaheuristic-based model.
MethodsAccuracy (%)
WOA-RBF91.52%
SAWOA-RBF95.09%
GWO-RBF95.54%
ABC-RBF92.86%
SSA-RBF95.09%
SOA-RBF93.30%
CASAWOA-RBF97.78%
Table 10. The accuracy of each ANN-Based model.
Table 10. The accuracy of each ANN-Based model.
MethodsAccuracy (%)
BP-NN86.61%
ELM91.07%
PNN93.30%
CASAWOA-RBF97.78%
Table 11. Comparison of the recall for each model.
Table 11. Comparison of the recall for each model.
ECT FaultsRBF-NNWOA-RBFSAWOA-RBFGWO-RBFABC-RBFSSA-RBFSOA-RBFBPNNELMPNNCASAWOA-RBF
F189.29%92.86%92.86%96.43%89.29%96.43%92.86%89.29%92.86%92.86%100%
F289.29%89.29%96.43%96.43%92.86%92.86%89.29%89.29%92.86%92.86%96.43%
F389.29%89.29%96.43%92.86%96.43%96.43%92.86%85.71%85.71%92.86%96.43%
F492.86%92.86%96.43%89.29%92.86%96.43%96.43%85.71%89.29%92.86%100%
F589.29%92.86%96.43%92.86%96.43%96.43%92.86%78.57%89.29%92.86%96.43%
F689.29%89.29%96.43%100%89.29%92.86%89.29%89.29%92.86%92.86%96.43%
F785.71%92.86%89.29%100%92.86%92.86%92.43%85.71%92.86%96.43%96.43%
F889.29%92.86%96.43%96.43%92.86%92.86%92.86%89.29%92.86%92.86%100%
Table 12. Comparison of the precision for each model.
Table 12. Comparison of the precision for each model.
ECT FaultsRBF-NNWOA-RBFSAWOA-RBFGWO-RBFABC-RBFSSA-RBFSOA-RBFBPNNELMPNNCASAWOA-RBF
F1100.00%89.66%96.30%96.43%96.15%93.10%96.30%80.65%89.66%92.86%96.55%
F283.33%83.33%84.38%90.00%83.87%96.30%89.29%83.33%83.87%86.67%96.43%
F392.59%89.29%96.43%92.86%90.00%93.10%92.86%82.76%92.31%96.30%100%
F481.25%89.66%93.10%96.15%92.86%96.43%93.10%88.89%92.59%89.66%96.55%
F592.59%100.00%96.43%100.00%96.43%96.43%96.30%88.00%92.59%92.86%100%
F686.21%92.59%100.00%93.33%100.00%89.66%89.29%92.59%100.00%100.00%93.10%
F788.89%92.86%96.15%96.55%92.86%96.43%93.10%85.71%86.67%93.10%100%
F892.59%96.30%100.00%100.00%92.86%100.00%96.43%92.59%92.86%96.30%100%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, P.; Wang, T.; Yang, H.; Meng, C.; Zhang, H.; Cheng, L. The Performance of Electronic Current Transformer Fault Diagnosis Model: Using an Improved Whale Optimization Algorithm and RBF Neural Network. Electronics 2023, 12, 1066. https://doi.org/10.3390/electronics12041066

AMA Style

Yang P, Wang T, Yang H, Meng C, Zhang H, Cheng L. The Performance of Electronic Current Transformer Fault Diagnosis Model: Using an Improved Whale Optimization Algorithm and RBF Neural Network. Electronics. 2023; 12(4):1066. https://doi.org/10.3390/electronics12041066

Chicago/Turabian Style

Yang, Pengju, Taoyun Wang, Heng Yang, Chuipan Meng, Hao Zhang, and Li Cheng. 2023. "The Performance of Electronic Current Transformer Fault Diagnosis Model: Using an Improved Whale Optimization Algorithm and RBF Neural Network" Electronics 12, no. 4: 1066. https://doi.org/10.3390/electronics12041066

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop