Next Article in Journal
Investigation and Analysis of Acoustojets by Spectral Element Method
Next Article in Special Issue
Classification of Monkeypox Images Based on Transfer Learning and the Al-Biruni Earth Radius Optimization Algorithm
Previous Article in Journal
Modeling and Optimization of Triple Diode Model of Dye-Sensitized Solar Panel Using Heterogeneous Marine Predators Algorithm
Previous Article in Special Issue
The Development of PSO-ANN and BOA-ANN Models for Predicting Matric Suction in Expansive Clay Soil
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature Selection and Classification of Transformer Faults Based on Novel Meta-Heuristic Algorithm

by
El-Sayed M. El-kenawy
1,
Fahad Albalawi
2,
Sayed A. Ward
3,4,
Sherif S. M. Ghoneim
2,
Marwa M. Eid
5,
Abdelaziz A. Abdelhamid
6,7,
Nadjem Bailek
8 and
Abdelhameed Ibrahim
9,*
1
Department of Communications and Electronics, Delta Higher Institute of Engineering and Technology, Mansoura 35111, Egypt
2
Electrical Engineering Department, College of Engineering, Taif University, P.O. Box 11099, Taif 21944, Saudi Arabia
3
Electrical Engineering Department, Shoubra Faculty of Engineering, Benha University, 108 Shoubra St., Cairo 11629, Egypt
4
Faculty of Engineering, Delta University for Science and Technology, Mansoura 11152, Egypt
5
Faculty of Artificial Intelligence, Delta University for Science and Technology, Mansoura 11152, Egypt
6
Department of Computer Science, College of Computing and Information Technology, Shaqra University, Shaqra 11961, Saudi Arabia
7
Department of Computer Science, Faculty of Computer and Information Sciences, Ain Shams University, Cairo 11566, Egypt
8
Energies and Materials Research Laboratory, Faculty of Sciences and Technology, University of Tamanghasset, Tamanrasset 11001, Algeria
9
Computer Engineering and Control Systems Department, Faculty of Engineering, Mansoura University, Mansoura 35516, Egypt
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(17), 3144; https://doi.org/10.3390/math10173144
Submission received: 29 June 2022 / Revised: 21 August 2022 / Accepted: 29 August 2022 / Published: 1 September 2022

Abstract

:
Detecting transformer faults is critical to avoid the undesirable loss of transformers from service and ensure utility service continuity. Transformer faults diagnosis can be determined based on dissolved gas analysis (DGA). The DGA traditional techniques, such as Duval triangle, Key gas, Rogers’ ratio, Dornenburg, and IEC code 60599, suffer from poor transformer faults diagnosis. Therefore, recent research has been developed to diagnose transformer fault and the diagnostic accuracy using combined traditional methods of DGA with artificial intelligence and optimization methods. This paper used a novel meta-heuristic technique, based on Gravitational Search and Dipper Throated Optimization Algorithms (GSDTO), to enhance the transformer faults’ diagnostic accuracy, which was considered a novelty in this work to reduce the misinterpretation of the transformer faults. The robustness of the constructed GSDTO-based model was addressed by the statistical study using Wilcoxon’s rank-sum and ANOVA tests. The results revealed that the constructed model enhanced the diagnostic accuracy up to 98.26% for all test cases.

1. Introduction

The power transformer considers one of the most vital elements in the electrical power system, as the wrong or repeated disconnection leads to the loss of a lot of profits for the electricity companies [1,2]. The undesirable outage of the transformer from the electrical power system occurs as a result of the exposure of the insulation system, which is the insulation oil and insulation paper, to various stresses, whether electrical, thermal, or mechanical. These stresses lead to insulation damage and rapid deterioration, which requires early prediction of the deterioration of the insulation condition inside the transformer. Dissolved gas analysis (DGA) is a common technique that was used for detecting and exploring the transformer faults based on the dissolved gases concentrations [3,4,5].
Several traditional DGA techniques, such as Key gases, Dornenburg, Rogers’ ratios, IEC code, and Duval triangle, contribute to interpreting the cause of transformer faults [6,7,8,9,10]. The DGA methods above were developed to diagnose the transformer faults based on the combustible gases’ ratios [4]. The combustible gases are Hydrogen ( H 2 ), Methane ( C H 4 ), Ethan ( C 2 H 6 ), Ethylene ( C 2 H 4 ), and Acetylene ( C 2 H 2 ). Some of these methods used four ratios between the five gases, such as Dornenburg and Rogers’ four ratios and the other methods used three ratios, such as IEC60599 code [11,12,13]. The Duval triangle method is one of the most common DGA techniques which has been used until now for diagnosing transformer faults, which is based the diagnose technique on the three gases (Methane ( C H 4 ), Ethylene ( C 2 H 4 ), and Acetylene ( C 2 H 2 )) [8,9]. The limitation of the traditional DGA techniques is the failure to interpret the transformer faults when the cases are out of the code. New graphical DGA techniques were developed to enhance the diagnostic accuracy of the transformer faults, such as Pentagon [14,15,16] and heptagon [3].
Researchers have recently merged artificial intelligence with the traditional DGA techniques to enhance its diagnostic accuracy. Artificial Neural Networks (ANNs) were utilized to build a smart system to enhance the diagnostic accuracy by comparing the output of five traditional DGA techniques to identify the dominant transformer faults. In addition, the ANN was used in research for the same purpose [17,18,19,20,21]. Fuzzy logic was also merged with traditional DGA techniques to enhance diagnostic accuracy, and several publications were reported [22,23]. A neuro-fuzzy was developed in [24] to enhance the transformer fault diagnostic accuracy. Support vector machine (SVM) and other classifiers were also used based on the feature selection for each fault to diagnose the transformer faults [25,26].
Several kinds of research addressed the utilization of optimization techniques with DGA to adapt the ratio limits of the traditional IEC Code and Rogers’ ratio methods to enhance the diagnostic accuracy of the transformer faults. The Particle Swarm Optimization (PSO) algorithm in [27] was used to modify the limits of the ratios among the gases to overcome the diagnostic failure of the transformer faults. It is also addressed in several works, as in [28,29] that used PSO with the DGA techniques to enhance the transformer faults’ diagnostic accuracy. Some other optimization algorithms were merged with the DGA to increase the fault diagnostic accuracy, such as Teaching learning-based optimization [30], adaptive dynamic meta-heuristics [2], and Grey Wolf Optimization [12].
This paper uses a novel Meta-heuristic algorithm, based on the Gravitational Search and Dipper Throated Optimization Algorithms (GSDTO), to enhance the diagnostic accuracy of the transformer faults’ misinterpretation with the other techniques, whether the traditional DGA or that merged the AI with the traditional DGA methods. As the Dipper Throated Optimization (DTO) algorithm relies on numerous variables in the optimization process, its performance degrades. In addition, the convergence of the algorithm is premature. However, the satisfactory balance between exploration and exploitation is a considerable advantage. To take advantage of this benefit, the suggested approach employs the DTO algorithm. Gravitational Search Algorithm (GSA), despite its simplicity and outstanding balance between exploration and exploitation, has disadvantages such as a low exploration rate and performance decrease when a large number of local optimum solutions exist. Utilizing the dipper throated optimizer, this study uses this algorithm to take advantage of its benefits while compensating for its limitations.
The proposed model uses the raw gases’ concentration regardless of the gases’ ratios considered in the conventional DGA techniques, which resulted in a wrong diagnosis of the fault due to the out-of-code cases. ANOVA statistical analysis is applied to investigate the stability of the proposed model with the uncertainty that occurred in the input data. The results and the statistical analysis indicated the robustness of the constructed algorithm and enhanced the diagnostic accuracy of the transformer faults for the testing samples. The dataset employed in this work consists of 460 samples collected from the literature and the Egyptian Electricity Holding Company central chemical laboratory.
The proposed GSDTO enhances the Long Short Term Memory (LSTM) classification method parameters. Excellent diagnostic accuracy of the transformer faults is achieved based on the proposed GSDTO+LSTM classification model. A binary version of the proposed (GSDTO) algorithm is first used for feature selection from the tested dataset. The binary GSDTO (bGSDTO) algorithm is tested first compared to PSO [31], Grey Wolf Optimizer (GWO) [32], Whale Optimization Algorithm (WOA) [33], Biogeography-Based Optimizer (BBO) [34], Firefly Algorithm (FA) [35], Genetic Algorithm (GA) [36], and Bat Algorithm (BA) [37]. Then, a classifier based on the suggested GSDTO algorithm and LSTM method is examined for the tested dataset. Comparative analysis is performed between the GSDTO+LSTM algorithms and WOA+LSTM, GWO+LSTM, GA+LSTM, and PSO+LSTM algorithms. The GSDTO+LSTM algorithm’s diagnostic accuracy is also examined using randomly selected data from the total number of samples.
This work’s main contributions can be expressed as follows:
  • A novel Gravitational Search Dipper Throated Optimization Algorithm (GSDTO) is proposed.
  • A binary GSDTO algorithm, a binary version of the proposed algorithm, is applied for feature selection from the tested dataset.
  • A GSDTO+LSTM classifier, based on the proposed GSDTO algorithm and LSTM method, is developed to improve the tested dataset classification accuracy.
  • The GSDTO algorithm’s statistical difference is tested by Wilcoxon’s rank-sum and ANOVA tests.
  • The GSDTO algorithm is used to improve the LSTM classification method performance for classifying purposes which can be applied in a new high voltage engineering application.
  • The binary GSDTO algorithm and the LSTM-based classification algorithm can be generalized and tested for various types of datasets.
The organization of this paper is as follows. Materials and methods as presented in Section 2. Section 3 discusses the proposed GSDTO algorithm, binary GSDTO algorithm, and the GSDTO+LSTM-based model. Section 4 shows the experimental results and analysis. The validation and discussion of the proposed GSDTO-based model compared to state-of-the-art models are discussed in this section. In Section 6, the conclusion of this work and the future directions are presented.

2. Materials and Methods

2.1. Distribution of the Data

A total number of 460 samples were used in this study. A number of 386 samples for training and 74 samples for testing purposes were used to investigate the robustness of the constructed model. The collection of samples was from the central chemical laboratory of Egyptian Electricity Holding Company and literature. The dataset has several attributes including Hydrogen ( H 2 ), Methane ( C H 4 ), Ethane ( C 2 H 6 ), Ethylene ( C 2 H 4 ), Acetylene ( C 2 H 2 ), Power factor, Interfacial V, Dielectric rigidity, Water content, Health index, Life expectation, Status, Category (ACT), and Time. Table 1 shows the distribution of training data samples. The data samples in the testing process are categorized as 43 samples for Partial Discharge (PD), 69 samples for low energy discharge (D1), 115 samples for high energy discharge (D2), 81 samples for a low thermal fault (T1), 24 samples for a medium thermal fault (T2), and 54 samples for a high thermal fault (T3). Table 2 illustrates the distribution of testing samples based on the source of the data and the fault types. The data samples in the testing process are categorized as seven samples for PD, 13 samples for D1, 24 samples for D2, 16 samples for T1, 4 samples for T2, and 10 samples for T3.

2.2. Machine Learning

One type of classification and prediction model is an artificial neural network (ANN). Complex interactions between data patterns or sets of cause-and-effect variables are modeled with artificial neural networks (ANN). These techniques may include transient detection and pattern recognition. ANN is an information processing pattern composed of neurons that work together to solve problems similar to the brain. When developing an algorithmic solution and extracting the structure from existing data, a neural network is useful [48].
Random forest, a technique based on statistical learning theory, offers a number of benefits, including fewer configurable parameters, greater prediction precision, and increased generalization capacity. It takes several samples from the original sample using the bootstrap sampling method, constructs decision tree models based on each bootstrap sample, integrates multiple decision tree predictions, and determines the outcome through a voting procedure [48].
The conclusions of the prediction algorithm utilizing the k-NN technique are based on historical events similar to the current state based on a measure of distance. The k-NN output values, simple average or weighted averaging, are used to generate forecasts. Consequently, specialists can evaluate the k-NN method’s outcomes. In k-NN numerical prediction, the object’s predictable variable is the average value of its k nearest neighbors. The k-NN algorithm is a fundamental and effective tool for machine learning [48].
The LSTM model is an improved artificial neural network model that may be used to solve a variety of problems, as detailed in [49]. The primary advantage of the LSTM is its ability to retain information for an extended period of time. Figure 1 illustrates the LSTM design. The LSTM model’s initial step is to determine which data from the cell state should be disregarded. As seen in Equation (1), this is accomplished using a forget gate or sigmoid layer.
f t = σ ( b f + W f [ h t 1 , x t ] )
The new data to be stored in the cell state should be determined. The sigmoid layer determines which values require updating, and the tanh layer adds the new candidate to the produced state, as described in Equations (2) and (3).
i t = σ ( b i + W i [ h t 1 , x t ] )
C t = tanh ( b i + W i [ h t 1 , x t ] )
The old cell state, C t 1 , is updated into the new cell state, C t , based on the following equation using Equations (1)–(3).
C t = i t × C t + f t × C t 1
The cell state will then multiply the sigmoid gate output by tanh and force values between [−1, 1].
h t = o t × tanh ( C t ) , o t = σ ( b o + W o [ h t 1 , x t ] )

2.3. Dipper Throated Optimization (DTO)

Among passerines, dipper throated birds are unusual. They are excellent swimmers, divers, and hunt underwater. Additionally, their flexible and small wings enable them to fly straight and fast without glides or pauses. The algorithm based on the dipper throated birds’ (DTO) behavior assumes that birds fly and swim in search of food supplies, with  N f s representing the number of birds. Birds’ positions are marked by BP , while BV denotes their velocities. The parameters BP and BV are denoted as follows [50]:
BP = B P 1 , 1 B P 1 , 2 B P 1 , 3 B P 1 , d B P 2 , 1 B P 2 , 2 B P 2 , 3 B P 2 , d B P 3 , 1 B P 3 , 2 B P 3 , 3 B P 3 , d B P n , 1 B P n , 2 B P n , 3 B P n , d
BV = B V 1 , 1 B V 1 , 2 B V 1 , 3 B V 1 , d B V 2 , 1 B V 2 , 2 B V 2 , 3 B V 2 , d B V 3 , 1 B V 3 , 2 B V 3 , 3 B V 3 , d B V n , 1 B V n , 2 B V n , 3 B V n , d
where B P i , j denotes ith position of bird in the jth dimension and B V i , j denotes ith bird velocity in the jth dimension. The initial locations of B P i , j are distributed uniformly between the lower and higher boundaries. The fitness values f n are calculated as shown in the accompanying array.
f = f 1 ( B P 1 , 1 , B P 1 , 2 , B P 1 , 3 , , B P 1 , d ) f 2 ( B P 2 , 1 , B P 2 , 2 , B P 2 , 3 , , B P 2 , d ) f 3 ( B P 3 , 1 , B P 3 , 2 , B P 3 , 3 , , B P 3 , d ) f n ( B P n , 1 , B P n , 2 , B P n , 3 , , B P n , d )
Following that, these values are arranged ascendingly. BP b e s t is declared as the first best solution. The remainder of the answers is expected to be regular birds BP n d for follower birds. BP G b e s t is declared to be the global best solution. The term R is a random value within [ 0 , 1 ] and is employed in the DTO algorithm to switch between swimming and flying birds. If  R < 0.5 , the birds are considered to be swimming birds, and their positions are changed as
BP n d ( t + 1 ) = BP b e s t ( t ) C 1 . | C 2 . BP b e s t ( t ) BP n d ( t ) |
where BP b e s t ( t ) is the best bird’s position at iteration t and BP n d ( t ) is a bird’s position. The parameters of C 1 and C 2 are calculated as follows.
C 1 = 2 C . r 1 C , C 2 = 2 r 1 , C = 2 ( 1 t T m a x 2 )
where C changes exponentially from 2 to 0, parameter of r 1 has a random value within [ 0 , 1 ] and maximum number of iterations is denoted as T m a x .
If R 0.5 , the birds are considered to be flying birds and their positions are updated as follows:
BP n d ( t + 1 ) = BP n d ( t ) + BV ( t + 1 )
The updated bird’s velocity, denoted as BV ( t + 1 ) , is calculated as
BV ( t + 1 ) = C 3 BV ( t ) + C 4 r 2 ( BP b e s t ( t ) BP n d ( t ) ) + C 5 r 2 ( BP G b e s t BP n d ( t ) )
where C 3 is a weight value, C 4 and C 5 are constants. BP G b e s t represents global best position (best of all agents in the population) and r 2 parameter value is within [ 0 ; 1 ] . The DTO algorithm is explained step by step in Algorithm 1 [50].

2.4. Gravitational Search Algorithm (GSA)

The GSA algorithm was proposed based on Newton’s laws of motion and gravity and was applied for Solar Radiation Forecasting in [51]. The position and gravitational masses (inertial, active, and passive) represent the main attributes for each agent in this algorithm. A problem’s solution is described by these attributes and decided by a fitness function. In GSA, the agent’s position is defined as in the following equation.
x i = x i 1 , , x i d , , x i N , i = 1 , 2 , , n
where x i d indicates ith agent position in d t h dimension of N dimensions and n is number of agents (masses). The  x i d ( t ) , the position of each agent, is updated as
x i d ( t + 1 ) = x i d ( t ) + v i d ( t + 1 ) ,
where v i d ( t ) , the velocity of an agent, is calculated as
v i d ( t + 1 ) = rand i . v i d ( t ) + a i d ( t )
where rand i is a random value in [ 0 , 1 ] . The acceleration of agent i in d t h direction, denoted as a i d ( t ) , is updated using the ith agent inertial mass, denoted as M i i ( t ) , as 
a i d ( t ) = F i d ( t ) M i i ( t )
Algorithm 1 DTO Algorithm.
1:
Initialization positions of agents BP i ( i = 1 , 2 , , n ) with n agents, velocities of agents BV i ( i = 1 , 2 , , n ) , maximum iterations T m a x , objective function f n , parameters of r 1 , r 2 , C 1 , C 2 , C 3 , C 4 , C 5 , C, R, t = 1
2:
Obtain f n for each agent BP i
3:
Find best agent BP b e s t
4:
while t T m a x do
5:
   for ( i = 1 : i < n + 1 ) do
6:
     if ( R < 0.5 ) then
7:
         Update current swimming agent position as
 
            BP n d ( t + 1 ) = BP b e s t ( t ) C 1 . | C 2 . BP b e s t ( t ) BP n d ( t ) |
8:
     else
9:
         Update current flying agent velocity as
 
            BV ( t + 1 ) = C 3 BV ( t ) + C 4 r 2 ( BP b e s t ( t ) BP n d ( t ) ) + C 5 r 2 ( BP G b e s t BP n d ( t ) )
10:
       Update current flying agent position as
 
            BP n d ( t + 1 ) = BP n d ( t ) + BV ( t + 1 )
11:
     end if
12:
   end for
13:
   Obtain  f n for each agent BP i
14:
   Update  C 1 , C 2 , C, R
15:
   Find best agent BP b e s t
16:
   Set  BP G b e s t = BP b e s t
17:
   Set  t = t + 1
18:
end while
19:
Return  BP G b e s t
Total gravitational force, denoted as F i d ( t ) , is calculated as
F i d ( t ) = j = 1 , j i N rand j F i j d ( t )
where rand j is a random value in [ 0 , 1 ] . The force acting on mass i from mass j, denoted as F i j d ( t ) , is updated by the following equation.
F i j d ( t ) = G ( t ) M p i ( t ) × M a j ( t ) | | x i ( t ) , x j ( t ) | | 2 + ε ( x j d ( t ) x i d ( d ) )
where M p i represents passive gravitational mass for agent i and M a j represents active gravitational mass for agent j. The  G ( t ) parameter indicates the gravitational constant at time t and ε represents a constant value. The term | | x i ( t ) , x j ( t ) | | 2 denotes Euclidean distance between agent i and agent j.
The fitness evaluation computes gravitational and inertia masses. The values of masses are computed using the fitness map, assuming that the gravitational and inertia mass are identical. The gravitational and inertial masses are revised by using the following equations:
M p i = M a j = M i i = M i , i = 1 , 2 , , N
M i ( t ) = m i ( t ) j = 1 N m j ( t ) , m i ( t ) = fit i ( t ) w o r s t ( t ) b e s t ( t ) w o r s t ( t )
where fit i ( t ) indicates the agent i fitness value at iteration t. For a minimization problem, w o r s t ( t ) and b e s t ( t ) are calculated as in the following equations.
b e s t ( t ) = min j { 1 , , N } fit j ( t )
w o r s t ( t ) = max j { 1 , , N } fit j ( t )
The GSA algorithm is explained step by step in Algorithm 2. Since all possible solutions are used to update the position of each solution, the GSA algorithm exhibits an extremely exploratory nature. Each solution in a rotation can affect the others depending on their distances and quality. However, the precision of this approach is frequently suboptimal.
Algorithm 2 GSA Algorithm.
1: Initialization positions of agents x i ( i = 1 , 2 , , n ) with n agents, maximum iterations T m a x , objective function f n , parameters of rand i , rand j , t = 1
2: Obtain f n for each agent x i
3: Find best agent x b e s t
4: while t T m a x   do
5:       for ( i = 1 : i < n + 1 ) do
6:           Update gravitational and inertia masses by Equations (19) and (20)
7:           Update acceleration of current agent by
                a i d ( t ) = F i d ( t ) M i i ( t )
8:          Update acceleration of current agent by
              v i d ( t + 1 ) = rand i . v i d ( t ) + a i d ( t )
9:           Update position of current agent by
              x i d ( t + 1 ) = x i d ( t ) + v i d ( t + 1 )
10:       end for
11:      Obtain  f n for each agent x i
12:       Update  rand i , rand j , t = t + 1
13:       Find best agent x b e s t
14:       Set  x G b e s t = x b e s t
15: end while
16: Return x G b e s t

3. Proposed Methodology

3.1. Proposed GSDTO Algorithm

The proposed Gravitational Search Dipper Throated optimization (GSDTO) algorithm is presented in Algorithm 3 step by step. The GSDTO algorithm covers the disadvantages of the DTO and GSA algorithms and combines their advantages to achieve the best global solution. The algorithm starts with initializing the positions of predetermined n agents x i ( i = 1 , 2 , , n ) and their velocities v i ( i = 1 , 2 , , n ) . It sets the allowed iterations for the execution process as T m a x , the objective function f n , the DTO parameters of r 1 , r 2 , C 1 , C 2 , C 3 , C 4 , C 5 , C, R, and the GSA parameters of rand i and rand j . The term r a n d G S D T O is a random value within [ 0 , 1 ] .
Algorithm 3 Proposed GSDTO Algorithm.
1: Initialization positions of agents x i ( i = 1 , 2 , , n ) with n agents, velocities of agents v i ( i = 1 , 2 , , n ) , maximum number of iterations T m a x , objective function f n , parameters of r 1 , r 2 , C 1 , C 2 , C 3 , C 4 , C 5 , C, R, rand i , rand j , r a n d G S D T O , t = 1
2: Obtain f n for each agent x i
3: Find best agent x b e s t
4: Set x G b e s t = x b e s t
5: while  t T m a x do
6:    if ( r a n d G S D T O > 0.5 ) then
7:      for ( i = 1 : i < n + 1 ) do
8:         if ( R < 0.5 ) then
9:          Update current swimming agent position by
               x ( t + 1 ) = x b e s t ( t ) C 1 . | C 2 . x b e s t ( t ) x ( t ) |
10:        else
11:           Update current flying agent velocity by
                 v ( t + 1 ) = C 3 v ( t ) + C 4 r 2 ( x b e s t ( t ) x ( t ) ) + C 5 r 2 ( x G b e s t x ( t ) )
12:           Update current flying agent position by
                 x ( t + 1 ) = x ( t ) + v ( t + 1 )
13:         end if
14:      end for
15:    else
16:      for ( i = 1 : i < n + 1 ) do
17:         Update gravitational and inertia masses by Equations (29) and (30)
18:         Update acceleration of current agent by a ( t ) = F ( t ) M i i ( t )
19:         Update velocity of current agent by Equation (27)
20:         Update position of current agent by Equation (26)
21:      end for
22:    end if
23:    Obtain  f n for each agent x i
24:    Update  C 1 , C 2 , C, R, rand i , rand j , r a n d G S D T O
25:    Find best agent x b e s t
26:    Set  x G b e s t = x b e s t
27:    Set  t = t + 1
28: end while
29: Return best agent x G b e s t
If r a n d G S D T O > 0.5 , the  GSDTO algorithm starts to update the positions and velocities of agents as in the following equations. The positions will be updated considering swimming agent if R < 0.5 by
x ( t + 1 ) = x b e s t ( t ) C 1 . | C 2 . x b e s t ( t ) x ( t ) |
Otherwise, agents will be considered as flying agents and the positions is updated as
x ( t + 1 ) = x ( t ) + v ( t + 1 )
where v ( t + 1 ) , updated velocity of each agent, will be calculated as
v ( t + 1 ) = C 3 v ( t ) + C 4 r 2 ( x b e s t ( t ) x n d ( t ) ) + C 5 r 2 ( x G b e s t x ( t ) )
If r a n d G S D T O 0.5 , the GSDTO algorithm will update the positions and velocities of the agents according to the following equations. The  x ( t ) , position of each agent at iteration t, will be calculated as follows.
x ( t + 1 ) = x ( t ) + v ( t + 1 ) ,
where v ( t ) , velocity of each agent at iteration t, is changed as
v ( t + 1 ) = rand i . v ( t ) + a ( t )
where the parameter a ( t ) , acceleration of each agent at iteration t, changes as
a ( t ) = F ( t ) M i i ( t )
The values of masses are computed using the fitness map for a minimization problem, assuming that the gravitational and inertia mass are identical by the following equations:
M p i = M a j = M i i = M i , i = 1 , 2 , , N
M i ( t ) = m i ( t ) j = 1 N m j ( t )
m i ( t ) = fit i ( t ) max j { 1 , , N } fit j ( t ) min j { 1 , , N } fit j ( t ) max j { 1 , , N } fit j ( t )
where fit i ( t ) indicates the agent i fitness value at iteration t.
The GSDTO algorithm’s computational complexity in this work is expressed as follows. For iterations T m a x and n number of agents, the complexity is defined as
  • Initialize parameters of the GSDTO algorithm, x i ( i = 1 , 2 , , n ) , v i ( i = 1 , 2 , , n ) , T m a x , C 1 , C 2 , C 3 , C 4 , C 5 , r 1 , r 2 , C, R, rand i , rand j , r a n d G S D T O , and  t = 1 : O(1).
  • Calculate f n for each agent x i : O(n).
  • Obtain the best agent x b e s t : O (n).
  • Update current swimming agent position: O( T m a x × n ).
  • Update current flying agent velocity: O( T m a x × n ).
  • Update current flying agent position: O( T m a x × n ).
  • Update acceleration of current agent a i ( t ) : O( T m a x × n ).
  • Update velocity of current agent v i ( t + 1 ) : O( T m a x × n ).
  • Update position of current agent x i ( t + 1 ) : O( T m a x × n ).
  • Calculate f n for each agent x i : O( T m a x ).
  • Update C 1 , C 2 , C, R, rand i , rand j , r a n d G S D T O : O( T m a x ).
  • Obtain best agent x b e s t : O( T m a x ).
  • Set x G b e s t = x b e s t : O( T m a x ).
  • Set t = t + 1 : O( T m a x ).
  • Obtain global best agent x G b e s t : O(1)
Based on the above analysis of the GSDTO algorithm, the computation complexity is set to O( T m a x × n ) and it will be O( T m a x × n × d ) for d dimension.

3.2. Proposed Binary GSDTO Algorithm

The solutions of the GSDTO algorithm will be strictly binary, with values of 0 or 1, in case of feature selection issues. Thus, the continuous values of the proposed GSDTO algorithm will be transformed to binary [0, 1] to facilitate the feature selection process from the dataset. This study employs the following equation, which is based on the S i g m o i d function [52].
x d t + 1 = 1 if S i g m o i d ( m ) 0.5 0 o t h e r w i s e , S i g m o i d ( m ) = 1 1 + e 10 ( m 0.5 ) ,
where x d t + 1 , at iteration t and dimension d, denotes the binary solution. The  S i g m o i d function can scale the output solutions to be binary ones. If  S i g m o i d ( m ) 0.5 , this will change the value to 1; otherwise, the value will be 0. The m parameter reflects the algorithm’s selected features. Figure 2 shows how the S i g m o i d function scales the output solutions to binary [0, 1]. The binary GSDTO algorithm is described in detail in Algorithm 4. By analyzing the bGSDTO algorithm, the computation complexity is set to O( t m a x × n ) and it will be O( t m a x × n × d ) for d dimension.
Algorithm 4 Proposed Binary GSDTO Algorithm.
1:
Initialization parameters of the GSDTO
2:
Convert solution to binary [0 or 1]
3:
Obtain fitness for each agent and Find best agent
4:
while t T m a x do
5:
   if ( r a n d G S D T O > 0.5 ) then
6:
     for ( i = 1 : i < n + 1 ) do
7:
        if ( R < 0.5 ) then
8:
          Update current swimming agent position
9:
        else
10:
          Update current flying agent velocity and position
11:
        end if
12:
     end for
13:
   else
14:
     for ( i = 1 : i < n + 1 ) do
15:
        Update gravitational and inertia masses
16:
        Update acceleration of current agent
17:
        Update velocity and position of current agent
18:
     end for
19:
   end if
20:
   Convert updated solution to binary
21:
   Obtain fitness for each agent and Find best agent
22:
   Update parameters
23:
end while
24:
Return best solution

3.3. Proposed GSDTO+LSTM Based Model

Figure 3 shows the proposed model based on the presented GSDTO algorithm. The proposed GSDTO+LSTM-based model is constructed based on two main phases: feature selection and classification. The first phase of the proposed model focuses on preprocessing and feature selection. The preprocessing is interesting in cleaning and normalizing the tested dataset, including correlation analysis, feature scaling, and removing null values. Feature selection in this phase is performed after cleaning the dataset. The binary version of the proposed GSDTO algorithm is employed in this step for selecting the best number of features from the total number of attributes in the tested dataset. The validation of this phase, to confirm the quality of the bGSDTO algorithm in feature selection, is performed in the experiments using feature selection performance metrics, including average error and standard deviation.
The second phase, the intermediate phase in Figure 3, which is utilized to compare results, includes applying the base models of NN, k-NN, and RF for the selected features in phase one. The output results of the three base classifiers in phase two are recorded to be compared with the GSDTO+LSTM model classification results in the third phase. The last stage of the proposed GSDTO+LSTM-based model, the classification phase, includes optimizing the LSTM model parameters by the presented GSDTO algorithm. The classification in the third phase of the model is applied based on the selected features in phase one, and the output results are compared with the results of phase two as the final results of this model.

4. Experimental Results

This section explains the results of this study in detail. The experiments are divided into three scenarios. The first scenario discusses the feature selection ability of the proposed bGSDTO algorithm for the tested dataset, while the second scenario shows the presented algorithm’s ability for classification purposes. In the third and last scenario, the GSDTO+LSTM algorithm’s diagnostic accuracy is examined using randomly selected data from the total number of samples.

4.1. Feature Selection Scenario

The binary version of the proposed (GSDTO) algorithm is used for feature selection from the tested dataset as mentioned in Figure 3. In the first scenario, the feature selection results of this study’s presented GSDTO algorithm are discussed. Table 3 shows the GSDTO algorithm configuration of all parameters used in the experiment, while Table 4 presents the compared algorithms configuration. The binary GSDTO (bGSDTO) algorithm is tested compared to PSO [31], GWO [32], WOA [33], BBO [34], FA [35], GA [36], and BA [37].
The objective equation, f n , is used to determine the quality of a solution in the binary GSDTO algorithm. f n is employed in the following equation for a classifier’s error rate, E r r , a number of selected features, s, and a number of missing features, S, as follows.
f n = α E r r + β | s | | S |
where α [ 0 , 1 ] and β = 1 α indicate the population relevance of the given feature. The approach is satisfactory if it can provide a subset of features capable of producing a low classification error rate. K-nearest neighbor (k-NN) is a widely used, straightforward classification technique. The goodness of the chosen features is ensured in this method by using the k-NN as a classifier. The only factor used to determine classifiers is the smallest distance between the query instance and the training examples; no model for K-nearest neighbors is used in this experiment.
In Table 5, the performance metrics employed to measure the feature selection results of the proposed algorithm are presented. The variables mentioned in Table 5 are: number of runs of the optimizer, M, best solution at run j, g j * , the g j * vector size, s i z e ( g j * ) , number of features, D, and N as number points, the output label of the classifier for a point i, C i , the label of class for a point i, L i , and the M a t c h function for matching calculation between two inputs.
The suggested and compared algorithms’ feature selection results, based on 20 runs and 80 iterations for ten agents as mentioned in Table 3, are provided in Table 6. The minimum average error of (0.1969) and the standard deviation of (0.0824) show the performance of the presented bGSDTO algorithm. The second best algorithm in feature selection of the tested data, minimum average error, is bGWO with (0.2141), followed by bBBO with (0.2161) and then bGA with (0.2277). The worst algorithm in feature selection is the bBA algorithm, with an average error of (0.2575). The selected features by the proposed algorithm from the tested dataset total features are named ( H 2 ), ( C H 4 ), ( C 2 H 6 ), ( C 2 H 4 ), ( C 2 H 2 ), and ( A C T ).
The statistical analysis, using a one-way analysis of variance (ANOVA) and Wilcoxon Signed Rank tests, is conducted to determine the performance of the proposed binary GSDTO algorithm based on the average error. Wilcoxon’s test is utilized to determine the p-values comparing the suggested algorithm to other algorithms. This statistical test can assess whether or not there is a significant difference between the results of the proposed algorithm and those of other algorithms with a p-value < 0.05. To determine whether there is a statistically significant difference between the suggested algorithm and other algorithms analyzed, an ANOVA test was also conducted. The ANOVA test results for the suggested versus compared algorithms are shown in Table 7, while Table 8 compares the proposed and compared algorithms using the Wilcoxon Signed-Rank test. The statistical analysis is performed using 20 runs of the presented and compared algorithms for fair comparisons.
Figure 4 uses the convergence curve for comparing the provided and comparable algorithms in feature selection. During the population initialization of the GSDTO algorithm and compared algorithm, the number of agents is set to 10, and the number of iterations is set to 80. The distribution of agents is then set randomly for fair comparisons, as shown in Figure 4. The proposed GSDTO algorithm shows fast convergence than other algorithms. Figure 5 also depicts the average error versus objective function for the provided and compared algorithms. According to Figure 5, the proposed GSDTO algorithm shows better results than other algorithms based on average error.
The residual, QQ (quantile-quantile), homoscedasticity plots, and heat map of the presented and comparative binary methods are shown in Figure 6. As opposed to the plot of the original dataset, the potential issues can be seen in the residual values and plots. Some datasets are poor categorization candidates. A residual plot places the independent variable on the horizontal axis and the residual values on the vertical axis. If the residual values are evenly and randomly distributed across the horizontal axis, the ideal scenario is realized. When the mean and the total of the residuals are equal to zero, the residual value is determined as (Actual value-Predicted value). The residual plot is displayed in Figure 6. A residual plot’s plot patterns can be used to identify whether a model is linear or nonlinear and which one is most appropriate. The homogeneity of variance or heteroscedasticity is visually inspected together with the projected scores for the dependent variable. A scenario known as homoscedasticity occurs when the error term, also known as noise or random disturbance in the relationship between the dependent and independent variables, is constant across all values of the independent variables. The accuracy of the research findings is increased by the heteroscedasticity plot, which is depicted in Figure 6. It can immediately and readily identify any violation.
In Figure 6, the QQ plot is also displayed. An example is a probability plot. It is mostly used to compare two probability distributions by graphing the quantiles against one another. The point distributions in the QQ plot may be seen to roughly fit on the line in the figure. As a result, the relationship between the actual and anticipated residuals is linear, supporting the effectiveness of the suggested approach. The provided and compared algorithms’ heat maps are displayed in Figure 6 as a data visualization tool. A two-dimensional color scale’s intensity reveals an algorithm’s level of sophistication. With regard to how the proposed method is superior to the comparable algorithms, the color variation provides clear visual indications. These figures confirm the quality of the bGSDTO algorithm in feature selection as mentioned in Figure 3.

4.2. Classification Scenario

The experiments’ second scenario discusses the classification results of the presented GSDTO algorithm based on the LSTM classifier for 20 runs and 80 iterations using ten agents as mentioned in Table 3. The basic classifiers of NN, k-NN, and RF models and the proposed GSDTO algorithm-based LSTM method are applied to the selected features from the tested dataset in phase 1 as mentioned in Figure 3. The classification results of the proposed algorithm are compared with the WOA+LSTM, GWO+LSTM, GA+LSTM, and PSO+LSTM based models and the basic models to show the performance of the presented algorithm. There are four hyperparameters to train the LSTM model, which are fed to the proposed algorithm (GSDTO). These hyperparameters are the size of attention weights set N a , encoding length for each attention weights L e , size of champion attention weights subset W a , and number of epochs T e .
The performance metrics used in this part are the Mean Square Error (MSE) and the Area Under the ROC Curve (AUC). Table 9 presents the configuration parameters of basic classification models of NN, k-NN, and RF employed in this scenario. Table 10 shows the signal/basic classification model results of AUC and MSE for the basic classification models. The best results with the maximum AUC value of (0.797) and the minimum MSE value of (0.04887) are achieved by the RF model, followed by the NN model and then the k-NN model.
The classification results of the proposed and compared algorithms, based on optimizing the parameters of the LSTM model, are shown in Table 11. The results with the maximum AUC value of (0.9826) and the minimum MSE value of (0.00001413) are achieved by the GSDTO+LSTM model. The results show the model’s superiority compared to the best basic classifier of the RF model, shown in Table 10, and the other models based on the LSTM method in Table 11. The WOA+LSTM-based model achieves the second best classification results with AUC of (0.957) and MSE of (0.000319), followed by the PSO+LSTM-based model, GWO+LSTM-based model, and the GA+LSTM-based model achieves the worst results.
The presented GSDTO+LSTM and compared classifiers statistical descriptions and the Wilcoxon Signed-Rank test results are shown in Table 12 and Table 13, respectively, based on 20 runs and 80 iterations for 10 agents as mentioned in Table 3, for fair comparison. This statistical test shows the significant difference between the results of the proposed algorithm (GSDTO+LSTM) and those of other algorithms with a p-value < 0.05.
Figure 7 compares the presented and compared algorithms convergence curves based on the LSTM model. The proposed GSDTO+LSTM algorithm shows faster convergence than other algorithms. The box plot in Figure 8 indicates the MSE of the presented and compared algorithms-based LSTM model. This figure shows that the GSDTO+LSTM algorithm achieved minimum MSE results. Figure 9 shows the histogram of MSE for the presented and compared algorithms using the LSTM model, based on the number of values with the Bin Center range (0.0–0.00072), which confirms the stability of the proposed algorithm. Figure 8 and Figure 9 present the visualization of changing the agents impact for the tested dataset for each run. Figure 10 shows the ROC curve for the presented GSDTO with PSO and WOA algorithms based LSTM model.
The residual, QQ, homoscedasticity plots, and heat map of the presented and compared algorithms based on the LSTM model are shown in Figure 11. The residual plot places the independent variable on the horizontal axis and the residual values on the vertical axis. The residual plot is displayed in Figure 11. The homogeneity of variance or heteroscedasticity is visually inspected together with the projected scores for the dependent variable and can identify any violation. The accuracy of the research findings is increased by the heteroscedasticity plot, which is depicted in Figure 11. In Figure 11, the QQ plot is displayed, and it is used to compare two probability distributions by graphing the quantiles against one another. The point distributions in the QQ plot may be seen to roughly fit on the line in the figure. As a result, the relationship between the actual and anticipated residuals is almost linear, supporting the effectiveness of the suggested approach. The provided and compared algorithms’ heat maps are displayed in Figure 11. A two-dimensional color scale’s intensity reveals an algorithm’s level of sophistication. These figures confirm the quality of the GSDTO+LSTM model as mentioned in Figure 3.

4.3. Validation and Discussion

The diagnostic accuracy of the suggested classification algorithm is compared to that of various DGA algorithms in the literature in Table 14. As shown in Table 14, the proposed classification algorithm has an overall diagnostic accuracy of 98.26 percent, which is higher than the diagnostic accuracy of other DGA techniques. The adaptive and NPR methods, which provide 94.6 and 90.54 percent accuracy, respectively, come close to the suggested algorithm. On the other hand, other classic DGA procedures, such as the Rogers Ratio method (45.95 percent), IEC 60599 (50 percent), and the Duval triangle method, have low overall diagnostic accuracies (66.27 percent). The diagnostic accuracy results for the suggested classification method indicated that it possesses a high capacity for correctly diagnosing transformer defects.

5. Sensitivity Analysis of the GSDTO Parameters

This section explores the GSDTO’s parameter sensitivity analysis. The GSDTO consists of five parameters: the R-Parameter, the exploration percentage, the population size, the number of iterations, and the C-Parameter. These settings determine the performance of the algorithm when solving the evaluated optimization problem in this work. Any change to a parameter can influence the optimization technique. As a result, a sensitivity analysis of these parameters is conducted.

5.1. One-at-a-Time Sensitivity Analysis

The sensitivity analysis has been conducted using the One-at-a-Time (OAT) sensitivity measure [54]. OAT is recognized as one of the simplest strategies for sensitivity analysis. OAT examines the performance of an algorithm by varying a single parameter while leaving the others unchanged. Table 15 and Table 16 detail the observed variations in GSDTO’s time and fitness values as parameter values are adjusted. As shown in Table 15 and Table 16, we selected 20 distinct values within the interval of each parameter by adding 5% to its length to generate a new evaluation value. Each of these variables underwent ten runs of the algorithm, and the averages for time and fitness are displayed in the table.

5.2. Regression Analysis

It has been determined through regression analysis how the algorithm’s parameters can explain algorithm performance variations. Regression analysis is appropriate when we need to predict the value of a dependent variable (algorithm performance) based on the value of an independent variable (parameter). The results of the regression analysis for the GSDTO parameters, convergence time, and fitness are shown in Table 17. The value of R Square indicates the proportion of the total variance in time or fitness that can be explained by the parameter values. According to Table 17, the number of iterations and population size have the highest R Square value for convergence time, indicating that they can explain any variation in convergence time exceptionally well. In contrast, the exploration percentage and the R-Parameter can account for 60.3% and 78.5% of the variance in convergence time, respectively. In Table 17, significance F values less than 0.05 indicate that the regression model significantly predicts the algorithm’s performance. The GSDTO’s parameters convergence time and minimum fitness versus objective function are shown in Figure 12.

5.3. Statistical Significance

To determine whether there is a statistically significant difference between the means of the observations in Table 15 and Table 16, we conducted an ANOVA on convergence time and fitness values while adjusting the GSDTO’s parameters. The results of the GSDTO’s ANOVA test are shown in Table 18. All p-values in Table 18 are less than 0.05. There is a statistically significant difference between the means of the five groups of convergence time and the five groups of minimal fitness observed by adjusting parameter values. A T-test with one tail was run at a significance level of 0.05. Table 19 provides the T-Test results for each observed pair of convergence time and minimum fitness parameters for GSDTO. In the table, p-values less than 0.05 indicate a statistically significant difference between groups.

6. Conclusions and Future Work

This research proposed a novel meta-heuristic technique based on a dataset from real-technical systems to classify the dissolved gas analysis (DGA) for transformer faults diagnosis, which is one of the most vital elements in the electrical power system. Initially, the suggested binary (GSDTO) technique is employed to choose features from the evaluated dataset. The binary GSDTO (bGSDTO) method is evaluated against PSO, GWO, WOA, BBO, FA, GA, and BA. A classifier based on the proposed GSDTO algorithm and LSTM approach is then used on the tested dataset. Comparing the classification results with WOA+LSTM, GWO+LSTM, GA+LSTM, and PSO+LSTM. The GSDTO+LSTM algorithm’s diagnostic accuracy is also examined using randomly selected data. Based on the statistical investigation, the robustness of the built model was examined. The results demonstrated that the developed model increased the diagnostic accuracy for all test cases to 98.26%. The sensitivity analysis of the GSDTO’s parameters, R-Parameter, exploration percentage, population size, number of iterations, and C-Parameter, confirm the performance of the algorithm. In the future, the binary GSDTO method and the GSDTO+LSTM-based classification algorithm can be generalized and evaluated on various datasets. Some additional experiments will be done to evaluate the scalability, runtime, and memory for the presented GSDTO+LSTM algorithm.

Author Contributions

Conceptualization, E.-S.M.E.-k., A.I. and A.A.A.; methodology, A.I.; software, E.-S.M.E.-k.; validation, S.S.M.G., N.B. and A.A.A.; formal analysis, S.S.M.G.; investigation, M.M.E.; resources, M.M.E.; data curation, S.S.M.G.; writing—original draft preparation, A.A.A. and A.I.; writing—review and editing, S.A.W., S.S.M.G. and N.B.; visualization, E.-S.M.E.-k.; supervision, S.A.W.; project administration, F.A.; funding acquisition, F.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Taif University Researchers Supporting Project TURSP-2020/97, Taif University, Taif, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors appreciate the Taif University Researchers Supporting Project TURSP-2020/97, Taif University, Taif, Saudi Arabia, for funding this research.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Badawi, M.; Ibrahim, S.A.; Mansour, D.E.A.; El-Faraskoury, A.A.; Ward, S.A.; Mahmoud, K.; Lehtonen, M.; Darwish, M.M.F. Reliable Estimation for Health Index of Transformer Oil Based on Novel Combined Predictive Maintenance Techniques. IEEE Access 2022, 10, 25954–25972. [Google Scholar] [CrossRef]
  2. Ghoneim, S.S.M.; Farrag, T.A.; Rashed, A.A.; El-Kenawy, E.S.M.; Ibrahim, A. Adaptive Dynamic Meta-Heuristics for Feature Selection and Classification in Diagnostic Accuracy of Transformer Faults. IEEE Access 2021, 9, 78324–78340. [Google Scholar] [CrossRef]
  3. Gouda, O.E.; El-Hoshy, S.H.; El-Tamaly, H.H. Proposed heptagon graph for DGA interpretation of oil transformers. IET Gener. Transm. Distrib. 2018, 12, 490–498. [Google Scholar] [CrossRef]
  4. Ward, S.A.; El-Faraskoury, A.; Badawi, M.; Ibrahim, S.A.; Mahmoud, K.; Lehtonen, M.; Darwish, M.M.F. Towards Precise Interpretation of Oil Transformers via Novel Combined Techniques Based on DGA and Partial Discharge Sensors. Sensors 2021, 21, 2223. [Google Scholar] [CrossRef]
  5. Benmahamed, Y.; Kherif, O.; Teguar, M.; Boubakeur, A.; Ghoneim, S.S.M. Accuracy Improvement of Transformer Faults Diagnostic Based on DGA Data Using SVM-BA Classifier. Energies 2021, 14, 2970. [Google Scholar] [CrossRef]
  6. Std C57.104-2008; IEEE Guide for the Interpretation of Gases Generated in Oil-Immersed Transformers. IEEE: Piscataway, NJ, USA, 2009; pp. 1–36. [CrossRef]
  7. IEC 60599: Mineral Oil-Filled Electrical Equipment in Service—Guidance on the Interpretation of Dissolved and Free Gases Analysis, Edition 2.1; IEC: Geneva, Switzerland, 2007.
  8. Duval, M. A review of faults detectable by gas-in-oil analysis in transformers. IEEE Electr. Insul. Mag. 2002, 18, 8–17. [Google Scholar] [CrossRef]
  9. Duval, M.; dePabla, A. Interpretation of gas-in-oil analysis using new IEC publication 60599 and IEC TC 10 databases. IEEE Electr. Insul. Mag. 2001, 17, 31–41. [Google Scholar] [CrossRef]
  10. Gouda, O.E.; El-Hoshy, S.H.; Ghoneim, S.S.M. Enhancing the Diagnostic Accuracy of DGA Techniques Based on IEC-TC10 and Related Databases. IEEE Access 2021, 9, 118031–118041. [Google Scholar] [CrossRef]
  11. Faiz, J.; Soleimani, M. Assessment of computational intelligence and conventional dissolved gas analysis methods for transformer fault diagnosis. IEEE Trans. Dielectr. Electr. Insul. 2018, 25, 1798–1806. [Google Scholar] [CrossRef]
  12. Hoballah, A.; Mansour, D.E.A.; Taha, I.B.M. Hybrid Grey Wolf Optimizer for Transformer Fault Diagnosis Using Dissolved Gases Considering Uncertainty in Measurements. IEEE Access 2020, 8, 139176–139187. [Google Scholar] [CrossRef]
  13. Li, X.; Wu, H.; Wu, D. DGA Interpretation Scheme Derived From Case Study. IEEE Trans. Power Deliv. 2011, 26, 1292–1293. [Google Scholar] [CrossRef]
  14. The duval pentagon-a new complementary tool for the interpretation of dissolved gas analysis in transformers. IEEE Electr. Insul. Mag. 2014, 30, 9–12. [CrossRef]
  15. Cheim, L.; Duval, M.; Haider, S. Combined Duval Pentagons: A Simplified Approach. Energies 2020, 13, 2859. [Google Scholar] [CrossRef]
  16. Mansour, D.E.A. Development of a new graphical technique for dissolved gas analysis in power transformers based on the five combustible gases. IEEE Trans. Dielectr. Electr. Insul. 2015, 22, 2507–2512. [Google Scholar] [CrossRef]
  17. Miranda, V.; Castro, A. Improving the IEC Table for Transformer Failure Diagnosis with Knowledge Extraction from Neural Networks. IEEE Trans. Power Deliv. 2005, 20, 2509–2516. [Google Scholar] [CrossRef]
  18. Souahlia, S.; Bacha, K.; Chaari, A. MLP neural network-based decision for power transformers fault diagnosis using an improved combination of Rogers and Doernenburg ratios DGA. Int. J. Electr. Power Energy Syst. 2012, 43, 1346–1353. [Google Scholar] [CrossRef]
  19. Guardado, J.; Naredo, J.; Moreno, P.; Fuerte, C. A comparative study of neural network efficiency in power transformers diagnosis using dissolved gas analysis. IEEE Trans. Power Deliv. 2001, 16, 643–647. [Google Scholar] [CrossRef]
  20. Equbal, M.D.; Khan, S.A.; Islam, T. Transformer incipient fault diagnosis on the basis of energy-weighted DGA using an artificial neural network. Turk. J. Electr. Eng. Comput. Sci. 2018, 26, 77–88. [Google Scholar] [CrossRef]
  21. Ou, M.; Wei, H.; Zhang, Y.; Tan, J. A Dynamic Adam Based Deep Neural Network for Fault Diagnosis of Oil-Immersed Power Transformers. Energies 2019, 12, 995. [Google Scholar] [CrossRef] [Green Version]
  22. Abu-Siada, A.; Hmood, S.; Islam, S. A new fuzzy logic approach for consistent interpretation of dissolved gas-in-oil analysis. IEEE Trans. Dielectr. Electr. Insul. 2013, 20, 2343–2349. [Google Scholar] [CrossRef]
  23. Noori, M.; Effatnejad, R.; Hajihosseini, P. Using dissolved gas analysis results to detect and isolate the internal faults of power transformers by applying a fuzzy logic method. IET Gener. Transm. Distrib. 2017, 11, 2721–2729. [Google Scholar] [CrossRef]
  24. Mulyodinoto, K.U.; Prasojo, R.A.; Abu-Siada, A. Applications of ANFIS to Estimate the Degree of Polymerization Using Transformer Dissolve Gas Analysis and Oil Characteristics. Polym. Sci. 2018, 14, 1–9. [Google Scholar]
  25. Bacha, K.; Souahlia, S.; Gossa, M. Power transformer fault diagnosis based on dissolved gas analysis by support vector machine. Electr. Power Syst. Res. 2012, 83, 73–79. [Google Scholar] [CrossRef]
  26. Zhang, Y.; Li, X.; Zheng, H.; Yao, H.; Liu, J.; Zhang, C.; Peng, H.; Jiao, J. A Fault Diagnosis Model of Power Transformers Based on Dissolved Gas Analysis Features Selection and Improved Krill Herd Algorithm Optimized Support Vector Machine. IEEE Access 2019, 7, 102803–102811. [Google Scholar] [CrossRef]
  27. Taha, I.B.M.; Hoballah, A.; Ghoneim, S.S.M. Optimal ratio limits of rogers’ four-ratios and IEC 60599 code methods using particle swarm optimization fuzzy-logic approach. IEEE Trans. Dielectr. Electr. Insul. 2020, 27, 222–230. [Google Scholar] [CrossRef]
  28. Illias, H.A.; Chai, X.R.; Bakar, A.H.A.; Mokhlis, H. Transformer Incipient Fault Prediction Using Combined Artificial Neural Network and Various Particle Swarm Optimisation Techniques. PLoS ONE 2015, 10, e0129363. [Google Scholar] [CrossRef]
  29. Illias, H.A.; Chai, X.R.; Bakar, A.H.A. Hybrid modified evolutionary particle swarm optimisation-time varying acceleration coefficient-artificial neural network for power transformer fault diagnosis. Measurement 2016, 90, 94–102. [Google Scholar] [CrossRef]
  30. Ghoneim, S.S.M.; Mahmoud, K.; Lehtonen, M.; Darwish, M.M.F. Enhancing Diagnostic Accuracy of Transformer Faults Using Teaching-Learning-Based Optimization. IEEE Access 2021, 9, 30817–30832. [Google Scholar] [CrossRef]
  31. Bello, R.; Gomez, Y.; Nowe, A.; Garcia, M.M. Two-Step Particle Swarm Optimization to Solve the Feature Selection Problem. In Proceedings of the Seventh International Conference on Intelligent Systems Design and Applications (ISDA 2007), Rio de Janeiro, Brazil, 20–24 October 2007; pp. 691–696. [Google Scholar] [CrossRef]
  32. El-Kenawy, E.S.M.; Eid, M.M.; Saber, M.; Ibrahim, A. MbGWO-SFS: Modified Binary Grey Wolf Optimizer Based on Stochastic Fractal Search for Feature Selection. IEEE Access 2020, 8, 107635–107649. [Google Scholar] [CrossRef]
  33. Eid, M.M.; El-kenawy, E.S.M.; Ibrahim, A. A binary Sine Cosine-Modified Whale Optimization Algorithm for Feature Selection. In Proceedings of the 2021 National Computing Colleges Conference (NCCC), Taif, Saudi Arabia, 27–28 March 2021. [Google Scholar] [CrossRef]
  34. Simon, D. Biogeography-Based Optimization. IEEE Trans. Evol. Comput. 2008, 12, 702–713. [Google Scholar] [CrossRef]
  35. Fister, I.; Yang, X.S.; Fister, I.; Brest, J. Memetic Firefly Algorithm for Combinatorial Optimization. arXiv 2012, arXiv:1204.5165. [Google Scholar]
  36. Kabir, M.M.; Shahjahan, M.; Murase, K. A new local search based hybrid genetic algorithm for feature selection. Neurocomputing 2011, 74, 2914–2928. [Google Scholar] [CrossRef]
  37. Karakonstantis, I.; Vlachos, A. Bat algorithm applied to continuous constrained optimization problems. J. Inf. Optim. Sci. 2021, 42, 57–75. [Google Scholar] [CrossRef]
  38. Egyptian Electricity Holding Company (EEHC) Reports. Available online: http://www.moee.gov.eg/english_new/report.aspx (accessed on 18 April 2022).
  39. Agrawal, S.; Chandel, A.K. Transformer incipient fault diagnosis based on probabilistic neural network. In Proceedings of the 2012 Students Conference on Engineering and Systems, Allahabad, India, 16–18 March 2012; IEEE: Piscataway, NJ, USA, 2012. [Google Scholar] [CrossRef]
  40. Wang, M.H. A novel extension method for transformer fault diagnosis. IEEE Trans. Power Deliv. 2003, 18, 164–169. [Google Scholar] [CrossRef]
  41. Zhu, Y.; Wang, F.; Geng, L. Transformer Fault Diagnosis Based on Naive Bayesian Classifier and SVR. In Proceedings of the TENCON 2006–2006 IEEE Region 10 Conference, Hong Kong, China, 14–17 November 2006; IEEE: Piscataway, NJ, USA, 2006. [Google Scholar] [CrossRef]
  42. Siva Sarma, D.V.S.S.; Kalyani, G.N.S. ANN approach for condition monitoring of power transformers using DGA. In Proceedings of the 2004 IEEE Region 10 Conference TENCON 2004, Chiang Mai, Thailand, 24 November 2004; Volume 3, pp. 444–447. [Google Scholar] [CrossRef]
  43. Hu, J.; Zhou, L.; Song, M. Transformer Fault Diagnosis Method of Gas Hromatographic Analysis Using Computer Image Analysis. In Proceedings of the 2012 Second International Conference on Intelligent System Design and Engineering Application, Sanya, China, 6–7 January 2012; pp. 1169–1172. [Google Scholar] [CrossRef]
  44. Seifeddine, S.; Khmais, B.; Abdelkader, C. Power transformer fault diagnosis based on dissolved gas analysis by artificial neural network. In Proceedings of the 2012 First International Conference on Renewable Energies and Vehicular Technology, Nabeul, Tunisia, 26–28 March 2012; pp. 230–236. [Google Scholar] [CrossRef]
  45. Rajabimendi, M.; Dadios, E.P. A hybrid algorithm based on neural-fuzzy system for interpretation of dissolved gas analysis in power transformers. In Proceedings of the TENCON 2012 IEEE Region 10 Conference, Cebu, Philippines, 19–22 November 2012; pp. 1–6. [Google Scholar] [CrossRef]
  46. Zhang, G.; Yasuoka, K.; Ishii, S.; Yang, L.; Yan, Z. Application of fuzzy equivalent matrix for fault diagnosis of oil-immersed insulation. In Proceedings of the 1999 IEEE 13th International Conference on Dielectric Liquids (ICDL’99) (Cat. No. 99CH36213), Nara, Japan, 25 July 1999; pp. 400–403. [CrossRef]
  47. Gouda, O.E.; Saleh, S.M.; El-hoshy, S.H. Power Transformer Incipient Faults Diagnosis Based on Dissolved Gas Analysis. Indones. J. Electr. Eng. Comput. Sci. 2016, 17, 10–16. [Google Scholar] [CrossRef]
  48. El-Kenawy, E.S.M.; Mirjalili, S.; Ibrahim, A.; Alrahmawy, M.; El-Said, M.; Zaki, R.M.; Eid, M.M. Advanced Meta-Heuristics, Convolutional Neural Networks, and Feature Selectors for Efficient COVID-19 X-Ray Chest Image Classification. IEEE Access 2021, 9, 36019–36037. [Google Scholar] [CrossRef]
  49. Ibrahim, A.; Mirjalili, S.; El-Said, M.; Ghoneim, S.S.M.; Al-Harthi, M.M.; Ibrahim, T.F.; El-Kenawy, E.S.M. Wind Speed Ensemble Forecasting Based on Deep Learning Using Adaptive Dynamic Optimization Algorithm. IEEE Access 2021, 9, 125787–125804. [Google Scholar] [CrossRef]
  50. Takieldeen, A.E.; El-kenawy, E.S.M.; Hadwan, M.; Zaki, R.M. Dipper Throated Optimization Algorithm for Unconstrained Function and Feature Selection. Comput. Mater. Contin. 2022, 72, 1465–1481. [Google Scholar] [CrossRef]
  51. El-kenawy, E.S.M.; Mirjalili, S.; Ghoneim, S.S.M.; Eid, M.M.; El-Said, M.; Khan, Z.S.; Ibrahim, A. Advanced Ensemble Model for Solar Radiation Forecasting using Sine Cosine Algorithm and Newton’s Laws. IEEE Access 2021, 9, 115750–115765. [Google Scholar] [CrossRef]
  52. Emary, E.; Zawbaa, H.M.; Hassanien, A.E. Binary grey wolf optimization approaches for feature selection. Neurocomputing 2016, 172, 371–381. [Google Scholar] [CrossRef]
  53. Taha, I.B.; Dessouky, S.S.; Ghoneim, S.S. Transformer fault types and severity class prediction based on neural pattern-recognition techniques. Electr. Power Syst. Res. 2021, 191, 106899. [Google Scholar] [CrossRef]
  54. Confalonieri, R.; Bellocchi, G.; Bregaglio, S.; Donatelli, M.; Acutis, M. Comparison of sensitivity analysis techniques: A case study with the rice model WARM. Ecol. Model. 2010, 221, 1897–1906. [Google Scholar] [CrossRef]
Figure 1. Architecture of LSTM Neural Network.
Figure 1. Architecture of LSTM Neural Network.
Mathematics 10 03144 g001
Figure 2. Sigmoid function to scale the output solutions to binary [0, 1].
Figure 2. Sigmoid function to scale the output solutions to binary [0, 1].
Mathematics 10 03144 g002
Figure 3. Proposed feature selection and classification model based on the proposed GSDTO algorithm.
Figure 3. Proposed feature selection and classification model based on the proposed GSDTO algorithm.
Mathematics 10 03144 g003
Figure 4. Feature selection comparison of the presented and compared algorithms based on convergence curve.
Figure 4. Feature selection comparison of the presented and compared algorithms based on convergence curve.
Mathematics 10 03144 g004
Figure 5. The bGSDTO and compared algorithms’ average error versus objective function.
Figure 5. The bGSDTO and compared algorithms’ average error versus objective function.
Mathematics 10 03144 g005
Figure 6. Residual, QQ, homoscedasticity plots, and heat map of the bGSDTO and compared algorithms.
Figure 6. Residual, QQ, homoscedasticity plots, and heat map of the bGSDTO and compared algorithms.
Mathematics 10 03144 g006
Figure 7. Comparison of the presented and compared algorithms convergence curves based LSTM model.
Figure 7. Comparison of the presented and compared algorithms convergence curves based LSTM model.
Mathematics 10 03144 g007
Figure 8. MSE for the presented and compared algorithms based LSTM model.
Figure 8. MSE for the presented and compared algorithms based LSTM model.
Mathematics 10 03144 g008
Figure 9. Histogram for the presented and compared algorithms based LSTM model.
Figure 9. Histogram for the presented and compared algorithms based LSTM model.
Mathematics 10 03144 g009
Figure 10. ROC curve for the presented GSDTO with PSO and WOA algorithms based LSTM model.
Figure 10. ROC curve for the presented GSDTO with PSO and WOA algorithms based LSTM model.
Mathematics 10 03144 g010
Figure 11. Residual, QQ, homoscedasticity plots, and heat map of the presented and compared algorithms based on LSTM model.
Figure 11. Residual, QQ, homoscedasticity plots, and heat map of the presented and compared algorithms based on LSTM model.
Mathematics 10 03144 g011
Figure 12. The GSDTO’s parameters convergence time and fitness versus objective function.
Figure 12. The GSDTO’s parameters convergence time and fitness versus objective function.
Mathematics 10 03144 g012
Table 1. Distribution of the 386 training samples according to the fault types and the references.
Table 1. Distribution of the 386 training samples according to the fault types and the references.
Ref.PDD1D2T1T2T3Total
[8]2003005
[9]92448001899
[25]0211318
[38]274255701828240
[39]1052019
[40]30443519
[41]1121016
Total4369115812454386
Table 2. The distribution of the 74 testing samples according to the fault types and the references.
Table 2. The distribution of the 74 testing samples according to the fault types and the references.
Ref.PDD1D2T1T2T3Total
[9]1681 117
[38]6611113239
[40] 21 3
[41] 11
[42] 22
[43] 1 1
[44] 111 3
[45] 11
[46] 2136
[47] 1 1
Total713241641074
Table 3. Configuration parameters of the GSDTO algorithm.
Table 3. Configuration parameters of the GSDTO algorithm.
Parameter (s)Value (s)
# Agents10
# Iterations80
# Repetitions20
Dimension# features
C [ 0 , 2 ]
R [ 0 , 1 ]
α of F n 0.99
β of F n 0.01
Table 4. Configuration parameters of the compared algorithms.
Table 4. Configuration parameters of the compared algorithms.
AlgorithmParameter (s)Value (s)
PSO W m a x , W m i n [0.9, 0.6]
C 1 , C 2 [2, 2]
GWOa2 to 0
WOAa2 to 0
r[0, 1]
BBOHabitat modification Probability1.0
Mutation Probability0.05
Immigration Probability[0, 1]
Migration rate1.0
Max immigration1.0
Step size1.0
BAPluse rate0.5
Loudness0.5
Frequency[0, 1]
GACrossover0.9
Mutation ratio0.1
Mechanism of SelectionRoulette wheel
FA# Fireflies10
Table 5. Feature selection performance metrics.
Table 5. Feature selection performance metrics.
MetricValue
Average Error 1 1 M j = 1 M 1 N i = 1 N M a t c h ( C i , L i )
Average Select Size 1 M j = 1 M s i z e ( g j * ) D
Average Fitness 1 M j = 1 M g j *
Best Fitness M i n j = 1 M g j *
Worst Fitness M a x j = 1 M g j *
Standard Deviation 1 M 1 ( g j * M e a n ) 2
Table 6. Presented bGSDTO and compared algorithms feature selection results.
Table 6. Presented bGSDTO and compared algorithms feature selection results.
bGSDTObGWObPSObBAbWOAbBBObFAbGA
Average error0.19690.21410.24790.25750.24770.21610.24630.2277
Average Select size0.14970.34970.34970.48910.51310.51350.38420.2921
Average Fitness0.26010.27630.27470.29760.28250.28040.32660.2877
Best Fitness0.16190.19660.2550.18730.24660.27010.24530.191
Worst Fitness0.26040.26350.32270.28890.32270.35660.34290.3061
Standard deviation Fitness0.08240.08710.08650.09640.08870.13140.12330.0887
Table 7. Presented bGSDTO versus compared algorithms ANOVA test results.
Table 7. Presented bGSDTO versus compared algorithms ANOVA test results.
SSDFMSF (DFn, DFd)p Value
Treatment (between columns)0.0626270.008946F (7, 152) = 173.9p < 0.0001
Residual (within columns)0.0078191520.00005144--
Total0.07044159---
Table 8. Presented bGSDTO and compared algorithms Wilcoxon Signed-Rank test results.
Table 8. Presented bGSDTO and compared algorithms Wilcoxon Signed-Rank test results.
bGSDTObGWObPSObBAbWOAbBBObFAbGA
Theoretical median00000000
Actual median0.19690.21410.24790.25750.24770.21610.24630.2277
# Values2020202020202020
Wilcoxon Signed Rank Test
Signed ranks’ sum210210210210210210210210
Positive ranks’ sum210210210210210210210210
Negative ranks’ sum00000000
P value (two tailed)<0.0001<0.0001<0.0001<0.0001<0.0001<0.0001<0.0001<0.0001
Exact or estimate?ExactExactExactExactExactExactExactExact
Significant (alpha = 0.05)?YesYesYesYesYesYesYesYes
Discrepancy0.19690.21410.24790.25750.24770.21610.24630.2277
Table 9. Parameters of basic classification models.
Table 9. Parameters of basic classification models.
ClassifierParameter (s)Value (s)
NNbeta_10.6
beta_20.899
epsilon1 × 10 6
validation_fraction0.1
learning_rate_init0.007
hidden_layer_sizes20
k-NNleaf_size20
p2
n_neighbors3
RFmin_weight_fraction_leaf0.0
min_samples_leaf1
n_estimators40
min_samples_split2
Table 10. Classification results of the three basic models for the tested dataset.
Table 10. Classification results of the three basic models for the tested dataset.
ModelAUCMSE
NN0.7440.080226
K-NN0.73870.098999
RF0.7970.04887
Table 11. Proposed and compared algorithms classification results based on LSTM model.
Table 11. Proposed and compared algorithms classification results based on LSTM model.
GSDTO+LSTMWOA+LSTMGWO+LSTMGA+LSTMPSO+LSTM
AUC0.98260.9570.9430.9340.949
MSE0.000014130.0003190.00021580.00073820.0004196
Table 12. Proposed and compared classifiers description.
Table 12. Proposed and compared classifiers description.
GSDTO+LSTMWOA+LSTMGWO+LSTMGA+LSTMPSO+LSTM
Number of values2020202020
Minimum0.00001330.0001240.0001150.0006410.000323
25% Percentile0.00001330.0003240.0002150.0007410.000423
Median0.00001330.0003240.0002150.0007410.000423
75% Percentile0.00001330.0003240.0002150.0007410.000423
Maximum0.00001330.0004240.00033150.00078540.000454
Range00.00030.00021650.00014440.000131
10% Percentile0.00001330.0003240.0002150.0007410.000423
90% Percentile0.00001330.0003240.0002150.0007410.000423
Mean0.00001330.0003190.00021580.00073820.00042
Std. Deviation00.000051040.00003521 2.494 × 10 5 2.38 × 10 5
Std. Error of Mean00.00001141 7.874 × 10 6 5.576 × 10 6 5.32 × 10 6
Coefficient of variation0.000%16.00%16.32%3.378%5.666%
Geometric mean0.00001330.0003130.00021290.00073780.000419
Geometric SD factor11.2541.191.0361.065
Harmonic mean0.00001330.00030310.00020960.00073730.000418
Quadratic mean0.00001330.00032290.00021850.00073860.00042
Skewness −2.7510.7003−3.067−3.734
Kurtosis 13.149.72914.0716.45
Sum0.0002660.006380.0043170.014760.008391
Table 13. Proposed and compared classifiers Wilcoxon Signed-Rank Test results.
Table 13. Proposed and compared classifiers Wilcoxon Signed-Rank Test results.
GSDTO+LSTMWOA+LSTMGWO+LSTMGA+LSTMPSO+LSTM
Theoretical median00000
Actual median0.00001330.0003240.0002150.0007410.000423
# values2020202020
Wilcoxon Signed-Rank
Signed ranks’ Sum (W)210210210210210
Positive ranks’ Sum210210210210210
Negative ranks’ Sum00000
P value (two tailed)<0.0001<0.0001<0.0001<0.0001<0.0001
Estimate or Exact?ExactExactExactExactExact
Significant ( α = 0.05 )?YesYesYesYesYes
How big is the discrepancy?
Discrepancy0.00001330.0003240.0002150.0007410.000423
Table 14. The 74 testing samples diagnostic accuracy of the suggested algorithm compared to other techniques.
Table 14. The 74 testing samples diagnostic accuracy of the suggested algorithm compared to other techniques.
Fault TypeSamplesGSDTO+LSTMAdaptive [2]IEC-60599 [7]IEC 60599 Modified [27]Rog. Modified [27]
PD710010028.57100100
D11310092.3130.7761.5461.54
D22495.8391.6741.6787.579.17
T11693.7593.9568.75100100
T2410010075100100
T31010010070100100
Overall7498.2694.65089.1986.49
Fault TypeSamplesDuval [8,9]NPR [53]SVM [53]Rog. 4 Ratios [6]
PD742.8610085.7114.29
D11369.2376.9276.920
D2247587.591.6750
T11656.25100100100
T240752525
T31010010010040
Overall7466.2790.5487.8445.95
Table 15. Convergence time results for different values of the GSDTO’s parameters.
Table 15. Convergence time results for different values of the GSDTO’s parameters.
R-ParameterExploration PercentagePopulation SizeIterations CountC-Parameter
ValuesTimeValuesTimeValuesTimeValuesTimeValuesTime
0.053.35353.257100.5100.6620.13.371
0.13.145103.717201.523200.2440.23.328
0.153.018153.442303.095301.1460.33.262
0.23.503203.249404.933402.0380.43.246
0.253.251253.048506.28502.9350.53.302
0.33.114303.049607.704603.8480.63.259
0.353.071353.044709.253704.7860.73.289
0.43.06403.0818010.817806.6490.83.22
0.453.037453.0399012.324907.4630.93.285
0.53.041503.03910013.88110010.22613.304
0.553.028553.05911015.43911011.1231.13.259
0.63.037603.06712017.96212011.9951.23.208
0.653.129653.05813019.29213013.0111.33.278
0.76.341703.03414020.2114013.7981.43.277
0.756.272753.04115021.58215015.8921.53.226
0.86.082803.04816023.17316016.8641.63.213
0.854.279853.04217025.76717017.0831.73.218
0.94.57903.0618030.93718017.8721.83.211
0.953.34953.05819035.50219019.4151.93.195
13.794953.04220035.07720020.26723.186
Table 16. Minimum Fitness results for different values of the GSDTO’s parameters.
Table 16. Minimum Fitness results for different values of the GSDTO’s parameters.
R-ParameterExploration PercentagePopulation SizeIterations CountC-Parameter
ValuesFitnessValuesFitnessValuesFitnessValuesFitnessValuesFitness
0.05−11.28165−9.138610−9.650650−7.52860.1−8.0656
0.1−11.281610−9.137620−12.3506100−6.99260.2−8.0656
0.15−11.818615−11.284630−11.2826150−8.60060.3−6.9926
0.2−11.820620−10.748640−11.2856200−8.60260.4−8.0626
0.25−11.283625−11.280650−10.2126250−8.06560.5−8.0656
0.3−11.284630−10.748660−12.3586300−8.06560.6−6.9926
0.35−10.747635−10.748670−12.3586350−9.67560.7−8.0656
0.4−11.285640−11.285680−12.3586450−8.06560.8−8.0616
0.45−10.748645−11.284690−11.8226500−9.13860.9−6.9926
0.5−11.284650−10.2126100−12.3586650−10.21261−8.0656
0.55−11.284655−11.2856110−12.3586700−8.60261.1−8.0656
0.6−12.355660−11.8206120−12.3586750−10.21261.2−8.0656
0.65−11.284665−11.2836130−12.3586800−10.74861.3−9.1386
0.7−10.747670−11.2846140−12.3586850−9.13861.4−10.2116
0.75−10.746675−11.2856150−12.3586900−9.67561.5−11.2836
0.8−12.354680−10.7476160−12.3586950−10.74861.6−9.1386
0.85−11.236685−12.3556170−12.35861000−10.21261.7−10.2116
0.9−11.809690−11.2716180−12.35861050−10.74861.8−12.3576
0.95−12.319695−11.2746190−12.35861150−9.67561.9−11.2846
1−12.330695−11.8096200−12.35861200−10.21262−12.3586
Table 17. Results of regression analysis for the GSDTO’s parameters.
Table 17. Results of regression analysis for the GSDTO’s parameters.
Convergence TimeMinimum Fitness
ParametersR SquareSignificance FR SquareSignificance F
R-Parameter 7.85 × 10 1 1.44 × 10 3 8.85 × 10 1 9.16 × 10 4
Exploration Percentage 6.03 × 10 1 1.67 × 10 5 6.40 × 10 1 3.91 × 10 3
Population Size 8.06 × 10 1 1.70 × 10 9 9.80 × 10 1 4.59 × 10 3
Iterations Count 8.23 × 10 1 1.70 × 10 11 7.44 × 10 1 1.44 × 10 5
C-Parameter 4.06 × 10 1 1.24 × 10 6 9.57 × 10 1 9.16 × 10 4
Table 18. ANOVA test results for the GSDTO’s parameters of Convergence time and Fitness.
Table 18. ANOVA test results for the GSDTO’s parameters of Convergence time and Fitness.
SSDFMSF (DFn, DFd)p Value
Treatment (between columns)11.5491.282F (9, 190) = 2483p < 0.0001
Residual (within columns)0.09811900.0005163--
Total11.64199---
Table 19. T-test with one tail was run at a significance level of 0.05 for different values of the GSDTO’s parameters.
Table 19. T-test with one tail was run at a significance level of 0.05 for different values of the GSDTO’s parameters.
Convergence Time
R-ParameterExploration PercentagePopulation SizeIterations CountC-Parameter
Theoretical mean00000
Actual mean0.77750.5930.80250.8280.421
Number of values2020202020
One sample t test
t, dft = 93.03, df = 19t = 48.00, df = 19t = 72.62, df = 19t = 93.97, df = 19t = 38.47, df = 19
p value (two tailed)<0.0001<0.0001<0.0001<0.0001<0.0001
p value summary********************
Significant (alpha = 0.05)?YesYesYesYesYes
How big is the discrepancy?
Discrepancy0.77750.5930.80250.8280.421
SD of discrepancy0.037380.055250.049420.03940.04894
SEM of discrepancy0.0083570.012350.011050.0088110.01094
95% confidence interval0.7600 to 0.79490.5671 to 0.61890.7794 to 0.82560.8096 to 0.84640.3981 to 0.4439
R squared (partial eta squared)0.99780.99180.99640.99790.9873
Minimum Fitness
R-ParameterExploration PercentagePopulation SizeIterations CountC-Parameter
Theoretical mean00000
Actual mean0.8850.64120.97640.7590.937
Number of values2020202020
One sample t test
t, dft = 122.0, df = 19t = 87.20, df = 19t = 188.8, df = 19t = 57.81, df = 19t = 80.10, df = 19
p value (two tailed)<0.0001<0.0001<0.0001<0.0001<0.0001
p value summary********************
Significant (alpha = 0.05)?YesYesYesYesYes
How big is the discrepancy?
Discrepancy0.8850.64120.97640.7590.937
SD of discrepancy0.032440.032890.023130.058710.05231
SEM of discrepancy0.0072550.0073530.0051720.013130.0117
95% confidence interval0.8698 to 0.90020.6258 to 0.65660.9656 to 0.98720.7315 to 0.78650.9125 to 0.9615
R squared (partial eta squared)0.99870.99750.99950.99430.997
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

El-kenawy, E.-S.M.; Albalawi, F.; Ward, S.A.; Ghoneim, S.S.M.; Eid, M.M.; Abdelhamid, A.A.; Bailek, N.; Ibrahim, A. Feature Selection and Classification of Transformer Faults Based on Novel Meta-Heuristic Algorithm. Mathematics 2022, 10, 3144. https://doi.org/10.3390/math10173144

AMA Style

El-kenawy E-SM, Albalawi F, Ward SA, Ghoneim SSM, Eid MM, Abdelhamid AA, Bailek N, Ibrahim A. Feature Selection and Classification of Transformer Faults Based on Novel Meta-Heuristic Algorithm. Mathematics. 2022; 10(17):3144. https://doi.org/10.3390/math10173144

Chicago/Turabian Style

El-kenawy, El-Sayed M., Fahad Albalawi, Sayed A. Ward, Sherif S. M. Ghoneim, Marwa M. Eid, Abdelaziz A. Abdelhamid, Nadjem Bailek, and Abdelhameed Ibrahim. 2022. "Feature Selection and Classification of Transformer Faults Based on Novel Meta-Heuristic Algorithm" Mathematics 10, no. 17: 3144. https://doi.org/10.3390/math10173144

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop