Next Article in Journal
Research, Application and Future Prospect of Mode Decomposition in Fluid Mechanics
Next Article in Special Issue
An Equivalent-Perceptional Intertemporal Choice Heuristics Model for Electric Operation Vehicle Charging Behavior
Previous Article in Journal
A Gradient-Based Algorithm with Nonmonotone Line Search for Nonnegative Matrix Factorization
Previous Article in Special Issue
A Novel Detection and Identification Mechanism for Malicious Injection Attacks in Power Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Attack–Defense Confrontation Analysis and Optimal Defense Strategy Selection Using Hybrid Game Theoretic Methods

1
Institute of Engineering, Yanshan University, Qinhuangdao 066004, China
2
School of Mathematics and Statistics, Taishan University, Tai’an 271000, China
3
College of Electric Engineering, Nanjing Xiaozhuang University, Nanjing 210023, China
*
Authors to whom correspondence should be addressed.
Symmetry 2024, 16(2), 156; https://doi.org/10.3390/sym16020156
Submission received: 25 December 2023 / Revised: 19 January 2024 / Accepted: 25 January 2024 / Published: 29 January 2024

Abstract

:
False data injection attacks are executed in the electricity markets of smart grid systems for financial benefits. The attackers can maximize their profits through modifying the estimated transmission power and changing the prices of market electricity. As a response, defenders need to minimize expected load losses and generator trips through load and power generation adjustments. The selection of strategies of the attacking and defending sides turns out to be a symmetric game process. This article proposes a hybrid game theory method for analyzing the attack–defense confrontation: firstly, a micro-grid-based power market model considering false data injection attacks is established using the Nash equilibrium method; secondly, the attack–defense game function is constructed and solved via the Stackelberg equilibrium algorithm. The Markov game algorithm and distributed learning algorithm are used to update equilibrium function; finally, a dynamic game behavior model of the two players is constructed through simulating the attack–defense probability. The evolutionary game method is used to select the optimal defense strategy for dynamic probability changes. Modified IEEE standard bus systems are illustrated to certify the effectiveness of the proposed model.

1. Introduction

Recently, smart grid (SG) systems with networking characteristics and uncertainty distributed power supplies have become increasingly complex [1]. Cyber-attack problems have become the main security issue for the reliability and security of SG operation [2]. False data injection (FDI) attacks are able to bypass the online monitoring of state estimation, steal energy, and gain economic benefits through false scheduling [3]. Attackers establish their own behavior through monitoring and injecting attacks, which may interfere with operations and modify estimated transmission power [4]. The payoff function is established to maximize their economic benefit with the least attack cost [5]. Meanwhile, the defenders of the system need to establish corresponding multi-level defense measures to deal with attacks and mitigate risks. The payoff function of defenders is established to minimize expected load losses and generator trips based on the current attack situations [6]. The selection of strategies by the attackers and defenders whose motives and emergency responses are contradictory can be modeled with a symmetric game process [7]. The game method is constructed to admit equilibrium to enable the two players to maximize their respective minimum rewards [8].
Relying on the ability to make multifaceted decisions on power grid security issues, game theory is widely applied in the field of network security [9,10,11,12,13,14,15,16]. In [9], a security resource allocation game model is established to discuss the load redistribution attacks problem. Both system operators and attackers maximize their own payoffs through selecting the optimal strategy; in [10], the game composed of attackers and defenders is constructed with a probability-weighting algorithm, and the optimal investment strategies of the players are obtained using the Nash Equilibrium. However, refs. [9,10] are inadequate in describing how both offensive and defensive strategies change with each other; ref. [11] investigates the Stackelberg game through having legitimate launchers act as leaders, and disruptors acting as followers. A genetic algorithm was used to obtain the optimal frequency hopping speed and optimal transmission power; ref. [12] formulates a Stackelberg game model through making the Cloud provider the leader and making attackers the second player, and obtains the utility function through applying artificial neural networks. Refs. [11,12] establish the master–slave change of offensive and defensive strategies. However, the model lacks in describing the cumulative profits in a dynamic, changing game process; ref. [13] establishes a Markov anti-attack model, compares the effectiveness of network space simulation defense under different attack types, and uses dynamic games with incomplete information to determine the optimal strategy; ref. [14] formulates a multistage optimization model for the deployment of a mobile target defense mechanism under a Markov decision, maximizing the profits under environmental constraints. However, each iteration solution of [13,14] depends on the revenue in the current unit time, ignoring the previous revenue, bringing in an exponential increase in complex calculations; in [15], a Markovian–Stackelberg game is proposed to simulate the sequential actions of attackers and defenders, and a secure constrained optimal power flow is given, which preserves the safety margin of key components to minimize the power outage scale and potential future risks; in [16], the adversarial interaction of the attacker and defender is modeled as a resource-constrained game, and a linear-time algorithm is used to obtain the optimal defense strategy. However, the game process in the above works assumes that both sides (attacker and defender) of the game are in pure rationality. In practice, the players will constantly adjust their strategies and dynamically pursue profits. Therefore, common rational game models overlook the limitations of bounded rationality, which can lead to deviations between attack and defense behaviors and actual situations, thereby weakening the accuracy and guiding value of security defense strategy selection methods. Regarding the above problem, evolutionary game method [17] is put in to analyze the ability of limited rationality in offensive and defensive behavior. The evolutionary game takes bounded rational players as the basis of game analysis. Based on the idea of biological dynamic evolution, it depicts both sides of attack and defense through a learning mechanism, and constantly improves the internal drive of behavior strategy [18,19]. The evolutionary game in the above works has established the evolutionary game between attackers and defenders according to system dynamics. The goal of optimal overall network performance is achieved. However, the method of solving evolutionary game equilibrium and the specific strategy selection method are not designed, which makes it difficult to guide the security defense decision.
Therefore, in view of the selection of the security defense strategy in network attack–defense confrontation, this paper analyzes the evolutionary trend of offensive and defensive behavior from limited rationality in reality. On this basis, the evolutionary stability strategy solution method is proposed to achieve the selection of optimal defense strategies. At the same time, low-income players in the game process are constantly learning from the strategies of the high-income players. This reflects the dynamic evolutionary trend of offensive and defensive confrontation under bounded rationality constraints. In addition, the law of forming an evolutionary stable equilibrium in different situations is also analyzed and summarized.
Further, this article considers the impact of the optimized output of micro-grid (MG) energy management on the dynamic strategy selection of two participants. The MG combines distributed energy, storage devices, and corresponding loads in a reduced space through advanced control systems [20]. The MG can participate in the electricity market through online trading with the main power grid, and can also operate independently in island mode to balance internal supply and demand [21,22]. The MG has played an increasingly important role in maintaining the power balance of the entire SG power system and reducing generation costs. The optimal operation results of the electricity market and generation output are considered in the game equilibrium solution.
The main contributions of this paper are listed as follows:
(1)
MG-based electricity markets model of an SG power system is considered for FDI attacks. The electricity market is established with a double-sided bidding mechanism using the Nash equilibrium game theory.
(2)
A hybrid game model of payoff function between attack and defense is established. Game theoretic methods are proposed to discuss the interaction behavior between attack and defense sides.
(3)
The benefits according to attack and defense strategies are quantified using the evolutionary game method, and the dynamic evolutionary learning of attack and defense probability is discussed.
(4)
The optimal defense strategy is selected after the evolutionary stable equilibrium solution is solved, and the dynamic confrontation trend of both sides is studied.
This article proposes a hybrid game theory method for analyzing the attack–defense confrontation: firstly, a micro-grid-based power market model considering false data injection attacks is established using the Nash equilibrium method; secondly, the attack–defense game function is constructed and solved via the Stackelberg equilibrium algorithm. The Markov game algorithm and distributed learning algorithm are used to update equilibrium function; finally, a dynamic game behavior model of the two players is constructed through simulating the attack–defense probability. The evolutionary game method is used to select the optimal defense strategy for dynamic probability changes.
The rest of this paper is organized as follows: The power market attack model with micro-grid participation is provided in Section 2; in Section 3, the game model of payoff function between attack and defense is established. Game theoretic methods are proposed to discuss the interaction behavior between the attacking and defending sides. An attack–defense evolutionary game model is constructed to obtain optimal strategies against the dynamic changing probability in Section 4. The optimal selection strategies for the defender are also listed in this section; numerical examples and discussions of the modified IEEE standards for 14 bus systems and 118 bus systems are presented in Section 5; the concluding remarks and future works are submitted in Section 6.

2. Electricity Markets Attack Model

In an electricity power market, the net power injection is expressed as the difference between generation and load. The state variable x is expressed by the relationship between power generation vector P and demand vector L as below.
x = L T , P T T
Then the line flow vector F will be expressed as below.
F = H f x = H f L T , P T T
H f is the measurement matrix of the distribution factor for the transmission line flow vector.
In an FDI attack process, attackers modify node prices to gain profits in the electricity market [23]. In this section, we will introduce the general structure of MG-based electricity market models and discuss real-time market attack models. The price is determined using a DC model that ignores reactive power and marginal losses.

2.1. Electricity Markets Model with MGs

In the electricity market, MGs, as participants who pursue maximum profits through exchanging electricity, can establish a bidirectional bidding mechanism model through calculating the market clearing price (MCP) and location marginal price (LMP) [24]. The Nash equilibrium game theory [25] is proposed in this section to make decisions on bidding and quotation strategies.
(1)
Strategy for MGs: In the bidding process, the strategy of participants is formulated based on the range of power generation capacity in a specific strategic space:
P M G i , τ Φ M G i , τ = P M G i , τ min , P M G i , τ max
where P M G i , τ is the generation scheduling of the i th MG player M G i , τ , i = 1 , 2 , , M g . P M G i , s min and P M G i , s max are the minimum and maximum capacities of P M G i , τ .
(2)
Bidding profit function B i τ P M G i , τ for the i th player:
B i τ P M G i , τ = λ r e f , i τ P M G i , τ ( 1 + θ i τ ) C i τ P M G i , τ
where C i τ P M G i , τ = a 0 i τ + a 1 i τ P M G i , τ + a 2 i τ P M G i , τ 2 . a 0 i τ , a 1 i τ , and a 2 i τ are cost coefficients. θ i τ is the scale factor of each player adding to the bidding cost function. λ r e f , i τ stands for the market unified trading price for players during the bidding period. In the day-ahead market, the problem is solved in a linearized DC model as below.
maxmize           B i τ P M G i , τ s u b j e c t   t o         i = 1 M g P M G i , τ + i = 1 M g P e x c h a n g e i , τ = j = 1 D L j τ                                   P M G i , τ min P M G i , τ P M G i , τ max                      
where M g is the number of buses. P M G i , τ is the generation at bus i . D is the number of loads. P e x c h a n g e i , τ is the transmitted power from/to the main grid. L j τ is the forecast load at bus j . P M G i , τ min and P M G i , τ max   are the lower and upper capacity bounds of P M G i , τ . The Lagrangian function for real-time optimal problems can be constructed as follows:
L L a g r a n g e i , τ P M G i , τ = B i τ P M G i , τ + μ i τ P M G i , τ min P M G i , τ + υ i τ P M G i , τ P M G i , τ max
where μ i τ and υ i τ denote the Lagrangian multipliers for the upper and lower limits of generator capacity.
(3)
Nash equilibrium game: The Nash equilibrium P M G 1 , τ , P M G 2 , τ , , P M G M g , τ means that when P M G i , τ is implemented in the game, no player can gain additional profits through changing their power generation scheduling. The strategy set of all players can be calculated through the following optimal iterations:
L L a g r a n g e i , τ P M G i , τ P M G i , τ = B i τ P M G i , τ P M G i , τ μ i τ + υ i τ   = 0
After the above optimal function is calculated, the MCP of the i th MG can be obtained when μ i τ and υ i τ converge into an allowed range [26]:
λ r e f , i τ = ( 1 + θ i τ ) a 1 i τ + a 2 i τ P M G i , τ
where λ r e f , i τ is the day-ahead locational marginal price of the i th player. The MCP λ r e f , τ can be represented as the highest bid price for all players during the bidding period:
λ r e f , τ = max λ r e f , i τ ,   i = 1 , 2 , , M g

2.2. Attack Model of the Real-Time Market

In the real-time market, due to the randomness and real-time dynamic characteristics of load demand, the runtime state is different from the optimal value:
x * = x + ω *
F * = H f x * = H f x + ω f
where x * and F * are the state variables and line flow vector in real time. ω * and ω f are the measurement noise. The attacker manipulates the electricity prices through modifying the power flow measurement with a vector of false data. With the correction of power flow, the economic dispatch of the power system will also change. The attack will succeed when the following physical characteristics of the system are met:
Δ F = H f Δ x = H f Δ L T , Δ P T T
The strategy equation of attackers in real-time electricity trading can be expressed as follows:
Δ i = 1 M U g , i t P g , i t Δ L i t + Δ P b u y , i t Δ P s e l l , i t = 0
Δ F l t = H f Δ x t = Δ P i t = Δ P j t
where Δ p g , i t is the change in injection power at node i and U g , i t is the scale factor. Δ P b u y , i t and Δ P s e l l , i t are the change powers for buying and selling, respectively. Δ P i t and Δ P j t represent changes of power in node i and j , respectively. Δ F l t is the change in branch power flow l . The LMP λ l m p , t can be obtained through calculating the minimum cost added in the power system:
λ l m p , t = λ r e f , τ + P l o s s , t P i , t λ r e f , τ + λ c o n g e s t i o n , t
where λ r e f , τ is the reference marginal obtained from (9). P l o s s , t P i , t λ r e f , τ is the marginal loss component. λ c o n g e s t i o n , t is the congestion component, which is further explained. P i t * , L j t * , and F l t * are defined as the optimal values of solution P i t , L j t , and F l t observed in the day-ahead market. P ^ i t and L ^ j t are the real-time state variables. The economic dispatch model can be obtained as follows:
minimize P g , i t   i = 1 N t = 1 T C i t P ^ g , i t + Δ P ^ g , i t + Δ P s e l l , i t   s u b j e c t   t o   i = 1 N t = 1 T Δ P ^ g , i t   + Δ P b u y , i t Δ P s e l l , i t   = 0     Δ P g , i t min Δ P ^ g , i t Δ P g , i t max     i = 1 N t = 1 T Δ P i t = j = 1 N t = 1 T Δ L j t     Δ F l t 0               l c o n g s e t l + : F ^ l t F l t max     Δ F l t 0               l c o n g s e t l : F ^ l t F l t min
where Δ P i t = P ^ i t P i t * , Δ L j t = L ^ j t L j t * . If the estimated power flow exceeds the line flow limit, the line is defined as congested. Δ F l is discussed during line congestion l c o n g s e t . The real-time Lagrange function is described in (17).
L L a g r a n g e , t = i = 1 N C i t Δ P i t j = 1 N C j t Δ L j t   λ i = 1 N Δ P i t j = 1 N Δ L j t   + i = 1 N μ i t max Δ P i t Δ P i t max   + i = 1 N υ i t min Δ P i t min Δ P i t + l l + η l + Δ F l t + l l η l Δ F l t
where λ c o n g e s t i o n , t = H j t T l η l + η l . μ i t max and υ i t min are the multipliers of the upper and lower limits of generator capacity. Solve (17) and calculate the Lagrange multiplier, and λ l m p , j t in the real-time market will be obtained in (18). In the DC lossless model, F l t L j t = H l j t , and then λ l m p , j t can be obtained in (19).
λ l m p , j t = L L a g r a n g e , t L j t = λ r e f , τ + l η l + η l F l t L j t
λ l m p , j t = λ r e f , τ + H j t T l η l + η l
where H l j t is the element on the l th row and j th column of matrix H j t . η l + and η l are the multipliers of the maximum and minimum limits of power flow at line l , respectively. It is clear that when η l + = η l and λ l m p , j t = λ r e f , τ , the congested routes have been alleviated. Assuming the attacker purchases electricity P l at bus i with prices λ r e f , i s in the current market, and sells it at bus j with prices λ r e f , j s ; then purchases equal quantities of electricity at bus j with prices λ l m p , j t in the real-time market and sells it at bus i with prices λ l m p , i t , the profit of the attacker can be obtained when the value of the following formula is positive:
Profit att = λ r e f , i τ λ r e f , j τ P l + λ l m p , j t λ l m p , i t P l

3. Behavior Model of Attackers and Defenders

In the attack and defense game model, attackers formulate allocation strategies for attack resources to achieve higher economic benefits. The defender then reduces the loss of the power grid to the minimum through the method of power flow distribution and generator output dispatching. In this section, the payoff game model of attackers and defenders with objective function are provided.

3.1. Attack Payoff Modeling

Electrical equipment is protected by hardware and software protection systems. The ultimate goal of the attack is to cause varying degrees of damage, such as a power outage within the system, through executing erroneous destructive variables on the system control. The motive of the attacker is to pay little cost and obtain as much revenue as possible. Usually, the rational attacker will choose the best strategy to achieve the target according to the profit and loss principle. Then, the payoff for an attacker U A can be expressed as follows:
max   U A = max Profit att EL att
where EL a t t is the expected loss when implementing an attack. In this paper, the illustrative model in [27] is employed to establish the success attack function. The cost of attack efforts can be described as follows:
EL att = C a 1 P Esc + C a 2 P Suc | Esc
where P Escape   detection = exp A D V is the probability of an attacker escaping detection when executing an attack. V is the number of known vulnerabilities for an attacker in the system. A is the number of available exploits of vulnerabilities for the attacker. D is the selection number of defenses by defenders against the attack. P Success | Escape   detection = A 2 A 2 + D 2 represents the probability that an attack can be successfully executed without being detected. The probability of obtaining a successful attack is as below.
P Suc = P Esc   P Suc | Esc = exp A N D N V N A N 2 A N 2 + D N 2
V N , A N , and D N are influenced by the skill levels of attackers and defenders, respectively. C a 1 and C a 2 are the costs of conducting attacks. C a 1 is the escaping detection cost that ties to the technical difficulty and duration of protective measures against attacking targets. C a 2 is the succeeding attack cost, responding to a defense adjustment after an attack.

3.2. Defense Payoff Modeling

The defense reaction is made through readjusting the generator output to maintain system balance. When it still cannot meet the load demand, measures of shedding loads are taken to prevent cascading failures of the power grid and ensure the stable operation of the power grid. The defense model is quantified with an objective function while expecting the minimal cost of shedding loads and generator tripping. The defender’s payoff function can be expressed as follows:
min   U D = min C D + C G P Suc + C M P Esc - P Suc | Esc
where C D and C G are the costs for load loss and generator tripping after the attack is successful established. C M is the maintenance cost of protection against a successful attack. The system operator minimizes the system operation cost considering operating constraints (25)–(28).
0 j = 1 N t = 1 T S j t j = 1 N t = 1 T L j t + Δ L j t
P G , i t =   i = 1 N t = 1 T 1 b G , i t P G C ,   b G , i t 0 , 1
P i = i = 1 N j = 1 N V i V j g i j cos θ i j + b i j sin θ i j
P i Δ P i i = 1 N j = 1 N V i b i j V j θ i j Δ θ i j = 0
(25)–(28) show the operating and capacity constraints of load and generator, where S j t is the shedding load of bus i at time t . θ i j = θ i θ j is the voltage phase angle difference between bus i and j . P G , i t is the tripped generators of bus i at time t . P G C is the capacity of generator G . b G , i t is a binary variable indicating whether the generator has tripped. V i and V i are the plural voltages of bus i and bus j . P i and Δ P i are the active injection power and power change of node i . b i j and g i j are the susceptance and conductance between bus i and j .

3.3. Hybrid Game Method

In this section, a hybrid game model is proposed to describe the interaction process of attack and defense competition. The game model played over a finite state space, defined as S , Ω A , Ω D , U A , U D , p A , q D , with the players of attacker and defender. The components in the game model are described as follows:
(1)
Ω A = Ω A 1 , Ω A 2 , , Ω A n represents the game strategy space of attackers;
(2)
Ω D = Ω D 1 , Ω D 2 , , Ω D m represents the game strategy space of defenders; n and m are the numbers of offense and defense, respectively, and n, m  2;
(3)
U A = U A 1 , U A 2 , , U A n represents the payoff function of attackers corresponding to the strategies of Ω A , Ω D ;
(4)
U D = U D 1 , U D 2 , , U D m is the payoff function of defenders;
(5)
p A = p A 1 , p A 2 , , p A n is the probability of attackers corresponding to Ω A , Ω D ;
(6)
q D = q D 1 , q D 2 , , q D m is the probability of defenders
where S is the state space of the game.

3.3.1. Stackelberg Equilibrium

In the attack–defense confrontation, the attacker makes the strategy of obtaining higher economic benefits with a lower cost of attacking. The defender takes the defense strategy through reducing the loss. In the attack–defense confrontation, each player aims to maximize their rewards while keeping the rewards of other players minimal. (21) and (24) become a Stackelberg equilibrium problem.
Let β * be a solution for (21) and η * be a solution for (24). Then, the point ( β * , η * ) becomes a Stackelberg equilibrium solution of the proposed game model, if for any ( β , η ) with β 0 and η 0 , the following conditions are satisfied:
max   U A β * , η * max   U A β , η *
min   U D η * , η * , β * min   U D η , η * , β *
where η * is the other equilibrium strategies. When all game participants’ strategies are in Stackelberg balance, the effectiveness of all players’ profits reach the maximum, and any participants cannot achieve greater benefits through changing their own strategies alone.

3.3.2. Markov Game Solution

In the network attack–defense confrontation, different participants have different levels of security knowledge, leading to different decision making. Meanwhile, with the time going on and the driving force of the learning mechanism, low-income participants continue to learn the strategies of the high-income participants and improve their behaviors. The dynamic game process of the attack and defense depends on the previous game process and the actions taken by all players. In this section, we use the minmax method of the Markov game and Q-learning algorithm in [28] with expected immediate rewards and expected long-term rewards equations to update the payoff functions. The optimal discounted sum of expected rewards for the attacker under a pair of strategies Ω A , Ω D can be represented as follows:
Q A s , Ω A , Ω D = R A s , Ω A , Ω D + γ s T s , Ω A , Ω D , s V A s
V A s = min Ω D max p A Q A s , Ω A , Ω D p A
where s , s S . 0 γ 1 is a discount factor, which gives the discount factor of future rewards on the optimal decision. R A s , Ω A , Ω D and R D s , Ω A , Ω D are the expected immediate rewards for the attacker and defender in state s . V A s is the payoff function for the attacker in state s . Similarly, the optimal discounted sum of expected rewards function for the defender is obtained:
Q D s , Ω A , Ω D = R D s , Ω A , Ω D + γ s T s , Ω A , Ω D , s V D s
V D s = min Ω A max q D Q D s , Ω A , Ω D q D
where Q A s , Ω A , Ω D and Q D s , Ω A , Ω D are the expected long-term rewards for the attacker and defender in state s . V D s is the payoff function for the defender in state s . T s , Ω A , Ω D , s is the probability of the state transition from s to s .

3.3.3. Distributed Learning Algorithm

The payoff function is updated and iterated through the Markov game algorithm. However, each iteration depends on the reward obtained at the current time step, ignoring previous rewards obtained. To find the optimal equilibrium of the Markov game, in this subsection, a distributed learning algorithm is provided. The algorithm updates the mixed strategy vectors according to the framework of learning automata as below.
q ω t + 1 = q ω t + ε r ω t e ω t q ω t
In this subsection, each player ω firstly initializes the strategy vector q ω t at time instant t . Then, each player randomly and independently selects a strategy based on the probability distribution of their strategy vector. The set of actions taken by players at different times results in a reward for each player expressed as r ω t . ε is an arbitrarily small positive constant. e ω t is a column vector with a length equal to the size of the player’s action set ω . The updated strategies are employed in the Markov game equations and Stackelberg game equations for the optimal equilibrium.

4. Design of the ADEG strategy

In the game process, the probability of both sides adopting different strategies varies. As the attack is random and the probability changes with time under a learning mechanism, the selection of attack/defense strategy forms a dynamic change process. In order to simulate the dynamic game process and obtain the optimal defense strategy against differential attack strategies, this section designs an attack–defense evolutionary game (ADEG) model to discuss the dynamic changes in the probability of players with complete and partial system information.

4.1. ADEG Model and Analysis

4.1.1. ADEG Model Construction

In this section, it is assumed that the attacker selects two attack strategies dynamically, and the defender selects two strategies for defense accordingly. Define a u v , b u v as the attack and defense payoffs when they take Ω A u , Ω D v . The dynamic change rate is expressed by the solution of the replication dynamic equation.
(1)
Profits of all the players:
U A u = v = 1 m q D v a u v , U A ¯ = u = 1 n p A u U A u
U D v = u = 1 n p A u b u v , U D ¯ = v = 1 m q D v U D v
where U A u and U D v are the expected profits for the attacker and defender. U A ¯ and U D ¯ are the corresponding average profits.
(2)
Replication dynamic equations for the profits:
d p A u d t = p A u U A u U A ¯
d q D v d t = q D v U D v U D ¯
(3)
Stable equilibrium solution:
d p A u d t = 0 d q D v d t = 0

4.1.2. Evolutionary Stable Solution

In this section, a specific example with four strategies ( Ω A 1 , Ω D 1 ), ( Ω A 1 , Ω D 2 ), ( Ω A 2 , Ω D 1 ), and ( Ω A 2 , Ω D 2 ) are given for the solution of evolutionary stable equilibrium. The attacking and defending sides choose strategies with different probabilities, and generate different benefits. The game tree is shown in Figure 1.
The process of solution for the defenders and attackers can be described as follows:
(1)
Expected and average profit of a defender:
U D 1 = p A 1 b 11 + p A 2 b 21
U D 2 = p A 1 b 12 + p A 2 b 22
U ¯ D = q D 1 U D 1 + q D 2 U D 2
where U D 1 and U D 2 represent the expected profits in defense strategies 1 and 2, respectively; U ¯ D represents the average profit of the defender. The replication dynamic equation of the defender over time can be represented by (44).
d q D 1 ( t ) d t = q D 1 ( U D 1 U ¯ D ) d q D 2 ( t ) d t = q D 2 ( U D 2 U ¯ D )
(2)
Expected and average profit of an attacker:
U A 1 = q D 1 a 11 + q D 2 a 21
U A 2 = q D 1 a 12 + q D 2 a 22
U ¯ A = p A 1 U A 1 + p A 2 U A 2
where U A 1 and U A 2 are the expected profits in attack strategies 1 and 2, respectively; U ¯ A represents the average profit of the attacker. The replication dynamic equation of the attacker over time can be represented by (48).
d p A 1 ( t ) d t = p A 1 ( U A 1 U ¯ A ) d p A 2 ( t ) d t = p A 2 ( U A 2 U ¯ A )
(3)
Based on the proposed ADEG model, when d q D 1 ( t ) d t = d q D 2 ( t ) d t d p A 1 ( t ) d t = d p A 2 ( t ) d t is satisfied, the solution can be obtained only through calculating d q D 1 ( t ) d t = 0 d p A 1 ( t ) d t = 0 . Through the equations, five sets of solutions for p A 1 and q D 1 can be obtained in Table 1.

4.2. ADEG-Based Optimal Defense Strategy Selection

In addition to the defense strategy that the defender can choose according to the attack probability, it can also influence the attack probability selection and send messages to the attacker. In this section, the dynamic relationship between the two players is discussed through the evolutionary stable equilibrium solutions. The replication dynamic model with three situations is shown in Figure 2.
(1)
For the defender, when p A 1 = b 22 b 21 b 11 b 21 b 12 + b 22 , with any probability selection of defense strategy q D 1 , there is d q D 1 ( t ) d t = 0 . However, once the value of p A 1 shifts, d q D 1 ( t ) d t will change dramatically, indicating that the state represented by the graph is not stable; once p A 1 b 22 b 21 b 11 b 21 b 12 + b 22 , q D 1 = 0 and q D 1 = 1 are two stable states. According to these, when p A 1 > b 22 b 21 b 11 b 21 b 12 + b 22 , q D 1 = 0 is the evolution stable strategy; when p A 1 < b 22 b 21 b 11 b 21 b 12 + b 22 , q D 1 = 1 is the strategy.
(2)
For the attacker, when q D 1 > a 22 a 21 a 11 a 21 a 12 + a 22 , p A 1 = 1 is the strategy; when q D 1 < a 22 a 21 a 11 a 21 a 12 + a 22 , p A 1 = 0 is the strategy; when q D 1 = a 22 a 21 a 11 a 21 a 12 + a 22 , for any p A 1 , there is d p A 1 ( t ) d t = 0 . However, once the value of q D 1 is shifted, d p A 1 ( t ) d t will undergo significant changes, leading to an unstable state.
Define p A , q D as the selection probability of attack strategies and defense strategies. After the update of the proposed hybrid game method above, the final optimal objective function of game model can be obtained as follows:
max p A U A * β * , η * max p A U A * β , η *  
min q D U D * η * , η * , β * min q D U D * η , η * , β *
The attack–defense strategies are developed based on simulated probabilities, providing a new defense strategy selection probability in advance to cope with possible attack situations. The simplified process of obtaining the optimal defense strategy is shown in Algorithm 1.
Algorithm 1. Process of obtaining the optimal defense strategy
Phase 1-:
(1)
Define the state space of power grid S .
(2)
Define attacker’s and defender’s strategy space Ω A , Ω D .
(3)
Derive state transition matrix T s , Ω A , Ω D , s .
(4)
Find the Stackelberg equilibrium solution ( β * , η * ).
(5)
Obtain the expected rewards from the attacker and defender R A s , Ω A , Ω D , R D s , Ω A , Ω D according to ( β * , η * ).
Phase 2-
(1)
Set initial

   Q A s , Ω A , Ω D , Q D s , Ω A , Ω D , V A s , V D s
(2)
Initialize q ω 0 .
  while Not Converged do
   Collect the payoff r ω t
   Update the strategy q ω t + 1 = q ω t + ε r ω t e ω t q ω t
   Check Convergence
   if Converged then
     break
    end if
   end while
(3)
Return the strategy vector q ω t .
(4)
Update Q A s , Ω A , Ω D , Q D s , Ω A , Ω D according to (31), (33).
(5)
Update V A s , V D s according to (32), (34).
(6)
Define the ADEG model.
(7)
Calculate the replication dynamic equation of both of the players.
(8)
Derive the optimal Stackelberg equilibrium solution according to (49), (50).
(9)
Derive the final optimal strategy of the defenders through solving the equilibrium solution.
The simplified process of obtaining the optimal network defense strategy based on the hybrid game algorithm is shown in Figure 3.

5. Discussion

In this section, a simulation case is constructed to demonstrate the electricity markets attack case in an SG with MGs connected. We will verify the proposed network attack and defense game model and discuss the optimal operation. In this section, we introduced simulations using IEEE 14 bus systems and IEEE 118 bus systems.

5.1. Optimal Attack–Defense Strategies Game Model

According to the above model, system dynamics are used to verify the effect of the ADEG model. In this paper, the evolutionary stable solution is in four situations of p A = 0 ,   q D = 0 ; p A = 0 ,   q D = 1 ; p A = 1 ,   q D = 0 ; and p A = 1 ,   q D = 1 . The two attack–defense strategies selection with different initial probabilities are listed in Table 2 and Table 3 as below.
In the real-time market, prices will change after modifying the estimated line flow. The system power will be reduced below the thermal limit of the transmission line, and congested lines will be alleviated. The impact of different attack probabilities on LMP is shown in Figure 4. The proposed method will be validated using the MATPOWER 4.0 software package for age verification [29].
For further comparison, 100,000 experiments are generated in both transmission lines 2–3 and 4–5 of the IEEE-14 bus system. The ratio of attacks on two routes is 1:1. The detection performances of the four attack strategies are discussed and compared in Table 4.
Based on the proposed game methods, the optimal payoff value of the two players under different game methods can be calculated and shown as follows.

5.1.1. Players with Complete Information

a.
IEEE 14-bus system
The proposed method was tested on an improved IEEE 14 bus testing system, where some of the generators were replaced by MGs with distributed and renewable generator units as well as storage units. The power value and demand are maintained under the same basic conditions as the IEEE standard 14 bus structure [30]. The system consists of one utility grid generator, four MGs, and eleven loads. MGs are located on buses 2, 3, 6, and 8, with 11 loads located on buses 2, 3, 4, 5, 6, 9, 10, 11, 12, 13, and 14. The game solution when the players obtain complete information is shown in Table 5.
From Table 5, it can be concluded that:
(1)
When the attacker chooses Ω A 1 , the defender will choose Ω D 1 to defend as the cost of C D in Ω D 2 is high;
(2)
When the attacker chooses Ω A 2 , the defender will choose Ω D 2 to defend as the cost C G , C M in Ω D 1 is high;
(3)
To achieve the best defense effect, the defender will choose Ω D 1 when the attacker selects Ω A 1 , and the defender will choose Ω D 2 when the attacker selects Ω A 2 ;
(4)
When the players choose the Markov game algorithm, the profit values both increase, but the costs for both the players are also increased;
(5)
When the distributed learning algorithm is employed, the costs for two players are decreased obviously. The optimal payoff values are obtained.
Table 6 displays the different power generation outputs of MGs under different initial probabilities in the IEEE-14 system. Pg2~8 represent the generators of MG2~MG8:
(1)
When the defender chooses Ω D 1 , the power generation costs of MG2 and MG3 are low, thus the power generation is close to full output. The power generation of MG8 is lower, and MG6 keeps maintaining an isolated mode as the power generation cost is high.
(2)
When the attacker chooses Ω D 2 , the outputs of MG2 and MG6 remain at the same level of value. MG2 needs to purchase electricity to meet sudden changes in load demand, and MG8′s output decreases to prevent a sharp increase in the cost of the entire power system. It is obvious that when MG3 cannot purchase electricity, the output of MG8 increases sharply.
b.
IEEE 118-bus system
The IEEE 118-bus system includes 19 generators, 177 transmission lines, 9 transformers, and 91 loads [31]. The generator G1 is the reference bus, and G2–G19 that connected to buses 10, 12, 25, 26, 31, 46, 49, 54, 59, 61, 65, 66, 69, 80, 87, 89, 100, 103, and 111 are replaced by MGs. The game solution when the players obtain complete information is shown in Table 7.
From Table 7, it can be concluded that:
(1)
Compared with the IEEE-14 system, the effect of the distributed learning algorithm in the IEEE-118 system is better, but the computational complexity increases, thus the value of C M is higher;
(2)
The IEEE-118 system is more complex with many generators and loads, and C D and C G are higher, but the profit of the attacker is lower, and the defense effect is better;
(3)
The difference between defense cost is smaller in the IEEE-118 system, and the proposed game method is numerically better than other game methods at the equilibrium point, indicating that the method proposed in this paper is suitable for complex large systems.

5.1.2. Players with Partial Information

The optimal payoff value for the two players in partial information with system topology is shown in Table 8. The game solution is obtained in both the IEEE-14 system and the IEEE-118 system. Table 9 shows the generation outputs of MGs considering different game methods.
From Table 8 and Table 9, it can be concluded that:
(1)
When the players have partial information, payoff on both sides is reduced, as the attack is not established completely and the defense is not using the optimal strategy;
(2)
In this situation, the cost of C G is higher than C D , and defenders choose to readjust the generator output instead of shedding loads;
(3)
When the mixed strategy is chosen, the defense effect of the IEEE-118 system is better than the IEEE-14 system because the generator and load are more complex.
For maximum attack–defense benefits, the selection of the optimal defense strategy should refer to the continuous evolution of attack and defense probability. In the next subsection, the final defense strategy is established according to the ADEG model.

5.2. ADEG-Based Optimal Defense Strategy Selection

5.2.1. Effectiveness of the ADEG Model

The purpose of this case is to verify the influence of the dynamic evolution of probability on the strategy, and the selection of the optimal defense strategies under different initial probabilities shown above. Based on the evolutionary stable solutions, experimental results of the evolution trend can be obtained as in Figure 5a–j.
(1)
From Figure 5a–d, when p A = 1 or p A = 0 , the final selection probability of the attacker is always 1 or 0. Similarly when q D = 1 or q D = 0 , the final selection probability of the defender is always 1 or 0;
(2)
From Figure 5e–h, when the mixed probability strategies are selected, the stable solution will be b 22 b 21 b 11 b 21 b 12 + b 22 = 0.51 and a 22 a 21 a 11 a 21 a 12 + a 22 = 0.59 . Then, when the initial attack probability is larger than 0.51, the final selection probability of the defender will eventually change to 1; when the initial attack probability is less than 0.51, the final selection probability of the defender will eventually change to 0. Similarly for the attacker, when the initial defense probability is larger than 0.59, the final selection probability of the attacker will eventually change to 1; when the initial defense probability is less than 0.59, the final selection probability of the attacker will eventually change to 0;
(3)
To further illustrate the dynamic effect of ADEG model, and verify the probability that the two players can be affected by each other, the following scenarios are proposed as shown in Figure 5i,j:
(a)
Situation 1: the initial probability of p A and q D are selected as fixed values of 0.2 and 0.9. When t = 5 s, we make p A suddenly rise to 0.8, and observe the variation of q D ;
(b)
Situation 2: the initial probability of p A and q D are selected as fixed values of 0.5 and 0. When t = 5 s, we make q D suddenly rise to 0.9, and observe the variation of p A .

5.2.2. Final Defense Strategies Selection

From the cases, the conclusion can be drawn that the selection probability of attack and defense strategy will change in real time, and the two players will change the probability of the attack–defense strategy choice considering the current game situation. The purpose of the defender is to select the final optimal defense strategies according to the evolutionary stable solutions. Define Ω D 1 = q D = 0 , Ω D 2 = q D = 1 , then the final optimal defense strategies can be chosen as follows:
(1)
For the fixed value of the evolutionary stable solutions, the final optimal defense strategies are selected in Table 10;
(2)
For the mixed selection of initial probabilities, the final optimal defense strategies are selected in Table 11;
(3)
When the mixed selection of initial probabilities changes in real time, the final optimal defense strategies are selected in Table 12;
(4)
To illustrate the effectiveness of the ADEG model in selecting strategies, dynamic situations with a constant value of initial probabilities have been shown in Table 13.

6. Conclusions

To analyze the dynamic game behavior of attackers and defenders in the electricity markets and choose the optimal defense strategy for the defender, this paper has carried out the following work through using hybrid game methods: firstly, the power market model with MG participation while considering FDI attack is established using the Nash equilibrium method. Influence on the impact of the optimization output for MGs energy management is discussed. Meanwhile, the behaviors of both the attackers and defenders are discussed. The payoff functions of the two players are constructed and solved via the Stackelberg equilibrium algorithm. The Markov game algorithm and distributed learning algorithm are used to update the payoff function; secondly, evolutionary game theory is used to discuss the dynamic game behavior of the two players, according to the analysis of the attack–defense probability. The optimal defense strategy is selected according to the dynamic changing probability; finally, based on the evolutionary stability strategy, the final optimal defense strategies selection algorithm is designed. The strategies are considered with different initial attack probabilities and evolutionary stable solutions. Modified IEEE standard bus systems are illustrated to certify the effectiveness of the proposed model. The simulation results have shown the game relationship between the benefits of both attack and defense sides, and the important role of the optimal output of MGs. In addition, the proposed ADEG model is proved to be effective with different initial probabilities, and the optimal defense strategy can be derived from the evolutionary model.
In the future, defense strategy selection when attacks are unpredictable will be studied.

Author Contributions

B.J. wrote the manuscript, and performed the data analysis and validation; X.Z. performed the data analysis and validation; D.Y. performed the formal analysis and methodology. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Shandong Provincial Natural Science Foundation under project ZR2023QF044 (2024.1–2026.12).

Data Availability Statement

The raw/processed data required to reproduce these findings cannot be shared at this time as the data also forms part of an ongoing study.

Acknowledgments

I would like to express my sincere gratitude to Yanshan University for their support.

Conflicts of Interest

The authors declare that there are no conflicts of interest in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Shahid, K.; Nainar, K.; Olsen, R.L.; Iov, F.; Lyhne, M.; Morgante, G. On the Use of Common Information Model for Smart Grid Applications—A Conceptual Approach. IEEE Trans. Smart Grid 2021, 12, 5060–5072. [Google Scholar] [CrossRef]
  2. Habibi, M.R.; Baghaee, H.R.; Blaabjerg, F.; Dragičević, T. Secure MPC/ANN-Based False Data Injection Cyber-Attack Detection and Mitigation in DC Microgrids. IEEE Syst. J. 2022, 16, 1487–1498. [Google Scholar] [CrossRef]
  3. Li, B.; Lu, R.; Xiao, G.; Li, T.; Choo, K.-K.R. Detection of False Data Injection Attacks on Smart Grids: A Resilience-Enhanced Scheme. IEEE Trans. Power Syst. 2021, 37, 2679–2692. [Google Scholar] [CrossRef]
  4. Rodríguez, M.; Betarte, G.; Calegari, D. A Process Mining-based approach for Attacker Profiling. In Proceedings of the 2021 IEEE URUCON, Montevideo, Uruguay, 24–26 November 2021; pp. 425–429. [Google Scholar]
  5. Sandal, Y.S.; Pusane, A.E.; Kurt, G.K.; Benedetto, F. Reputation Based Attacker Identification Policy for Multi-Access Edge Computing in Internet of Things. IEEE Trans. Veh. Technol. 2020, 69, 15346–15356. [Google Scholar] [CrossRef]
  6. Rappoport, J.S. The Problem of Approach of Controlled Objects in Dynamic Game Problems with a Terminal Payoff Function. Cybern. Syst. Anal. 2020, 56, 820–834. [Google Scholar] [CrossRef]
  7. Zhang, B.; Dou, C.; Yue, D.; Park, J.H.; Zhang, Y.; Zhang, Z. Game and Dynamic Communication Path-Based Pricing Strategies for Microgrids under Communication Interruption. IEEE/CAA J. Autom. Sin. 2023, 10, 1032–1047. [Google Scholar] [CrossRef]
  8. Aydeger, A.; Manshaei, M.H.; Rahman, M.A.; Akkaya, K. Strategic Defense Against Stealthy Link Flooding Attacks: A Signaling Game Approach. IEEE Trans. Netw. Sci. Eng. 2021, 8, 751–764. [Google Scholar] [CrossRef]
  9. Liu, Z.; Wang, L. Defense Strategy Against Load Redistribution Attacks on Power Systems Considering Insider Threats. IEEE Trans. Smart Grid 2021, 12, 1529–1540. [Google Scholar] [CrossRef]
  10. Pirani, M.; Nekouei, E.; Sandberg, H.; Johansson, K.H. A Graph-Theoretic Equilibrium Analysis of Attacker-Defender Game on Consensus Dynamics under H2 Performance Metric. IEEE Trans. Netw. Sci. Eng. 2021, 8, 1991–2000. [Google Scholar] [CrossRef]
  11. Li, Y.; Bai, S.; Gao, Z. A Multi-Domain Anti-Jamming Strategy Using Stackelberg Game in Wireless Relay Networks. IEEE Access 2020, 8, 173609–173617. [Google Scholar] [CrossRef]
  12. Jakóbik, A. Stackelberg Game Modeling of Cloud Security Defending Strategy in the Case of Information Leaks and Corruption. Simul. Model. Pract. Theory 2020, 103, 102071. [Google Scholar] [CrossRef]
  13. Chen, Z.; Cui, G.; Zhang, L.; Yang, X.; Li, H.; Zhao, Y.; Ma, C.; Sun, T. Optimal Strategy for Cyberspace Mimic Defense Based on Game Theory. IEEE Access 2021, 9, 68376–68386. [Google Scholar] [CrossRef]
  14. Zhou, Y.; Cheng, G.; Zhao, Y.; Chen, Z.; Jiang, S. Toward Proactive and Efficient DDoS Mitigation in IIoT Systems: A Moving Target Defense Approach. IEEE Trans. Ind. Inform. 2022, 18, 2734–2744. [Google Scholar] [CrossRef]
  15. Zhang, Z.; Huang, S.; Chen, Y.; Li, B.; Mei, S. Cyber-Physical Coordinated Risk Mitigation in Smart Grids Based on Attack-Defense Game. IEEE Trans. Power Syst. 2022, 37, 530–542. [Google Scholar] [CrossRef]
  16. Emadi, H.; Clanin, J.; Hyder, B.; Khanna, K.; Govindarasu, M.; Bhattacharya, S. An Efficient Computational Strategy for Cyber-Physical Contingency Analysis in Smart Grids. In Proceedings of the 2021 IEEE Power & Energy Society General Meeting (PESGM), Washington, DC, USA, 26–29 July 2021; pp. 1–5. [Google Scholar]
  17. Shi, Y.; Rong, Z. Analysis of Q-Learning Like Algorithms Through Evolutionary Game Dynamics. IEEE Trans. Circuits Syst. II Express Briefs 2022, 69, 2463–2467. [Google Scholar] [CrossRef]
  18. Zhang, H.; Tan, J.; Liu, X.; Huang, S.; Hu, H.; Zhang, Y. Cybersecurity Threat Assessment Integrating Qualitative Differential and Evolutionary Games. IEEE Trans. Netw. Serv. Manag. 2022, 19, 3425–3437. [Google Scholar] [CrossRef]
  19. Chen, G.; Yu, Y. Convergence Analysis and Strategy Control of Evolutionary Games with Imitation Rule on Toroidal Grid. IEEE Trans. Autom. Control. 2023, 68, 8185–8192. [Google Scholar] [CrossRef]
  20. Zhang, B.; Dou, C.; Yue, D.; Zhang, Z.; Zhang, T. A Packet Loss-Dependent Event-Triggered Cyber-Physical Cooperative Control Strategy for Islanded Microgrid. IEEE Trans. Cybern. 2021, 51, 267–282. [Google Scholar] [CrossRef] [PubMed]
  21. Monica, P.; Kowsalya, M.; Guerrero, J.M. Logarithmic droop-based decentralized control of parallel converters for accurate current sharing in islanded DC microgrid applications. IET Renew. Power Gener. 2021, 15, 1240–1254. [Google Scholar]
  22. Mohammadhassani, A.; Teymouri, A.; Mehrizi-Sani, A.; Tehrani, K. Performance Evaluation of an Inverter-Based Microgrid Under Cyberattacks. In Proceedings of the 2020 IEEE 15th International Conference of System of Systems Engineering (SoSE), Budapest, Hungary, 2–4 June 2020. [Google Scholar] [CrossRef]
  23. Liu, C.; Zhou, M.; Wu, J.; Long, C.; Kundur, D. Financially Motivated FDI on SCED in Real-Time Electricity Markets: Attacks and Mitigation. IEEE Trans. Smart Grid 2019, 10, 1949–1959. [Google Scholar] [CrossRef]
  24. Hasankhani, A.; Hakimi, S.M. Stochastic energy management of smart microgrid with intermittent renewable energy resources in electricity market. Energy 2021, 219, 119668. [Google Scholar] [CrossRef]
  25. Razmi, P.; Buygi, M.O.; Esmalifalak, M. A Machine Learning Approach for Collusion Detection in Electricity Markets Based on Nash Equilibrium Theory. J. Mod. Power Syst. Clean Energy 2021, 9, 170–180. [Google Scholar] [CrossRef]
  26. Dou, C.; Yue, D.; Li, X.; Xue, Y. Mas-based management and control strategies for integrated hybrid energy system. IEEE Trans. Ind. Inform. 2016, 12, 1332–1349. [Google Scholar] [CrossRef]
  27. Major, J.A. Advanced techniques for modeling terrorism risk. J. Risk Financ. 2002, 4, 15–24. [Google Scholar] [CrossRef]
  28. Ma, C.Y.T.; Yau, D.K.Y.; Lou, X.; Rao, N.S. Markov Game Analysis for Attack-Defense of Power Networks under Possible Misinformation. IEEE Trans. Power Syst. 2012, 28, 1676–1686. [Google Scholar] [CrossRef]
  29. Zimmerman, R.D.; Murillo-Sanchez, C.E.; Thomas, R.J. MATPOWER: Steady-State Operations, Planning, and Analysis Tools for Power Systems Research and Education. IEEE Trans. Power Syst. 2011, 26, 12–19. [Google Scholar] [CrossRef]
  30. Foo Eddy, Y.S.; Gooi, H.B.; Chen, S.X. Multi-agent system for distributed management of microgrids. IEEE Trans. Power Syst. 2015, 30, 24–34. [Google Scholar] [CrossRef]
  31. Hassan, M.H.; Kamel, S.; El-Dabah, M.A.; Khurshaid, T.; Domínguez-García, J.L. Optimal Reactive Power Dispatch with Time-Varying Demand and Renewable Energy Uncertainty Using Rao-3 Algorithm. IEEE Access 2021, 9, 23264–23283. [Google Scholar] [CrossRef]
Figure 1. The proposed attack–defense game tree.
Figure 1. The proposed attack–defense game tree.
Symmetry 16 00156 g001
Figure 2. The replication dynamic model curve. (a) The variation of the defender when p 1 < d 22 d 21 d 11 d 21 d 12 + d 22 . (b) The variation of the defender when p 1 > d 22 d 21 d 11 d 21 d 12 + d 22 . (c) The variation of the defender when p 1 = d 22 d 21 d 11 d 21 d 12 + d 22 . (d) The variation of the attacker when q 1 > a 22 a 21 a 11 a 21 a 12 + a 22 . (e) The variation of the attacker when q 1 < a 22 a 21 a 11 a 21 a 12 + a 22 . (f) The variation of the attacker when q 1 = a 22 a 21 a 11 a 21 a 12 + a 22 .
Figure 2. The replication dynamic model curve. (a) The variation of the defender when p 1 < d 22 d 21 d 11 d 21 d 12 + d 22 . (b) The variation of the defender when p 1 > d 22 d 21 d 11 d 21 d 12 + d 22 . (c) The variation of the defender when p 1 = d 22 d 21 d 11 d 21 d 12 + d 22 . (d) The variation of the attacker when q 1 > a 22 a 21 a 11 a 21 a 12 + a 22 . (e) The variation of the attacker when q 1 < a 22 a 21 a 11 a 21 a 12 + a 22 . (f) The variation of the attacker when q 1 = a 22 a 21 a 11 a 21 a 12 + a 22 .
Symmetry 16 00156 g002
Figure 3. The selection of the optimal network defense strategy.
Figure 3. The selection of the optimal network defense strategy.
Symmetry 16 00156 g003
Figure 4. LMPs with different attack probabilities.
Figure 4. LMPs with different attack probabilities.
Symmetry 16 00156 g004
Figure 5. (ad) The evolution curve of attack and defense strategies. (a) The evolution curve with the initial probability ( p 1 = 0 , q 1 = 0 ). (b) The evolution curve with the initial probability ( p 1 = 0 , q 1 = 1 ). (c) The evolution curve with the initial probability ( p 1 = 1 , q 1 = 0 ). (d) The evolution curve with the initial probability ( p 1 = 1 , q 1 = 1 ). (eh). The evolution curve of attack and defense strategies. (e) The evolution curve with the initial probability ( p 1 = 0.4 , q 1 = 0.5 ). (f) The evolution curve with the initial probability ( p 1 = 0.4 , q 1 = 0.6 ). (g) The evolution curve with the initial probability ( p 1 = 0.6 , q 1 = 0.5 ). (h) The evolution curve with the initial probability ( p 1 = 0.6 , q 1 = 0.6 ). (i,j). Real-time evolution curve of attack and defense strategies. (i) The evolution curve in situation 1. (j) The evolution curve in situation 2.
Figure 5. (ad) The evolution curve of attack and defense strategies. (a) The evolution curve with the initial probability ( p 1 = 0 , q 1 = 0 ). (b) The evolution curve with the initial probability ( p 1 = 0 , q 1 = 1 ). (c) The evolution curve with the initial probability ( p 1 = 1 , q 1 = 0 ). (d) The evolution curve with the initial probability ( p 1 = 1 , q 1 = 1 ). (eh). The evolution curve of attack and defense strategies. (e) The evolution curve with the initial probability ( p 1 = 0.4 , q 1 = 0.5 ). (f) The evolution curve with the initial probability ( p 1 = 0.4 , q 1 = 0.6 ). (g) The evolution curve with the initial probability ( p 1 = 0.6 , q 1 = 0.5 ). (h) The evolution curve with the initial probability ( p 1 = 0.6 , q 1 = 0.6 ). (i,j). Real-time evolution curve of attack and defense strategies. (i) The evolution curve in situation 1. (j) The evolution curve in situation 2.
Symmetry 16 00156 g005aSymmetry 16 00156 g005bSymmetry 16 00156 g005c
Table 1. Selection of the attack and defense strategies for solutions.
Table 1. Selection of the attack and defense strategies for solutions.
RDE SolutionsDefense StrategyWith
Probability
Attack StrategyWith
Probability
p A 1 = 0 , q D 1 = 0 Ω D 1 0 Ω A 1 0
p A 1 = 0 , q D 1 = 1 Ω D 2 1 Ω A 1 0
p A 1 = 1 , q D 1 = 0 Ω D 1 0 Ω A 2 1
p A 1 = 1 , q D 1 = 1 Ω D 2 1 Ω A 2 1
p A 1 = b 22 b 21 b 11 b 21 b 12 + b 22 , q D 1 = a 22 a 21 a 11 a 21 a 12 + a 22 Ω D 1 ,
Ω D 2
( a 22 a 21 a 11 a 21 a 12 + a 22 , 1 a 22 a 21 a 11 a 21 a 12 + a 22 ) Ω A 1 , Ω A 2 ( b 22 b 21 b 11 b 21 b 12 + b 22 , 1 b 22 b 21 b 11 b 21 b 12 + b 22 )
Table 2. Attack strategies selection.
Table 2. Attack strategies selection.
Initial Attack StrategiesAttacked LinePower Flow (MW)Attack Level (%)Initial Probability
Without AttackWith Attack
Ω A 1 2–354.727.3550 p A = 1
Ω A 2 2–3, 4–554.749.310 p A = 0
Ω A 1 , Ω A 2 2–3, 4–554.743.920 p A = 0.4
Ω A 1 , Ω A 2 2–354.716.4180 p A = 0.6
Table 3. Defense strategies selection.
Table 3. Defense strategies selection.
Initial Defense StrategiesDetection Mode SelectionInitial Probability
MGs in Grid-Connected ModeLoad Shedding
Ω D 1 withwithout q D = 1
Ω D 2 withoutwith q D = 0
Ω D 1 , Ω D 2 withwith q D = 0.5
Ω D 1 , Ω D 2 withwith q D = 0.6
Table 4. Final detection result with ratio of 1:1.
Table 4. Final detection result with ratio of 1:1.
Initial Attack StrategiesDetection DelayDetection Rate
Ω A 1 2885.03%
Ω A 2 1290.98%
Ω A 1 , Ω A 2 992.01%
Ω A 1 , Ω A 2 1095.35%
Table 5. Optimal attack–defense payoff value in 14-bus.
Table 5. Optimal attack–defense payoff value in 14-bus.
Initial Probabilities Attack   Payoff   $ Defense   Payoff   $
P r o f i t a t t E L a t t C D   +   C G C M
Stackelberg method p A = 0 q D = 0 29.386.3527.32 + 0.855.06
p A = 0 q D = 1 59.827.3137.01 + 20.2112.10
p A = 1 q D = 0 67.3313.0632.09 + 10.5014.88
p A = 1 q D = 1 36.4812.4548.53 + 1.219.35
Markov
Algorithm
p A = 0 q D = 0 34.418.2528.44 + 0.915.73
p A = 0 q D = 1 63.128.8139.33 + 20.4513.00
p A = 1 q D = 0 37.5212.5049.44 + 1.8110.01
p A = 1 q D = 1 69.4113.8633.09 + 11.2314.90
Proposed algorithm p A = 0 q D = 0 33.317.0527.00 + 0.715.01
p A = 0 q D = 1 62.057.0236.52 + 19.1111.99
p A = 1 q D = 0 36.8810.5046.55 + 1.059.12
p A = 1 q D = 1 68.2211.1230.09 + 9.1512.76
Table 6. MG outputs (MW) under different initial probabilities in 14-bus.
Table 6. MG outputs (MW) under different initial probabilities in 14-bus.
Pg2Pg3Pg6Pg8
no attack28.0128.920.0411.93
p A = 0 , q D = 0 2833.970.017.16
p A = 0 , q D = 1 2825.320.0115.93
p A = 1 , q D = 0 2840.010.011.72
p A = 1 , q D = 1 2830011.94
Table 7. Optimal attack–defense payoff value in 118-bus.
Table 7. Optimal attack–defense payoff value in 118-bus.
Initial ProbabilitiesAttack Payoff $ Defense Payoff $
Profit a t t EL a t t C D + C G C M
Stackelberg method p A = 0 q D = 0 217.4240.35200.93 + 8.6237.21
p A = 0 q D = 1 415.3146.52215.66 + 140.3776.06
p A = 1 q D = 0 288.8879.31274.59 + 9.3340.15
p A = 1 q D = 1 428.5180.92212.11 + 70.3880.75
Markov
Algorithm
p A = 0 q D = 0 235.7842.62202.11 + 9.3138.40
p A = 0 q D = 1 438.4248.11227.72 + 142.1578.11
p A = 1 q D = 0 310.5380.56280.30 + 10.1542.53
p A = 1 q D = 1 451.6683.33223.11 + 70.5681.15
Proposed algorithm p A = 0 q D = 0 230.6930.00189.12 + 7.7846.35
p A = 0 q D = 1 421.8332.45205.38 + 130.9678.12
p A = 1 q D = 0 300.9668.91260.45 + 8.8948.49
p A = 1 q D = 1 440.7270.83208.30 + 68.5589.70
Table 8. Optimal attack–defense payoff value.
Table 8. Optimal attack–defense payoff value.
Initial Probabilities Attack   Payoff   $ Defense   Payoff   $
P r o f i t a t t E L a t t C D   +   C G C M
IEEE-14 system p A = 0.4 q D = 0.5 28.125.1124.00 + 6.157.12
p A = 0.4 q D = 0.6 36.334.8928.30 + 23.5213.00
p A = 0.6 q D = 0.5 29.008.3244.12 + 8.3511.54
p A = 0.6 q D = 0.6 40.159.0629.11 + 15.3014.98
IEEE-118 system p A = 0.4 q D = 0.5 189.4520.83170.35 + 14.5650.34
p A = 0.4 q D = 0.6 405.6121.42200.11 + 160.5084.12
p A = 0.6 q D = 0.5 289.3360.35230.25 + 16.3550.66
p A = 0.6 q D = 0.6 410.5468.44198.45 + 82.3194.30
Table 9. MG outputs (MW) under different initial probabilities in 118-bus.
Table 9. MG outputs (MW) under different initial probabilities in 118-bus.
Pg2Pg3Pg6Pg8
no attack28.0128.920.0411.93
p A = 0.4 , q D = 0.5 2829.990.0112.58
p A = 0.4 , q D = 0.6 28.0129.990.0413.15
p A = 0.6 , q D = 0.5 2840.010.0111.68
p A = 0.6 , q D = 0.6 28.0140.010.0414.10
Table 10. Optimal defense strategies selection 1.
Table 10. Optimal defense strategies selection 1.
Initial ProbabilitiesEvolutionary Stable SolutionsFinal Defense Strategies
p A = 0 , q D = 0 p A = 0 , q D = 0 Ω D 1
p A = 0 , q D = 1 p A = 0 , q D = 1 Ω D 2
p A = 1 , q D = 0 p A = 1 , q D = 0 Ω D 1
p A = 1 , q D = 1 p A = 1 , q D = 1 Ω D 2
Table 11. Optimal defense strategies selection 2.
Table 11. Optimal defense strategies selection 2.
Initial ProbabilitiesEvolutionary Stable SolutionsFinal Defense Strategies
p A = 0.4 , q D = 0.5 p A = 0 , q D = 0 Ω D 1
p A = 0.4 , q D = 0.6 p A = 0 , q D = 1 Ω D 2
p A = 0.6 , q D = 0.5 p A = 1 , q D = 0 Ω D 1
p A = 0.6 , q D = 0.6 p A = 1 , q D = 1 Ω D 2
Table 12. Optimal defense strategies selection 3.
Table 12. Optimal defense strategies selection 3.
Initial ProbabilitiesProbabilities ChangeFinal Defense Strategies
p A = 0.4 , q D = 0.5 p A = 0.5 0.9 Ω D 1 Ω D 2
p A = 0.4 , q D = 0.6 p A = 0.6 0.2 Ω D 2 Ω D 1
p A = 0.6 , q D = 0.5 p A = 0.5 0.9 Ω D 1 Ω D 2
p A = 0.6 , q D = 0.6 p A = 0.6 0.2 Ω D 2 Ω D 1
Table 13. Optimal defense strategies selection 4.
Table 13. Optimal defense strategies selection 4.
Initial ProbabilitiesProbabilities ChangeFinal Defense Strategies
p A = 0.4 (fixed value), q D = 0.5 p A = 0.4 0.9 Ω D 1 Ω D 2
p A = 0.4 (fixed t value), q D = 0.6 p A = 0.4 0.1 Ω D 1 Ω D 1
p A = 0.6 (fixed value), q D = 0.5 p A = 0.6 0.1 Ω D 2 Ω D 1
p A = 0.6 (fixed value), q D = 0.6 p A = 0.6 0.9 Ω D 2 Ω D 2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jin, B.; Zhao, X.; Yuan, D. Attack–Defense Confrontation Analysis and Optimal Defense Strategy Selection Using Hybrid Game Theoretic Methods. Symmetry 2024, 16, 156. https://doi.org/10.3390/sym16020156

AMA Style

Jin B, Zhao X, Yuan D. Attack–Defense Confrontation Analysis and Optimal Defense Strategy Selection Using Hybrid Game Theoretic Methods. Symmetry. 2024; 16(2):156. https://doi.org/10.3390/sym16020156

Chicago/Turabian Style

Jin, Bao, Xiaodong Zhao, and Dongmei Yuan. 2024. "Attack–Defense Confrontation Analysis and Optimal Defense Strategy Selection Using Hybrid Game Theoretic Methods" Symmetry 16, no. 2: 156. https://doi.org/10.3390/sym16020156

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop