
On the whole, the present microgrid constitutes numerous actors in highly decentralized environments and liberalized electricity markets. The networked microgrid system must be capable of detecting electricity price changes and unknown variations in the presence of rare and extreme events. The networked microgrid system comprised of interconnected microgrids must be adaptive and resilient to undesirable environmental conditions such as the occurrence of different kinds of faults and interruptions in the main grid supply. The uncertainties and stochasticity in the load and distributed generation are considered. In this study, we propose resilient energy trading incorporating DC-OPF, which takes generator failures and line outages (topology change) into account. This paper proposes a design of Long Short-Term Memory (LSTM) - soft actor-critic (SAC) reinforcement learning for the development of a platform to obtain resilient peer-to-peer energy trading in networked microgrid systems during extreme events. A Markov Decision Process (MDP) is used to develop the reinforcement learning-based resilient energy trade process that includes the state transition probability and a grid resilience factor for networked microgrid systems. LSTM-SAC continuously refines policies in real-time, thus ensuring optimal trading strategies in rapidly changing energy markets. The LSTM networks have been used to estimate the optimal Q-values in soft actor-critic reinforcement learning. This learning mechanism takes care of the out-of-range estimates of Q-values while reducing the gradient problems. The optimal actions are decided with maximized rewards for peer-to-peer resilient energy trading. The networked microgrid system is trained with the proposed learning mechanism for resilient energy trading. The proposed LSTM-SAC reinforcement learning is tested on a networked microgrid system comprised of IEEE 14 bus systems.
Citation: Desh Deepak Sharma, Ramesh C Bansal. LSTM-SAC reinforcement learning based resilient energy trading for networked microgrid system[J]. AIMS Electronics and Electrical Engineering, 2025, 9(2): 165-191. doi: 10.3934/electreng.2025009
[1] | Dipak R. Swain, Sunita S. Biswal, Pravat Kumar Rout, P. K. Ray, R. K. Jena . Optimal fractional sliding mode control for the frequency stability of a hybrid industrial microgrid. AIMS Electronics and Electrical Engineering, 2023, 7(1): 14-37. doi: 10.3934/electreng.2023002 |
[2] | Olayanju Sunday Akinwale, Dahunsi Folasade Mojisola, Ponnle Akinlolu Adediran . Mitigation strategies for communication networks induced impairments in autonomous microgrids control: A review. AIMS Electronics and Electrical Engineering, 2021, 5(4): 342-375. doi: 10.3934/electreng.2021018 |
[3] | Sebin J Olickal, Renu Jose . LSTM projected layer neural network-based signal estimation and channel state estimator for OFDM wireless communication systems. AIMS Electronics and Electrical Engineering, 2023, 7(2): 187-195. doi: 10.3934/electreng.2023011 |
[4] | D Venkata Ratnam, K Nageswara Rao . Bi-LSTM based deep learning method for 5G signal detection and channel estimation. AIMS Electronics and Electrical Engineering, 2021, 5(4): 334-341. doi: 10.3934/electreng.2021017 |
[5] | Vuong Quang Phuoc, Nguyen Van Dien, Ho Duc Tam Linh, Nguyen Van Tuan, Nguyen Van Hieu, Le Thai Son, Nguyen Tan Hung . An optimized LSTM-based equalizer for 100 Gigabit/s-class short-range fiber-optic communications. AIMS Electronics and Electrical Engineering, 2024, 8(4): 404-419. doi: 10.3934/electreng.2024019 |
[6] | K.V. Dhana Lakshmi, P.K. Panigrahi, Ravi kumar Goli . Machine learning assessment of IoT managed microgrid protection in existence of SVC using wavelet methodology. AIMS Electronics and Electrical Engineering, 2022, 6(4): 370-384. doi: 10.3934/electreng.2022022 |
[7] | Mehdi Tabasi, Pouyan Asgharian . Optimal operation of energy storage units in distributed system using social spider optimization algorithm. AIMS Electronics and Electrical Engineering, 2019, 3(4): 309-327. doi: 10.3934/ElectrEng.2019.4.309 |
[8] | Saumen Dhara, Pradip Kumar Sadhu, Alok Kumar Shrivastav . Analysis of over current relay and hybrid filter including the utilization of SFCL in a distribution network with DG. AIMS Electronics and Electrical Engineering, 2022, 6(1): 81-107. doi: 10.3934/electreng.2022006 |
[9] | Assila Yousuf, David Solomon George . A hybrid CNN-LSTM model with adaptive instance normalization for one shot singing voice conversion. AIMS Electronics and Electrical Engineering, 2024, 8(3): 292-310. doi: 10.3934/electreng.2024013 |
[10] | Cherechi Ndukwe, M. Tariq Iqbal, Xiaodong Liang, Jahangir Khan, Lawrence Aghenta . LoRa-based communication system for data transfer in microgrids. AIMS Electronics and Electrical Engineering, 2020, 4(3): 303-325. doi: 10.3934/ElectrEng.2020.3.303 |
On the whole, the present microgrid constitutes numerous actors in highly decentralized environments and liberalized electricity markets. The networked microgrid system must be capable of detecting electricity price changes and unknown variations in the presence of rare and extreme events. The networked microgrid system comprised of interconnected microgrids must be adaptive and resilient to undesirable environmental conditions such as the occurrence of different kinds of faults and interruptions in the main grid supply. The uncertainties and stochasticity in the load and distributed generation are considered. In this study, we propose resilient energy trading incorporating DC-OPF, which takes generator failures and line outages (topology change) into account. This paper proposes a design of Long Short-Term Memory (LSTM) - soft actor-critic (SAC) reinforcement learning for the development of a platform to obtain resilient peer-to-peer energy trading in networked microgrid systems during extreme events. A Markov Decision Process (MDP) is used to develop the reinforcement learning-based resilient energy trade process that includes the state transition probability and a grid resilience factor for networked microgrid systems. LSTM-SAC continuously refines policies in real-time, thus ensuring optimal trading strategies in rapidly changing energy markets. The LSTM networks have been used to estimate the optimal Q-values in soft actor-critic reinforcement learning. This learning mechanism takes care of the out-of-range estimates of Q-values while reducing the gradient problems. The optimal actions are decided with maximized rewards for peer-to-peer resilient energy trading. The networked microgrid system is trained with the proposed learning mechanism for resilient energy trading. The proposed LSTM-SAC reinforcement learning is tested on a networked microgrid system comprised of IEEE 14 bus systems.
Renewable Energy (RE) is vital to building a resilient and secure future energy system. The penetration of distributed energy resources (DER) into the entire power system facilitates the liberalization of electricity markets, the availability of standby capacity for peak demand, the enhancement in reliability and power quality, the augmentation of the local electricity network, the support to the existing grid, the combined generation of electricity and heat, the efficient use of low-priced fuel, etc. [1]. Now, one question that arises is, "What should be the business and technology platforms that can manage the variability of a generation of DER, diversity, and complexity in the transfer of electricity to end-users?" [2]. New technologies, which are also to be developed, should have capabilities to interface the augmented power system infrastructure with power market stakeholders. System operators and other entities who deal with electricity markets should possess the controllability and visibility of different DERs; moreover, there should be requisite interfacing between these system components to achieve optimized monitoring and control operations [3]. New acceptable business models need to be created along with an expansion in power system elements due to trading entities as per the repercussion of the enhancement in investment in DER, wholesale electricity market, demand response programs, energy storages, and plug-in-electric vehicles (PEVs) technologies. Consumers are gaining a prime importance in whole power system developments due to the following objectives: a) to estimate and predict the proportion of customer participation in demand reduction; b) to identify the localized demand response and its impact on a utilities' distribution system; c) to include the demand response and distributed energy resources into the utilities' action plan; and d) to identify the effect of the reduction of load on the utilities' procurement plan [4,5].
There must be some ways to utilize demand reduction by consumers during peak and non-peak hours. The possibilities are to be examined such that a DER and a combination of DER at the local level not affecting that does not affect the power grid can have the generation capacity to cater to the consumers' demand, at most, during peak hours, and to identify whether the production capacity of DER is surplus during peak hours and non-peak hours [6]. It is a challenging task to sum up the surplus power generation of different DERs that posses the stochastic generation capacity. It is not clear whether and by which ways it is possible to utilize the aggregated power available from dispersedly interconnected DER and demand reduction of numerous heterogeneous loads and then postpone new power system infrastructure developments with an uprise of the power demand. The aggregated power from DERs should be available for sale in electricity markets with many bids/offers [7].
In the cyber-physical system of microgrids, distributed control schemes are found to be better in solving economic dispatch problems compared to central control schemes because distributed control schemes are robust, easily scalable, and possess a lower cost of implementation. Though distributed control schemes show advantages over central control schemes, these are prone to cyber-attacks. Therefore, to mitigate the effects of cyber-attacks, an attack-robust distributed economic dispatch strategy is designed. In a transactive energy framework, microgrid (MG) aims to schedule an optimal hourly strategy in the day-ahead market to maximize profits. Next, MGs try to minimize the imbalance of cost in the real-time market. The installation of a microgrid provides a lucrative solution to dependency on the supply of power from the main grid and obtains economic benefits from locally generated power [8]. The basic architecture of the microgrid is shown in Figure 1.
Attack-resilient intelligent power management has been developed such that the efficient operation of an emergency power system in the presence of a cyber-attack can be ensured. An adaptive neuro-fuzzy inference system (ANFIS) based methodology can validate the integrity of critical data about an energy management system and is capable of detecting the occurrence of cyber-attacks [9]. The various energy management systems (EMS) available in the literature have pros and cons based on the minimum operating cost, customer privacy, flexibility, computation power reduction, reliability, and resiliency. In islanded mode, a microgrid operates independently; hence, there is a possibility that the microgrid may become less tolerant to faults and its resiliency gets affected.
A mechanism has to be developed to improve the resiliency of the independent microgrid [10]. By promoting the dispersion of power resources, the resiliency of the microgrid can be improved [11]. The feasibility of the resilient operation is to be analyzed with three actions unit commitments, energy storage schedules, load curtailment, and adjustable load schedules. The resiliency-oriented optimal scheduling model for the microgrid was developed in [12]. A mechanism must be developed to ensure the economically optimal operation, a robustness against uncertainties present in the system, and a fast-islanding operation with minimum consumer inconvenience. The mathematical modeling of optimal scheduling that considers resiliency for microgrids is not widely available. To build a resilient and environment-friendly microgrid that emphasizes customers, an intelligent and distributed autonomous framework was developed [13]. In this model, the occurrence of major outages due to natural disasters such as floods, tsunamis, earthquakes, heavy rains, etc. was considered. The severity and occurrence of these events may increase in the future [14]. In the grid-connected mode, a proactive operation model was implemented while considering a scheduling horizon of 24 hours. This proactive scheme works on initial warnings created well before the occurrence of major outages. The resilient-oriented operation scheme for microgrids considers the major outages. A cyber-physical system of the multi-energy system is prone to Denial of Service (DoS) attacks; consequently, the implementation of distributed energy management algorithms is affected [15]. In the load-shedding algorithm, the possibility of maximum load curtailment has to be identified. In the ramp-down algorithm, the maximum possible generation and the remaining amount of surplus power at different time instants has to be identified.
The approach of reinforcement learning (RL) is goal-directed while learning from the given environment. In reinforcement learning, the agents are software programs. First, these agents discern the environment and make decisions accordingly for an action. The actions of the agents must be optimal and superior. The environment of an RL space should be either deterministic, observable, discrete, or continuous, and either single or multi-agent. The variables of the RL algorithm known as hyperparameters are set for the model. These are different from the parameters of the model. The hyperparameters are epsilon, alpha, and gamma, where epsilon is a greedy factor, alpha is the learning rate, and gamma is the discount factor [16,17]. RL-based energy management has been implemented in the creation of an energy management system for an electrified powertrain [18]. The design of a resilient multi-energy microgrid system faces the challenges of stochastic uncertainties of renewable generation, the development of model-free control schemes under undesirable conditions, and the achievement of robust and efficient operations [19]. A multi-agent deep RL algorithm has been designed for the resilience-driven routing and scheduling operations of the mobile energy storage system [20]. The resilience quantification and planning of the power distribution grid were analyzed with a zigzag topological approach [21].
In order to achieve several goals, such as lowering operating expenses and guaranteeing power supply dependability, a deep reinforcement learning-based energy scheduling approach is suggested [22]. To enhance the control performance of energy management system (EMS), SAC is applied to the EMS of an electric vehicle (EV) equipped with a hybrid energy storage system (HESS) [23]. The deep reinforcement learning algorithm is used to identify and calibrate problematic parameters using PMU measurements [24]. By maintaining the variance of the state-action returns within a tolerable range, a deep off-policy actor-critic variation is proposed to learn a continuous return distribution [25]. In order to maximize their financial benefits and enhance system reliability, a study suggests a residential demand response strategic bidding approach for load aggregators with deep reinforcement learning [26]. To harness energy flexibility, a major commercial building's cooling setpoint has been controlled using deep reinforcement learning based on Soft Actor Critic [27]
The components and stakeholders of a cyber-physical system of microgrids are shown in Figure 1. In a networked microgrid, a resilient energy trading system is crucial to ensure the energy security and price stability, especially as the world faces increasing climate risks and the transition to more sustainable energy sources. A networked microgrid consists of multiple interconnected microgrids that can operate either autonomously or in coordination with a main power grid. This structure enhances the resilience, flexibility, and efficiency in energy management. Peer-to-peer energy trading in networked microgrids enables prosumers (producers + consumers) to trade excess energy directly with consumers. When resilience is incorporated, the trading mechanism ensures the fault tolerance, cybersecurity, and adaptability against disruptions (e.g., cyberattacks, extreme weather, or grid failures), as seen in Figure 2.
In energy trading, resilience is key to ensuring the continued flow of electricity, even in the face of disruptions such as supply shocks, demand volatility, technological changes, or environmental challenges. To build resilience in energy trading, the following strategies must be adopted:
● Sourcing energy from a variety of regions and suppliers to reduce the dependence on any one source;
● Using AI and machine learning to predict market trends and quickly respond to disruptions;
● Modernizing grids and storage solutions to effectively handle fluctuations in the supply and demand; and
● Creating policies that encourage market transparency and provide support during crises.
Resilient energy trading ensures stable, secure, and adaptive transactions, even under uncertain conditions. However, several factors can impact its efficiency and robustness. These causes can be categorized into technical, economic, cyber, environmental, and regulatory factors. Resilient energy trading is influenced by multiple factors, including technical, economic, cyber, and environmental challenges, as shown in Table 1.
Sr No | Fault Type | Impact on trading |
1 | Line Faults | Power congestion, price fluctuations |
2 | Generator Failure | Demand-supply imbalance, blackouts |
3 | Bus Outages | Complete trading failure, cascading blackouts |
4 | Cyber Attacks | Price manipulation, fake trading data |
A. Impact on Trading due to Transmission line faults (Short Circuits, Overloads)
● Power flow disruptions affect the energy prices and trading stability.
● Some buses become isolated, thus causing supply shortages.
● Increased congestion in unaffected transmission lines.
B. Impact on Trading due to Generator Failure (Bus Outage, Frequency Instability)
● Sudden loss of generation capacity causes price spikes.
● Microgrid-dependent areas face severe power shortages.
● Market participants hoard energy, thus leading to unfair pricing.
The Long Short-Term Memory (LSTM) network is part of the deep recurrent neural network (RNN) class with backpropagation and does not use gradients. Various memory gates are added in the LSTM network, which possess the blocks that are comprised of gates to manage the block's state and output [28,29]. The text classification tasks have been solved using optimal LSTM topologies. The soft actor-critic (SAC) algorithm, including maximum entropy with LSTM, has been considered for the energy management of multi-energy systems [30]. The workload prediction model was designed considering the LSTM model [31]. In time series forecasting, the LSTM model was used in [32]. The spiking neural P (SNP) based LSTM model was created to process sequential data [33]. The stock market values were predicted using the LSTM model [34]. A method was suggested to predict the parts with the highest frequency with the Random Forest (RF), while LSTM predicted the remaining parts [35]. A multi-scale FCN (MFCN) and LSTM network were considered for learning spatial and temporal features [36]. In the error detection of electroencephalography (EEG), a bidirectional LSTM neural network was implemented. The electrocardiogram signals were partitioned into normal and abnormal using LSTM [37]. The multi-series classification has been performed with a full convolution network- LSTM [38]. The levelized cost of electricity (LCOE) and payback period approach was used to assess the solar-powered microgrid's economic viability [39]. A technique for transferring tokens between networked microgrids with interoperable blockchains has been created. A technique for token transfers between networked microgrid interoperable blockchains has been devised for safe and private smart power contracts between electricity suppliers and customers [40]. Under grid-to-vehicle and vehicle-to-grid systems, the electric vehicle network hub enables blockchain-based safe and robust energy trading [41].
Resilient energy trading ensures the secure, adaptive, and efficient energy exchange in decentralized networks. However, several research gaps hinder its full potential. The major points in the research gap are highlighted below:
● Existing optimization models fail to adapt to dynamic price fluctuations, grid failures, and adversarial conditions;
● Current energy trading frameworks struggle with data privacy, non-independent and identically distributed (IID) data distribution, and communication failures;
● Vulnerabilities to DDoS attacks, data poisoning, and blockchain smart contract exploits threaten the resilience of energy trading;
● Current models do not address climate-induced disruptions (e.g., wildfires, hurricanes) that affect the energy trading resilience;
● There is a lack of holistic models that integrate renewable energy, microgrids, and energy storage; and
● Limited research has been done on the design of quantification techniques for real-time energy resilience.
Addressing these research gaps requires integrating RL, federated learning, blockchain security, and real-world validation. Advancing resilient energy trading frameworks will enhance the stability, security, and efficiency in decentralized energy markets.
In this paper, LSTM-soft actor-critic (SAC) RL-based resilient energy management is suggested for a networked microgrid system. In the networked microgrid system, the peer-to-peer resilient energy trading mechanism is adaptive and robust to undesirable scenarios such as the occurrence of faults in transmission lines, the failure of generators, and the failure of supply from the main grid. The LSTM-SAC reinforcement learning is proposed to train the networked MG system to make it resilient and adaptive. The state observations comprise the expected values of the electrical load demand, power generations by the distributed generators, and the state of charge (SoC) and state of health (SoH) of energy storage devices and electric vehicles. The action part contains either buying or selling rates for surplus and deficit power, charging and discharging rates of energy storage devices, and electric vehicles. The LSTM-SAC RL finds the actions for optimal Q values that maximize the reward for different scenarios of undesirable events in a microgrid.
The proposed LSTM-SAC for Resilient Energy Trading is comprised of the following major features, particularly in dynamic and uncertain energy markets:
● Unlike standard SAC, integrating LSTM enables the model to learn from past energy trading patterns, grid conditions, and demand fluctuations;
● SAC-LSTM recognizes trends and adjusts strategies based on historical sequences, thus improving decision-making in volatile energy markets;
● Multiple microgrids use SAC-LSTM, and they dynamically learn optimal trading strategies while ensuring resilience against failures; and
● By considering past failures and recovery mechanisms, SAC-LSTM enhances the resilience of energy trading strategies.
The major contributions of this paper are highlighted below:
● This paper formulates the resilient energy trading process by incorporating DC-OPF that considers line outages (topology change) and generator failures;
● The proposed LSTM-SAC network does not use the gradient descent technique to estimate Q-values, while the policy is maximized and avoids overestimates and underestimates of Q-values;
● The networked microgrid is trained with the proposed LSTM-SAC algorithm that formulates the reward function comprised of microgrid losses and profits, which are determined based on the surplus and deficit power and the time of occurrence of fault; and
● A Markov Decision Process (MDP) is used to develop the RL-based resilient energy trade problem.
This paper is organized as follows: Section 1 covers the introduction; Section 2 includes the basic concepts of RL; Section 3 focuses on the resilient energy trading problem formulation; Section 4 discusses RL-based secure and resilient energy trading; Section 5 covers the proposed LSTM-SAC RL scheme for resilient energy trading systems; Section 6 includes a simulation and results; and Section 7 concludes the paper.
The approach of RL is goal-directed while learning from the given environment. In RL, the agents are software programs. First, these agents discern the environment and make decisions accordingly for an action. The actions of the agents must be optimal and best. The environment of the RL space should be deterministic, observable, discrete or continuous, and single or multi-agent. There are four main parts of RL such as a policy, a reward signal, a value function, and/or a model function [42,43].
a. Policy: A policy is the learning process and behavior of the agent in the environment. A policy defines the action to be taken out of discerned states from the environment. The policy is the core part of the RL and maybe a simple function or a look-up table.
b. Reward: The reward is the goal of the RL problem. At each discrete time step, an agent receives a reward, which is a number from the environment. The objective of the agent is to maximize its reward. The reward the agents receive depends on the good or bad events. The reward of any number to the agent depends on the stochastic functions of the environment and the agent's actions.
c. Value function: A value function indicates the good things that will occur in the long run. The value function depends on an agent who expects to collect the reward in future actions. The environment directly gives a reward to the agents with an estimation of values on the actions during their entire lifetime.
RL is based on mathematical principles from MDPs, probability, optimization, and dynamic programming. An MDP is defined by a tuple (S,A,P,R,γ),
where S is the set of state, A is the set of actions, P(s′|s,a) is the transition probability from state s to s′ given action, r(s,a) is the reward function, and γ is the discount factor (0 ≤γ≤1) .
The state value function is shown below:
Vπ(s)∈Eπ[∑∞t=0γtr(st,at)|s0=s]. | (1) |
It represents the expected cumulative reward when starting from state s and following the policy.
The action –value function is shown below:
Qπ(s,a)=Eπ[∑∞t=0γtr(st,at)|s0=s,a0=a]. | (2) |
It represents the expected cumulative reward starting from state s , taking action a , and then following the policy π . The Bellman equation expresses the recursive relationship of the value function:
Vπ(s)=∑aπ(a|s)∑s′P(s′|s,a)[r(s,a)+γVπ(s′). | (3) |
The Bellman Equation for the Q-Function is as follows:
Qπ(s,a)=r(s,a)+γ∑s′P(s′|(s,a)∑a′π(a′|s′)Qπ(s′,a′). | (4) |
For the optimal policy π∗ , the value function satisfies the following:
V∗(s)=maxaQ∗(s,a). | (5) |
The optimal Q-function satisfies the Bellman optimality equation:
Q∗(s,a)=r(s,a)+γ∑s′P(s′|s,a)a′maxQ∗(s′,a′). | (6) |
Using the Q-function, the policy is improved using the following:
π′(s)=argmaxaQπ(s,a). | (7) |
This interprets the action that maximizes the expected reward. The temporal difference learning updates the value function as follows:
V(s)←V(s)+α(r+γV(s′)−V(s)), | (8) |
where α is the learning rate. Q-learning is an off-policy learning method with the following update rule:
Q(s,a)←Q(s,a)+α(r+γmaxa′Q(s′,a′)−Q(s,a)), | (9) |
where maxa′Q(s′,a′) represents the best possible future value.
An adaptive neuro-fuzzy inference system is developed to make the attack-resilient energy management of a microgrid. In a multi-energy system, the optimal operation of each energy hub and optimal energy trading path are considered during DoS attacks [9]. A load shedding of a non-critical load is implemented in a microgrid that is islanded after a disaster in a networked microgrid system [10]. In the islanded mode, a microgrid gets disconnected from the main grid and, consequently, becomes less tolerant to faults; then, the resiliency of the microgrid is affected [11,12]. The system must be smart enough to anticipate such as electricity price changes and unknown fluctuations. However, some events are rare and extreme, which deeply affects the system. Resiliency is described as the capability of the system to adapt itself during the occurrence of these rare and extreme events and be robust enough to these events [13]. A resiliency-oriented microgrid optimal scheduling model has been developed. The operating model employed a normal and resilient operation. The capability of the microgrid has to be identified to supply loads during main grid supply interruptions. The networked microgrid system must be robust and adaptive to operational uncertainties of loads and non-dispatchable generation. The worst-case scenarios with prevailing uncertainties are to be considered to make the system robust and resilient [14,15].
Considering the uncertainty of the DG, the total generation cost of MG k is given below:
TCk(t)=∑i∈NgCki(E(Pkg,i(t)), | (10) |
where Ng is the total number of distributed generating units in the network microgrid system, T is the total number of time intervals, E(Pkg,i(t))=ˉPkg,i(t)+ΔPkg,i(t) is the expected value of power generation, and ˉPkg,i(t) is the forecasted value of power generation by DG i.
ΔPkg,i<max(ΔPk,g,i) represents an uncertainty part of DG i and is limited by a maximum value in the worst-case scenario.
The operating limits (superscript k is dropped for simplicity) are as follows:
Pming,i<Pg,i<Pmaxg,i,i∈Ng, | (11) |
Pg,i(t)−Pg,i(t−1)≤Δup, | (12) |
Pg,i(t−1)−Pg,i(t)≤Δdown, | (13) |
∑i∈NgPg,i≥Lt, | (14) |
where Lt is the total demand. The state of charge of the energy storage system is as follows:
SOCbt,i(t)=SOCbt,i(t−1)−(Pcbt,i(t)ηc−Pdbt,i(t)(t)ηd)ΔT, | (15) |
where Pcbt and Pdbt are the charging and discharging power of energy storage devices, respectively, with constraints Pcmin≤Pcbt(t)≤Pcmax and Pcmin≤Pcbt(t)≤Pcmax .
The state of charge of an electric vehicle is shown below:
SOCev,i(t)=SOCev,i(t−1)−(Pcev,i(t)ηc−Pdev,i(t)ηd)ΔT, | (16) |
where Pcev and Pdev are the charging and discharging power of electric vehicles, respectively, with the constraints Pcmin≤Pcev(t)≤Pcmax and Pdmin≤Pdev(t)≤Pdmax .
It is required to completely charge the battery after different operations in a day so that a fully charged battery will be available for the next day's operations. To obtain a fully charged battery, a set of intervals is reserved for charging, as shown below:
Ωrc={Tc,Tc+1,…….,T}whereTc=T−Nfc, | (17) |
where Nfc is the cardinality of set Ωrc and can be found as follows:
q={(SOCmax−SOCmin)Ebt,maxηchargePbt,max | (18) |
andNfc=⌈q⌉. | (19) |
The operation and maintenance cost in a microgrid is formulated as follows:
COM(t)=CPVOMPpv(t)+CWTOMPwt(t)+CMgTOMPMgT(t)+CesOMPes(t). | (20) |
At a particular time instant t , the surplus or deficit power in MG is described as follows, while the uncertainty of the system is considered. At time instant t , let ΔPk be the surplus or deficit power in MG k ; then,
ΔPk(t)=∑i∈Ng(E(Pkg,i(t)))−E(Pkagg(t))−∑j∈Nb(αbt,jPkbt,j(t))−∑l∈Nev(αev,lPkev,l(t)), | (21) |
where αbt,j and αev,l are the binary parameters defined for energy storage system j and electric vehicle l . These parameters are 1 for charging and -1 for the discharging mode. E(Pkagg(t)) is the expected aggregated load demand and E(Pkg,i(t)) is the expected power generation by DG i at time t in MG, and Ng is the number of distributed generators in the microgrid.
The objective is to minimize the overall cost of the microgrid. For each node in the microgrid, the overall cost includes the following components:
a. The cost of power purchased from the utility grid or prosumers, and
b. The cost of the power generated by the distributed generator.
The operational and maintenance costs include the cost of the battery wear due to charging and discharging. If the microgrid is in surplus power, then it can sell energy to the distribution network operator (DNO), other MG, prosumers, and utility grid and generate revenues and follows:
MGprof(t)orMGloss(t)=MGkrev(ΔPk(t))−MGkcost(t)+Cres(t), | (22) |
where MGkcost=TCk(t)+COM(t) , MGprof(t) is the profit, MGloss(t) is the loss of the microgrid, and Cres is resilient energy trading.
The security of the electrical distribution network is an important criterion. The network must be capable to withstand any sudden loss of a part of the network. One of the goals of the work is to design a resilient energy management system. Resilient energy management develops a mechanism to withstand the different sorts of faults such as line faults, distributed grid faults, power outages, and communication network faults.
A. Transmission line fault
During the occurrence of faults, the power could not flow through the faulty transmission line. In this energy management system, this power flow is described as follows:
PL(t)=0,∀t∈[t0L,tfL]. | (23) |
The power flow has been set equal to zero through the faulty line. The t0L is the occurrence time of the fault, and tfL is the final time of the fault duration. If the information of the fault duration is not known, then a future time instant constraint must be imposed. Consequently, the power has to be transferred through the other transmission lines to avoid overloading the other transmission lines. If the fault occurs in the radial transmission, then it is cumbersome to deliver power to the disconnected MG.
B. Distribution Grid Fault
During a distribution grid fault, the MG is unable to draw power from the utility grid until recovery. This is described as a constraint:
Ptg=0,∀i∈{1,…,Ng},∀t∈[tig,tfg], | (24) |
where tig is the initial time of grid fault, and tfg is the final fault time. In all cases of a fault occurrence, the constraints related to the faults are included in the optimization problem.
C. Communication network fault
A communication network fault may occur if the communication link is physically disconnected or if the link may fail due to a cyber attack. The sort of fault can be described in the adjacency matrix, B , as if the communication link between node/agent fails; then,
bi,j=bj,i=0, | (25) |
where bi,j∈B represents the communication link between two nodes i and j in the networked microgrid system. Due to the occurrence of faults, the performance of the energy management algorithm diminishes and convergence problems may occur.
Resilient Energy Management (REM) ensures stable and efficient power distribution under uncertainties such as cyberattacks, equipment failures, or natural disasters. DC Optimal Power Flow (DC-OPF) plays a key role in optimizing the power dispatch while enhancing the system resilience. DC-OPF is a linear approximation of the AC power flow problem and is widely used in power system operations for economic dispatch and congestion management. In the presence of faults, the topology of the network may change due to line outages, bus faults, or generator failures, thus requiring modifications in the standard DC-OPF formulation.
The classical DC-OPF problem aims to minimize the generation cost while satisfying a power balance and transmission constraints. In the presence of faults (e.g., line tripping, generator failure, or bus outage), the network topology and constraints change of DC-OPF is written below.
The standard DC-OPF objective is to minimize the total power generation cost as follows:
min∑i∈GCi(Pgi), | (26) |
where the quadratic cost function for generator i is as follows:
Ci(Pi)=aiP2gi+biPgi+ci. | (27) |
Subject to constraints
a. Generator limits
Pmingi≤Pgi≤Pmaxgi,∀i∈G | (28) |
b. Transmission Line Flow Limits
−Pmaxij≤Bij(θi−θj)≤Pminij | (29) |
c. Line outages (Topology Change)
If a line (i,j) is outaged due to a fault, it is removed from the system, thereby modifying the admittance matrix and power flow equations as follows:
Pij=∑j∈N∖{j′}θi−θjxij,∀(i,j′)∈F, | (30) |
where F is the set of faulted lines, θi and θj are the voltage phase angles at buses i and j , respectively, xij is the reactance of the transmission line between buses i and j . The reactance matrix X and the incidence matrix B are updated to reflect the removal of the affected lines.
To ensure robustness,
−Pmaxij(F)≤Pij(F)≤−Pminij(F). | (31) |
d. If a generator at bus i fails, then
Pgi=0,∀i∈Gfaulted, | (32) |
where Gfaulted represents the set of failed generators. The system needs dispatching of the remaining generators to maintain a power balance as follows:
Pgi(F)+∑j∈DPdj(F)=0,∀F∈С, | (33) |
where С is the possible contingencies, and Pdj is the power demand at the bus.
The line power flow has to be calculated using the following DC power flow equation:
PtL=bALbB−1Ptg, | (34) |
where PtL∈RNL is the vector obtained by stacking the power flows in each line at time t . b∈RNL×NL is a diagonal matrix in which each element b(i,j) is the susceptance of line i . B∈RNB×NB is the admittance matrix, and Pg∈RNB is the vector obtained by stacking all the bus power injections. The elements of the adjacency matrix of the microgrid ALb∈RNL×NL are ALb(i,j)∈{0,1−1} if line i and bus are not connected, line i starts at bus j , or line i ends at the bus, respectively.
The microgrid seeks to minimize the cost of buying energy and maximize the revenue from selling energy. The Resilient energy trading is formulated while considering the DC-OPF in the presence of faults as follows:
Cres=maxPtp,Pts,Ptst[T∑t=1[λ1CtsPts−λ2CtbPtb+λ3CtstPtst+λ3CtpPtp]], | (35) |
subject to (26),
where Cts , Ctb are the selling and buying market prices, respectively, Ctst is the price to store the energy, Ctp is the peer-to-peer trading price, Pts and Ptb are the power sold and bought to the main grid, respectively, and λ1,λ2,λ3,λ4 are the weighting factors.
PSO optimizes DC-OPF by iteratively updating a population of candidate solutions (particles) based on their velocity and position updates as follows:
1. Initialize the particles (each particle represents a possible set of generator outputs Pgi )
2. Evaluate the fitness function (generation cost)
3. Update the particle velocity and position
vt+1i=wvti+c1r1(pbest,i−Ptgi)+c2rr(gbest−Ptgi), | (36) |
Pt+1gi=Ptgi+vt+1i, | (37) |
where w is the inertia weight, c1 and c2 are acceleration coefficients, r1 and r2 are random numbers in [0, 1], pbest,i is the personal best position of particle i , and gbest is the global best position among all the particles. Repeat until the convergence criteria is met.
The RL-based resilient energy trading problem is formulated as an MDP:
M=(S,A,P,r,γ,Rt,Nrel,Oltotal) | (38) |
S is the state space representing the energy market, demand, price, and grid conditions.
A is the action space for trading decisions (buy/sell/store energy).
P is the state transition probability.
r is the reward function modeling profits, resilience, and stability.
γ is the discount factor balancing short-term vs. long-term rewards.
Nrel= Network reliability
Oltotal= Operating loss in resilient energy trading
Rt is grid resilience index that measures how well the energy trading system maintains functionality and restores operations after a fault. The Grid Resilience index in energy trading quantifies the power grid's ability to maintain the stability, efficiency, and reliability while enabling seamless energy transactions under normal and stressed conditions. It ensures that the energy market remains functional despite disruptions such as cyberattacks, equipment failures, demand surges, or renewable intermittency. Before any fault or disruption occurs, the grid operates under normal trading conditions with optimal pricing, demand-supply balance, and stable grid parameters. Once a disruption (fault, cyberattack, or market instability) occurs, post-fault resilience determines the system's ability to either maintain or restore trading operations with a minimal impact.
It is expressed as follows:
Rt=EpostfaultEprefault×100%, | (39) |
where:
Epostfault: Total energy traded before the fault
Eprefault : Total energy traded after the fault
If Rt=100% , then the system is fully resilient (no impact of faults). If Rt <100% , then there is degradation in trading due to faults.
Pre-Fault Energy Trading is affected by the following factors:
● Availability of power from conventional and renewable sources,
● Diversity of generation sources (coal, nuclear, solar, wind, hydropower),
● Real-time energy consumption patterns of industries, households, and businesses,
● Demand-response mechanisms to prevent extreme price hikes, and
● Consistent pricing to ensure fair transactions and minimal volatility.
The post fault energy trading is affected by the following factors:
● Amount of energy that can be quickly restored after a fault,
● Availability of backup power sources such as microgrids or battery storage,
● Sudden price spikes or fluctuations due to power shortages,
● Impact on energy buyers and sellers, and
● How quickly the grid can isolate faults and reroute power.
It refers to the ability of the energy trading systems to maintain stable, secure, and continuous operations despite faults, cyber threats, market fluctuations, and physical grid failures. Several factors influence the network reliability, which are categorized based on technical, economic, and security aspects. Network reliability in energy trading is defined as the probability that all critical components (power grid, trading platform, cybersecurity, and market operations) function properly. The network reliability index is defined as follows:
Nrel=Pg.Ptrad.Pcyb.Pmar1+Pg+Ptrad+Pcyb+Pmar, | (40) |
where Pg = Probability that the power grid is operational,
Ptrad = Probability that the energy trading platform is functioning,
Pcyb= Probability that the cybersecurity system is intact, and
Pmar= Probability that the energy market remains stable.
The state transition probability is defined below.
The operating loss in resilient energy trading refers to the financial and energy losses that occur due to system failures, cyber threats, market disruptions, and grid instabilities. These losses are categorized into direct financial losses, energy loss, opportunity cost, and recovery expenses. The total operating loss in a resilient energy trading system is expressed as follows:
Oltotal=Olenergy+Olmarket+Olcyber+Olrev, | (41) |
where:
Olenergy= Loss due to untraded energy
Olmarket= Loss due to market price fluctuations
Olcyber= Loss from cyberattacks or data breaches
Olrev= Cost of restoring operations after failure
P(t)=eAtP(0), | (42) |
where A is the transition rate matrix and is defined as follows:
A=[−σ0σ00⋯0μ1−(σ1+μ1)λ1⋯00⋮0μ2⋮0−(σ2+μ2)⋮0⋯⋱μn00−μn], | (43) |
where λi= failure rate at state i , and μi= Recovery rate due to self-healing.
Let us consider the nodes in the microgrid defined as an agent. At some discrete-time interval t, the variables considered are load demand, distributed generation, SoC and SoH, and profit. Under extreme and normal conditions, the system state for an MG is described as follows:
sk(t)={(E(Pkg,i(t))).E(SOCkbt,j(t)),E(SOCkbt,l(t)),E(Pkagg)}, | (44) |
where i=1,2,3….,Ndg , j=1,2,3….,Nbt , l=1,2,3….,Nev , and sk∈S and S={s1,s2…….,sn} .
S represents the set of all feasible system states in the system of MG. sk stands for a feasible state, and n is the number of all feasible system states. All feasible system states concerns all possible values of the electricity demand, distributed power generation, and SOC/SOH of the battery. Next, the action of MG is defined on the following decision variables:
ak(t)={Cst,Cbt,Cstt,Cpt}, | (45) |
where ak(k)∈U , U={a1 , a2 , ……., an} , U is the set of all feasible actions of corresponding states, and Nbt and Nev are the numbers of the energy storage devices and electric vehicles, respectively.
The reward function is formulated for resilient energy trading (44). At time t , the reward function is comprised of three components: the profit or loss of MG, the surplus or deficit power, and the duration of the occurrence of a fault,
r(t)=βp×MGkprof(t)−βl×MGkloss(t)+ω×f(ΔPk(t))−φfaultTfault, | (46) |
where (ΔPk)=ΔPk/(1+ln(|ΔPk|) Tfault=|tffault−tifault| , Tfault is the duration of the fault, tifault is the time of occurrence of fault, and tffault is the time of the end of the fault. βp is 1 if MG is in profit, else 0 similarly, βl=1 if MG is in loss, else it is zero, and φfault is 1 if fault occurs else it is 0.
The reward function is defined based on either the surplus or deficit power. The first component is related to the profit of the microgrid, and the second component is related to the loss incurred in the microgrid. The third component is related to the surplus or deficit power. The β and ω are the weighting factors to scale the components. There is an objective to choose Q values that generate greater values of reward.
At a time index, the LSTM cell is comprised of three types of gates, such as Forget ft , an Input gate with an input it and an update gt , and an output ot . The LSTM focuses on time-series forecasting. Let, {xt−1,xt−2,…..,xt−n} denote the input sequence. The suitable weight matrices are considered for corresponding inputs of the network activation functions. The internal memory cell state Ct−1 defines elements of the internal state vector that need to be maintained, updated, or crashed through interaction with outputs of the previous time step ht−1 and the inputs of the current time step xt . The sigmoid activation function is used in the forget gate. The sigmoidal function (or Relu function) and hyperbolic tangent functions are used for the input and the update of the Input gate. The output gate decides what to forget from the previous state ht−1 and updates cell state ct , and which value will pass the output gate. ft , it , gt, and ot are defined as follows:
ft=sigm(Wfxxt+Wfhht−1+bf), | (47) |
it=sigm(Wixxt+Wihht−1+bi), | (48) |
gt=tanh(Wgxxt+Wghht−1+bg), | (49) |
ot=sigm(Woxxt+Wohht−1+bo). | (50) |
The cell memory is recursively updated by interactions with the previous time step value and with the values of ft and gt gates. The process continues to repeat for the next time step.
Ct=ft⊙Ct−1+it⊙gt | (51) |
In the proposed scheme, the Q-function and policy are simultaneously learned. It considers off-policy data and the Bellman equation to learn the Q-function that is used in computing the policy. The optimal action is chosen by taking argmax over the Q-values of all actions. Thus, the actor is the policy network that directly outputs the action. For exploration promotion, some Gaussian noise is included in the action that is determined by the policy. The actor output is fed to the Q network to calculate the Q-value. The target networks are created for both the critics and actors. The target networks are updated based on the main networks. The actor (policy network) loss is the sum of the Q-values of the states. The critic network is used to compute the Q-values. The action network computes the action that is passed to the critic network. Using the reward function (44), the error propagates back to update the LSTM CRITIC and LSTM ACTOR networks, as shown in Figure 3.
SAC is a model-free, off-policy RL algorithm that optimizes both reward maximization and entropy regularization for better exploration. Integrating LSTM into SAC improves its ability to handle sequential dependencies in energy trading, thus making it more resilient to uncertainties in decentralized markets.
Energy trading in decentralized markets is highly dynamic due to the following:
● Fluctuations in the energy demand and supply;
● Uncertain energy prices; and
● Cyber threats and adversarial conditions.
The proposed SAC-LSTM algorithm has the following features:
Temporal awareness: LSTM captures past energy trading patterns.
Exploration-exploitation balance: SAC's entropy regularization ensures robust decision-making.
Resilience: SAC can adapt to adversarial trading conditions and market fluctuations
The policy network outputs a stochastic action distribution using LSTM embeddings:
πϑ(at,st)=N(μθ(ht),σθ(ht)), | (52) |
where ht=LSTM(st,ht−1) represents the hidden state, and μθ(ht) and σθ(ht) are the mean and variance of the active distribution, respectively. The objective function includes entropy regularization,
where H(πθ)=−∑πθlogπθ is the entropy that controls exploration (automatically tuned in SAC).
Critic (Q-Value) Network with LSTM
The Q-function is learned using two critics to reduce overestimation bias.
The target Q-function follows soft Bellman updates,
where the target value is as follows:
yt=rt+γ[mini=1,2Qϕ′(st+1,at+1)−αH(π(at+1,,st+1)]. | (53) |
Automatic Temperature (α\alphaα) Adjustment
SAC dynamically adjusts the entropy weight α :
(α)=Eat π[−αlogπ(at|st)−αH0], | (54) |
where H0 is the target entropy.
Steps of Proposed LSTM- SAC algorithm for resilient energy trading
1. Input: Consider States in the environment
sk(t)={(E(Pkdg,i(t))).E(SOCkbt,j(t)),E(SOCkbt,l(t)),E(Pkagg)} |
2. Output: Action taken in the environment
ak(t)={Cst,Cbt,Cstt,Cpt} |
3. Set βp , βl , ω , φfault and Qth
4. Initialize the replay buffer D
5. Set two different target network parameters as follows:
Φtarg,1←Φ1 and Φtarg,2←Φ2 and
6. For episode 1, 2, 3, ……N do
Initialize states of peer-to-peer resilient energy trading parameters of microgrid for different scenarios,
7. For time slot t=1,2,3………..,T,do
Evaluate reward r(t) where
r(t)=βp×MGkprof,i(t)−βl×MGkloss,i(t)+ω×f(ΔPk(t))−φfaultTfault |
8. For different (s,a) pair
Derive the optimal control action a∗=argmaxaQ∗(s,a)
Execute the control action in the environment
9. Update the state S′ , store the transition in replay buffer D
10. Building LSTM network for the actor and critic for the optimal Q∗(s,a)
11. Check for overestimates Q∗(s,a)≤Qth
Model = sequential();
Model.add (LSTM (neurons), activation = RELU, return_sequences = True)
Model.compile (Loss = RMSE, optimizer = ADAM);
12. While epoch = 1, 2, ….. DO
Train model;
Validate model;
Return model
End
Evaluate future state optimal target Q-function Q∗(s′,a′)
t←t+1
End
A networked microgrid comprised of 04 IEEE 14 bus system is considered with different capacities of distributed generations, energy storage devices, and electric vehicles. The stochastic values of the generation capacity of distributed generators and electricity demand are considered. In an MG, 10 distributed generators (DGs), 04 energy storage devices, and 02 electric vehicles are available in the system. The minimum and maximum power generation by the distributed generators are in the range of 45 MW and 3500 MW.
It is assumed that the MG is facilitated to also get power from the main grid. The simulation work was performed using the GAMS software and MATLAB. The adjacency matrix varies during normal conditions and the occurrence of extreme events. During different scenarios, state-action pairs are generated and Q(s,a) values are approximated using the LSTM network for SAC RL. In the environment, the optimal Q∗(s,a) value and a∗ , which maximize the reward, are obtained.
At a time instant with an aggregated load demand, the generations by different distributed generators are obtained under normal conditions. The optimal power generations of different DGs have been obtained using (26) to (33) and (36) to (37); then, the surplus or deficit power changes were calculated during normal and extreme conditions. The charging schedules of Electric vehicles and energy storage devices were obtained using (17). Due to the occurrence of extreme events, there may be several scenarios on networked microgrids. Out of various scenarios, only one scenario was considered for resilient peer-to-peer energy trading, which is shown in Table 2. Let extreme events occurs at time t=10hours , and the system is recovered at t=20hours so the duration of the fault is Tfault=10hours.
Scenario | MG1 | MG2 |
1 | • Occurrence of fault at transmission line 6-8 | • Occurrence of fault at transmission line 3-4 |
2 | • Failure of DGs 9 and 10 | • Failure of DG 5 and 6 |
Accordingly, the profit or loss of the MG is evaluated in Equation (22). The MG is trained for different scenarios using a LSTM-SAC RL algorithm. The optimal Q values (9) are obtained and the corresponding optimal action (45) and reward (46) are evaluated. The optimal action a∗ is obtained with optimal Q∗ values, while rewards at different time intervals are generated and maximized. To avoid a gradient descent, the LSTM neural network is implemented to generate an optimal Q∗ while the reward is maximized. For different scenarios, such as the occurrence of a fault on a transmission line, the failure of distributed generators, or the failure of supply from the main grid, the networked microgrid system is trained for optimal actions while the reward is maximized. In a scenario, in MG1, the distributed generators 9 and 10 fail, and in MG2, the distributed generators 5 and 6 fail. As shown in Table 2, during the scenario of extreme conditions, the optimal power generation (36)-(37) using the PSO algorithm by microgrid 1 and microgrid 2 are obtained, and are shown in Figures 4 and 5, respectively. The state of charge of the energy storage devices and electric vehicles are computed and shown in Figures 6 and 7, respectively. For microgrid 1 and microgrid 2, during this scenario, the surplus and deficit powers are evaluated at different time intervals, which is shown in Figure 8. The rewards of microgrids 1 and 2 are calculated and shown in Figure 9 during time intervals of extreme conditions. For scenario 1, the grid resilient factor was calculated to 90% (39).
In the learning process, the rewards were obtained for microgrid 1 and microgrid 2, and the time intervals were identified while the rewards were negative. The negative rewards show that the microgrid has a loss during the occurrence of extreme events for peer-to-peer energy trading.
The LSTM network comprises hidden units: 200, training options: adam, MaxEpochs: 250, and Error: Root mean square error (RMSE). The convergence in the LSTM network is shown in Figure 10. Stochastic scheduling has been implemented in peer-to-peer energy trading while energy storages are incorporated to improve resiliency.
In this paper, a multi-LSTM-SAC RL for the resilient energy management of microgrids was proposed for secure and resilient peer-to-peer energy trading. This proposed mechanism was tested in a networked microgrid system comprised of an IEEE 14 bus with energy storage devices and electric vehicles. The MG was trained with LSTM-SAC RL for different scenarios, such as the failure of generation by one or more than one generator or the occurrence of a fault in the transmission line. During the extreme conditions, secure and resilient peer-to-peer energy trading was designed. At different time intervals of extreme events, the optimal power generation, the surplus or deficit power, the state of charge of energy storage devices, and the electric vehicle power were evaluated. The convergence in the proposed learning scheme is obtained and optimal actions were obtained for different extreme events. This learning scheme did not use gradient descent; otherwise, the LSTM algorithm was used to estimate the Q values. Furthermore, this learning scheme avoids overestimating the Q values for resilient energy management. The convergence in the learning mechanism was obtained and loss was minimized. The proposed LSTM-SAC resilient peer-to-peer energy trading creates a platform to maximize profit under extreme conditions.
Desh Deepak Sharma: Conceptualization, methodology, investigation, Ramesh C. Bansal: Validation, visualization; All authors: Writing ‒ original draft, writing ‒ review and editing. All authors approved the final manuscript.
The authors declare that they have not used Artificial Intelligence (AI) tools in the creation of this article.
This work is supported by research project granted by Council of Science and Technology, Uttar Pradesh, India (project id 3246), and Department of Higher Education, U.P. (Project Id :78/2024/1015).
The authors declare that there are no conflicts of interest in this paper.
[1] |
Pudjianto D, Ramsay C, Strbac G (2007) Virtual power plant and system integration of distributed energy resources. IET Renew Power Gen 1: 10‒16. https://doi.org/10.1049/iet-rpg:20060023 doi: 10.1049/iet-rpg:20060023
![]() |
[2] |
Hanna R, Ghonima M, Kleissl J, Tynan G, Victor DG (2017) Evaluating business models for microgrids: Interactions of technology and policy. Energy Policy 103: 47‒61. https://doi.org/10.1016/j.enpol.2017.01.010 doi: 10.1016/j.enpol.2017.01.010
![]() |
[3] | Rahman S (2008) Framework for a resilient and environment-friendly microgrid with demand-side participation. Proceedings IEEE Power Eng Soc Gen Meeting—Convers Del Electr Energy 21st Century. https://doi.org/10.1109/PES.2008.4596108 |
[4] |
Hirsch A, Parag Y, Guerrero J (2018) Microgrids: A review of technologies, key drivers, and outstanding issues. Renewable and Sustainable Energy Reviews 90: 402‒411. https://doi.org/10.1016/j.rser.2018.03.040 doi: 10.1016/j.rser.2018.03.040
![]() |
[5] |
Mengelkamp E, Gä rttner J, Rock K, Kessler S, Orsini L, Weinhardt C (2018) Designing microgrid energy markets: A case study: The Brooklyn Microgrid. Applied Energy 210: 870‒880. https://doi.org/10.1016/j.apenergy.2017.06.054 doi: 10.1016/j.apenergy.2017.06.054
![]() |
[6] |
Shahgholian G (2021) A brief review on microgrids: Operation, applications, modeling, and control. Int T Electr Energy Syst 31: e12885. https://doi.org/10.1002/2050-7038.12885 doi: 10.1002/2050-7038.12885
![]() |
[7] |
Baringo A, L. Baringo L (2016) A stochastic adaptive robust optimization approach for the offering strategy of a virtual power plant. IEEE T Power Syst 32: 3492–3504. https://doi.org/10.1109/TPWRS.2016.2633546 doi: 10.1109/TPWRS.2016.2633546
![]() |
[8] |
Mishra DK, Ray PK, Li L, Zhang J, Hossain M, Mohanty A (2022) Resilient control based frequency regulation scheme of isolated microgrids considering cyber attack and parameter uncertainties. Applied Energy 306: 118054. https://doi.org/10.1016/j.apenergy.2021.118054 doi: 10.1016/j.apenergy.2021.118054
![]() |
[9] | Kamal MB, Wei J (2017) Attack- resilient energy management architecture hybrid emergency power system for more-electric aircrafts. 2017 IEEE Power & Energy Society Innovative Smart Grid Technologies Conference (ISGT). https://doi.org/10.1109/ISGT.2017.8085993 |
[10] |
Hussain A, Bui V, Kim H (2018) A Resilient and Privacy-Preserving Energy Management Strategy for Networked Microgrids. IEEE T Smart Grid 9: 2127‒2139. https://doi.org/10.1109/TSG.2016.2607422 doi: 10.1109/TSG.2016.2607422
![]() |
[11] |
Khodaei A (2014) Resiliency-oriented microgrid optimal scheduling. IEEE T Smart Grid 5: 1584–1591. https://doi.org/10.1109/TSG.2014.2311465 doi: 10.1109/TSG.2014.2311465
![]() |
[12] |
Li Y, Li T, Zhang H, Xie X, Sun Q (2022) Distributed resilient Double –Gradient-Descent Based Energy Management Strategy for Multi-Energy System under DoS attacks. IEEE T Netw Sci Eng 9: 2301‒2316. https://doi.org/10.1109/TNSE.2022.3162669 doi: 10.1109/TNSE.2022.3162669
![]() |
[13] | Liu X (2017) Modelling, analysis, and optimization of interdependent critical infrastructures resilience. Ph. D. thesis, CentraleSupelec. |
[14] | Mumbere SK, Matsumoto S, Fukuhara A, Bedawy A, Sasaki Y, Zoka Y, et al. (2021) An Energy Management System for Disaster Resilience in Islanded Microgrid Networks. 2021 IEEE PES/IAS Power Africa, 1‒5. https://doi.org/10.1109/PowerAfrica52236.2021.9543282 |
[15] |
Gholami A, Shekari T, Grijalva S (2019) Proactive management of microgrids for resiliency enhancement: An adaptive robust approach. IEEE T Sustain Energ 10: 470–480. https://doi.org/10.1109/TSTE.2017.2740433 doi: 10.1109/TSTE.2017.2740433
![]() |
[16] |
Watkins CJ, Dayan P (1992) Q-learning. Machine Learning 8: 279‒292. https://doi.org/10.1007/BF00992698 doi: 10.1007/BF00992698
![]() |
[17] | Nandy A, Biswas M (2019) Reinforcement learning with open AI Tensor Flow, keras using python, Springer Science (Apress.com). |
[18] |
Ganesh AH, Xu B (2022) A review of reinforcement learning based energy management systems for electrified powertrains: Progress, challenge, and potential solution. Renew Sust Energ Rev 154: 111833. https://doi.org/10.1016/j.rser.2021.111833 doi: 10.1016/j.rser.2021.111833
![]() |
[19] |
Zhang T, Sun M, Qiu D, Zhang X, Strbac G, Kang C (2023) A Bayesian Deep Reinforcement Learning-based Resilient Control for Multi-Energy Micro-gird. IEEE T Power Syst 38: 5057‒5072. https://doi.org/10.1109/TPWRS.2023.3233992 doi: 10.1109/TPWRS.2023.3233992
![]() |
[20] |
Wang Y, Qiu D, Strbac G (2022) Multi-agent deep reinforcement learning for resilience-driven routing and scheduling of mobile energy storage systems. Appl Energ 310: 118575. https://doi.org/10.1016/j.apenergy.2022.118575 doi: 10.1016/j.apenergy.2022.118575
![]() |
[21] | Chen Y, Heleno M, Moreira A, Gel YR (2023) Topological Graph Convolutional Networks Solutions for Power Distribution Grid Planning. In: Kashima, H., Ide, T., Peng, WC. (eds) Advances in Knowledge Discovery and Data Mining. PAKDD 2023, 123‒134. https://doi.org/10.1007/978-3-031-33374-3_10 |
[22] |
Zhang B, Hu W, Cao D, Li T, Zhang Z, Chen Z, et al. (2021) Soft actor-critic–based multi-objective optimized energy conversion and management strategy for integrated energy systems with renewable energy. Energ Convers Manage 243: 114381. https://doi.org/10.1016/j.enconman.2021.114381 doi: 10.1016/j.enconman.2021.114381
![]() |
[23] |
Xu D, Cui Y, Ye J, Cha SW, Li A, Zheng C (2022) A soft actor-critic-based energy management strategy for electric vehicles with hybrid energy storage systems. J Power Sources 524: 231099. https://doi.org/10.1016/j.jpowsour.2022.231099 doi: 10.1016/j.jpowsour.2022.231099
![]() |
[24] |
Wang S, Diao R, Xu C, Shi D, Wang Z (2021) On Multi-Event Co-Calibration of Dynamic Model Parameters Using Soft Actor-Critic. IEEE T Power Syst 36: 521‒524. https://doi.org/10.1109/TPWRS.2020.3030164 doi: 10.1109/TPWRS.2020.3030164
![]() |
[25] |
Duan J, Guan Y, Li SE, Ren Y, Sun Q, Cheng B (2022) Distributional Soft Actor-Critic: Off-Policy Reinforcement Learning for Addressing Value Estimation Errors. IEEE T Neur Net Lear Syst33: 6584‒6598. https://doi.org/10.1109/TNNLS.2021.3082568 doi: 10.1109/TNNLS.2021.3082568
![]() |
[26] |
Zhang Z, Chen Z, Lee WJ (2022) Soft Actor-Critic Algorithm Featured Residential Demand Response Strategic Bidding for Load Aggregators. IEEE T Ind Appl 58: 4298‒4308. https://doi.org/10.1109/TIA.2022.3172068 doi: 10.1109/TIA.2022.3172068
![]() |
[27] |
Kathirgamanathan A, Mangina E, Finn DP (2021) Development of a Soft Actor-Critic deep reinforcement learning approach for harnessing energy. Energy and AI 5: 100101. https://doi.org/10.1016/j.egyai.2021.100101 doi: 10.1016/j.egyai.2021.100101
![]() |
[28] |
Ergen T, Kozat SS (2020) Unsupervised Anomaly Detection With LSTM Neural Networks. IEEE T Neur Net Lear Syst 31: 3127‒3141. https://doi.org/10.1109/TNNLS.2019.2935975 doi: 10.1109/TNNLS.2019.2935975
![]() |
[29] |
Bataineh AA, Kaur D (2021) Immunocomputing-Based Approach for Optimizing the Topologies of LSTM Networks. IEEE Access 9: 78993‒79004. https://doi.org/10.1109/ACCESS.2021.3084131 doi: 10.1109/ACCESS.2021.3084131
![]() |
[30] |
Zhou Y, Ma Z, Zhang J, Zou S (2022) Data-driven stochastic energy management of multi-energy system using deep reinforcement learning. Energy 261: 125187–125202. https://doi.org/10.1016/j.energy.2022.125187 doi: 10.1016/j.energy.2022.125187
![]() |
[31] |
Kumar J, Goomer R, Singh AK (2018) Long Short-Term Memory Recurrent Neural Network (LSTM-RNN) Based Workload Forecasting Model For Cloud Datacenters. Procedia Computer Science 125: 676‒682. https://doi.org/10.1016/j.procs.2017.12.087 doi: 10.1016/j.procs.2017.12.087
![]() |
[32] |
Abbasimehr H, Paki R (2022) Improving time series forecasting using LSTM and attention models. J Amb Intell Human Comput 13: 673–691. https://doi.org/10.1007/s12652-020-02761-x doi: 10.1007/s12652-020-02761-x
![]() |
[33] |
Liu Q, Long L, Yang Q, Peng H, Wang J, Luo X (2022) LSTM-SNP: A long short-term memory model inspired from spiking neural P systems. Knowledge-Based Systems 235: 107656. https://doi.org/10.1016/j.knosys.2021.107656 doi: 10.1016/j.knosys.2021.107656
![]() |
[34] |
Moghar A, Hamiche M (2020) Stock Market Prediction Using LSTM Recurrent Neural Network. Procedia Computer Science 170: 1168‒1173. https://doi.org/10.1016/j.procs.2020.03.049 doi: 10.1016/j.procs.2020.03.049
![]() |
[35] |
Karijadi I, Chou S (2022) A hybrid RF-LSTM based on CEEMDAN for improving the accuracy of building energy consumption prediction. Energy and Buildings 259: 111908. https://doi.org/10.1016/j.enbuild.2022.111908 doi: 10.1016/j.enbuild.2022.111908
![]() |
[36] |
Zhao L, Mo C, Ma J, Chen Z, Yao C (2022) LSTM-MFCN: A time series classifier based on multi-scale spatial–temporal features. Comput Commun 182: 52‒59. https://doi.org/10.1016/j.comcom.2021.10.036 doi: 10.1016/j.comcom.2021.10.036
![]() |
[37] |
Ariza I, Tardón LJ, Barbancho AM, De-Torres I, Barbancho I (2022) Bi-LSTM neural network for EEG-based error detection in musicians' performance. Biomed Signal Process Control 78: 103885. https://doi.org/10.1016/j.bspc.2022.103885 doi: 10.1016/j.bspc.2022.103885
![]() |
[38] |
Karim F, Majumdar, S, Darabi H, Harford S (2019) Multivariate LSTM-FCNs for time series classification. Neural Networks 116: 237‒245. https://doi.org/10.1016/j.neunet.2019.04.014 doi: 10.1016/j.neunet.2019.04.014
![]() |
[39] |
Saini KK, Sharma P, Mathur HD, Gautam AR, Bansal RC (2024) Techno-economic and Reliability Assessment of an Off-grid Solar-powered Energy System. Appl Energy 371: 123579. https://doi.org/10.1016/j.apenergy.2024.123579 doi: 10.1016/j.apenergy.2024.123579
![]() |
[40] |
Sharma DD (2024) Asynchronous Blockchain-based Federated Learning for Tokenized Smart Power Contract of Heterogeneous Networked Microgrid System. IET Blockchain 4: 302‒314. https://doi.org/10.1049/blc2.12041 doi: 10.1049/blc2.12041
![]() |
[41] |
Sharma DD, Singh SN, Lin J (2024) Blockchain-enabled secure and authentic Nash Equilibrium strategies for heterogeneous networked hub of electric vehicle charging stations. Blockchain: Research and Applications 5: 100223. https://doi.org/10.1016/j.bcra.2024.100223 doi: 10.1016/j.bcra.2024.100223
![]() |
[42] |
Foruzan E, Soh L, Asgarpoor S (2018) Reinforcement Learning Approach for Optimal Distributed Energy Management in a Microgrid. IEEE Trans Power Syst 33: 5749‒5758. https://doi.org/10.1109/TPWRS.2018.2823641 doi: 10.1109/TPWRS.2018.2823641
![]() |
[43] | Suttan RS, Barto AG (2018) Reinforcement Learning: An introduction, the MIT Press London, England. |
Sr No | Fault Type | Impact on trading |
1 | Line Faults | Power congestion, price fluctuations |
2 | Generator Failure | Demand-supply imbalance, blackouts |
3 | Bus Outages | Complete trading failure, cascading blackouts |
4 | Cyber Attacks | Price manipulation, fake trading data |
Scenario | MG1 | MG2 |
1 | • Occurrence of fault at transmission line 6-8 | • Occurrence of fault at transmission line 3-4 |
2 | • Failure of DGs 9 and 10 | • Failure of DG 5 and 6 |
Sr No | Fault Type | Impact on trading |
1 | Line Faults | Power congestion, price fluctuations |
2 | Generator Failure | Demand-supply imbalance, blackouts |
3 | Bus Outages | Complete trading failure, cascading blackouts |
4 | Cyber Attacks | Price manipulation, fake trading data |
Scenario | MG1 | MG2 |
1 | • Occurrence of fault at transmission line 6-8 | • Occurrence of fault at transmission line 3-4 |
2 | • Failure of DGs 9 and 10 | • Failure of DG 5 and 6 |