Sharma, Desh DeepakBansal, Ramesh C.2025-04-162025-04-162025-03Sharma, D.D. & Bansal, R.C. 2025, 'LSTM-SAC reinforcement learning based resilient energy trading for networked microgrid system', AIMS Electronics and Electrical Engineering, vol. 9, no. 2, pp. 165-191, doi : 10.3934/electreng.2025009.2578-1588 (online)10.3934/electreng.2025009http://hdl.handle.net/2263/102140On the whole, the present microgrid constitutes numerous actors in highly decentralized environments and liberalized electricity markets. The networked microgrid system must be capable of detecting electricity price changes and unknown variations in the presence of rare and extreme events. The networked microgrid system comprised of interconnected microgrids must be adaptive and resilient to undesirable environmental conditions such as the occurrence of different kinds of faults and interruptions in the main grid supply. The uncertainties and stochasticity in the load and distributed generation are considered. In this study, we propose resilient energy trading incorporating DC-OPF, which takes generator failures and line outages (topology change) into account. This paper proposes a design of Long Short-Term Memory (LSTM) - soft actor-critic (SAC) reinforcement learning for the development of a platform to obtain resilient peer-to-peer energy trading in networked microgrid systems during extreme events. A Markov Decision Process (MDP) is used to develop the reinforcement learning-based resilient energy trade process that includes the state transition probability and a grid resilience factor for networked microgrid systems. LSTM-SAC continuously refines policies in real-time, thus ensuring optimal trading strategies in rapidly changing energy markets. The LSTM networks have been used to estimate the optimal Q-values in soft actor-critic reinforcement learning. This learning mechanism takes care of the out-of-range estimates of Q-values while reducing the gradient problems. The optimal actions are decided with maximized rewards for peer-to-peer resilient energy trading. The networked microgrid system is trained with the proposed learning mechanism for resilient energy trading. The proposed LSTM-SAC reinforcement learning is tested on a networked microgrid system comprised of IEEE 14 bus systems.en© 2025 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0).Long short-term memory (LSTM)Networked microgridsReinforcement learning (RL)Resilient energy tradingSoft actor-critic (SAC)Renewable energyDistributed energy resources (DER)DC optimal power flow (DC-OPF)SDG-07: Affordable and clean energyLSTM-SAC reinforcement learning based resilient energy trading for networked microgrid systemArticle