TY - GEN
T1 - Network Reconfiguration for Enhanced Operational Resilience using Reinforcement Learning
AU - Abdelmalak, Michael
AU - Gautam, Mukesh
AU - Morash, Sean
AU - Snyder, Aaron F.
AU - Hotchkiss, Eliza
AU - Benidris, Mohammed
N1 - Funding Information:
reconfigurations might exist, the ACA selects the decision based on their corresponding probability of success. In other words, connecting SW1 and SW3 will result in feasible reconfiguration solutions. However, this decision is associated with less probability value within the trained ACA model. V. CONCLUSION This paper has proposed a distribution network reconfiguration approach to control tie-switches of distribution systems for enhanced operational resilience. The proposed method leverages ACA to determine set of tie-switches to be connected due to multiple line outages. An MDP was formulated to train a single agent actor-critic model. The process was repeated for diverse failure scenarios to train the ACA networks. The proposed methodology was tested on the 33-node distribution feeder. The results showed the effectiveness of the proposed ACA to determine the set of tie-switches that allow feasible network reconfiguration maintaining traverse and radiality constraints. The trained ACA was tested against single, double, and multiple failure scenarios and showed accuracy of almost 97%. The proposed algorithm provides the system operators with a fast-acting algorithm to restore curtailed loads in distribution networks after an extreme event. In the future, the characteristics of power systems including loads, voltages, and currents will be considered as well as scalability to large-scale systems. ACKNOWLEDGEMENT This work was supported by the U.S. National Science Foundation (NSF) under Grant NSF 1847578.
Publisher Copyright:
© 2022 IEEE.
PY - 2022/9/28
Y1 - 2022/9/28
N2 - This paper proposes a reinforcement learning-based approach for distribution network reconfiguration(DNR) to enhance the resilience of the electric power supply. Resilience enhancements usually require solving large-scale stochastic optimization problems that are computationally expensive and sometimes infeasible. The exceptional performance of reinforcement learning techniques has encouraged their adoption in various power system control studies, specifically resilience-based real-time applications. In this paper, a single agent framework is developed using an Actor-Critic algorithm (ACA) to determine statuses of tie-switches in a distribution feeder impacted by an extreme weather event. The proposed approach provides a fast-acting control algorithm that reconfigures the feeder topology to reduce or even avoid load shedding. The problem is formulated as a discrete Markov decision process in such a way that a system state captures the system topology and its operational characteristics. An action is made to open or close a specific set of tie-switches after which a reward is calculated to evaluate the practicality and advantage of that action. The iterative Markov process is used to train the proposed ACA under diverse failure scenarios and is demonstrated on the 33-node distribution feeder system. Results show the capability of the proposed ACA to determine proper switching action of tie-switches with accuracy exceeding 93%.
AB - This paper proposes a reinforcement learning-based approach for distribution network reconfiguration(DNR) to enhance the resilience of the electric power supply. Resilience enhancements usually require solving large-scale stochastic optimization problems that are computationally expensive and sometimes infeasible. The exceptional performance of reinforcement learning techniques has encouraged their adoption in various power system control studies, specifically resilience-based real-time applications. In this paper, a single agent framework is developed using an Actor-Critic algorithm (ACA) to determine statuses of tie-switches in a distribution feeder impacted by an extreme weather event. The proposed approach provides a fast-acting control algorithm that reconfigures the feeder topology to reduce or even avoid load shedding. The problem is formulated as a discrete Markov decision process in such a way that a system state captures the system topology and its operational characteristics. An action is made to open or close a specific set of tie-switches after which a reward is calculated to evaluate the practicality and advantage of that action. The iterative Markov process is used to train the proposed ACA under diverse failure scenarios and is demonstrated on the 33-node distribution feeder system. Results show the capability of the proposed ACA to determine proper switching action of tie-switches with accuracy exceeding 93%.
KW - Actor critic
KW - Markov decision process
KW - network reconfiguration
KW - reinforcement learning
KW - resilience
UR - http://www.scopus.com/inward/record.url?scp=85140835053&partnerID=8YFLogxK
UR - https://www.mendeley.com/catalogue/8fb5bc47-1704-359d-976d-f72a2f26c9a4/
U2 - 10.1109/SEST53650.2022.9898469
DO - 10.1109/SEST53650.2022.9898469
M3 - Conference contribution
AN - SCOPUS:85140835053
SN - 9781665405577
T3 - SEST 2022 - 5th International Conference on Smart Energy Systems and Technologies
BT - SEST 2022 - 5th International Conference on Smart Energy Systems and Technologies
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 5th International Conference on Smart Energy Systems and Technologies, SEST 2022
Y2 - 5 September 2022 through 7 September 2022
ER -