TY - GEN
T1 - Reinforcement Learning-Based Observability-Aware Cyber Restoration of Power Grid
AU - Edib, Shamsun Nahar
AU - Vokkarane, Vinod M.
AU - Lin, Yuzhang
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - The cyber resilience of cyber-physical power grids relies on swift restoration of cyber domain components following major disturbances such as, natural disasters or man-made attacks. The cyber domain restoration problem is inherently stochastic due to uncertainties surrounding initial outage conditions and restoration action failures. Traditionally, optimization-based methods, such as heuristics and mixed-integer linear programming (MILP), are utilized for solving restoration problems. However, these methods suffer from time-consuming processes and limited adaptability to dynamic conditions. To address these challenges, this paper formulates the observability recovery problem (ORP) as a Markov decision process and uses deep reinforcement learning (DRL) to solve the problem. Numerical simulations on the IEEE 30-bus system demonstrate that our proposed approach outperforms the heuristic approach in terms of both performance and computational efficiency. Moreover, when compared to the MILP approach, our method achieves comparable performance while requiring significantly less computation time.
AB - The cyber resilience of cyber-physical power grids relies on swift restoration of cyber domain components following major disturbances such as, natural disasters or man-made attacks. The cyber domain restoration problem is inherently stochastic due to uncertainties surrounding initial outage conditions and restoration action failures. Traditionally, optimization-based methods, such as heuristics and mixed-integer linear programming (MILP), are utilized for solving restoration problems. However, these methods suffer from time-consuming processes and limited adaptability to dynamic conditions. To address these challenges, this paper formulates the observability recovery problem (ORP) as a Markov decision process and uses deep reinforcement learning (DRL) to solve the problem. Numerical simulations on the IEEE 30-bus system demonstrate that our proposed approach outperforms the heuristic approach in terms of both performance and computational efficiency. Moreover, when compared to the MILP approach, our method achieves comparable performance while requiring significantly less computation time.
KW - Cyber-physical system
KW - deep reinforcement learning
KW - Markov decision process
KW - observability
KW - power system restoration
UR - http://www.scopus.com/inward/record.url?scp=105000830218&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=105000830218&partnerID=8YFLogxK
U2 - 10.1109/GLOBECOM52923.2024.10901146
DO - 10.1109/GLOBECOM52923.2024.10901146
M3 - Conference contribution
AN - SCOPUS:105000830218
T3 - Proceedings - IEEE Global Communications Conference, GLOBECOM
SP - 1599
EP - 1604
BT - GLOBECOM 2024 - 2024 IEEE Global Communications Conference
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2024 IEEE Global Communications Conference, GLOBECOM 2024
Y2 - 8 December 2024 through 12 December 2024
ER -