Adversary emulation is an offensive exercise that
provides a comprehensive assessment of a system’s resilience
against cyber attacks. However, adversary emulation is typically
a manual process, making it costly and hard to deploy in cyber-physical
systems (CPS) with complex dynamics, vulnerabilities,
and operational uncertainties. In this paper, we develop an
automated, domain-aware approach to adversary emulation for
CPS. We formulate a Markov Decision Process (MDP) model to
determine an optimal attack sequence over a hybrid attack graph
with cyber (discrete) and physical (continuous) components and
related physical dynamics. We apply model-based and model-free
reinforcement learning (RL) methods to solve the discrete-continuous
MDP in a tractable fashion. As a baseline, we also
develop a greedy attack algorithm and compare it with the RL
procedures. We summarize our findings through a numerical
study on sensor deception attacks in buildings to compare the
performance and solution quality of the proposed algorithms.
Revised: January 19, 2021 |
Published: November 16, 2020
Citation
Bhattacharya A., T. Ramachandran, S. Banik, C.P. Dowling, and S. Bopardikar. 2020.Automated Adversary Emulation for Cyber-Physical Systems via Reinforcement Learning. In IEEE International Conference on Intelligence and Security Informatics (ISI 2020), November 9-10, 2020, Arlington, VA, 1-6. Piscataway, New Jersey:IEEE.PNNL-SA-156040.doi:10.1109/ISI49825.2020.9280521