November 16, 2020
Conference Paper

Automated Adversary Emulation for Cyber-Physical Systems via Reinforcement Learning

Abstract

Adversary emulation is an offensive exercise that provides a comprehensive assessment of a system’s resilience against cyber attacks. However, adversary emulation is typically a manual process, making it costly and hard to deploy in cyber-physical systems (CPS) with complex dynamics, vulnerabilities, and operational uncertainties. In this paper, we develop an automated, domain-aware approach to adversary emulation for CPS. We formulate a Markov Decision Process (MDP) model to determine an optimal attack sequence over a hybrid attack graph with cyber (discrete) and physical (continuous) components and related physical dynamics. We apply model-based and model-free reinforcement learning (RL) methods to solve the discrete-continuous MDP in a tractable fashion. As a baseline, we also develop a greedy attack algorithm and compare it with the RL procedures. We summarize our findings through a numerical study on sensor deception attacks in buildings to compare the performance and solution quality of the proposed algorithms.

Revised: January 19, 2021 | Published: November 16, 2020

Citation

Bhattacharya A., T. Ramachandran, S. Banik, C.P. Dowling, and S. Bopardikar. 2020. Automated Adversary Emulation for Cyber-Physical Systems via Reinforcement Learning. In IEEE International Conference on Intelligence and Security Informatics (ISI 2020), November 9-10, 2020, Arlington, VA, 1-6. Piscataway, New Jersey:IEEE. PNNL-SA-156040. doi:10.1109/ISI49825.2020.9280521