Bus 14
Bus 14
T R A N S M I S S I O N S Y S T E M U S I N G S E C U R I T Y A N A LY T I C S
A Dissertation
Presented in Partial Fulfillment of the Requirements for the
DEGREE OF DOCTOR OF PHILOSOPHY
with a
Major in Computer Science
in the
College of Graduate Studies
University of Idaho
by
MOHAMMAD ASHRAFUZZAMAN
Major Professor
FREDERICK T. SHELDON, PH.D.
Committee Members
HASAN M. JAMIL, PH.D., ROBERT B. HECKENDORN, PH.D.,
DANIEL C O N T E D E L E O N , P H . D . , M I C H A E L A . H A N E Y, P H . D . ,
C L I N T O N L . J E F F E R Y, P H . D .
Department Administrator
TERENCE SOULE, PH.D.
D E C E M B E R 2020
ii
A U T H O R I Z AT I O N TO S U B M I T D I S S E RTAT I O N
Major Professor:
Frederick T. Sheldon, Ph.D. Date
Committee Members:
Hasan M. Jamil, Ph.D. Date
Department
Administrator:
Terence Soule, Ph.D. Date
iii
ABSTRACT
The electric smart grid, a critical national infrastructure and among the largest and
intelligently change the values in the measurement matrix used to compute state
estimation. These data integrity attacks can potentially disrupt the critical control
false data injection (SFDI) attacks against SE cannot be detected by the conventional
SE measurement data is presented. A threat model that identified three possible attack
models was developed, and synthetic datasets corresponding to these attack models
were generated for standard IEEE 14-bus and 57-bus systems. After normalizing and
and stacking ensemble machine learning models were trained and tested for model
and cross-validation, trained models were identified that can detect the SFDI attacks
accurately and reliably. Evaluation of the models using standard metrics shows that
supervised artificial neural networks with four hidden layers and 1200 hidden units
per layer can detect 98.24% of the attacks with a false alarm rate of 1.25%. Among
the unsupervised models, elliptic envelope performs the best with 73% detection rate
with 3% false alarm rate. It was also found that the detection rate is the same for all
iv
the machine learning methods for all the six datasets corresponding to different attack
The core contributions of this dissertation are the demonstration that a machine
learning based security analytics framework can successfully detect the SFDIA attacks
and the identification of artificial neural network with the right set of hyper-parameter
values as the best performing model. Additional contributions include a survey and a
taxonomy of false data injection attacks on different parts of the power grid for the first
time in the literature, an exhaustive survey of machine learning based approaches for
detecting SFDI attacks, implementation of a software for running the machine learning
ACKNOWLEDGEMENTS
The formulation of this research proposal has been possible because of the support
Dr. Frederick T. Sheldon for his encouragement, guidance and support in undertak-
ing and proceeding with this research. I extend my heartfelt gratitude towards the
Dr. Hasan Jamil, Dr. Daniel Conte de Leon, and Dr. Michael Haney for their valuable
advice during the research, for reviewing this dissertation and for providing insightful
feedback. I am grateful to Dr. Yacine Chakhchoukh for his support in generating the
all-important datasets for this research, and to Dr. Brian Johnson for his many im-
portant advice. I gratefully acknowledge the research collaborations with Dr. Ananth
Jillepalli and Saikat Das and their encouragement. My sincere thanks go to Victor
House and John McFarland for providing technical support in setting up the research
infrastructure, and to Susan Branting and Arleen Furedy for providing administrative
2016 (Grant Number IGEM17-001). I gratefully acknowledge this support from the
State of Idaho.
vi
D E D I C AT I O N
and
and
TA B L E OF CONTENTS
A U T H O R I Z AT I O N TO S U B M I T D I S S E R TAT I O N . . . . . . . . . . . . . ii
ABSTRACT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iii
ACKNOWLEDGEMENTS . . . . . . . . . . . . . . . . . . . . . . . . v
D E D I C AT I O N . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vi
TA B L E OF CONTENTS . . . . . . . . . . . . . . . . . . . . . . . . . vii
LIST OF FIGURES. . . . . . . . . . . . . . . . . . . . . . . . . . . x
LIST OF TA B L E S . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii
LIST OF A B B R E V I AT I O N S . . . . . . . . . . . . . . . . . . . . . . . xv
1 INTRODUCTION . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1 The Context . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 The Research Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.3 Security Analytics Approach . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.4 Scope of the Research . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.5 Contributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.6 Author’s Related Publications . . . . . . . . . . . . . . . . . . . . . . . . 8
1.7 Organization of the Dissertation . . . . . . . . . . . . . . . . . . . . . . . 10
2 R E L AT E D W O R K . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
2.2 Detection using Supervised Learning . . . . . . . . . . . . . . . . . . . . 12
2.3 Detection using Unsupervised Learning . . . . . . . . . . . . . . . . . . . 16
2.4 Detection using Semi-supervised Learning . . . . . . . . . . . . . . . . . 19
2.5 Detection using Reinforcement Learning . . . . . . . . . . . . . . . . . . 21
2.6 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
3 S TAT E E S T I M AT I O N A N D S T E A L T H Y F A L S E D ATA I N J E C T I O N A T TA C K S . 23
3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
3.2 State Estimation in Power Transmission Systems . . . . . . . . . . . . . . 23
3.3 Formulation of State Estimation and False Data Injection Attacks . . . . . 25
3.4 Stealthy False Data Injection Attacks . . . . . . . . . . . . . . . . . . . . 27
3.5 Stealthy False Data Injection Attack Process . . . . . . . . . . . . . . . . 28
3.6 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
4 T A X O N O M Y O F F A L S E D ATA I N J E C T I O N A T TA C K S . . . . . . . . . . . 32
4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
4.2 Related Works . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
4.3 Component-based Attack Classification . . . . . . . . . . . . . . . . . . . 33
viii
LIST OF FIGURES
FIGURE 3.1 A flow chart of the state estimation process (from [105]). . . . 24
FIGURE 3.2 Diagram showing relationship between bad data, false data and
stealthy false data. . . . . . . . . . . . . . . . . . . . . 28
FIGURE 3.3 A simplified diagram of a power grid with its SCADA and communication
systems showing the FDI attack vectors. . . . . . . . . . . . 29
FIGURE 4.1 Diagram showing taxonomy of stealthy false data injection attacks
on the power grid. . . . . . . . . . . . . . . . . . . . . 34
FIGURE 7.1 Five individual classifiers combining with one of six ensemble
classifiers to form six ensembles. . . . . . . . . . . . . . . 82
FIGURE A.1 Graph of the evaluation metrics values for supervised individual
and ensemble models for the 14-bus-AM1 dataset. . . . . . . 124
FIGURE A.2 Graph of the evaluation metrics values for unsupervised individual
and ensemble models for the 14-bus-AM1 dataset. . . . . . . 124
FIGURE A.3 ROC curves for the supervised individual and ensemble models
for 14-bus-AM1 dataset. . . . . . . . . . . . . . . . . . . 125
FIGURE A.4 ROC curves for the unsupervised individual and ensemble models
for 14-bus-AM1 dataset. . . . . . . . . . . . . . . . . . . 125
FIGURE A.5 Graph of the evaluation metrics values for supervised individual
and ensemble models for the 14-bus-AM2 dataset. . . . . . . 127
FIGURE A.6 Graph of the evaluation metrics values for unsupervised individual
and ensemble models for the 14-bus-AM2 dataset. . . . . . . 127
FIGURE A.7 ROC curves for the supervised individual and ensemble models
for 14-bus-AM2 dataset. . . . . . . . . . . . . . . . . . . 128
FIGURE A.8 ROC curves for the unsupervised individual and ensemble models
for 14-bus-AM2 dataset. . . . . . . . . . . . . . . . . . . 128
xi
FIGURE A.9 Graph of the evaluation metrics values for supervised individual
and ensemble models for the 14-bus-AM3 dataset. . . . . . . 130
FIGURE A.10 Graph of the evaluation metrics values for unsupervised individual
and ensemble models for the 14-bus-AM3 dataset. . . . . . . 130
FIGURE A.11 ROC curves for the supervised individual and ensemble models
for 14-bus-AM3 dataset. . . . . . . . . . . . . . . . . . . 131
FIGURE A.12 ROC curves for the unsupervised individual and ensemble models
for 14-bus-AM3 dataset. . . . . . . . . . . . . . . . . . . 131
FIGURE A.13 Graph of the evaluation metrics values for supervised individual
and ensemble models for the 57-bus-AM1 dataset. . . . . . . 133
FIGURE A.14 Graph of the evaluation metrics values for unsupervised individual
and ensemble models for the 57-bus-AM1 dataset. . . . . . . 133
FIGURE A.15 ROC curves for the supervised individual and ensemble models
for 57-bus-AM1 dataset. . . . . . . . . . . . . . . . . . . 134
FIGURE A.16 ROC curves for the unsupervised individual and ensemble models
for 57-bus-AM1 dataset. . . . . . . . . . . . . . . . . . . 134
FIGURE A.17 Graph of the evaluation metrics values for supervised individual
and ensemble models for the 57-bus-AM2 dataset. . . . . . . 136
FIGURE A.18 Graph of the evaluation metrics values for unsupervised individual
and ensemble models for the 57-bus-AM2 dataset. . . . . . . 136
FIGURE A.19 ROC curves for the supervised individual and ensemble models
for 57-bus-AM2 dataset. . . . . . . . . . . . . . . . . . . 137
FIGURE A.20 ROC curves for the unsupervised individual and ensemble models
for 57-bus-AM2 dataset. . . . . . . . . . . . . . . . . . . 137
FIGURE A.21 Graph of the evaluation metrics values for supervised individual
and ensemble models for the 57-bus-AM3 dataset. . . . . . . 139
FIGURE A.22 Graph of the evaluation metrics values for unsupervised individual
and ensemble models for the 57-bus-AM3 dataset. . . . . . . 139
FIGURE A.23 ROC curves for the supervised individual and ensemble models
for 57-bus-AM3 dataset . . . . . . . . . . . . . . . . . . 140
FIGURE A.24 ROC curves for the unsupervised individual and ensemble models
for 57-bus-AM3 dataset . . . . . . . . . . . . . . . . . . 140
FIGURE A.25 Graph of metrics values for neural networks models with varying
hidden layers and hidden units on 14-Bus-AM1 dataset. . . . . 143
FIGURE A.26 Graph of metrics values for neural networks models with varying
hidden layers and hidden units on 14-Bus-AM2 dataset. . . . . 145
FIGURE A.27 Graph of metrics values for neural networks models with varying
hidden layers and hidden units on 14-Bus-AM3 dataset. . . . . 146
FIGURE A.28 Graph of metrics values for neural networks models with varying
hidden layers and hidden units on 57-Bus-AM1 dataset. . . . . 149
FIGURE A.29 Graph of metrics values for neural networks models with varying
hidden layers and hidden units on 57-Bus-AM2 dataset. . . . . 150
FIGURE A.30 Graph of metrics values for neural networks models with varying
hidden layers and hidden units on 57-Bus-AM3 dataset. . . . . 152
xii
LIST OF TA B L E S
TA B L E 6 . 1 Evaluation metrics values for the three supervised methods for all
the six datasets. . . . . . . . . . . . . . . . . . . . . . . 73
LIST OF A B B R E V I AT I O N S
CHAPTER 1
INTRODUCTION
The power grid1 , including generators, transmission systems, distribution systems, and
numerous other devices, is one of the largest and most complex critical infrastruc-
tures. The power grid and its various components have, for decades, been undergoing
evolution. Significant changes came when the components of the grid and the super-
visory control and data acquisition (SCADA) systems that monitor and control these
power grids into cyber-physical systems (CPS), with the downside of inheriting issues
Nonetheless, the industry has attempted to “air-gap” operational technology (OT) from
information technology (IT) networks toward protecting valuable CPS assets critical to
stable operations. Unfortunately, many OT networks are still not fully insulated from
the IT networks [16] and are vulnerable to both internal and external threats [77]. A
study by the Ponemon Institute reports that 90% of organizations relying on OT have
the report [92]. As a result the power grid has been subjected to a new set of exploits,
The vulnerabilities of power grids are illustrated with a few well-known incidents.
In January of 2008 the Central Intelligence Agency (CIA) reported that a number of
1 throughout this dissertation, the terms “smart grid” and “power grid” are used interchangeably.
2
non-US cities were under cyber-attack affecting distribution systems causing a wide-
spread blackout [1]. In December 2015, three power distribution companies were
taken down in a coordinated cyber-attack where seven substations were isolated for
7 hours and operators were forced to switch to manual mode in order to gain control
back to the grid, resulting in a power outage for about 225,000 Ukrainians [61].
One of the many ways smart grids can be attacked over the cyber network is using
stealthy false data injection (SFDI) attacks, which are described as a new class of
cyber-attacks in power grids by Liu et al. [74]. In SFDI attacks, also called deception
attacks or covert data integrity attacks, adversaries intrude into different parts of the
power grids, for example smart meters, substation meters or sensors, any part of
control system computers, etc. and then inject or modify any measurement data with
the goal of causing disruption in the grid operation, power theft, or any other malicious
intent. As shown by Xiang et al. [116], false data injection is can be an important
element of a coordinated attack on the power grid and represents an important class
One such SFDI attack targets the state estimation process in the power transmission
system (EMS) at the power grid control center. The SE computes voltage magnitudes
and phase angles at all of the different buses of the power system after collecting
measurements that are communicated to the control center from remote terminal
units (RTUs) equipped with SCADA units [2]. State estimation process is described
3
in Section 3.2. If the incorrect measurement data affect the outcome of state esti-
mation, the resulting misinformation can reduce the control center operators’ level
of situational awareness [6]. This potentially forces the operators to take corrective
actions under the false assumption that the state variable values obtained from state
estimation are correct. This may cause disruption in the real-time operation of the
transmission system by adversely impacting tools for contingency analysis, unit com-
mitment, optimal power flow and computation of locational marginal pricing (LMPs)
for electricity markets. Cyber-attacks that impact the SE results have been presented
The state estimation process includes a step called “bad data processing” whereby
any anomalous data, be it due to telemetry malfunction or false data injection (FDI)
stealthy false data injection (SFDI) attacks that cannot be detected by the traditional
constitutes “stealthy” false data injection attacks on the state estimation in power
1.3 S E C U R I T Y A N A LY T I C S A P P R O A C H
Rapid advancement in machine learning algorithms have enabled them to find natural
patterns in data that generate insight and enable better decisions and predictions. The
use of data analytics to predict, detect, and prevent security threats is termed security
analytics [25]. The incorporation of cyber capabilities into smart grid functionality
has led to the proliferation of new data sources. The availability of abundant data
4
After using traditional statistical approaches and physics of state estimation, re-
searchers have embarked on using machine learning based approaches to detect SFDI
attacks. A literature review on machine learning based stealthy false data injection
attack (SFDIA) detection approaches given in Chapter 2 shows that this research area
is still far from being mature, and a natural progression seems toward using machine
learning (ML) based approaches to develop an effective way to detect the SFDI attacks
to investigate and identify suitable machine learning methods that will reliably and
accurately detect presence of SFDI attacks in the state estimation measurement data.
1. As part of threat modeling, three attack models for stealthy false data injection
2. Standard IEEE power transmission systems with 14 buses and 57 buses were
measurement data used by state estimation under normal operation of the sys-
tems.
3. An attack generation module was used to generate spoofed data that replaced
actual data in some of the normal measurement data, simulating stealthy false
data injection attacks. For each of the three attack models, datasets were gen-
5
erated and attack data were injected in the SE measurement data, obtaining six
4. The features (i.e., the sensor measurements) in the datasets were ranked accord-
ing to their importance (in training machine learning models) using random
forest classifier and eliminated the features that have minimum contribution on
model training.
5. Three supervised machine learning models, namely GLM, GLB, and DRF, were
trained, tested, and evaluated for performance using standard metrics to find out
models were developed. The ensemble models were trained, tested, and eval-
uated for performance using standard metrics to find out the efficacy of these
was trained, tested and evaluated for performance using standard metrics to find
trained, tested and evaluated for performance using standard metrics to find out
9. After analyzing and comparing the performance of the different models in de-
tecting SFDI attacks, the best-performing models were identified. It was found
that artificial neural network performs the best, among supervised models, with
98.25% detection rate and 1.25% false alarm rate and elliptic envelope performs
6
the best among unsupervised models with 73% detection rate and 3% false alarm
transmission systems.
2. The research considered time-discrete data only, and not time-series data.
of real data.
1.5 CONTRIBUTIONS
tem state estimation measurement data to detect presence of stealthy false data
injection attacks.
2. Using the framework, achieving over 98% detection rates with negligible false
learning models on a dataset. The framework runs the ML methods with differ-
for the methods, and prints out the collated results in tables and as graphs.
5. Model selection and evaluation of SFDI attack detection efficiency of three super-
using five supervised methods in the first stage of the stacks, and another using
five unsupervised models. Six binary classifiers were used in the second stage as
ensemble classifiers.
neural network with varying hyper-parameter values to find out the best-performing
1.6 A U T H O R ’ S R E L AT E D P U B L I C AT I O N S
tecting stealthy false data injection attacks in the smart grid using ensemble-
based machine learning,” Journal of Computers & Security, Elsevier, August 2020.
doi:10.1016/j.cose.2020.101994
Sheldon. “Supervised learning for detecting stealthy false data injection attacks
Nature, 2020.
“Elliptic envelope for detecting stealthy false data injection attacks in the smart
power grids using deep learning,” in IEEE Wireless Communications and Mobile
doi:10.1109/IWCMC.2018.8450487
9
damage mitigation model for cyber-attacks on smart grids,” 2018 IEEE 42nd
The author’s other collaborative publications not directly related to this dissertation
are:
2020. (accepted)
choukh, M. Haney, F. Sheldon, and D. Conte de Leon, “ISAAC: The Idaho CPS
smart grid cybersecurity testbed,” IEEE Texas Power and Energy Conference
(TPEC), 2019.
cyber-physical system model for IEEE 14-bus power system security,” 13th Inter-
and risk assessment for cyber-physical control systems,” 14th IEEE International
High-level and extensible system for training and infrastructure risk assessment,”
1.7 O R G A N I Z AT I O N O F T H E D I S S E R TAT I O N
related work that used machine learning methods for detecting SFDI attacks on the SE.
SFDI attacks on state estimation, and the SFDI attack process. Chapter 4 presents a
taxonomy and a survey of SFDI attacks in smart grid. Chapter 5 describes the security
analytics based SFDIA detection framework put forward in this dissertation. Chapter 6
describes the use of supervised methods GLM, GLB, and DRF to detect SFDIA and
presents the corresponding results. Chapter 7 describes the use of stacking ensemble
models to detect SFDIA and presents the corresponding results. Chapter 8 describes
the use of artificial neural networks and corresponding results. Chapter 9 describes
the use of of unsupervised elliptic envelope method and corresponding results. The
dissertation is concluded in Chapter 10, followed by the References. Detail results from
11
the machine learning model selections and evaluations in Chapters 7–9 are presented
CHAPTER 2
R E L AT E D W O R K
2.1 INTRODUCTION
The literature is replete with detection of SFDIA using different machine learning
methods, however none has described the complete process using security analyt-
ics. Therefore, in this chapter, a review of the literature for machine learning-based
approaches to detect stealthy false data injection attacks on the state estimation of
power grids is presented. Table 2.1 shows the learning class, algorithms used for
feature selections, algorithms used for training, and how the datasets were generated
or obtained for each of the works reviewed. The table also lists the metrics used by
2.2 D E T E C T I O N U S I N G S U P E RV I S E D L E A R N I N G
In supervised learning, labeled data, i.e., a training set of examples with correct re-
sponses or ground truths, is provided and based on this training the machine learning
algorithm generalizes (i.e., learns the patterns in the training data) respond correctly
to input sets that are unlabeled. In this section, the supervised methods used for
detecting SFDIA on the state estimation in power transmission systems are described.
al. [88] is one of the early works that used machine learning for detection of SFDIA.
They ran multi-layer perceptron (MLP), k-nearest neighbours (kNN), support vector
machine (SVM), sparse logistic regression (SLR), and AdaBoost and gave comparative
T A B L E 2 . 1 : Literature review of machine learning-based detection of false data injection attacks on the state estimation.
performances of these models. They used simulated data from MATPOWER [132].
ing on the sparsity the performance for different models change a lot. They showed
that with the least sparse attack model AdaBoost can achieve 100% precision.
He et al. [43] proposed a conditional deep belief network (CDBN), one of var-
temporal behavior features of the stealthy FDI attacks. The CDBN they employed
the first hidden layer to extract the high-dimensional temporal features. In all other
hidden layers conventional RBM were used. They used IEEE 118-bus system simulated
using MATPOWER to validate their proposed solution. They showed that the attack
detection increases with increase in number of buses. The highest accuracy value of
98.10% is obtained by a CDBN with 5 hidden layers, which is only a 0.5% improvement
Wang et al. [110] applied the concept of “first difference”, borrowed from eco-
attacks on the measurements. The “first-difference aware” data is then trained using
supervised models kNN, artificial neural network (ANN), SVM, naïve Bayes (NB), and
IEEE 14-bus system. They achieved 99.73% overall accuracy with an ANN model.
scheme with different configurations for detecting stealthy FDI attacks on the state
estimation of a power grid. They used random forest for feature selection and com-
pared performances of the deep learning scheme with three other machine learning
(GLM) and distributed random forests classifier (DRF). Like others they conducted
the experiments using data generated for an IEEE 14-bus system using MATPOWER.
Ahmed et al. [3] proposed two Euclidean distance-based anomaly detection schemes.
The first scheme utilizes unsupervised-learning over unlabeled data to detect outliers
over labeled data to detect the deviations in the measurements. The authors used a
genetic algorithm for feature selection. They have tested the proposed methods on
IEEE 14-, 39-, 57- and 118-bus systems using MATPOWER generated data.
The proposed framework by Niu et al. [87] has two detectors: a network anomaly
detector and an FDI attack detector. For detecting the FDI attacks, they formulated
network (CNN). At the same time a recurrent neural network with LSTM cell is de-
ployed to capture the dynamic behavior of the cyber-networks in the power system.
randomized trees (ERT) algorithm and kernel principal component analysis (KPCA) for
Mohammadpourfard et al. [80] used the idea of concept drift, unpredictable shifts
attacks. They tested their idea using kNN with PCA as feature extractor and evaluated
Ganjkhani et al. [33] took into account the high correlation between the power
flow measurements data as well as among the state variables and proposed a method
16
that uses a recurrent architecture of ANN configured with nonlinear auto-regressive ex-
ogenous (NARX). They simulated an IEEE 14-bus system and used the corresponding
Hamlich et al. [39] used five classifiers, namely kNN, random forest (RF), decision
trees (DT), MLP and SVM, for detecting SFDIA. They collected power flow data from
a physical bus feeding system connected to a power system stabilizer and then used
2.3 D E T E C T I O N U S I N G U N S U P E RV I S E D L E A R N I N G
In this section, the unsupervised methods used for detecting SFDIA on the state estima-
tion in power transmission systems are described. Unsupervised methods are trained
with unlabeled datasets. Instead of relying on ground truths, these methods attempt
to learn the intrinsic properties of the data, and separates the data into regions or
clusters of data that have similar properties. In other words, the algorithms attempt to
learn the hidden patterns in the input data, and later predicts responses to test inputs
Chaojun et al. [20] proposed a method that tracks the dynamics of variations in
the measurement data. They used Kullback-Leibler distance (KLD), that calculates the
for tracking that dynamics. The method detects presence of SFDIA when KLD is larger
than those for the historical data. They generate the attack data for an IEEE 14-bus
system by modifying load data from the New York independent system operator. The
method fails to detect SFDIA in system buses where the change in measurement data
is very small.
17
It’s a two step sparse model. First, the lower and upper bounds of each state variable
are modeled as a dual optimization problem that aims to maximize the variation
intervals of the system variable. Then stacked auto-encoder (SAE), is used to extract
the nonlinear and non-stationary features in electric load data. The 6-layer SAE had
logistic regression as the final layer. They also used MATPOWER to simulate 9-bus,
14-bus, 30-bus and 118-bus systems, and generated attack data using MATLAB. Their
dataset is also quite small, 35,000, for any deep learning training. The training phase
takes just over 12 minutes. While this work deals with data in the power systems, the
goal of the paper was to predict electric load forecasting more accurately, not to detect
SFDI attacks.
Hao et al. [41] proposed a sparse PCA-approximation based model to detect stealthy
FDI attacks. In this model, identification of real measurements with the availability of
sparse data sets is achieved by using recovery functions. The recovery function’s accu-
racy is inversely proportional to the sparsity of available data. As such, this model falls
short at identifying FDI attacks when data is too sparse to produce reliably accurate
recovery functions. They evaluated their approach on IEEE 9-, 14-, and 57-bus systems
simulated with MATPOWER. They showed that the probability of successful detection
of attack increases if the system’s error tolerance performance is relaxed. If the system
can tolerate a false alarm rate of 10%, then attack detection accuracy approaches
100%.
Ahmed et al. [5] utilized unsupervised learning method isolation forest to detect
FDI attacks–they call it covert data integrity assault–using simulated data generated by
MATPOWER. They reduce the dimensionality of the data using principal component
analysis (PCA). In order to demonstrate that isolation forest performs better, they
18
compare the results with performance of a few other learning models namely SVM,
kNN, NB and MLP. They did not report how long it took to train the models. It is
unexpected that their results of isolation forest is better than the other models which
are all supervised models. That is because supervised models generally perform better
in terms of accuracy than unsupervised models on the same dataset. They showed
Mohammadpourfard et al. [79] used Fuzzy C-Means (FCM) too detect SFDIA by
measuring and comparing deviations in the probability distributions in the data. They
used PCA for feature reduction. They generated the attack data for IEEE 14-bus and
IEEE 9-bus systems by replacing load data from the New York independent system
operator with spoofed data. They reported that their method can achieve 92.88%
detection rate.
Ahmed et al. [5] utilized unsupervised learning method isolation forest (ISOF) to
detect FDI attacks using simulated data generated by MATPOWER. They reduced the
that ISOF performs better, they compared their results with those of a few other
learning methods namely support vector machines (SVM), k-nearest neighbors (k-NN),
naive Bayes (NB) and multilayer perceptron (MLP). They did not report how long it
took to train the models. They reported only accuracy, precision and F1-score values.
It is unexpected that their results of ISOF are better than the other models which
are all supervised models. Generally, supervised models perform better in terms of
accuracy and precision than unsupervised models on the same dataset because they are
trained with labeled data. In a separate work, Ahmed et al. [3] proposed a Euclidean
distance-based anomaly detection scheme. The authors used a genetic algorithm for
19
feature selection. They tested the proposed methods on IEEE 14-, 39-, 57- and 118-bus
Yang et al. [121] used one-class SVM (OCSVM), robust covariance, ISOF and local
outlier factor (LOF) methods. They ran these methods using data from a simulated
IEEE 14-bus system. However, the dataset uses only 1000 set of measurements. They
reported only accuracy and precision values for the algorithms which can be mis-
leading metrics for anomaly detection. The most relevant metrics for sparse attack
2.4 D E T E C T I O N U S I N G S E M I - S U P E RV I S E D L E A R N I N G
In semi-supervised learning, the models are trained using a small amount of labeled
data along with a large amount of unlabeled data during training. These methods are
suitable when labeled data is sparse and the training with small amount of labeled
data can considerably improve training accuracy. In this section, the works that used
Esmalifalak et al. [29] attempted to use two methods for detecting SFDIA. The first
of the two models they proposed utilizes the multivariate Gaussian semi-supervised
learning algorithm and the second model utilizes a distributed support vector machine
(SVM) based algorithm, which requires no supervised learning. Both models use
non-linear classifier with a Gaussian kernel to define the attacked and the safe modes’
boundary. In their simulation to generate data using MATPOWER, they collect active
power measurements from each transmission line. However, the dataset they use is
too small, namely 1000. They report that the semi-supervised model can attain up to
20
82% and the SVM-based method can attain up to about 78% of F1-score. This work is
SFDI attacks on SE in transmission systems, and they demonstrate that this approach
the dimension of the data using PCA, 2) used a positively labeled set to build mixed
and finally 4) used an unlabeled dataset for evaluation. They used an IEEE 118-
solution by comparing its performance with those of SVM and MLP. They reported
Ozay et al. [88] used semi-supervised SVM (S3VM), variation of supervised SVM,
to detect SFDIA. S3VM is based on the assumption that the input data in the same
cluster have the same labels and the difference in number of input data in sub-clusters
is not large. However, since SFDIA is sparse, yielding an imbalanced dataset, S3VM
does not work very well when sparsity is high. The authors conducted experiments on
Chakhchoukh et al. [19] proposed a detection method using a newly developed ma-
chine learning technique known as the density ratio estimation (DRE). The DRE [101]
sion or an attack model. Given two sets of samples from two different distributions,
DRE learns the ratio between the two probability density functions. Here the authors
use data from “normal” operation of the power system as one set of samples and the
data that might be contaminated with noise, bad-data or false data injection attacks as
another. They demonstrated the method using simulation of an IEEE 118-bus system.
21
They reported that the DRE method can detect 100% of the attacks. However, they
In reinforcement learning, the machine learning model uses an agent that facilitates
learning in an interactive environment by trial and error using feedback from its own
actions and outcomes. In reinforcement learning the goal is to find a suitable action
model that would maximize the total cumulative rewards and punishments of the
agent.
tested their proposed solution using MATPOWER-generated data for an IEEE 14-bus
system.
2.6 S U M M A RY
In this chapter, a review of the works that use machine learning to detect SFDIA is
1. None of the works combined both supervised and unsupervised models together
in one solution.
2. A few works mentioned here used multiple classifier models as individual models,
so far.
4. Almost all the works used simulated datasets for testing their models. Three
works used power flow data from power grid, but they added synthetic attacks
5. Most of the attack data are generated randomly. However, in real-life an ad-
versary would craft the SFDIA intelligently considering the system dynamics.
Unless tested with real attack data, the performance of the methods against such
CHAPTER 3
3.1 INTRODUCTION
This chapter briefly describes state estimation in power transmission system and the
mathematical formulation for stealthy false data injection attacks on static state esti-
mation in power systems. The chapter also describes the SFDI attack process.
3.2 S TAT E E S T I M AT I O N I N P O W E R T R A N S M I S S I O N S Y S T E M S
State estimation (SE) at the transmission system in electric power grids is a key
the best estimate of the values of the system’s unknown state variables, i.e., voltage
magnitudes and phase angles of the system buses, from the measurements available
from the network model and sent by the SCADA system to the control center. State
estimation is run in every few seconds to a few minutes. The functions of the state
estimator include identifying and correcting anomalies in the data, suppressing any
bad data, and refining the measurements. Finally it gives a set of state variables that
energy management system (EMS). Figure 3.1 gives the data flow in a typical state
estimator.
abnormal measurements by checking for residuals that do not obey the Gaussian
24
assumption. While the standard residual analysis tests can identify the presence of
errors, it may not detect “stealthy” FDIA because an attacker familiar with the power
transmission system topology information can carefully craft the data amounts to be
injected in a way that the residual of the original measurement vector remains the
same as the residual of the measurement vector with the injected data.
The following EMS tools are dependent on the state variable values estimated by
the SE.
of the most important tasks of the Energy Management Systems (EMS). Used
25
for the purpose of fast estimation of system stability right after outages, by bulk
ule called Unit Commitment Schedule which tells us beforehand when and which
units to start and shut down during the operation over a pre-specified time, such
that the total operating cost for that period becomes minimum.
power system and find the combination of the flows that is operationally and
economically optimal.
Locational marginal pricing tools are used to price out the cost of electricity for
This section introduces the mathematical formulation for stealthy false data injection
from the SCADA units at chosen time snapshots and communicating those to the
control center every few seconds to a few minutes. These measurements are power
flows and injections, as well as voltage magnitudes. The AC static SE estimates the
(state) vector x ∈ Rn that contains phase angles and voltage magnitude at the different
buses, where n = 2k − 1 and k is the number of buses in the system. The slack bus
phase angle is assumed to be the reference and is fixed to 0. The state vector obeys
26
z = h( x) + e (3.1)
The nonlinear vector function h(·) is computed from the grid topology and the pa-
rameters for transmission lines, transformers and other devices. The error vector
from SCADA units. The AC SE is executed using an iterative algorithm based on the
weighted least squares (WLS) [2] to compute and estimate the vector x, i.e.,
where Hk] = ( Hk> R−1 Hk )−1 Hk> R−1 , the matrix Hk is the Jacobian of h with respect to
Let ∆zk = zk − h( xk−1 ) be the kth residual vector. After the convergence of the
algorithm, i.e., once k x̂k − x̂k−1 k < δ for some chosen threshold δ > 0, the obtained
checking for residuals that do not obey the Gaussian assumption. These abnormal or
bad data could be due to natural failures such as sensor or communication misbehavior,
or to FDI attacks. The most practical bad data detection rules are known as the chi-
square test (χ2 ) and the ”3σ” rejection rule [2]. The iterative algorithm is equivalent
to an estimation that is run iteratively after linearizing the regression in each step. The
z = Hx + e (3.3)
27
If contamination occurs due to an FDI attack then the measurement vector z re-
ceived at the control center is replaced by z a with z a = z + Hc. The obtained new state
detect such contamination by analyzing the residual (i.e., the difference between the
measurement vector z and the calculated value from the state estimation, i.e., z − H x̂).
In the largest normalized residual test, if the largest absolute value of the elements in
3.4 S T E A L T H Y F A L S E D ATA I N J E C T I O N AT TA C K S
In the case of FDI attacks, if the injected data are large enough the conventional
residual tests can detect them and these are called non-stealthy FDI attacks. In the
non-stealthy case, the measurement matrix H is not known to the attackers and they
simply generates random attack vectors and manipulate the meter readings.
On the other hand, if the attackers are familiar with the power system topology
information or know the measurement matrix H, they can carefully craft the data
amounts to be injected in a way that the residual r of the original measurement vector
z remains the same as the residual r a of the measurement vector z with the injected
data z.
r a = z a − H x̂ a = z − H x̂ = r (3.4)
These are called stealthy FDI attacks as they cannot be detected using the conventional
Figure 3.2 shows the rough relationship between bad data and false data. The
middle band can have both bad data and false data. The false data that fall into the
middle band are detected by the state estimation’s bad-data detector. The false data
that fall in the bottom band, i.e., the stealthy false data, cannot be detected by the
bad-data detector.
3.5 S T E A L T H Y F A L S E D ATA I N J E C T I O N AT TA C K P R O C E S S
False data injection attacks can be carried on different parts of the power grid, e.g.,
However in this research proposal, SFDI attacks only on the state estimation in the AC
power transmission system are considered. Figure 3.3 shows a simplified diagram of
29
the power grid with the transmission system, the SCADA and the wireless commu-
nication links. It also indicates the possible attack vectors. An attacker can break
into the remote sensors associated with the buses and modify the measurement data
The following are typical steps an adversary may follow for an SFDI attack:
(a) If the adversary is an outsider, they will hack into the system using one
system either using the means above or using social engineering ploys. This
malware may have the ability to steal system information, particularly the
system topology.
(c) If the adversary is a trusted insider, then s/he may already have the access
(a) After the adversary has gained access into the system and obtained neces-
sary system information, they can now surreptitiously change the measure-
(b) The operator and the state estimator assume that the data is correct and
(c) Since the state variable values do not represent the actual state of the
system, calculation by any of the post-SE tools will be incorrect. This will
major disruptions.
The goal of the attacker is to disrupt the operation of the transmission system
leading to a failure in one or more component or bus, which may even trigger a cascade
of failures, i.e., tripping of breakers because of power overload, and causing localized
3.6 S U M M A RY
In this chapter, the static state estimation process for power transmission systems,
the mathematics of bad-data in PTS, what constitutes SFDIA, and the SFDIA process
are described. State estimation process calculates the values for bus system variables
using the matrix of measurements from different RTUs. The bad-data detection tools
can identify and rectify anomalous data and spoofed data. However, there is a class of
spoofed data that cannot be detected by bad-data detectors. These stealthy false data
injection attacks can be a second line of attack in a coordinated attack on the power
grids.
32
CHAPTER 4
TA X O N O M Y OF FA L S E D ATA I N J E C T I O N AT TA C K S
4.1 INTRODUCTION
Extensive research have been done since the identification of false data injection (FDI)
attacks by Liu et al. [73]. Research on FDI attacks mainly focuses on three aspects:
The theoretical research is concerned with developing the SFDI attack schemes,
i.e., the construction of vectors to be injected into actual data matrices. These vectors,
capable of evading detection by the control center, are developed for different compo-
nents of the grid, situations, and constraints. In application research, the focus is on
analyzing the impacts of SFDI attacks on power system operation, mainly on energy
management systems (EMS) and market management systems (MMS), for instance,
towards proposing detection mechanisms and defense strategies from the viewpoint
In this chapter, a survey of the theoretical research and a taxonomy of SFDI attack
models is presented.
4.2 R E L AT E D W O R K S
There are only a few works in the literature that attempted to classify false data
injection attacks on smart grids. None of these works claimed to have developed a
Guan et al. [37] in a limited survey of FDI attacks, described centralized and
and a superficial discussion of FDI attacks on the control system. They did not cover
many other FDI attacks and they limited their survey to only a few works.
Liang et al. [62] did a review of FDI attacks on the power grids. They summarized
the different strategies of constructing valid FDI attacks under various constraints.
They described FDI attacks when 1) attacker has full knowledge of the system, 2)
attacker has incomplete knowledge of the system, 3) attacks are on the system topol-
ogy, and 4) attacks are on the AC power flow. They did not cover attacks on other
parts of the transmission system and on other parts of the power grid. Their work is a
Liu and Li [72] did a broader survey of FDI attacks covering attacks on the trans-
mission system, distribution system and microgrid. However, the paper is more of a
survey of FDIA models, impact analyses and defense strategies, than a taxonomy of
FDIA.
attacks classified based on the power grid components targeted and based on the kind
4.3 C O M P O N E N T - B A S E D AT TA C K C L A S S I F I C AT I O N
In the following few sections, a taxonomy of FDI attacks based on the targeted com-
ponents of the power grid is presented. Section 4.4 presents FDIAs on transmission
systems, Section 4.5 on distribution systems, Section 4.6 on microgrids and Section 4.7
F I G U R E 4 . 1 : Diagram showing taxonomy of stealthy false data injection attacks on the power grid.
34
35
4.4 AT TA C K S O N T R A N S M I S S I O N S Y S T E M
False-data injections in power grid was identified as cyber-attacks by Liu et al. [73] for
the direct current (DC) power flow model in the transmission system. They argued that
if an attacker can construct a false data injection attack such that the overall residue
of the system will not increase, the attack on measurements can bypass the residual
test for bad-data detection and constitute a successful attack. Rahman and Mohsenian-
Rad [81] formulated the false-data injection attacks for the alternating current (AC)
The FDI attacks on the transmission system can be broadly categorized in two
groups depending on whether the attacker has full knowledge of the system topology
including system parameters, power grid topology, state estimation algorithm, bad-
data detection mechanism, and transmission-line admittance values, etc., and has the
ability to modify data at all the meters, they can adjust the false data injection attack
vector in a way that the attack remains undetected and successfully passes the residue-
After the initial identification by Liu et al. [73] for the DC model, a number of works
followed. Kosut et al. [56] proposed a graph-theory based algorithm to construct false-
data injection attacks. Kim et al. [55] proposed a data-framing FDI attack where it
was shown that an attacker can inject data in such a way that the energy management
injection source. Yang et al. [123] proposed a collective sparse attack strategy where
state variables in the same cluster are attacked by the same attack vector. Wang
and Ren [111] studied the situation when FDIA is on transmission system and the
topology is not fixed. Assuming that only one line can be cut off per time interval,
they developed necessary and sufficient condition under which successful stealthy FDI
Hug and Giampapa [48] analyzed stealthy false data attacks against the AC state
estimation and showed that an attacker can launch a successful stealth attack if they
have the full network information of a power grid. They developed a graph-theoretic
measurement at a bus. Since the injected power at a zero-injection bus must be zero
to ensure power flow feasibility, any malicious modification of the readings of these
measurements will be detected with a relative high probability. Thus they concluded
that the zero-injection buses can improve the resilience of a power system to false data
injection attacks.
37
Chakhchoukh and Ishii [18] identified two attack scenarios on both AC and DC
state estimations. In one scenario, called masked attacks, attacks are possibly uniden-
tified with a convergence to an arbitrary state unknown to the attacker. In the second
attacker.
systems, the status of each transmission line, whether the line is in service or not,
is sent to the control center in real time. Under normal operations, the topology
lines. However, during transmission of this status data, an attacker can intercept
and modify the status of a line sent to the control center exploiting the vulnerability
of communication networks. This alteration in the status data can lead the control
center to erroneously believe the system is operating under a topology different from
that in reality. This can lead to serious implications: a grid that is under stress may
measures to ensure stability. On the other hand, an operator may deem a normally-
operating grid to be under stress and can take costly remedial actions including load
shedding. To construct stealthy attacks, the attacker modifies the meter data for
transmission systems and the network topology data simultaneously in such a way
Kim and Tong [54] derived a necessary and sufficient condition for the existence
of an undetectable attack for strong adversaries who has knowledge of the complete
38
system, in other words can observe all meter and network data. They presented a
simple undetectable attack, called state-preserving attack, where the attack intention-
ally preserves the state in order to have a sparse attack vector. Under this attack
an adversary can simulate the physical outage of a transmission line without actually
severing the line where a pair of additional power increments is injected into the power
measurements (and not the phase angles) at the terminal buses of the targeted line.
The topology attack model in Kim and Tong [54] assumes that the injected false
power at a bus is infinite which is quite impractical since a control center operator
usually has some knowledge about the load distribution of a power grid and can
predict future loads using load forecasting. To overcome this drawback, Liu and
for determining a feasible attack region, thereby a local topology attack model that
Liu and Li [71] and Li et al. [60] proposed a topology attack model to mask the
physical outage of a transmission line injecting false data into some measurements
such that the new power flow is consistent with the case in which the line is on outage.
This will cause the control center to believe that this line is still in service although it
is physically on outage.
Phasor measurement units (PMU) are used to estimate real-time phasor and sequence
magnitude and phase angle of an electrical phasor quantity, such as voltage or current,
system (GPS). These measurements are transmitted to the control centre and are used
mote points on the grid with high accuracy. However, dependence of synchronized
measurements on GPS signals has been identified as an attack surface. Two kinds of
attacks have been identified. Zhang et al. [131] and Jiang et al. [50] demonstrated
that PMUs are vulnerable to spoofing attacks that provide false time stamps on the
PMU measurements. The spoofed time stamps cause two types of errors: the phase
angle error and the time-of-arrival (TOA) error. The phase angle error will render the
operator unable to or to wrongly detect the outages of transmission lines; the TOA
error will result in miscalculation of the locations of disturbance events. The spoofing
can also attack the clock offset of a PMU to disrupt voltage stability monitoring.
Liu and Li [68] showed that the line outage detection using the PMU data can
maximize the residual of the outage line in the detection algorithm. When a line
outage occurs, the operator calculates the expected phase angle changes due to this
outage based on received real-time measurements and compares the values to these
real phase angle changes measured by PMUs to locate the fault line. With the PMU
measurement data changed, the operator is forced to take incorrect remedial actions
causing disruptions.
So far the review dealt with FDI attacks that are based on the assumption that the
attacker knows the complete configuration information of the power grid and power
information and therefore it is not possible to launch a successful FDI attack on power
grids. However, research has established that it is still possible for an adversary to
launch a successful FDI attack even if they have incomplete information about the grid.
In this section, a review of research that developed FDI attack models with incomplete
system information will be presented. The focus of this line of research is to ascertain
enough system information from grid operation to launch a successful FDI attack.
Esmalifalak et al. [30] observed that topology information is embedded into the corre-
lations among power flow measurements when the system parameters (e.g., active or
passive loads) vary within a small dynamic range, and that the topology information
can be extracted from this. In practice, power system topology remains the same unless
there is a reconfiguration of the system. Hence, due to the slow dynamic nature for a
short period of time, the equivalent knowledge of the topology can be revealed using
component analysis (ICA) based algorithm to obtain topology information from power
flow measurements.
Rahman and Mohsenian-Rad [93] formulated FDI attacks with incomplete infor-
mation from both the attacker’s and system operator’s viewpoints and introduced a
novel vulnerability measure that can compare and rank different power grid topologies
against such attacks. Rahman and Mohsenian-Rad [93] and also Giani et al. [34]
developed an approach to construct FDI attacks by splitting the grid into several
subnetworks, and increasing or decreasing phase angles by the same amount incre-
mentally in each subnetwork. Since the bus angles in the subnetworks change by
the same degrees, the power flows in the area covered by the subnetworks will not
41
change. They termed this as “local attack” where the modified phase angles constitute
the attacks.
Kekatos et al. [52] showed a way to identify grid topology using electricity prices.
algorithm to estimate the grid Laplacian matrix from the locational market prices
(LMPs) using a regularized maximum likelihood estimator, since under the DC model,
patch problem. The authors solved the problem using an ADMM-based algorithm and
demonstrated that the estimated Laplacian matrix is close to the real topology. Moya
et al. [84] introduced the concept of correlation index for wide-area measurement
Yang et al. [123] developed three attack models, namely least-effort attack model,
optimal attack model and optimal attack model for large grid. First an adversary can
launch the least-effort attack in order to find the sparsest attack vector so that the
After gaining knowledge of some state variables, the adversary can launch the linear
transformation based optimal attack to launch a full-fledged FDI attack on the grid.
The third model is based on a heuristic algorithm that can be used to launch a full-
Yu and Chin [124] proposed a method to generate sparse false data injection
attacks of the general kind on both the DC and AC cases by applying the principal
component analysis (PCA) to the measurement data matrix into a new subspace, pre-
technique which can transform the correlated observations into uncorrelated variables
42
known as principal components. These orthogonal principal components are the linear
combinations of the original observations. The authors construct FDI attacks based
on the data of the new projected space. These attacks are generated without the
knowledge of the measurement matrix and the assumption regarding the distribution
of state variables. In order to generate the false data injection attack vector, they
applied PCA on the vector of actual measurements, and multiplied it with the transpose
Anwar et al. [8, 9] extended the idea of Yu and Chin [124] to construct an attack
from only the measurement signals even in the presence of gross, non-Gaussian, errors.
They further developed a sparse optimization based blind (no system knowledge)
stealthy attacks construction strategy that is stealthy and successful even in the pres-
tion errors.
AC power flow equations are highly nonlinear and it is much more difficult to formu-
late an attack for the AC model with incomplete network knowledge. Rahman and
Liu and Li [70] extended the local attack model for DC systems in [67] to the
AC case by replacing the estimated voltage magnitudes at the boundary buses in the
attacking region with the corresponding voltage measurements and setting the power
flows on the tie-lines calculated using the estimated voltage magnitudes and phase
Kim and Tong [54] proposed a heuristic method for constructing FDI attack on network
topology for the case when the adversary has only limited knowledge of the system.
They demonstrated that the physical outage of a line can be simulated by launching the
this line. Kim and Tong [54] also demonstrate that a topology attack has significant
Tan et al. [102, 103] formulated stealthy false-data injection attacks on the sensor
used in power grids to maintain the grid frequency at a nominal value. The authors
show that, if an attacker can gain access to sensor data and a few system constants,
they can learn the attack impact model and achieve the optimal attack. False data
injection attacks on the sensor measurements for AGC can cause frequency excursion
that triggers remedial actions, by control center operators, such as disconnecting loads
or generators, that may lead to blackouts and potentially costly equipment damage.
Chen et al. [21] investigated the impact of FDI attacks on the static security assessment
(SSA), a process that determines if the power system is in a secure or alert (insecure)
state; the secure state implies the load is satisfied and no limit violation will occur
(i.e., outages of one or several lines, transformers or generators). They showed that
44
successful FDI attacks can manipulate the results of SSA. They identified two scenarios.
In fake secure signal attack, the system operator is made to believe that the system is
operating in a secure condition when actually it is not. In fake insecure signal attack,
4.5 AT TA C K S O N D I S T R I B U T I O N S Y S T E M S
The existing research on FDI attacks against state estimation in transmission systems
on the distribution system, the attacker must know the estimated state of the system.
Lim et al. [63] was among the first to investigate false data injection attacks in the
power distribution systems, even though they didn’t call it FDI attacks. FDI attacks
on the distribution systems are realized by tampering with the messages between the
server and the feeder remote terminal units (FRTU). These messages contain voltage
and current data and control commands. The authors categorized the attacks into
three distinctive attacks. First, altering the contents of these messages and sending to
the FTRUs. The altered messages can control automatic switches in the system ma-
liciously, eventually causing power outages. Second, creating spoofed messages and
to the FRTUs. Third, sending a previous message again, thus effectively delivering
incorrect time-varying information that reflects system status and actions. This is a
replay attack.
45
Guo et al. [38] showed that an attacker on the distribution system can attack
the feeders or FRTUs, subsystems or the customers. Deng et al. [27] showed that
an adversary can approximate the system state based on the power flow or injection
measurements and construct attacks on the state estimation of the distribution system.
4.6 AT TA C K S O N M I C R O G R I D S
Microgrids are local and decentralized low-voltage electric power grids that can dis-
connect from traditional power grid for autonomous and self-sufficient operations.
Lin et al. [64] investigated FDIA against the distributed energy routing process
in microgrids. They considered several general attacks, in which the adversary may
manipulate the quantity of energy supply, the quantity of energy response, and the
link state of energy transmission. They quantitatively demonstrated that the false data
injected by those attacks cause imbalanced demand and supply, increase the cost for
Zhang et al. [128] investigated the impacts of FDIA on the dynamic microgrid
partition process. They identified three FDI attacks on microgrids: i) claim less energy
than what can be provided, ii) claim more energy than what is required, and iii) both
i) and ii) at the same time. They show that for all these situations injected false data
Hao et al. [40] observed aberrant operation of a microgrid when false data is
injected into the voltage controller of the substation. Liu et al. [71] proposed an
FDI attack model against control signals of solar photovoltaic (PV) and energy storage
system (ESS) controls in microgrids, Li et al. [59] studied FDIA against the control
46
signals of the inverters in microgrids, and Beg et al. [14] considered false data injection
in microgrids by categorizing the attacks by their utilization levels and monitored the
Liu et al. [66] considered FDI attacks on communication links in a microgrid and
theoretically formulated the conditions when the data injection constitutes i) a false
attack, and ii) a missed or undetected attack. They further defined two classes of
undetectable attacks: i) zero trace undetectable, and ii) non-zero trace undetectable.
4.7 AT TA C K S O N A D V A N C E D M E T E R I N G I N F R A S T R U C T U R E
tween smart meters located at the consumers and utility companies, allowing infor-
mation about consumption, outages, and electricity rates to be shared reliably and
efficiently and for efficient, accurate and advanced monitoring and control.
Grochocki et al. [36] investigated FDI attacks as part of larger scheme of cyber-
attacks on smart meters, neighborhood area networks (NAN) and home area networks
(HAN). They developed an elaborate attack tree after performing threat analysis. FDI
Anwar et al. [10] described a relationship between the power system stability
indices and the FDI attacks when the attacks are injected through smart meters. They
identified the level of vulnerabilities of each smart meter in terms of different degrees
of FDI attacks.
47
4.8 I M PA C T - B A S E D AT TA C K C L A S S I F I C AT I O N
Yuan et al. [125] developed the concept of load redistribution (LR) attack that can
(SCED). The SCED minimizes total system operation cost (e.g., generation cost, load
a successful LR attack, the attacker makes sure that the attack amount at a load
measurement must be limited within a small range to reduce the situational awareness
of the system. Under an LR attack, the falsified SCED solution leads the operator to
redistribute the loads in a way that may drive the system to an uneconomic operating
immediate or a delayed fashion. The authors classified the LR attacks into i) immediate
attack and ii) delayed attack. Yuan et al. [126] further developed the concept of LR
attack by developing models to quantify the damage of both immediate and delayed
attacks.
Liu and Li [69] and Liu et al. [67] investigated local load redistribution attacks
with incomplete information. Xiang et al. [115] demonstrated that the LR attacks
have a non-negligible impact on the power system reliability. The authors modeled
the FDI attacks using semi-Markov modeling for both non-encrypted and encrypted
communication and quantified the influence of load redistribution (LR) attack on the
long-term power supply reliability. Xiang et al. [116] showed that LR attacks can
Garcia [81] showed that the direct load control command signals and indirect load
control price signals can be attacked by launching load altering FDI attacks.
48
FDI attacks can be constructed to manipulate the locational market pricing (LMP) by
modifying the measurement matrix of the state estimation, because the real-time LMP
is calculated using estimated congestion pattern obtained from the state estimation.
Xie et al. [117, 118] were the first to study this kind of attacks on the electricity
vector to maximize profit in the real-time market. They also developed a heuristic-
Jia et al. [49] took into consideration the effects of both the RTU measurements
and the system topology state on real-time LMP and proposed formulation to calculate
the injected false-data vectors for causing congestion pattern that affect the LMP most.
Choi and Xie [22] studied the impacts of topology errors on LMPs. Esmalifalak et
al. [31] studied the effect of FDIA on power market using a game-theoretic approach.
impact of topology attacks on the integrity of optimal power flow (OPF) and, hence,
It was seen before that the smart meters are susceptible to FDI attacks and an adver-
sary can cause a major disruption in the system by attacking a set of critical meters.
McLaughlin et al. [78] demonstrated how energy theft can be achieved by altering
While investigating FDI attacks against the distributed energy routing process in
microgrids, Lin et al. [64] identified the energy deceiving attack. The authors showed
49
how an attacker can manipulate the memory of smart meters in the microgrids and in-
ject erroneous energy demand and supply messages to cause energy deceiving attacks.
4.9 S U M M A RY
False data injection attacks or covert data integrity attacks can be targeted on different
parts of the smart grid, and for different purposes. This chapter presented a compre-
hensive review and a taxonomy of different FDI attack models classified based on the
CHAPTER 5
S E C U R I T Y A N A LY T I C S F R A M E W O R K
5.1 INTRODUCTION
In this chapter, the security analytics framework that is used to detect SFDI attacks
has been around for just about a decade, mostly to protect from threats in the cyber
systems. In this dissertation research, security analytics is being used to detect cyber-
5.2 S E C U R I T Y A N A LY T I C S
occurs. Security analytics often, but not always, employs big data analytics.
SFDI attacks. The central part of the framework is to use machine learning models
for attack detection where machine learning models are first developed by letting
the models learn the data behavior, i.e., relations in different types of data instances,
through the training process. Once the models exist, real-time data is classified as
machine learning is that the training process typically requires significant time and
51
1. Risk Analysis
2. Threat Modeling
3. Data Generation
4. Data Engineering
5. Feature Reduction
A process flow diagram of the scheme is given in Figure 5.1. The process flow
starts with a brief risk analysis, followed by threat modeling where the attack models
are identified. These attack models are used when data generation module creates the
synthetic datasets. Data engineering normalizes the data with standard scaling and
balances the datasets for the supervised models. The scaled and balanced datasets are
used by the feature reduction module. All the machine learning methods separately
use the feature-reduced datasets to train. The trained models are tested and evaluated
using standard evaluation metrics. This model selection phase produces the best per-
forming models which are to be deployed to detect SFDI attacks. The deployed models
52
monitor and analyze the oncoming SE measurement data vectors for the presence of
The following sections describe the components of the security analytics frame-
work.
5.3 R I S K A N A LY S I S
Risk is the level of impact on the organization’s operations, operational assets, and
personnel given the potential impact of a threat and the likelihood of the threat occur-
53
ring. In other words, risk is the possibility of damage or harm and the likelihood that
damage and harm will be realized. Risk analysis is an effort to identify vulnerabilities
and their related threats, assess the potential costs of exploitation, and determine
In this dissertation, it is considered that SFDI attacks have already been identified
as threats to the power transmission systems, since it is assumed that the adversaries
have already intruded into the system by way of compromising sensors or communica-
tion channels and have adequate knowledge of the system topology to craft effective
SFDI attacks. As shown in Section 4.8, the impacts of the SFDI attacks on the system
depend on the particular kind of SFDI attack being constructed. A risk analysis of SFDI
attacks will have to consider the type, scope, and intensity of the attacks, the assets and
services affected, the duration of the affected services, severity of physical, economic,
and intangible damages, etc. Risk analysis of SFDI attacks on state estimation in power
5.4 T H R E AT M O D E L I N G
A threat is an agent that can potentially compromise the normal operation of a system.
The threat agent can be a human actor, or an event or circumstance. The agent can be
a malicious actor or a benign but naive actor accidentally causing a security incident.
Threats are identified, categorized, and prioritized using the threat modeling process.
For the case of SFDI attacks, the threat agents are malicious, purposeful, and
highly resourceful. In this limited threat modeling, the different ways adversaries
can constitute attacks on the measure data are identified. These attack models are
described below.
54
1. The attacks are only on the static state estimation measurement data in AC power
transmission systems.
2. The attacks are targeted attacks and not random attacks. That means the same
set of compromised RTUs are attacked for the entire duration of the attack.
3. The attacker has partial knowledge of the system topology. They know the
4. The attacks are general attacks without intending any particular impact.
As described in Chapter 4, the SFDI attacks can occur at different points of the trans-
mission system enabling the attackers according to the access point. An attacker
can break into one of the remote sensors associated with the buses and modify the
and modify the network packets. In order to cover this broad scope of SFDI attacks,
three attack models for the investigation in the proposed research are considered.
These attack models are described with the help of Figure 5.2. It shows a simplified
diagram of a 14-bus transmission system, along with the cyber network that consists
In this scenario, it is assumed that sensors in one of the substations in the system
are compromised by an attacker and hence the attacker can change the measurement
values corresponding to the sensors in that Substation only. For example, if Substation
SS-1 in Figure 3.1 is hacked then values from all the sensors in that substation can be
modified.
A control center in the system collects the measurement values from the sensors in
a number of buses, collates those and relays to the central SCADA center. If an
attacker can intrude into one of the control centers, they can modify the measurements
values that go through this control center. Referring to Figure 5.2, if the attacker gets
access into control center CC-1, they can modify the measurement values from sensors
This scenario covers the cases when an outside adversary gets access into the enterprise
SCADA center or when the attacker is a trusted insider hereby having full access to the
matrix of measurement values. This later scenario can also be a malware that has gone
into the SCADA center and has the ability to change the measurement values. In this
model, since the attack happens at the central SCADA center, the adversary has the
capability to change any or all the measurement values from all the sensors.
57
5.5 D ATA G E N E R AT I O N
For this research, instead of collecting, data was generated because of lack of availabil-
ity of real data. Real SFDIA data, even if they exist, will be very few, because SFDI
The generation of data for this research is a two-step process: 1) State Estimation
(SE) measurement data representing normal operation of the power system and 2)
For generating power flow SE measurement data, standard IEEE 14-bus and 57-bus
systems were simulated using MATPOWER [132]. The measurements are obtained
from solving power flows using the MATPOWER and adding Gaussian measurement
noise. The loads are considered to vary randomly around their average values.
The IEEE 14-bus system has 5 generators and 11 loads [106], as shown in Fig-
a total of 122 measurement features. The IEEE 57-bus system has 7 generators and
41 loads [106], as shown in Figure 5.4. The measurements for the 57-bus system are:
188 active power-flows, 57 active power-injections, 188 reactive power flows, 57 reac-
features.
58
The attack generation program is written using MATLAB. The buses considered com-
promised are selected randomly following the three attack models described earlier.
Then the measurement values corresponding to those buses are replaced by attack
data generated by this program. 10% of ‘normal’ data were randomly chosen to be
‘attack’ data.
The generated measurement vectors with attacks, i.e., modifications in the mea-
surements, were used for state estimation using state estimation simulation in MAT-
POWER. As the state estimation successfully processed the data vectors, it demon-
strated that bad-data detector could not detect the attacked data vectors. This shows
5.6 D ATA E N G I N E E R I N G
During the data engineering, also called data wrangling or data preprocessing, phase,
The dataset generated did not have any missing data or invalid data; consequently,
no data imputation or cleaning was required. The data types of all the features in the
dataset are numeric, except for the class label which is either ‘normal’ or ‘attack’. As
part of preprocessing, all the ‘normal’ strings were changed to 0s and ‘attack’ strings
to 1s.
5.6.1 Scaling
Standard scaling was applied to the dataset. In standard scaling, the features are
normalized by removing the mean and scaling to unit variance. Standard scaling
replaces the data values in a feature by their z score. For a value x, the z score is
calculated as: z = ( x − µ)/σ, where µ is the mean and σ is the standard deviation of
5.6.2 Balancing
The dataset generated contains 90% normal data and 10% attack data implying that
the dataset is imbalanced. Supervised classifiers perform poorly when trained with
imbalanced datasets, especially for the minority class. In this case, the ‘attacks’ are
in the minority class, and the goal is to detect these precisely. Therefore detection
of these minor class ‘attack’ instances in an imbalanced dataset will not be accurate.
To overcome this problem, two popular techniques to balance datasets, namely the
synthetic minority over-sampling technique (SMOTE) and the edited nearest neighbor
(ENN), were applied to over-sample the ‘attack’ sets of data and under-sample the
‘normal’ sets of data respectively [13]. After this balancing act the ratio of major and
minor class samples in the dataset was 3:2. Since the unsupervised models function
61
as outlier or anomaly detectors, the dataset does not need to be balanced for the
5.7 F E AT U R E R E D U C T I O N
The complexity and time for model training increase sharply with an increase in the
number of features, i.e., feature dimensionality, in the dataset, because of the so-
called “curse of dimensionality” [107]. The number of data items, i.e., features, in
SE measurement data increases with the number of buses in the system. For example,
the number of features in the measurement data for an IEEE 9-bus system is 27, for
an IEEE 14-bus system it is 122, for an IEEE 57-bus system is 547, and for an IEEE
It is often observed that not all the features in a dataset contribute equally in
training the models. Hence, the features with the least discriminating properties can
be safely eliminated from the dataset without compromising the model performance.
Feature selection may lead the trained model to maximizing its performance while
In the security analytics scheme presented in this dissertation, random forest classifier
(RFC) [46] was used to rank and select the features in the SE dataset according to
Random forest, a popular classification and regression method, can also be used
ensemble of a large number of decision trees. Each of the trees is constructed over
62
a random extraction of the instances from the dataset and a random extraction of
the features. These trees are uncorrelated since they can’t access all features or all
split the dataset into two such that similar response values end up in the same set. At
the time of this split, a measure of how much each feature contributes in making this
decision is taken. This measure forms the basis of ranking the features according to
their importance. Then, a number of the most important features are retained from
the ordered list of features while the others are deleted from the dataset to obtain a
feature-reduced dataset.
The random forest classifier was used on the dataset to obtain an ordering of the
features according to their importance. A plot showing the feature importance for
the 14-bus-AM1 dataset is given in Figure 5.5. The figure shows that the first 21
features have the largest variances, and therefore only these features were retained in
the dataset as the predictor variables, plus the target variable. Figure 5.6 shows the
feature importance for the 57-bus-AM2 dataset showing that the first 43 features have
largest variances, and therefore these features were retained in the dataset.
64
In summary, the datasets generated in Section 5.5 are scaled and balanced (for
supervised methods only) in Section 5.6, and finally feature-reduced in Section 5.7.
The properties of the twelve datasets after scaling, balancing, and feature selection are
given in Table 5.1. These datasets will be used for model selection, i.e., for training
Table 5.1 shows that the ratio of normal data and attack data, i.e., data tampered
with by the attackers, is roughly 3:2 for datasets used in supervised model selections
and roughly 9:1 for datasets used in unsupervised model selections. It also shows
that the total number of data instances for supervised datasets are more than 100,000.
This is because when SMOTEEN method is used to balance a dataset, the method
increases the minor class instances and reduces the major class instances. At the end,
train on a particular dataset. It is then said that the ‘model’ has learned the patterns or
65
relationships in the data instances using the ‘method’. Many models can be built using
one method by using different datasets or different sets of parameters for the method.
Even if only one dataset and one set of parameters are used, different models can be
built using different ways of training, i.e., depending on how the dataset was split or
Model selection is the process of training and evaluating machine learning methods
and selecting the best model from among a collection of machine learning models for
methods (e.g., decision tree, logistic regression, SVM, kNN, etc.), 2) models of the
same method configured with different hyper-parameter values (e.g., different hidden
layers and units for ANN), and 3) models of the same method and same set of hyper-
In this security analytics framework, all of these three types are used to train the
most robust models. For each of the different machine learning methods, a set of
hyper-parameter values were chosen and the models were trained using 10-fold cross-
validation over randomly divided (resampled) training data. In 10-fold cross valida-
tion, the training dataset is divided into 10 equal subsets. The training is repeated
10 times with 9 subsets and the remaining 1 subset, called the validation dataset, is
used to test the trained model. In every repetition, a different sets of training subsets
and validation subset are used. Figure 5.7 shows the k-fold cross-validation process.
process yields the best model out of several models of one ML method with one set
of hyper-parameter values. Then this process is repeated for different sets of hyper-
parameter values for that one ML method. The best models obtained from each of
these sets of hyper-parameter values are then compared using standard evaluation
66
metrics described in Section 5.8.1. Finally, this entire process is applied to all the
different methods yielding one best model for each of the different methods. The final
best model is then selected by comparing these models using evaluation metrics.
The best models are selected after the models are trained with cross-validation and
their performance are measured against the testing dataset using standard evaluation
metrics described below. Since the goal is to detect attack data, which is in the
minor class, the appropriate metrics to be used to select optimum models are F1-score,
sensitivity, and specificity. The ideal situation would be when one model exhibits
highest values for all these three metrics. However, that is not always the case. Then,
a model is chosen as the best one that gives highest sensitivity along with highest or
very close to highest values for F1-score and specificity. These best-performing models
Chapters 6–9 will elaborate on this section in describing the model selection with
different machine learning methods and will present the results showing how good
A machine learning classification model predicts class labels as output for a given
input dataset. In this case, the class labels are “normal” and “attack”. Depending on
the predicted class labels, the outcomes for binary classification can be categorized as:
1) True positive (TP): when the model correctly identifies an attack, 2) True negative
(TN): when it correctly identifies a normal or non-attack, 3) False positive (FP): when
Predicted Predicted
Normal Attack
Actual Normal TN FP
Actual Attack FN TP
To evaluate the models in this paper, the following metrics derived from the confu-
1. Accuracy = ( TP + TN )/TotalPopulation
5. Specificity = TN/( FP + TN )
6. F1-score = 2TP/(2TN + FP + FN )
68
Accuracy is the percentage of correct identification over total data instances. Pre-
cision, also known as the positive predictive value, represents how often the model
correctly identifies an attack. Sensitivity, also known as the true positive rate (TPR),
recall, or detection rate, indicates how many of the attacks the model does identify
correctly. Sensitivity intuitively gives the ability of the classifier to find all the positive
samples and the precision intuitively gives the ability of the classifier not to label
as positive a sample that is negative. Specificity, also known as the true negative
rate, measures the proportion of actual negatives, i.e., non-attacks, that are correctly
identified as such. F1-score, also known as F-measure, provides the harmonic average
The ROC curve plots FPR on the X-axis and TPR on the Y-axis. This means that
the top left corner of the plot is the “ideal” point, where the FPR = 0, and TPR = 1.
The larger the area under the curve (AUC) the better. The red dotted line indicates the
random classification and has an AUC of 0.5. The ROC AUC score is a measure of the di-
of different models over all possible thresholds, the ROC curves are plotted.
In addition, the run times (i.e., elapsed times) are usually measured for compar-
ing the speed of different training models running the all-feature dataset versus the
reduced-feature dataset.
One or more best models based on a comparison of the standard evaluation metrics
across all models that were tested are then deployed in the system being protected to
detect security issues in real-time. For this case, the model(s) are to be deployed at
69
the power control center within the state estimation process and right after the bad-
data detection. In this dissertation research, the model deployment phase was not
5.10 AT TA C K M O N I T O R I N G
The SE measurement data matrix goes through the steps in data engineering phase,
where it is normalized and feature-reduced. Then it goes through the deployed models.
If any of the models identifies the presence of a potential SFDI attack then the mech-
anism raises an alert for the control center operator to take necessary action. Since,
the models were not deployed in this research, attack monitoring is also left out as a
future work.
5.11 T H E S O F T WA R E
A software was implemented to perform most of the phases of the security analytics
process presented above. This software was used to do data generation, data engineer-
The data generation module was implemented using MATPPOWER [132] and MAT-
LAB, and all the other functionalities were implemented using the Python program-
ming language. Python library scikit-learn [90] was used to write the feature reduction
At this time, the implemented software supports RFC for feature reduction; GLM,
GBM, DRF, ANN, SVM, NB, DT, and LR for supervised learning; OCSVM with two
kernels, ISOF, LOF, and EE for unsupervised learning; stacking ensembles using the
the model performance and printing out the results in tabular forms; and plotting
performance graphs.
5.12 S U M M A RY
In this chapter, the security analytics framework used for detecting SFDI attacks on
the steps in the framework and describes what were done for different steps for
this research, except the “Machine Learning Model Selection” phase. The following
four chapters discuss machine learning model selection using supervised, ensemble,
artificial neural networks, and elliptic envelope methods, and demonstrate how good
the models will be in detecting SFDI attacks using standard evaluation metrics.
71
CHAPTER 6
S U P E RV I S E D M O D E L S
6.1 INTRODUCTION
This is the first of four chapters that represents machine learning model selection phase
(Section 5.8) in the security analytics framework. These chapters present building of
machine learning models using different machine learning methods using the datasets
generated and prepared in Chapter 5. The model selection process ends with deter-
mining the efficacy of the models in detecting SFDI attacks using standard evaluation
This chapter describes the model selection using three popular supervised meth-
ods, namely generalized linear model (GLM), gradient boosting machine (GBM), and
distributed random forests (DRF) and resultant detection strengths of the models.2
6.2 M O T I V AT I O N
and reliably detect SFDI attacks was started with three popular and simple to use
methods. For that reason, a powerful linear method, GLM, a boosting method, GBM,
6.3 T H E S U P E RV I S E D M E T H O D S
This section briefly describes the three machine learning methods used in this chapter.
Generalized linear models (GLM) were first proposed as a general framework for uni-
fying various other statistical models, including linear regression, logistic regression
and Poisson regression [86]. A few other linear regression and classification methods
have been added to GLMs since then. An implementation of the GLM suite generally
includes Gaussian (i.e. normal), Poisson, Gamma and Ordinal regressions as well as
binomial and multinomial classifications. The GLM allows the linear model to relate
to the response variable using a link function and the magnitude of the variance of
ble is a collection of weak classifiers. The mean of predictions by all these classifiers
is deemed as the final prediction. GBM trains many models in a gradual, additive and
sequential manner. GBM overcomes the limitations of the weak classifiers by using
gradients in the loss function which is a measure of how good the model’s coefficients
Distributed random forest (DRF) is another ensemble-based method that works by gen-
tree, using the given dataset. Each of the trees in the forest is a weak classifier built
on a subset of rows and columns. More trees will obviously reduce the variance. The
average prediction over all of these trees is used to make a final prediction.
Each of the six datasets was split into a 7:3 ratio for training and testing dataset
respectively. The training datasets are used to train the three models and the testing
datasets are used to evaluate the models for selection using the standard metrics. 10-
fold cross-validation over randomly divided training data was used during training of
the models.
6.5 R E S U LT S
The numerical results in terms of the evaluation metrics for testing the models using
6.6 D I S C U S S I O N O F R E S U LT S
The following observations are made from Table 6.1 for the evaluation metrics values
1. The performance for all the three ML models are effectively same.
2. The performance of these models for all the six datasets are also effectively the
same.
3. The overall accuracy values for the models are 87%, whereas precision values are
about 98%. However, in a classification problem where the goal is to detect the
minor class occurrences, the most important metrics are the sensitivity which, in
this case, measures the proportion of actual “attacks” that are correctly identified
as such; and the specificity which measures the proportion of actual “non-attacks”
that are correctly identified as such. For these models, the sensitivity values
are about 72%. This indicates that even the any of these model would be able
to detect about 72% of the attacks and the rest 28% will go undetected. The
specificity values for the models are about 98% meaning that the models are able
The corresponding full-feature datasets (122 features for 14-bus system and 547
features for 57-bus system) were also used to train the three models. It was found
that the performance numbers for the models do not change at all for these cases.
However, it was found that the elapsed times for training with the full-feature datasets
for 14-bus system take up to 400% more time, and for 57-bus system take up to 600%
more time.
75
6.7 S U M M A RY
In this chapter, use of the security analytics framework with supervised learning in the
model selection phase was described for detection of SFDI attacks. The supervised
models used are GLM, GBM and DRF. It cannot be confidently said that the models
performed well in detecting the attacks as the average detection rate obtained by
these models is only around 72%. It was found that using the datasets with all the
features do not improve the performance values, however it slows the training down
significantly.
76
CHAPTER 7
S TA C K I N G E N S E M B L E S
7.1 INTRODUCTION
This chapter presents the model selection part of the security analytics framework
using stacking ensemble. It also describes the ensemble framework3 and building of
machine learning models using machine learning methods in the ensemble framework
using the datasets generated and prepared in Chapter 5. The model selection process
ends with determining the efficacy of the ensemble models in detecting SFDI attacks
The ensemble framework employs two sets of machine learning stacking ensembles.
One set of the ensembles uses only supervised methods while the other one uses only
7.2 M O T I V AT I O N
Different classifiers usually perform differently on the same data. In stacking ensemble-
based machine learning, multiple different classifiers are used together and the results
given by these constituent classifiers are further classified by another (second stage)
3 The ensemble framework is a part of the security analytics framework.
4 The work presented in this chapter has been published in the following two papers:
1. M. Ashrafuzzaman, S. Das, Y. Chakhchoukh, S. Shiva, and F. T. Sheldon. “Detecting stealthy false
data injection attacks in the smart grid using ensemble-based machine learning,” Journal of Computers
& Security, Elsevier, August 2020. doi:10.1016/j.cose.2020.101994
2. M. Ashrafuzzaman, S. Das, Y. Chakhchoukh, S. Duraibi, S. Shiva, and F. T. Sheldon. “Supervised
learning for detecting stealthy false data injection attacks in the smart grid,” Transactions on
Computational Science and Computational Intelligence, Advances in Security, Networks, and Internet
of Things, Springer Nature, 2020. (accepted)
77
classifier [28, 91]. This approach is called a two-layer ensemble or stacking [114].
In supervised learning, labeled data, i.e., a training set of examples with correct
responses, is provided and based on this training the machine learning method gen-
eralizes (i.e., learns the patterns in the training data) to classify the unlabeled input
sets. In unsupervised learning, the method is trained with unlabeled data to identify
similarities between the inputs that have something in common. These similar inputs
are then categorized together. In other words, the method attempts to learn the hidden
patterns in the input data, and later predicts responses to test inputs based on the
learned patterns.
better than unsupervised models. However, the additional and often cumbersome
data engineering needed to label raw data is painstaking and challenging. Moreover,
attack data is very sparse. Consequently, the availability of labeled data is not always
guaranteed in a timely manner. That is why most datasets are synthetic. Because
supervised models are trained with labeled data, they learn the patterns associated
with “attacks” very well and can detect such patterns more consistently. However,
if the attack pattern is not one of the learned patterns or if the attack is a new
one, then the performance of supervised models is highly degraded. In these cases,
must be considered. For these reasons, both supervised and unsupervised models
This section briefly describes the machine learning methods used in the ensemble
framework. More information on these learning methods can be found in the books
Logistic regression (LR) is a classification algorithm that performs very well on linearly
separable classes. LR builds on ideas from the field of statistics where the logistic
model is used to discern the probability of a true/false class or event. This can be
extended to models with more than two classes of events. Each of the events would
be assigned a probability value between 0 and 1, where the sum of all probabilities
is unity. The coefficients of the logistic regression algorithm must be estimated from
the training data which is done by using maximum-likelihood estimation. The best
coefficients would result in a model that would predict a value very close to 1 for the
default class and value very close to 0 for the other class.
Support vector machine (SVM) is a group of supervised learning methods that identify
the patterns for data classification or regression analysis based on finding a separating
hyperplane in the feature space between two classes, in such a way that the distance
between the hyperplane and the closest data points for each class is maximized. The
SVM algorithm is based on probabilistic statistical learning theory [24] whereby the
approach favors a minimized classification risk rather than optimal classification. Var-
79
ious types of dividing classification surfaces can be realized by applying a kernel, such
tangent [23].
Naïve Bayes is a simple classifier used in many machine learning problems. Based on
the Bayes theorem this probabilistic classifier helps define the probability of an event
based on some prior knowledge of certain conditions associated with that event. The
name naïve Bayes originates from the fact that the input features are assumed to be
Decision tree (DT) is a non-parametric supervised learning method and is used both for
classification and regression. DT builds a tree in which each branch shows a probability
between a number of possibilities and each leaf shows a decision. The paths from the
root of the tree to the leaves represent classification rules. The algorithm collects
information and applies the rules for the purpose of the decision to take a particular
path. In DT, each level splits the data according to different attributes and attempts to
The computational architecture of neural networks mimics the neural structure and
function of the brain forming the interconnected groups of artificial neurons. Each of
these artificial neurons is a set of input values and associated weights that trigger the
80
neuron beyond a threshold. The neural network is organized in layers of neurons. The
first layer is the input layer and the last one is the output layer. The layers in between
these two are called hidden layers. The neural networks attempt to hierarchically
learn deep features and correlations in input data by adjusting the weight associated
with the neurons. Neural network architectures have many variants with each finding
set containing only the objects of that class. Trained in this way, the classifier flags
anomaly [53]. OCSVM [99] is a support vector machine based anomaly detector. Like
the supervised SVM models, unsupervised one-class versions also work with different
kernels. In the ensemble framework, one-class SVMs with a linear kernel and a
Each tree is trained on a random subset of the training data. Isolation forest is used to
a split value between the maximum and minimum values of the selected features
equivalent to the path length from the root to the leaf. The path length from the
root to leaf, averaged over a forest of such random trees, is a measure of normality
and the decision function. Random partitioning produces noticeably shorter paths for
anomalies. Therefore, the trees that collectively produce shorter paths for particular
Elliptic envelope (EE) attempts to ‘draw’ an ellipse putting the normal class members
inside the ellipse. Any observation outside the ellipse is then classified as an outlier or
anomaly. EE models the data as a high dimensional Gaussian distribution with possi-
ble covariances between feature dimensions and uses the FAST-Minimum Covariance
Determinant [47, 97] to estimate the size and shape of the ellipse.
The local outlier factor (LOF) [15] first computes the local densities of the data objects
using distances given by k-nearest neighbors (k-NN) algorithm. LOF then identifies
regions of similar density by comparing the local density of a given data object to the
local densities of its neighbors. The data objects that have substantially lower densities
This section provides an overview of the stacking ensemble-based SFDI attack de-
tection scheme. The ensemble framework (a smaller framework within the overall
In this ensemble-based approach, two different ensembles are used: one using five
supervised classifiers and another one five unsupervised classifiers. These individual
classifiers when run together, simultaneously or one after another, form the first phase
The supervised classifiers included in the ensemble framework are: decision tree
(DT), logistic regression (LR), naïve Bayes (NB), support vector machine (SVM), and
The unsupervised classifier models included in the ensemble framework are: one-class
support vector machine with polynomial kernel (OCSVM_P), one-class support vector
machine with linear kernel (OCSVM_L), isolation forest (ISOF), elliptic envelope (EE),
Stacking ensemble is a learning method where different classifiers are combined into
a meta-classifier that has better generalization performance than each individual clas-
sifier alone [28, 91]. Stacking ensemble is a two-stage process. The first stage consists
of different classification methods, for example, DT, SVM, LR, and so on, or one base
classification algorithm can be used repeatedly with different subsets of the training
data. These individual classifiers are run together, simultaneously or one after another.
In the second stage, the decisions given by individual classifiers are fed as input to
another classifier, called the ensemble classifier, for the final decision. Majority voting
is the most popular ensemble method, which selects the class label that has been
predicted by the majority of the constituent classifiers. Instead of majority voting, any
binary classification method can be used as the ensemble classifier. The ensemble
framework uses six classifiers as the ensemble classifier: majority voting (Ens_MV), lo-
84
gistic regression (Ens_LR), naïve Bayes (Ens_NB), neural network (Ens_NN), decision
This section presents a set of model selection experiments that were performed using
the ensemble framework presented above. The goal of the experiments is to find
one or more machine learning models in the ensemble framework that can detect
SFDI attacks accurately and reliably. The experiments were conducted with individual
classification first and then ensemble classification. The experiments ran the data
through five supervised and five unsupervised learning models. Six ensemble models
were run with the outcomes of the supervised models, and six ensemble models with
the outcomes of the unsupervised models. The hyper-parameters used for different
models are given in Table 7.1. The dataset was split into two subsets: 70% for training
and 30% for testing. To avoid over-fitting and to obtain robust models, 10-fold cross
validation over randomly divided training data was used. Then the test data was used
7.6 R E S U LT S
Table 7.4 lists representative individual and ensemble models for different datasets
and their corresponding evaluation metrics values. Detail results of all the individual
and ensemble models using different datasets and the associated evaluation metrics
T A B L E 7 . 2 : Training times for supervised individual and ensemble models for the
14-bus-AM1 dataset.
Elapsed Time
Models 21 Features 122 Features Without SVM
LR 0.56s 1.02s
NB 0.27s 1.03s
NN 0.57s 0.83s
DT 1.59s 1m 55s
SVM 45m 14s 2h 28m
Ens_MV 45m 19s 2h 30m 6.07s
Ens_LR 45m 17s 2h 30m 4.15s
En_NB 45m 17s 2h 30m 4.12s
Ens_NN 45m 17s 2h 30m 4.33s
Ens_DT 45m 17s 2h 30m 4.08s
Ens_SVM 45m 33s 2h 31m 8.55s
T A B L E 7 . 3 : Training times for unsupervised individual and ensemble models for the
14-bus-AM1 dataset.
Elapsed Time
Models 21 Features 122 Features Without LOF
Dataset Learning Classifier Models F1-Score Accuracy Precision Sensitivity Specificity ROC Elapsed
Type Category AUC Time
14-Bus-AM1 Supervised Individual All Models 0.8439 0.8931 0.9991 0.7304 0.9997 0.8650 0.57s
Ensemble All Models 0.8472 0.8961 0.9993 0.7353 0.9997 0.8675 45m 17s
Unsupervised Individual EE 0.6318 0.9214 0.5606 0.7237 0.9418 0.8327 12.39s
Ensemble Ens_NN 0.6218 0.9216 0.5428 0.7278 0.9505 0.8341 20m 11s
14-Bus-AM2 Supervised Individual All Models 0.8472 0.8913 0.9989 0.7354 0.9987 0.8951 0.59s
Ensemble All Models 0.8477 0.8911 0.9991 0.7361 0.9989 0.8686 45m 17s
Unsupervised Individual EE 0.6534 0.9379 0.7031 0.6107 0.9728 0.7923 12.4s
Ensemble Ens_SVM 0.6521 0.9321 0.7083 0.6048 0.9701 0.7893 20m 11s
14-Bus-AM3 Supervised Individual All Models 0.8479 0.8974 0.9994 0.7363 0.9998 0.8954 0.62s
Ensemble All Models 0.8482 0.8964 0.9996 0.7367 0.9999 0.8682 45m 18s
Unsupervised Individual EE 0.6540 0.9392 0.7034 0.6110 0.9733 0.7921 12.37s
Ensemble Ens_NB 0.6529 0.9395 0.7089 0.6051 0.9743 0.7896 20m 11s
57-Bus-AM1 Supervised Individual All Models 0.8369 0.8826 0.9897 0.7288 0.9921 0.8591 0.60s
Ensemble All Models 0.8424 0.8893 0.9903 0.7261 0.9921 0.8591 45m 18s
Unsupervised Individual EE 0.6331 0.9189 0.5564 0.7198 0.9389 0.8327 12.4s
Ensemble Ens_NN 0.6202 0.9188 0.5375 0.7201 0.9486 0.8341 20m 12s
57-Bus-AM2 Supervised Individual All Models 0.8449 0.8862 0.9991 0.7305 0.9994 0.9056 0.62s
Ensemble All Models 0.8449 0.8862 0.9991 0.7305 0.9994 0.8652 45m 18s
Unsupervised Individual EE 0.6420 0.9388 0.7115 0.5841 0.9751 0.7803 12.39s
Ensemble Ens_SVM 0.6457 0.9402 0.7353 0.5757 0.9788 0.7771 20m 11s
57-Bus-AM3 Supervised Individual All Models 0.8453 0.8866 0.9996 0.7322 0.9998 0.9059 0.61s
Ensemble All Models 0.8453 0.8866 0.9996 0.7322 0.9998 0.8655 45m 19s
Unsupervised Individual EE 0.6421 0.9390 0.7119 0.5848 0.9756 0.7802 12.38s
Ensemble Ens_SVM 0.6462 0.9402 0.7359 0.5760 0.9791 0.7772 20m 10s
87
88
7.7 D I S C U S S I O N O F R E S U LT S
1. For a particular dataset, the performance values for all the supervised individual
classifiers and all the supervised ensemble classifiers are effectively the same for
2. Moreover, the performance values for all the supervised individual classifiers and
all the supervised ensemble classifiers are effectively the same for all the metrics
for all six datasets, i.e., the performance numbers do not vary for different
datasets.
3. The supervised ensemble models do not exhibit better performance than the
4. The precision values for the supervised models are very close to 100%, whereas
accuracy values are about 90%. However, in a classification problem where the
goal is to detect the minor class occurrences, the most important metrics are
the sensitivity which, in this case, measures the proportion of actual “attacks”
that are correctly identified as such; and the specificity which measures the
supervised models, the sensitivity values for all the models are very similar,
varying between 73.53%. This indicates that the models would be able to detect
at most 73% of the attacks and the rest 27% will go undetected. The specificity
values for the models are 99.97% meaning that the models are able to identify a
“non-attack” as such almost always, and will seldom raise a false alert.
89
5. For unsupervised models, the performance for different models vary widely for
any dataset.
6. For six different datasets also the performance of unsupervised models vary.
ensemble models are very close to each other for any particular dataset.
8. Out of six datasets, the best-performing ensemble models are a bit better than
cluded either way that ensembles are better than individuals or vice versa.
9. Among the unsupervised individual models, EE performed the best for all six
datasets.
10. Among the unsupervised models, the best sensitivity of 72.78% is given by Ens_NN
11. Among the six datasets, the lowest sensitivity of 58.41% is given by EE. The
Table 7.2 shows the elapsed time for training the different supervised models. It
is notable that not only the ensemble models do not perform any better, but they also
take more time to run than the individual models. This is because the ensembles first
run all the five individual models and then run the ensemble classifier, and the accumu-
lated elapsed time is therefore higher. The same can be said about the unsupervised
As seen in Chapter 2, the SVM is a popular model among the researchers working
on the problem of detecting stealthy FDI attacks on SE. However, the experiment
shows that SVM performs the same as the other models. Moreover, SVM takes much
90
more time to train. Whereas the other individual models take less than 2 seconds
to train, SVM takes more than 2700 seconds or 45 minutes on the feature-reduced
dataset. On the original dataset with 122 features, SVM takes an astounding 8900
seconds or 2.47 hours. This also exemplifies the “curse of dimensionality” and how it
can be handled by reducing the feature set. It was also observed that if SVM is taken
out as a constituent individual model, then the times taken by the ensemble models
were reduced drastically without any reduction in performance. The last columns in
Tables 7.2 and 7.3 show times taken by the ensemble models when SVM or LOF models
The experiment also used the corresponding full-feature dataset (122 features for
14-bus system) to train the models. It was found that the performance numbers for the
models do not change at all for these cases. However, it was found that the elapsed
times for training with the full-feature datasets for 14-bus system take over 300%
more time. For unsupervised models, the elapsed times are 900% higher respectively
7.8 S U M M A RY
In this chapter, use of the security analytics framework with ensemble learning in the
model selection phase was described for detection of SFDI attacks. The performance
of the ensemble models were compared with the performance of individual models. It
was found that the performance of individual supervised models are the same as those
of the ensemble models. However, for the unsupervised models, the ensemble models
performed better for two datasets, and the individual models performed better for the
other four datasets. The best detection rate found for any of the models is 73%. For
91
both the supervised and unsupervised models, reducing the feature set increases the
training speeds by many folds without suffering any degradation in detection rates.
92
CHAPTER 8
A RT I F I C I A L N E U R A L N E T W O R K S
8.1 INTRODUCTION
In this chapter, the model selection part of the security analytics framework using
artificial neural networks (ANN) is presented. The ANN models are built using ANN
methods training on the datasets generated and prepared in Chapter 5. The model
selection process ends with determining the efficacy of the ANN models in detecting
8.2 M O T I V AT I O N
So far in the previous two chapters, eight supervised models, five unsupervised mod-
els, and six stacking ensemble models were used to detect SFDI attacks with decent
detection rate. However, none of the models showed any acceptable performance.
Moreover, all the supervised models returned effectively the same detection rate of
about 73%, with artificial neural network showing a little better results than the
others. Therefore, as a next logical step, a decision to treat artificial neural networks
through full range of model selection was taken because ANNs have been known to
The artificial neural network, also known as feedforward neural networks (FFNN)
ing models. The main idea for deep learning came from Hecht-Nielsen who proved
the universal expressive power of a three-layer neural network back in 1989 [44].
make way for harnessing the power of this model and for practical implementation
architectures [45].
The computational architecture of neural networks mimics the neural structure and
function of the brain forming the interconnected groups of artificial neurons. Each of
these artificial neurons is a set of input values and associated weights that trigger the
neuron beyond a threshold. The neural network is organized in layers of neurons. The
first layer is the input layer and the last one is the output layer. The layers in between
these two are called hidden layers. The basic architecture has an input layer of nodes
equal to the number of input data, one or more hidden layers with varying number of
nodes or neurons, and an output layer with number of nodes equal to the number of
responses. Figure 8.1 shows a basic architecture of an artificial neural network with
the input layer, 2 hidden layers and the output layer. The neural networks attempt
94
to hierarchically learn deep features and correlations in input data by adjusting the
weight associated with the neurons. Neural network architectures have many variants
This section presents the model selection process with artificial neural networks. The
goal of the process is to identify an ANN model that can detect SFDI attacks accurately
and reliably. To find out the best-performing ANN model, the ANN method was
run through a grid-search with hidden layers and hidden units as the varying hyper-
parameters. The hidden layers were chosen from 2, 4, and 6; and with each of
these hidden layers, hidden units were chosen between 50 and 2500. Other hyper-
batch_size=‘auto’,
beta_1=0.9,
activation= ‘relu’,
solver=‘adam’,
beta_2=0.999,
early_stopping=False,
epsilon=1e-08,
alpha=0.0001,
learning_rate_init=0.001,
max_iter=500,
momentum=0.9,
95
power_t=0.5,
random_state=None,
n_iter_no_change=10,
shuffle=True,
tol=0.0001,
validation_fraction=0.1,
warm_start=False.
T A B L E 8 . 1 : Evaluation metrics values for the best-performing ANN models for the
six datasets.
Datasets Hidden Hidden F1-Score Accuracy Precision Sensitivity Specificity ROC Elapsed
Layers Units AUC Time
14-Bus-AM1 4 1200 0.9834 0.9853 0.9744 0.9824 0.9878 0.9956 23h 50m
14-Bus-AM2 4 1200 0.9832 0.9855 0.9747 0.9824 0.9878 0.9956 22h 12m
14-Bus-AM3 4 1200 0.9836 0.9855 0.9746 0.9826 0.9887 0.9956 23h 56m
57-Bus-AM1 4 1200 0.9827 0.9847 0.9739 0.9819 0.9872 0.9951 1d 9h
57-Bus-AM2 4 1200 0.9829 0.9850 0.9740 0.9820 0.9874 0.9954 1d 3h
57-Bus-AM3 4 1200 0.9831 0.9853 0.9742 0.9822 0.9876 0.9956 1d 6h
8.5 R E S U LT S
Table 8.1 lists the best ANN models for different datasets and their corresponding eval-
uation metrics values. Detail results of ANN model selection using different datasets
8.6 D I S C U S S I O N O F R E S U LT S
1. For all the six datasets, the best-performing ANN models are the ones with
4 hidden layers and 1200 hidden units in each of the hidden layers. These
96
particular models were chosen because these models have highest sensitivity,
2. The sensitivity values for the best-performing models are around 98.25%, mean-
ing the models can correctly detect over 98% of the attacks. The specificity values
are around 98.75%, meaning that on less than 1.25% times the models would
raise a false alarm. These results are by far the better results that were seen with
other models in this research. As a matter of fact, these results are among the
3. Models with more than 4 layers and 1200 units were also trained. However, it
was observed that the performance ratings actually go down with higher layers
4. The time taken to train the best models are around a day and a half, i.e., about
[email protected] and 20GB memory running a 64-bit High Sierra macOS. This is not
the machine was not dedicated to this task. The training time will reduce dras-
units (GPU) is used. Moreover in practice, the training is seldom done in real-
time. Instead, the training is used off-line to obtain a robust model and then this
8.7 S U M M A RY
In this chapter, use of the security analytics framework with artificial neural networks
in the model selection phase was described for detection of SFDI attacks. Model
97
selection was performed using artificial neural network models with wide range of
hyper-meter settings. It was found that models with 4 hidden layers and 1200 hidden
units in each hidden layers can detect 98.25% attacks while raising only 1.25% false
alarms. These results are the best it was seen in this research and are among the best
in the literature.
98
CHAPTER 9
ELLIPTIC ENVELOPE
9.1 INTRODUCTION
In this chapter, the model selection part of the security analytics framework using ellip-
tic envelope (EE) method is presented. The EE models are built using EE methods with
different hyper-parameter values and training on the datasets generated and prepared
in Chapter 5. The model selection process ends with determining the efficacy of the
Section 5.8.1.5
9.2 M O T I V AT I O N
from normal data using a tagged or labeled dataset. Unsupervised methods, on the
other hand, sort data into different clusters where the strength of the clustering lies
within the algorithm itself. Among unsupervised methods, novelty and outlier detec-
tion strategies have shown significant promise in detecting otherwise hidden anoma-
lies [58, 26]. The supervised methods require large amounts of data to be curated as
labeled data (i.e., with ground truth). Unsupervised methods do not require labeled
While it may not be very difficult to collect data during normal operation of trans-
mission systems, “attack” data is very rarely available because attacks do not happen
often. In addition to that the process of curating the data to create labeled datasets is
an onerous task. Instead of being trained on data labeled with ground truth, unsuper-
vised models work by finding the hidden similarities in different data components and
attempt to group similar data instances in regions or clusters. For bi-modal data like SE
measurement data with sparse attacks, unsupervised models create one cluster fitting
most of the non-attack data (major data) and treat instances lying far outside the
distribution data [96]. Elliptic envelope tries to make an elliptical cluster and fits the
major class instances in that. Instances far away from the cluster are then considered
systems. The elliptic envelope method models data as a high dimensional Gaussian
an ellipse so that majority of the data instances fit into the ellipse. Data instances
lying far outside the ellipse are then considered anomalies or outliers and, for the
current context, are marked as an attack. The elliptic envelope method uses the FAST-
100
minimum covariance determinant (FAST-MCD) [97] to estimate the shape and size
of the ellipse. The FAST-MCD algorithm selects non-intersecting sub-sets of data and
computes the mean µ, and the covariance matrix C, in each data-feature for each sub-
set. The Mahalanobis distance, d MH , a measure of the distance between a point P and
set and the data are ordered in ascending order by d MH . The Mahalanobis distance
obtained from this estimate is used to define the threshold for determining outliers or
q
d MH = ( x − µ ) T C −1 ( x − µ ) (9.1)
where C is the covariance matrix. If the covariance matrix is the identity matrix,
then d MH reduces to the Euclidean distance and to the normalized Euclidean distance
how many σ (standard deviation) a data point is from the mean of a distribution.
The FAST-MCD algorithm selects sub-sets from the original dataset, with small values
This procedure is iterated until the determinate of the covariance matrix converges.
The covariance matrix with the smallest determinate from all sub-set forms an ellipse
To find an optimized elliptic envelope model, the models were trained with varying val-
ues for the hyper-parameter “contamination rate”. Contamination rate represents the
approximately how much of the data instances should sit outside of the enclosing high-
dimensional ellipse that contains the majority of the data instances. In this model
selection experiment, the values of contamination rate started at 0.001 and gradually
increased to 0.5.
The dataset was split in 7:3 ratio into training subset and test subset retaining the
same distribution of normal and attack data. To obtain robust models without over-
fitting, 10-fold cross-validation over randomly divided sub-sets of training data during
training of the models was used. Then, the test data was used for prediction and for
9.5 R E S U LT S
Table 8.1 lists the best EE models for different datasets and their corresponding evalu-
ation metrics values. Detail results of elliptic envelope model selection using different
datasets and associated evaluation metrics values are given in Appendix A .2.
102
9.6 D I S C U S S I O N O F R E S U LT S
In this section, the results from the evaluations with different elliptic envelope models
are discussed. The observations from Tables A .19- A .24 and Table 9.1 are:
0.03 for 14-Bus-AM1 dataset, and 0.035 for five other datasets.
• The best sensitivity value varies from 70.07% to 72.79% for the six datasets.
• The best specificity value varies from 96.44% to 97.87% for the six datasets.
9.7 S U M M A RY
In this chapter, use of the security analytics framework with elliptic envelope in the
model selection phase was described for detection of SFDI attacks. For the model
selection process a wide range of hyper-parameter values were used for the elliptic
envelope models. It is found that the best-performing elliptic envelope model can
CHAPTER 10
CONCLUSION
10.1 P R I M A R Y R E S U LT S
1. Successful use of security analytics to detect stealthy false data injection attacks
on the static state estimation process used by power transmission control systems,
2. Achieving a 98.24% detection rate and a 1.25% false alarm rate using supervised
artificial neural networks with 4 hidden layers and 1200 hidden units per layer.
These results are among the best results in the related works in the literature.
3. Among the unsupervised models, achieving a 73% detection rate and a 3% false
4. Detection rates are the same for datasets with different attack models and differ-
10.2 D I S S E R TAT I O N S U M M A R Y
Research has shown that intruders in the power transmission systems can surrep-
titiously change the RTU sensor measurements in such a way that can upend the
situational awareness of the operators of power control centers and force them to act
in ways that may cause electricity theft, changes in the electricity prices, or outages.
These subtle changes in the state estimation measure data, called stealthy false data
104
injection attacks, cannot be detected by the conventional measures in the power trans-
This research aimed at developing a scheme to accurately and reliably detect SFDI
attacks on the static state estimation in AC power transmission systems. The scheme
employs machine learning models to detect SFDI attacks on the static state estimation
developed and three attack models were identified, based on which synthetic data
simulating standard IEEE 14-bus and 57-bus systems using MATPOWER were gener-
ated. A software for managing machine learning model selection and evaluation was
developed using the Python programming language and scikit-learn machine learning
ensemble (Section 7.4.2) methods. These models were parameter-tuned, trained and
The experiments to detect SFDIA were started with three supervised models, GLM,
GBM, and DRF, and it was found that the results, with a detection rate of 72%, were
models were used with the hope that it will improve the detection rates. However,
even the best results from the ensembles were also not very good, with the detection
rate being 73%. The next attempt was to use artificial neural networks with wide
ranges of hidden layers and hidden units. It was found that when trained with the
right hyper-parameter settings (i.e., 4 hidden layers and 1200 hidden units per layer),
ANN can detect 98.24% of the attacks with only 1.25% false alarm rate. This result is
105
among the best results found in the literature for SFDI attack detection for simulated
Since, unsupervised models do not need labeled data to train, it was also impera-
tive to find out which unsupervised model performs the best with these datasets. It
was found that the elliptic envelope model can detect 73% of the attacks with a false
alarm rate of 3%. It was also found that the detection rates for all the methods used in
this dissertation are the same for the datasets corresponding to different bus sizes and
different attack models. This means that irrespective of whether the attacker compro-
mised one sensor, multiple sensors, or all sensors, the machine learning methods will
Therefore, this dissertation shows that machine learning based security analytics
can accurately and reliably detect SFDI attacks on the static state estimation in power
transmission systems.
In addition to the security analytics framework, the software, and the results ob-
tained with that, a survey of different types of false data injection attacks on different
parts of the entire power grid and a taxonomy of such attacks were presented in
the dissertation. Also presented was a survey of literature of the machine learning
based approaches to detect SFDI attacks on the static state estimation of the power
transmission systems. The ideas listed as future work below show that the research will
open up a number of immediate research questions in this urgent and active area. The
software and the datasets will be made available via GitHub so that the experiments
The future work items of this research can be categorized as in the sub-sections below.
• The security analytics software implemented does not have a graphical user
the software.
• At this time the software supports GLM, GBM, DRF, DT, LR, NB, ANN, SVM,
OCSVM with different kernels, ISOF, LOF, and EE methods. More methods can
urable. An extensible and configurable framework will ensure that such a tech-
nique could be standardized for systematic use in transmission systems and other
problem areas.
• The methods supported are all for time-discrete data analytics. Addition of time-
series analytics methods like recurrent neural networks will enable the software
• Support for adversarial machine learning can be another useful future work.
• At this time, the software uses only RFC for feature selection. More feature
• More visualization support, e.g., interactive graphs, 3-D graphs, etc., can be
added.
• It was seen in chapters 6 and 7 that the performance of all supervised methods
for all the six datasets are essentially the same. This is not expected as different
mance with the same dataset. An important future work will be to investigate
why all the supervised methods resulted in the same performance metrics values.
approach with the internal operations of the methods hidden to the human. IML
stand the reasons behind the ML models’ decisions (i.e., predictions). Alternately,
supervised models giving same results, datasets with new parameters can be
generated and/or the datasets can be scrambled before employing the machine
learning methods. This may uncover additional insight into the properties of
datasets pertaining to stealthy false data injection attacks. Combining IML ap-
proach with this may result in formulating a science of SFDIA modeling and
testing to ensure 360 degree fidelity and trustworthiness of the machine learning
model predictions.
• This dissertation used one boosting (GBM) and one bagging (DRF) methods.
None of these methods yielded good classification results. Boosting and Bagging
puting systems with graphical processing units (GPU) will reduce the training
• This research used 14-bus and 57-bus system to validate the security analytics
framework. Further testing with systems having large number of buses will shed
• This research shows 98% detection rate can be achieved for the synthetic datasets.
before-seen attacks since unsupervised models do not need labeled data. This
to take into account the correlation in power transmission data with seasons
and time of the day. Correlated time-series data needs to be analyzed using ML
• This research dealt with measurement data associated with static state estima-
tion. More and more power systems are now using PMUs and hence dynamic
be used for analytics in many other areas, including security analytics of other
power grid security issues, ICS and CPS security issues, computational biology,
• SFDI attacks are still limited to academic research. Except for the 2015 Ukraine
attack, there is no other known instance of SFDI attack. Consequently, data cor-
responding to actual attacks are not available. However, it may be possible to use
power system testbeds and real-time digital simulation (RTDS) to generate more
• Deployment of the trained models to find out how efficiently the models detect
SFDI attacks, possibly using a power system testbed, is another future work.
110
this dissertation (in Appendix B ). This conceptual model can be fleshed out,
Stealthy false data injection attacks currently exist only in academic research labs.
causing power theft or even power system failure which could be deadly and extremely
costly. Studies such as this one, help the community to understand more completely
how detection, prevention, and countermeasure mechanisms of this type can play a
role. Ideally, defense-in-depth or layered defenses should prevent the adversaries from
spoofing the system to enable such SFDI attacks. However, there is no guarantee that
adversaries will not be successful in penetrating layered defenses. Moreover, the SFDI
attack is also a threat from the inside. Users, with proper credentials constitute a
significant portion of the threat in this SFDI attack surface. Therefore, active research
is necessary to devise even more effective mechanisms to quickly detect and deter
such attacks. The research presented in this dissertation is a step towards achieving a
BIBLIOGRAPHY
[12] Abdelrahman Ayad, Hany EZ Farag, Amr Youssef, and Ehab F El-Saadany.
Detection of false data injection attacks in smart grids using recurrent neural
networks. In 2018 IEEE Power & Energy Society Innovative Smart Grid
Technologies Conference (ISGT), pages 1–5. IEEE, 2018. doi:10.1109/ISGT.
2018.8403355.
[13] Gustavo E A P A Batista, Ronaldo C Prati, and Maria Carolina Monard. A study
of the behavior of several methods for balancing machine learning training
data. ACM SIGKDD Explorations Newsletter, 6(1):20–29, 2004. doi:10.1145/
1007730.1007735.
[14] Omar Ali Beg, Taylor T Johnson, and Ali Davoudi. Detection of false-
data injection attacks in cyber-physical DC microgrids. IEEE Transactions on
industrial Informatics, 13(5):2693–2703, 2017.
[15] Markus M Breunig, Hans-Peter Kriegel, Raymond T Ng, and Jörg Sander.
LOF: Identifying density-based local outliers. In Proceedings of the 2000 ACM
SIGMOD International Conference on Management of Data, pages 93–104, 2000.
doi:10.1145/342009.335388.
[16] Eric Byres. The air gap: SCADA’s enduring security myth. Communications of
the ACM, 56(8):29–31, 2013. doi:10.1145/2492007.2492018.
[17] Mario R Camana-Acosta, Saeed Ahmed, Carla E Garcia, and Insoo Koo.
Extremely randomized trees-based scheme for stealthy cyber-attack detection
in smart grid networks. IEEE Access, 8:19921–19933, 2020. doi:10.1109/
ACCESS.2020.2968934.
[18] Yacine Chakhchoukh and Hideaki Ishii. Coordinated cyber-attacks on the
measurement function in hybrid state estimation. IEEE Transactions on
Power Systems, 30(5):2487–2497, September 2015. doi:10.1109/TPWRS.2014.
2357182.
[19] Yacine Chakhchoukh, Song Liu, Masashi Sugiyama, and Hideaki Ishii. Statistical
outlier detection for diagnosis of cyber attacks in power state estimation. In
2016 IEEE Power and Energy Society General Meeting (PESGM), pages 1–5. IEEE,
2016.
[20] Gu Chaojun, Panida Jirutitijaroen, and Mehul Motani. Detecting false data
injection attacks in AC state estimation. IEEE Transactions on Smart Grid,
6(5):2476–2483, 2015. doi:10.1109/TSG.2015.2388545.
[21] Jiongcong Chen, Gaoqi Liang, Zexiang Cai, Chunchao Hu, Yan Xu, Fengji Luo,
and Junhua Zhao. Impact analysis of false data injection attacks on power
system static security assessment. Journal of Modern Power Systems and Clean
Energy, 4(3):496–505, 2016.
[22] Dae-Hyun Choi and Le Xie. Ramp-induced data attacks on look-ahead dispatch
in real-time power markets. IEEE Transactions on Smart Grid, 4(3):1235–1243,
2013.
[23] Nello Christianini and John Shawe-Taylor. Support vector machines and other
kernel-based learning methods. Cambridge UP, 2000.
113
[36] David Grochocki, Jun Ho Huh, Robin Berthier, Rakesh Bobba, William H
Sanders, Alvaro A Cárdenas, and Jorjeta G Jetcheva. AMI threats, intrusion
detection requirements and deployment recommendations. In 2012 IEEE Third
International Conference on Smart Grid Communications (SmartGridComm),
pages 395–400. IEEE, 2012.
[37] Zhitao Guan, Nan Sun, Yue Xu, and Tingting Yang. A comprehensive survey of
false data injection in smart grid. International Journal of Wireless and Mobile
Computing, 8(1):27–33, 2015.
[38] Yonghe Guo, Chee-Wooi Ten, and Panida Jirutitijaroen. Online data validation
for distribution operations against cybertampering. IEEE Transactions on Power
Systems, 29(2):550–560, 2013.
[39] Mohamed Hamlich, Abdelkarim El Khantach, and Tarik Ibn Ziad. Machine
learning methods against false data injection in smart grid. International
Journal of Reasoning-based Intelligent Systems, 2(1), 2020. doi:10.1504/IJRIS.
2020.104991.
[40] Jianye Hao, Eunsuk Kang, Jun Sun, Zan Wang, Zhaopeng Meng, Xiaohong
Li, and Zhong Ming. An adaptive markov strategy for defending smart grid
false data injection from malicious attackers. IEEE Transactions on Smart Grid,
9(4):2398–2408, 2016.
[41] Jinping Hao, Robert J Piechocki, Dritan Kaleshi, Woon Hau Chin, and Zhong
Fan. Sparse malicious false data injection attacks and defense mechanisms in
smart grids. IEEE Transactions on Industrial Informatics, 11(5):1–12, 2015. doi:
10.1109/TII.2015.2475695.
[42] Trevor Hastie, Robert Tibshirani, and Jerome Friedman. The Elements of
Statistical Learning: Data Mining, Inference, and Prediction. New York: Springer-
Verlag, 2nd edition, 2008.
[43] Youbiao He, Gihan J Mendis, and Jin Wei. Real-time detection of false data
injection attacks in smart grid: A deep learning-based intelligent mechanism.
IEEE Transactions on Smart Grid, 2017. doi:10.1109/TSG.2017.2703842.
[44] R. Hecht-Nielsen. Theory of the backpropagation neural network. In
International Joint IEEE Conference on Neural Networks, pages 593–605, 1989.
doi:10.1109/IJCNN.1989.118638.
[45] Geoffrey E Hinton, Simon Osindero, and Yee-Whye Teh. A fast learning
algorithm for deep belief nets. Neural computation, 18(7):1527–1554, 2006.
[46] Tin Kam Ho. Random decision forests. In Proceedings of 3rd international
conference on document analysis and recognition, volume 1, pages 278–282.
IEEE, 1995.
[47] Mia Hubert and Michiel Debruyne. Minimum covariance determinant. Wiley
Interdisciplinary Reviews: Computational Statistics, 2(1):36–43, 2010. doi:10.
1002/wics.61.
115
[59] Yan Li, Peng Zhang, Liang Zhang, and Bing Wang. Active synchronous detection
of deception attacks in microgrid control systems. IEEE Transactions on Smart
Grid, 8(1):373–375, 2017.
[60] Zhiyi Li, Mohammad Shahidehpour, Ahmed Alabdulwahab, and Abdullah
Abusorrah. Bilevel model for analyzing coordinated cyber-physical attacks on
power systems. IEEE Transactions on Smart Grid, 7(5):2260–2272, 2015.
[61] Gaoqi Liang, Steven R Weller, Junhua Zhao, Fengji Luo, and Zhao Yang Dong.
The 2015 Ukraine blackout: Implications for false data injection attacks. IEEE
Transactions on Power Systems, 32(4):3317–3318, 2017. doi:10.1109/TPWRS.
2016.2631891.
[62] Gaoqi Liang, Junhua Zhao, Fengji Luo, Steven R Weller, and Zhao Yang Dong.
A review of false data injection attacks against modern power systems. IEEE
Transactions on Smart Grid, 8(4):1630–1638, 2016. doi:10.1109/TSG.2015.
2495133.
[63] IH Lim, S Hong, MS Choi, SJ Lee, TW Kim, SW Lee, and BN Ha. Security
protocols against cyber attacks in the distribution automation system. IEEE
Transactions on Power Delivery, 25(1):448–455, 2010.
[64] Jie Lin, Wei Yu, Xinyu Yang, Guobin Xu, and Wei Zhao. On false data injection
attacks against distributed energy routing in smart grid. In Proceedings of the
2012 IEEE/ACM Third International Conference on Cyber-Physical Systems, pages
183–192. IEEE Computer Society, 2012.
[65] Fei Tony Liu, Kai Ming Ting, and Zhi-Hua Zhou. Isolation-based anomaly
detection. ACM Transactions on Knowledge Discovery from Data (TKDD), 6(1):1–
39, 2012. doi:10.1145/2133360.2133363.
[66] Mengxiang Liu, Chengcheng Zhao, Ruilong Deng, Peng Cheng, Wenhai Wang,
and Jiming Chen. Nonzero-dynamics stealthy attack and its impacts analysis in
DC microgrids. In 2019 American Control Conference (ACC), pages 3922–3927.
IEEE, 2019.
[67] Xuan Liu, Zhen Bao, Dan Lu, and Zuyi Li. Modeling of local false data injection
attacks with reduced network information. IEEE Transactions on Smart Grid,
6(4):1686–1696, 2015. doi:10.1109/TSG.2015.2394358.
[68] Xuan Liu, Zhiyi Li, Xingdong Liu, and Zuyi Li. Masking transmission line
outages via false data injection attacks. IEEE Transactions on Information
Forensics and Security, 11(7):1592–1602, 2016.
[69] Xuan Liu and Zuyi Li. Local load redistribution attacks in power systems with
incomplete network information. IEEE Transactions on Smart Grid, 5(4):1665–
1676, 2014. doi:10.1109/TSG.2013.2291661.
[70] Xuan Liu and Zuyi Li. False data attacks against ac state estimation with
incomplete network information. IEEE Transactions on smart grid, 8(5):2239–
2248, 2016.
117
[71] Xuan Liu and Zuyi Li. Local topology attacks in smart grids. IEEE Transactions
on Smart Grid, 8(6):2617–2626, 2016.
[72] Xuan Liu and Zuyi Li. False data attack models, impact analyses and defense
strategies in the electricity grid. The Electricity Journal, 30:35–42, 2017. doi:
10.1016/j.tej.2017.04.001.
[73] Y. Liu, M. K. Reiter, and P. Ning. False data injection attacks against state
estimation in electric power grids. In Proc. 16th ACM Conf. on Computer and
Communications Security, pages 21–32, 2009.
[74] Yao Liu, Peng Ning, and Michael K Reiter. False data injection attacks against
state estimation in electric power grids. ACM Transactions on Information and
System Security, 14(1):13:1–13:33, 2011. doi:10.1145/1952982.1952995.
[75] Prasanta Chandra Mahalanobis. On the generalized distance in statistics. In
Proceedings of the National Institute of Sciences of India, volume 2:1, pages 49–
55. National Institute of Science of India, 1936.
[76] Stephen Marsland. Machine Learning: An Algorithmic Perspective. CRC press,
2015.
[77] Stephen McLaughlin, Charalambos Konstantinou, Xueyang Wang, Lucas
Davi, Ahmad-Reza Sadeghi, Michail Maniatakos, and Ramesh Karri. The
cybersecurity landscape in industrial control systems. Proceedings of the IEEE,
104(5):1039–1057, 2016. doi:10.1109/JPROC.2015.2512235.
[78] Stephen McLaughlin, Dmitry Podkuiko, and Patrick McDaniel. Energy theft in
the advanced metering infrastructure. In International Workshop on Critical
Information Infrastructures Security, pages 176–187. Springer, 2009.
[79] Mostafa Mohammadpourfard, Ashkan Sami, and Ali Reza Seifi. A statistical
unsupervised method against false data injection attacks: A visualization-based
approach. Expert Systems with Applications, 84:242–261, 2017.
[80] Mostafa Mohammadpourfard, Yang Weng, Mykola Pechenizkiy, Mohsen
Tajdinian, and Behnam Mohammadi-Ivatloo. Ensuring cybersecurity of smart
grid against data integrity attacks under concept drift. International Journal of
Electrical Power & Energy Systems, 119:105947, 2020. doi:10.1016/j.ijepes.
2020.105947.
[81] Amir-Hamed Mohsenian-Rad and Alberto Leon-Garcia. Distributed internet-
based load altering attacks against smart power grids. IEEE Transactions on
Smart Grid, 2(4):667–674, 2011. doi:10.1109/TSG.2011.2160297.
[82] Christoph Molnar. Interpretable Machine Learning. Lean Publishing, 2020.
[83] Nour Moustafa, Benjamin Turnbull, and Kim-Kwang Raymond Choo. An
ensemble intrusion detection technique based on proposed statistical flow
features for protecting network traffic of internet of things. IEEE Internet of
Things Journal, 6(3):4815–4830, 2018. doi:10.1109/JIOT.2018.2871719.
118
[84] Christian Moya, Chih-Che Sun, Jiankang Wang, and Chen-Ching Liu. Defensing
against measurement attacks on sub-transmission level. In Power and Energy
Society General Meeting (PESGM), 2016, pages 1–5. IEEE, 2016. doi:10.1002/
9781118755471.sgd055.
[85] Ahmed S Musleh, Guo Chen, and Zhao Yang Dong. A survey on the detection
algorithms for false data injection attacks in smart grids. IEEE Transactions on
Smart Grid, 11(3):2218–2234, 2019.
[86] J. A. Nelder and R. W. M. Wedderburn. Generalized linear models. Journal of
the Royal Statistical Society. Series A (General), 135(3):370–384, 1972.
[87] Xiangyu Niu, Jiangnan Li, Jinyuan Sun, and Kevin Tomsovic. Dynamic detection
of false data injection attack in smart grid using deep learning. In 2019 IEEE
Power & Energy Society Innovative Smart Grid Technologies Conference (ISGT),
pages 1–6. IEEE, 2019. doi:10.1109/ISGT.2019.8791598.
[88] Mete Ozay, Inaki Esnaola, Fatos Tunay Yarman Vural, Sanjeev R Kulkarni, and
H Vincent Poor. Machine learning methods for attack detection in the smart grid.
IEEE Transactions on Neural Networks and Learning Systems, 27(8):1773–1786,
2015. doi:10.1109/TNNLS.2015.2404803.
[89] Sarita Paudel, Paul Smith, and Tanja Zseby. Attack models for advanced
persistent threats in smart grid wide area monitoring. In Proceedings of the
2nd Workshop on Cyber-Physical Security and Resilience in Smart Grids, pages
61–66. ACM, 2017. doi:10.1145/3055386.3055390.
[90] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel,
M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos,
D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. Scikit-learn: Machine
learning in Python. Journal of Machine Learning Research, 12:2825–2830, 2011.
[91] Robi Polikar. Ensemble learning. In Ensemble Machine Learning, pages 1–34.
Springer, 2012. doi:10.1007/978-1-4419-9326-7_1.
[92] Ponemon Institute. Cybersecurity in operational technology: 7 insights you
need to know, April 2019. accessed on August 16, 2020. URL: https:
//lookbook.tenable.com/ponemonotreport/ponemon-OT-report.
[93] Md Ashfaqur Rahman and Hamed Mohsenian-Rad. False data injection attacks
with incomplete information against smart power grids. In 2012 IEEE Global
Communications Conference, pages 3153–3158. IEEE, 2012. doi:10.1109/
GLOCOM.2012.6503599.
[94] Md Ashfaqur Rahman and Hamed Mohsenian-Rad. False data injection attacks
against nonlinear state estimation in smart power grids. In 2013 IEEE Power &
Energy Society General Meeting, pages 1–5. IEEE, 2013.
[95] Mohammad Ashiqur Rahman, Ehab Al-Shaer, and Rajesh Kavasseri. Impact
analysis of topology poisoning attacks on economic operation of the smart
power grid. In 2014 IEEE 34th International Conference on Distributed
Computing Systems, pages 649–659. IEEE, 2014.
119
[96] Peter J Rousseeuw. Least median of squares regression. Journal of the American
statistical association, 79(388):871–880, 1984.
[97] Peter J Rousseeuw and Katrien Van Driessen. A fast algorithm for the minimum
covariance determinant estimator. Technometrics, 41(3):212–223, 1999.
[98] Jürgen Schmidhuber. Deep learning in neural networks: An overview. Neural
networks, 61:85–117, 2015. doi:10.1016/j.neunet.2014.09.003.
[99] Bernhard Schölkopf, Robert C Williamson, Alex J Smola, John Shawe-Taylor,
and John C Platt. Support vector method for novelty detection. In Advances in
Neural Information Processing Systems, pages 582–588, 2000.
[100] Marina Sokolova and Guy Lapalme. A systematic analysis of performance
measures for classification tasks. Information Processing & Management,
45(4):427–437, 2009. doi:10.1016/j.ipm.2009.03.002.
[101] Masashi Sugiyama, Taiji Suzuki, and Takafumi Kanamori. Density Ratio
Estimation in Machine Learning. Cambridge University Press, 2012.
[102] Rui Tan, Hoang Hai Nguyen, Eddy Foo, Xinshu Dong, David KY Yau, Zbigniew
Kalbarczyk, Ravishankar K Iyer, and Hoay Beng Gooi. Optimal false data
injection attack against automatic generation control in power grids. In
Proceedings of the 7th International Conference on Cyber-Physical Systems, page 2.
IEEE Press, 2016. doi:10.1109/ICCPS.2016.7479109.
[103] Rui Tan, Hoang Hai Nguyen, Eddy YS Foo, David KY Yau, Zbigniew Kalbarczyk,
Ravishankar K Iyer, and Hoay Beng Gooi. Modeling and mitigating impact of
false data injection attacks on automatic generation control. IEEE Transactions
on Information Forensics and Security, 12(7):1609–1624, 2017. doi:10.1109/
TIFS.2017.2676721.
[104] Song Tan, Debraj De, Wen-Zhan Song, Junjie Yang, and Sajal K Das. Survey of
security advances in smart grid: A data driven approach. IEEE Communications
Surveys & Tutorials, 19(1):397–422, 2017. doi:10.1109/COMST.2016.2616442.
[105] Mini S Thomas and John Douglas McDonald. Power System SCADA and Smart
Grids. CRC press, 2015.
[106] University of Washington. Power System Test Case Archive. [Online]. Available:
http://www.ee.washington.edu/research/pstca/, 2018.
[107] Michel Verleysen and Damien François. The curse of dimensionality in data
mining and time series prediction. In International Work-Conference on Artificial
Neural Networks, pages 758–770. Springer, 2005. doi:10.1007/11494669_93.
[108] Defu Wang, Xiaojuan Wang, Yong Zhang, and Lei Jin. Detection of power
grid disturbances and cyber-attacks based on machine learning. Journal of
Information Security and Applications, 46:42–52, 2019. doi:10.1016/j.jisa.
2019.02.008.
120
[109] Huaizhi Wang, Jiaqi Ruan, Guibin Wang, Bin Zhou, Yitao Liu, Xueqian Fu, and
Jian-Chun Peng. Deep learning based interval state estimation of AC smart grids
against sparse cyber attacks. IEEE Transactions on Industrial Informatics, 2018.
doi:10.1109/TII.2018.2804669.
[110] Jingxuan Wang, Wenting Tu, Lucas CK Hui, Siu-Ming Yiu, and Eric Ke Wang.
Detecting time synchronization attacks in cyber-physical systems with machine
learning techniques. In 2017 IEEE 37th International Conference on Distributed
Computing Systems (ICDCS), pages 2246–2251. IEEE, 2017. doi:10.1109/
ICDCS.2017.25.
[111] Shaocheng Wang and Wei Ren. Stealthy false data injection attacks against state
estimation in power systems: Switching network topologies. In Proceedings of
the 2014 American Control Conference. IEEE, 2014. doi:10.1109/ACC.2014.
6858904.
[112] Yi Wang, Mahmoud Amin, Jian Fu, and Heba Moussa. A novel data analytical
approach for false data injection cyber-physical attack mitigation in smart grids.
IEEE Access, 2017. doi:10.1109/ACCESS.2017.2769099.
[113] John Warsinske. The Official (ISC)2 CISSP CBK Reference. Sybex, 2019.
[114] David H Wolpert. Stacked generalization. Neural networks, 5(2):241–259,
1992.
[115] Yingmeng Xiang, Zhilu Ding, Yichi Zhang, and Lingfeng Wang. Power system
reliability evaluation considering load redistribution attacks. IEEE Transactions
on Smart Grid, 8(2):889–901, 2016.
[116] Yingmeng Xiang, Lingfeng Wang, and Nian Liu. Coordinated attacks on electric
power systems in a cyber-physical environment. Electric Power Systems Research,
149:156–168, 2017. doi:10.1016/j.epsr.2017.04.023.
[117] Le Xie, Yilin Mo, and Bruno Sinopoli. False data injection attacks in
electricity markets. In 2010 First IEEE International Conference on Smart Grid
Communications, pages 226–231. IEEE, 2010.
[118] Le Xie, Yilin Mo, and Bruno Sinopoli. Integrity data attacks in power market
operations. IEEE Transactions on Smart Grid, 2(4):659–666, 2011. doi:10.
1109/TSG.2011.2161892.
[119] Ruzhi Xu, Dawei Chen, and Rui Wang. Data integrity attack detection for
node voltage in cyber-physical power system. Arabian Journal for Science and
Engineering, 2020. doi:10.1007/s13369-020-04813-y.
[120] J. Yan, B. Tang, and H. He. Detection of false data attacks in smart grid with
supervised learning. In 2016 International Joint Conference on Neural Networks
(IJCNN), pages 1395–1402, July 2016. doi:10.1109/IJCNN.2016.7727361.
[121] Can Yang, Yong Wang, Yuhao Zhou, Jiongming Ruan, Wei Liu, et al. False data
injection attacks detection in power system using machine learning method.
Journal of Computer and Communications, 6(11):276, 2018. doi:10.4236/jcc.
2018.611025.
121
[122] Liqun Yang, Yuancheng Li, and Zhoujun Li. Improved-ELM method for detecting
false data attack in smart grid. International Journal of Electrical Power & Energy
Systems, 91(Supplement C):183 – 191, 2017. doi:10.1016/j.ijepes.2017.03.
011.
[123] Qingyu Yang, Jie Yang, Wei Yu, Dou An, Nan Zhang, and Wei Zhao. On
false data-injection attacks against power system state estimation: Modeling
and countermeasures. IEEE Transactions on Parallel and Distributed Systems,
25(3):717–729, 2014. doi:10.1109/TPDS.2013.92.
[124] Zong-Han Yu and Wen-Long Chin. Blind false data injection attack using
PCA approximation method in smart grid. IEEE Transactions on Smart Grid,
6(3):1219–1226, 2015. doi:10.1109/TSG.2014.2382714.
[125] Yanling Yuan, Zuyi Li, and Kui Ren. Modeling load redistribution attacks in
power systems. IEEE Transactions on Smart Grid, 2(2):382–390, 2011.
[126] Yanling Yuan, Zuyi Li, and Kui Ren. Quantitative analysis of load redistribution
attacks in power systems. IEEE Transactions on Parallel and Distributed Systems,
23(9):1731–1738, 2012.
[127] Heng Zhang, Wenchao Meng, Junjian Qi, Xiaoyu Wang, and Wei Xing Zheng.
Distributed load sharing under false data injection attack in an inverter-based
microgrid. IEEE Transactions on Industrial Electronics, 66(2):1543–1551, 2018.
[128] Xialei Zhang, Xinyu Yang, Jie Lin, and Wei Yu. On false data injection attacks
against the dynamic microgrid partition in the smart grid. In 2015 IEEE
International Conference on Communications (ICC), pages 7222–7227. IEEE,
2015.
[129] Xiaocai Zhang, Zhixun Zhao, Yi Zheng, and Jinyan Li. Prediction of taxi
destinations using a novel data embedding method and ensemble learning.
IEEE Transactions on Intelligent Transportation Systems, 2019. doi:10.1109/
TITS.2018.2888587.
[130] Yichi Zhang, Lingfeng Wang, Yingmeng Xiang, and Chee-Wooi Ten. Power
system reliability evaluation with SCADA cybersecurity considerations. IEEE
Transactions on Smart Grid, 6(4):1707–1721, July 2015. doi:10.1109/TSG.
2015.2396994.
[131] Zhenghao Zhang, Shuping Gong, Aleksandar D Dimitrovski, and Husheng
Li. Time synchronization attack in smart grid: Impact and analysis. IEEE
Transactions on Smart Grid, 4(1):87–98, 2013.
[132] Ray Daniel Zimmerman, Carlos Edmundo Murillo-Sánchez, and Robert John
Thomas. MATPOWER: Steady-state operations, planning, and analysis tools for
power systems research and education. IEEE Transactions on Power Systems,
26(1):12–19, 2011. doi:10.1109/TPWRS.2010.2051168.
122
APPENDIX A
A.1 R E S U L T S O F S TA C K I N G E N S E M B L E E X P E R I M E N T S
In this section, the performance metrics values for all six datasets from the experiments
with the ensemble framework in Chapter 7 are presented in tables, graphs and ROC
curves.
The evaluation metric values for supervised and unsupervised individual and en-
semble models for 14-bus-AM1 dataset are tabulated in Tables A .1 and A .2. The
corresponding bar-graphs are shown in Figures A .1 and A .2. The ROC curves are
F I G U R E A . 3 : ROC curves for the supervised individual and ensemble models for
14-bus-AM1 dataset.
F I G U R E A . 4 : ROC curves for the unsupervised individual and ensemble models for
14-bus-AM1 dataset.
126
The evaluation metric values for supervised and unsupervised individual and en-
semble models for 14-bus-AM2 dataset in Tables A .3 and A .4. The corresponding bar-
graphs are shown in Figures A .5 and A .6. The ROC curves are shown in Figures A .7
and A .8.
F I G U R E A . 7 : ROC curves for the supervised individual and ensemble models for
14-bus-AM2 dataset.
F I G U R E A . 8 : ROC curves for the unsupervised individual and ensemble models for
14-bus-AM2 dataset.
129
The evaluation metric values for supervised and unsupervised individual and en-
semble models for 14-bus-AM3 dataset in Tables A .5 and A .6. The corresponding bar-
graphs are shown in Figures A .9 and A .10. The ROC curves are shown in Figures A .11
and A .12.
F I G U R E A . 1 1 : ROC curves for the supervised individual and ensemble models for
14-bus-AM3 dataset.
F I G U R E A . 1 2 : ROC curves for the unsupervised individual and ensemble models for
14-bus-AM3 dataset.
The evaluation metric values for supervised and unsupervised individual and en-
semble models for 57-bus-AM1 dataset in Tables A .7 and A .8. The corresponding bar-
132
graphs are shown in Figures A .13 and A .14. The ROC curves are shown in Figures A .15
and A .16.
F I G U R E A . 1 5 : ROC curves for the supervised individual and ensemble models for
57-bus-AM1 dataset.
F I G U R E A . 1 6 : ROC curves for the unsupervised individual and ensemble models for
57-bus-AM1 dataset.
The evaluation metric values for supervised and unsupervised individual and en-
semble models for 57-bus-AM2 dataset in Tables A .9 and A .10. The corresponding bar-
135
graphs are shown in Figures A .17 and A .18. The ROC curves are shown in Figures A .19
and A .20.
F I G U R E A . 1 9 : ROC curves for the supervised individual and ensemble models for
57-bus-AM2 dataset.
F I G U R E A . 2 0 : ROC curves for the unsupervised individual and ensemble models for
57-bus-AM2 dataset.
The evaluation metric values for supervised and unsupervised individual and en-
semble models for 14-bus-AM3 dataset in Tables A .11 and A .12. The corresponding
138
bar-graphs are shown in Figures A .21 and A .22. The ROC curves are shown in Fig-
F I G U R E A . 2 3 : ROC curves for the supervised individual and ensemble models for
57-bus-AM3 dataset
F I G U R E A . 2 4 : ROC curves for the unsupervised individual and ensemble models for
57-bus-AM3 dataset
141
A.2 R E S U LT S O F A R T I F I C I A L N E U R A L N E T W O R K
EXPERIMENTS
In this section, the performance metrics values for all six datasets from the experiments
Table A .13 and corresponding graph in Figure A .25 show the performance metrics
values for artificial neural network models with different hidden layers and different
ROC
Layers Units F1-Score Accuracy Precision Sensitivity Specificity AUC Time
143
AM1 dataset.
144
Table A .14 and corresponding graph in Figure A .26 show the performance metrics
values for artificial neural network models with different hidden layers and different
ROC
Layers Units F1-Score Accuracy Precision Sensitivity Specificity AUC Time
145
AM2 dataset.
F I G U R E A . 2 7 : Graph of metrics values for neural networks models with varying hidden layers and hidden units on 14-Bus-
146
AM3 dataset.
147
Table A .15 and corresponding graph in Figure A .27 show the performance metrics
values for artificial neural network models with different hidden layers and different
ROC
Layers Units F1-Score Accuracy Precision Sensitivity Specificity AUC Time
Table A .16 and corresponding graph in Figure A .28 show the performance metrics
values for artificial neural network models with different hidden layers and different
ROC
Layers Units F1-Score Accuracy Precision Sensitivity Specificity AUC Time
149
AM1 dataset.
F I G U R E A . 2 9 : Graph of metrics values for neural networks models with varying hidden layers and hidden units on 57-Bus-
150
AM2 dataset.
151
Table A .17 and corresponding graph in Figure A .29 show the performance metrics
values for artificial neural network models with different hidden layers and different
ROC
Layers Units F1-Score Accuracy Precision Sensitivity Specificity AUC Time
152
AM3 dataset.
153
Table A .18 and corresponding graph in Figure A .30 show the performance metrics
values for artificial neural network models with different hidden layers and different
ROC
Layers Units F1-Score Accuracy Precision Sensitivity Specificity AUC Time
A.3 R E S U LT S O F E L L I P T I C E N V E L O P E E X P E R I M E N T S
In this section, the performance metrics values for all six datasets from the experiments
In this section, the results from the model selection process in terms of the evalua-
Table A .19 and corresponding graph in Figure A .31 show the effect of contamina-
tion rates for elliptic envelope models for the 14-Bus-AM1 dataset.
Contamination Elapsed
Rate F1-Score Accuracy Precision Sensitivity Specificity Time
Table A .20 and corresponding graph in Figure A .32 show the effect of contamina-
tion rates for elliptic envelope models for the 14-Bus-AM2 dataset.
Contamination Elapsed
Rate F1-Score Accuracy Precision Sensitivity Specificity Time
Table A .21 and corresponding graph in Figure A .33 show the effect of contamina-
tion rates for elliptic envelope models for the 14-Bus-AM3 dataset.
Contamination Elapsed
Rate F1-Score Accuracy Precision Sensitivity Specificity Time
Table A .22 and corresponding graph in Figure A .34 show the effect of contamina-
tion rates for elliptic envelope models for the 57-Bus-AM1 dataset.
Contamination Elapsed
Rate F1-Score Accuracy Precision Sensitivity Specificity Time
Table A .23 and corresponding graph in Figure A .35 show the effect of contamina-
tion rates for elliptic envelope models for the 57-Bus-AM2 dataset.
Contamination Elapsed
Rate F1-Score Accuracy Precision Sensitivity Specificity Time
Table A .24 and corresponding graph in Figure A .36 show the effect of contamina-
tion rates for elliptic envelope models for the 57-Bus-AM3 dataset.
Contamination Elapsed
Rate F1-Score Accuracy Precision Sensitivity Specificity Time
APPENDIX B
A D A M A G E M I T I G AT I O N M O D E L
B.1 INTRODUCTION
In this chapter, a model for damage mitigation when SFDI attacks take place without
being detected and damage is either imminent or on the way is proposed.6 The
rationale for such a model is that despite the availability of sophisticated defense
mechanisms, smart grid security breaches may happen causing serious damages. In
the age of cheap and popular internet-enabled commodity products such as smart
home devices, health monitors, vehicles, etc., attackers are finding increasingly more
When fast changing attack signatures are used as a strategy to force the security
apparatuses to relearn the attack properties and rethink countermeasures, critical time
is lost and substantial damage can result. An attack strategy is considered that has
the potentiality to breach smart grid security for which there is limited defense, and
the detection of the attack takes substantially long time. It is proposed that in such
B.2 T H E G R I D W AT C H M O D E L
The assumption is that in the potentially compromised SG, the incidence response plan
(IRP) no longer is reflective of a true defense or damage control plan for the entire SG
although the IRP was developed meticulously with considerable expense. However, it
may still be an effective plan for part of network and so it is worth implementing the
plan for the pertinent sub-network. Because for those sub-networks, the IRP is still a
valid response to any real failure. The proposed GridWatch damage mitigation model
deployed due to misleading state estimation (SE) by the main controller. Since the
actual state of the SG is unreliable, a predicted time T is used that takes into account
the time needed to detect attacks and estimated grid failure time, to isolate the SG into
so that the global state information does not affect the local operations. The choice of
k will depend on the time-dependent importance or value of the sub-network, its role
in the global SG, and its ability to sustain valuable services, within the predicted time
T. In essence, the most valuable sub-networks is chosen to be saved first even if that
means that a few low priority sub-networks have been compromised that could have
of vertices V . A vertex can represent any component of the smart grid, such as a
communication system, etc. The properties of a vertex will depend on the component
163
it represents. Some of the common properties include capacity, type, criticality index,
tional way. There can be two kinds of edges: i) power-flow and ii) ICT connection.
and the capacity of the power-line. Similarly, an ICT connection will represent if
components are connected over network communication line. The FDI attacks can
propagate only through the ICT edges, and cascading failures can only travel via the
power-flow edges.
The set G is a set of graphs g of the form hV, Ei over the sets V ⊆ V and E ⊆ E such
that for any two graphs g1 = hV1 , E1 i and g2 = hV2 , E2 i, g1 , g2 ∈ G, the sets of vertices
and edges are pairwise disjoint, i.e., V1 ∩ V2 = ∅ and E1 ∩ E2 = ∅. This implies that
the graphs gi s are sub-networks that are not inter-connected. Furthermore, the sets E
and Ei s are pairwise disjoint as well, i.e., E ∩ Ei = ∅, for all i. That means, the edges
in E help connect the sub-networks within the SG. Figure B .1(a) shows a GridWatch
b, and c) shown in blue. There are three sub-networks – red, blue and green (shown
using edge colors). In this figure, red nodes are more valuable than the greens, and
greens more than the yellow nodes. Figure B .1(a) shows three embedded graphs –
blue, red and green. The thicker solid lines are active, and the thinner dashed lines
are inactive but alternate connections. Figure B .1(b) shows re-routing of nodes a and
The SG is also a tuple of the form hG , C, Σ, I, ∆i, and so are each of the sub-networks
such that ⊆ I, i.e., the IRP for the SG is equal to or more that the union of all the
S
j Ij
164
individual sub-network IRPs. The damage mitigation plan ∆ is a function which aims
π within a time budget T. The list S is computed such that for every member si ∈ S,
for i = 1, . . . , |S|, ϕ(si ) ≥ ϕ(si+1 ) at time t of the valuation. The valuation function ϕ
IRP I can remain effective), the importance of the network si in G , and topology of G .
However, the partition function π excludes all elements in G that are potentially
compromised (based on sensor data and other network information), and creates
able if they are spatially clustered, and includes all required SG components that are
functional. The partition function potentially alters the topology of G since it considers
from the set of partitions returned by π, possibly fusing multiple partitions into one
that share connecting edges or bottleneck partitions to maximize the value of the target
and distribution systems sufficient to meet the energy demands of the consumers in
165
the sub-network, has at least one controller, an SE, and sufficient IRPs. Also, the
for all values of i and j. To limit the possible damage to the sub-network, controller
Figure B .1(b) shows the SG in Figure B .1(a) in which it was determined that the
node marked black has been potentially compromised. To limit the potential damage,
with the help of the functions π and ϕ, ∆ reassigned node 1 under the control of node
B.3 C U R R E N T S TAT E O F G R I D W AT C H
GridWatch is a conceptual model that aims to limit physical damage to SG, self-inflicted
or otherwise, in real time. GridWatch recognizes the fact that there are many attack
types, and not all attacks are recognizable, especially FDI attacks. It aims to buy time
by isolating the grid into autonomous but smaller SGs so that proper detection and
counter measures may be taken. It also admits some damage, if it must, to save the
that the spoofed data is a precursor of an attack, and the stealthy intent is to force the
controller sustain self-inflicted wounds and make it damage the most-valuable sub-
network.
166
B.4 S U M M A RY
An outline of the conceptual model for a best-effort damage mitigation scheme is pre-
sented in this chapter. This conceptual model needs to be implemented and deployed
APPENDIX C
PERMISSIONS TO R E U S E P U B L I S H E D PA P E R S