Adversarial Attacks Agaisnt Pe Survey2112.12310
Adversarial Attacks Agaisnt Pe Survey2112.12310
Xiang Linga , Lingfei Wub , Jiangyu Zhangd , Zhenqing Qud , Wei Dengd , Xiang Chend ,
Yaguan Qianc , Chunming Wud , Shouling Jid , Tianyue Luoa , Jingzheng Wua,∗, Yanjun Wua,∗
a
Institute of Software, Chinese Academy of Sciences, Beijing, 100190, Beijing, China
b
Pinterest, New York, 10018, NY, USA
c
Zhejiang University of Science and Technology, Hangzhou, 310023, Zhejiang, China
d
Zhejiang University, Hangzhou, 310027, Zhejiang, China
arXiv:2112.12310v5 [cs.CR] 17 Feb 2023
Abstract
Malware has been one of the most damaging threats to computers that span across multiple operat-
ing systems and various file formats. To defend against ever-increasing and ever-evolving malware,
tremendous efforts have been made to propose a variety of malware detection that attempt to
effectively and efficiently detect malware so as to mitigate possible damages as early as possible.
Recent studies have shown that, on the one hand, existing machine learning (ML) and deep learn-
ing (DL) techniques enable superior solutions in detecting newly emerging and previously unseen
malware. However, on the other hand, ML and DL models are inherently vulnerable to adversarial
attacks in the form of adversarial examples, which are maliciously generated by slightly and care-
fully perturbing the legitimate inputs to misbehave. Adversarial attacks are initially studied in the
domain of computer vision like image classification, and then quickly extended to other domains,
including natural language processing, audio recognition, and even malware detection.
In this paper, we focus on malware with the file format of portable executable (PE) in the
family of Windows operating systems, namely Windows PE malware, as a representative
case to study the adversarial attack methods in such adversarial settings. To be specific, we
start by first outlining the general learning framework of Windows PE malware detection based
on ML/DL and subsequently highlighting three unique challenges of performing adversarial at-
tacks in the context of Windows PE malware. Then, we conduct a comprehensive and system-
atic review to categorize the state-of-the-art adversarial attacks against PE malware detection,
as well as corresponding defenses to increase the robustness of Windows PE malware detec-
tion. Finally, we conclude the paper by first presenting other related attacks against Windows
PE malware detection beyond the adversarial attacks and then shedding light on future research
directions and opportunities. In addition, a curated resource list of adversarial attacks and de-
fenses for Windows PE malware detection is also available at https://github.com/ryderling/
adversarial-attacks-and-defenses-for-windows-pe-malware-detection.
Keywords: Portable Executable, Malware Detection, Machine Learning, Adversarial Machine
Learning, Deep Learning, Adversarial Attack
∗
Corresponding Author.
Email addresses: [email protected] (Xiang Ling), [email protected] (Lingfei Wu),
1. Introduction
With the rapid development and advancement of information technology, computer systems
are playing an indispensable and ubiquitous role in our daily lives. Meanwhile, the cyberattack
that attempts to maliciously exploit the computer system with malicious intentions (e.g., dam-
aging computers or gaining economic profits) has been an important type of ever-increasing and
constantly evolving security threat in our society. Malware (i.e., short for Malicious software) is
one of the most common and powerful cyberattacks for attackers to perform malicious activities
in computer systems, such as stealing confidential information without permissions, compromis-
ing the whole system, and demanding for a large ransom. While malware spans across multiple
operating systems (e.g., Windows, Linux, macOS, Android, etc.) with various file formats, such
as portable executable (PE), executable and linkable format (ELF), Mach-O, Android application
package (APK), and portable document format (PDF), we focus on malware with the PE files in
the family of Windows operating systems (namely Windows PE malware) in this paper due to
the following two reasons. First, malware analysis techniques (e.g., detection methods) for Win-
dows PE files are mostly different from those for other files like APK and PDF files because their
underlying operating systems, the file format, and execution modes are significantly different from
each other. Research shows there is no universal malware detection that can satisfactorily detect
all kinds of malware, and thus existing literature papers on malware analysis commonly point out
what specific operating system they target and what file format they are [1, 2]. That is the very
first and most important reason why we focus on Windows PE malware in this paper. Second,
Windows is the most worldwide popular and long-standing operation system for end users while
the malware in the file format of PE constitutes the earliest and maximum studied threat in the
wild [3]. According to the statistics of Kaspersky Lab at the end of 2020, there are an average
of 360,000 malware detected by Kaspersky per day and over 90% of which are Windows PE mal-
ware [4]. Similar statistical trends have been reported by Kaspersky Lab at the end of 2021 [5],
indicating Windows PE files are still not sufficiently protected until now.
To mitigate and address the ever-increasing number of security threats caused by Windows PE
malware, there are tremendous research efforts have been made to detect Windows PE malware
effectively and efficiently [6, 7, 1, 2, 8]. In particular, traditional malware detection can be traced
back to signature-based malware detection, which determines whether a given suspicious software
is malicious or not (i.e., malware or goodware) by comparing its signature with all signatures
from the maintained database of malware that has been previously collected and confirmed. It is
obvious that the fatal flaw of signature-based malware detection is that it can only detect previously
collected and known malware due to the heavy reliance on the malware database. In the last few
decades, inspired by the great successes of ML and DL in various long-standing real-world tasks
(e.g., computer vision, natural language processing, speech recognition, etc.), a variety of ML/DL-
based malware detection methods [7, 1, 2, 8] that leverage the high capacity of ML/DL models
have been adapted and presented to detect Windows PE malware. In general, these ML/DL-based
malware detection methods claim that they can generalize well to predict the new and previously
unseen (i.e., zero-day) malware instances due to the inherent generalizability of ML/DL models.
2
Accepted by ELSEVIER Computers & Security.
Unfortunately, recent studies have demonstrated that existing ML/DL models are inherently
vulnerable to adversarial attacks in the form of adversarial examples, which are maliciously gen-
erated by slightly and carefully perturbing the legitimate inputs to confuse the target ML/DL
models [9, 10]. Since the creation of adversarial attacks, most research papers focused on studying
adversarial attacks in the domain of computer vision [11], e.g., slightly and carefully perturbing
a “Persian cat” image x such that the resulting adversarial image x0 can be misclassified as a
“lionfish” by the target image classifier. Normally, in the context of images, most proposed adver-
sarial attacks resolve to the feature-space attack like various gradient-based methods, which can
be directly applied to generate adversarial images. Until now, there have been a large number
of adversarial attack methods and corresponding defense mechanisms being proposed by security
researchers and practitioners in both academia and industry [12, 13, 14, 15, 16].
Inspired by those studies of adversarial attacks in the context of images, a natural question
arises is that, is it feasible to perform adversarial attacks against existing malware detec-
tion methods, especially against ML/DL based PE malware detection? To answer the
aforementioned question, in recent five years, security researchers and practitioners have proposed
lots of adversarial attacks in the context of malware [17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28,
29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55],
requiring that the generated adversarial malware file should not only be misclassified as the “good-
ware” by the target malware detection model, but also behaves exactly the same as the original
malware file. Compared to the adversarial attack in the context of images, exploring the adver-
sarial attack in the context of malware is completely different and extremely challenging due to
the highly structured nature of software files like PE files. To put it simply, even though we can
employ existing feature-space attacks to generate the “adversarial features of malware”, it is sig-
nificantly challenging to find the corresponding “adversarial malware file” that can preserve the
format, executability, and maliciousness the same as the original malware file.
Related work to this survey. In fact, for the general adversarial attacks and defenses,
there are lots of surveys being done on the image, audio, video, text, and graph [11, 12, 13, 14,
15, 16, 56, 57, 58, 59], but very few surveys focusing on the adversarial attacks in the context of
malware [60, 61, 62]. Here, we introduce those closely related surveys and highlight their limitations
and differences compared with our survey paper as follows.
• In [60], Pierazzi et al. are the first to present a general mathematical formalization of adver-
sarial attacks in the problem-space and further propose a novel problem-space attack against
Android malware detection. Although they identify four key constraints and commonalities
among different domains (i.e., image classification, face recognition, code attribution, PDF
malware, android malware, etc.) in terms of the problem-space adversarial attack, this pa-
per [60] is not a survey paper as it neither extensively collects all existing research efforts
nor systematically categorizes and summarizes these efforts in this research direction.
• In [61], Park and Yener actually conduct an incomplete and improper survey in re-
viewing existing adversarial attacks against malware classifiers, since the authors mistakenly
categorize them into two categories: gradient-driven and problem-driven approaches, which
are clearly not sufficient to cover all existing adversarial attacks against malware detec-
tion. For instance, the semantics-preserving reinforcement learning (SRL) attack proposed
by Zhang et al. [38] is neither a gradient-driven nor a problem-driven method. In addition,
this paper [61] lacks surveying existing defense methods against such adversarial attacks.
3
Accepted by ELSEVIER Computers & Security.
• In [62], Li et al. first make a series of formulations and assumptions in the context of ad-
versarial malware detection and then attempt to survey this research field of adversarial
malware detection in a broad spectrum of malware formats, including Windows PE, Android
Package Kit (APK), and Portable Document Format (PDF), which is supposed to be too
coarse-grained to fully understand the unique challenges adversarial attacks and defenses
for different malware formats. It is well-known that malware detection relies heavily on
the specific file formats of malware, and thus existing literature papers on malware analysis
commonly point out what specific operating system they target and what file format they
are [1, 2]. Therefore, our paper instead focuses on the malware format of Windows PE,
which allows us to specifically identify the distinct characteristics of Windows PE malware
and further gain a deeper and thorough understanding of adversarial attacks and defenses in
terms of Windows PE malware.
• To the best of our knowledge, this is the first work that summarizes and highlights three
unique challenges of adversarial attacks in the context of Windows PE malware in the wild,
i.e., format-preserving, executability-preserving, and maliciousness-preserving.
• We conduct a comprehensive and systematic review for adversarial attacks against Windows
PE malware detection and propose a complete taxonomy to categorize the state-of-the-art
adversarial attack methods from different viewpoints.
• We summarize the existing adversarial defenses for PE malware detection against the pro-
posed adversarial attacks. In addition, we discuss other types of attacks against Windows
PE malware detection beyond adversarial attacks.
Organization. The rest of this paper is organized as follows. We introduce the general layout
of PE files and outline the ML/DL-based learning framework of PE malware detection in § 2. In § 3,
we manifest three unique challenges of adversarial attacks against PE malware detection compared
with the general adversarial attacks in the context of images. § 4 first presents our taxonomy of
adversarial attacks against PE malware detection and then gives a detailed introduction to the
state of the art. We summarize the existing adversarial defenses for PE malware detection in § 5.
In § 6, we first discuss other types of attacks against PE malware detection beyond the adversarial
attacks and then point out some research directions and possible opportunities for future work.
We conclude our survey paper in § 7.
4
Accepted by ELSEVIER Computers & Security.
This section aims to provide the basics to understand how to take advantage of ML and DL
for malware detection in terms of PE files in the family of Windows operating systems (OSs). In
particular, we first introduce the general layout of PE files and PE malware in § 2.1, and then
outline the general learning framework of PE malware detection models based on ML/DL in § 2.2.
Header Information
dows OSs to execute code and store essential data that is nec- DOS Header
information to ensure the executable is positioned and executed correctly. The “.rsrc” section
contains all resources (e.g., icons, menus, etc.). The last group is the other un-mapped data
which will not be mapped into memory. In particular, the un-mapped data refers to the chunk of
unused bytes, like debug information, at the end of PE files.
Within the family of Windows OSs, PE mainly has two typical and most commonly used file
types, i.e., EXEcutable (EXE) and Dynamic Link Library (DLL), which are generally ended with
“.exe” and “.dll” as the suffix name. Normally, an “.exe” file can be run independently while a “.dll”
file contains the library of functions that other executables can use in the Windows platform [64].
2.1.2. PE Malware
Malicious software, i.e., malware, is purposely designed and implemented to satisfy the mali-
cious goals and intentions of attackers, e.g., accessing the system without user permissions, stealing
private or confidential information, asking for a large ransom, etc. Since the PE file format was first
created in the family of Windows OSs, PE files have been widely leveraged by malicious attackers
to build PE malware. Until now, according to the security reports from AV-TEST Institute [66]
and Avira Protection Labs [67], PE malware still remains the predominant threat for both personal
users and business users in the wild for the following two major reasons. First, the worldwide pop-
ularity of the family of Windows OSs and the commonality of PE files inside make the family of
Windows OSs, especially the PE file that can be executed, become the main target of attackers for
benefit maximization. Second, unlike other file types, PE files can be self-contained, which means
that PE malware can include all needed data and does not require additional data to launch the
attack. In addition, based on different types of proliferation and different malicious intentions, PE
malware can be further briefly classified as viruses, trojans, PUA, worms, adware, ransomware,
etc. For more details, we refer interested readers to [68, 69, 7].
6
Accepted by ELSEVIER Computers & Security.
2.2 Learning Framework for PE Malware Detection
Data Acquisition (Input) Feature Engineering Learning from Models and Predictions
Naïve Bayes
Machine Learning
Byte Sequence
EXE EXE EXE
Static
Readable String Decision Tree
…… SVM
Dynamic
File Status MLP
EXE EXE EXE Problem Space Registry Status Feature Space Label Space
Deep Learning
CNN
…… RNN = {0,1}
EXE EXE EXE
Hybrid Opcode LSTM
Table 1: Common Features and Corresponding Learning Models for PE Malware Detection.
Category
Representative ML/DL Representative PE Malware
Features
Static
Byte Sequence X Naı̈ve Bayes [82], SVM [83], Schultz et al. [3], Kolter and
DT [84], MLP [85], CNN [86], Maloof [88], Saxe and Berlin
LightGBM [87]. [89], Gibert et al. [90],
MalConv [91], EMBER [92].
Readable String X Naı̈ve Bayes [82], SVM [83], Schultz et al. [3], SBMDS [94],
MLP [85], DT [84], RF [93], Islam et al. [95], Saxe and
LightGBM [87]. Berlin [89], EMBER [92].
Header
X MLP [85], LightGBM [87], Saxe and Berlin [89],
Information
Naı̈ve Bayes [82], SVM [83], EMBER [92], PE-Miner [96].
DT [84].
Grayscale Image X kNN [97], CNN [86], Nataraj et al. [98], Kim [99],
SVM [83]. Visual-AT [100].
CPU/Memory/IO
X CNN [86], kNN [97], Rieck et al. [101], AMAL [102],
etc. Status
SVM [83]. Abdelsalam et al. [103].
File Status X Hierarchical Clustering [104], Bailey et al. [106], Rieck et al.
CNN [86], kNN [97], [101], AMAL [102],
SVM [83], GNN [105]. Abdelsalam et al. [103],
MatchGNet [107].
Registry Status X Hierarchical Clustering [104], Bailey et al. [106], Rieck et al.
kNN [97], SVM [83]. [101], AMAL [102].
Network Status X Hierarchical Clustering [104], Bailey et al. [106], Rieck et al.
CNN [86], kNN [97], [101], AMAL [102],
SVM [83], GNN [105]. Abdelsalam et al. [103],
MatchGNet [107].
Opcode X X kNN [97], SVM [83], DT [84], AMCS [109], Santos et al.
RNN [108], CNN [86], [110], IRMD [111],
Hierarchical Clustering [104]. RMVC [112].
System or API
X X RIPPER [113], SVM [83], Schultz et al. [3], SBMDS [94],
Calls
Hierarchical Clustering [104], Rieck et al. [101], Qiao et al.
CNN [86], LSTM [114]. [115], Zhang et al. [116].
Control Flow
X X Naı̈ve Bayes [82], SVM [83], Kapoor and Dhavale [117],
Graph
RF [93], GNN [105]. MAGIC [118], MalGraph [119].
Function Call
X X RF [93], AutoEncoder [120], Hassen and Chan [121],
Graph
CNN [86], GNN [105]. DLGraph [122], DeepCG [123],
MalGraph [119].
∗ If the paper does not clearly name the PE malware detection, we use the author name(s) of the paper with its reference.
8
Accepted by ELSEVIER Computers & Security.
2.2 Learning Framework for PE Malware Detection
of properties of PE files, there is a large body of work on extracting various features, which can
be generally categorized into three broad category: static features, dynamic features and hybrid
features [7, 2, 8] and summarized in Table 1.
First of all, static features are directly extracted from the PE samples themselves without
actually running them. For instance, byte sequence, readable string, header information, and the
grayscale image are commonly used static features for PE malware detection.
• Readable String: A PE file might contain readable strings that reflect it intentions or se-
mantics, like file names, IP addresses, domain names, author signatures, etc. After ex-
tracting readable strings in a PE file, their numeric feature representation can be a set
of binary attributes (i.e., whether the string exists), frequencies, or even 2D histogram fea-
tures [3, 94, 95, 89].
• Grayscale Image: Since the value range of bytes in a PE file is the same as the pixel value in
an image, a visualization-based feature engineering approach is to transform a PE file into a
grayscale image, for which each byte in a PE file corresponds to a pixel in an image [126].
Inspired by the recent great successes of image classification methods, a lot of visualization-
based methods have also been proposed for PE malware detection [98, 99, 100].
Second, dynamic features refer to those features that can be extracted by first running the
executable in an isolated environment (e.g., sandbox, virtual machine, etc.) and then monitoring
their runtime status in terms of system resources, files, registries, network, and others.
• System Resource Status: The execution of malware inevitably occupies system resources
(e.g., CPU, memory, IO, etc.), whose runtime status can be considered as dynamic features for
malware detection, as a variety of malware within one specific family might follow a relatively
fixed pattern of system resources during execution. In particular, CPU usage, memory usage,
and I/O request packets are commonly monitored as dynamic features [101, 102, 103].
• File Status: Malware normally needs to operate on files of target users for reaching malicious
intentions by attackers. Thus, logging and counting for the files accessed, created, modified,
or deleted are commonly used dynamic features in malware detection [106, 101, 102, 103, 107].
• Registry Status: Registries that store the system/application-level configurations are im-
portant for the family of Windows OSs. The malware could operate on registries with
malicious intentions, like self-starting malware. Similar to file status, registry status like
9
Accepted by ELSEVIER Computers & Security.
2.2 Learning Framework for PE Malware Detection
counting the registries created, modified, and deleted can also be regarded as dynamic fea-
tures [106, 101, 102].
• Network Status: The spread of malware like trojans and ransomware mainly depends on the
network. Taking trojans as an example, they are likely to connect remote servers with certain
network ports. Therefore, when diving into the specific aspects of network status, there is
a variety of network-level information that can be used for creating a rich set of dynamic
features [106, 101, 102, 103, 107], such as the number of distinct IP addresses or certain ports,
the number of different HTTP requests (e.g., POST, GET, HEAD, PUT, etc.), the number
of common DNS record types (e.g., PTR, CNAMN, SOA, etc.), to name just a few.
Finally, we exemplify four commonly used hybrid features, i.e., opcode, system/API calls,
control flow graph (CFG), and function call graph, which can be extracted from executables with
either static analysis methods or dynamic analysis methods. For instance, opcodes of executables
can be obtained by either extracting from their disassembled instructions or monitoring their
runtime instructions in memory.
• Opcode: Executables, including malware, can be generally considered as a collection of in-
structions that are executed in a specific order. In machine assembly language, an instruction
consists of an opcode and several operands, in which the opcode specifies the operation to be
executed and the operand refers to the corresponding data or its memory location. As prior
studies suggest, the opcode distributions of malware statistically differ from goodware, and
thus various features are constructed from the opcodes, such as their frequency, n-grams of
opcode sequences, or even opcode images [109, 110, 111, 112].
• System/API Calls: System/API calls refer to how executables interact with system-level
or application-level libraries in the family of Windows OSs. Similar to the opcode, various
feature representations are thus constructed from system/API calls, such as the frequency of
system/API calls, and n-grams of system/API call sequences [3, 94, 101, 115, 116].
• CFG: The CFG is a graph-based feature representation that is commonly used to characterize
the control flow of executables, including PE malware [117, 118]. Building from assembly in-
structions of executables, each node in the CFG represents a sequence of instructions without
branching and each edge represents the control flow path between two nodes [127].
• Function Call Graph: The function call graph [128] that attempts to build the caller-callee
relation between different functions (including system/API or user-implemented functions), is
regarded as a more coarse-grained graph representation compared with CFG [121, 122, 123].
It is worth noting that, on the one hand, we just briefly review and categorize the commonly
used features in PE malware detection, and do not attempt to cover all, which is not the goal of
our paper. On the other hand, all the features mentioned above are not separate or independent,
they are actually mixed for PE malware detection in the wild. In essence, the process of feature
engineering can be broadly expressed as a feature mapping function that maps the problem space
into the feature space (i.e., the numeric features), which is formulated in Def. 1 as follows.
Def. 1 (Feature Mapping Function). A feature mapping function φ is formulated as φ : Z →
X, in which Z denotes the problem space from a specific real-world application, and X denotes
the corresponding feature space, numerically describing the intrinsic properties of objects in the
problem space.
10
Accepted by ELSEVIER Computers & Security.
The training process of a malware detection model is to learn the model parameters based on
a large number of training samples, so that the malware detection model can approximate the
real relationship function between the feature patterns of executables and their binary detection
labels. After that, to predict whether a given executable is malware or not, the malware detection
model with learned parameters can effectively and efficiently compute the probabilities assigned to
both classes of malware and goodware. In order to find the most applicable model, it is actually
quite common to test different ML/DL models for PE malware detection depending on the specific
task. In Table 1, the last two columns present the representative ML/DL model architectures and
corresponding PE malware detection methods with references.
In this section, we first introduce the general concept and taxonomy of adversarial attacks that
have been originally and extensively studied in the domain of image classification tasks, and then
manifest the most unique challenges of adversarial attacks for PE malware when compared with
other fields like images, audios, texts, etc.
11
Accepted by ELSEVIER Computers
3.1 & Adversarial
Security. Attacks: The General Concept and Taxonomy
z x
Images Executables
Feature Mapping
𝝓𝝓
EXE
Malware: 94.3%
Persian Cat: 95.7% Goodware: 5.7% Problem-space Feature-space
Attack Attack
EXE
𝝓𝝓−𝟏𝟏
Malware: 10.0%
Lion Fish: 99.8%
Goodware: 90.0%
Inverse
Feature Mapping
z' x'
Figure 3: Illustration of the connection between the feature-space attack and the problem-space attack, in
which the feature mapping function φ and the inverse feature mapping function φ−1 act as bridges for transitions
between the feature-space and the problem-space.
min
0
cost(z 0 , z)
z
s.t. z 0 ∈ Z (2)
0 0
f (φ(z )) = y 6= y
12
3.2 Three Unique Challenges of Adversarial Attacks for PE Malware: From Feature-space to
Accepted by ELSEVIER Computers & Security. Problem-space
in which cost(z2 , z1 ) denotes any cost function that transforms z1 into z2 in the problem-space of
the specific application.
When comparing the problem-space attack in Eq. (2) with the feature-space attack in Eq. (1),
it is easy to find the most fundamental and noticeable difference between them is, the problem-
space attack involves a feature mapping function φ that maps the problem-space into the feature-
space, which is usually neither invertible nor differentiable. Therefore, the problem-space attack
can hardly use gradient-based methods directly to generate adversarial examples. In Fig. 3, we
illustrate the connection between the feature-space attack and the problem-space attack.
3.2. Three Unique Challenges of Adversarial Attacks for PE Malware: From Feature-space to
Problem-space
Originally, adversarial attacks are explored in the domain of image classification tasks and a
variety of feature-space attacks are subsequently proposed to generate adversarial examples for the
malicious purpose of misclassification, e.g., misclassifying a Persian cat into a lionfish with a high
probability of 99.8% as depicted in Fig. 3. Actually, the main reason for the success of directly
performing the feature-space attack to generate adversarial examples of images is that, it is easy
to find the corresponding image z 0 from the generate adversarial feature x0 via the inverse feature
mapping function φ−1 (i.e., z 0 = φ−1 (x0 )), as indicated in Fig. 3. However, when considering the
adversarial attacks for the PE files, the circumstance becomes completely different and extremely
challenging due to the problem-feature space dilemma [130], which is mainly manifested in
the following two aspects.
1. The feature mapping function φimage for images is relatively fixed (i.e., an image can be
formatted as a two-dimensional array of pixels where each pixel value is a three-dimensional
RGB vector with a continuous value between 0 to 255), while the feature mapping function
φpe for PE files is not fixed and can take various and diverse approaches of feature engineering
as detailed in § 2.2.3. Especially in the setting of black-box attacks, the attacker cannot know
the specific feature mapping function φpe for PE files, which greatly increases the difficulty
of adversarial attacks for PE files.
1
There is no unified view on whether to treat the scenario of knowing the classification label with probability as
the black-box attack.
13
3.2 Three Unique Challenges of Adversarial Attacks for PE Malware: From Feature-space to
Accepted by ELSEVIER Computers & Security. Problem-space
2. For images, although the inverse feature mapping function φ−1 image is not exactly bi-injective
(e.g., the pixel value might not be in the range of 0 to 255), it is continuously differentiable,
and thus the feature-space attack based on gradients can directly apply on images to generate
adversarial examples. However, for various different feature mapping functions of PE files,
to map a feature vector in the feature-space into an executable in the problem-space, it is
almost impossible to find an exact or approximate function of inverse feature mapping φ−1 pe
that is either bi-injective or differentiable.
As depicted in Fig. 4, in order to generate adversarial examples for PE files, although there
is a variety of adversarial attacks that exploit the feature-space attacks based on gradients have
been proposed, we argue that these adversarial attacks are impractical and unrealistic against
PE malware detection in the wild world. This is because what these adversarial attacks generate
is the “adversarial PE feature” rather than the “adversarial PE malware” in the end, and an
“adversarial PE feature” does not guarantee to correspond to an “adversarial PE malware” due
to the following two reasons. On the one hand, it is almost impossible to find a corresponding
adversarial PE malware z 0 based on the generated adversarial PE feature x0 , as the inverse feature
mapping function φ−1 pe is normally neither bi-injective nor differentiable. On the other hand, even
though we could find the exact PE malware z 0 in the problem-space that corresponds to the
generate adversarial feature x0 in the feature-space, there is no guarantee that the found z 0 is also
“adversarial”. Taking the x03 in Fig. 4 as an example, although its feature representation x03 in
feature-space is misclassified as benign (i.e., f (x03 ) = 0), but its corresponding PE malware object
z30 in problem-space is still detected as malicious (i.e., f (φ(z30 )) = 1 6= 0).
Therefore, to further generate practical and realistic adversarial PE malware against malware
detection in the wild, one of the possible or even the only way so far is to seek for the problem-
space attack to generate adversarial PE malware in the problem-space as defined in Eq. (2). To be
specific, as depicted in Fig. 4, current problem-space attacks normally attempt to find and apply
a series of self-defined problem-space transformations (i.e., T1 , T2 , T3 and T4 ) that sequentially
T T
transform the original PE malware z into the desired adversarial PE malware z 0 (i.e., z −−→
1
z10 −−→
2
T T
z20 −−→
3
z30 −−→
4
z 0 ), such that ¶ z 0 is no longer detected as malicious by the target malware detection,
0
and · z maintains the same semantics as the original z. In the following parts, we detail the three
unique challenges of maintaining the semantics of adversarial PE malware for practical and realistic
adversarial attacks against PE malware detection and present the relationship between the three
challenges in Fig. 5.
Benign
x'
x3
x1
x
x2
malicious
Malicious
3.2 Three Unique Challenges of Adversarial Attacks for PE Malware: From Feature-space to
Accepted by ELSEVIER Computers & Security. Problem-space
Problem-space benign
malicious Format-preserving
Maliciousness-
preserving
In order to explore the most promising advances of adversarial attacks against PE malware
detection, in this section, we comprehensively and systematically categorize state-of-the-art ad-
versarial attacks from different viewpoints, i.e., adversary’s knowledge, adversary’s space, target
malware detection, and attack strategy. Fig. 6 illustrates the general category of adversarial at-
tacks against PE malware detection of this paper. In the following subsections, in terms of the
adversary’s knowledge (i.e., white-box versus black-box), we will first introduce the white-box
adversarial attacks against PE malware detection in § 4.1, and then introduce the black-box ad-
versarial attacks in § 4.2. Finally, we highlight the summary of state-of-the-art adversarial attacks
against PE malware detection in § 4.3.
Section 4.1
……
Problem-Space
Raw Byte based Malware Detectors
Attacks
Section 4.1.2 Explaination-based
Method
Adversarial Attack
Attacks Strategy Reinforcement Learning
API Call List based Malware Detectors
continuous bytes of sections in the middle of PE malware. First, they iteratively generate the
adversarial payload with the gradient-based method of FGSM [9] in the continuous feature space.
Then, to generate the practical adversarial PE malware, they attempt to reconstruct back to the
input problem space by directly searching the closest neighbor to the generated adversarial payload.
Similarly, Suciu et al. [18] extend the FGSM-based adversarial attacks with two previously
proposed strategies (i.e., append-FGSM and slack-FGSM [17]), and further perform a systematic
evaluation to compare the effectiveness of both append and slack strategies against MalConv. Their
experimental results show that slack-FGSM outperforms append-FGSM with a smaller number of
modified bytes. Possible reasons are that the appended bytes of append-FGSM might exceed the
maximum size of the model input (e.g., 2MB for MalConv), or that slack-FGSM can make use
of surrounding contextual bytes to amplify the power of FGSM since the CNN-based MalConv
detector requires the consideration of the contextual bytes within the convolution window.
Chen et al. [19] suggest that all those adversarial attacks [17, 18] append or inject adversarial
bytes that are first initialized by random noises and further iteratively optimized, which might
lead to inferior attack performance. To address this issue, Chen et al. propose two novel white-
box attacks (i.e., BFA and Enhanced-BFA) with the saliency vector generated by the Grad-CAM
approach [131]. For BFA, it selects the data blocks with significant importance from benign PE
files using computed saliency vectors and then appends those data blocks to the end of the original
PE malware. Besides that, Enhanced-BFA is presented to use FGSM to iteratively optimize these
perturbations generated by BFA. Experimental results show that Enhanced-BFA and BFA have
comparative attack performances when the number of appending bytes is large, but Enhanced-BFA
is ten times more effective than BFA when the number of appending bytes is small.
Qiao et al. [20] propose a white-box adversarial attack against raw bytes based malware detec-
tors like MalConv. In particular, it first generates a prototype sample to maximize the output of
17
Accepted by ELSEVIER Computers
4.1 White-box
& Security.
Adversarial Attacks against PE Malware Detection
malware detection model towards the target class (i.e., malware) by directly applying the gradient
descent algorithm. Next, to ensure that the generated adversarial PE malware preserves both exe-
cutability and maliciousness as the input PE malware, it modifies the modifiable part of input PE
malware (i.e., bytes between sections, bytes at the end of PE files, and bytes in the newly added
section) in a fine-grained manner under the guidance of the generated prototype sample.
API Call List based Malware Detectors. By flipping the bits of the binary feature vector
of malware (“1” denotes the presence of one Windows API call and “0” denotes the absence),
Al-Dujaili et al. introduce four kinds of white-box adversarial attacks with k multi-steps, namely
dFGSMk , rFGSMk , BGAk , and BCAk [21], to attack API call list based malware detectors. To
be specific, dFGSMk and rFGSMk are two white-box adversarial attacks that are adapted mainly
from the FGSM attack in the continuous feature-space [9, 132] but extended for the binary feature
space via deterministic or randomized rounding, respectively. BGAk and BCAk are two gradient
ascent based attacks that update multiple bits or one bit in each step, respectively.
As the winner of “Robust Malware Detection Challenge” [133] in both attack and defense tracks,
Verwer et al. [134] propose a novel white-box adversarial attack with greedy random accelerated
multi-bit search, namely GRAMS, which generates functional adversarial API call features and
also builds a more robust malware detector in a standard adversarial training setting. The main
idea of GRAMS is to perform a greedy search procedure that explores gradient information as the
heuristic to indicate which bits to flip among all the binary search space (i.e., 22761 API calls).
At each iteration, GRAMS flips k bits of API calls that have the largest absolute gradient and
exponentially increases or decreases the value of k depending on whether GRAMS finds a better
or worse solution. To ensure the functionality of the generated adversarial malware, both [21]
and [134] limit the attack to flipping ‘0’ to ‘1’, meaning both of them only add irrelevant API calls.
Visualization based Malware Detectors. Differently, to attack the visualization-based
malware detectors, Liu et al. [22] propose the first white-box adversarial attack approach, namely
Adversarial Texture Malware Perturbation Attack (ATMPA), based on adversarial attacks in the
domain of image classification tasks [9, 10]. In particular, ATMPA first converts the malware sample
to a binary texture grayscale image and then manipulates the corresponding adversarial example
with subtle perturbations generated from two existing adversarial attack approaches - FGSM [9] and
C&W [10]. However, the major limitation of ATMPA is that the generated adversarial grayscale
image of the malware sample destroys the structure of the original malware and thus cannot be
executed properly, which makes ATMPA unpractical for real-world PE malware detection.
Similar to ATMPA, Khormali et al. present an adversarial attack COPYCAT [23] against vi-
sualization based malware detectors with CNNs. COPYCAT also makes use of existing generic
adversarial attacks (e.g., FGSM, PGD, C&W, MIM, DeepFool, etc.) to generate an adversarial
image. After that, COPYCAT appends the generated adversarial image to the end of the original
image of malware rather than directly adding it to the original malware image.
Differently, to evade visualization based malware detectors, Park et al. [24] propose another
adversarial attack based on the adversarial malware alignment obfuscation (AMAO) algorithm.
Specifically, a non-executable adversarial image is first generated by the off-the-shelf adversarial
attacks in the field of image classification [9, 10]. Then, in order to attempt to preserve the exe-
cutability, the adversarial PE malware is finally generated by the AMAO algorithm that minimally
inserts semantic N OP s at the available insertion points of the original malware such that the
modified PE malware is as similar as possible to the generated no-executable adversarial image.
Other Miscellaneous Malware Detectors. In [25], Li et al. first train ML-based malware
18
Accepted by ELSEVIER Computers
4.1 White-box
& Security.
Adversarial Attacks against PE Malware Detection
detection models based on OpCode n-gram features, i.e., the n-gram sequence of operation codes
extracted from the disassembled PE file. Then, the authors employ an interpretation model of
SHAP [135] to assign each n-gram feature with an importance value and observe that the 4-gram
“move + and + or + move” feature is a typical malicious feature as it almost does not appear
in the benign PE samples. Thus, based on this observation, the authors consider a generation
method of adversarial PE malware by instruction substitution. For instance, the “move + and +
or + move” in 10 sampled malware samples can be replaced with “push + pop + and + or + push
+ pop”, which can be used to bypass the malware detectors in their evaluation.
19
Accepted by ELSEVIER Computers
4.2 Black-box
& Security.
Adversarial Attacks against PE Malware Detection
Table 2: Summary of State-of-the-Art Adversarial Attacks against PE Malware Detection. WB/BB is short for the white-box
attack and the black-box attack, FS/PS is short for the problem-space attack and the feature-space attack, and / denotes
emptily/fully preserving the related property. In particular, as it is almost impossible to theoretically prove both properties
of executability-preserving and maliciousness-preserving, we thus use to denote that related property is preserved neither
conceptually nor empirically, to denote it is preserved both conceptually and empirically, and to denote it is only preserved
conceptually but not empirically without experimental verification.
Adversary’s Space
Format
empirical verification)
empirical verification)
Executability (with
Maliciousness (with
Attack Names Year Category Detection Name Transformation Strategy
Append or inject
Kreuk et al.
2018 WB FS Static MalConv the adversarial FGSM
[17]
payload
Append or inject
Suciu et al.
2019 WB FS Static MalConv adversarial FGSM
[18]
payload
Append the
BFA,
selected or
Enhanced- Grad-CAM
2019 WB FS Static MalConv optimized bytes
BFA or FGSM
from benign PE
[19]
files
Append or inject
Qiao et al. Gradient-
2022 WB FS Static MalConv adversarial
[20] based
payload
dFGSMk ,
rFGSMk , API call list based Add irrelevant Gradient-
2018 WB FS Static
BGAk , BCAk malware detectors API calls based
[21]
API call list based Add irrelevant Gradient-
GRAMS [134] 2020 WB FS Static
malware detectors API calls based
Visualization- Add adversarial
ATMPA [22] 2019 WB FS Static based malware noise to the FGSM, C&W
detectors malware image
Visualization- Append FGSM, PGD,
COPYCAT
2019 WB FS Static based malware adversarial noise C&W, MIM,
[23]
detectors generated DeepFool
Visualization-
Insert the
AMAO [24] 2019 WB FS Static based malware FGSM, C&W
semantic N OP s
detectors
Opcode
Opcode-based Interpretation
Li et al. [25] 2020 WB FS Static instruction
malware detectors model SHAP
substitution
Append Gradient-
AMB [26] 2018 WB PS Static MalConv
adversarial bytes based
Modify specific
Demetrio Gradient-
2019 WB PS Static MalConv regions in the PE
et al. [27] based
header
MalConv, DOS Header
Gradient-
RAMEn [28] 2020 WB PS Static Byte-based DNN Extension,
based
Model [125] Content Shifting
Binary
Lucas et al. MalConv, Gradient-
2021 WB PS Static diversification
[29] AvastNet [124] based
techniques
Table to be continued.
20
Accepted by ELSEVIER Computers
4.2 Black-box
& Security.
Adversarial Attacks against PE Malware Detection
Continued Table.
BB
API call list based Add irrelevant
MalGAN [30] 2017 w/o FS Static GAN
malware detectors API calls
prob.
BB
Improved API call list based Add irrelevant
2019 w/o FS Static GAN
MalGAN [31] malware detectors API calls
prob.
BB
API call list based Add or Remove Greedy
EvnAttack [32] 2017 w FS Static
malware detectors API calls Algorithm
prob.
BB API call sequence
Hu and Tan Insert irrelevant Generative
2017 w/o FS Dynamic based malware
[33] API calls Model
prob. detectors
API call sequence Insert irrelevant
Transferability,
GADGET [34] 2018 BB FS Dynamic based malware API calls with
Heuristics
detectors IAT Hooking
BB API call sequence Insert API calls
Greedy
ELE [35] 2019 w FS Dynamic based malware with IAT
Algorithm
prob. detectors Hooking
API call sequence Insert API calls
Evolutionary
BADGER [36] 2020 BB FS Dynamic based malware with IAT
Algorithm
detectors Hooking
BB Predefined Transferability,
Rosenberg
2020 w/ FS Static EMBER modifiable Explainable
et al. [37]
prob. features ML
BB Inject semantic
CFG-based Reinforcement
SRL [38] 2020 w/o FS Static N OP s into CFG
malware detectors Learning
prob. blocks
BB
gym-malware format-preserving Reinforcement
2017 w/o PS Static –
[39, 137] modifications Learning
prob.
BB
format-preserving Reinforcement
gym-plus [40] 2018 w/o PS Static –
modifications Learning
prob.
gym-malware- BB
format-preserving Reinforcement
mini 2020 w/o PS Static –
modifications Learning
[41] prob.
BB
format-preserving Reinforcement
DQEAF [42] 2019 w/o PS Static –
modifications Learning
prob.
BB
RLAttackNet format-preserving Reinforcement
2020 w/o PS Static –
[43] modifications Learning
prob.
BB
AMG-VAC format-preserving Reinforcement
2021 w/o PS Static –
[44] modifications Learning
prob.
BB
AIMED-RL format-preserving Reinforcement
2021 w/o PS Static –
[45] modifications Learning
prob.
BB
format-preserving Reinforcement
AMG-IRL [46] 2021 w/o PS Static –
modifications Learning
prob.
BB
format-preserving
ARMED [47] 2019 w/ PS Static – Randomization
modifications
prob.
BB Append strings
Dropper [48] 2019 w/o PS Static – from goodware & Randomization
prob. Packing
Experience-
BB Append bytes
Chen et al. based
2019 w/o PS Static MalConv from benign PE
[19] Randomiza-
prob. files
tion
Table to be continued.
21
Accepted by ELSEVIER Computers
4.2 Black-box
& Security.
Adversarial Attacks against PE Malware Detection
Continued Table.
BB format-preserving Weighted
Song et al.
2020 w/o PS Static – (macro & micro) Randomiza-
[49]
prob. modifications tion
BB
format-preserving Genetic
AIMED [50] 2019 w/ PS Static –
modifications Programming
prob.
BB
format-preserving Genetic
MDEA [51] 2020 w/ PS Static MalConv
modifications Algorithm
prob.
BB Inject and pad
Genetic
GAMMA [52] 2020 w/ PS Static – sections from
Algorithm
prob. benign PE files
BB
Append bytes to
GAPGAN [53] 2020 w/o PS Static MalConv GAN
the end
prob.
BB
Obfuscation-like Convolutional-
MalFox [54] 2020 w/o PS Static –
techniques GAN
prob.
Targeted BB
Occlusion of Binary
occlusion 2018 w/ PS Static –
important bytes Search
attack [55] prob.
BB Binary
Lucas et al. MalConv, Hill-climbing
2021 w/ PS Static diversification
[29] AvastNet [124] algorithm
prob. techniques
160 system-level API calls) of the target malware detector and considers only adding some irrelevant
API calls into the original malware sample for generating the adversarial malware samples in the
feature space. MalGAN first builds a differentiable substitute detector to fit the target black-box
malware detector and then trains a generator to minimize the malicious probability of generated
adversarial malware predicted by the substitute detector. Subsequently, Kawai et al. [31] further
present an Improved-MalGAN after addressing several issues of MalGAN from a realistic viewpoint.
For instance, Improved-MalGAN trains the MalGAN and the target black-box malware detector
with different API call lists while the original MalGAN trains with the same API call list.
In [32], Chen et al. introduce another black-box adversarial attack, namely EvnAttack. EvnAt-
tack first employs Max-Relevance [138] to calculate the importance of each API call in classifying
PE malware or goodware based on the training set and then ranks those API calls into two sets: M
and B. In particular, M contains API calls that are highly relevant to malware, while B contains
API calls that are highly relevant to goodware. Intuitively, EvnAttack is a simple and straight-
forward attack method that manipulates the API call list by either adding the API calls in B or
removing the ones in M . Specifically, EvnAttack employs a bidirectional selection algorithm that
greedily selects API calls for the manipulation of addition or removal based on the fact that how
the manipulation influences the loss of the target PE malware detector.
API Call Sequence based Malware Detectors. Aiming at attacking RNN-based malware
detection models that take the API call sequence as the input, Hu and Tan [33] propose a generative
model based black-box adversarial attack to evade such RNN-based PE malware detectors. In
particular, a generative RNN is trained based on PE malware to generate an irreverent API call
sequence that will be inserted into the original API call sequence of the input PE malware, while
a substitute RNN model is trained to fit the target RNN-based malware detector based on both
benign samples and the gradient information of malware samples from the generative RNN model.
In [34], Rosenberg et al. propose a generic end-to-end attack framework, namely GADGET,
against state-of-the-art API call sequence-based malware detectors under black-box settings by
the transferability property. GADGET is carried out in three steps: i) GADGET first trains a
surrogate model to approximate the decision boundaries of the target malware detector by using the
Jacobian-based dataset augmentation method [139]. ii) it then performs a white-box attack on the
surrogate model to generate the adversarial API call sequence by restricting the insertion of API
calls into the original API call sequence. In more detail, GADGET first randomly selects an insert
position and then uses a heuristic searching approach to iteratively find and insert the API calls
such that the generated adversarial sequence follows the direction indicated by the Jacobian. iii) to
generate practical adversarial malware samples from the adversarial API call sequence, GADGET
uses a proxy wrapper script to wrap the original malware by calling the additional APIs with valid
parameters in the corresponding position based on the generated adversarial API call sequence.
Fadadu et al. [35] propose an executable level evasion (ELE) attack under black-box settings
to evade PE malware detectors based on the API call sequence. The manipulation of ELE is
restricted only to the addition of new API calls, which are chosen by maximizing the fraction of
sub-sequences that have the added API call in the domain of benign samples and minimizing the
fraction of sub-sequences that have the added API call in the domain of malware samples. To
further make the modified PE malware can be executed properly, ELE uses a novel IAT (i.e.,
Import Address Table) hooking method to redirect the control in the adversarial code that is
attached to the PE malware. In particular, the adversarial code contains a wrapper function that
not only has identical arguments and returns values with the original API function, but also invokes
23
Accepted by ELSEVIER Computers
4.2 Black-box
& Security.
Adversarial Attacks against PE Malware Detection
the anti-virus engines while do not break the format and functionality. Particularly, with only
knowledge of the binary detection output, they propose a completely black-box adversarial attack
based on reinforcement learning (RL), namely gym-malware. The gym-malware first defines 10
kinds of format-preserving and functionality-preserving modifications for Windows PE files as the
action space available to the agent within the environment. Then, for any given PE malware,
gym-malware tries to learn which sequences of modifications in the action space can be used to
modify the PE malware, such that the resulting PE malware is most likely to bypass the static
anti-virus engines. Although, gym-malware has demonstrated its effectiveness against PE malware
detectors, its experimental results also show that RL with an agent of deep Q-network (DQN) or
actor-critic with experience replay (ACER) [141] offers limited improvement compared with the
random policy.
On the basis of gym-malware, there are multiple follow-up work [40, 41, 42, 43, 44, 45, 46]
proposing problem-space black-box adversarial attacks against static PE malware detection models.
In particular, Wu et al. [40] propose gym-plus based on gym-malware with the improvement
of adding more format-preserving modifications in the action space and their experimental results
show that gym-plus with DQN obtains a higher evasion rate than gym-plus with the random
policy. Differently, Chen et al. [41] propose gym-malware-mini based on gym-malware with a
limited and smaller action space. Based on the observation of most of the format-preserving
modifications of gym-malware and gym-plus are stochastic in nature (e.g., the appending bytes
to the new section are chosen at random for simplicity, etc.) and those modifications are not
exactly repeatable, gym-malware-mini makes 6 kinds of random format-preserving modifications
to deterministic modifications, making the RL algorithms easier to learn better policies among
limited action space.
Besides that, Fang et al. [42] present a general framework using DQN to evade PE malware
detectors, namely DQEAF, which is almost identical to gym-malware in methodology except for
three implementation improvements as follows. 1) DQEAF uses a subset of modifications employed
in gym-malware and guarantees that all of them would not lead to corruption in the modified mal-
ware; 2) DQEAF uses a vector with 513 dimensions as the observed state, which is much lower than
that in gym-malware; 3) DQEAF makes priority into consideration during the replay of past transi-
tions. Fang et al. [43] also observe that the modifications in the action space of gym-malware have
some randomness and further found that most effective adversarial malware from gym-malware are
generated by UPX pack/unpacked modifications, which could lead to some training problems with
RL due to the non-repeatability of those modifications. Thus, they first reduce the action space
to 6 categories having certain deterministic parameters and then propose an improved black-box
adversarial attack, namely RLAttackNet, based on the gym-malware implementation.
Ebrahimi et al. [44] suggest that the RL-based adversarial attacks against PE malware de-
tectors normally employ actor-critic or DQN, which are limited in handling environments with
combinatorially large state space. Naturally, they propose an improved RL-based adversarial at-
tack framework of AMG-VAC on the basis of gym-malware [39, 137] by adopting the variational
actor-critic, which has been demonstrated to be the state-of-the-art performance in handling en-
vironments with combinatorially large state space. As previous RL-based adversarial attacks tend
to generate homogeneous and long sequences of transformations, Labaca-Castro et al. [45] thus
present an RL-based adversarial attack framework of AIMED-RL as well. The main difference
between AIMED-RL and other RL-based adversarial attacks is that AIMED-RL introduces a
novel penalization to the reward function for increasing the diversity of the generated sequences of
25
Accepted by ELSEVIER Computers
4.2 Black-box
& Security.
Adversarial Attacks against PE Malware Detection
transformations while minimizing the corresponding lengths. Li and Li [46] suggest that existing
RL-based adversarial attacks [39, 137, 42] employ the artificially defined instant reward function
and environment, which are highly subjective and empirical, potentially leading to non-converge of
the RL algorithm. Therefore, in order to address the issue of the subjective and empirical reward
function, they present an Inverse RL-based adversarial malware generation method, namely AMG-
IRL, which could autonomously generate the flexible reward function according to the current
optimal strategy.
In short, compared to [40, 41, 44, 45], the adversarial malware samples generated by [42, 43, 46]
are verified not only for executability within the Cuckoo sandbox [142], but also verified for the
original maliciousness via comparing the function call graph between the before and after malware
samples with IDA Pro [143].
Randomization based attacks. To fully automatize the process of generating adversarial
malware without corrupting the malware functionality under the black-box setting, Castro et al. [47]
propose ARMED – automatic random malware modifications to evade detection. ARMED first
generates the adversarial PE malware by randomly applying manipulations among 9 kinds of
format-preserving modifications from [39, 137], and then employs the Cuckoo sandbox to test the
functionality of the generated adversarial malware samples. In case the functionality test fails, the
above steps would be re-start with a new round until the functionality test successes.
Ceschin et al. [48] find that packing the original PE malware with a distinct packer [144, 145]
or embedding the PE malware in a dropper [146, 147] is an effective approach in bypassing ML-
based malware detectors when combined with appending goodware strings to malware. However,
some of the generated adversarial PE malware suffer from either not being executed properly or
being too large in size. To solve the challenges, the authors implemented a lightweight dropper,
namely Dropper, which first creates an entire new PE file to host the original PE malware and
then randomly chooses goodware strings to be appended at the end of the newly created PE file.
Similar to the white-box attack of Enhanced-BFA, Chen et al. [19] also introduce another black-
box version of adversarial attack against MalConv. First, it continuously selects data blocks at
random from goodware and appends them to PE malware to generate adversarial PE malware.
After performing multiple random attacks as above, it then calculates the contribution degree of
each data block based on the experience of successful trajectories of data blocks. Finally, it appends
the data blocks to the end of PE malware according to the order of their contribution degrees.
In [49], Song et al. propose an automatic black-box attack framework that applies a sequence
of actions to rewrite PE malware for evading PE malware detectors. In particular, to generate
adversarial malware samples with a minimal set of required actions from macro/micro actions,
the authors employ an action sequence minimizer that consists of three steps. First, it randomly
selects macro-actions according to the previously updated weights of actions as the action sequence
to rewrite the original malware. Second, it tries to remove some unnecessary macro-actions from the
action sequence to generate a minimized adversarial malware, and increases the weights of effective
actions for updating. Finally, for every macro-action in the minimized adversarial malware, it
attempts to replace the macro-action with a corresponding micro-action. Besides that, the proposed
framework can also help explain which features are responsible for evasion as every required action
in adversarial malware samples corresponds to a type of affected feature.
Evolutionary algorithm based attacks. Following the black-box adversarial attack frame-
work of ARMED [47], Castro et al. [50] propose AIMED, which employs a genetic programming
approach rather than randomization methods to automatically find optimized modifications for
26
Accepted by ELSEVIER Computers
4.2 Black-box
& Security.
Adversarial Attacks against PE Malware Detection
27
Accepted by ELSEVIER Computers
4.3 Summary
& Security.
of Adversarial Attacks against PE Malware Detection
crimination model in GAN. First, the API call list (i.e., DLL and system functions) of the original
malware sample is extracted as a binary feature vector by PE Parser; Second, the Generator takes
both the malware feature vector and a sampled 3-dimensional Gaussian noise as input to produce a
3-dimensional perturbation path, indicating whether each of the three perturbation methods (i.e.,
Obfusmal, Stealmal, and Hollowmal) is adopted. Third, following the produced perturbation path,
the PE editor generates the adversarial malware sample with corresponding perturbation methods.
Finally, the Generator will stop training until the generated adversarial PE malware fails to be
recognized by Discriminator.
Heuristic based attacks. Aiming to quantify the robustness of PE malware detectors ranging
from two ML-based models to four commercial anti-virus engines, Fleshman et al. [55] propose one
novel targeted occlusion black-box attack for comparing with three pre-existing evasion techniques,
i.e., random-based gym-malware [39, 137], obfuscation through packing [145, 148], and malicious
ROP injection [149]. For the proposed targeted occlusion attack, it first uses the occlusion binary
search method to identify the most important byte region according to the changes of the malicious
probability for a given PE malware detector, and then replaces the identified region with completely
random bytes or a contiguous byte region selected randomly from benign samples. However, we
believe that adversarial malware samples generated by the targeted occlusion attack are destructive
because the replacement could prevent the generated malware from being executed, not to mention
maintain the original maliciousness.
Based on the two kinds of functionality-preserving transformations (i.e., in-place randomization
and code displacement) that manipulate the instructions of binaries in a fine-grained function level,
Lucas et al. [29] also propose another black-box version of adversarial attack based on a general
hill-climbing algorithm. This black-box attack is basically similar to the white-box version and the
only difference is how the attempted transformation is selected. Specifically, the black-box attack
first queries the model after attempting to apply one of the transformations to the PE malware,
and then accepts the transformation only if the corresponding benign probability increases.
Compared with the white-box attacks, the black-box attacks are more realistic and practical in
the wild due to their minimal reliance on knowledge about the target malware detector. For the
feature-space black-box attacks, as they are actually performed in the feature space of PE files,
existing adversarial attack methods generally devise corresponding feasible transformations (e.g.,
add irrelevant API calls) for PE malware detectors with different feature spaces (e.g., API call list
based malware detectors), indicating the black-box attacks are normally malware detector specific.
However, for the problem-space black-box attacks with the most strict requirements due to their
manipulations in the problem-space, most of them are malware detector agnostic, meaning that
these adversarial attack methods can be used to attack any kind of PE malware detectors in theory.
In terms of property preservation (i.e., format, executability, and maliciousness), for all kinds of
adversarial attacks against PE malware detection, it is also observed from Table 2 that most of them
can only guarantee the format-preserving rather than executability-preserving and maliciousness-
preserving. In particular, several adversarial attack methods (e.g., ATMPA [22], COPYCAT [23]
and the target occlusion attack [55]) might destroy the fixed layout and grammar of the PE format
that is necessary to load and execute the PE file. On the other hand, for those adversarial attacks
like [29, 42, 43, 46, 47, 48, 49, 50], they are verified not only for the executability, but also verified
experimentally whether the generated adversarial PE malware keeps the same maliciousness as the
original PE malware, which is strongly recommended and advocated in our opinion.
As various adversarial attacks continue to be proposed and evaluated, adversarial defense meth-
ods are meanwhile proposed accordingly. In fact, the rapid development of adversarial attacks and
counterpart defenses constitutes a constant arms race, meaning a new adversarial attack can easily
inspire the defender to devise a novel corresponding defense, and a newly proposed defense method
will inevitably lead the attacker to design a new adversarial attack for profit. Therefore, it is im-
portant to explore the most promising advances of adversarial defenses for Windows PE malware
detection against adversarial attacks. Although there are currently few researchers that specifi-
cally and exclusively propose adversarial defenses for Windows PE malware detection, most of the
aforementioned research efforts on adversarial attacks might more or less present corresponding
defense methods. In this section, we summarize the state-of-the-art adversarial defenses for PE
malware detection in recent years, mainly including adversarial training and several other defenses.
API calls generated by the adversarial attack of MalGAN [30]. Anderson et al. [137] first exploit
the gym-malware to generate adversarial PE malware, and then re-train the malware detection
model of EMBER based on the original PE files and the generated adversarial PE malware. Dif-
ferently, in addition to adversarial training with adversarial API calls generated by EvnAttack,
Chen et al. [32] also present a new secure-learning framework for PE malware detection, namely
SecDefender, which adds a security regularization term by considering the evasion cost of feature
manipulations by attackers.
To sum up, for those adversarial defenses based on adversarial training, it is generally observed
that, 1) adversarial training is experimentally demonstrated to mitigate one or several adversarial
attacks to some extent; 2) adversarial training inevitably introduces significant additional costs in
generating adversarial examples during the training process.
6. Discussions
The previous sections of § 4 and § 5 enable interested readers to have a better and faster under-
standing with regard to the adversarial attacks and defenses for Windows PE malware detection.
In the following subsections, we first present the other related attacks against PE malware detec-
tion beyond the aforementioned adversarial attacks in § 6.1 and then shed some light on research
directions as well as opportunities for future work in § 6.2.
30
Accepted by ELSEVIER Computers & Security. 6.1 Beyond Adversarial Attacks
31
Accepted by ELSEVIER Computers & Security. 6.2 Future Directions and Opportunities
in a problem space, Shapira et al. propose a novel instance poisoning attack by first selecting
the goodware that is most similar to the target malware instance and then adding sections to
the goodware for adversarially training the poisoned model. Actually, the manipulation of adding
sections acts as the backdoor trigger, which can remain the functionality of the associated goodware
as well as the malware instance. During the testing phase, the target malware instance associated
with the backdoor trigger will be misclassified as benign by the poisoned model.
6.2.2. Practical and Efficient Adversarial Attacks against Commercial Anti-viruses in the Wild
As introduced and summarized in § 4 and Table 2, the current adversarial attack methods
against PE malware detection have been devoted to developing problem-space black-box adversarial
attacks, which usually take a similar and typical attack procedure. In general, the attack procedure
first defines a set of available transformations (e.g., inject adversarial payload, insert the semantic
N OP s, etc.) in the problem-space, and then employs a variety of search strategies (e.g., gradient-
based, reinforcement learning, genetic algorithm, etc.) to choose a sequence of transformations
which can be applied to the original PE malware for generating the adversarial PE malware. Based
on the above observation, we argue there is still much room for improving both the effectiveness
and efficiency of adversarial attacks against PE malware detection in two aspects: i) devising
32
Accepted by ELSEVIER Computers & Security.
and defining more practical and stealthier transformations for PE malware. For example, instead
of simply inserting the N OP s in the blocks of CFGs [38] that is easily noticed and removed by
defenders, the transformation of splitting one block of CFGs into multiple iteratively called blocks
is much stealthier to be noticed and removed. ii) designing and implementing more efficient search
strategies to accelerate the generation of adversarial PE malware. We argue that it is quite time-
consuming for both RL and genetic algorithm based search strategies.
In addition, it is clearly observed that most existing adversarial attacks target PE malware
detection based on static analysis rather than dynamic analysis, which is particularly unknown for
both attackers and defenders. However, the mainstream commercial anti-virus software/service
used by end users of laptops and servers normally employs a hybrid defense solution with both
static analysis and dynamic analysis. Therefore, it is extremely important and urgently demanded
to devise and implement practical and efficient adversarial attacks against PE malware detection
based on dynamic analysis and commercial anti-viruses.
7. Conclusion
Acknowledgement
This project is supported by the Strategic Priority Research Program of the Chinese Academy
of Sciences under Grant No.XDA0320000, the National Natural Science Foundation of China under
No.62202457 and No.U1936215, and the project funded by China Postdoctoral Science Foundation
under No.2022M713253. Yaguan Qian is also supported by the Key Program of Zhejiang Provincial
Natural Science Foundation of China under No.LZ22F020007.
References
[1] D. Ucci, L. Aniello, R. Baldoni, Survey of machine learning techniques for malware analysis, Computers &
Security 81 (2019) 123–147.
[2] E. Raff, C. Nicholas, A survey of machine learning methods and challenges for windows malware classification,
2020. ArXiv preprint arXiv:2006.09271.
[3] M. G. Schultz, E. Eskin, F. Zadok, S. J. Stolfo, Data mining methods for detection of new malicious executables,
in: IEEE Symposium on Security and Privacy, IEEE, Oakland, California, USA, 2001, pp. 38–49.
[4] Kaspersky Lab, The number of new malicious files detected every day increases
by 5.2% to 360,000 in 2020, https://www.kaspersky.com/about/press-releases/2020_
the-number-of-new-malicious-files-detected-every-day-increases-by-52-to-360000-in-2020, 2020.
Online (last accessed October 1, 2021).
[5] Kaspersky Lab, New malicious files discovered daily grow by 5.7% to
380,000 in 2021, https://www.kaspersky.com/about/press-releases/2021_
new-malicious-files-discovered-daily-grow-by-57-to-380000-in-2021, 2020. Online (last accessed
September 14, 2022).
[6] N. Idika, A. P. Mathur, A survey of malware detection techniques, Purdue University 48 (2007) 32–46.
[7] Y. Ye, T. Li, D. Adjeroh, S. S. Iyengar, A survey on malware detection using data mining techniques, ACM
Computing Surveys 50 (2017) 1–40.
[8] F. Ceschin, H. M. Gomes, M. Botacin, A. Bifet, B. Pfahringer, L. S. Oliveira, A. Grégio, Machine learning (in)
security: A stream of problems, 2020. ArXiv preprint arXiv:2010.16045.
[9] I. J. Goodfellow, J. Shlens, C. Szegedy, Explaining and harnessing adversarial examples, in: International
Conference on Learning Representations, OpenReview.net, San Diego, CA, USA, 2015, pp. 1–11.
[10] N. Carlini, D. Wagner, Towards evaluating the robustness of neural networks, in: IEEE Symposium on Security
and Privacy, IEEE, San Jose, CA, USA, 2017, pp. 39–57.
[11] N. Akhtar, A. Mian, Threat of adversarial attacks on deep learning in computer vision: A survey, IEEE Access
6 (2018) 14410–14430.
[12] A. Chakraborty, M. Alam, V. Dey, A. Chattopadhyay, D. Mukhopadhyay, Adversarial attacks and defenses:
A survey, arXiv preprint arXiv:1810.00069 (2018).
[13] J. Zhang, C. Li, Adversarial examples: Opportunities and challenges, IEEE Transactions on Neural Networks
and Learning Systems 31 (2019) 2578–2593.
[14] A. Serban, E. Poll, J. Visser, Adversarial examples on object recognition: A comprehensive survey, ACM
Computing Surveys (CSUR) 53 (2020) 1–38.
[15] G. R. Machado, E. Silva, R. R. Goldschmidt, Adversarial machine learning in image classification: A survey
toward the defender’s perspective, ACM Computing Surveys (CSUR) 55 (2021) 1–38.
[16] T. Long, Q. Gao, L. Xu, Z. Zhou, A survey on adversarial attacks in computer vision: Taxonomy, visualization
and future directions, Computers & Security 121 (2022) 102847.
[17] F. Kreuk, A. Barak, S. Aviv-Reuven, M. Baruch, B. Pinkas, J. Keshet, Deceiving end-to-end deep learning
malware detectors using adversarial examples, 2018. ArXiv preprint arXiv:1802.04528.
[18] O. Suciu, S. E. Coull, J. Johns, Exploring adversarial examples in malware detection, in: IEEE Security and
Privacy Workshops, IEEE, San Francisco, CA, USA, 2019, pp. 8–14.
[19] B. Chen, Z. Ren, C. Yu, I. Hussain, J. Liu, Adversarial examples for CNN-based malware detectors, IEEE
Access 7 (2019) 54360–54371.
34
Accepted by ELSEVIER Computers & Security. REFERENCES
[20] Y. Qiao, W. Zhang, Z. Tian, L. T. Yang, Y. Liu, M. Alazab, Adversarial malware sample generation method
based on the prototype of deep learning detector, Computers & Security (2022) 102762.
[21] A. Al-Dujaili, A. Huang, E. Hemberg, U.-M. O’Reilly, Adversarial deep learning for robust detection of binary
encoded malware, in: IEEE Security and Privacy Workshops, IEEE, San Francisco, CA, USA, 2018, pp. 76–82.
[22] X. Liu, J. Zhang, Y. Lin, H. Li, ATMPA: Attacking machine learning-based malware visualization detection
methods via adversarial examples, in: IEEE/ACM International Symposium on Quality of Service, IEEE,
Phoenix, AZ, USA, 2019, pp. 1–10.
[23] A. Khormali, A. Abusnaina, S. Chen, D. Nyang, A. Mohaisen, COPYCAT: practical adversarial attacks on
visualization-based malware detection, 2019. ArXiv preprint arXiv:1909.09735.
[24] D. Park, H. Khan, B. Yener, Generation & evaluation of adversarial examples for malware obfuscation, in:
International Conference on Machine Learning and Applications, IEEE, Boca Raton, FL, USA, 2019, pp.
1283–1290.
[25] X. Li, K. Qiu, C. Qian, G. Zhao, An adversarial machine learning method based on opcode n-grams feature
in malware detection, in: International Conference on Data Science in Cyberspace, IEEE, Hong Kong, China,
2020, pp. 380–387.
[26] B. Kolosnjaji, A. Demontis, B. Biggio, D. Maiorca, G. Giacinto, C. Eckert, F. Roli, Adversarial malware bina-
ries: Evading deep learning for malware detection in executables, in: European Signal Processing Conference,
IEEE, Roma, Italy, 2018, pp. 533–537.
[27] L. Demetrio, B. Biggio, G. Lagorio, F. Roli, A. Armando, Explaining vulnerabilities of deep learning to
adversarial malware binaries, 2019. ArXiv preprint arXiv:1901.03583.
[28] L. Demetrio, S. E. Coull, B. Biggio, G. Lagorio, A. Armando, F. Roli, Adversarial EXEmples: A survey and
experimental evaluation of practical attacks on machine learning for windows malware detection, 2020. ArXiv
preprint arXiv:2008.07125.
[29] K. Lucas, M. Sharif, L. Bauer, M. K. Reiter, S. Shintre, Malware makeover: Breaking ml-based static analysis
by modifying executable bytes, in: ASIA Conference on Computer and Communications Security, ACM,
Hongkong, China, 2021, pp. 744–758.
[30] W. Hu, Y. Tan, Generating adversarial malware examples for black-box attacks based on GAN, 2017. ArXiv
preprint arXiv:1702.05983.
[31] M. Kawai, K. Ota, M. Dong, Improved MalGAN: Avoiding malware detector by leaning cleanware features,
in: International Conference on Artificial Intelligence in Information and Communication, IEEE, Okinawa,
Japan, 2019, pp. 40–45.
[32] L. Chen, Y. Ye, T. Bourlai, Adversarial machine learning in malware detection: Arms race between evasion
attack and defense, in: European Intelligence and Security Informatics Conference, IEEE, Athens, Greece,
2017, pp. 99–106.
[33] W. Hu, Y. Tan, Black-box attacks against RNN based malware detection algorithms, 2017. ArXiv preprint
arXiv:1705.08131.
[34] I. Rosenberg, A. Shabtai, L. Rokach, Y. Elovici, Generic black-box end-to-end attack against state of the
art API call based malware classifiers, in: International Symposium on Research in Attacks, Intrusions, and
Defenses, Springer, Heraklion, Crete, Greece, 2018, pp. 490–510.
[35] F. Fadadu, A. Handa, N. Kumar, S. K. Shukla, Evading API call sequence based malware classifiers, in:
International Conference on Information and Communications Security, Springer, Beijing, China, 2019, pp.
18–33.
[36] I. Rosenberg, A. Shabtai, Y. Elovici, L. Rokach, Query-efficient black-box attack against sequence-based
malware classifiers, in: Annual Computer Security Applications Conference, ACM, Virtual Event, 2020, p.
611–626.
[37] I. Rosenberg, S. Meir, J. Berrebi, I. Gordon, G. Sicard, E. O. David, Generating end-to-end adversarial
examples for malware classifiers using explainability, in: International Joint Conference on Neural Networks,
IEEE, Glasgow, United Kingdom, 2020, pp. 1–10.
[38] L. Zhang, P. Liu, Y.-H. Choi, Semantic-preserving reinforcement learning attack against graph neural networks
for malware detection, 2020. ArXiv preprint arXiv:2009.05602.
[39] H. S. Anderson, A. Kharkar, B. Filar, P. Roth, Evading machine learning malware detection, in: Black Hat
USA, blackhat.com, Las Vegas, NV, USA, 2017, pp. 1–6.
[40] C. Wu, J. Shi, Y. Yang, W. Li, Enhancing machine learning based malware detection model by reinforcement
learning, in: International Conference on Communication and Network Security, ACM, Qingdao, China, 2018,
pp. 74–78.
[41] J. Chen, J. Jiang, R. Li, Y. Dou, Generating adversarial examples for static PE malware detector based on
35
Accepted by ELSEVIER Computers & Security. REFERENCES
deep reinforcement learning, Journal of Physics: Conference Series 1575 (2020) 012011.
[42] Z. Fang, J. Wang, B. Li, S. Wu, Y. Zhou, H. Huang, Evading anti-malware engines with deep reinforcement
learning, IEEE Access 7 (2019) 48867–48879.
[43] Y. Fang, Y. Zeng, B. Li, L. Liu, L. Zhang, DeepDetectNet vs RLAttackNet: An adversarial method to improve
deep learning-based static malware detection model, PLoS One 15 (2020) e0231626.
[44] M. Ebrahimi, J. Pacheco, W. Li, J. L. Hu, H. Chen, Binary black-box attacks against static malware detectors
with reinforcement learning in discrete action spaces, in: IEEE Security and Privacy Workshops, IEEE, Virtual
Event, 2021, pp. 85–91.
[45] R. Labaca-Castro, S. Franz, G. D. Rodosek, AIMED-RL: Exploring adversarial malware examples with re-
inforcement learning, in: Joint European Conference on Machine Learning and Knowledge Discovery in
Databases, Springer, Bilbao, Spain, 2021, pp. 37–52.
[46] X. Li, Q. Li, An irl-based malware adversarial generation method to evade anti-malware engines, Computers
& Security 104 (2021) 102118.
[47] R. L. Castro, C. Schmitt, G. D. Rodosek, ARMED: How automatic malware modifications can evade static
detection?, in: International Conference on Information Management, IEEE, Cambridge, United Kingdom,
2019, pp. 20–27.
[48] F. Ceschin, M. Botacin, H. M. Gomes, L. S. Oliveira, A. Grégio, Shallow security: On the creation of
adversarial variants to evade machine learning-based malware detectors, in: Reversing and Offensive-Oriented
Trends Symposium, ACM, Vienna, Austria, 2019, pp. 1–9.
[49] W. Song, X. Li, S. Afroz, D. Garg, D. Kuznetsov, H. Yin, Automatic generation of adversarial examples for
interpreting malware classifiers, 2020. ArXiv preprint arXiv:2003.03100.
[50] R. L. Castro, C. Schmitt, G. Dreo, AIMED: Evolving malware with genetic programming to evade detection,
in: International Conference on Trust, Security and Privacy in Computing and Communications / International
Conference on Big Data Science and Engineering, IEEE, Rotorua, New Zealand, 2019, pp. 240–247.
[51] X. Wang, R. Miikkulainen, MDEA: Malware detection with evolutionary adversarial learning, 2020. ArXiv
preprint arXiv:2002.03331.
[52] L. Demetrio, B. Biggio, G. Lagorio, F. Roli, A. Armando, Functionality-preserving black-box optimization of
adversarial windows malware, 2020. ArXiv preprint arXiv:2003.13526.
[53] J. Yuan, S. Zhou, L. Lin, F. Wang, J. Cui, Black-box adversarial attacks against deep learning based malware
binaries detection with GAN, in: European Conference on Artificial Intelligence, volume 325, IOS Press,
Santiago de Compostela, Spain, 2020, pp. 2536–2542.
[54] F. Zhong, X. Cheng, D. Yu, B. Gong, S. Song, J. Yu, MalFox: Camouflaged adversarial malware example
generation based on C-GANs against black-box detectors, 2020. ArXiv preprint arXiv:2011.01509.
[55] W. Fleshman, E. Raff, R. Zak, M. McLean, C. Nicholas, Static malware detection & subterfuge: Quantifying
the robustness of machine learning and current anti-virus, in: International Conference on Malicious and
Unwanted Software, IEEE, Nantucket, MA, USA, 2018, pp. 1–10.
[56] B. Alshemali, J. Kalita, Improving the reliability of deep neural networks in nlp: A review, Knowledge-Based
Systems 191 (2020) 105210.
[57] J. Lan, R. Zhang, Z. Yan, J. Wang, Y. Chen, R. Hou, Adversarial attacks and defenses in speaker recognition
systems: A survey, Journal of Systems Architecture 127 (2022) 102526.
[58] G. Zeng, F. Qi, Q. Zhou, T. Zhang, Z. Ma, B. Hou, Y. Zang, Z. Liu, M. Sun, OpenAttack: An open-source
textual adversarial attack toolkit, 2020. ArXiv preprint arXiv:2009.09191.
[59] L. Sun, Y. Dou, C. Yang, J. Wang, P. S. Yu, L. He, B. Li, Adversarial attack and defense on graph data: A
survey, 2018. ArXiv preprint arXiv:1812.10528.
[60] F. Pierazzi, F. Pendlebury, J. Cortellazzi, L. Cavallaro, Intriguing properties of adversarial ml attacks in the
problem space, in: IEEE Symposium on Security and Privacy, IEEE, Virtual Event, 2020, pp. 1332–1349.
[61] D. Park, B. Yener, A survey on practical adversarial examples for malware classifiers, in: Reversing and
Offensive-oriented Trends Symposium, ACM, Vienna, Austria, 2020, pp. 23–35.
[62] D. Li, Q. Li, Y. Ye, S. Xu, Arms race in adversarial malware detection: A survey, ACM Computing Surveys
(CSUR) 55 (2021) 1–35.
[63] Microsoft, Inc., PE format, https://docs.microsoft.com/en-us/windows/win32/debug/pe-format, 2020.
Online (last accessed October 22, 2020).
[64] M. Pietrek, Inside Windows: An in-depth look into the Win32 portable executable file for-
mat, MSDN Magazine: https://docs.microsoft.com/en-us/archive/msdn-magazine/2002/february/
inside-windows-win32-portable-executable-file-format-in-detail, 2020. Online (last accessed Octo-
ber 22, 2020).
36
Accepted by ELSEVIER Computers & Security. REFERENCES
37
Accepted by ELSEVIER Computers & Security. REFERENCES
tional neural networks, in: AAAI Conference on Artificial Intelligence, AAAI Press, New Orleans, Louisiana,
USA, 2018, pp. 7759–7764.
[91] E. Raff, J. Barker, J. Sylvester, R. Brandon, B. Catanzaro, C. Nicholas, Malware detection by eating a whole
EXE, 2017. ArXiv preprint arXiv:1710.09435.
[92] H. S. Anderson, P. Roth, EMBER: an open dataset for training static PE malware machine learning models,
2018. ArXiv preprint arXiv:1804.04637.
[93] T. K. Ho, Random decision forests, in: International Conference on Document Analysis and Recognition,
volume 1, IEEE, Montreal, Canada, 1995, pp. 278–282.
[94] Y. Ye, L. Chen, D. Wang, T. Li, Q. Jiang, M. Zhao, Sbmds: An interpretable string based malware detection
system using svm ensemble with bagging, Journal in Computer Virology 5 (2009) 283–293.
[95] R. Islam, R. Tian, L. Batten, S. Versteeg, Classification of malware based on string and function feature
selection, in: Cybercrime and Trustworthy Computing Workshop, IEEE, Ballarat, Australia, 2010, pp. 9–17.
[96] M. Z. Shafiq, S. M. Tabish, F. Mirza, M. Farooq, PE-Miner: Mining structural information to detect malicious
executables in realtime, in: International Workshop on Recent Advances in Intrusion Detection, Springer,
Saint-Malo, France, 2009, pp. 121–141.
[97] N. S. Altman, An introduction to kernel and nearest-neighbor nonparametric regression, The American
Statistician 46 (1992) 175–185.
[98] L. Nataraj, V. Yegneswaran, P. Porras, J. Zhang, A comparative assessment of malware classification using
binary texture analysis and dynamic analysis, in: ACM Workshop on Security and Artificial Intelligence,
ACM, Chicago, IL, USA, 2011, pp. 21–30.
[99] H.-J. Kim, Image-based malware classification using convolutional neural network, in: Advances in Computer
Science and Ubiquitous Computing, Springer, Taichung, Taiwan, China, 2017, pp. 1352–1357.
[100] X. Liu, Y. Lin, H. Li, J. Zhang, A novel method for malware detection on ml-based visualization technique,
Computers & Security 89 (2020) 101682.
[101] K. Rieck, T. Holz, C. Willems, P. Düssel, P. Laskov, Learning and classification of malware behavior, in:
International Conference on Detection of Intrusions and Malware, and Vulnerability Assessment, Springer,
Paris, France, 2008, pp. 108–125.
[102] A. Mohaisen, O. Alrawi, M. Mohaisen, AMAL: high-fidelity, behavior-based automated malware analysis and
classification, Computers & Security 52 (2015) 251–266.
[103] M. Abdelsalam, R. Krishnan, Y. Huang, R. Sandhu, Malware detection in cloud infrastructures using convo-
lutional neural networks, in: International Conference on Cloud Computing, IEEE, San Francisco, CA, USA,
2018, pp. 162–169.
[104] J. Franklin, The elements of statistical learning: data mining, inference and prediction, The Mathematical
Intelligencer 27 (2005) 83–85.
[105] T. N. Kipf, M. Welling, Semi-supervised classification with graph convolutional networks, in: International
Conference on Learning Representations, OpenReview.net, Toulon, France, 2017, pp. 1–14.
[106] M. Bailey, J. Oberheide, J. Andersen, Z. M. Mao, F. Jahanian, J. Nazario, Automated classification and
analysis of internet malware, in: International Workshop on Recent Advances in Intrusion Detection, Springer,
Gold Goast, Australia, 2007, pp. 178–197.
[107] S. Wang, Z. Chen, X. Yu, D. Li, J. Ni, L. Tang, J. Gui, Z. Li, H. Chen, P. S. Yu, Heterogeneous graph matching
networks for unknown malware detection, in: International Joint Conference on Artificial Intelligence, ijcai.org,
Macao, China, 2019, pp. 3762–3770.
[108] K. Cho, B. Van Merriënboer, C. Gulcehre, D. Bahdanau, F. Bougares, H. Schwenk, Y. Bengio, Learning
phrase representations using rnn encoder–decoder for statistical machine translation, in: Empirical Methods
in Natural Language Processing, ACL, Doha, Qatar, 2014, pp. 1724–1734.
[109] Y. Ye, T. Li, Y. Chen, Q. Jiang, Automatic malware categorization using cluster ensemble, in: ACM SIGKDD
Conference on Knowledge Discovery and Data Mining, ACM, Washington, DC, USA, 2010, pp. 95–104.
[110] I. Santos, F. Brezo, X. Ugarte-Pedrero, P. G. Bringas, Opcode sequences as representation of executables for
data-mining-based unknown malware detection, Information Sciences 231 (2013) 64–82.
[111] J. Zhang, Z. Qin, H. Yin, L. Ou, Y. Hu, Irmd: malware variant detection using opcode image recognition, in:
International Conference on Parallel and Distributed Systems, IEEE, Wuhan, China, 2016, pp. 1175–1180.
[112] G. Sun, Q. Qian, Deep learning and visualization for identifying malware families, IEEE Transactions on
Dependable and Secure Computing 18 (2018) 283–295.
[113] W. W. Cohen, Learning trees and rules with set-valued features, in: AAAI/IAAI, AAAI Press, Portland,
Oregon, USA, 1996, pp. 709–716.
[114] S. Hochreiter, J. Schmidhuber, Long short-term memory, Neural Computation 9 (1997) 1735–1780.
38
Accepted by ELSEVIER Computers & Security. REFERENCES
[115] Y. Qiao, Y. Yang, L. Ji, J. He, Analyzing malware by abstracting the frequent itemsets in api call sequences,
in: IEEE International Conference on Trust, Security and Privacy in Computing and Communications, IEEE,
Melbourne, Australia, 2013, pp. 265–270.
[116] Z. Zhang, P. Qi, W. Wang, Dynamic malware analysis with feature engineering and feature learning, in: AAAI
Conference on Artificial Intelligence, AAAI Press, New York, NY, USA, 2020, pp. 1210–1217.
[117] A. Kapoor, S. Dhavale, Control flow graph based multiclass malware detection using bi-normal separation,
Defence Science Journal 66 (2016) 138–145.
[118] J. Yan, G. Yan, D. Jin, Classifying malware represented as control flow graphs using deep graph convolu-
tional neural network, in: IEEE/IFIP International Conference on Dependable Systems and Networks, IEEE,
Portland, OR, USA, 2019, pp. 52–63.
[119] X. Ling, L. Wu, W. Deng, Z. Qu, J. Zhang, S. Zhang, T. Ma, B. Wang, C. Wu, S. Ji, MalGraph: Hier-
archical graph neural networks for robust Windows malware detection, in: IEEE Conference on Computer
Communications, IEEE, Virtual Event, 2022, pp. 1998–2007.
[120] P. Vincent, H. Larochelle, I. Lajoie, Y. Bengio, P.-A. Manzagol, L. Bottou, Stacked denoising autoencoders:
Learning useful representations in a deep network with a local denoising criterion, Journal of Machine Learning
Research 11 (2010) 3371–3408.
[121] M. Hassen, P. K. Chan, Scalable function call graph-based malware classification, in: ACM on Conference on
Data and Application Security and Privacy, ACM, Scottsdale, AZ, USA, 2017, pp. 239–248.
[122] H. Jiang, T. Turki, J. T. Wang, Dlgraph: Malware detection using deep learning and graph embedding,
in: International Conference on Machine Learning and Applications, IEEE, Orlando, FL, USA, 2018, pp.
1029–1033.
[123] S. Zhao, X. Ma, W. Zou, B. Bai, Deepcg: classifying metamorphic malware through deep learning of call
graphs, in: International Conference on Security and Privacy in Communication Systems, Springer, Orlando,
FL, USA, 2019, pp. 171–190.
[124] M. Krčál, O. Švec, M. Bálek, O. Jašek, Deep convolutional malware classifiers can learn from raw executables
and labels only, in: International Conference on Learning Representations – Workshop Track, OpenReview.net,
Vancouver, BC, Canada, 2018, pp. 1–4.
[125] S. E. Coull, C. Gardner, Activation analysis of a byte-based deep neural network for malware classification,
in: IEEE Security and Privacy Workshops, IEEE, San Francisco, CA, UAS, 2019, pp. 21–27.
[126] L. Nataraj, S. Karthikeyan, G. Jacob, B. S. Manjunath, Malware images: visualization and automatic classi-
fication, in: International Symposium on Visualization for Cyber Security, ACM, Pittsburgh, PA, USA, 2011,
pp. 1–7.
[127] X. Ling, L. Wu, S. Wang, T. Ma, F. Xu, A. X. Liu, C. Wu, S. Ji, Multilevel graph matching networks for deep
graph similarity learning, IEEE Transactions on Neural Networks and Learning Systems (TNNLS) (2021).
[128] B. G. Ryder, Constructing the call graph of a program, IEEE Transactions on Software Engineering 5 (1979)
216–226.
[129] X. Ling, L. Wu, S. Wang, G. Pan, T. Ma, F. Xu, A. X. Liu, C. Wu, S. Ji, Deep graph matching and searching
for semantic code retrieval, ACM Transactions on Knowledge Discovery from Data (TKDD) 15 (2021).
[130] E. Quiring, A. Maier, K. Rieck, Misleading authorship attribution of source code using adversarial learning,
in: USENIX Security Symposium, USENIX Association, Santa Clara, CA, USA, 2019, pp. 479–496.
[131] R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, D. Batra, Grad-CAM: Visual explanations
from deep networks via gradient-based localization, in: International Conference on Computer Vision, IEEE,
Venice, Italy, 2017, pp. 618–626.
[132] A. Kurakin, I. Goodfellow, S. Bengio, Adversarial examples in the physical world, in: International Conference
on Learning Representations, OpenReview.net, Toulon, France, 2017, pp. 1–14.
[133] MIT-IBM Watson AI Lab, Robust malware detection challenge, 1st Workshop on Adversarial Learning Meth-
ods for Machine Learning and Data Mining in KDD 2019 https://sites.google.com/view/advml/Home/
advml-2019/advml19-challenge, 2019. Online (last accessed October 15, 2020).
[134] S. Verwer, A. Nadeem, C. Hammerschmidt, L. Bliek, A. Al-Dujaili, U.-M. O’Reilly, The robust malware detec-
tion challenge and greedy random accelerated multi-bit search, in: ACM Workshop on Artificial Intelligence
and Security (AISec), ACM, Virtual Event, 2020, pp. 61–70.
[135] S. M. Lundberg, S.-I. Lee, A unified approach to interpreting model predictions, in: Advances in Neural
Information Processing Systems, Curran Associates, Inc., Long Beach, CA, USA, 2017, pp. 4765–4774.
[136] M. Sundararajan, A. Taly, Q. Yan, Axiomatic attribution for deep networks, in: International Conference on
Machine Learning, PMLR, Sydney, NSW, Australia, 2017, pp. 3319–3328.
[137] H. S. Anderson, A. Kharkar, B. Filar, D. Evans, P. Roth, Learning to evade static PE machine learning malware
39
Accepted by ELSEVIER Computers & Security. REFERENCES
40
Accepted by ELSEVIER Computers & Security. REFERENCES
clean-label poisoning attacks on neural networks, in: Advances in Neural Information Processing Systems,
Curran Associates, Inc., Montréal, Canada, 2018, pp. 6106–6116.
[161] A. Ali, B. Eshete, Best-effort adversarial approximation of black-box malware classifiers, in: EAI International
Conference on Security and Privacy in Communication Networks, Springer, Washington DC, USA, 2020, pp.
318–338.
[162] N. Papernot, F. Faghri, N. Carlini, I. Goodfellow, R. Feinman, A. Kurakin, C. Xie, Y. Sharma, T. Brown,
A. Roy, et al., Technical report on the cleverhans v2.1.0 adversarial examples library, 2016. ArXiv preprint
arXiv:1610.00768.
[163] X. Ling, S. Ji, J. Zou, J. Wang, C. Wu, B. Li, T. Wang, DEEPSEC: A uniform platform for security analysis
of deep learning model, in: IEEE Symposium on Security and Privacy, IEEE, San Francisco, USA, 2019, pp.
673–690.
[164] Y. Li, W. Jin, H. Xu, J. Tang, DeepRobust: A pytorch library for adversarial attacks and defenses, 2020.
ArXiv preprint arXiv:2005.06149.
41