Eforensics Magazine 2022 01 Malware Forensic Analysis
Eforensics Magazine 2022 01 Malware Forensic Analysis
Editor-in-Chief
Joanna Kretowicz
[email protected]
Managing Editor:
Michalina Szpyrka
[email protected]
Editors:
Marta Sienicka
[email protected]
Marta Strzelec
[email protected]
Bartek Adach
[email protected]
Magdalena Jarzębska
[email protected]
Senior Consultant/Publisher:
Paweł Marciniak
CEO:
Joanna Kretowicz
[email protected]
Marketing Director:
Joanna Kretowicz
[email protected]
DTP
Michalina Szpyrka
[email protected]
Cover Design
Hiep Nguyen Duc
Publisher
Hakin9 Media Sp. z o.o.
02-511 Warszawa
ul. Bielawska 6/19
Phone: 1 917 338 3631
www.eforensicsmag.com
All trademarks, trade names, or logos mentioned or used are the property of their respective owners.
The techniques described in our articles may only be used in private, local networks. The editors hold no responsibility for misuse of the
presented techniques or consequent data loss.
Word from the team
Dear Readers,
We wish you all the best in the New Year! We hope that it will be fruitful for you and
will allow you to develop your interests and skills. We sincerely want to help you
with the latter, which is why we present to you our latest magazine on Malware
Forensic Analysis. We have prepared a real compendium of knowledge on this topic
for you. From the articles, you will learn what the main forms of malware are, how
to analyze them, which type of analysis (static or dynamic) is appropriate for a
given case, and what tools to use! Therefore, if you are interested in this topic, you
must take a look at our magazine.
Did you think that was it? Well no! In the magazine you will also find:
So if you want to start the New Year productively and invest in your development,
reach for our latest magazine!
Check out our Table of Contents below for more information about each article (we
included short leads for you).
We hope that you enjoy reading this issue! As always, huge thanks to all the
authors, reviewers, to our amazing proofreaders, and of course you, our readers,
for staying with us! :)
Due to the increase in the number of connected computing devices, the distribution of malicious
programs associated with criminal practice grows daily. Consequently, the presence of malware in
forensic examinations is increasingly frequent. In addition, the high diversity of classes and different
methods of malware performance make the forensic examinations performed on these types of
programs create new challenges for specialists in computer forensics. The purpose of this article is
to present fundamental and objective concepts about malware analysis for professionals in the area,
along with basic tools and techniques that will help in the discovery of the activities performed by
malicious programs.
The Malware
The amount of digital pests spread on the internet increases every year. Every day, countless computer
programs are developed with the strict purpose of performing illicit activities on computing devices.
Our personal information and financial data has never been more valuable. Criminal organizations see
this illegal practice as an extremely profitable and low-risk activity, moving millions of dollars a year.
Consequently, in the forensic context, the need to perform malware scans increases proportionally due
to the high number of contaminated devices. However, although malicious programs are a nightmare in
the lives of users, they can be the solution in a forensic examination, since many malwares store
valuable information that would not be obtained without their presence, which allows proving the
5
In the course of this article, some techniques and tools used in malware exams will be presented that
help to understand the behavior and activities performed by the program under analysis. Our biggest
challenge in developing the article was trying to provide as much information on the topic as possible
in the short space of an article, since malware analysis and program reverse engineering are extremely
extensive and complex subjects. Because of this, throughout the article, we will focus on the analysis of
software developed for the Windows platform, as this operating system is the target of the vast majority
of current malware.
The article consists of the presentation of some essential concepts of malware analysis followed by the
demonstration of some tools used to obtain information about the functioning of a program in the
forensic context. It is important for the reader to download the aforementioned tools and test them so
that there is greater use of the concepts studied. It is also critical to download the latest versions of
tools used to minimize the risk of existing vulnerabilities. Throughout the article, most of the programs
What is a Malware?
Malware is malicious software developed with the purpose of infiltrating a computer system and
collecting information without authorization or simply causing damage. Once installed, and depending
on the complexity of the malware, these programs can have access to files present on the computer,
traffic transmitted on the network, keys pressed and even environmental eavesdropping or webcam
image capture.
The main motivation in the development of malware is to obtain financial resources, however, other
Malware uses numerous techniques to infect computing devices, making prevention and detection
• Exploitation of vulnerabilities: Many malware have exploits used to gain access through security holes
contained in outdated programs on the victim's computer;
6
• Auto-execution of external media: Functions such as Windows autorun, used by optical media and
removable devices such as USB sticks, are a great gateway for malicious software, as they run
programs without the user's consent and, as a rule, are inhabited by default;
• Social Engineering: It is the psychological manipulation employed by the attacker to persuade the
victim and lead him to execute the malware. Social engineering techniques often use means to
arouse the victim's curiosity, abuse their ambition or even their innocence;
• User Conscious Execution: In some cases, the user installs malware on purpose in order to obtain
access information about other users who share the use of the machine.
Malware Classes
There are numerous malware classifications according to their method of operation, purpose and
complexity. In an expert examination, the prior identification of the category to which the malware
belongs facilitates the search for traces and allows a more targeted analysis. Currently, due to the
evolution of malware, these classes are not mutually exclusive, that is, malicious software can belong to
more than one category. The main types are described below.
Spyware
This type of malware has as its main purpose the collection of information from users who make use of
the computer without their consent. They usually send users' personal information to external servers
on the internet. It is a class of software widely used for espionage and also by scammers to obtain
passwords to access services such as internet banking, credit card numbers, e-mail or access to
systems. Often this type of malware assists in forensic analysis, as it collects information that would not
normally be stored by operating system logs or other applications. Spyware can still be divided into a
Adware
Its main activity is the display of advertisements, but they are developed with the aim of collecting
information about the user's browsing on the web to later offer advertisements for products and
7
Keyloggers
Used to capture and store the keys that are pressed on the keyboard by the user. Most keyloggers store
the collected data in encoded or encrypted files, which makes malware detection and forensic analysis
difficult. In fact, if the entire hard drive is indexed, the files containing the words typed and stored by
the keylogger will not be detected in a keyword search. In these cases, a deeper analysis of the
Screenloggers
Usually working together with keyloggers, this type of malware captures some screens presented by the
victim's computer. Most also store the mouse position at the time of capture in the image. Many of
them start running after a certain event, such as when the user enters their internet banking page. They
are widely used to capture passwords entered on virtual keyboards. Most screenloggers do not encrypt
the generated images, which makes it easier to detect malware in forensic analysis.
Sniffers
Sniffers are used to capture all traffic transmitted by a network card. They are usually used for lawful
purposes by network administrators who make use of this type of tool to detect possible problems in
their network. However, they can be used for malicious purposes by capturing everything that passes
through the victim's network card, making it possible to obtain unencrypted passwords or even the
Backdoors
Usually used after a successful attack, the backdoor is intended to guarantee later access to the
attacker in an easy way, without the need to resort to the methods initially used for the invasion. Allows
remote execution of commands on the local system, in addition to having little or no authentication for
access. The detection and analysis of a backdoor in a forensic examination can exonerate an accused, if
it is proved that his computer was accessed remotely and used for the practice of illicit activities.
8
Worms
Its main purpose is to spread across a network by infecting different machines. They exploit
vulnerabilities present in programs or network services to gain access to the victim's machine and install
Bots
It is identical to the worm in terms of the means of propagation and infection, however, they differ in
the possibility of receiving external commands, similar to a backdoor. The possibility of remote control
makes the bot guarantee access to the victim's operating system, be updated with new exploits or even
allow its use for joint attacks. A network made up of several computers infected with bots is called a
botnet.
Botnets are used by hackers for DDoS (Distributed Denial of Service) attacks. This type of attack causes
multiple bot-infected machines (called zombies) to try to access a given service simultaneously, causing
Trojans
This type of program performs seemingly harmless functions, however, it performs secondary malicious
activities without the user's consent, constituting a gateway to other types of malware. Screensavers,
animated cursors, software cracks and keygen are examples of programs commonly used as trojan
horses.
Rootkits
Unlike other malware categories that focus on infection or malicious activity, the rootkit is concerned
with hiding the presence of malicious code on an infected computer, ensuring the presence of an
attacker or the persistence of malware. Rootkits often hide themselves by removing system logs,
encrypting their data and inserting them in a distributed way into operating system folders, altering task
manager programs so that they do not appear in active processes lists, and running at the kernel level,
9
Ransomware
A type of malware that restricts access to the infected system with a kind of lock and charges a ransom
(as in a kidnapping) so that access can be restored, which makes it practically impossible to track the
criminal who may receive the value. This type of "hijacking virus" works by encoding the operating
system's data so that the user no longer has access. Once a file is infected, the malware will encode
user data in the background without anyone noticing. When ready, it will issue a pop-up warning that
the machine is locked and that the user will no longer be able to use it, unless they pay the amount
required to obtain the key that gives access to the data again.
Vírus
It consists of malicious software that propagates by making copies of itself or infecting files/programs
present on the computer. Viruses need to be explicitly executed to start their malicious activities and
their infection process, as they do not make use of vulnerabilities in software present on the victim's
computer to spread. The scope of action of this type of malware is predominantly within the same
As with any forensic examination, the main objective of a malware analysis in the criminal context is to
answer the questions contained in the report request, as well as clarify the facts under investigation so
that the materiality of a crime can be confirmed or refuted, describing its dynamics and identify its
authorship, pointing out any other data deemed useful for the promotion of justice. However, we can
cite three major scenarios in which malware analysis is essential to draw conclusions about the criminal
practice.
Suspicious Software
In this scenario, there is software already known and used by users, developed internally or acquired
from third parties, in which there is a suspicion that this program performs secondary and malicious
functions without the administrators' knowledge. In this type of case, forensics must seek evidence
about the behavior of the analyzed software, identifying attributes that resemble a trojan.
10
Some of the goals of malware analysis in this context are:
• Analyze whether the software in question sends information to external environments not provided
for in its original documentation;
• Check if the program downloads executables or plug-ins from unauthorized sites in the software
documentation;
• Determine whether the signatures of the binary files that make up the software match the same
signatures of the binary files after a new installation, as the analyzed software may perform malicious
• Check if the suspicious program logs keystrokes in other applications or stores screenshots under
unauthorized conditions;
• Identify the existence of TCP/UDP ports opened by the program that are not authorized.
Malware Attacks
There are indications of an attack on a computing device or network infrastructure involving the use of
malicious software to gain access. In these circumstances, there is already a compromised and malware-
contaminated environment that must be examined for traces that determine the purpose, materiality
and, if possible, authorship of the attack. This type of examination is also known as post-mortem
analysis, which essentially boils down to studying a program's activity by analyzing its effects after
execution. They are part of the scope of malware analysis in this context:
• Identify the software that possibly contributed to the success of the attack;
• Check if there are open ports associated with the programs under analysis on the hosts or servers
present in the forensic environment;
• Investigate the existence of reverse connections related to suspicious software (a technique widely
used to circumvent security elements such as firewalls, proxies or IDS);
• Analyze the operating system logs in search of users created by the programs under analysis;
11
Malware as Secondary Element
This is the scenario with the highest number of cases in forensic examinations involving malware
analysis. In such cases, malicious software is not the main focus of forensic examination, but acts as an
additional provider of information that would not be obtained through conventional means. Currently,
most malware found have spyware-like functionality, capturing keystrokes or images displayed on the
monitor, which allows for obtaining valuable information for the exam, such as passwords for encrypted
volumes and records of instant conversations, among other data. The execution of certain activities is
• Verify that the malware is not the element responsible for the criminal practice present on the
computer device examined.
Types of Analysis
Malware analysis involves the adoption of a large number of tools and techniques in order to find out
what malicious activities are performed by the program in question. The specialized literature divides
malware analysis into two groups: static analysis and dynamic analysis. Next, we will study some
characteristics and subdivisions of these approaches and later, in the course of the article, we will delve
deeper into the skills and auxiliary tools involved in adopting each analysis separately. Additionally,
some characteristics of post-mortem analysis and some methods of anti-analysis will be discussed.
Static Analysis
The static analysis encompasses all possible ways to obtain information about the functioning of the
malware without executing it. Techniques such as string analysis of the executable file, identification of
used APIs and disassembling are examples of procedures adopted in static analysis. Static analysis is
further subdivided into two levels: basic static analysis and advanced static analysis.
The basic static analysis consists of examining the executable with the help of specific tools, but without
analyzing the machine instructions contained in the binary file. String parsing, PE32 [Data structure
12
present in MS-Windows binary files that contains essential information for the initialization and
functioning of an executable code] information, and displaying binary resources are examples of
techniques in basic static parsing. This type of analysis can provide some information about malicious
software features, but it is inefficient in more sophisticated malware that use methods to obfuscate the
code.
The advanced static analysis works by reverse engineering the code using a disassembler to examine
the instructions belonging to the binary file. Its advantage is that by examining the instructions that will
be executed on the CPU, advanced static analysis allows you to understand exactly what the software
does. On the other hand, the major disadvantage of this analysis is the scarcity of efficient tools that
automate the decompilation process, allowing the analyzed software to return to the high-level
programming language that gave rise to the binary file, which would simplify the understanding of its
functions.
This difficulty is caused by the numerous layers that a code written in a high-level language goes
through until it becomes an executable file. In addition, there are numerous compilers belonging to
different manufacturers, containing several versions and frameworks with infinite libraries. Thus, this
analysis requires a certain degree of knowledge in assembly language (Assembly), understanding of the
machine's architecture and concepts about the operating system on which the program was compiled,
making advanced static analysis unfeasible in some cases because it is quite expensive. Because of this,
we will not go into details about advanced static analysis, as it would take an entire book to understand
Dynamic Analysis
Dynamic analysis is the study of software functioning with the program running, that is, with the
malware allocated in memory and consuming CPU resources. In this analysis, changes made by the
program to the operating system, hard disk files, and network are examined. In addition, the assembly
language instructions present in the binary file are parsed at runtime using debuggers. In dynamic
analysis, it is extremely important to prepare a controlled environment before the execution of the
13
malware, since its execution without due care can cause serious side effects in the operating system
and in the network. Dynamic analysis is also subdivided into two levels according to its complexity.
The basic dynamic analysis essentially consists of observing the software in a computational
environment after its execution. This analysis has a more behavioral focus, checking what changes are
made by the malware, but without debugging the low-level instructions at runtime. For this, some tools
are essential that allow analyzing the system calls invoked by the program, the changes in the files, the
insertion of keys in the Windows registry, and the attempts to connect via the network, among other
The advanced dynamic analysis is based on the use of a debugger to examine each instruction applied
by the software at runtime. In this way, it is possible to collect more detailed information about the
program, being ideal in cases where good results are not obtained with the basic analyses. It is also a
very efficient procedure in situations where the malware makes use of anti-analysis techniques to make
the code difficult to understand. However, the advanced dynamic analysis is quite complex and for the
same reasons explained in the advanced static analysis we will not go into much detail in this analysis
Post-mortem Analysis
The post-mortem analysis comprises the behavioral analysis of malware based on the traces left by it
after its execution in a computing environment. In most cases, these traces are the only evidence of the
presence of malware in the environment, as the malicious program may have been removed
accidentally by a network administrator or on purpose by the attacker. In this scenario, the operating
system and other application logs are essential to understand how the malware works, since the
executable file is no longer present on the devices, which makes it impossible to analyze the program's
behavior in a controlled environment. Firewall logs, restore points, the event viewer, prefetch and the
Windows registry are examples of tools that can provide valuable information about the functioning of
14
The big problem with this analysis is that, in some cases, the identification of the presence of malware,
and the consequent detection of the attack, only occurs after a long period of time, which compromises
a large part of the traces left by the program on the devices. Over time, files and logs can be deleted,
metadata is changed, unallocated space is overwritten; all these actions compromise the evidence that
would allow a more incisive conclusion about the malicious activity of the software.
Anti-analysis
The anti-analysis involves numerous techniques adopted by malware to prevent its malicious activities
from being analyzed in the course of its execution. These techniques range from the simple behavior
change in the malicious activities of the software presenting an alternative behavior to the
Executable Compression
The compression technique, also known as packaging, basically consists of compressing an executable
file, generating compressed data that is combined with the decompression code into a single, self-
extracting executable file. At the time of executing the self-extracting archive, the decompression code
recreates the original executable file before executing it. This entire process takes place imperceptibly
to the user, since the decompression process takes place implicitly and without the need for additional
programs. Because of this, this technique differs from the use of external compression programs, such
as WinZip or WinRar, which require greater user interaction, in addition to the installation of these
programs.
Executable compression is designed to decrease the space taken up by executable files on secondary
storage devices, as well as reduce the time and bandwidth required for distributing this type of file over
the network. However, this technique has also been used by malware developers to make static analysis
of the binary file more difficult. When an executable file goes through a compression algorithm, it
becomes a smaller file with very different values from the original file. In this way, the vast majority of
information that could be obtained from the executable file through static analysis cannot be visualized
after using compression. Attempts to parse features, search for strings, or modify via hex editors will be
frustrated. Only the code responsible for decompressing the original file will be readable for static
15
analysis. In addition, compression modifies the signatures of the executable file, which makes it difficult
While compression makes malware analysis more difficult, it does not make it impossible. Many
malware use well-known packers to mask their signatures: UPX, ASPack, Petite, Armadillo. Identifying
the compressor used in the executable file compression process helps find tools that assist in extracting
the original file without running it, which will make it possible to analyze malware later.
There is a very efficient tool called PEiD (Portable Executable iDentificator) that helps identify if an
executable file has been compressed with a known packer. Furthermore, if the executable file has not
been compressed by a packer, PEiD allows identification of which compiler was used at the time of its
creation. This tool uses a database containing signatures from several packers and compilers, these
signatures are compared with snippets of the code present in the executable file, which makes it
possible to identify the program used in the production of the executable file. PEiD is a free tool and
Figure 1 - PEiD program interface demonstrating that the cbzvl.exe file was compressed with the UPX packer.
Obfuscation
Obfuscation is a technique that adopts several methods to make it difficult to reverse engineer a given
program. This practice was created with the aim of preventing commercial and legitimate programs
from being “cracked” by malicious users, making it difficult to use unlicensed software with fake serials
Although obfuscation is a feature that adds security to commercial software, this practice has been
adopted in some sophisticated malware. In these cases, most of the traditional methods and tools used
16
for malware analysis are discarded, requiring advanced skills in the use of debuggers and assembly
language for successful analysis. Another problem encountered with obfuscation is that, unlike what
happens when compressing the executable, there are few known programs that perform the
obfuscation of executable files in an automated way. As a rule, each software that has its code
obfuscated adopts methods implemented in different ways, which makes the analysis work even more
difficult. Even with all these obstacles, no obfuscated software is entirely immune from reverse
The main anti-cracking methods used in program obfuscation are described below:
• Elimination of Symbolic Information: Strings present in text boxes, which are displayed to the user
during the program, help the analyst find certain pieces of code of interest within the executable file.
Furthermore, Java programs contain a vast amount of symbolic information within the JAR file, such
as names of classes, methods, and libraries. Therefore, the main focus of this method is to eliminate
any and all textual information present in the software binary file(s). A simple mechanism used to hide
the textual information of a binary file is to create a function to encode/decode strings at runtime,
that is, the message will be decoded and intelligible when displayed to the user, but it will be
• Obfuscation and Code Encryption: These are methods that modify the structure of the program in
order to make it as hard to understand as possible, making analysis difficult through debuggers and
disassemblers, but without compromising the normal functionality of the software. Code obfuscation
basically involves changes in the logic and data handled by the program, making it difficult to
understand its normal flow of execution. Code scrambling encrypts parts of the program, using a key,
• Anti-debugger techniques: This method aims to prevent the program from being executed in a
debugger, harming the dynamic analysis of the software. A technique used to detect the presence of
debuggers is to create checksums in some code snippets and, after execution, perform an integrity
check on these snippets. This technique works because the analysts, when making use of debuggers,
change some parts of the binary file when they insert breakpoints or include patches in the code; this
allows the program to realize that it has been changed and terminate its execution.
17
A great additional source of information and bibliographic references on anti-analysis techniques can
be found in the research carried out by Murray Brand under the title “Analysis Avoidance Techniques of
Having finished the theoretical basis involving the terms and concepts related to malware analysis
presented at the beginning of the article, we will now move on to a more practical approach to malware
exams, presenting tools and techniques that help to understand the basic functioning of possible
malicious software.
In some cases in malware analysis exams, one of the first challenges for the computer expert is to
identify the presence of malicious software on the computing storage device. In scenarios such as
“search for malware” and “malware as a secondary element” presented in the topic “Objectives of
malware analysis in the forensic context”, the first task of the expert is to identify the presence of
suspicious software present on the forensic storage device to start to malware analysis.
One of the most basic ways to identify the presence of malware on a computer is to simply run an
antivirus on the storage device. Depending on the antivirus used, this tactic usually brings good results.
The correct procedure for running an antivirus on a hard drive that is being scanned is described below:
1. Create the physical hard disk image in a file. One way to do this is to use the Linux dd command or
2. Mount the created image file in read mode, avoiding unwanted changes on the disk. Both the
mount command on Linux and the FTK Imager perform image mounting;
It is extremely important to mount the created image file in read mode before running the antivirus,
because when detecting the presence of a threat, some antiviruses remove the file without issuing any
message to the user, which would compromise the integrity of the evidence and violate the chain of
18
VirusTotal
The antivirus to be used to scan the storage device will depend on the expert's taste. Currently, there
are numerous antivirus solutions, both commercial and free, including online tools that allow you to
upload a binary file to a web page where information about the uploaded file is displayed, including its
classification as malicious or harmless software. One of these online tools worth mentioning is the
When uploading a binary to VirusTotal, this file will be scanned by more than 50 antiviruses from the
most varied manufacturers. After the scan, a report is presented containing the number of antiviruses
that classified the file as malicious, the name of the malware and additional information about the
malware, if present (Figure 2). This approach involving multiple diagnoses will allow a more concrete
conclusion regarding the classification of the software, reducing cases of false positives or false
negatives presented by some antiviruses. In addition, if the identified malware is reasonably known, it
will be possible to obtain information about its characteristics and behavior through internet searches,
which will facilitate the tests to be performed and the traces to be found.
Figure 2 - Report provided by the VirusTotal tool after analyzing an executable file.
19
While scanning the storage device using antivirus is a simple and efficient approach, in some situations
it may not be so interesting. There are scenarios in which, after running the antivirus, some potentially
malicious software is found, but these programs were only stored on the examined machine and, for
some reason, were never executed. In most cases, because of its greater harmful capacity, it is more
important to try to find malware in an active condition on the computer, that is, malware that has been
executed at some point and remains operative on the system, in order to subsequently and in a
subsidiary way look for inactive malware simply stored on the device. For this, it is essential to find out
which executables, services, drivers, DLLs are launched together with the operating system, since most
malware makes use of bootstrap resources present in the operating system to stay active and to carry
Autoruns
When the Windows operating system is loaded, there are several locations that can be used to start
programs automatically. These bootstrap locations are present in some specific folders, registry keys,
system files, and scheduled tasks, among other areas of the operating system. Malware references are
often found in these locations to keep malicious software activities up and running. The great diversity
of bootstrap locations present in Windows creates the possibility that the execution of some malware
There is an extremely useful tool for identifying binaries loaded together with the operating system
called Autoruns. This tool pretty much consolidates all forms of auto-initialization of binary files in
Windows, including means of booting at logon time, running services, loading drivers, etc. Autoruns is
part of the Sysinternals suite of tools developed by Mark Russinovich. Sysinternals tools are approved
the course of this article, we will use some of the Sysinternals tools due to their great usefulness for
Autoruns, when opened, loads all program startup references in Windows, as shown in Figure 3. It
separates each startup entry into different tabs according to the auto-execution method, although all
entries can be viewed in the “Everything” tab. Autoruns is a very simple tool to use and its main
20
• Allows you to check boot entries for an operating system offline. To do so, just mount the image file
in read mode and point to the directory path containing the operating system root and user profile in
my File -> Analyze Offline System. The offline verification feature is very important in the
forensic context, since the operating system present on the forensic physical storage device must not
• Checks the digital signature of binary files referenced in boot entries. This functionality is very
important in situations where a binary file is infected with a virus and has its integrity violated. In such
cases, the digital signature verification will detect the improper change and the binary file will be
• Highlights in yellow the existing boot records in which the binary files were not found;
• Highlights in pink the records that do not have information from the software distributor, do not have
a digital signature of the file or the digital signature present is invalid;
• It has integration with the VirusTotal online tool, enabling the sending of hashes of the binary files
that are automatically initialized, which allows it to easily detect known malware;
• Allows you to save the records of Windows startup entries in a text file, which makes it possible to
compare it with a later startup state so that discrepancies in the records are identified, such as the
21
Figure 3 - Display of all startup entries in Autoruns.
A good strategy to identify malware using Autoruns is to look for files that have some number of hits in
the “VirusTotal” column. In Figure 3 for example, the entry lb02 (present in the registry key
lb02.exe in the System32 directory of Windows, which was classified as malware by 41 of the 56
antivirus used in the analysis by VirusTotal. By clicking on the value 41/56 present in the VirusTotal
column, the default browser will open and the VirusTotal tool report will be displayed containing
Other suspicious binary files are those that have entries highlighted in pink, which basically represent
the absence or invalidity of the software vendor's digital signature. For more details, it is important to
verify that a digital signature exists or that the digital signature is valid. To do so, just right-click on the
entry and then click on Properties..., a window containing the properties of the binary file will be
displayed. Click on the Digital Signatures tab (this tab is only displayed if there is a signature for the file)
and then on the Details button. A new window will be presented containing the signer's data, including
the signature validity information according to Figure 4. According to Figure 4, the babylon.exe
22
executable does not have a valid digital signature, which may be caused by infection by a virus, the use
of a crack to violate the license to use the software, or an expired digital certificate.
Most malware does not have digital signatures, not even information about its distributor (Description
and Publisher columns), as is the case with the lb02.exe malware. Because of this, binary files with these
Prefetch Files
Prefetch is a Windows component that is part of the memory manager and its main purpose is to
reduce the time needed to start programs. It creates a file with a .pf extension in the C:
\Windows\Prefetch folder for each program it runs. This file contains information about the components
needed to run the program, such as DLLs, data/configuration files, other executables, etc. With this
information already centralized in a cache file, Windows is able to load programs faster.
Although Prefetch was developed with the aim of improving operating system performance, it is an
excellent source of data for forensic analysis and, consequently, for malware scanning .pf files hold
valuable information about the executables they represent. Data about last run date, number of runs,
and path of binary files can be obtained by reading these files. In some cases, these files make it
possible to obtain information even from executable files that have already been deleted.
Prefetch files are binary files, that is, they cannot be read by a simple text editor. For this, it is necessary
to use a parser to interpret the .pf files. An excellent parser for prefetch files is Nirsoft's
23
WinPrefetchView, this program can be downloaded for free at http://www.nirsoft.net/utils/
win_prefetch_view.html.
To load WinPrefetchView, open a command prompt pointing to the directory where the program is
located and type: WinPrefetchView.exe /folder <path to prefetch folder>. The characters < and >
represent program parameters. If the /folder parameter is not specified, WinPrefetchView will interpret
the prefetch files from the current machine on which the program is running and not the prefetch files
WinPrefetchView presents the interface as shown in Figure 5. In the upper frame, the tool lists all the
prefetch files found in the directory specified as a parameter, containing information about the last
execution date, number of executions, executable path, etc. In the lower frame, the files accessed by
Prefetch when loading a given executable file are listed, along with their respective paths on the hard
disk.
24
When using prefetch files to identify the presence of malware, the most important information to be
analyzed in this context is the data on the hard disk path of executables that have associated .pf files
(WinPrefetchView's “Process Path” field). Most malware is not installed in folders normally used by
normal programs (e.g., C:\Program Files), they make use of directories that usually do not host
executable files. Because of this, it is important to identify programs running under the conditions listed
below:
• In Program Data (C:\ProgramData) or generic users (All Users, Default User, Guest);
• Windows programs run in alternate paths (e.g., C:\Windows\svchost.exe, the correct one
would be C:\Windows\System32\svchost.exe).
It is also important to check the path of files loaded by a suspicious program at the time of its launch
There are situations where an executable has more than one prefetch file; in other situations, it may be
that the executable has no associated .pf file. These cases will depend on how the operating system
manages Prefetch, in addition to the use of external applications that clean the disk in order to increase
free space.
Static Analysis
Once the suspicious software is identified and extracted from the image file of the examined storage
device, the first step in malware analysis is to get as much information about the program as possible
without running it. For this, techniques and tools used in static analysis are of paramount importance.
Strings
A very simple and fast way to obtain information about the functionality of a program is to visualize the
strings present in a binary file. Through the strings, it is possible to identify warning messages, specific
paths of files on the hard disk, and URLs used for connection, among other data relevant to the exam.
25
The sysinternals tool suite has a program called strings (https://technet.microsoft.com/en-us/
sysinternals/bb897439), which displays the text present in a binary file, encoded in both ASCII and
UNICODE.
To view the strings present in an executable, open a command prompt pointing to the directory where
the program is located and type: strings.exe <executable_name.exe>. The characters < and >
addition to the name, the full path of the file on the hard disk. You can also search for specific words
within the executable by combining the Strings tool with the Windows Findstr command using the
If the result of the Strings tool, after its execution in a binary file, is composed in its vast majority of
small strings, apparently meaningless, this may indicate the use of a packer in the binary file. As studied
in the topic “Executable Compression”, the best way to detect if a binary file makes use of a
compressor is to use the PeiD tool. Depending on the packer used, extracting the original executable
file can be quite simple, an example of this is UPX. To extract the original executable from an
executable compressed with UPX, just download the packer (http://upx.sourceforge.net/) and type at a
represents the difference between the strings presented in a compressed executable and its respective
original file.
Table 1 – First lines of the strings.exe tool bringing words with at least 8 characters in two different executable files.
26
Portable Executable Format
In order for us to understand and interpret the information provided by the more specific tools used in
static analysis, it is essential to understand the format of an executable file on the Windows operating
The PE (Portable Executable) format is a data structure present in certain binary files (executables, DLLs,
object code, etc.). Analyzing this data structure, it is possible to obtain information about imported
libraries and functions, exported functions, and resources used (menus, screens, icons, etc.). The PE file
format contains a few headers followed by a series of sections. Headers have metadata about the file
itself. Following the headers are sections of the file, which contain useful information. Listed below are
• .text – This section contains the instructions that the CPU executes. As a rule, it is the only section that
has the program execution code itself, the other sections store data and supporting information;
• .rdata – Typically contains the program's import and export information, including libraries and
functions. In some cases, a file may contain .idata and .edata sections which store import and export
data, respectively;
• .data – The .data section holds the program's global data, that is, information that can be accessed
from anywhere in the program and at any time;
• .rsrc – This section stores the resources used by the program that are not considered part of the
executable, such as icons, menus, images, screens, cursors, and strings. Strings may be stored
elsewhere in the program, but are often present in the .rsrc section.
In some cases, these sections may have different names due to the use of different compilers or the
complexity deliberately used by obfuscation, however the vast majority of PE files use the section
A great complementary MSDN reference on the PE format can be found in Matt Pietrek's article
“Peering Inside the PE: A Tour of the Win32 Portable Executable File Format” at https://
msdn.microsoft.com/en-us/library/ms809762.aspx.
27
PEview
A simple and practical program used to interpret the information present in a PE file is PEview. This
Some metadata contained in PEview are worth mentioning. The IMAGE_FILE_HEADER header, present
inside the IMAGE_NT_HEADERS header, contains a field called Time Date Stamp, which displays the
compilation date of the executable. This information helps to understand the popularity of malware.
For example, old malware is more likely to have a detailed analysis published on the internet about the
program, making it easier to understand its malicious activities; in newly developed malware, the
probability of success in internet searches is drastically reduced in addition to the lower chances of
detection through antivirus. Unfortunately, the build date can easily be changed by a malware
programmer. If a completely meaningless date is identified, it is very likely that it has been changed.
Another interesting data is present in the IMAGE_OPTIONAL_HEADER header in the Subsystem field.
Through this field, it is possible to know if the program has a graphical interface (GUI) or is based on the
command line (CLI). Programs with a graphical interface are represented by the value
the value IMAGE_SUBSYSTEM_WINDOWS_CUI. The absence of the .rsrc section is also an indication that
the program works in a console or composes a library (DLL), as these types of applications usually do
28
Information about the size occupied by each section within the binary file can be obtained from the
IMAGE_SECTION_HEADER headers for each section. The value present in the Virtual Size field
represents the size occupied in memory by the section after its loading. The Size of Raw Data field
displays the size occupied on the hard disk by the corresponding section. Generally, these sizes tend to
be the same, with slight variations due to the size of the blocks in the file system. However, large
variations in the sizes present in the .text section can mean the use of packers in the execution of the
Although PEview displays all the metadata contained in a PE file, there are other more specialized tools
that focus on some specific sections of the PE format as we will see below.
Dependency Walker
Aiming at software modularization, programmers develop libraries that can be used by several different
programs, which facilitates programming and code reuse. However, for the functions provided by a
library to be used by programs, linkage mechanisms between the two are necessary:
• Static Linking – In static linking, the entire contents of the library are copied to the code area of the
executable at compile time. This practice is less common as it makes the executable file much larger
than other binding practices as it merges the executable code with the library code into a single file.
Static binding makes malware analysis more complex due to the increase in the number of
• Dynamic Link – This type of link loads all the libraries used by the program at the time of its
execution. The libraries are usually present in DLLs (Dynamic Link Library) files, which export functions
that can be used by external programs. External DLL functions used by a program are usually present
in the .rdata section in the PE header of the binary file. Dynamic binding is the most used form of
engineering because, based on the functions used by the program in system DLLs, it is possible to
have an idea of the activities performed by the application without executing it. For example, if the
InternetOpenUrl function (belonging to the Wininet.dll Windows library) is present in the import table
of a program, it is possible to assume that this program at some point makes an internet connection,
opening a specific URL and making use of some high level protocol like HTTP, HTTPS, FTP, etc;
29
• Runtime linking – Unlike dynamic linking, in which all the libraries used are linked at program startup,
in runtime linking, the program links to the libraries only when these functions are needed in the
program’s course of execution. Therefore, it is not possible to identify the functions of imported
libraries by reading the .rdata section of the PE header. The loading of these libraries during
execution is done manually by the programmer, generally making use of the LoadLibrary and
GetProcAddress functions (present in the Kernel32.dll library), which allows accessing any function
belonging to the external libraries. Because of this, when an executable adopts this type of binding, it
is not possible to categorically state which functions are bound to the program. Runtime binding is
often found in binary files that make use of packers or in programs that adopt obfuscation techniques,
although it is also possible to find this type of binding in legitimate programs due to the peculiarities
of some compilers.
Despite the existence of other linking methods, dynamic linking remains the most used way by
programs to link to external libraries. Therefore, a more detailed analysis of the import table found in
the .rdata section of an executable file can provide useful information about the functioning of the
The Dependency Walker tool (http://www.dependencywalker.com/) has the purpose of exposing the
dynamic links of a binary file, that is, it presents all the functions of imported external libraries present in
the .rdata section of the PE header. In addition, it also points out the functions exported by a binary file,
if they exist. Figure 7 presents the tool's interface when analyzing possible spyware.
30
Figure 7 - Dependency Walker tool interface with enumerated panels.
The upper left panel (Panel 1) displays a tree containing as root element the analyzed executable file
(FILE.EX_) followed by the DLLs imported by this program at a lower level (KERNEL32.DLL,
one of the DLLs, the functions imported by the executable program present in the selected library
(GDI32.DLL) are displayed in the upper right panel (Panel 2). Based on these functions, it is possible to
have an idea of what the program can do. The BitBlt function highlighted in Figure 7 belonging to
GDI32.DLL, for example, is often used to generate user screen captures, which would allow the
program to act as a screenlogger, recording the screens accessed by the user in image files. It is also
important to note that the program does not import functions from suspicious DLLs, as these may be
part of the malware present in another binary file. In the right central panel (Panel 3), all the functions
exported by the selected DLL that can be imported by a program are displayed.
To check if the file under analysis exports any function, just click on the root element of the tree in Panel
In some cases, binary files do not import functions from DLLs by name, but using the ordinal number of
the function. In these cases, the names of the imported functions will not be displayed in the Function
31
field on Panel 2, only the respective function number will be displayed in the Ordinal field. To identify
the name of a function having only its ordinal number, just right-click on the desired function in Panel 2
and click on “Highlight Matching Export Function”. The function corresponding to the selected ordinal
The central panel (Panel 4) displays information about the DLL versions that can be loaded along with
the program. The bottom pane (Panel 5) displays warning and error messages, if any.
Detailed information about the libraries and functions provided by Windows can be consulted on the
85%29.aspx.
Additionally, Lohit Mehta member of InfoSec Institute ranked the Windows functions that are commonly
found in malware analysis. This list is divided into two parts and can be accessed at:
• http://resources.infosecinstitute.com/windows-functions-in-malware-analysis-cheat-sheet-part-1/,
• http://resources.infosecinstitute.com/windows-functions-in-malware-analysis-cheat-sheet-part-2/.
Resource Hacker
Another important part of the PE header is the .rsrc section. As previously studied, this section contains
various features of the application, such as dialog boxes, menus, and strings, among others. The
Resource Hacker tool allows you to graphically view the resources present in a binary file in PE format.
This tool is free and can be downloaded from the official address http://www.angusj.com/
resourcehacker/.
Resource Hacker helps understand in advance which are the possible interactions between program
and user. Based on the menus and dialog boxes displayed by the tool, it is possible to have an idea of
the options available to the user for manipulating the program under analysis. In addition, it is possible
in some cases to identify hidden screens or menus that only after a certain action, such as pressing
certain keys together, would be displayed. It is also important to examine the strings classified by the
Resource Hacker, as they may contain messages presented to the user or some additional coding done
via script or markup language (e.g. assembling an HTML page for later submission).
32
Figure 8 - Resource Hacker program interface showing a dialog box present in the binary file.
Figure 8 presents the Resource Hacker interface. The left side panel lists the resources available in the
file by category (Menu, Dialog, Icon, String Table, etc.), each category represents a resource type as per
the name. When expanding a category, all resources present are listed according to their ID number.
When a resource is selected, its content will be displayed in the right-hand pane.
Some programs, aiming to provide support for multiple languages, use Microsoft's technology called
MUI - Multilingual User Interface. When this technology is used, the .rsrc section of the PE header of the
binary file will only present basic features such as icons or binary version information. In these cases, the
other resources will usually be present in a file that has the same name as the original file, but with the
addition of the .mui extension. The .mui file will contain the most relevant features for reverse
Dynamic Analysis
After obtaining all possible information about how the binary file works through static analysis, the next
step is to examine its behavior during execution. Great care must be taken during this procedure due to
the risks caused by the execution of the malware to the operating system and the network
infrastructure. Side effects are unwanted, so it is essential to prepare a controlled and safe environment
33
for the execution of activities. There are two ways to study how malware works safely: setting up a
Setting up a physical environment is a great way to analyze malware, since the anti-analysis techniques
used to detect the execution of the malicious program in virtual machines would not work. However,
this is a very expensive practice, as in many cases an environment composed of more than one machine
is required, which would lead to the assembly of a segregated physical network. In addition, the
physical environment is not very flexible, as in some cases it is necessary to reinstall the operating
system to return to the state before the malware was executed. For these and other reasons, running
malware in a controlled physical environment can become unfeasible, although in some cases it is the
only alternative.
The creation of virtual machines is the most used method in the creation of safe environments for the
execution of malware due to the ease provided in the manipulation of virtual hard disks through
snapshots, also known as snapshots. Snapshots allow you to save the entire state of a hard disk at a
given time, making it possible to return to the state of the hard disk at the time of snapshot creation
after several changes to files. In addition, many virtualization tools allow the creation of multiple
snapshots, which allows saving several points at different times of the malware analysis, allowing the
Process Explorer
Process Explorer presents information about all processes running on the operating system. It is
analogous to the Windows Task Manager (taskmgr), however, it is much more robust and contains
additional information and resources that are of interest to malware analysis. This tool displays active
processes in a hierarchical manner, making it possible to identify the relationship between parent and
child processes. Process Explorer is part of the Sysinternals suite of tools and can be downloaded for
free at https://technet.microsoft.com/en-us/sysinternals/bb896653.aspx.
One of the main features of Process Explorer is the possibility to view the handles and DLLs loaded by a
process. Handles are operating system resources that are being handled by a process during its
execution. Open files and directories, Windows registry keys, mutexes and threads are examples of
handles that can be used by a process. By analyzing the handles, it is possible to identify the directories
34
in which the files created by the malware are present. In addition, it is possible to identify other DLLs
loaded by the malware at runtime (runtime linking), which could not be identified by analyzing
the .rdata section using Dependency Walker, due to the use of the LoadLibrary and
Process Explorer also makes it possible to search by name of handles or DLLs; in this way, it is possible
to identify which processes have loaded a particular DLL or which processes have a particular file open.
The tool's interface is shown in Figure 9. The Process Explorer divides the interface into two panels, the
upper panel displays the running processes and the lower panel displays the Handles or DLLs used by
the process. It is possible to add new fields in the upper panel with more specific information about the
process (Menu View -> Select Columns...). Blue processes represent user processes, pink
processes represent service processes. Also, new objects are initially highlighted in green and newly
deleted objects are highlighted in red. Additionally, if any process is displayed in purple, it means that
its executable file is packaged. The information displayed by the tool is updated every second by
default.
35
Figure 9 - Process Explorer tool interface.
When double-clicking on an active process, detailed information about the process is displayed, such
as data about the executable file, performance data, open TCP/UDP ports, threads created by the
Like the Autoruns tool, Process Explorer performs the verification of the digital signatures of running
program files and loaded DLLs. In addition, it also has integration with the VirusTotal online scanning
tool. These features make it easy to identify virus-infected files as well as find new running malware.
Process Monitor
Unlike Process Explorer, which displays all running processes, Process Monitor captures some events
performed by processes in a certain period. This tool captures four classes of events performed by a
process on Windows: activities in the registry, activities in the file system, activities in the network, and
activities involving processes and threads. Process Monitor is analogous to a sniffer, only instead of
36
capturing network packets, it captures system calls executed by processes, such as querying registry
keys, writing to a file on the hard disk, or creating a thread, among many other activities.
Process Monitor is also a program that is part of the Sysinternals suite of tools and, like the other tools
in this suite, it is free and can be downloaded by accessing the address https://technet.microsoft.com/
en-us/library/bb896645.aspx.
Figure 10 represents the Process Monitor interface. When running, Process Monitor starts capturing the
events that are occurring in the operating system. To pause capturing events, just click on the File menu
and then on Capture Events or use the shortcut keys Ctrl+E. Its interface presents a timeline
containing information about the events of a process. Each column displays specific data about the
event, such as the date and time of the event, the name of the process, the type of operation, the path
in which the event took place, the result of the event and detailed information about the event.
The tool captures thousands of events per second from many different processes, which makes data
analysis and interpretation difficult. To facilitate the visualization and interpretation of captured events,
Process Monitor has a filtering resource in which it is possible to create filters using a series of specific
characteristics such as process PID, category, process user, and operation, among others. Using the
filtering feature, only events that match the filter expression will be displayed. To access the filtering
37
feature, just click on the Filter menu and then on Filter... or use the shortcut keys Ctrl+L. It is also
possible to highlight specific events displayed in the analysis interface, making it easier to identify
certain program behaviors. To access this feature, just click on the Filter menu and then on Highlight...
or use the shortcut keys Ctrl+H. When the filtering resource is being used, it only performs the
filtering of events already registered, so it is not a capture filter and cannot be used in order to reduce
the amount of captured events, which would decrease the amount of memory consumed by the tool,
The interface shown in Figure 10 shows the mapping of events from a keylogger (while monitoring,
random keys were pressed). In this capture, the filtering feature is used so that only the process with
PID 6536 is displayed. In addition, the events that query registry values (RegQueryValue) and write to
files (WriteFile) are highlighted. Based on the capture, it is possible to infer that, when performing
search of the keyboard layout used by the user. The value of this registry key contains the name of the
DLL responsible for the keyboard layout being used at the time of capture (KBDBR.DLL). The malware
loads the KBDBR.DLL library to configure its layout according to the layout used. Subsequently, several
writings are carried out in the file called TUT.001, present in the path C:\ProgramData\TUT\, which
most likely stores the keys captured by the keylogger that are typed by the user.
Although Process Monitor captures certain network activities, its use is not recommended for
performing scans that aim to identify malware behavior in a network environment, since it only presents
superficial information about connections made by processes in addition to not performing capturing
trafficked frames. For this type of exam, other more specific tools will be presented throughout this
article.
Network Monitor
Many malware use a network to receive external commands, updates or even send information to third
parties. Based on the data trafficked by the malware, it is possible to discover from attempts to
propagate through the local network to the identification of servers that store stolen personal data from
38
users. Because of this, it is essential to examine the network activities performed by the program under
analysis.
To capture the data transmitted in a computational device effectively, it is necessary to use a network
sniffer. For a sniffer to work in the virtual environment created for the dynamic analysis of malware, it is
necessary to create a virtual network or, as a last resort, to provide internet access to the program under
analysis.
Wireshark is currently, without a doubt, the most used sniffer in the world, having a series of extremely
efficient features and functionalities. However, for malware analysis in Windows environments, it is
recommended to use the Microsoft sniffer called Network Monitor (NetMon). NetMon's great
differential in relation to Wireshark is its simplicity in performing the traffic analysis of a specific process,
ignoring the traffic of the other processes, which facilitates the interpretation of the frames sent and
received from a given program in execution. With NetMon, just click on the desired process and all
network traffic associated with that process will be displayed by the tool. This sniffer is free and can be
downloaded at https://www.microsoft.com/en-us/download/details.aspx?id=4865.
When starting Network Monitor, the user must select the network device on which he wants to capture
frames and then click on New Capture. Then, the main interface of the program will be displayed, like
the one shown in Figure 11. To start capturing frames, just click on the Start button.
39
Figure 11 - Network Monitor tool interface with enumerated panels.
NetMon's interface is divided into panels with distinct information about data in transit as detailed
below:
• Network Conversations (Panel 1) displays the processes that manifested some network activity during
the capture period. The number in parentheses next to the process name represents its PID number.
By clicking on the + sign, all process flows separated by source and destination IP address will be
displayed. The value <Unknown> represents all the traffic generated by the network to which the
machine belongs that does not have any associated process, for example, broadcast traffic, wi-fi
• Display Filter (Panel 2) allows adding display filters to captured frames involving regular expressions.
The filter applied in this panel is just a visualization filter. To apply capture filters, click the Capture
Settings button;
• Frame Summary (Panel 3) lists all frames stored during capture. If a particular process is selected in
Network Conversations, Frame Summary will only display frames that are involved with that process.
40
This panel has a lot of information about the frames, such as timestamp, source and destination
addresses, and protocol, among others. To add or delete the columns displayed, just right-click on
• Frame Details (Panel 4) displays detailed information about all protocol headers that make up the
frame selected in Panel 3. Panel 4 displays protocol headers organized in a bottom-up manner (lower
layer first and then lower layer protocol headers, highest layer). By clicking on the + sign of one of
the headers, detailed information about the selected header will be displayed;
• Hex Details (Panel 5) displays the frame in hex without having been parsed by a parser. On the right
side of this panel, the ASCII equivalent of the hexadecimal values shown on the left side is displayed.
When clicking on any header value in Panel 4, the respective value in hexadecimal is highlighted in
Hex Details.
Panel 3 of Figure 11 displays the capture of all frames sent/received by the Firefox browser according
to the selection present in Panel 1. The frame selected in Panel 3 is displayed in detail in Panel 4,
containing the request for the main page of the URL www. google.com.br through the HTTP protocol,
which makes use of the GET command for the request. Panel 5 presents the selected frame in
hexadecimal format, highlighting the HTTP protocol request as selected in frame 4. In the capture
recorded in the figure, visualization filters were not applied, as shown in Panel 2.
When capturing network data sent by malware with NetMon, it is important to note which IP address or
URL the program is trying to connect to, which transport and application protocol is being used, which
data is being sent. With this information in mind, it is possible to understand the malware's intentions
TCPView
Using a sniffer in a virtual machine with internet access, if a given process does not transmit any packets
during the capture, it will not be possible to detect whether this process makes use of the network,
even if it keeps open ports or persistent connections. To identify this type of behavior, it is necessary to
consult the states of active TCP connections in the operating system. A quick way to find out the TCP
states of connections is to run the netstat command at the Windows command prompt. While netstat is
useful in many cases, it is not as efficient for malware analysis as it does not dynamically display state
41
information, that is, it does not automatically update changes in connection states. Also, netstat is not
To identify persistent connections and ports opened by a process, TCPView is a great solution. With this
tool it is possible to identify all processes that keep ports open in the system, both UDP and TCP
(LISTENING state), in addition to displaying all established TCP connections, even if they are not
transmitting data. Furthermore, the information presented by the program is updated every second by
default. This program is free and is also part of the sysinternals toolkit and can be found at https://
technet.microsoft.com/en-us/library/bb897437.aspx.
The TCPView interface can be seen in Figure 12. Its interface is very simple and displays information
related to the process responsible for the connection, transport protocol, ports and local and remote
addresses, and the connection status, among other information. By default, TCPView resolves DNS
names to IP addresses and ports. To disable this functionality, just access the Options menu and
uncheck the Resolve Addresses option. To display only active connections, access the Options menu
42
It is important to note whether the program under analysis keeps any UDP ports open, as some
backdoors do not use the TCP transport protocol, which, consequently, does not generate connection
states. It is also interesting to verify that the suspicious program does not establish persistent TCP
connections with external servers with port 80 as the destination port, as this is a technique used to
hide remote connections to circumvent firewalls that filter other types of known ports.
Advanced Analysis
If even after employing basic malware analysis tools and techniques it is still not enough to understand
the functioning of the program and identify possible malicious activities, then the application of
advanced techniques on reverse engineering will be inevitable. Depending on the complexity and
sophistication of the malware, it will take many hours of work to get effective results. In addition to
time, it will be essential to use advanced tools such as disassemblers and debuggers. Although there is
no space for a more in-depth analysis of this matter, below, two widely used tools for software reverse
IDA
A very powerful and widely used disassembler and debugger is the IDA (Interactive Disassembler) by
multiple processors and multiple executable file formats. IDA converts all machine instructions present
in the binary file back into assembly language. In addition, it has a series of features such as function
identification, stack analysis, and search for local variables, among many others.
IDA tries to be as interactive as possible. Among its greatest resources is the graphic mode, which
creates a kind of flowchart which facilitates the visualization of the calls and the course of the program
flow. It also allows adding comments, labeling data and naming functions, and then saving this
information in its internal database for further analysis. Another outstanding feature is FLIRT (Fast
Library Identification and Recognition Technology). This technology allows the recognition of basic
functions generated by compilers supported by IDA and present in libraries of a certain high-level
language. For example, in a program created in C using FLIRT, IDA makes it possible to easily identify
standard functions from C-language libraries (stdio.h, stdlib.h, string.h) as the functions
43
There are commercial and free versions of IDA. Version 5.0 of IDA is free for non-commercial use,
however, it has limited functionality and low support for processor families. Consequently, commercial
OllyDbg
OllyDbg is a debugger for x86 architecture that allows you to analyze programs while they are running.
OllyDbg was widely used by crackers to break codes, circumventing serials and software licenses. It is
currently popular for reverse engineering malware due to its ease of manipulation, addition of plug-ins
and, most importantly, for being free. This tool can be downloaded via the link http://www.ollydbg.de/
download.htm.
Its interface is very intuitive, making it possible to track registers, recognize procedures, analyze the
stack, or search for strings, among other applications. It also allows the creation of patches in the
executable file, which basically consists of saving the changes made to the program code, providing
the modification of instructions present in the binary file and in the execution flow. Currently, its biggest
disadvantage is that it does not support binary files compiled for x64 processors, which makes it
Final Considerations
Malware analysis becomes an increasingly frequent reality in forensic examinations. Understanding the
concepts on the subject, knowing methods to identify a malicious program and understanding its basic
functioning are tasks that must be present in the daily life of any criminal expert.
As portrayed throughout the article, malware are resources that provide additional elements to the
forensic examination, making it possible to find new information about the device user (e.g., spyware
analysis), or even start new investigations into cybercrimes (e.g., bot or backdoor analysis). Because of
this, when malicious programs are not the focus of forensic examination, they should not be
References:
• AQUILINA, J.; CASEY, E.; MALIN, C. Malware Forensics: Investigating and Analyzing Malicious Code.
EUA: Syngress, 2008.
44
• BRAND, M. Analysis Avoidance Techniques of Malicious Software. Available at: <http://ro.ecu.edu.au/
cgi/viewcontent.cgi?article=1138&context=these>.
• EAGLE, C. The IDA Pro Book: The Unofficial Guide to the World's Most Popular Disassembler. EUA:
No Starch Press, 2011.
• HOGLUND, G; MCGRAW, G. Exploiting Software: How to Break Code. EUA: Addison-Wesley, 2004.
• INFOSEC INSTITUTE. Windows Functions in Malware Analysis: Cheat Sheet – Part 1. Available at:
<http://resources.infosecinstitute.com/windows-functions-in-malware-analysis-cheat-sheet-part-1>.
• INFOSEC INSTITUTE. Windows Functions in Malware Analysis: Cheat Sheet – Part 2. Available at:
<http://resources.infosecinstitute.com/windows-functions-in-malware-analysis-cheat-sheet-part-2>.
• LIGH, M.; ADAIR, S.; HARTSTEIN, B.; RICHARD, M. Malware Analyst’s Cookbook and DVD: Tools and
Techniques for Fighting Malicious Code. EUA: Wiley, 2011.
• MSDN. Peering Inside the PE: A Tour of the Win32 Portable Executable File Format. Available at:
<https://msdn.microsoft.com/en-us/library/ms809762.aspx>.
• SIKORSKI, M.; HONIG, A. Practical Malware Analysis: The Hands-On Guide to Dissecting Malicious
Software. EUA: No Starch Press, 2012.
45
About the Authors
Forensic Sciences (SBCF). C|EH, C|HFI, DSFE and ISO 27002 Senior Manager. Author and technical
reviewer of the book “Treatise of Computer Forensics”. Reviewer and editorial board member of the
free software culture. He has worked in the technology area for more than 20
Analyst and Ad-hoc Forensic Computer Expert. Leader of the OWASP Belém
46
IPED, An Amazing Tool
To Analyze A Disk
Image
by Marcus Fábio Fontenelle
This article will give you a brief introduction to IPED, an open source tool developed and
maintained by the Federal Police of Brazil to analyze digital evidence.
Introduction
IPED, Digital Evidence Processor and Indexer (translated from Portuguese) is a tool developed and
maintained by the Brazilian Federal Police. Its development was started in 2012. Initially, it was
supervised by the digital forensic expert Luís Felipe Nassif, and now there is a team of digital forensic
At first, the idea was to facilitate the use of digital attachments on forensic reports, index them, and
offer a search tool. It was developed in Java, working through an intuitive command line, and
As of 2014, it began to be widely used in Operation Car Wash, a huge Brazilian criminal investigation
focused on corruption and led by Brazilian Federal Police. At that time, IPED started also to be widely
used to process and analyze digital evidence, often seized at crime scenes by Brazilian law enforcement
or in a corporate investigation by private examiners. Besides, this tool is listed in INTERPOL Innovation
Centre (https://github.com/INTERPOL-Innovation-Centre/IPED).
47
Although it has always been open source, it was only in 2019 that its code was officially published and
• Customizable: it has several setting files that allow suitability to the best evidence acquisition scenario
based on hardware and software available;
• High Portability: developed in Java allowing compatibility with Windows and Linux environments;
• Efficient Processing: supports millions of items and smart use of the idle time of processors/cores
(known as workers);
• Full Extraction: Support for multiple file formats (DD, 001, E01, ISO, VHD, VMDK, UFDR, AD1)
performing data carving, expanding containers (files that have other files inside them, for instance,
compressed files), indexing, search for keywords in the content and properties of files, and
• Intuitive Interface of Analysis: The developers tried to suit the interface so it was familiar to those that
had already used known commercial software, such as FTK or Encase;
• Code: It is available free for use and may have its features extended and improved.
IPED has several functionalities similar to the main existing commercial tools. Among these features, we
can highlight:
• Image thumbnails;
• Video thumbnails;
• Hexadecimal viewer;
• Visualization of the most used types of documents (Office, LibreOffice, PDF, etc);
• Georeferencing;
• Nudity detection;
48
• Regex;
• Tags;
• HTML report.
IPED Configuration
Next, the basic configuration of IPED will be presented. However, it is important that you analyze
carefully all the configuration options of the software, because in a real case you may need some “fine-
Before configuring IPED, Java (64-bit version) must be installed on the computer. After installing Java,
go to the Windows search box and search for CONFIGURE JAVA. On the Java tab, click the View
button. In the Runtime Parameters type –Xmx3G exactly as shown in Figure 1 below. This
49
After installing and configuring Java, download IPED and unzip it in the folder of your choice. We must
Go to the folder where you unzipped the IPED, open the LocalConfig.txt file, and define the
To enable carving and file recovery by IPED, in the folder where the software was unzipped, in folder
iped-<version>\profiles\en\default, open the file IPEDConfig.txt and enable (if they are
In order to expand container files, detect possible encrypted files, index metadata, generate previews,
etc., and in addition to that, show the thumbnails of images and extract images from videos, it is also
Running IPED
Once you’ve finished the basic configuration of IPED, you are ready to run it. Its syntax is as follows:
50
It is worth noticing that IPED accepts more than one image file as input, allowing the analysis of copies
of different media belonging to the same case in the same interface. Right after running the Java
In Figure 3, we can see that there is a progress bar at the top indicating the number of processed items
and the number of items to be processed, as well as the estimated completion time of processing. On
the left, several statistics are presented, the most interesting being the elapsed and estimated
processing time ("Processing Time" and “Estimated Finished" options) and the average processing
speed ("Average Speed" option). On the right, the number of processing cores used (workers) is shown
and which tasks are being performed in each one of them. In the center, the tasks in execution are
51
Understand the Interface and Main Features
After processing, in the target folder, the file IPED-SearchApp.exe will be found. When executed,
In the upper left corner, there are several predefined filters (Figure 5) that can help the investigator in
his/her analysis.
52
Right below the filters section, two ways of viewing the indexed information is presented: Categories
The Categories tab summarizes and categorizes the information that is usually most relevant in an
investigation, such as users registered in the system, installed software, Internet search and browsing
history, Windows trash files, cookies and cache of browsers, downloads carried out, etc. New releases
of the IPED could bring new categories. The Evidences tab presents the disk partitions and the entire
file hierarchy, as well as the file system metadata found in the processed image file.
In the center of the IPED interface, a list of files is shown according to the selection made in the
Categories or Evidences tab. The Table tab (Figure 8) shows the files in detail, presenting information
such as name, type, size, MAC times, if the file was deleted and recovered, when was the last change in
the Windows registry, the path where the file was found in the file system and its hash. Usually, the
53
Figure 8 – List of Files
When a folder with image files is selected, is possible on the Gallery tab to visualize the thumbnails of
If a certain image has geolocation metadata, a map indicating the georeferenced location could be
viewed on the Map tab. To be able to use georeferencing properly, it is necessary to set a Google API
key, but it's possible to leave the field blank and use a limited version of this feature (Figure 10).
Figure 10 – Georeferencing
54
In the lower right corner of the interface, you can access more detailed information about the selected
file. If a preview of the file is possible, it will be displayed on the Preview tab (Figure 11). Video files may
In the Metadata tab, additional information is presented according to the type of file, such as the date
and time of file creation (Figure 13), latitude and longitude (Figure 14), or make and model of the
device that captured the image (Figure 15) and much more.
55
Figure 13 – Metadata: creation date and time, and time zone
56
In the Links tab (Figure 16) it is possible to analyze the relationships (if any) between the users in the
IPED also allows us to perform searches using regular expressions. For example, in Figure 17, a search
is performed for JPEG files, which contain the letter P, followed by one or more numbers. Note that
TYPE is one of the columns displayed in the interface. It is possible to use the information displayed on
the columns of the interface to perform filters like the one in Figure 17.
57
IPED searches both the name and the content of the file, including unallocated spaces. If we type just
the first part of the expression before the AND operator, all files with a JPG extension will be displayed.
If we leave just the last part of the expression (after the AND operator) all files that in their name or
content have a string of characters that contain the letter P followed by one or more numbers will be
displayed.
It is possible to tag the files of interest found during the investigation to facilitate the analysis and
generation of the expert report. Suppose you want to tag the JPEG files that have been deleted (Figure
18). To do this, we must select all the displayed files; right-click on one of them and select the Manage
Next, the screen shown in Figure 20 will be displayed. Let’s create a Bookmark called Deleted JPG Files
58
Figure 20 – Bookmark Creation
Right after closing the window, notice that in the lower left corner of the IPED interface you could see
To generate a report, just right-click on any file and select the option Create Indexed Report. A window
similar to Figure 22 should be displayed and you should select the bookmarks that you want to list in
the report and the folder where the report will be created. After that, a processing window shows up.
59
When the processing is finished, the file report.htm will be created in the root of the folder indicated as
the output folder. The information in the Case Information section, as well as other parameters of the
report (Figure 23), can be edited in the file HTMLReportConfig.txt located in the folder iped-
<version>\profiles\en\default\conf.
Figure 23 – Report
Final Considerations
IPED is a flexible tool that should be considered by any digital investigator as an option for processing
digital evidence. Despite the lack of beauty of its interface, the performance during the processing of a
forensic image file and the speed of searching for evidence is much better than any other similar tool.
Now that you know a little bit about IPED you have to try it and create your own opinion. Have fun!
60
About the Author
HTCIA (High Technology Crime Investigation Association) and ISSA (Information Systems Security
marcusfabio
61
Freely Distributed Open
Source Image Forensics
Tools
by James A (Jim) McCoy, Jr.
I have learned not to trust the internet, especially not photos. Social media is littered with fake and
altered images. I have seen legitimate news sources duped by fake images. Images are often
“shopped” for harmless reasons; just for fun, or to generate attention on social media. The
problems come when fake or altered images are used for malicious purposes.
I like to use freely distributed Open Source applications designed to examine images. They can be
I like to start by looking at the metadata to learn what I am looking at before I look at it. Actually, if
someone gives me an image to examine, I ask them for details before even looking at the metadata. I
get the WWWWW&H from them, then begin to look at it myself. My examination begins with the
metadata.
Simply stated, metadata is data about data or information about the image that is recorded by the
camera. It is a set of data describing and providing information about rights and administration of the
image that is transported with the image file, in a way that can be understood by other software and
62
The pixels of image files are created by automated capture from cameras or scanners. Metadata is
Internally – embedded in the image file, formats include JPEG, DNG, PNG, TIFF …
Externally – outside the image file in a digital asset management system (DAM) (XMP data) or
Descriptive – information about the visual content. May include headline, caption, keywords. Can
Rights – identification of the creator, copyright information, credits, and underlying rights in the
Administrative – creation date, location, instructions for users, job identifiers, etc.
It is essential that the metadata stored in an image file stays with the image. Metadata is critical for
identification and copyright protection. Metadata is also key to smoothing workflow, easily finding
digital images via search – online or offline – and tracking image usage. It is useful in authenticating an
image in that it can serve to confirm that the image is what the creator says that it is or cast doubt on
the authenticity. If the image creator claims that the photo was taken by them on a beach in Barbados,
but the metadata shows the location as New Jersey, then there is a problem.
The problem with metadata is that it can be manipulated. The location of that photo of Barbados that is
actually a beach in New Jersey can be changed, leaving the investigator searching for clues to fakery in
the image.
As I said before, authenticating an image can be as simple or complicated as the level of sophistication
of the creator. If the creator of the image is unfamiliar with metadata then they will not have the ability
to modify it. A more sophisticated creator may know how to manipulate metadata, posing a challenge
for the e-forensic investigator. If the metadata is manipulated, all is not necessarily lost; there can be
other facets of the image that disagree with the metadata that can cast doubt on the authenticity. The
metadata supports the image but the image should also support the metadata.
63
Let's take a look at an example of metadata and a couple of tools used for reading it.
The volume of information contained in the metadata (date captured, date taken, date original, date
created, date edited, date modified, date digitized, date accessed, etc.) can be overwhelming. What’s
with all of those dates? Those dates are where you will find clues to fakery. The dates are not the only
storytellers, but they are the items that most “move the needle,” so to speak.
Now that we are familiar with the metadata that can verify the date, time, location, and other
supporting documentation, and we know that it can be edited, how do we examine, evaluate, and
I like Open Source, freely distributed tools. In part, because they are free of cost but also because they
are not proprietary, and there is usually good support from the user community. I always look for Open
Some of the available tools are made for editing metadata, but, naturally, they can display the data in
The investigator should look at the metadata to see that it supports the claims made by the image
creator. Does the date match when the creator says that the photo was taken? Does the geo location
support the stated location? If not, then we have some doubts cast upon the authenticity. Even if the
64
metadata does appear to support the claims, we still don't stop there. We want to see if the photo
Before examining the image using the tools, the investigator should look closely for signs of obvious
fraud and deception. Is there something that looks out of place? Is there a person or object in the
photo without a shadow? Does it appear to be a different time of day or season than told by the
metadata? Does the weather report from the alleged day of the photo say that it rained but the ground
is dry? Look closely at the image, write a detailed description of the contents, then use your tools to
Again, I first looked for Open Source tools to examine the image. I also like simple things. If I can find
something that is easy to use and quickly identify a fake, I will use that for my initial examination.
I like to use forensically.com, a set of free tools for digital image forensics that includes clone detection,
error level analysis, meta data extraction and more. I see it as a free and easy way to take a quick look
at an image. With forensically, I can quickly debunk an image, or at least begin to cast doubt on its
authenticity but I may still need to take a closer look at it. There are more sophisticated online and
downloadable tools.
http://fotoforensics.com is another online tool that can be used to examine details of a photo including
metadata and image analysis. This could help identify areas of an image that have been manipulated
65
These are two online tools. I may not want to upload a file to server or not have access at that time, so I
For this article, I decided to take a look at JPEGsnoop, a freely distributed tool for Windows.
JPEGsnoop examines and decodes details of JPEG, MotionJPEG, AVI and Photoshop files, as well as
analyze the source the authenticity of the image. There was no installation, as it is a downloadable .exe
file. I especially like that when I want to test something out or use it to take a quick look.
Another helpful feature is the internal database that compares an image against compression
signatures, identifying the digital camera or software that was likely used to generate the image. This
feature is extremely useful in determining whether or not a photo has been edited or tampered with. If
the compression signature matches Photoshop, then it is a safe bet that the photo is not original.
JPEGsnoop reports a large amount of information, that includes quantization table matrix (chrominance
and luminance), chroma subsampling, JPEG Quality, JPEG resolution, Huffman tables, EXIF metadata,
The app does what the web site says it does, quickly and easily.
Most tools have the same basic functionality, it is best for each investigator to experiment with tools
and decide which they like working with the best. User interfaces and intuitiveness may vary, some may
be more complicated to install. I do not go looking for new tools every day, but I read and periodically
do a Google search for the top rated tools. If something is new or improved, I download a copy and try
66
it. I am always using the best available tools. This is another reason why I like Open Source tools, I don't
James (Jim) McCoy, Jr - I have been living and teaching in China for nine
interests in terrorism grew out of my MA studies and the partial PhD studies that pursued in
Homeland Security. Technology has always had a place in my heart and I found a way to combine my
interests in technology and terrorism. I am currently studying Cybersecurity, Ethical Hacking, and
Cyberforensics. As I continue to teach, I am developing an online business that includes writing and
partner Claire and our two daughters YoYo and Li Li (age 2 ½ and 5).
67
Malware - The
Nightmare Time
by Wilson Mendes
The evolution of technology and the rapid growth of the internet have brought the threat of cyber
attacks to private and public infrastructure. Words like backdoors, spyware, worm, keylogger, trojan,
miners, botnet, rootkit, and ransomware have permanently entered to general use.
With the arrival of the Internet of Things, we have many more devices available, significantly
increasing connections to access information; watches, refrigerators, lamps, smart homes, smart TVs
and various devices that are already part of our daily lives, bringing great changes to our lives,
increasing the ability of cybercriminals to act. This has revolutionized the way organizations do
business.
Introduction
Governments, the financial sector, the public sector, large corporations, electric and energy companies,
the data centers of hundreds of segments around the world are at this very moment compromised.
A data breach is more and more common in our lives, it is not difficult to find a lot of information about
the company, about you, about me, about us. In the internet space, we have a very big opportunity to
get a credential.
68
Data centers in hundreds of social segments around the world are currently compromised through
security breaches.
Many countries invest in APT, hacker armies with advanced technological capability, sponsored by the
• national defense,
• intellectual property,
• military plans,
• manufacturing,
69
• government agencies,
• business organizations,
• foreign embassies,
• telecommunications companies,
• universities,
• multi-system infringement,
• healthcare organizations,
• businessman,
Some of the most famous APT (Advanced Persistent Threat) groups in the world:
APT/Titanium,
Carbanak,
Equation Group,
Grupo Lazarus,
APT28/Fancy Bear,
APT29/Cozy Bear,
APT34, APT37/Reaper,
These cyber attacks that make use of malicious software, also known as malware, represent a great
challenge for people and institutions. Advanced zero-day vulnerability exploits, social engineering,
targeted spear phishing, and cyber espionage are some of the techniques used by these groups
The United States, Canada, United Kingdom, Russia, Norway, Ukraine, Mexico, Germany, France, Brazil,
England and dozens of other countries have already detected the fingerprints of these groups with the
70
aim of stealing information since companies located in these countries deal with confidential and high-
value information.
What Is Malware?
Malware is a broad term that refers to different types of programming code; it can take the form of an
executable, which performs malicious actions with intent to steal, encrypting your data and selling your
personal information. For example: viruses, trojan, scareware horses, trojans, spyware, rootkits and
worms.
Usually, it enters your system without your consent. These cyber attacks use malware to steal
confidential information. That's why without the knowledge, skills, and tools needed to analyze
malicious software, it's nearly impossible to recognize the symptoms of an infected device. These
specialties are essential for detecting, investigating, and defending against these attacks. These cyber
attacks are undoubtedly on the rise, and they focus on extracting valuable information.
While performing malware analysis, you will often come across various types of malicious programs;
some of these malicious programs are categorized based on their functionality and attack vectors as
mentioned here:
Rootkit: Malware that provides the attacker with privileged access to the infected system and
Trojan: Malware that disguises itself as a regular program to trick users to install it on their systems.
Once installed, it can perform malicious actions such as stealing sensitive data, uploading files to the
Virus or Worm: Malware that is capable of copying itself and spreading to other computers. A virus
needs user intervention, whereas a worm can spread without user intervention.
Botnet: This is a group of computers infected with the same malware (called bots), waiting to receive
instructions from the command-and-control server controlled by the attacker. The attacker can then
71
issue a command to these bots, which can perform malicious activities such as DDOS attacks or
Backdoor/Remote Access Trojan (RAT): This is a type of Trojan that enables the attacker to gain
Adware: Malware that presents unwanted advertisements (ads) to the user. They usually get
delivered via free downloads and can forcibly install software on your system.
Information stealer: Malware designed to steal sensitive data such as banking credentials or typed
keystrokes from the infected system. Some examples of these malicious programs include key loggers,
Ransomware: Malware that holds the system for ransom by locking users out of their computer or by
Malware classification based on its definitions is not always possible because a single malware
containing multiple functionalities may include a worm component, a backdoor and a ransomware for
example, discarding some of these components after successful exploitation of the target.
• VirusShare: https://virusshare.com,
• VirusBay: https://beta.virusbay.io,
• MalwareBazaar: https://bazaar.abuse.ch/browse/,
• Vx-underground: https://vx-underground.org/samples.html,
• AVCaesar: https://avcaesar.malware.lu,
• Malwr: https://malwr.com,
• TheZoo: http://thezoo.morirt.com.
72
Searching and scanning online for the suspect file with multiple anti-virus and malware scanning
service:
• VirusTotal: http://www.virustotal.com,
• VirSCAN: http://www.virscan.org,
• Metadefender: https://www.metadefender.com.
So that better defenses can be built and protect an organization's network, malware analysis involves,
through reverse engineering, analysis of suspicious binaries to identify their characteristics and
functionality, determining identifiable patterns through different techniques of how the system was
compromised and impacted and can be used to cure and prevent future infections.
These techniques include dynamic analysis, fingerprinting the malware, code analysis, file type, memory
and static analysis, determining malware's target architecture, extracting strings, functions, and
metadata, obfuscation techniques, classifying and comparing the malware samples, and can reveal
different information about the attack and the context around the suspicious file.
Malware developers can modify code using obfuscation techniques to circumvent these detections
because they are always looking for new ways to attack systems whose vulnerabilities have been
patched. Remember that the engines rely on signatures and heuristics to detect malicious files.
The program Pegasus, a famous spyware whose sale is not illegal, developed by the Israeli company
NSO Group, and whose main customers are governments around the world, allows you to capture
audio and video calls, text messages, and photos, contact a network, activate the microphone and
camera, as well as remotely allowing full control of the device, sending the collected files on a schedule
to its servers without being detected. Facebook has denounced NSO for using its platform to distribute
The owner of Amazon and The Washington Post, Jeff Bezos, was invited in 2018 by the Crown Prince of
Saudi Arabia, Mohammed bin Salman, to a dinner in Los Angeles. On the night of the dinner, Jeff
73
Bezos and Mohammed Salman exchanged phones and, moments later, Prince Mohammed started
According to an investigation contracted by the entrepreneur himself, a few weeks later, Jeff Bezos
received an MP4 video file, which showed a photogram with a Saudi and Swedish flag and text overlaid
Hours later, his iPhone, which on a daily average consumed kbytes of data, after receiving the video for
allegedly containing some type of malicious code that managed it remotely, behaved strangely,
sending data thousands of times more than usual, reaching at more than 126Mb daily traffic, stabilizing
The details are that the newspaper, The Washington Post, owned by Bezos, became very interested in
the case, publishing investigative articles by the famous dissident Saudi journalist Jamal Khashoggi,
This does not mean that the message was sent from Prince bin Salman's device, as the accounts of this
application are linked to a telephone number, which may be contained in other devices controlled by
other people. Please, for a better understanding, read the article: Mobile Service Breach https://
eforensicsmag.com/product/osint-and-social-media-forensics/.
Emails, messages, photos, videos, audio and all kinds of files that were transiting that smartphone
started to be monitored in an unwanted way. The spying continued until 2019, and data output reached
4.6 gigabytes. Hackers had gained access to the files and applications of the richest man in the world.
Weeks after the murder of Khashoggi, November 2018, Jeff Bezos receives, according to the UN
report, through the Saudi prince’s account, a photo of his alleged lover Lauren Sanchez accompanied
by a message: “Arguing with a woman is like reading a software license agreement. In the end you
At the same time Jeff Bezos was negotiating a secret divorce settlement with his ex-wife, released
months later by the American tabloid National Enquirer. Jeff Bezos accused the tabloid of attempted
extortion by threatening him with posting sexual photos and messages captured from his device.
74
Newspaper
RedCurl uses hacking techniques similar to well-known groups such as Red October and CloudAtlas.
These would have stolen confidential corporate documents including contracts, financial documents,
employee records and legal records they employ various techniques to cover up their activities
US renews program that allows indiscriminate spying outside its borders. An authorization that allows
you, without the need for permission from companies such as Google or AT&T, to intercept from emails
to phone calls outside the US, even if the authors of the messages are talking to a US citizen. Hackers
from the DarkSide group forced the temporary shutdown of the Colonial Pipeline network, which
Political espionage virus denounced by WhatsApp was used in Brazil. In 2018, researchers found a
digital trace of a mechanism sold by an Israeli company to governments as a weapon of war. APT29 is
using malware known as WellMess and WellMail to target various organizations worldwide. Secret
report links North Korea to cyber attack WannaCry - The US National Security Agency links the
Pyongyang regime to the virus. North Korea was behind the WannaCry virus, which affected 300,000
Government of President Rafael Correa of Ecuador spied on politicians and journalists to control
opponents. The Intelligence Service had individualized files on surveillance of politicians, journalists,
businessmen, indigenous groups, tweeters and even members of its own team. TikTok Geopolitics,
privacy and security flaws motivate campaign against the Chinese app. Citing national security reasons,
Former Colombian Army chief ordered illegal wiretaps of politicians, judges and journalists. Semana
magazine reveals that General Nicacio Martínez's departure from the Armed Forces took place when
the espionage was discovered. A platform called Invisible Man, which was bought from a Spanish
company, could access WhatsApp and Telegram Web conversations, and deleted photos and
North Korean 'hackers' steal South Korea's plans for an eventual war. The documents include details of
a plot to assassinate Kim Jong-un, according to a South Korean lawmaker. Germany suffers the biggest
attack of 'hackers' in its history. Personal data of more than 400 politicians was leaked. Only the far-right
75
Alternative for Germany party was saved from the invasion, which also affects Chancellor Angela
Merkel.
Spanish justice investigates the movements of Russian spies in Catalonia. Police find the presence of
members of an elite military unit specializing in destabilization operations in Europe. UK accuses Russia
of trying to steal data on the coronavirus vaccine. Several organizations, including some investigating a
drug against the virus, were attacked with malicious 'software' to obtain information.
Chinese 'hackers' stole information on Spanish vaccine against covid-19. US court accuses Beijing of
cyber attacks in 11 countries, including Spain. Hacker group leaks stolen NSA spy programs. Former
agency analyst Edward Snowden points to Russia as responsible for the attack.
Mexican government declares secrecy on contracts for the acquisition of software used to spy on
journalists. The Attorney General has ruled that data on alleged contracts with the security firm NSO
Group cannot be consulted until 2021. Nearly 13,000 programs bypass Android's permissions to collect
data from cell phone owners. Thousands of applications circumvent limitations and spy, even without
Note: The author of the article and the eForensics Magazine are not responsible for any damages
Final considerations
Cybersecurity is an ongoing activity that requires constant updates and changes in the behavior of
technology use. We are facing a cyber arms race, with lower barriers to entry, in which any nation can
With the constant growth of data leakage and commercialization by the billion-dollar industry of data
brokers, multiplied by the fragility in the use of secure unencrypted communications and the failed laws
to protect the privacy of users, governments, people, companies, hospitals, and justice systems, any
device connected to the internet is vulnerable, running the risk of being spied on.
The information you store on a smartphone these days is bigger and more valuable than the
information you store on your personal computer. In the world of the internet, a profitable business for
76
Malware analysis requires deep knowledge with mandatory skills in different subjects. Learning malware
analysis takes a lot of time and patience. The malicious and increasingly sophisticated use of emerging
and disruptive technologies continually makes the search for cybersecurity professionals to combat
Facebook, Instagram, Gmail, Tik Tok, Pokemon go, WhatsApp and dozens of other apps can remotely
activate our cameras, microphones, access our photos, read messages, determine our geolocation…
Noble reader, what would be the difference between these applications accepted and used by billions
www.wicasame.com
LinkedIn https://www.linkedin.com/in/wilson-mendes-72161016/
77
Reversing Malware
Anonymously
by Anderson Sales
Anonymity on the internet is important for many different purposes, and malware analysis is one of
them. The malware analyst mindset before starting the activity should be: you are the perimeter to
be protected!
During your analysis, it is not good for you to directly interact with the malware; caution must be
used in anonymization as the threat actor behind an active campaign may collect your personal
information, such as geolocation, operating system and others. Check below for a list with the most
used techniques, remembering that depending on the scenario you will need to use them all at the
same time.
Windows x Linux
There are situations where, during hands-on analysis, the malware code escapes from the browser
sandbox; after all, one of the malware’s goals is to create a file on disk or code execution in memory, so
To mitigate this problem, if the malware does not need to run and the analysis performed is non-
dynamic, it is recommended that if the artifact is a Linux malware, use Windows operating system and
vice versa, except for some exceptions: for example, dynamic and interactive analysis of a malware
script that uses PowerShell must be performed in the Windows operating system because it is
78
Virtual Machines
Configure the virtual machine's network interface in host-only mode and then renew the IP using DHCP.
Another important tactic is to hide your public IP. There are enterprise solutions (NordVPN, AirVPN and
others) that cover this functionality, but it is possible to do it using opensource solutions.
The most used operating system for malware analysis is Remnux and it has the TOR tool by default,
which is one of the most popular for anonymous browsing on the internet.
Terminal + TOR
TOR alone is not enough to start reverse malware activities, because there are situations where you
need to use commands via terminal: for these cases, tools like Torify and Proxychains are used (both are
79
This query returns your new information about public IP, geolocation, connection country, etc.
Proxychains is the most used for this need (terminal + TOR), its usability is similar to torify. The
command below automatically opens Firefox using the TOR network by default.
Example:
#proxychains firefox
IP Leak
It is important to verify that the DNS is not leaking your operating system and other network
• https://www.dnsleaktest.com,
• https://ipleak.net.
When browsing malicious websites, if the page requests your geolocation through popup, don't click
the "Allow location Access" button, otherwise your real location will be discovered.
80
User agent and language spoof
A very important (and very difficult to do) task is attribution; in other words, figuring out who's attacking
you and assigning that threat actor to the attack. But just as it is difficult to attribute an attack, we can
also hide our personal information to make it difficult to attribute the malware investigation to our real
person.
One of these ways is checking browser, user agent and language details as they each provide clues as
to who your opponent is, and this technique can be used to our advantage. There are some Firefox
plugins that make this kind of change, like “User Agent Switcher Switcher” and “Language Switch”.
For interactions using wget command, it is possible to change manually by typing the parameter "--
user-agent":
https://developers.whatismybrowser.com/useragents/explore/operating_system_name/windows/
For convenience, and to keep from having to type "wget --user-agent" every time you need to
use it, Remnux also has a configuration file called “wgetrc” that allows you to change the user agent
81
HTTP header spoof
There are situations in which the malware uses information that is in the header to communicate with
C2 or check some data that is in the cookie, etc.; in the HTTP headers of a web page there is important
information. To analyze this kind of information, the curl command is used to save a page or collect
82
We can see that the requested page is redirecting (HTTP/1.1 301 Moved Permanently) and to which
Remnux also has a configuration file that allows you to change the default curl parameters
(~/.curlrc) to hide your personal information during analysis. This is important because there are
situations where it is necessary to automate some tasks, such as handling digital evidence that uses the
HTTP header, creating a script and applying it in an information collection flow that uses the SOAR tool,
doing a first response and checking a lot of websites at the same time.
OSINT Malware
This is the process that uses open source information to understand and analyze malware. From OSINT,
it is possible to discover (if the malware has already been submitted by other analysts) information such
as the reputation of the IP and/or URL that is hosting the artifact, kind of C2, malware environment, etc.
If someone has already done the analysis, you just check the analysis, you don't have to do the analysis
Below is a list of websites about OSINT to find out more about malware threats:
• https://www.urlvoid.com,
• https://global.sitesafety.trendmicro.com,
• https://zeltser.com/lookup-malicious-websites/,
• https://www.virustotal.com.
83
WHOIS
Why is collecting WHOIS information important? Because usually, malware campaigns have domains
that have recently been registered; they are new domains, that is, newly created domains are highly
Automater - This tool will search several sites like Ipvoid.com, Robtex, alienvault. VT, VxVault,
Online Sandboxes
The sandboxes that currently exist have a high level of maturity in their automated analyses, but there is
• https://any.run,
• https://www.hybrid-analysis.com,
• https://www.joesandbox.com.
Conclusion
Every malware has a threat actor behind it who is monitoring the campaign closely and, nowadays,
malware is a profitable source of money for criminals. When you take down or disrupt a campaign, you
are causing a financial loss and that is bad for the malware business; that's why the malware analyst's
mindset should always be focused on anonymity before starting the hands-on; in addition, your actions
84
About the Author
Master's Degree Student, has most 15 years experience with SOC/MSS and
85
Cyber-Savvy Board
Of Directors
by Longinus Timochenco
Dear Readers,
with your permission, before we start reading, I like to reinforce a few things so that we can start from
It is important to highlight that Information Security is everyone's responsibility. Security is part of our
virtual education, respecting limits and preserving the integrity and availability of information!
86
It is important to emphasize the need for digital security in our lives today, we cannot respond only with
IT-Technology, the issue must be dealt with in a more comprehensive and in-depth manner. Let's not
Current security is high level and strategic for predictive and real time performance. Think about it
responsibly.
Security must exist to protect us and not limit and intimidate us, but for that we have to collaborate with
It is high time for information security to stop saying “bit & byte” and for boards to just talk about
business.
Digital is embedded in all aspects of the business. Cyber risk affects the entire organization: it impacts
business activities at all levels and can be a factor in other relevant risks such as operational, regulatory
and reputational. The concern with cyber security and privacy and data protection must be on the
87
But, truth be told, we need to improve our understanding of information security, and this is not just a
task for senior management. Expert teams must know how to explain the impact of security breaches
For this, strengthen “Periodic Education and Governance”, it will translate security for those who make
strategic decisions in the business and frame this dynamic issue in an appropriate way.
Cyber security is a way to protect people and companies against cyber attacks, which take advantage
of digital vulnerabilities to invade, steal and manipulate data or files, but in my view we can no longer
guide or target our efforts and visions only in IT, believing that our problems will be solved by
technology. Our vulnerabilities are much deeper, broader and more sensitive. I believe we must quickly
make this an important topic on the agenda of your daily strategies. How can you think about
are no longer talking about options, but the need for survival of your business and even lives, adopt a
88
Cyber Security x Information Security
In summary:
Cyber security: involves prevention and protection acting only in cyber space, that is, one connected to
the internet or to networks that link one computer (and other devices) to another.
Information security: involves the prevention and protection against all types of risk, whether physical or
digital, controlling people's access to places, file access permissions, among others.
Be prepared and step forward to face the C-Level with equality because now the CISO/BISO is part of
this decision chain, and if applied, they can add a lot of value to the business at the right time:
• Clear and Objective Goal: "The business is the most important” -> The board does not want to know
the technical details of cybersecurity, but the real situation of your company's business risk.
• Simple language -> The board will never hear technical language, you need to be able to present an
objective and clear language. Ex: "Our security controls ensure a high level of compliance and
• Car metrics -> CISO needs to focus on strategic and business-relevant metrics for a better
understanding of everyone and their degree of importance.
• Adverse inquiries, prepare -> Executives are smart to ask tough questions, especially when they're
cornered. Ex: "What is my risk of loss now in $?" “How many customers can be affected?”
Cybersecurity has become a more frequent topic on the board, but how should you report failures?
Analysis carried out during Infosecurity Europe points out a great difficulty for Security Officers to face
the executive board to report losses, attacks and information leaks. According to the survey, more than
half of information security professionals, Security Officers, prefer to face a dental root canal treatment
than report failures to the executive board. The good news is that cybersecurity has become a more
In the last 12 or 18 months, due to several global security incidents, the issue of cybersecurity has
attracted more attention from the executive board. The regulations of some sectors, such as the
financial sector, for a long time guide and make demands on the sector, requiring greater executive
89
supervision. However, other sectors, such as health, are beginning to move to safeguard their
customers' information.
However, the Security Officer's difficulty in reporting failures and threats without being classified as
“terrorist” is still notorious, so how can you report effectively without causing general panic? Be
consultative and become necessary to the board for decision making, CISOs make a difference and add
The tip from the InfoSecurity Magazine website, which the author of this news has already experienced,
is to create a periodic schedule of reports and status updates, with the most relevant reporting points
Let's give an example: reporting to the board how many calls have been handled by the access control
team; it is no more than area control numbers, but reporting the history of answered calls,
demonstrating a growth curve, compared to the amount of operational resources in FTEs that work in
access control can demonstrate productivity gains or the need for investment in automation or team
For this reason, the strategy of creating a monthly report, which reaches the executive's desk, with
metrics carefully structured and thought out under the executive vision, can create a means of showing
trends that may generate impacts in the future or even demonstrate the actions of analysis of risk and
demands for decision-making by the board to assume risk or release investments to reduce the level of
The company's objective is to make a profit for its shareholders and not invest in security.
Let's understand that we never had and will never have a priority for investments in the security area,
not even in the financial sector. Investments are made based on the need for regulatory compliance
and financial, productive gains or operational risk reduction for the company. Thus, it is imperative that
the security professional be strategic in his interaction with the board, understanding their objectives
and seeking to adapt their language to the executives who will be their interlocutors. Periodic,
consistent and continuous communication is necessary, so that the Security Officer is not called only in
times of crisis and problems, when the spirits, patience and willingness to deal with the matter are the
90
worst possible. In these moments, the Security Officer will certainly be seen as a hindrance and not as a
Conclusion
You already understand that cybersecurity must be a top priority in the company, as it ensures that your
Therefore, it is essential to be open, holistic, and flexible to new Cyber Security management models to
defend information, periodically assess strengths, weaknesses, opportunities and threats, be bold and
don't be afraid to take a stand, but structure yourself. Raise the maturity of your organization on the
Cyber Security theme. I've been through huge organizations and I can guarantee that with good
Dear readers,
I hope I have contributed and that this way we can increase the maturity of our companies and that we
make a difference!
91
Reflection readers
I recommend that the industry launch its products and solutions mainly technological, reinforcing the
item "digital education". We know that this raises the cost of the product, but minimizes mass losses for
everyone.
Together we can provide Convenience with Security, I guarantee that we will increase the difficulties for
I strongly recommend that you strengthen the physical and logical security boundaries in this order and
1. PEOPLE;
2. PROCESSES;
3. TECHNOLOGY;
4. INTELLIGENCE.
Every day we are seeing more cases of data leaks, major losses and threats to humanity and our
business. So I ask you, how long are we going to allow it? We must unite to fight cyber warfare in an
organized and intelligent way, otherwise we will see great damage and loss.
We must tell the world what level of security we want in our lives, think about it and turn it into attitude.
92
About the Author
Longinus Timochenco
Contact:
LinkedIn https://www.linkedin.com/in/longinustimochenco/
e-mail – [email protected]
93
Malware Analysis
And Its Forms
by Daniele Giomo
Malware Analysis is the process of studying how a particular malware works and its side effects. It is
divided into two approaches that are useful for better understanding the functioning of the malware
being analyzed, namely Static Analysis and Dynamic Analysis.
• Before going into the details of the two analysis methods, it is necessary to specify some of the most
common types of malware:
• RAT (Remote Access Tool): a type of malware that create a virtual back door on the machine, which
• Spyware: used to collect information from the system on which they are installed and to transmit it
to the person concerned. The information obtained may vary according to the attacker's needs.
• Adware: software that presents advertising messages to the user during use, for a reduced or no
price. They can cause damage, such as PC slowdowns, and privacy risks as they communicate
• Backdoor/bot: programs that allow unauthorized access to the system they are running on. Typically,
they spread in combination with a Trojan or worm, or they constitute a form of legitimate emergency
access to a system, inserted to allow, for example, the recovery of a forgotten password.
• Rootkits: consist of a driver and sometimes modified copies of programs normally present in the
system. Rootkits are not harmful in themselves, but have the function of hiding, both from the user
94
and from anti-virus programs, the presence of particular files or system settings. They are, therefore,
• Ransomware: Virus that encrypts all data on a disk, according to a complex encryption key; to
obtain it and decrypt the computer, you have to pay the attacker who infected the PC and then
• Cryptojacking: malware that aims to steal the computing power of the CPU, to mine
cryptocurrencies.
The purpose of malware analysis is to study its functioning, extract its characteristics and write rules for
its detection. When a malware analysis is performed, it will be unreadable to the human eye. To
understand its behavior, it will be necessary to use specific tools and some tricks in order to identify the
necessary information.
Static Analysis
Analysis consists of examining the file without running it. Static analysis can confirm whether a file is
malicious, provide information on its functionality, and sometimes provide information that will allow
simple signatures to be produced. Basically, static analysis is simple and can be quick, but it is largely
ineffective against sophisticated malware and can miss important behavior. It also consists of the
process of reverse engineering the malware code, loading the file into a disassembler and looking at
the program instruction by instruction in order to find out how it works. The instructions are executed
by the CPU, so a static analysis will tell you exactly what the program does.
• Inspecting the file format: The metadata of the file can provide very useful information, for example,
Windows PE (portable executable) can provide a lot of information about compile time, imported and
• Extraction of strings: This refers to examining the output that the software generates and information
regarding the operations done by the malware.
• Fingerprinting: This step includes hashes of the encrypted file, find artifacts, such as file name or log
strings.
95
• AV scanning: If the file being scanned carries with it the signatures of a family of malware already
scanned, or it is known malware, many antivirus will be able to detect it.
• Disassembly: In this phase, reverse engineering is performed on the assembly code of the file in order
to deduce the logic and intentions of the analyzed file.
The steps listed above can be carried out through the use of some useful tools for this type of analysis,
such as:
• PEStudio: a tool that allows the static analysis of the file being analyzed, extracting a lot of useful
information, such as the signature and the format. It allows the analysis with VirusTotal and
extrapolates all the strings present in the file, which is very useful.
It supports numerous executable file formats for different processors and operating systems. IDA
provides simplified tools to support reverse engineering activities, obtaining information on cross-
references (XREFs) between the various sections, on the parameters of API calls, etc. This tool is
96
Dynamic Analysis
Dynamic analysis involves the execution of the malware and the observation of its behavior on the
system in order to understand what it does. Before you can run the malware safely, you need to set up
an environment that allows you to perform this type of analysis, without the risk of damaging your
system or network. Dynamic analysis may also include using a debugger to examine the running status
of the malicious file. This type of analysis allows the extraction of further detailed information about the
malicious file that is difficult to extract with other techniques. Some tools used during a dynamic
• Process Monitor: an advanced monitoring tool for Windows that shows the real-time file system, log
and process/thread activity. This tool combines the functionality of two legacy Sysinternals utilities,
Filemon and Regmon, allows filtering of various processes, complete properties of events such as
session IDs and usernames, reliable process information, complete thread stack with integrated
97
• Wireshark: an open source sniffer, a packet capture tool that intercepts and logs network traffic.
Wireshark provides visualization, packet flow analysis, and in-depth analysis of individual packets.
It can be used to analyze internal networks and network usage, application debugging problems and
study protocols in action. But it can also be used to capture passwords, reverse engineer network
98
In addition to these tools, there are also automatic dynamic analysis systems, called Sandbox, which
allow the execution of the sample on a machine in order to monitor in detail the behavior of the sample
• Cuckoo: an advanced, modular, automated malware analysis system and is open source. Basically, it
allows you to analyze different types of malicious files (executables, Office documents, pdf files, e-
mails, etc.). Tracks API calls and general behavior of the file being parsed and extracts information
and signatures. Downloads and analyzes network traffic, even if encrypted with SSL/TLS. It also allows
• Any.run: an interactive online service for static and dynamic analysis of malware, it is very similar to
the service made available with Cuckoo, the only difference is that this service has a free version and
a full paid version of all the services necessary for a detailed analysis.
Malware Detection
Malware detection includes various techniques that are used with the aim of detecting malware and
preventing infection of the computer system, protecting it from potential information loss or system
compromise.
99
Here are some techniques used for malware detection:
• Signature-based detection: The most common method used in malware detection is signature-based
detection. A signature is a string of bits that uniquely identifies a specific virus. Signature-based
detection scans files on your computer and cross-checks their contents with a dictionary of virus
signatures. With this method the antivirus software can quarantine, repair or delete the infected file
and fix the virus problem. However, this technique can identify a small subset of emerging threats but
• Anomaly Detection: Responsible for observing any unusual or potentially harmful behavior. Anomaly
detection can be described as an alert for strange system behavior. This process can be performed in
two phases: the training phase and the discovery phase. First, the software is trained to learn normal
host behavior, while for the second, the trained model is used to recognize any unexpected or
unusual behaviors that occur in the system. One of the advantages of anomaly detection is that it may
• Change Detection/Integrity Checking: The main purpose of the integrity check is to detect any
changes in the files that may have been caused by some malware. To achieve this, all files on the
system are hashed and monitored periodically. If a file has been modified, the calculated hash will not
Machine Learning
Machine learning is a branch of computer science that is also linked to artificial intelligence (AI), it deals
with creating systems that learn or improve performance based on the data that is given to it to
process. In this way, the algorithm can behave accordingly in the face of situations never encountered
before. In the context of malware detection, an example could be a new malware sample never
analyzed before where its actions are therefore unknown. It was initially possible, and sometimes still is,
to use the file signature as a method of detecting whether it is a malicious file or not. Unfortunately, this
is no longer enough as the attackers have started developing malware capable of modifying their
signature using the polyformation technique, or using packing, thus making the above analysis useless.
To deal with these new threats, it was necessary to move to a more complex analysis, which can also
include the use of machine learning, thus analyzing the behavior of the malware during execution,
100
looking for malicious behavioral patterns: file changes suspicious hosts, registry keys and connections.
The main advantage of this method is the ability to identify if a sample is malicious or not, but also,
through specific clustering techniques, the family to which the malware might belong.
In order to build a malware detector with machine learning, the following steps are generally followed:
• Collection of malware samples used to train the system to recognize future malware.
• Extraction of certain characteristics for each sample in order to make the detection system more
precise.
• Training of the machine learning system to recognize malware through the extracted functionalities.
• Verification through tests on some samples not present in the original reference dataset in order to
see if they are effective or not.
In machine learning, there are two main algorithms: supervised and unsupervised machine learning
algorithms, their difference lies in the way they approach the data made available to them to arrive at a
Supervised machine learning: a machine learning technique that has the purpose of instructing a
computer system in such a way as to allow it to automatically elaborate predictions or conclusions with
respect to an input, based on a series of examples initially provided consisting of pairs of input and
output.
Unsupervised machine learning: a machine learning technique which consists of providing the
computer system with a series of inputs that the machine will reclassify and organize on the basis of
common characteristics in order to try to make conclusions and predictions on subsequent inputs.
Unlike supervised learning, uncategorized examples are provided as this step must be performed
Training
In order to implement a model that can detect malware and non-malicious files, it is first necessary to
collect malware and benign samples in order to have several examples to prevent any false positives.
Thanks to the opensource community, finding these samples is quite simple, the malware dataset can
be found online or through the Virus Total paid platform. Then we move on to extracting the features,
101
which depend on the format of the file you are using and the type you can get. Finally, we move on to
machine learning, dividing the dataset into two parts (malicious and not malicious), one for training and
the other that will be used later to train the model and test how efficient it is in detecting malicious files.
Based on the results obtained, it will be evaluated whether it is appropriate to add features or remove
them.
systems analysis since 2007, when I worked closely with the Italian
CAS (Certificate of Advanced Studies) in Advanced Digital Forensics at the SUPSI University of
Manno (Switzerland).
102
A Systematic Literature Review
For OSINT (Open Source
Intelligence) Evaluation As A Tool
To Aid Combating Cybercrime
by Francisco de Assis F. Henriques
Information has great importance in modern society. In this context, we have technological
developments and information from the use of this technology, impacting our society globally and
characterizing it as an information society, making information systems fundamental for its
operation. This universe of content and environments is subject to a growing increase in people and
cyber fraud.
The objective of this article was to highlight, through a systematic review, the use of OSINT as a
necessary instrument to support security/intelligence agents in the fight against digital crimes.
In parallel with other authors, OSINT fonts are legally accessible by the public without violating any
copyright or privacy laws and distinguished from other forms of intelligence for that reason. That's
why they are considered "publicly available", this allows the use of OSINT to go beyond security
services.
Introduction
According to Hassan and Hijazi (2018), society has been transformed by the advent of the internet with
billions of people communicating and exchanging information. The author makes it clear that the
benefits of the digital age have brought different types of risks. Malicious actors, like terrorist groups,
103
What exists in open sources can compose a large body of knowledge for a specific situation. With
structured methods of research in open sources, the process of digital intelligence emerges, a process
that uses all technological, digital, telematic and signal interception means to obtain data and analyze
1. Can OSINT be used as a tool to support security forces in cases involving Cyber Crimes?
1.1. For which functions does OSINT apply within the context of research in Computer Forensics?
1.2. What are the opportunities in the area envisaged in the study?
In this context, this work proposal aims to present the concepts of OSINT and how it is used to support
The research method adopted in this work was the systematic literature review, which consists of
gathering evidence from previously published material, consisting mainly of books and journal articles
Research, from the point of view of objectives, can be: exploratory, descriptive and explanatory. The
present study is characterized as a descriptive research, whose main objective is to “describe the
variables. It involves the use of standardized data collection techniques: questionnaire and systematic
This article is organized as follows: section “Summary”, where we will address the general objective,
exposing the research methodology and describing the review method, with its criteria.
The “Introduction” section will provide a contextualization of all the topics covered. The “Theoretical
Framework” presents the concepts relevant to the theme, OSINT concepts and its role in Intelligence,
the historical context of cybersecurity and OSINT tools that help investigations and fight against
cybercrime.
104
Then, in “Research Methodology” we will show the entire protocol followed for the elaboration of the
systematic review. The results of the review, together with a detailed analysis and discussion of each
research question, share space in the “Results and Discussion” section, where we will discuss the scope
of this systematic literature review, as well as point out new research to be explored on the use of the
OSINT in the acquisition of intelligence against cybercrime, and the "Final Considerations" section that
Theoretical Framework
The systematic review of the literature shows us that the researcher seeks up-to-date sources on
debates related to the field of knowledge studied. The theoretical framework that guides this article
addresses the concepts of OSINT and references research that addresses the methodologies for using
OSINT in collecting information, as well as other works that will serve to demonstrate the efficiency of
OSINT in helping security forces to combat crimes, as well as concepts used by researchers in the
creation of methodologies and frameworks that use the information obtained by OSINT as a basis for
their work.
Cybersecurity Ventures estimates the global costs of cybercrime to grow 15% per year over the next
five years, reaching $10.5 trillion annually by 2025, prompting governments to invest in developing
open source intelligence (OSINT) tools and techniques to combat these crimes. (MINDSECBLOG, 2021)
As shown by Barreto and Wendt (2020:2), the numerous data and information available are not used by
public security operators, more specifically intelligence and police investigation agents. Although
elaborated in a coordinated way with all the knowledge phases elaborated, due to some lack of
knowledge of the security agent in the collection and search of the data, the knowledge is generated in
According to Silva and Menezes (2005:38), the literature review/bibliographic research will contribute
to: obtain information about the current situation of the researched topic or problem; knowing existing
publications on the topic and aspects that have already been addressed; check for similar and different
opinions about the topic or aspects related to the research topic or problem.
105
OSINT - Basic Concepts
The concept of OSINT is broad, its applicability will depend on how each researcher uses open sources
for queries on a given subject. Source is "any data or knowledge that is of interest to the intelligence or
investigation professional for the production of knowledge" (BARRETO and WENDT, 2020:4).
Second Evangelist et al., (2020), the concept of OSINT is a concept that addresses the search,
collection, processing, analysis and use of information from open sources that can be legally accessed
by any individual or organization. As Cepik assures us (2003:32), the OSINT concept is analysis based
on “legally obtaining official documents without security restrictions, direct and non-clandestine
observation of political, military and economic aspects of the internal life of other countries or targets,
monitoring the media, legal acquisition of books and specialized magazines of technical-scientific
character”.
OSINT includes all public accesses for obtaining information, including: the Internet, traditional media,
specialized newspapers or geospatial information. It can be said that open sources have been around
for many years, but with the explosion of the internet there have been many cybersecurity professionals
and researchers publishing journals and articles on cybercrime threats, as well as ordinary people
In this context, it is clear that much predictive intelligence can be obtained from public and unclassified
sources. Intelligence data must be collected from different sources, in this way, as advocated by
Yeboah-ofori and Brimicombe (2018:88), the importance of OSINT has become a conflict between the
private sector, the government and the military over how the data intelligence must be collected from
different sources.
Collecting, exploiting and disseminating them correctly and in a timely manner for the purpose of
addressing specific intelligence requirements has been a major challenge. IDC Research, in its 2020
survey, reports that the “total amount of digital data created worldwide will reach 44 zetabytes and the
number will increase faster within five years to reach 180 zetabytes in 2025” (DNA DATA STORAGE
ALLIANCE, 2021).
106
As explained above, we conclude that the increase in the number of people using the Internet to do
their jobs and, consequently, the growing volume of digital data, will make online sources the main
source of OSINT for both governments and business corporations in the future.
According to Hassan and Hijazi (2018:10), several authors can benefit from OSINT and their motivations
can be the most diverse. The biggest consumers of OSINT sources are the military departments,
government and government agencies, owing this consumption to the huge technological
development and the widespread use of the Internet around the world.
Governments use OSINT sources for a variety of purposes, from national security to understanding
International organizations use OSINT to protect their supply chain from terrorist groups by analyzing
social media sites and internet messaging apps to predict future terrorist actions.
The authors make it clear that all methodologies have some limitations and challenges, among them
It can be said that OSINT brings great responsibilities to the agent that makes use of its tools, in this
context, it is clear that there are legal concerns in many cases. The most worrying thing, however, is that
there are ways for someone to acquire information through illegal means and how the legal system
should deal with this. Another concern is when some forms of hidden public information are collected
It is important to consider that the benefits of OSINT are in several areas and no one should
underestimate its use. Collecting from open sources does not provide risks when compared to other
forms of intelligence and its cost is much lower when compared to other sources, for example, the use
of spy satellites.
Hasan and Hijazi (2008:341, our translation), shows us the importance of OSINT when they say:
The information age has resulted in an explosive amount of potential sources of intelligence and will
shape the future of OSINT collection. In the intelligence arena, the practice of online data collection to
combat terrorism and solve crime is predicted to increase. In addition, OSINT will continue to offer an
inexpensive method to acquire intelligence on any community around the globe [The information age
has resulted in an explosive amount of potential intelligence sources and will shape the future OSINT
107
gathering. In the intelligence arena, it is predicted that the practice of harvesting online data to counter
terrorism and solve crime will increase. In addition, OSINT will continue to offer a cheap method to
acquire intelligence about any community around the globe. (HASSAN; HIJAZI, 2018:341)].
The author makes it clear that OSINT is the preferred method for obtaining information from agencies
around the world. Importantly, OSINT is not limited to security forces and intelligence services alone,
as civil society.
In this way, OSINT is expected to be increasingly inserted into the daily life of ordinary citizens so that
they can obtain information beyond the most common sources - often with manipulated information -
in order to seek knowledge about how criminals act in the digital world and have access to tools that
The origins of Cybersecurity date back to the 1970s, when in 1977 the US government recognized that
open access to computer systems could lead to security breaches, at that time the proposed federal
computer system protection bill had not passed the Congressional review (KREMLING and PARKER,
2018:57).
According to Lynett (2015), network computing was emerging, because until the late 1980s the internet
as we know it had not yet materialized. Large organizations, especially governments, were starting to
connect computers via telephone lines even though there was no worldwide network. Recognizing this,
people started looking for ways to get into phone lines connected to computers so they could steal
In the 1980s, the film WarGames was released and in 1983, attempts to hacking increased, in part
thanks to its release. In 1987, the Computer Security Law was created to strengthen security measures
The 1990s presents us with the beginning of the Information Security industry. The networks based on
Internet Protocol (IP) shifted the focus to availability, with threats such as viruses and denial-of-service
108
The malicious activity of the Internet morphed into the first decade of the 21st century and financial
gain was seen as a lucrative business. Threats like Code Red, Nimda, among others, began to take
Today's crimes include enhanced identity theft attacks, malware, social engineering and denial of
Cybersecurity professionals need to be able to handle incidents in different areas. Alencar (2010) tells
us that the approach is multidisciplinary and worked by different areas of knowledge such as
Currently, the number of cyber threats is continuously growing and the techniques used for the
development of illicit acts have become increasingly intelligent and advanced. We must understand the
link between cyber attacks by analyzing the relationship of data and the techniques used.
KIM, N. et al., (2018) shows us that OSINT is an invaluable tool for collecting this data when he
proposes in his work Design of a Cyber Threat Information Collection System for Cyber Attack
Correlation a system whose function is to collect the infrastructure attack data from various open data
sources (OSINT) and uses the collected data as an input value to collect more data recursively.
A cyber threat intelligence gathering system was developed and tested based on the structure and
functions of the proposed system. Twelve types of information related to cyberattacks were collected.
About two million data items related to cyber attacks were collected over a one-month data collection
period.
With the new tools and resources available, open source investigation has been an invaluable source of
information for anyone investigating a wide variety of topics and for a wide range of reasons.
OSINT Tools
As Bielska et al. assure us (2020:3), considering that OSINT was once exclusive to intelligence analysts
and national security professionals, it is currently observed that there is a growing performance of
professionals in areas such as journalism, cybersecurity, human rights and advocacy. In recent years,
organizations, human rights activists and journalists have embraced these new tools and resources.
109
Open source research will become a basic part of the work of many researchers, regardless of their
background (HIGGINS, 2016:195). There are a number of OSINT related websites that have a
considerable number of specialized tools to perform open source and different source searches to get
information on a regular basis. The key is knowing how to differentiate the information sought from the
Pedersen (2021:6) tells us that Open Source Intelligence is not a tool, although many excellent tools are
available as data aggregators to facilitate the collection phase. No tool will be able to achieve what a
With the advent of new tools, open source research has become a valuable source of information for
According to Bielska et al., (2020) the right tool can determine if you gather the right information and
that the more tools you have in your portfolio, the more flexible your OSINT capabilities will be. Here
• DuckDuckGo (https://duckduckgo.com/): a tool that gathers results from over 400 sources,
including Yahoo, Bing and Wikipedia.
• Swisscows (https://swisscows.com): Located in Switzerland, the search engine uses its own
private servers and is not dependent on third-party infrastructure, with datacenter protected by
Swiss data privacy laws. privatelee (https://privatelee.com): Search the web and images privately.
• CheckUserNames (https://checkusernames.com/): online tool that can help you find usernames
on 170+ social networks.
• maltego (https://www.maltego.com/): tool for recognition on the internet and that allows you to
obtain the results for the specified target, such as IP, domains, etc.
110
• theHarvester (https://github.com/laramies/theHarvester): Python-based tool to be used in the
early stages of an investigation, taking advan
Methodology
Gil, (2009:17) states that research is a rational and systematic procedure that aims to provide answers to
proposed problems. When we do not have information to solve a proposed problem, research is
required. Scientific research can be classified according to its nature (basic or applied), its objectives
(exploratory, descriptive and explanatory) and its method or approach (qualitative, quantitative or
mixed). The author tells us that descriptive research describes a phenomenon or object of study and
The article is characterized likede scriptive, aiming at the analysis of OSINT as a tool to support
The main objective of descriptive research is “to describe the characteristics of a given population or
Cervo and Bervian (2002) tell us that the bibliographic reference “seeks to know and analyze the
existing cultural or scientific contributions on a certain subject, theme or problem.” It seeks to explain a
This article emerged from a bibliographic survey on open source intelligence and its role in research
and studies published in newspapers and specialized magazines. We sought to carry out a quantitative
The objective of this systematic review is to demonstrate through primary studies how OSINT is used as
a support tool in solving cyber crimes. In this way, to complete, we intend to answer the main research
question:
1. Can OSINT be used as a support tool for security forces in cases involving Cyber Crimes?
1.1. For which functions does OSINT apply within the context of Computer Forensics and
Cybersecurity research?
111
1.2. What are the opportunities in the area envisaged in the study?
Dieste et al. (2009) tells us that the systematic review identifies relevant empirical studies based on a
search strategy. An adequate strategy must be defined to detect relevant empirical studies involving
several decisions: selecting the appropriate information sources (i.e. bibliographic databases or digital
libraries), selecting the article fields in which to search for the terms, defining the research to identify
The first step consisted of determining the keywords to search for related works. The keywords defined
were: “OSINT”, “Digital Forensics” and “Threat Intelligence”. The OR and AND operators, used
respectively for synonymous terms and alternative terms for each keyword, were defined. From there,
the following result was obtained: string generic search query: (OSINT) OR (OSINT AND "DIGITAL
The search string was adjusted to suit the characteristics of each electronic database. Systematic
literature searches were performed to find relevant studies based on the following databases:
• Science@Direct (http://www.sciencedirect.com);
• Scopus (http://www.scopus.com).
The choice of electronic databases was guided by the study by Dieste et al. (2009), which defines some
criteria as available primary studies, relevant conferences in the research area, search for studies in
The selection process of the works was elaborated with the help of the tool parsifal (https://parsif.al).
That, according to definitions ofAbout Parsifal, (2021) is: „an online tool designed to support
researchers in carrying out systematic literature reviews in the context of Software Engineering.
Geographically distributed researchers can work together in a shared workspace, designing the
112
protocol and conducting the research. In addition to providing a way to document the entire process,
the tool will help you remember what is important during a systematic literature review.”
The selected studies addressed aspects of using OSINT as a tool or methodology for collecting
information about digital crimes, systems developed with OSINT methodologies, as well as studies that
The initial selection of articles was made by analyzing the title of the work, abstract and keywords. After
the initial selection, inclusion and exclusion criteria were applied, in order to extract the works of
interest at the end. The following inclusion criteria for the works were defined:
• Inclusion criterion 4: Studies dealing with OSINT in Digital Forensics. The exclusion criteria for the
works were defined as follows:
➡Duplicate studies;
After string search has been adapted, according to the characteristics of each electronic database
consulted, 487 publications were found in the study selection stage. Of the 487 papers obtained from
four digital libraries published in the years 2017 to 2021, 43 articles were obtained after applying
The result extracted from these electronic databases was imported into the Parsifal tool and presented
113
• IEEE Digital Library: 76 publications;
Of these works, 444 results were disregarded, with 133 duplicate works and 311 works rejected
because they are subjects that do not fit the OSINT theme, studies prior to 2017, studies in languages
other than English and Portuguese - the latter included in the search to evaluate the existence of works
developed in Brazil related to the theme - and Short Papers (publications with less than five pages).
114
The criterion used for the first selection of articles was to search for the occurrence of the OSINT
terminology in the title, abstract and/or keywords of national publications in Journals in the database.
Figure 2 shows the number of publications on the OSINT topic found in the databases of publications
consulted.
After the selection steps, 43 relevant studies were found, as shown in Table 1, below:
115
116
Analyzing the results of Figure 2, we can see that the databases with the highest concentration of
A temporal analysis was carried out to identify the period in which the largest number of publications
on OSINT is found. This was verified in Figure 3 that the highest concentration of publications occurred
in the period between 2018 and 2020, and in 2018 is the largest amount of published works with the
OSINT theme.
117
Figure 3 – Number of articles per year.
In addition to the inclusion and exclusion criteria, an assessment of the quality of the studies was also
carried out through a questionnaire that sought to assess the methodology of the studies, the objective
of the research, practical aspects, limitations of the study and whether the study was cited by other
researchers. After analyzing the studies, a quality assessment list was answered with the following
questions:
1. Does the study provide an experimental model to evaluate the presented framework or
methodology?
3. Did the study carry out a well-described practical experiment to evaluate the proposal?
For each study evaluated, according to Table 2, the following scores were assigned based on the sum
Weight: 0.0.
118
119
Results and Discussion
This section seeks to show the evidence found in the studies that contribute to answering the following
1. Can OSINT be used as a support tool for security forces in cases involving Cyber Crimes?
It was possible to verify that all 43 studies surveyed are related to OSINT and its use as a tool to help
fight cyber crimes; it is possible to perceive methodologies, applications and theoretical knowledge
that aim to increase the use of OSINT as an analysis tool and search for information. Several studies
were found whose main theme is the processing of information to combat cybersecurity threats.
1.1. For which functions does OSINT apply within the context of research in Computer Forensics?
120
Among the studies analyzed, there are several works that develop topics related to research in
computer forensics. Studies raise questions relating to the analysis of the TOR networktwo,
according to Narayanan et al., (2020) and delay et al., (2018). Forensic intelligence is evidenced in
studies by Quick and Choo (2018). Data analysis through OSINT tools to provide behavioral
information on a particular group can provide excellent inputs for behavioral analysis of extremist
1.2. What are the opportunities in the area envisaged in the study?
We are in a society influenced by data and information. The internet has revolutionized the way
information travels, how companies trade and how data is produced. OSINT – Open source
Intelligence – came to explore these revolutions, through the wide range of research sources.
Technologies that allow companies to collect information about competitors impact the corporate
environment. State intelligence technologies that allow knowing more about individuals or
corporations, these and other opportunities make OSINT a broad terrain to be explored with
countless opportunities, as seen in Pellet studies et al. (2019) and Eldridge et al. (2018).
The enormous supply of information covering all areas of humanity's knowledge provides a major
challenge for intelligence activities and the fight against cybercrime. Dependence on quality data
collection generates difficulties that permeate physical, logical and human resources structures.
The results of the study showed that several works are being developed to minimize the difficulties
Final considerations
The scope of this work was limited to highlighting the importance of OSINT from primary studies and
demonstrating the sources of research for future works in the area of intelligence in open sources.
At the end of this work, we consider it to be an attempt to observe and understand the OSINT universe
through the analysis of scientific literature, as well as to bring light to common doubts to people who
121
Constant advances in information technologies change the way data is generated and collected. The
interest of the scientific community in the subject will allow the creation of tools and methodologies
Data from studies of this RSL highlight the importance of OSINT in the fight against cyber crimes and
the immense field of research that can be developed. It is up to each of us to reflect on the paths that
must be traced to expand the questions and hypotheses raised in the field of open data.
References:
• About Parsifal. parsifal, 2021. Available at: <https://parsif.al/about/>. Access on: 27 Oct. 2021.
• ALENCAR, G.D. Strategies for mitigating internal threats. 2010. 137 p. Dissertation (Computer
Science) — Federal University of Pernambuco UFPE, Pernambuco.
• BIELSKA, A. et al. OPEN SOURCE INTELLIGENCE TOOLS AND RESOURCES HANDBOOK 2020.
2020. ed. [Sl]: i-intelligence, 2020.
• BARRETO, A.G.; WENDT, E. Intelligence and Criminal Investigation in open sources. 3. ed. Rio de
Janeiro: Brasport, 2020.
• DAWSON, M.; LIEBLE, M.; ADEBOJE, A. Open Source Intelligence: Performing Data Mining and Link
Analysis to Track Terrorist Activities. in: LATIFI, S. (Org.). Information Technology - New Generations.
Advances in Intelligent Systems and Computing. Cham: Springer International Publishing, 2018, V.
558, p. 159–163.
• DELONG, M. et al. OSINT Analysis of the TOR Foundation. arXiv:1803.05201 [cs], 24 Mar. 2018.
Available at: <http://arxiv.org/abs/1803.05201>. Accessed on: 19 Apr. 2021.
dnastoragealliance.org/dev/wpcontent/uploads/2021/06/ DNA-Data-Storage-Alliance-An-
122
• ELDRIDGE, C.; HOBBS, C.; MORAN, M.Fusing algorithms and analysts: opensource intelligence in
the age of 'Big Data'. Intelligence and National Security, 16 Apr. 2018. v. 33, no. 3, p. 391–406.
• EVANGELIST, J.R.G. et al. Systematic Literature Review to Investigate the Application of Open Source
Intelligence (OSINT) with Artificial Intelligence. Journal of Applied Security Research, 7 May. 2020. p.
• GIL, A.C. How to design research projects. Sao Paulo: Atlas, 2009.
• HASSAN, N.A.; HIJAZI, R. Open Source Intelligence Methods and Tools: A Practical Guide to Online
Intelligence. in: HASSAN, N.A.; HIJAZI, R. (Org.). Open Source Intelligence Methods and Tools: A
• KIM, N. et al. Design of a Cyber Threat Information Collection System for Cyber Attack Correlation.
in: 2018 INTERNATIONAL CONFERENCE ON PLATFORM TECHNOLOGY AND SERVICE
(PLATCON), 2018, Jeju. Electronic Annals... Jeju: IEEE,2018.p. 1–6.Available in: <https://
• KREMLING, J.; PARKER, A.M.S. Cyberspace, cybersecurity, and cybercrime. First Edition ed. Los
Angeles: SAGE Publications, 2018.
• MINDSECBLOG. Cybercrime will cost the world $10.5 trillion annually by 2025. Security Minute,
March 16. 2021. Available at: <https://minutodaseguranca.blog.br/ crime-cibernetico-custara-ao-
• NAKAMURA, E.T.; GEUS, P.L. De. Network security in cooperative environments. São Paulo (SP):
Novatec, 2007.
• NARAYANAN, P.S.; ANI, R.; KING, A.T.L. TorBot: Open Source Intelligence Tool for Dark Web. in:
RANGANATHAN, G.; CHEN, J.; ROCHA, A. (Org.). Inventive Communication and Computational
Technologies. Lecture Notes in Networks and Systems. Singapore: Springer Singapore, 2020, v. 89, p.
187–195.
123
• PELLET, H.; SHIAELES, S.; STAVROU, S. Locating social network users and profiling their movement.
Computers & Security, Mar. 2019. v. 81, p. 49–57. Available at: <https://linkinghub.elsevier.com/
• QUICK, D.; CHOO, K.-K.R. Digital forensic intelligence: Data subsets and Open
• Source Intelligence (DFINT + OSINT): A timely and cohesive mix. Future Generation Computer
Systems, Jan. 2018. v. 78, p. 558–567. Available at: <https:// linkinghub.elsevier.com/retrieve/pii/
• SILVA, E.L.; MENEZES, I.N. Research methodology and dissertation writing. 4. ed. Florianópolis:
UFSC, 2005.
• YEBOAH-OFORI, A.; ALLAN BRIMICOMBE. Cyber Intelligence and OSINT: Developing Mitigation
Techniques Against Cybercrime Threats on Social Media. International Journal of Cyber-Security and
intelligence-and-osintdeveloping-mitigation-techniques-against-cybercrime-threats-on-social-
Brazil.
124
Applying Zero Trust
Principles To Critical
Controls
by Paulo Pereira, PhD
The emergence of the Zero Trust network concept involves the evolution of cyber-attacks,
considering, essentially, that a traditional network cannot keep up with this evolution.
Critical examples
Taking as a first example, last year, the attack on the fuel pipeline in the United States, we now know
what occurred in greater depth: attackers were able to scale privileges on the network because of
administrative compromised credentials (for example, old passwords still in use). That is, at some point,
there was no continuous verification of the credentials of these administrators and users accessing the
network that controls fuel distribution in this region of the United States. In addition, traffic was not
monitored across the entire length of the network (or at least in the most critical segments), which
A second example perfectly illustrates the lack of these security layers provided in a Zero Trust network.
The attack against the City Hall of a mid-sized town, which occurred at the end of August 2021, shows a
scenario in which:
125
• There is no information available for local businesses to know if their registration data was captured
by the attackers (via the Lockbit 2.0 ransomware) and even if they were leaked or if there was a
repayment payment.
• Essential services have been stopped, servers and systems were affected, blocking the issuance of
Invoice and Service Center (which concentrates 90% of the documents of companies and citizens of
the city), showing that network traffic is not properly monitored and that the network of the city has
Contrary to the above attack scenario, considering these two examples described above, a Zero Trust
network considers some principles that are the pillars for deploying security layers in a network (BARTH,
• A traditional network has failures at specific points, either by administrators or on behalf of users and
non-homologated devices accessing the network.
• There is no inspection of traffic passing through network segments. In fact, few companies know what
is going through corporate network traffic.
• There is little flexibility in positioning hosts, making it easier to enumerate and identify connected
devices on this network.
• The network is always considered hostile, therefore, the verification should be continuous, with zero
confidence.
• External and internal threats exist on the network all the time.
• Each device, user, and network flow are authenticated and authorized.
• Policies must be dynamic and calculated from as many data sources as possible.
It is because of these existing security needs in a traditional network that one of the central aspects of a
Zero Trust network is the creation of policies dedicated to existing resources in a network. In other
words, there is a control plan under which accesses, identities, devices, and systems are referred to.
This means that there must be a policy defined where there are these resources are in use. This is called
126
a data plane that follows the control plan and is defined in each resource access or usage situation.
Therefore, the following important points should be followed in the implementation of the Zero Trust
network:
• There should not be general authentication, and access should always be verified and granted with
privilege limitations.
• Micro network segmentation should allow mapping of resources where policies will be defined.
• The company ensures that all its own systems are in their safest possible state.
• All communications are made securely, regardless of the location of the network. Possibly the use of
end-to-end encryption.
• Access to resources is determined by the policy defined in the control plan, including the observable
state of the user, system, and environment.
We can observe a concept that differs frontally from a traditional network architecture: micro
segmentation. In the conceptualization of a Zero Trust network, it is believed that traditional networks
cannot cover the entire perimeter and thus validate all devices and users who try to access the
network's resources. Micro segmentation would divide the network into smaller parts and in these parts
the control plan would be applied. That is, with a clear definition of the critical areas of the network in a
127
Figure 1: Micro Segment for Access to the Financial Server on a Zero Trust Network
The 0-Day log4j vulnerability (CVE-2021-4428) has already been much discussed in articles and other
communication channels. The key to this vulnerability is the possibility for attackers to access
administrative accounts and from these accounts deliver ransomware that will impact servers and
systems and, consequently, enable access to critical data, such as access to medical data, registrations
on refunds, military registrations, among other examples that have already occurred and are properly
reported.
128
Figure 1 above shows a micro network segment in which there are two devices accessing some network
resource, represented by the Finance Department server. The control plan must identify the host, the
user (including administrators’ accounts), and deliberate on authentication on the network with
Currently, the accounts of the donor are configured with a maximum of privileges, and the examples of
recent attacks show that it is time to review this condition by configuring these accounts with the Zero
Trust principles, delimiting the privileges in time and use of those accounts.
References:
• NIST: https://nvd.nist.gov/vuln/detail/CVE-2021-44228.
• MITRE: https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-44228.
Paulo Pereira, PhD - I was born in São Paulo, the big, boring and
and philosophic area. I never work in this area...but, one day, walking
University Nove de Julho (UNINOVE) and I work with forensic analysis and malware analysis (reverse
engineering of malware) as a free consultant. To escape from reality, in my spare time, I go to some
place to practice fly fishing in the rivers that cut through the mountains and I keep going to
129
Significance Of Image
Tools In Digital
Forensics
by Avirup Dutta, Gaurav Kumar Singh
Digital visual media represents one of the most important means for correspondence. Recently, the
unwavering quality of digital visual data has been addressed, because of the simplicity in
duplicating both its origin, or source, and content. The use of image tools in digital forensics is a
new field of research that targets approving the legitimacy of images by recovering data about their
set of experiences. Two primary issues are considered: the distinguishing proof of the imaging
device that caught the image, and the recognition of traces of forgeries. These days, because of the
promising outcomes achieved by early investigations and the continually developing number of
uses, digital image forensics addresses an engaging examination space for many analysts. This
overview is intended for investigating existing tools and giving a view of the past, the present, and
the eventual fate of image tools in digital forensics.
Introduction
Images and recordings have turned into the primary data transporters in this digital period. The
expressive capability of visual media and the simplicity in their securing, dissemination, and capacity is
with the end goal that they are increasingly more taken advantage of to pass on data. As a result,
images and recordings today address a common source of proof, both inconsistent disagreements and
in preliminary gathering of evidence. The least complex video in TV news is regularly acknowledged as
a certificate of the honesty of the detailed news. Essentially, surveillance video recordings can establish
130
Along with a huge number of benefits, the accessibility of digital visual media brings a significant
disadvantage. Image processing specialists can undoubtedly get to and adjust image content, and
hence affect its importance, without leaving visually noticeable traces. Besides, with the spread of
inexpensive, user-friendly, altering tools, the craft of altering and forging visual content is no longer
limited to specialists. As an outcome, the altering of images for malicious intent is currently more
widespread than at any other time. Digital Image Forensics is that branch of multimedia security that,
along with Digital Watermarking, targets differentiating and uncovering malevolent image control.
In July 2010, Malaysian legislator Jeffrey Wong Su En professed to have been knighted by Queen
Elizabeth II, as an acknowledgment for his commitment to the worldwide guide association Médecins
Sans Frontières. An image of him being granted the honor by the Queen of England went with his
assertion, and was discussed in local media (Fig. I.a). When addressed concerning the honor, however,
the British High Commission in Kuala Lumpur clarified that the name of Mr. Wong was not on the list of
knighthood recipients and that the picture conflicted with the typical convention for knighthood
functions. The picture was, finally, demonstrated to be a merge between a unique function photograph
(Fig. I.b) and Mr. Wong's face, which worked to build his popularity.
This sort of episode [1] contributed to making increasingly more problematic the utilization of digital
pictures as proof. [2] An affirmation of their realness is required, before further depending on their
substance. Consequently, two questions concerning the historical backdrop of the picture must be
answered:
The main question is of significant interest when the source of the picture is simply the proof, for
example, at the point when the responsibility for the recording camera is compromised, or when an
accusatory content is such provided that it was recorded by a particular device (for example, video
surveillance). The second inquiry is of more broad interest and can be straightforwardly applied to the
fake knighthood picture case. Addressing those inquiries is somewhat simple when the first picture is
known. On account of the fake knighthood, the basic accessibility of the first picture was adequate to
uncover the fraud. In reasonable cases, however, practically no data can be thought to be known
131
deduced about the first picture. Therefore, specialists need to validate the picture history
independently.
Picture: (I.a.) The doctored picture portraying Jeffrey Wong Su En while getting the honor from Queen Elizabeth and picture (I.b.) the original
image of Ross Brawn getting the Order of the British Empire from the Queen.
Digital Image Forensics (DIF) targets offering tools to help blind examination. This fresh-out-of-the-box
new discipline comes from existing sight and sound security-related examination spaces (for example,
watermarking and steganography) and takes advantage of picture handling and examination
apparatuses to recover data about the historical backdrop of a picture. Two ways of exploration
advance under the name of Digital Image Forensics. The first incorporates strategies that attempt to
reply to question a) by playing out some sort of comparison examination to determine the device that
caught the picture, or possibly to figure out which devices didn't catch it. These techniques will be
grouped in the article under the normal name of picture source device distinguishing proof methods.
Then the second group of strategies points at uncovering hints of semantic control (for example,
Both these fields and the DIF area overall are drawing in a developing interest from established
researchers. The first distributions in the theme concerning "DIF" date back to 2003, although past
work may have been made public a couple of years sooner, for example, Hani Farid's examination on
bicoherence highlights for altering detection [3]. These days, the topic is so well known that the analysts
are beginning to propose strategies to differentiate criminology techniques, taking advantage of their
shortcomings to more readily stow away or fake the control. These examinations address a valuable
132
commitment to the improvement of picture legal sciences, pushing specialists to make an ever-
Application of Image Tools in Image Processing: Digital image processing includes a wide assortment
pressure, spectral assessment, picture examination, versatile separating, and so on. To tackle these
assignments, a ton of algorithms and strategies have been created and are now under review, which
extends traditional and non-old style approaches. Fuzzy logic is a grounded discipline and many
amazing discoveries are accessible in the logical writing concerning fuzzy set hypothesis and related
A successful method for applying fuzzy logic in advanced picture handling is to deal with every pixel of
a picture based on rules applied to a bunch of pixels in its area. Without huge limits, we can accept
these pixels to have a place with a rectangular window, called the fuzzy mask.
It might be utilized to attempt to distinguish slanted lines. Without critical limits, we will accept the
Much of the time it is helpful to manage relative qualities rather than outright ones. From a practical
perspective, relative handling can be executed by accepting as fuzzy factors the contrasts between the
upsides of the pixels in the mask and the worth of the focal pixel. Contingent on the particular idea of
the issue, both relative and outright handling can be utilized in a similar standard: for example, an edge
location issue might require relative handling beforehand and outright handling afterward.
133
General structure of the tool
Exploring different avenues regarding fuzzy processing might require an l& of "test and modify" cycles,
the principle plan objective was to abbreviate the general time needed by the advancement grouping.
For this reason, specific strategies have been taken on to make the altering stage fast and user-
accommodating and to diminish the handling time also. The instrument is written in C language and
includes a local Graphical User Interface (GUI) with pull-down menus and mouse-driven user
connections. Menu-driven activities are diminished to a base, including basically documenting the
board and the choice of altering, test, and code age areas. The aspect N of the fuzzy mask is
3. one consequent;
Application examples
To represent the presence of the instrument for picture handling some models are given. The first tends
to edge detection issues. A notable edge detector that removes the edges of a picture is the Sobel
operator, which is utilized in old-style picture handling. An alternate methodology is presented by fuzzy
handling. For this reason, a 3 X 3 mask has been taken on and a relative-forerunner/outright resulting
handling has been executed. The point of the remarkable standard is to make the pixels of the picture
white that are encircled by pixels of comparative power, and to make the wide range of various ones
dark (the edge pixels). The point of the guidelines is to address pixels whose qualities are excessively
134
Conclusions
The strategies that we looked into in this overview address significant outcomes for media security,
particularly thinking that the issues they tackle were beforehand (nearly) neglected. An enormous
arrangement of tools is currently accessible to explore picture sources and to validate procurement
devices. Among them, tools that investigate design commotion were demonstrated to be promising for
distinguishing even various models of a similar model device. An even bigger number of strategies
have been created to identify picture altering, some of which are likewise ready to confine the
manufactured regions. Notwithstanding these accomplishments, significant difficulties stay open for
Digital Image Legal sciences. A first fundamental concern is the power of the current tools. Aside from
[11], no genuine similar review exists that assesses the real exactness of DIF techniques. This is
fundamental because of the absence of set up benchmarks and public testing information bases. Few
endeavors toward this path [12, 13] have been done; be that as it may, an enormous dataset exhaustive
of various scenes, light, and ecological conditions and assaults is as yet missing for altering discovery,
for example. Given the development of the area, it is sensible to accept that new information bases and
relative investigations will show up soon, also as open contests, like the BOSS test for steganalysis [14].
Such advancement is attractive both for further developing correspondence among analysts and for
A software tool committed to picture handling with fuzzy guidelines has been introduced. The
adequacy of the device depends on an exact plan which has zeroed in on user-invitingness and
intelligence. The handling execution depends on fuzzy derivation systems which have been shown to
have the option to give great outcomes on test pictures. According to an overall perspective, fuzzy
logic addresses an ideal interface between the user and the issue, permitting the idea of handling to be
instinctively indicated through common human-like terms. According to a more explicit perspective,
fuzzy logic offers an extremely clear method for carrying out versatile handling and, what is disturbing, a
novel compelling way to deal with an address by rules an expansive assortment of handling issues.
References:
• [1] Farid H (2006) Digital Doctoring: How to tell the real from the fake. Significance 3(4):162–166;
135
• [2] Photo tampering throughout history: http://www.cs.dartmouth.edu/farid/research/digital
tampering/;
• [3] Farid H (1999) Detecting digital forgeries using bispectral analysis. Technical Report, AIM-1657,
MITAI Memo;
• [4] D. Dubois and H. Prade: "Fuzzy Sets and Systems: Theory and Applications", Academic Press,
New York, 1980;
• [5] F. Russo, P. Russo, S. Broili: "A Graphical Prototyping System for Computer-Aided Development of
Expert Instrumentation Software", Proceedings of IEEE IMTC/91, Atlanta, GA, May 1991, pp.
495-500;
• [6] A. Kandel: "Fuzzy Techniques in Pattern Recognition", Wiley Inter-science, New York, 1982;
• [7] M. Mizumoto, and H. J. Zimmermann: "Comparison of fuzzy reasoning methods", Fuzzy Sets and
Systems, North-Holland Publishing Company, n.8, 1982, pp.253-283;
• [8] L. A. Zadeh: "Outline of a New Approach to the Analysis of Complex Systems and Decision
Processes", IEEE Trans. Systems, Man and Cybernetics, vo1.3, 1973, pp. 28-44;
• [9] Y. F. Li and C. C. Lau: "Development of Fuzzy Algorithms for Servo Systems", IEEE Control
Systems Magazine, April 1989, pp.65-72;
• [11] Shi YQ, Chen C, Chen W (2007) A natural image model approach to splicing detection. ACM
Workshop on Multimedia and Security (ACM MMSEC07). ACM, New York, NY, USA, pp 51–62;
• [12] Gloe T, Bohme R (2010) The Dresden Image Database for benchmarking digital image forensics.
SAC, Sierre;
• [13] Ng T-T, Chang S-F (2004) A data set of authentic and spliced image blocks. Columbia University
technical report, Available: http://www.ee.columbia.edu/trustfoto;
136
About the Authors
Avirup Dutta and Gaurav Kumar Singh - Department of Forensic Science, Chandigarh University,
137