Mohak RR
Mohak RR
1
Declaration
We,
Mohak (2020UCP1025)
Dommari Anitha (2020UCP1070)
Declare that this project titled ”Evolving Strategies in Android Malware Detection:Leveraging
Network Insights” and the work presented in it are our own. I confirm that:
• This project work was done wholly or mainly while in candidature for a B.Tech. De-
gree in the Department of Computer Science and Engineering at Malaviya National
Institute of Technology, Jaipur (MNIT).
• Where any part of this report has previously been submitted for a degree or any other
qualification at MNIT or any other institution, this has been clearly stated. Where we
have consulted the published work of others, this is always clearly attributed. Where
we have quoted from the work of others, the source is always given. With the exception
of such quotations, this project is entirely our own work.
i
Certificate
This is to certify that this B.Tech project titled ”Evolving Strategies in Android Malware De-
tection:Leveraging Network Insights” submitted by Mohak (2020UCP1070) and Dommari
Anitha (2020UCP1070) in partial fulfilment of the requirement for the award of Bachelor of
Technology(CSE) degree as a record of student’s project carried under my supervision and
guidance.
Date
Assistant Professor
Department of Computer Science and Engineering
Malaviya National Institute of Technology
ii
Abstract
iii
Acknowledgements
iv
Contents
Declaration 1
Certificate i
Abstract ii
Acknowledgements iii
Contents iv
List of Figures vi
1 Introduction 1
2 Related Work 2
v
3.4.2 Mean Absolute Loss . . . . . . . . . . . . . . . . . 19
vi
List of Figures
3.1 CICFlowMeter Working . . . . . . . . . . . . . . . . . . . 8
vii
List of Tables
4.1 Performance Metrics for six Algorithms (Macro Average) . . 37
4.2 Performance Metrics for Ensembled predictions (Ma Average) 37
viii
Chapter 1
Introduction
1
Chapter 2
Related Work
2
promising results, supported by a substantial dataset for further research and
evaluation.
Another paper titled ”Machine Learning Model’s for Network Traffic Clas-
sification in Programmable Logic” explores the application of machine learn-
ing models for network traffic classification, particularly focusing on the anal-
ysis of network packet payloads. It highlights the significance of accurate
classification and identification of anomalous payloads for network security.
The study investigates the implementation of several neural network mod-
els, including convolutional neural networks (CNNs), residual neural net-
works (ResNets), autoencoders, and variational autoencoders, on field pro-
grammable gate arrays (FPGAs) such as the Xilinx VC1902.
These models are evaluated for their inference speeds and accuracy in
classifying packet payloads, with a specific emphasis on achieving inference
speeds exceeding 10,000 packets per second. The performance of FPGA
implementations is compared against that of advanced graphics processing
units (GPUs), specifically the NVIDIA V100 and A100.
3
Chapter 3
Proposed methodology and algorithms
4.Data Extraction via ADB: Utilizing ADB once again, we extract the
captured network traffic data from the emulator environment. This data re-
trieval step ensures that we possess a comprehensive dataset containing the
network interactions of each application under scrutiny.
4
5. CICFlowMeter Integration for Flow Extraction: Employing CI-
CFlowMeter, a flow-based network traffic analysis tool, we extract relevant
flow information from the captured traffic data. Flows encapsulate sequences
of packets sharing common attributes, enabling us to discern patterns and
anomalies in application behavior more effectively.
6. Data Storage in CSV Format: The extracted flow data is then stored
in CSV (Comma-Separated Values) format, ensuring compatibility with var-
ious data analysis and machine learning tools. This format enables seamless
integration of the dataset into our analysis pipeline while maintaining flexi-
bility for future modifications and enhancements.
3.2 Architectures
2. TCP Dump Acquisition: TCP dumps are obtained from the installed
applications through the ADB interface. These dumps contain detailed
network traffic information.
5
3. Data Extraction: The TCP dumps are pulled from the device’s SD card
and processed using CICFlow to generate CSV files. This step ensures
the data is in structured format for further analysis.
Methodology Enhancements
To enhance the manual flow labeling process, several improvements have
been made to the existing Flow Labeler architecture:
2. Code Modifications: The Flow Labeler code has been updated to ac-
commodate the new labeling requirements and feature extraction meth-
ods. This ensures compatibility with the revised methodology and facil-
itates accurate flow classification.
6
These metrics provide insights into the distribution and variability
of flow characteristics.
(b) Protocol Specific Features: For each protocol (e.g., HTTP, Telnet,
SMTP), protocol-specific features are generated to capture unique
characteristics associated with different network protocols. For in-
stance, for HTTP flows, features like HTTP method (GET, POST),
status code, and content type are extracted
(c) Flow Rate Features: Flow rate features, such as packets per sec-
ond and bytes per second, are calculated to quantify the rate of data
transfer within each flow. These features provide insights into flow
intensity and help distinguish between normal and anomalous traf-
fic
3.2.2 CICFlowMeter
7
Figure 3.1: CICFlowMeter Working
8
cients between variables in a dataset are displayed in a table called a correla-
tion matrix. It is an essential tool for understanding the relationships among
variables, identifying multicollinearity, and guiding feature selection. The
correlation coefficient quantifies the strength and direction of the linear rela-
tionship between two variables, encompassing values that range from -1 to
1. A direct association is shown by a positive correlation coefficient (as one
variable rises, the other rises as well), whereas an inverse relationship is in-
dicated by a negative correlation coefficient (as one variable rises, the other
falls). There is no linear relationship between the variables, as shown by a
value of 0.
To compute the correlation matrix, a correlation measure such as Pear-
son’s correlation coefficient is used. Pearson’s correlation coefficient, de-
noted by ‘r’, is calculated as follows:
P
((xi − x̄)(yi − ȳ))
r = pP P
(xi − x̄)2 (yi − ȳ)2
where xi and yi are individual data points, x̄ and ȳ are the means of the
P
respective variables, and denotes the summation. Only linear relationships
between variables are sensitive to Pearson’s correlation coefficient.
The correlation matrix is symmetric, with diagonal elements equal to 1,
as the correlation of a variable with itself is always 1. The coefficients of
correlation between two pairs of variables are represented by the off-diagonal
elements.
9
cation, it assigns the majority class label among these neighbors, while for
regression, it calculates the average of their target values.
Various distance metrics can be employed to compute the distance be-
tween data points. For example, the Euclidean distance between two points
’x’ and ’y’ in an n-dimensional space can be computed using the formula:
p
d(x, y) = (x1 − y1 )2 + (x2 − y2 )2 + · · · + (xn − yn )2
3.3.2 AdaBoost
10
(c) Compute the learner’s weight in the ensemble by using the error
rate. A lower error rate results in a higher weight.
(d) Update the weights of the instances in the dataset. occurrences
that were incorrectly identified received heavier weights, while cor-
rectly classified occurrences received lighter weights. This ensures
that misclassified instances have a higher probability of being se-
lected in the next iteration.
(e) Normalize the instance weights so they sum up to 1.
M
!
X
H(x) = sign αm hm (x)
m=1
where M is the number of weak learners, αm is the weight of the m-th
weak learner, and hm (x) is the prediction of the m-th weak learner.
AdaBoost is effective in various problem domains and often achieves high
classification accuracy. It can, however, be sensitive to erratic data and out-
liers as it focuses on correcting misclassifications during the training process.
To perform at the highest level, hyperparameters like the number of iterations
and the type of weak learner must be carefully tuned.
11
1. For each tree in the ensemble:
(a) Select some random subset from the training data with replacement
(known as bootstrapping). This subset is used to train the current
decision tree.
(b) Random Forest introduces additional randomness by randomly se-
lecting a subset of features to consider for each decision tree split.
This random feature selection process helps to decorrelate the trees
within the ensemble, leading to a more robust model.
(c) Grow the decision tree until it reaches its maximum depth or until a
stopping requirement, such as a required minimum of samples per
leaf node, is satisfied.
2. For making predictions, input the new instance to all the decision trees
in the ensemble.
Numerous benefits of Random Forest include its capacity for handling big
datasets, highly dimensional feature spaces, and missing data. It also pro-
vides robust performance with less risk of overfitting compared to individual
decision trees. Moreover, it can be used for feature importance estimation,
as it calculates the average feature impurity decrease across all trees in the
ensemble.
However, Random Forest can be computationally expensive due to the
need to train multiple trees, and it may not be as interpretable as a single de-
cision tree given its ensemble nature. Despite these limitations, Random For-
est remains a popular and a versatile algorithm for various machine-learning
tasks.
12
3.3.4 Logistic Regression
1. Model Training:
2. Prediction:
13
feature distribution can affect performance in real-world datasets. Despite
these limitations, Logistic Regression remains a valuable tool due to its bal-
ance of simplicity, interpretability, and efficiency.
Decision trees are robust and intuitive models utilized for both classification
and regression tasks. They segment the feature space into regions and assign
a class label or forecast a target variable for each region. The ease of under-
standing and interpreting decision trees makes them widely preferred across
diverse domains, including healthcare and finance, where comprehending the
underlying decision process is vital. Key Steps of the Decision Tree Algo-
rithm:
1. Node Splitting:
(a) Decision trees choose the most suitable feature to split the data at
each node by employing a selected criterion, like Gini impurity or
information gain. This process entails assessing each feature’s ca-
pacity to segregate the data into homogeneous classes.
(b) In the case of numerical features, decision trees identify the ideal
split point that maximizes the homogeneity of the child nodes pro-
duced.
2. Recursive Partitioning:
(a) After making a split, decision trees proceed to partition the data re-
cursively into subsets using the chosen feature and split point. This
recursive process continues until a stopping criterion is satisfied,
such as attaining a maximum depth or reaching a minimum number
of samples per leaf..
14
(a) At each terminal node (leaf), decision trees assign a class label de-
termined by the majority class of the instances in that node. For
regression tasks, the leaf nodes may instead hold the mean or me-
dian of the target variable.
(a) Decision trees may undergo pruning to prevent overfitting and im-
prove generalization performance. Pruning involves removing nodes
that do not contribute significantly to reducing impurity or that re-
sult in minor improvements in predictive accuracy.
5. Prediction:
(a) For predicting new instances, decision trees navigate the tree start-
ing from the root node and moving to a leaf node based on the
feature values of the instance. The class label or predicted value
stored in the leaf node is then assigned as the final prediction.
15
its ability to model complex nonlinear relationships in data, MLPs have be-
come a cornerstone in machine learning research and applications.
Key Steps of the Multilayer Perceptron Algorithm:
5. Iteration: Steps 2-4 are repeated iteratively for multiple epochs until
the model converges to a satisfactory solution or a predefined stopping
criterion is met.
Multilayer Perceptron (MLP) neural networks offer remarkable flexi-
bility in modeling complex relationships within data, enabling accurate
predictions across diverse domains. With their ability to learn hierar-
chical representations of features from raw data, MLPs reduce the need
for manual feature engineering and excel in tasks requiring nonlinear
mappings. However, MLPs come with computational demands, espe-
cially in training deep architectures, and are sensitive to hyperparame-
16
ter settings, often requiring meticulous tuning for optimal performance.
Additionally, their black-box nature poses challenges for interpretabil-
ity, hindering insights into model reasoning. Despite these limitations,
MLPs remain indispensable in modern machine learning, powering ad-
vancements in artificial intelligence and data-driven decision-making.
2. Compute Class Means: LDA calculates the mean vector for each class
in the dataset. These class means act as representatives of the data dis-
tribution for each class.
17
5. Feature Transformation: LDA selects the top k eigenvectors correspond-
ing to the largest eigenvalues to create a transformation matrix. This
matrix is then utilized to project the data onto a lower-dimensional sub-
space while maximizing class separability.
The cross-entropy loss function, also known as logistic loss or log loss, is
a common cost function used in machine learning for classification tasks.
It measures the difference between the expected probabilities predicted by
the model and the actual labels of the output. This loss function is particu-
larly useful for training models to classify data into multiple categories.vskip
0.5cm The product of the true class label and the log of the predicted class
probability, expressed as a negative sum over all classes, is the cross-entropy
loss function. During training, it is preferred to keep this value as low as
possible because doing so increases the predictability of the model.
18
N
− N1
P
L= [yi log(pi ) + (1 − yi ) log(1 − pi )]
i=1
where:
L is the cross-entropy loss
N is the number of samples in the dataset
y is the true label (0 or 1)
p is the predicted probability of the positive class
Mean Absolute Loss, often referred to as Mean Absolute Error (MAE), serves
as a metric to evaluate the performance of a regression model. It computes
the average absolute difference between the predicted and actual values of a
dataset, providing insight into the model’s accuracy in predicting continuous
variables.
X
M AE = (1/n) ∗ |yi − ŷi|
where:
n is the total number of data points
yi is the actual value of the i-th data point
ŷi is the predicted value of the i-th data point
19
Chapter 4
Experimental Setup and Result Analysis
2. Handling missing values: The next step was to identify and remove
any instances with null or empty values. This was crucial to ensure
the integrity of the data and prevent any inaccuracies in the subsequent
analysis.
20
4.3 Output Labels
The final output labels for the processed dataset are ’benign’, ’general mal-
ware’, and ’adware’. Each row in the dataset would represent a specific in-
stance of network activity, categorized based on these labels. Here’s a clearer
representation:
These output labels help categorize the network activity captured in the CSV
file into different types, such as benign, general malware, or adware, enabling
further analysis and classification of the data.
4.3.1 Results
We train and test the KNN classifier on our dataset which we generated and
calculated the Accuracy, precision, F1 Score and Recall along with the confu-
sion matrix and Compared with the available dataset from Canadian Institute
for Cybersecurity-AAGM2017 .
1. Accuracy: 0.895
2. Precision: 0.895
3. Recall: 0.895
4. F1-Score: 0.895
21
Figure 4.1: Confusion Matrix -knn
22
2.Random Forest: We train and test the Random Forest classifier on our
dataset which we generated and calculated the Accuracy, precision, F1 Score
and Recall along with the confusion matrix and Compared with the available
dataset from Canadian Institute for Cybersecurity-AAGM2017
1. Accuracy: 0.913
2. Precision: 0.913
3. Recall: 0.913
4. F1-Score: 0.912
23
Figure 4.4: Canadian Institute for Cybersecurity-AAGM2017 (Dataset)
3.Decision tree: We train and test the Decision tree classifier on our
dataset which we generated and calculated the Accuracy, precision, F1 Score
and Recall along with the confusion matrix and Compared with the available
dataset from Canadian Institute for Cybersecurity-AAGM2017
1. Accuracy: 0.920
2. Precision: 0.919
3. Recall: 0.9202
4. F1-Score: 0.912
24
Figure 4.5: Application
25
4. Multi layer perceptron: We train and test the MLP classifier on our
dataset which we generated and calculated the Accuracy, precision, F1 Score
and Recall along with the confusion matrix and Compared with the available
dataset from Canadian Institute for Cybersecurity-AAGM2017
1. Accuracy: 0.829
2. Precision: 0.830
3. Recall: 0.829
4. F1-Score: 0.827
26
Figure 4.8: Canadian Institute for Cybersecurity-AAGM2017 (Dataset)
5. Ada Boost: We train and test the AdaBoost classifier on our dataset
which we generated and calculated the Accuracy, precision, F1 Score and Re-
call along with the confusion matrix and Compared with the available dataset
from Canadian Institute for Cybersecurity-AAGM2017
1. Accuracy: 0.741
2. Precision: 0.745
3. Recall: 0.741
4. F1-Score: 0.731
27
Figure 4.9: Application
28
6. Logistic regreesion: We train and test the Logistic regression classifier
on our dataset which we generated and calculated the Accuracy, precision,
F1 Score and Recall along with the confusion matrix and Compared with the
available dataset from Canadian Institute for Cybersecurity-AAGM2017
1. Accuracy: 0.672
2. Precision: 0.6525
3. Recall: 0.672
4. F1-Score: 0.646
29
Figure 4.12: Canadian Institute for Cybersecurity-AAGM2017 (Dataset)
1. Accuracy: 0.890
2. Precision: 0.890
3. Recall: 0.890
4. F1-Score: 0.889
30
Figure 4.13: Ensembled Confusion Matrix of All models
The main steps involved in this process to classify network packets as benign
or malicious are as follows:
31
min fpktl furg cnt burg cnt
fPktsPerSecond bPktsPerSecond flowBytesPerSecond
flowPktsPerSecond mean fpktl max bpktl
std flowpktl mean bpktl std bpktl
std fpktl min flowpktl mean flowpktl
min bpktl max flowiat duration
fpsh cnt max fiat total fiat
mean fiat min flowiat min fiat
max fpktl max biat bpsh cnt
total bpktl total bpackets total bhlen
total biat std biat total fhlen
total fpackets mean biat max flowpktl
min biat std fiat total fpktl
flow urg flow cwr flow ece
fAvgSegmentSize min seg size forward avgPacketSize
bAvgSegmentSize flow fin init win bytes forward
flow syn init win bytes backward downUpRatio
max active max idle mean active
mean idle min active min idle
mean flowiat flow psh bVarianceDataBytes
flow rst sflow bbytes std active
sflow bpacket std flowiat flow ack
std idle bAvgPacketsPerBulk RRT samples clnt
FFNEPD sflow fpacket Act data pkt forward
bAvgBulkRate fHeaderBytes fAvgPacketsPerBulk
fAvgBytesPerBulk sflow fbytes fAvgBulkRate
Once the features have been extracted, it is essential to identify the most
important ones that contribute significantly to the classification process. To
achieve this, we calculated the correlation matrix of the extracted features.
32
33
Feature selection based on correlation matrix
4.4.1 Application
34
Figure 4.15: Home screen of the application
35
Figure 4.17: counts of different categories
36
Performance Metrics
Algorithm Precision Recall F1-Score Accuracy
Random Forest 0.908 0.88 0.89 0.9135
Decision tree 0.90 0.89 0.90 0.92
Adaboost 0.75 0.64 0.66 0.74
KNN 0.87 0.86 0.86 0.90
LDA 0.55 0.53 0.0.53 0.64
MLP 0.81 0.78 0.79 0.8291
Table 4.1: Performance Metrics for six Algorithms (Macro Average)
Performance Metrics
Algorithm Precision Recall F1-Score Accuracy
Ensemble Predictions 0.88 0.85 0.86 0.8911
Table 4.2: Performance Metrics for Ensembled predictions (Ma Average)
37
Chapter 5
Conclusion and Future Scope
In conclusion, the report delved into the development and evaluation of effec-
tive methods to identify and thwart malicious applications through network
traffic analysis.we successfully examined the performance of three widely-
used machine learning algorithms, namely KNN, Decision tree, random for-
est ,mlp AdaBoost for the detection of malicious activities. To process the
dataset, we employed state-of-the-art tools, such as CICFlowMeter and FlowLa-
beler, which facilitated the proper extraction and labelling of network traffic
features.
Our comprehensive analysis of these algorithms, combined with the ad-
vanced tools utilized, has provided valuable insights into the capabilities and
limitations of network traffic-based techniques for detecting malicious An-
droid applications. However, there are several avenues for future research
and improvement in this area.
1. Future Scope
38
Bibliography
1. Abdul Kadir A, Habibi Lashkari A and Daghmehchi Firoozjaei M. (2024).
Android Operating System. Understanding Cybersecurity on Smart-
phones.
39
9. Kumari A, Dubey R and Sharma I. (2023). ShNP: Shielding Nuclear
Plants from Cyber Attacks Using Artificial Intelligence Techniques 2023
Annual International Conference on Emerging Research Areas: Interna-
tional Conference on Intelligent Systems (AICERA/ICIS).
40
18. Buriya S and Sharma N. (2023). Malware Detection Using 1d Con-
volution with Batch Normalization and L2 Regularization for Android
2023 International Conference on System, Computation, Automation
and Networking (ICSCAN).
41
Mohak-Report.pdf
ORIGINALITY REPORT
16 %
SIMILARITY INDEX
12%
INTERNET SOURCES
7%
PUBLICATIONS
11%
STUDENT PAPERS
PRIMARY SOURCES
1
Submitted to Malaviya National Institute of
Technology
2%
Student Paper
2
Submitted to Higher Education Commission
Pakistan
1%
Student Paper
3
Submitted to Aliah University
Student Paper 1%
4
Submitted to South Bank University
Student Paper 1%
5
Submitted to Aston University
Student Paper 1%
6
link.springer.com
Internet Source 1%
7
Submitted to University of West London
Student Paper <1 %
8
Md Firoz Hasan, Md. Hasan Moon, Dewan
Mamun Raza. "IoT Network Intrusion
<1 %
Detection Using Ensemble Learning
Approach", 2023 14th International
Conference on Computing Communication
and Networking Technologies (ICCCNT), 2023
Publication
9
www.coursehero.com
Internet Source <1 %
10
iq.opengenus.org
Internet Source <1 %
11
www.mdpi.com
Internet Source <1 %
12
Ton Duc Thang University
Publication <1 %
13
centrallibrary.cit.ac.in
Internet Source <1 %
14
ebin.pub
Internet Source <1 %
15
Sujatha M., Jaidhar C.D.. "Machine learning-
based approaches to enhance the soil fertility
<1 %
—A review", Expert Systems with
Applications, 2023
Publication
16
dokumen.pub
Internet Source <1 %
17
Submitted to Jawaharlal Nehru Technological
University
<1 %
Student Paper
18
Submitted to New Jersey Institute of
Technology
<1 %
Student Paper
19
Submitted to University College London
Student Paper <1 %
20
eurchembull.com
Internet Source <1 %
21
Submitted to De Montfort University
Student Paper <1 %
22
hal.univ-lorraine.fr
Internet Source <1 %
23
vdocument.in
Internet Source <1 %
24
Submitted to University of Surrey
Student Paper <1 %
25
litbangpemas.unisla.ac.id
Internet Source <1 %
26
Submitted to International University -
VNUHCM
<1 %
Student Paper
27
Submitted to Liverpool John Moores
University
<1 %
Student Paper
28
www.ijraset.com
Internet Source <1 %
<1 %
29
Arash Habibi Lashkari, Andi Fitriah A.Kadir,
Hugo Gonzalez, Kenneth Fon Mbah, Ali A.
Ghorbani. "Towards a Network-Based
Framework for Android Malware Detection
and Characterization", 2017 15th Annual
Conference on Privacy, Security and Trust
(PST), 2017
Publication
30
Submitted to University of Hertfordshire
Student Paper <1 %
31
eprints.whiterose.ac.uk
Internet Source <1 %
32
www.arxiv-vanity.com
Internet Source <1 %
33
fastercapital.com
Internet Source <1 %
34
Submitted to Multimedia University
Student Paper <1 %
35
Mustapha Habib, Thomas Ohlson Timoudas,
Yiyu Ding, Natasa Nord, Shuqin Chen, Qian
<1 %
Wang. "A hybrid machine learning approach
for the load prediction in the sustainable
transition of district heating networks",
Sustainable Cities and Society, 2023
Publication
36
Submitted to Nanyang Technological
University
<1 %
Student Paper
37
v1.overleaf.com
Internet Source <1 %
38
Submitted to National College of Ireland
Student Paper <1 %
39
Submitted to University of East London
Student Paper <1 %
40
Submitted to Concordia University
Student Paper <1 %
41
Submitted to University of Southern California
Student Paper <1 %
42
www.researchgate.net
Internet Source <1 %
43
coek.info
Internet Source <1 %
44
repository.charlotte.edu
Internet Source <1 %
45
Submitted to CSU Northridge
Student Paper <1 %
46
Wang, Jinzhen. "Toward Smart and Efficient
Scientific Data Management", New Jersey
<1 %
Institute of Technology, 2023
Publication
47
dspace.univ-ouargla.dz
Internet Source <1 %
48
"Computational Science and Its Applications –
ICCSA 2018", Springer Science and Business
<1 %
Media LLC, 2018
Publication
49
Dingyu Yang, Jian Cao, Cheng Yu, Jing Xiao. "A
Multi-step-ahead CPU Load Prediction
<1 %
Approach in Distributed System", 2012
Second International Conference on Cloud
and Green Computing, 2012
Publication
50
Mahendra Prasad, Sachin Tripathi, Keshav
Dahal. "An efficient feature selection based
<1 %
Bayesian and Rough set approach for
intrusion detection", Applied Soft Computing,
2020
Publication
51
Sibi Chakkaravarthy Sethuraman, Tharshith
Goud Jadapalli, Devi Priya Vimala Sudhakaran,
<1 %
Saraju P. Mohanty. "Flow based containerized
honeypot approach for network traffic
analysis: An empirical study", Computer
Science Review, 2023
Publication
52
berlmathges.de
Internet Source <1 %
53
fdocuments.net
Internet Source <1 %
54
wiredspace.wits.ac.za
Internet Source <1 %
55
www.math.mcmaster.ca
Internet Source <1 %
56
Abdullah Al Maruf, Zakaria Masud Ziyad, Md.
Mahmudul Haque, Fahima Khanam. "Emotion
<1 %
Detection from Text and Sentiment Analysis
of Ukraine Russia War using Machine
Learning Technique", International Journal of
Advanced Computer Science and
Applications, 2022
Publication
57
Kasim Tasdemir, Rafiullah Khan, Fahad
Siddiqui, Sakir Sezer, Fatih Kurugollu, Alperen
<1 %
Bolat. "An Investigation of Machine Learning
Algorithms for High-bandwidth SQL Injection
Detection Utilising BlueField-3 DPU
Technology", 2023 IEEE 36th International
System-on-Chip Conference (SOCC), 2023
Publication