PHD0273 Rajshree Srivastava 196204 Cse 2024
PHD0273 Rajshree Srivastava 196204 Cse 2024
DOCTOR OF PHILOSOPHY
By
RAJSHREE SRIVASTAVA
January, 2024
@ Copyright JAYPEE UNIVERSITY OF INFORMATION TECHNOLOGY (Established
under H.P.Legislative Assembly Act No. 14 of 2002 and Approved by UGC under section
2(f)) WAKHNAGHAT, SOLAN, H.P. (INDIA)
January, 2024
ALL RIGHTS RESERVED
SUPERVISOR’S CERTIFICATE
This is to certify that the work in the thesis entitled “Early identification and
classification of thyroid nodule in medical ultrasound images” submitted by
Rajshree Srivastava at Jaypee University of Information Technology,
Waknaghat, India, is a bonafide record of her original work carried out under
my supervision. This work has not been submitted elsewhere for any other degree
or diploma.
(Signature of Supervisor)
Date:
i
DECLARATION BY THE SCHOLAR
I hereby declare that the work reported in the Ph.D. thesis entitled “Early
identification and classification of thyroid nodule in medical ultrasound
images” submitted at Jaypee University of Information Technology,
Waknaghat, India, is an authentic record of my work carried out under the
supervision of Dr. Pardeep Kumar. I have not submitted this work elsewhere for
any other degree or diploma. I am fully responsible for the contents of my Ph.D.
thesis.
Rajshree Srivastava
Department of Computer Science & Engineering and Information
Technology, Jaypee University of Information Technology,
Waknaghat, Solan, India
Date:
ii
ACKNOWLEDGEMENT
With the divine grace of “Almighty God”, the journey of Ph.D comes to end. I would like take
this opportunity to thanks each and every person who has helped me to complete this journey.
I would like to express my gratitude to our Honorable Vice Chancellor Prof. (Dr.) Rajendra
Kumar Sharma, Dean (Academics and Research) Prof. (Dr.) Ashok Kumar Gupta and Head
Department of CSE/IT Prof. (Dr.) Vivek Kumar Sehgal to promote the research and facilitate
resources in the institution. I would also like to thank my doctoral committee members for
their valuable feedback and critical reviews during presentations and time to time help.
I am also grateful for the support received from JUIT, Waknaghat. In particular, I thank, all
staff members at the Department of Computer Science and Engineering and Information
Technology, JUIT, Waknaghat who has been extremely helpful at numerous occasions. I
would like to thank my parents for supporting me spiritually throughout my life. Last but not
the least, I thank my fellow Ph.D. friends for their consistent help and valuable discussion.
iii
ABSTRACT
In medical imaging, machine learning (ML) and deep learning (DL) play a significant role to
predict symptoms of early diseases. DL is one of the growing trends in general data analysis
since 2013. It is an improvement of artificial neural networks (ANN) as it consists of many
hidden layers to permit high level of abstraction from data. In particular, convolutional neural
networks (CNNs) have proven to be a potential tool for computer vision tasks. Deep CNNs
have a capability to automatically learn raw data especially images.
The accurate assessment or identification of disease depends on image interpretation as well as
acquisition. Due to the improvement in last decade in image acquisition, devices acquire data
at high rate with increased resolution. However, the interpretation of images has recently begun
to benefit from computer technology. Mostly these are made by the radiologist, physicians and
senior doctors but limited due to its subjectivity and high skilled physicians/ doctors.
Computerized tools in the medical imaging field are the key enablers to improve diagnosis by
facilitating the findings.
Analysis of thyroid ultrasonography (USG) images via visual inspection and manual
examination for early identification and classification of thyroid nodule has always been
cumbersome. This manual examination of thyroid USG images in order to identify benign and
malignant thyroid nodule can be tedious and time-consuming. Various deep learning models
have emerged in medical field especially in thyroid nodule classification with the rapid
advancement in technology and increase in computational resources. Early identification of
such nodules can improve the effectiveness of clinical interventions and treatments. Therefore,
many researchers now advocate the use of computer diagnostic system (CDS) to objectively
and quantitatively analyze the USG images of thyroid nodules. This helps the radiologists to
solve the differences in interpretation of results.
Incidents of thyroid tumor have dramatically increased in recent years. However, early
ultrasound diagnosis can reduce morbidity and mortality. The work in clinical situations relies
heavily on the subjective experience of the sonographer. Thus, developing of a reliable thyroid
nodule identification model based on ultrasound images is needed to ensure the accuracy and
efficiency of thyroid nodule diagnosis. The key motive of this thesis is to develop efficient
models using ML and DL techniques for thyroid nodule identification and classification. The
first phase of this thesis proposes two hybrid models namely ANN-SVM and CNN-SVM. These
models work in four stages namely data collection, preprocessing, classification using
iv
hybridization of ML and DL classifiers and analysis of results. The second phase of this thesis
proposes an optimized CNN based model using segmentation and boundary detection
techniques. In this work, Alex-Net, visual geometry group (VGG-16), deep neural network
(DNN) and Res-Net-50 models are optimized using grid search optimization (GSO) technique.
The third phase of this thesis proposes a deep-generative adversial network (Deep-GAN) model
to improve the accuracy of the VGG and Alex-Net models. VGG-GAN and Alex-GAN. The
proposed model will be useful to physicians, researchers, practitioner and small hospitals
having limited resources can benefit from using this model to assist with thyroid ultrasound
examination diagnosis.
v
TABLE OF CONTENTS
CONTENT PAGE NO
Supervisor’s Certificate i
Acknowledgement iii
Abstract iv-v
2.5 Summary 25
vi
Chapter 3: ANN-SVM hybrid model for the 26-34
identification and classification of thyroid nodules
3.1.1Data Collection 27
3.1.2 Pre-processing 27
3.1.3 Feature-extraction 27
3.1.4 Classification 28
3.1.4.1 SVM 28
3.1.4.2 KNN 29
3.1.4.3DT 29
3.1.4.4ANN 29
3.1.4.5RF 30
3.1.4.6 Hybridization 30
3.2 Summary 34
3.1.2 Pre-Processing 36
4.1.3.1 CNN 37
4.1.3.2 SVM 40
vii
4.1.3.3 Hybridization of CNN-SVM Model 41
4.1.4 Result Analysis 42
4.2 Summary 45
5.1.2 Pre-Processing 48
5.1.3.2 Segmentation 49
5.1.4.1 Res-Net-50 51
5.1.4.2 Alex-Net 51
5.1.4.3 VGG-16 52
5.1.4.4 DNN 53
viii
6.1.1 Data Collection and Pre-processing 62
6.1.2 Data Augmentation 62
References 74
List of publications 88
ix
LIST OF TABLES
x
5.3 Results of GSO technique obtained on 57
learning rate and dropout factor parameters
on GSO-CNN model, Alex-Net, Res-Net-50,
VGG-16 and DNN
xi
LIST OF FIGURES
Figure Caption Page
Number Number
xii
4.10 The epoch-accuracy graph for Dset-1 45
xiii
6.8(a-d) Training for VGG-16 architecture 69
xiv
LIST OF ACRONYMS & ABBREVIATIONS
AI Artificial Intelligence
AC Active Contour
CT Computed Tomography
D Discriminator
DL Deep Learning
DT Decision Tree
DA Data Augmentation
DRLSE Distance-Regularized-Level-Set-Evolution
xv
ELM Extreme Learning Machine
FP False Positive
FN False Negative
FE Feature Engineering
FD-SMOTE Farthest-Distance-Based-Synthetic-Minority-
Oversampling Technique
GB Giga bytes
G Generator
GM Generative Models
hyp Hyperparameter
LR Logistic Regression
MHz Megahertz
xvi
ML Machine Learning
NN Neural Network
P Probability
RF Random Forest
xvii
TSH Thyroid-stimulating hormone
T4 Thyroxine
T3 Triiodothyronine
TN True Negative
TP True Positive
TA Texture Analysis
USG Ultrasonography
xviii
CHAPTER 1
INTRODUCTION
Disease prediction is important for medical institution as well as for the patients to make
possible medical care decisions. Sometimes, incorrect decision might lead to serious
consequences. Hence, early identification and classification of disease is important. According
to the latest report of American Thyroid Association (ATA), more than 12 percent of the U.S.
population will develop a thyroid condition during their lifetime [1]. Thyroid dysfunction is the
most prevalent endocrine disorder worldwide having an epidemiology of 4.6% in the United
Kingdom, 11% in India and 2% in the U.S among the overall population [2]. Thyroid gland is
a small butterfly shaped endocrine gland located in neck region below Adam’s apple [3]. It
produces three types of hormones namely thyroxine(T4), calcitonin and triiodothyronine (T3)
to maintain the metabolism of the body. The function of these hormones depends on the right
amount of supply of iodine in diet [4]. The ductless gland located at the bottom of our brain
sends signals to secret hormones called as Thyroid Stimulating Hormone (TSH). It informs the
thyroid gland to secret the hormones as per the requirement [5]. Thyroid Nodules (TNDs) are
solid or fluid filled lumps formed within the thyroid region/gland. Figure 1.1 shows an image
of thyroid gland. There are various imaging processes like Ultrasonography (USG) and
Computed Tomography (CT) used for the identification of diseases [6]. USG is highly
recommended for the identification of TNDs as it is cost friendly, radiation free, real-time and
painless.
In last few decades, success of Machine Learning (ML) and Deep Learning (DL) techniques in
image identification/diagnosis tasks leads to solve challenges in diagnostic imaging. ML
techniques have a great potential to deal problems in the field of medicine, drug discovery and
decision making. Medical images are mostly analyzed by the radiologists who are highly
experienced and have knowledge in the respective domain. A delayed or erroneous diagnosis
causes harm to the patients [12]. In ML, learning new features are well described by the patterns
in images. The sparse representation (i.e., feature extraction/selection) in ML techniques finds
informative patterns from a shallow architecture but they lack in its representational power [13].
DL techniques use the feature engineering (FE) step into a learning step [14]. It works by
discovering informative representation in a self-taught manner with some pre-processing steps
without extracting features manually [15]. Thus, the manual feature engineering has been
shifted to computers thereby allowing the non-experts of ML to use DL techniques efficiently
especially in the medical diagnosis filed. Some of the popular reasons lead to success of DL
techniques are: (1) development of advance algorithms [16], (2) availability of large amount of
data [17], (3) easy derivation of hierarchical features from low-level to high-level [18], (4)
improvement in the traditional Artificial Neural Networks (ANN) i.e., increase in multiple layer
[19] and (5) advancement in high-tech processing units [20].
2
1.1 Thyroid Nodule Datasets
This section focuses on the datasets used for the identification and classification of TNDs in
this thesis. The dataset plays significant role while conducting any experiment. It contains
information that helps the model to have a high level of understanding. In this thesis, two
datasets are considered:
The criteria for the inclusion and exclusion of thyroid USG images are as follows:
Inclusion-
• Images having clear boundary of the nodules.
• Images having one or multiple nodules.
• Patients having age greater than 16 years.
• Patients with the confirmed thyroid nodules problem.
Exclusion –
• Excluded the images which were zoomed in or zoomed out.
• Patients having age less than 16 years.
• Patients without confirmed thyroid nodules problem.
3
1.2 Performance Parameters
There are four different parameters considered for evaluation of proposed models in this thesis:
• Accuracy: It is defined as the ratio of correctly classified TNDs to the total number of
examples and computed using equation 1.1:
𝑇𝑃+𝑇𝑁
𝐴𝑐𝑐𝑢𝑟𝑎𝑐𝑦 = ∗ 100 (1.1)
𝑇𝑃+𝑇𝑁+𝐹𝑃+𝐹𝑁
• Specificity: It is defined as the percentage (%) of people who test negative for a specific
disease among a group of people who do not have the disease and calculated using equation
1.2:
𝑇𝑁
𝑆𝑝𝑒𝑐𝑖𝑓𝑖𝑐𝑖𝑡𝑦 = ∗ 100 (1.2)
𝑇𝑁+𝐹𝑃
• Sensitivity: It is defined as the percentage (%) of people who test positive for a disease
that have disease and calculated using equation 1.3:
𝑇𝑃
𝑆𝑒𝑛𝑠𝑖𝑡𝑖𝑣𝑖𝑡𝑦 = ∗ 100 (1.3)
𝑇𝑃+𝐹𝑃
• F-measure: It measures the test accuracy of the model and calculated using equation 1.4:
2𝑇𝑃
𝐹 − 𝑚𝑒𝑎𝑠𝑢𝑟𝑒 = ∗ 100 (1.4)
2𝑇𝑃+𝐹𝑁+𝐹𝑃
4
1.4 Organization of Thesis
Chapter 1: This chapter provides the introduction about thyroid nodule problem along with
its types, datasets used, and performance parameters for the identification and classification
of TNDs.
Chapter 2: This chapter discusses the literature review of different ML and DL techniques
for identification and classification of TNDs.
Chapter 3: This chapter discusses the first part of objective one of the thesis i.e. design of a
hybrid model using hybridization of ML classifiers for the identification and classification
of TNDs.
Chapter 4: This chapter discusses the second part of objective one of the thesis i.e. design
of a hybrid model using hybridization of ML and DL classifier for the identification and
classification of TNDs.
Chapter 5: This chapter elaborates the second objective of the thesis i.e. design of an
optimized DL model for the identification and classification of TNDs.
Chapter 6: This chapter focuses on the third objective of the thesis i.e. design of an improved
DL model using Generative Adversial Network (GAN) for the identification and
classification of TNDs.
Chapter 7: This chapter concludes the thesis and throws light on future perspectives.
5
CHAPTER 2
LITERATURE REVIEW
This chapter describes literature review related to thyroid disorder. The method of
categorization of work is based on techniques related to identification, classification and feature
selection for thyroid identification and detection. For this study, the selected articles are
assorted into three categories namely simple, improved and hybrid techniques. Figure 2.1
shows the assortment of literature review.
6
Figure 2.2: Framework of simple techniques [23]
Zhu et al. proposed ML-based CAD system [25]. Sonographic features are extracted from the
USG images and ANN is used for discriminating USG images. The model has achieved
specificity of 81.80%, accuracy of 83.10% and sensitivity of 83.80%.
Song et al. proposed a Grey Level Co-Occurrence Matrix (GLCM) method for feature
extraction and different ML classifiers for classification [27]. The logistic regression has
achieved a good result from the rest of the classifiers on 155 USG images.
Ardakani et al. evaluated the CAD system with Texture Analysis (TA), Linear Discrimant
Analysis, Principal Component Analysis, (PCA+LDA) for the classification of TNDs [28].
Nugroho et al. used bilateral filters for the pre-processing step results in smoothen images and
also preserves the boundaries of the object followed by Active Contour Without Edges
(ACWE) for the segmentation of the thyroid nodules [29]. The result shows not only
improvement in the accuracy of the segmentation of active contours (AC) but also makes
localization of the TNDs clear.
Song et al. proposed a GLCM based texture feature model using different ML algorithms for
classification [30]. The model was evaluated on 147 USG images. The best accuracy result is
achieved by Support Vector Machine (SVM) with 76.6%, Random Forest (RF) with 74.1%,
ANN with 72.8% and Decision Tree (DT) with 67.1% accuracy.
Pan et al. proposed Principal Component Analysis (PCA) ensemble method for feature
7
reduction and found that it performs better than Bagging and AdaBoost [31].
Vanithamani et al. proposed CAD system using GLCM for feature extraction, segmentation
using fuzzy c- means clustering and classification using SVM on TDID dataset [33].
Nugroho et al. proposed a texture feature extraction using Gray Level Run Length Matrix
(GLRLM), histogram and GLCM. Multi-Level Perceptron (MLP) classifier was considered for
classification [34]. Model is evaluated on small public dataset.
Pasha et al. proposed Ensemble Gain Ratio Feature Selection (EGFS) model [36]. The model
ensemble RF & gain ratio algorithm to find the most efficient features followed by its alignment
with LR, K-Nearest Neighbor (KNN) and Naive Bayes.
Jothi et al. proposed a novel SVM linear+ SVM Quadratic+ Closest-matching CAD system i.e.
(SVM-L+SVM-Q+ CMR) [37]. The proposed CAD system has achieved a good result by
combining the advantages of SVM and CMR classifiers.
Jiang et al. proposed an intelligent model using ML-based techniques [38]. All ML-based
classifiers have achieved 84% accuracy except DT classifier.
Shankar et al. proposed a Multi-Kernel-Based SVM (MKSVM) system for thyroid disease
classification using the UCI and private datasets [39].
Ma et al. proposed a ML-based model (center clustering, deep neural network, SVM, KNN and
LR) for thyroid nodule classification [40]. A total of 508 USG images are used in the
experiment. The highest accuracy achieved is 87%.
8
Zheng et al. proposed the serum Raman spectroscopy technique combined with Artificial Fish
Coupled with Uniform Design (AFUD) –SVM discriminant model capable for the detection of
thyroid dysfunction [41].
Dandan et al. proposed CAD system using Wavelet Multi-Sub-Bands Co-Occurrence Matrix
(WMCM) and other feature extraction techniques [42]. The model is evaluated on 180 USG
images.
Poudel et al. proposed an autoregressive feature for thyroid ultrasound images without
modeling the images [43].
Prochazka et al. proposed a 2-threshold binary decomposition CAD system that uses direction
independent feature set [44]. Authors addressed the problem of direction-independent features
problem. Model was evaluated on 40 USG images and used histogram and segmentation for
classification. The results show that SVM gives good results in comparison with RF.
Wu et al. proposed texture analysis system using Local Binary Pattern (LBP) and Histogram
of Oriented Gradients (HOG) feature extraction techniques [45]. The model was evaluated on
4,574 USG images.
Colakoglu et al. proposed ML-based model to differentiate between BTND and MTND via
extracting 306 features from 235 USG images [46]. Among ML-based algorithms, the RF
classifier gives a good result with an accuracy of 86.8%.
Zhang et al. proposed ML-based thyroid nodule detection using sonographic features and Real-
Time Elastography (RTE) [47]. Nine different ML algorithms like RF, LR, SVM, Neural Net,
Adaptive Boosting, Naive Bayes and Convolutional Neural Network (CNN) were considered
but RF shows the best result with an accuracy of 84.1 % in comparison with other classifiers.
Kesarkar et al. proposed an ANN based CAD system [48]. The ACWE based segmentation
method and MLP for the classification of TNDs images.
Zhao et al. compared various ML classifiers for thyroid nodule detection using sonographic
features [49]. The model shows that the application of ML-based classifiers like DT, SVM,
9
KNN, RF, LR, MLP, Xgboost, Gradient Boosting Tree and Naive Bayes on sonographic
features has achieved the best result. The model was evaluated on 849 USG images.
Table 2.1 shows summarization of literature review using simple techniques.
10
Table 2.1: Summarization of literature review using simple techniques
11
sensitivity 88.89%.
Koundal et Automated ---- Spatial Public dataset (i) It is insensitive to nodule (i)It has strong resistance to
al. (2016) delineation of neutrosophic echogenicity. noise.
thyroid nodules distance (ii) It has achieved 95.45%
regularized level sensitivity.
set
Pasha et al. EGFS model EGFS LR, Naive UCI repository and local (i) It has achieved 96.45% (i) It has strong resistance to
(2017) Bayes, KNN dataset accuracy. noise.
(ii) The model has achieved (ii) Complex data pre-processing
good result with a few is used.
numbers of features.
Mustafa et FD-SMOTE Minimum Genetic UCI repository (i) The model works good (i)Work can be extended on
al. (2017) Technique Description Length algorithm on less dataset. rough theory problem.
Method, PCA (ii) It has achieved 84.12%
accuracy.
Jothi et al. SVM-L+SVM- PSO SVM Public dataset (i) The proposed CAD (i)Size of dataset is less.
(2017) Q+CMR CAD system differentiate PT C
System and NT H&E- stained
thyroid histo-pathology
images.
(ii) It has achieved accuracy
99.54% sensitivity 100%
and specificity 98.57%.
Jiang et al. ML based CAD Sonographic features SVM, KNN, LR, Public dataset (i)The model has achieved (i) Other feature extraction
(2017) system DT, RF 84% accuracy. techniques can be explored.
(ii) Size of the dataset is less.
Zheng et al. AFUD System Wavelet Multi-sub- SVM First Affiliated Hospital (i) It has achieved 82.74% (i)Size of the dataset used is less.
(2018) bands Co-occurrence accuracy.
matrix (ii) The proposed technique
can be used to develop a
portable device for
detecting TNDs function.
Dandan et al. ML based CAD WMCM SVM Public dataset (i)he model has achieved (i) Work can be extended on
(2018) system 87% accuracy. other different types of feature
extraction methods.
(ii) Size of the dataset is less.
Shankar et MKSVM method Auto-regressive SVM Public and UCI (i)It has achieved accuracy (i)High computation time.
al. (2019) features 97.49%, specificity 94.5%
12
and sensitivity 99.05%.
(ii)The model has achieved
good result with a few
numbers of features.
Ma et al. ML based CAD Sonographic features KNN, LR, SVM Public dataset (i)The model has achieved (i) Work can be extended on
(2020) System 87% accuracy. other different types of feature
extraction methods.
(ii) Size of the dataset is less.
Poudel et al. Autoregressive Histogram features, SVM, RF SurgicEye GmbH, (i) The proposed approach (i)Bi-Spectral model can be used
(2019) Feature based SFTA University Clinic of outperforms from DSC and in future work.
CAD model Magdeburg, Germany had similar SE and SP.
(ii) It has achieved 90%
accuracy.
Prochazka et CAD System HOG, LBP SVM, RF Our clinic (i) It uses direction- (i)Work can be extended on
al. (2019) independent features. large dataset.
(ii) It has achieve-d 94.64%
accuracy.
Wu et al. CAD System LBP, GTF, HOG, Second Affiliated (i) Can be used as 2nd (i)High computation time.
(2019) FOS, GLCM, HWF, Hospital of Fujan opinion for diagnosis of
Autoregressive Medical University TNDs.
model, GRLM (ii) It has achieved 88.58%
accuracy.
Colakoglu et ML based CAD Ultrasonography, ML classifiers --- (i)It has achieved accuracy (i)Work can be extended on
al. (2019) system shear wavelet texture 86.8%, sensitivity 85.2% large dataset.
and specificity 87.9%.
Zhang et al. ML based CAD ----- RF, LR, SVM, First Affiliated Hospital (i)The model has achieved (i)Different other feature
(2019) system Neural net, 84.6% accuracy, 84.2% selection techniques can be used
adaptive specificity and 88.1% for better result.
boosting, naive sensitivity.
bayes and CNN
Kesarkar et ANN model for Sonographic features MLP ----- (i)It has achieved 93.84% (i)Size of the dataset is less.
al. (2021) thyroid nodule accuracy, 97.82%
diagnosis sensitivity and 93.75%
specificity.
Zhao et al. ML based EGFS DT, SVM, KNN, ------- (i)It has achieved 88.8% (i)Work can be extended to other
(2021) techniques for RF, LR, MLP, accuracy, 92.9% specificity feature extraction methods.
classification of Xgboost, and 81.7% sensitivity.
thyroid USG gradient
13
images boosting tree and
Naive baye
14
2.2 Improved Techniques
Improved techniques are the DL techniques based on ANN with representation learning. It
uses various layers to progressively extract a higher level of features from the user input [50].
DL techniques applications include audio recognition, medical image analysis, speech
recognition, bioinformatics and Natural Language Processing (NLP) [51]. The term “deep”
means large number of features to be extracted from raw image as input. E.g., lower layers and
higher layers identify edges and letters [52]. DL reduces the cost of feature engineering and
retrieves a lot of information that helps the model to achieve better results [53]. One of the
major differences between ML and DL is that ML cannot extract features automatically; it
requires labelled parameters whereas in the case of DL, it extracts features automatically
without human intervention [54]. Figure 2.3 shows the working of the improved techniques. It
also works in three steps: (1) input, (2) feature extraction and classification and (3)
prediction.
Archarya et al. proposed Gabor transform features from high-resolution TNDs images [56].
The model was evaluated on 240 USG images.
Peng et al. show the first order features selection in TNDs ultrasound image recognition as
image biomarkers [57].
Zhu et al. proposed an image augmentation CAD system using transfer learning technique as
Generative Adversarial Network (GAN) [58].
15
Song et al. proposed a multi-scale DL single-shot detection network method guided by nodule
with the help of a CNN [59].
Zhang et al. proposed a multi-scale single-shot detection network guided by nodule [60]. The
model was evaluated on 45 USG images.
Sundar et al. explored different DL techniques like CNN, VGG-16 model and Inception-V3
model for the diagnosis of TNDSs [61].
Wang et al. proposed a novel EM-based model to train Alex-Net for the diagnosis of the TNDs
[62]. Authors addressed the problem of labeled data and evaluation was done on public and
private datasets.
Chen et.al proposed 2- Level Attention-Based Bi-Directional LSTM (Ab-LSTM) for TNDs
classification [63].
Li et al. proposed a Deep Convolutional Neural Network (DCNN) model for the detection of
TNDs [64]. The result shows good improvement in sensitivity, accuracy and specificity in
comparison with a group of skilled radiologists.
Ko et al. proposed a DCNN for TNDs diagnosis [65]. The model was evaluated on 589 USG
images dully approved by the radiologist and used 3 layers of CNN for the classification of
TNDs.
Wang et al. proposed a DL-based CAD system using YOLOv2 NN [66]. The proposed method
achieved a real-time and synchronous diagnosis CAD system for TNDs diagnosis.
Liu et al. proposed clinical-knowledge-guided CNN model for the TNDs classification [67].
Aboudi et al. proposed a CAD system combining different attributes after optimization to
achieve good accuracy [68].
Guo et al. proposed an improved DL technique for TNDs identification [69]. A new CNN
16
model is proposed which integrates Squeeze and Excitation (SE) and Maximum Retention of
Inter-Pixel Relations Module (CNN-SE-MPR). The model was evaluated on 407 USG images.
Ajilisa et al. used various pre-trained CNN network [70]. K-means clustering is employed to
deal with imbalanced datasets.
Nguyen et al. used a fusion of spatial and Fourier based methods for feature extraction and also
used two CNN models together to improve the accuracy [71].
Vasile et al. proposed an intelligent diagnosis model using ensemble DL methods [72]. Model
was designed with 5 CNN layers and then ensemble with pre-trained VGG-19 model. The
model was evaluated on 230 USG images.
Zhu et al. proposed an automatic DCNN CAD system [73]. The model was evaluated on 719
USG images.
Yang et al. proposed a Multi-Task Cascade DL (MCDLM) model [74]. It works in three steps
namely pre-processing, U-Net model for segmentation of nodules and semi-supervised method
for feature extraction and classification. The model has achieved good result on small dataset.
Table 2.2 shows summarization of literature review using improved techniques.
17
Table 2.2: Summarization of literature review using improved techniques
Article Proposed Feature Classification Dataset Findings Limitations
[Ref.] Model extraction/
Detection
techniques
Du et al. CAD System -- DLRSE Wuhan Tongji (i) It is insensitive to noise and (i)It requires manual initial contour.
(2015) Hospital, China nodule echogenicity.
(ii) Achieved 85% sensitivity.
Acharya CAD System Gabor MLP, SVM, Chiang Mai (i)It has achieved 94.3%. (i)Work can be extended to non-linear feature
et al. Transform KNN University Hospital, extraction algorithms.
(2016) Thailand
Peng et CAD System FOS SVM Ruian People (i)It has achieved specificity (i)Work can also extend on large thyroid USG
al. Hospital, Zhejiang, 93.3%, accuracy 88% and dataset.
(2017) China sensitivity 82.1%.
Zhu et. al DL based --- CNN TDID (i) It can be used as a 2nd opinion (i)Handcrafted and DL model can be used for
(2017) CAD system by the radiologist. better result.
(ii) It has achieved 93.75%
accuracy.
Song et. MC-CNN ---- VGG-16 TDID (i) It has achieved accuracy (i)Sample size is less
al (2019) Framework 92.1%, sensitivity 94.1% and
specificity 96.2%.
(ii) Embedded spatial pyramid
module into traditional CNNs.
Zhang et FAFL Model Multi- --- Local hospital (i) New feature association model (i) Improvement can be done by fusion of two
al. channel is proposed. or more classifiers.
(2018) feature (ii) It has achieved 83% accuracy. (ii) The dataset size is very small.
association
Sundar DL CAD VGG-16 and - Public and private (i) Achieved good result on a (i)The dataset size is very small.
et al. System Inception- smaller number of datasets.
(2018) V3 model (ii) It has achieved accuracy 93%
with Inception-v3 + CNN and
with VGG-16 with 89%.
Wang et Semi- VGG-16 ---- TDID (i)The model solves multi- (i)The dataset size is very small.
al. supervised instance problems and also
(2018) method improves the classification
accuracy.
18
(ii)It has achieved accuracy
80.91%, sensitivity 81.82% and
specificity 80%.
Chen et Ab-LSTM -- CNN Ruijin Hospital, (i) Model perform good. (i)The dataset size is very small.
al. Memory Shanghai (ii) It has achieved 86.18%
(2018) method accuracy.
Li et al. DCNN ---- CNN Traditional Chinese (i) It can be used as 2nd opinion (i)The dataset size is very small.
(2019) Weihai Municipal for diagnosis.
Hospital, Shandong, (ii) It has achieved sensitivity
Tianjin Cancer 93.4% and specificity 86.1%.
Hospital and Western
Medicine Hospital,
Jilin
Ko et al. DCNN --- CNN ----- (i)The model has achieved 86% (i)Sample size is less
(2019) accuracy.
Wang et DL- CAD ---- CNN Tertiary class A (i) Can be used as 2nd opinion for (i)Work can be extended on large database.
al. System hospital the diagnosis.
(2019) (ii) It has achieved sensitivity
90.5% and accuracy 90.31%.
Liu et al. DL- CAD --- CNN Cancer Hospital ofthe (i) The pixel annotation used in (i)Work can be extended on large database.
(2019) System Chinese Academy of this work is coarse.
Medical Sciences (ii) It has achieved accuracy on
dataset-I 97.5% and
accuracy on dataset-II 97.1%.
Aboudi CAD system Statistical, CNN CIM Laboratory of (i) Fusion of (intensity + textural (i)Work can be extended on large database.
et al. GLCM the National features + shape) is used.
(2020) University of (ii) It has achieved 96.13%
Columbia accuracy.
Guo et CNN-SE- --- CNN Co-operate hospitals (i) With the help of MPR module, (i)Work can be extended on big database.
al. MPR System loss of inter pixel relation is
(2020) solved.
(ii) It has achieved accuracy
90.17%, specificity 92.35% and
sensitivity 86.99%.
Ajilisa et CAD System ---- Alexnet, DNN, TDID (i) DNN performs best. (i) Work can be experiment on large database.
al. Inception v-3, (ii) Failed to address the problem of noise.
(2020) Googlenet,
VGG-19,
19
Xception,
Resnet-10,
VGG-16,
Resnet-50,
Nguyen DL based FFT CNN TDID (i) Fusion of Res-Net50 and (i)Work can be extended on large database.
et al. CAD system Inception Model is considered for
(2021) classification.
(ii) It has achieved 92.05%
accuracy.
Vasile et Ensemble ---- CNN Public dataset (i)Ensemble CNN and VGG-19. (i)Work can be extended on large database.
al. model (ii)It has achieved sensitivity
(2021) 95.75%, accuracy 97.35% and
specificity 98.43%.
(iii)It can be used as 2nd opinion
by the physician.
Zhu et al. DL based --- CNN Ethics committee (i) It has achieved 86.5% avg. (i)Hand-crafted feature extraction methods
(2021) model Hospital of Shanghai accuracy, 87.1% specificity and can also be used to improve accuracy.
86.6% sensitivity.
(ii) It can be used as 2nd opinion
by the physician.
Yang et DL CAD Sonographic CNN Partner hospitals (i)Used augmentation technique. (i)Size of the dataset is less.
al. system features (ii)It has achieved 90.01%
(2021) accuracy, 92.15% specificity and
87.47% sensitivity.
20
2.3 Hybrid Techniques
Hybrid techniques are the combination of two or more classifiers to improve performance of
model [75]. It works in three steps namely, (1) data pre-processing, (2) model selection and
(3) prediction. Figure 2.4 shows the working of hybrid techniques.
Xia et al. proposed a hybrid Extreme learning machines (ELM) based approach for TNDs
diagnosis [77]. The model was evaluated on 230 USG images.
Nguyen et al. proposed a model based on fusion of spatial and frequency domain for feature
extraction from USG images and achieved an accuracy of 90.88% [78].
Qin et al. proposed (conventional ultrasound+ elasticity imaging) model for the classification
of TNDs [79].
Xie et al. proposed a (DL +handcrafted features) model and achieved an accuracy of 85% [80].
The model was evaluated on 623 USG images.
Sun et al. proposed a (DL-based technique + statistical) features together for classification of
TNDs [81]. The model was evaluated on 490 USG images.
Hang et al. suggested a Res-GAN model [82]. In this work, authors have compared ResNet-
18 with Res-GAN model and Speeded Up Robust Features (SURF) with deep features along
21
with RF as classifier. The proposed Resnet-Generative Adversial Neural Network (Res-GAN)
model has achieved an accuracy of 95%.
Srivastava R., et al. proposed a hybrid model using ANN and SVM for the classification of
TNDs [83]. The model shows that when two classifiers hybrid together, the accuracy is
increased. Table 2.3 shows summarization of literature review using hybrid techniques.
22
Table 2.3: Summarization of literature review using hybrid techniques
Article [Ref.] Proposed Model Feature extraction/ Classification Dataset Findings Limitations
Detection
techniques
Liu et al. (2017) Hybrid method SIFT, HOG, VLAD, CNN Cancer Hospital of (i) The feature used (i)There is no
LBP the Chinese shows that both hand- comparison with the
Academy crafted features and DL other models.
features contribute to the
thyroid image
classification.
(ii) It has achieved
specificity 94.1%,
accuracy 92.9% and
sensitivity 90.8%.
Xia et al. (2017) Hybrid method Relief method SVM Wenzhou Central (i) It can be used as 2nd (i)Small dataset.
Hospital opinion for TNDs
diagnosis.
(ii) It has achieved
accuracy 87.72%,
sensitivity 78.89% and
specificity 94.55%.
Nguyen et al. (2019) DL-CAD System --- CNN TDID (i) Used the combination (i)Work can be
of spatial and frequency extended on other
domain problem. different type of
(ii) It has achieved feature extraction
90.88% accuracy. method.
Qin et al. (2019) Hybrid method ----- VGG-16 Huiying Medical (i)Use of end-to-end (i)The model is yet
Technology implementation leads to to be evaluated on
(Beijing) Co.Ltd good efficiency result. large dataset.
(ii)It has achieved
accuracy 94.70%,
specificity 92.7% and
sensitivity 97.96%.
23
Xie et al. (2020) Hybrid deep LBP CNN Shanghai Tenth (i) Can be used as 2nd (i)Different other
learning based People’s hospital opinion for the diagnosis. feature methods can
feature method (ii) It has achieved 85% be explored.
accuracy.
Sun et al. (2020) Hybrid method Statistical features CNN First Affiliated (i) A voting system is (i) Different other
Hospital of Nanjing used. feature methods can
Medical University (ii) The model has be explored.
achieved 86.5% (ii)The model is yet
accuracy. to be evaluated on
large dataset.
Hang et al. (2021) Res-GAN model Statistical features ResNet-18, Random TDID (i) The model has (i) Different other
Forest, AdaBoost performed well in less feature methods can
sample size. be explored.
(ii) The Res-GAN model (ii)Different other
performs well in DL techniques can
comparison with ResNet- be explored for the
identification of
18.
TNDs.
(iii) In this work, fusion
of SURF + deep
features+ RF is used for
classifier.
(iv) Accuracy: 95%
Srivastava, R et al. Hybrid model GLCM ANN, SVM, DT, TDID, collected (i) The model has (i) Different other
(2021) KNN, RF performed well in less feature methods can
sample size. be explored.
(ii) The issue of noise in (ii) DL models can
images is addressed. be explored.
(iii) Model shows that
accuracy of the model is
increased in comparison
with standalone classifier
when it is hybrid.
(iv) The model is
evaluated on 2 datasets.
24
2.4 Inferences Drawn from Literature Review
The literature review shows significant research carried out for the identification and
classification of TNDs. However, there are many gaps which are partially unexplored and needs
immediate attention. The following are the inferences drawn from the detailed review:
1. Most of the existing study relies on either ML or DL classifier for the identification and
classification of TNDs. Thus, there is a scope to build hybrid model.
2. This review sheds light on the possible use of segmentation, clustering, boundary and
edge detection techniques to enhance the performance of improved techniques.
3. It is inferred from that improved techniques model require continuous training and
adjustment of hyper parameters. It is still an unresolved issue as it requires complete
domain knowledge.
4. It is evident that less attention has been paid by the researchers for noise removal in
thyroid USG images.
5. This review sheds light on the importance of TNDs disorder. Such thyroid disorder has
also attracted a large number of health care experts subject to availability of USG images
with advancement in medical imaging. Hence, such images are a good inspiration for
the healthcare analysts to identify and detect TNDs disorder.
2.5 Summary
This chapter described the literature review conducted for TNDs identification and
classification. The inferences drawn are inked from the detailed literature review carried out in
this chapter. The coming chapters in the thesis will present effective and efficient models for
the identification and classification of TNDs.
25
CHAPTER 3
ML and DL techniques have been widely used in recent years in medical imaging. USG is one
of the suggested diagnostic methods for accurately differentiating BTND and MTND images.
In this chapter, ANN-SVM hybrid model is proposed for the identification and classification of
TNDs USG images.
3.1.2 Pre-processing
It is an important phase for the development of any model. In this phase, image resizing, noise
removal using median filter and RGB to gray scale conversion are performed to maintain the
uniformity of the dataset. All these steps are computed using equation 3.5-3.7:
where images: the RGB tensor to convert the images, name: name of the operation
The main reason for using median filter is that it removes the noise and also preserves the
edges of the images for the accurate prediction [84]. Benign images are labelled as “0” and
malignant images as “1”.
27
3.1.3.1 Contrast: It measures the local variance in GLCM and can be computed by using
equation 3.8:
𝑁−1
𝑁−1
𝐶𝑜𝑛𝑡𝑟𝑎𝑠𝑡 = ∑ 𝑘2 ∑ (𝑓(𝑎, 𝑏)) (3.8)
𝑎,𝑏=0
𝑘=0
3.1.3.2 Correlation: It computes the combined probability rate of the particular pixel pairs and
can be computed by using equation 3.9:
where µ is mean
3.1.3.3 Homogeneity: It measures the closeness of the distribution of the elements in the
GLCM to the GLCM diagonal using equation 3.10:
3.1.3.4 Energy: It computes the total number of squared elements in the GLCM and can be
computed by equation 3.11:
𝐸𝑛𝑒𝑟𝑔𝑦 = ∑𝑎,𝑏=0
𝑁−1 𝑓(𝑎, 𝑏)2 (3.11)
3.1.4 Classification
Four different types of ML classifiers are considered namely ANN, SVM, DT, KNN and RF.
These are discussed below:
3.1.4.1 SVM
It is one of the frequently used supervised ML classifiers, widely used in image processing
and pattern recognition [87]. In SVM, an optimal hyperplane decides the separation between
individual classes of patterns. It must be maximum so that the problem of underfitting and
overfitting can be reduced [88-89]. Figure 3.2 shows the optimal hyperplane of SVM,
28
C1(diamonds) and C2(circles). Here, SVM tries to fit a linear boundary known as a hyperplane
between two different classes and tries to maximize the distance between the boundary and its
nearest point of each class.
3.1.4.3 DT
Its structure is like a tree where the top most nodes is called as “root node”, leaf node as the
“class label”, branches as result on the test, decision node represent the test of an attribute. One
of the advantages of using DT classifier is its ability to give good classification result but it may
lead to the problem of overfitting [91].
3.1.4.4 ANN
It is inspired by the “nervous system” where nodes act as an artificial neuron and the directed
edges between neurons are defined by weights [92]. Figure 3.3 shows structure of ANN. In this
work, we have used three layered ANN.
29
Figure 3.3: Architecture of ANN [92]
3.1.4.5 RF
It is a kind of ensemble learning technique involving a large number of individual DT. It works
by first selecting the random samples from a given dataset and then constructs DT for each
sample. Thereafter, it performs prediction based on maximum voting. RF does not suffer from
the problem of overfitting and is used for both regression problems and classification [93].
3.1.4.6 Hybridization
An ANN- based 3-layer Feed Forward Neural Network (FFNN) is used along with
hybridization of SVM classifier for the classification of TNDs. A simple ANN having one
hidden layer with 4 neurons in the input layer was created. The neurons present in MLP
network are trained consequently on backpropagation algorithm until an output value is
obtained at each of the output neurons. ANN was trained with the extracted features and the
output of the hidden layer was used to train the SVM classifier. Figure 3.4 shows the ANN-
SVM hybrid model.
30
Figure 3.4 : The ANN-SVM hyrbid model
3.1.5 Result Analysis
This section summarizes the results of the ANN-SVM hybrid model for classification of TNDs.
The work is done on MATLAB 2016b software using intel core i5 8th generation computer.
The k-fold cross validation technique is used to validate the results of the ANN-SVM model.
In this work, 5 cross validation is considered i.e., D= [c1, c2,.……c5], where 4-fold is used for
training and 1-fold is used for testing. Once, the pre-processing phase is over and features are
extracted using GLCM method, various ML algorithms like RF, SVM, KNN, DT and ANN are
employed for the classification. For better representation, public dataset is named as: Dset-1
and collected dataset is named as: Dset-2. Figure 3.5 shows the parameter setting for ANN-
SVM hybrid model. The various parameters considered are batch size, optimizer, etc. Table
3.1 shows the standalone ML classifiers comparison on Dset-1. The ANN classifier has
achieved an accuracy of 79.28%, specificity of82%, F-measure of 80.9% and sensitivity of
77.16% and SVM has achieved an accuracy of 78.4%, specificity of 80.4%, sensitivity of
76.3% and F-measure of 79%. Table 3.2 shows the performance of the standalone ML
classifiers on Dset-2. The ANN classifier has achieved an accuracy of 84.96%, specificity of
85.88%, F-measure of 85.4% and sensitivity of 81.56% andSVM has achieved an accuracy of
82.65%, specificity of 83.50%, sensitivity of 80.12% and F-measure of 84%. From the table 3.2
and table 3.3, it can be seen that ANN performs better in comparison with the rest of the other
classifiers i.e., DT, KNN, RF followed by SVM classifier. Table 3.3 show classification results
of the best two classifiers and the proposed hybrid model on Dset-1 and Dset-2. It can be
analyzed from the table that when ANN and SVM classifiers are hybrid together then the
performance of the model is increased.
31
Thus, it can be said that the proposed ANN-SVM performs better in comparison with the
standalone ANN and SVM classifier. Table 3.4 shows comparison of the ANN-SVM hybrid
model with existing models.An improvement of 2% to 4% is reported with ANN-SVM hybrid
model in comparison with reported literature. The ANN-SVM hybrid model has achieved an
accuracy of 84.12%,specificity of 82.95%, F-measure of 85.23% and sensitivity of 85.14% on
Dset-1 and an accuracy of 90%, specificity of 87.5%, F-measure of 91.67% and sensitivity of
91.66% on Dset-2.
32
Table 3.3: Classification results of the best two classifiers and the proposed ANN-SVM hybrid
model on Dset-1 and Dset-2
Datasets Classifiers/ Accuracy Specificity F-measure Sensitivity
Models (%) (%) (%) (%)
ANN 79.28 77.16 80.9 82
Dset-1 SVM 78.4 76.3 79 80.4
ANN+SVM 84.12 82.95 85.23 85.14
(Hybrid model)
ANN 84.96 81.56 85.4 85.88
Dset-2 SVM 82.65 80.12 84 83.50
ANN+SVM 90 87.5 91.67 91.66
(Hybrid model)
Table 3.4: Comparison of the ANN-SVM hybrid model with existing models
Model Feature Classifier Accuracy Specificity Sensitivity F-measure
Extraction (%) (%) (%) (%)
Technique
(Zhu et. al Sonographic ANN 83.10 81.80 83.80 -
(2013)) features
33
3.2 Summary
This chapter presents a hybrid model using hybridization of two ML classifiers for
identification and classification of TNDs. The proposed ANN-SVM model is compared with
standalone classifiers as well with hybrid model. From the experiments, it can be inferred that
the performance of the proposed model is competitive when the two ML classifiers are hybrid
together.
34
CHAPTER 4
CNN and its variants are one of the most famous and commonly employed algorithms in the
field of medical imaging [94]. Some of the popular reasons include automatic identification of
relevant features from the images without any human supervision. Its structure is inspired from
neurons processing in human brains [95]. This chapter gives an elaborative explanation on the
hybridization of CNN with SVM for the identification and classification of TNDs. The
efficiency of the CNN-SVM hybrid model is tested on collected and TDID datasets.
The working of CNN-SVM model for the identification and classification of TNDs is discussed
in this section. The model works in four stages namely: (1) data collection, (2) pre-processing,
(3) classification and (4) result analysis. Figure 4.1 shows the systematic flow of the CNN-
SVM model.
In this work, public TDID dataset and collected dataset are considered. The public dataset has
188 malignant and 107 benign images i.e., a total of 295 thyroid USG images whereas collected
dataset has 226 malignant and 428 benign i.e., a total of 654 thyroid USG images. Figure 4.2
shows the sample of benign and collected dataset.
4.1.2 Pre-processing
It is defined as a way of converting raw data into a desired form [96-97]. This derived
information will be fed into the training model for the successful diagnosis [98]. The various
steps involve in pre-processing are image resizing, noise removal using gaussian blur function,
RGB to grayscale conversion and data augmentation respectively. Initially the size of image
was 560 X 360 pixels, for the uniformity it’s being resized to 256 X 256 pixels. In this work,
four types of data augmentation techniques are used namely like rescale, zoom, shear and
horizontal flip respectively. All these steps are computed using equation [4.12-4.15]:
𝑟𝑒𝑠𝑖𝑧𝑒 = 𝑡𝑓. 𝑘𝑒𝑟𝑎𝑠. 𝑆𝑒𝑞𝑢𝑒𝑛𝑡𝑖𝑎𝑙([𝑙𝑎𝑦𝑒𝑟𝑠. 𝑅𝑒𝑠𝑖𝑧𝑖𝑛𝑔(𝐼𝑀𝐺 _𝑆𝐼𝑍𝐸, 𝐼𝑀𝐺 _𝑆𝐼𝑍𝐸)]) (4.12)
where images: the RGB tensor to convert the images, name: name of the operation
36
𝑡𝑟𝑎𝑖𝑛_𝑑𝑎𝑡𝑎𝑔𝑒𝑛 = 𝐼𝑚𝑎𝑔𝑒𝐷𝑎𝑡𝑎𝐺𝑒𝑛𝑒𝑟𝑎𝑡𝑜𝑟(𝑁𝑎𝑚𝑒, 𝑉𝑎𝑙𝑢𝑒) (4.15)
4.1.3.1 CNN
The CNN architecture consists of four parts namely: (1): input layer, (2): convolutional layer,
(3): max-pooling layer, (4): fully connected layer and (5): output layer [99]. Figure 4.3 shows
the workflow of CNN.
1. Input Layer
It specifies a complete description of an image (Nh x Nw x Nc) were h: height, w: width and c:
channel size of an image [101].
2. Convolutional Layer
In this layer, 2-D kernels are learned during the initial phase. The coefficients(coeff.) of all
these kernels encodes the important information from the input image [102]. There are some
hyper-parameters like number of kernels, padding, size of kernels and stride to improve the
performance [103]. Convolution works by finding the sum of the dot(.) products between the
input image and the kernel [104]. The output of the convolution becomes the input of the next
layer. It works as the spatial filtering [105]. Here, the size of kernel (k) is defined by equation
16:
𝑊𝑥𝐻 (4.16)
where W: width and W=2a+1, a and b: positive integer, H: height of the image pixel and
H=2b+1
37
During the entire convolutional operation, the size of the image and the kernel are same.
Padding is one of the techniques of the convolution process to introduce zeros (0s) around the
border of the image to maintain the aspect ratio of the resultant image after convolution [106].
This can be computed using equation 4.17:
𝑁+2𝜌−𝑤 𝑁+2𝜌−𝑤
[𝑁, 𝑁, 𝑍] ∗ [𝑊, 𝑊, 𝑁 ] = ([ + 1] , [ + 1] , 𝑁 ) (4.17)
𝐶 𝑠 𝑠 𝑓
where [N, N, Z]: Size of an image, [W, W, N_C]: Size of a kernel, p: no. of padding, s: stride
length, N: filter depth, Z: image depth, Nf: filters of the kernel
Figure 4.4(a-e) explains the illustration of a convolutional operation with 3X3 kernel size. In
this case, the input to the image is 5X5 matrix with stride 1. After the 9th convolution operation,
3X3 feature map is obtained. This shows the size of the input image is reduced extracting some
useful global and local features.
Figure 4.4(a-e): The illustration of a convolutional operation with 3X3 kernel size [106]
3. Max-pooling Layer
The output of the convolutional layer becomes the input of the max-pooling layer [107]. It
38
extracts the most essential features from each feature map, thereby giving the best feature
output. The major benefits of using this layer are as follows: (1) it prevents overfitting, (2)
feature invariance and (3) dimensionality reduction of the features [108]. It can be computed
using equation 4.18 and 4.19:
𝑦𝑖 = max 𝑅𝑥 𝑅{𝑦𝑖𝑟𝑥𝑟}𝑓(𝑟, 𝑟) (4.18)
𝑓(𝑟, 𝑟) = 𝜀. 𝑦𝑘−1 𝑥 𝑣𝑘 + 𝑒𝑘 (419)
𝑖 𝑖,𝑗 𝑗
where max RxR: max-pooling in RxR, 𝑦𝑖𝑟𝑥𝑟: i-th output of rxr window, ϵ: trainable variant
Figure 4.5(a-b) explains the illustration of max-pooling operation. The complete operation
works from left(L) to right(R), the first picture is the input, next is the max-pooling and so on.
39
Figure 4.6: The flattening operation performed in fully connected layer [109]
5. Output Layer
In this layer we have the classified result which is mostly followed by softmax function (i.e.,
the normalized exponential function) which gives the output mapped into the range of [0,1]
[112]. It can be computed using equation 4.20:
𝑒𝑧𝑖
𝜎(𝑧 ) = (4.20)
𝑖 𝑛
𝛴𝑖=1 𝑒 𝑧𝑖
4.1.3.2 SVM
It is a class of ML algorithms works by finding (n-1) dimensional hyper-plane. It finds a plane
by maximizing the dist. between the data points available or given in the classes with the help
of support vectors [113]. SVM employs the use of kernels like sigmoid, radial basis function
(rbf), linear, polynomial and non-linear which takes the data as the input and gives output
accordingly [114-115].
Let there be a sample dataset input: {(𝑥1, 𝑦1 ), (𝑥2, 𝑦2), … … (𝑥𝑛, 𝑦𝑛)}, 𝑥𝑖 𝜖 𝑅𝑑, target output: 𝑖 =1,2,
… … 𝑁, 𝑦 ∈ {∓1}[116]. The optimal classification hyperplane is defined as given in equation 4.21:
𝑣𝑇𝑥𝑖 + 𝑏 = 0 (4.21)
where b: bias, 𝑣: weight vector
Then b and 𝑣 can be determined using equation 4.22:
𝑦𝑖(𝑤𝑇𝑥𝑖 + 𝑏) ≥ 1 − 𝑥𝜉𝑖 (422)
40
where ξi is the slack variable(var.).
According to Lagrange multiplier method, the solution to optimal classification hyperplane is
converted into optimization using equation 4.23:
𝑁
𝑁 𝑁
1
𝑄(𝑎) = ∑ 𝑎𝑖 − ∑ ∑ 𝑎𝑖 𝑎𝑗 𝑦𝑖 𝑦𝑗 𝐾 (𝑥𝑖 , 𝑥𝑗 ) (4.23)
2
𝑖=1 𝑗=1
𝑖=1
42
F-measure as 95.64% and specificity of 91.89% on Dset-1 and specificity of 93.93%, F-measure
of 98.33% and sensitivity of 97.80% on Dset-2. An improvement of 2% to 3% is reported by
proposed model in figure 4.9. Figure 4.10 shows the epoch-accuracy graph for Dset-1. Figure
4.11 shows the epoch-accuracy graph for Dset-2.
Table 4.1: Parameters set for training CNN-SVM hybrid model for Dset-1 and Dset-2
Parameter Batch Kernel Activation Stride Pooling Activation Optimizer Epoch Regularizer
Size Size layer l2
Values 32 3 Relu 2 Max- Linear Adam 100 0.01
pooling
layer
Table 4.2: The comparison of the CNN-SVM model on Dset-1 and Dset-2 with existing models
used for classifying TNDs
Models Accuracy F- Sensitivity Specificity
(%) measure (%) (%)
(%)
(Wang et al. 80.91 - 81.82 80
(2018))
(Nguyen et al. 90.88 - - -
(2019))
(Song et al. 92.1 - 94.1 96.2
(2019))
(Ajilisa et al. 89.93 89.43 92.76 -
(2020))
(Nguyen et al. 92.05 - - -
(2021))
(Proposed 94.57 95.64 96.70 91.89
CNN-SVM
hybrid
model on
Dset-1)
(Proposed 96 98.33 97.80 93.93
CNN-SVM
hybrid
model on
Dset-2)
43
100
96
94.57
95 92.1 92.05
90.88 89.93
90
Percentage
85
80.91
80
75
70
Wang et al. Song et al. Nguyen et Ajilisa et al. Nguyen et Proposed Proposed
(2018) (2019) al. (2019) (2020) al. (2021) CNN-SVM CNN-SVM
hybrid hybrid
model on model on
dataset-1 dataset-2
Models
Figure 4.8: Performance comparison of the accuracy parameter with the existing and CNN-
SVM model on Dset-1 and Dset- 2
85 81.82
80
80
75
70
65
60
Wang et al. (2018) Song et al. (2019) Ajilisa et al. (2020) Proposed CNN- Proposed CNN-
SVM hybrid model SVM hybrid model
on dataset-1 on dataset-2
Models
Figure 4.9: Performance comparison of specificity, f-measure and sensitivity parameters with
existing and CNN-SVM model on Dset-1 and Dset-2
44
Epoch-accuracy graph for dataset-1
100
97
94
91
Accuracy(%)
88
85
82
79
76
73
10 20 30 40 50 60 70 80 90 100
Epoch
90
88
86
84
82
80
78
10 20 30 40 50 60 70 80 90 100
Epoch
4.2 Summary
An effective CNN-SVM hybrid model is proposed for classification of BTND and MTND USG
images in this chapter. The work has been evaluated on 1180 and 2616 public TDID and
collected TNDs image datasets on Google Colaboratory. The proposed model is found to
perform well even when the sample size of the dataset is less. It has shown an improvement of
45
3% to 5% on Dset-1 and 4% to 6% on Dset-2 in comparison with existing models reported in
literature for the identification and classification of TNDs.
46
CHAPTER 5
47
Figure 5.1: Systematic flow of the proposed GSO-CNN model
5.1.2 Pre-processing
The various pre-processing steps involved in this work are image resizing, noise removal, RGB
to grayscale conversion as discussed in section 3.2.2. The various data augmentation techniques
involved in this work are RandRotation, RandXTranslation, RandYTranslation,
RandXReflection, RandYReflection and Padding.
48
Figure 5.2: The morphological operation [120]
In this work, dilation and erosion operation are considered. The structure used in this work is
“disk”. The dilation and erosion are computed using equation 5.24 and 5.25:
𝐼(𝑑𝑖𝑙𝑎𝑡𝑒)𝑆 = {𝐼 + 𝑆; 𝑓𝑜𝑟 𝑎𝑙𝑙 𝑝𝑖𝑥𝑒𝑙𝑠 𝑖𝑛 𝐼 ∈ 𝑆} (5.24)
𝐼(𝑒𝑟𝑜𝑑𝑒𝑑)𝑆 = {𝐼(𝑒𝑟𝑟𝑜𝑟𝑒𝑑)𝑆}(𝑒𝑟𝑜𝑑𝑒𝑑)𝑆 (5.25)
5.1.3.2 Segmentation
It is defined as the process of partitioning image into image segments formerly known as image
objects [121]. It is mostly used to change the representation into more meaningful form for easy
analysis [122]. Active Contour (AC) segmentation technique was first coined by “Kass” [123].
It is basically the curve in the image space whose deformation is based on energy minimization.
Let x and y be the position of co-ordinates of an 2D image I(x,y), and curve be:
49
𝑉(𝑠) = (𝑥(𝑠), 𝑦(𝑠)) (5.27)
where s є [0,1]
The energy function (E) is defined using equation 5.28 and 5.29:
1
𝐸𝑡𝑜𝑡𝑎𝑙 = ∫ 𝐸(𝑉(𝑠))𝑑𝑠 (5.28)
0
1
𝐸𝑡𝑜𝑡𝑎𝑙 = ∫ [𝐸𝑖𝑛𝑡 (𝑉(𝑠)) + 𝐸𝑒𝑥𝑙 (𝑉(𝑠))] 𝑑𝑠 (5.29)
0
To control the deformation and displacement, Eint and Eext are converted into internal and
external forces.
The Eint and Eext energy component can be defined as:
1 1
𝐸𝑖𝑛𝑡(𝑉 (𝑠)) = [ 𝛼(𝑠 )|𝑉𝑠 (𝑠)|2 + 𝛽(𝑠)|𝑉𝑠𝑠(𝑠)|2] (5.30)
2 2
where γ: weight factor, G_σ (x,y): 2D gaussian function, *:convolution operator, ∇: gradient
operator, σ: standard deviation
At equation 5.32, the Eint and Eext attain the minimum value. Hence goal is achieved.
𝐹𝑖𝑛𝑡+ 𝐹𝑒𝑥𝑡 = 0 (5.32)
5.1.4.2 Alex-Net
It was coined in 2012 and employed for generalized image classification for the first time [128].
In the Alex-Net, ReLU function is introduced to solve the problem of non-linearity and speed
up of the network [129]. Alex-Net consists of five convolutional(conv.) layer, three fully
connected layers and one output layer. There is a Local Response Normalization (LRN) layer
that follows the first as well as the second convolutional layer [130]. Figure 5.4 shows the
structure of Alex-Net.
51
Figure 5.4: Structure of Alex-Net [131]
5.1.4.3 VGG-16
The idea was coined from the “University of Oxford” and trained on ImageNet dataset [132].
A traditional VGG-16 consists of multiple “3X3 kernel-sized filters” that help the model in
learning complex features by increasing depth of the network [133]. Figure 5.5 shows the
architecture of VGG-16.
Optimization of drop-out and learning rate factor using GSO technique is discussed in this
section. Let the hyperparameters be hyp1,……..hypn of the DL model and α1,…..αn be domains.
The model is trained with hyp (DLtrain) on the training of TNDs images. The main aim behind
hyp is to find the best parameter hyp1*, hyp2*. Figure 5.7 shows the proposed GSO-CNN model.
53
Figure 5.7: Proposed GSO-CNN model
5.2 Proposed GSO-CNN Algorithm for the identification and classification of TNDs
Table 5.1: Parameter setting for GSO-CNN model for Dset -1 and Dset- 2
Parameters Value
Batch size 15
Activation function Softmax
Max-pooling 2X2
Drop-out factor 0.2
Max-epoch 10
Learning rate 0.01
Optimizer SGDM
Table 5.2: Performance comparison of the GSO-CNN model with Alex-Net, Res-Net-50,
VGG-16 and DNN on Dset -1 and Dset-2
Dset Techniques Models Accuracy Specificity F-measure Sensitivity
(%) (%) (%) (%)
GSO-CNN
Without segmentation and boundary
GSO-CNN
model 95.33 95 97.20 96.66
Dset -2
56
GSO-CNN
detection techniques
Res-Net-50 87.91 86.58 88.55 89
CASE 2 VGG-16 86.66 83.95 87.56 87.12
GSO-CNN
model 96.02 95 98.34 97.02
DNN 92.30 90.16 93.19 93.68
Dset -2
Table 5.3: Results of GSO technique obtained on dropout factor and learning rate parameters
obtained on GSO-CNN model, Alex-Net, Res-Net-50, VGG-16 and DNN
rate
Dropout 0.2 0.02 0.06 0.09 0.1 0.04 0.08
factor
(GSO-
CNN) 95.30 80.90 85.16 89 76.06 82.32 87.19
Model
DNN 91.01 69 76.88 82.76 62.3 71.1 81.1
Accuracy
Models
Dset-1
Models
Dset -2
57
Table 5.4: Comparison of the proposed GSO-CNN model and existing models on Dset -1 and
Dset-2
Models F-measure Accuracy Specificity Sensitivity
(%) (%) (%) (%)
(Wang et. al (2018)) - 80.91 80 81.82
100
95.3 96.02
95
92.1 92.05
90.88
89.93
90
Accuracy(%)
85
80.91
80
75
70
Wang et. al Song et. al Nguyen et. Ajilisa et. al Nguyen et. Proposed Proposed
(2018) (2019) al (2019) (2020) al (2020) model on model on
dataset-1 dataset-2
Models
Figure 5.8: Comparative analysis of the accuracy parameter with models on Dset -1 and Dset-
2
58
120
96.7 95 98.34
100 94.1 92.76 92.53
96.2 89.43 94.62 94.09
81.82
80 80
Percentage
60
40
20
0
Wang et. al (2018) Song et. al (2019) Ajilisa et. al (2020) Proposed model onProposed model on
dataset-1 dataset-2
Models
Figure 5.9: Comparative analysis of models based on sensitivity, f-measure and specificity on
Dset -1 and Dset-2
100 95.3
86.43
90
81
78.12
80 73.82
69.23
Accuracy (%)
70 63.59
55.62
60
50 43.73
40
3 0.75
30
1 2 3 4 5 6 7 8 9 10
Epoch
59
100 92.53 96.02
95
86.37
90
81
85
76.45
80
71.33
75
Accuracy(%) 65.69
70
65
55.67
60
55
45.67
50
45
40
0.33
353
30
1 2 3 4 5 6 7 8 9 10
Epoch
5.2 Summary
GSO-CNN model is proposed for the identification and classification of TNDs in this chapter.
The model is compared with Res-Net-50, DNN, VGG-16, Alex-Net and various reported
models. It is concluded from the results that the proposed GSO-CNN outperforms when
compared with and without segmentation and boundary detection techniques. The GSO
technique is used to optimize the drop-out factor and learning rate parameters to find the best
results.
60
CHAPTER 6
Health care is conventionally regarded as essential determinant in promoting the general mental,
physical, and social well-being of people around the world [140]. A reliable system forTNDs is
necessary for early identification and classification of BTND and MTND to save effort, time
and human life [141-145]. In this chapter, Deep- Generative Adversarial Network (Deep- GAN)
based model is proposed for the early identification and classification of TNDs. GSO technique
is used for tunning of the Deep-GAN (i.e., Alex-GAN and VGG-GAN) model by optimizing
the optimizer, batch size and learning rate parameters.
61
Figure 6.1: The systematic flow of Deep-GAN model
63
Figure 6.3: Proposed Alex-GAN model
64
3. Use GAN for data augmentation
4. Divide train and test part of TNDs images
5. Initialize the parameters of Deep-GAN model
-dropoutfactor
- maxepoch
- batchsize
- learning rate
- activation function
- optimizer
6. Use GSO technique for optimization of Deep-GAN.
7. Train the model
Classify type of TNDs
END
65
improvement of 2% to 4% with VGG-GAN and existing models. Figure 6.5 shows the generator
network and figure 6.6 shows the discriminator network. Figure 6.7(a-b) shows the training of
Alex-Net architecture whereas figure 6.8(a-d) shows the training of VGG-16 architecture.
Figure 6.9 and 6.10 show the accuracy epoch graph of Alex-GAN model on Dset-1 and Dset-
2. Figure 6.11 and 6.12 show the accuracy epoch graph of VGG-GAN model on Dset-1 and
Dset-2.
Table 6.2: The grid search results over different values of hyperparameters set for Alex-GAN
and VGG-GAN models
Results
Table 6.4: Performance comparison of the Deep- GAN model (i.e., Alex-GAN and VGG-GAN) with
existing models for classification of TNDs
Models F-measure Accuracy Sensitivity Specificity
(%) (%) (%) (%)
(Wang et al. (2018)) - 80.91 81.82 80
(Nguyen et al. (2019)) - 90.88 - -
(Song et al. (2019)) - 92.1 94.1 96.2
(Ajilisa et al. (2020)) 89.43 89.93 92.76 -
(Hang et al. (2021)) - 95 - -
(Nguyen et al. (2021)) - 92.05 - -
Proposed model 95.65 96 96.70 94.91
(Alex-GAN) on
Dset-1
Proposed model 98.26 97.03 97.70 95.83
(Alex-GAN) on
Dset-2
Proposed model 95.18 93.60 96.73 90.32
(VGG-GAN) on
Dset-1
Proposed model 96.7 94.28 95.98 91.48
(VGG-GAN) on
Dset-2
67
Figure 6.5: Generator network Figure 6.6: Discriminator network
68
Figure 6.8a: Generator network Figure 6.8b: Discriminator network
69
Accuracy-epoch graph for dataset-1
100
95
Accuracy(%) 90
85
80
75
70
10 20 30 40 50 60 70 80
Epoch
95
Accuracy(%)
90
85
80
75
10 20 30 40 50 60 70 80
Epoch
85
80
75
70
65
10 20 30 40 50 60 70 80
Epoch
70
Accuracy-epoch graph for dataset-2
100
95
90
Accuracy(%)
85
80
75
70
65
60
10 20 30 40 50 60 70 80
Epoch
6.2 Summary
The proposed Deep-GAN model has achieved promising results for the early identification and
classification of TNDs. GSO technique is used for tunning of the Deep-GAN (i.e. Alex-GAN
and VGG-GAN) model. The experimental results demonstrate that the proposed Alex-GAN
performs better in comparison with VGG-GAN and other DL models reported in literature with
an improvement of 2% to 4%. The developed models have been evaluated on TDID and
findings are generalized by evaluating such models on collected dataset. The proposed model
will help the radiologists for early identification and classification of TNDs in USG images.
71
CHAPTER 7
This chapter concludes research work written in the thesis. The major research contributions and
future research directions are discussed as follows.
▪ The proposed model has been evaluated on public TDID and collected dataset taken from
Kriti Scanning Center, Prayagraj, Uttar Pradesh, India duly approved by NABH and
Healthcare Providers. The duration of dataset collection was from July 2020 to March
2021.The goal behind using collected dataset is to generalize the results of the proposed
models for the identification and classification of benign and malignant TNDs.
▪ Two hybrid models are proposed: (1) hybridization of two ML classifiers i.e., ANN-SVM
hybrid model and (2) hybridization of DL and ML classifier i.e., CNN-SVM hybrid model.
The experimental results for these models show that hybridization of DL and ML classifier
outperforms in comparison with hybridization of two ML classifiers. The findings
demonstrate that the proposed CNN-SVM hybrid model performs better as compared to the
standalone classifier.
▪ A GSO-CNN model is proposed for the classification of TNDs in medical USG images. The
variants of CNN models like Res-Net-50, DNN, Alex-Net and VGG-16 are explored for the
classification of TNDs. The proposed model is optimized using GSO technique in terms of
hyperparameters like optimizer, batch size and learning rate. The proposed GSO-CNN model
experimental results are compared with Res-Net, DNN, Alex-Net, VGG-16 and other state-
of-the-art models. The results shows that the proposed GSO-CNN model is competitive in
comparison with the reported models in literature.
▪ Deep-GAN model is proposed to improve the accuracy for the identification and classification
of TNDs. GSO technique is used for tunning of the Deep-GAN (i.e. Alex-GANand VGG-
GAN) model by optimizing the optimizer, learning rate and batch size hyperparameters. The
The proposed models can be used by the clinicians, doctors, researchers and practitioners for the
diagnostic purpose. The proposed models will act as an early alarm for patients having TNDs.
73
REFERENCES
[1] Das, D., Banerjee, A., Jena, A. B., Duttaroy, A. K., & Pathak, S., “Essentiality, relevance,
and efficacy of adjuvant/combinational therapy in the management of thyroid
dysfunctions”, Biomedicine & Pharmacotherapy, vol. 146, 2022.
[2] https://www.thyroid.org/professionals/ata-professional-guidelines
[3] Nilsson, Mikael, and Henrik Fagman, “Development of the thyroid gland”, Development,
vol. 144, pp. 2123-2140, 2017.
[4] Zhang, X., Malik, B., Young, C., Zhang, H., Larkin, D., Liao, X. H., & Arvan , P.,
“Maintaining the thyroid gland in mutant thyroglobulin-induced hypothyroidism requires
thyroid cell proliferation that must continue in adulthood”, Journal of Biological
Chemistry, 2022.
[5] Darras, Veerle M., "Thyroid gland", In Sturkie's Avian Physiology, pp. 795-812, Academic
Press, 2022.
[6] Sadeghi, D., Shoeibi, A., Ghassemi, N., Moridian, P., Khadem, A., Alizadehsani, R., ... &
Acharya, U. R, “An overview of artificial intelligence techniques for diagnosis of
Schizophrenia based on magnetic resonance imaging modalities: Methods, challenges, and
future works”, Computers in Biology and Medicine, 2022.
[7] Abbasian Ardakani, A., Reiazi, R., & Mohammadi, A., “A Clinical Decision Support
System Using Ultrasound Textures and Radiologic Features to Distinguish Metastasis
from Tumor‐Free Cervical Lymph Nodes in Patients with Papillary Thyroid Carcinoma”,
Journal of Ultrasound in Medicine, vol. 37(11), pp. 2527-2535, 2018.
[8] Hegedus, L., Bianco, A. C., Jonklaas, J., Pearce, S. H., Weetman, A. P., & Perros, P.,
“Primary hypothyroidism and quality of life", Nature Reviews Endocrinology, vol. 18(4),
pp. 230-242, 2022.
[9] Demet, M. M., Ozmen, B., Deveci, A., Boyvada, S., Adiguzel, H., & Aydemir, O,
“Depression and anxiety in hyperthyroidism”, Archives of medical research, vol. 33(6),
pp. 552-556, 2002.
[10] Kwong, N., Medici, M., Angell, T. E., Liu, X., Marqusee, E., Cibas, E. S., ... & Alexander,
E. K., “The influence of patient age on TND formation, multinodularity, and thyroid cancer
risk”, The Journal of Clinical Endocrinology & Metabolism, vol. 100(12), pp. 4434-4440,
2015.
74
[11] Richman, D. M., Benson, C. B., Doubilet, P. M., Wassner, A. J., Asch, E., Cherella, C. E.,
& Frates, M. C, “Assessment of American college of radiology thyroid imaging reporting
and data system (TI-RADS) for pediatric thyroid nodules”, Radiology, vol. 294(2), pp.
415-420.
[12] Ker, J., Wang, L., Rao, J., & Lim, T, “Deep learning applications in medical image
analysis”, IEEE Access, vol. 6, pp. 9375-9389, 2017.
[13] Yap, P. T., Zhang, Y., & Shen, D., “Multi-Tissue Decomposition of Diffusion MRI Signals
via $\ell _ {0} $ Sparse-Group Estimation”, IEEE Transactions on Image Processing, vol.
25(9), pp. 4340-4353, 2016.
[14] Schmidhuber, J, “Deep learning in neural networks: An overview”, Neural networks, vol.
61, pp. 85-117, 2015.
[15] Bengio, Y., “Learning deep architectures for AI”, Now Publishers Inc., 2009.
[16] Hinton, G. E., & Salakhutdinov, R. R, “Reducing the dimensionality of data with neural
networks”, Science, vol. 313(5786), pp. 504-507, 2006.
[17] Vincent, P., Larochelle, H., Lajoie, I., Bengio, Y., Manzagol, P. A., & Bottou, L, “Stacked
denoising autoencoders: Learning useful representations in a deep network with a local
denoising criterion”, Journal of machine learning research, 2010.
[18] Nair, V., & Hinton, G. E, “Rectified linear units improve restricted boltzmann machines”,
In Icml, 2010.
[19] Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., & Salakhutdinov, R., “Dropout:
a simple way to prevent neural networks from overfitting”, The journal of machine
learning research, vol. 15(1), pp. 1929-1958, 2014.
[20] Ioffe, S., & Szegedy, C, “Batch normalization: Accelerating deep network training by
reducing internal covariate shift”, In International conference on machine learning, pp.
448-456, 2015.
[21] Pedraza, L., Vargas, C., Narváez, F., Duran, O., Munoz, E., & Romero, E., “An open
access thyroid ultrasound image database”, In 10th International Symposium on Medical
Information Processing and Analysis, vol. 9287, pp.188-193, SPIE, 2015.
[22] Quality Council of India, Accessed: 07- Sep. 2022, [Online], Available:
https://www.nabh.co/frmViewCGHSRecommend.aspx?Type=Diagnostic%20Centre&cit
yID=94
[23] Guvenir, H. A., Acar, B., Demiroz, G., & Cekin, A, “A supervised machine learning
algorithm for arrhythmia analysis”, In Computers in Cardiology, pp. 433-436, IEEE, 1997.
75
[24] Gupta, V., Sachdeva, S., & Dohare, N, “Deep similarity learning for disease prediction”,
Trends in Deep Learning Methodologies, pp.183-206, 2021.
[25] Zhu, L. C., Ye, Y. L., Luo, W. H., Su, M., Wei, H. P., Zhang, X. B., & Zou, C. L, “A
model to discriminate malignant from benign thyroid nodules using artificial neural
network”, PLoS One, vol. 8(12), 2013.
[26] Narayan, N. S., Marziliano, P., Kanagalingam, J., & Hobbs, C. G., “Speckle patch
similarity for echogenicity-based multiorgan segmentation in ultrasound images of the
thyroid gland”, IEEE journal of biomedical and health informatics, vol. 21(1), pp.172-
183, 2015.
[27] Song, G., Xue, F., & Zhang, C., “A model using texture features to differentiate the nature
of thyroid nodules on sonography”, Journal of Ultrasound in Medicine, vol. 34(10), pp.
1753-1760, 2015.
[28] Ardakani, A. A., Gharbali, A., & Mohammadi, A., “Application of texture analysis method
for classification of benign and malignant thyroid nodules in ultrasound images”, Iranian
journal of cancer prevention, vol. 8(2), 2015.
[29] Nugroho, H. A., Nugroho, A., & Choridah, L., “Thyroid nodule segmentation using active
contour bilateral filtering on ultrasound images”, In 2015 International Conference on
Quality in Research (QiR), pp. 43-46, 2015.
[30] Song, G., Xue, F., & Zhang, C, “A model using texture features to differentiate the nature
of thyroid nodules on sonography”, Journal of Ultrasound in Medicine, vol. 34(10), pp.
1753-1760, 2015.
[31] Pan, Q., Zhang, Y., Zuo, M., Xiang, L., & Chen, D., “Improved ensemble classification
method of thyroid disease based on random forest”, In 2016 8th International Conference
on Information Technology in Medicine and Education (ITME), pp. 567-571, 2016.
[32] Alrubaidi, W. M., Peng, B., Yang, Y., & Chen, Q., “An interactive segmentation algorithm
for thyroid nodules in ultrasound images”, In International Conference on Intelligent
Computing, pp. 107-115, Springer, Cham, 2016.
[33] Vanithamani, R., & Dhivya, R., “Thyroid Nodule Classification in Medical Ultrasound
Images”, In International Conference on Soft Computing and Pattern Recognition, pp.
509-514, Springer, Cham, 2016.
[34] Nugroho, H. A., Rahmawaty, M., Triyani, Y., & Ardiyanto, I., “Texture analysis for
classification of thyroid ultrasound images”, In 2016 International Electronics Symposium
(IES), pp. 476-480, 2016.
76
[35] Mustafa, N., Memon, R. A., Li, J. P., & Omer, M. Z., “A classification model for
imbalanced medical data based on PCA and farther distance based synthetic minority
oversampling technique”, Int. J. Adv. Comput. Sci. Appl, vol.8(1), pp.61-67, 2017.
[36] Pasha, S. J., & Mohamed, E. S., “Ensemble gain ratio feature selection (EGFS) model with
machine learning and data mining algorithms for disease risk prediction”, In 2020
International Conference on Inventive Computation Technologies (ICICT), pp. 590-596,
2020.
[37] Jothi, J. A. A., & Rajam, V. M. A., “Automatic classification of thyroid histopathology
images using multi-classifier system”, Multimedia Tools and Applications, vol. 76(18), pp.
18711-18730, 2017.
[38] Jiang, Y., Deng, Z., Chen, J., Wu, H., Choi, K. S., & Wang, S, “Intelligent diagnostic
methods for thyroid nodules”, Journal of Medical Imaging and Health Informatics, vol.
7(8), pp. 1772-1779, 2017.
[39] Shankar, K., Lakshmanaprabu, S. K., Gupta, D., Maseleno, A., & De Albuquerque, V. H.
C., “Optimal feature-based multi-kernel SVM approach for thyroid disease classification”,
The Journal of Supercomputing, vol. 76(2), pp. 1128-1143, 2020.
[40] Ma, X., Xi, B., Zhang, Y., Zhu, L., Sui, X., Tian, G., & Yang, J., “A Machine Learning-
based Diagnosis of Thyroid Cancer Using Thyroid Nodules Ultrasound Images.”, Current
Bioinformatics, vol.15(4), pp. 349-358, 2020.
[41] Zheng, X., Lv, G., Du, G., Zhai, Z., Mo, J., & Lv, X., “Rapid and low-cost detection of
thyroid dysfunction using Raman spectroscopy and an improved support vector machine”,
IEEE Photonics Journal, vol. 10(6), pp. 1-12, 2018.
[42] Dandan, L., Yakui, Z., Linyao, D., Xianli, Z., & Yi, S., “Texture analysis and classification
of diffuse thyroid diseases based on ultrasound images”, In 2018 IEEE International
Instrumentation and Measurement Technology Conference (I2MTC), pp. 1-6, 2018.
[43] Poudel, P., Illanes, A., Ataide, E. J., Esmaeili, N., Balakrishnan, S., & Friebe, M., “Thyroid
ultrasound texture classification using autoregressive features in conjunction with machine
learning approaches”, IEEE Access, vol. 7, pp. 79354-79365, 2019.
[44] Prochazka, A., Gulati, S., Holinka, S., & Smutek, D., “Classification of thyroid nodules in
ultrasound images using direction-independent features extracted by two-threshold binary
decomposition”, Technology in cancer research & treatment, vol. 18, 2019.
[45] Wu, Y., & Liu, P., “A Classification Algorithm of Ultrasonic Thyroid Standard Planes
Using LBP and HOG Features”, In 2019 IEEE 13th International Conference on Anti-
77
counterfeiting, Security, and Identification (ASID), pp. 103-107, 2019.
[46] Colakoglu, B., Alis, D., & Yergin, M., “Diagnostic value of machine learning-based
quantitative texture analysis in differentiating benign and malignant thyroid nodules”,
Journal of oncology, 2019.
[47] Zhang, B., Tian, J., Pei, S., Chen, Y., He, X., Dong, Y., & Zhang, S., “Machine learning–
assisted system for thyroid nodule diagnosis”, Thyroid, vol. 29(6), pp. 858-867, 2019.
[48] Kesarkar, X. A., & Kulhalli, K. V., “Thyroid Nodule Detection using Artificial Neural
Network”, In 2021 International Conference on Artificial Intelligence and Smart Systems
(ICAIS), pp. 11-15, 2021.
[49] Zhao, C. K., Ren, T. T., Yin, Y. F., Shi, H., Wang, H. X., Zhou, B. Y., ... & Xu, H. X., “A
comparative analysis of two machine learning-based diagnostic patterns with thyroid
imaging reporting and data system for thyroid nodules: diagnostic performance and
unnecessary biopsy rate”, Thyroid, vol. 31(3), pp. 470-481, 2021.
[50] Bengio, Y, “Learning deep architectures for AI”, Foundations and trends in Machine
Learning, vol. 2(1), pp. 1-127, 2009.
[51] Rusk, N, “Deep learning”, Nature Methods, vol. 13(1), pp. 35-35, 2016.
[52] Lauzon, F. Q, “An introduction to deep learning” In 2012 11th International Conference
on Information Science, Signal Processing and their Applications (ISSPA), pp. 1438-1439,
2012.
[53] Voulodimos, A., Doulamis, N., Doulamis, A., & Protopapadakis, E, “Deep learning for
computer vision: A brief review”, Computational intelligence and neuroscience, 2018.
[54] Ronao, C. A., & Cho, S. B, “Human activity recognition with smartphone sensors using
deep learning neural networks”, Expert systems with applications, vol. 59, pp. 235-244,
2016.
[55] Du, W., & Sang, N., “An effective method for ultrasound thyroid nodules segmentation”,
In 2015 International Symposium on Bioelectronics and Bioinformatics (ISBB), pp. 207-
210, 2015.
[56] Acharya, U. R., Chowriappa, P., Fujita, H., Bhat, S., Dua, S., Koh, J. E., ... & Ng, K. H,
“Thyroid lesion classification in 242 patient population using Gabor transform features
from high resolution ultrasound images”, Knowledge-Based Systems, vol. 107, pp. 235-
245, 2016.
[57] Peng, W., Liu, C., Xia, S., Shao, D., Chen, Y., Liu, R., & Zhang, Z., “Thyroid nodule
recognition in computed tomography using first order statistics”, Biomedical engineering
78
online, vol.16(1), pp. 1-14, 2017.
[58] Zhu, Y., Fu, Z., & Fei, J., “An image augmentation method using convolutional network
for thyroid nodule classification by transfer learning”, In 2017 3rd IEEE International
Conference on Computer and Communications (ICCC), pp. 1819-1823, 2017.
[59] Song, W., Li, S., Liu, J., Qin, H., Zhang, B., Zhang, S., & Hao, A., “Multitask cascade
convolution neural networks for automatic thyroid nodule detection and recognition”,
IEEE journal of biomedical and health informatics, vol. 23(3), pp. 1215-1224, 2018.
[60] Zhang, R., Liu, Q., Cui, H., Wang, X., Song, S., Huang, G., & Feng, D., “Thyroid
classification via new multi-channel feature association and learning from multi-modality
MRI images”, In 2018 IEEE 15th international symposium on biomedical imaging (ISBI
2018), pp. 277-280, 2018.
[61] Sundar, K. S., Rajamani, K. T., & Sai, S. S. S., “Exploring image classification of thyroid
ultrasound images using deep learning”, In International Conference on ISMAC in
Computational Vision and Bio-Engineering, pp. 1635-1641, Springer, Cham, 2018.
[62] Wang, J., Li, S., Song, W., Qin, H., Zhang, B., & Hao, A., “Learning from weakly-labeled
clinical data for automatic thyroid nodule classification in ultrasound images”, In 2018
25th IEEE International Conference on Image Processing (ICIP), pp. 3114-3118, 2018.
[63] Chen, D., Zhang, J., & Li, W., “Thyroid Nodule Classification Using Two Levels
Attention-Based Bi-Directional LSTM with Ultrasound Reports”, In 2018 9th
International Conference on Information Technology in Medicine and Education (ITME),
pp. 309-312, 2018.
[64] Li, Z., Qin, J., Zhang, X., & Wan, Y., “A Hybrid Intelligent Framework for Thyroid
Diagnosis”, In Cyberspace Data and Intelligence, and Cyber-Living, Syndrome, and
Health, pp. 441-451, Springer, Singapore, 2019.
[65] Ko, S. Y., Lee, J. H., Yoon, J. H., Na, H., Hong, E., Han, K., ... & Kwak, J. Y., “Deep
convolutional neural network for the diagnosis of thyroid nodules on ultrasound”, Head &
neck, vol. 41(4), pp. 885-891, 2019.
[66] Wang, L., Yang, S., Yang, S., Zhao, C., Tian, G., Gao, Y., ... & Lu, Y., “Automatic thyroid
nodule recognition and diagnosis in ultrasound imaging with the YOLOv2 neural
network”, World journal of surgical oncology, vol. 17(1), pp. 1-9, 2019.
[67] Liu, T., Guo, Q., Lian, C., Ren, X., Liang, S., Yu, J., ... & Shen, D., “Automated detection
and classification of thyroid nodules in ultrasound images using clinical-knowledge-
guided convolutional neural networks”, Medical image analysis, vol. 58, 2019.
79
[68] Aboudi, N., Guetari, R., & Khlifa, N., “Multi-objectives optimization of features selection
for the classification of thyroid nodules in ultrasound images”, IET Image Processing,
2020.
[69] Guo, X., Zhao, H., & Tang, Z., “An Improved Deep Learning Approach for Thyroid
Nodule Diagnosis”, In 2020 IEEE 17th International Symposium on Biomedical Imaging
(ISBI), pp. 296-299, 2020.
[70] Ajilisa, O. A., Jagathyraj, V. P., & Sabu, M. K, “Computer-Aided Diagnosis of Thyroid
Nodule from Ultrasound Images Using Transfer Learning from Deep Convolutional
Neural Network Models”, In 2020 Advanced Computing and Communication
Technologies for High Performance Applications (ACCTHPA), pp. 237-241, 2020.
[71] Nguyen, D. T., Kang, J. K., Pham, T. D., Batchuluun, G., & Park, K. R., “Ultrasound
image-based diagnosis of malignant thyroid nodule using artificial intelligence”, Sensors,
vol. 20(7), 2020.
[72] Vasile, C. M., Udriștoiu, A. L., Ghenea, A. E., Popescu, M., Gheonea, C., Niculescu, C.
E., ... & Alexandru, D. O., “Intelligent Diagnosis of Thyroid Ultrasound Imaging Using an
Ensemble of Deep Learning Methods”, Medicina, vol. 57(4), 2021.
[73] Zhu, Y. C., AlZoubi, A., Jassim, S., Jiang, Q., Zhang, Y., Wang, Y. B., ... & Hongbo, D.
U., “A generic deep learning framework to classify thyroid and breast lesions in ultrasound
images”, Ultrasonics, vol. 110, 2021.
[74] Yang, W., Dong, Y., Du, Q., Qiang, Y., Wu, K., Zhao, J., ... & Zia, M. B., “Integrate
domain knowledge in training multi-task cascade deep learning model for benign–
malignant thyroid nodule classification on ultrasound images”, Engineering Applications
of Artificial Intelligence, vol. 98, 2021.
[75] Dabowsa, N. I. A., Amaitik, N. M., Maatuk, A. M., & Aljawarneh, S. A., “A hybrid
intelligent system for skin disease diagnosis”, In 2017 International Conference on
Engineering and Technology (ICET), pp. 1-6, 2017.
[76] Liu, T., Xie, S., Zhang, Y., Yu, J., Niu, L., & Sun, W., “Feature selection and thyroid
nodule classification using transfer learning”, In 2017 IEEE 14th International Symposium
on Biomedical Imaging (ISBI 2017), pp. 1096-1099, 2017.
[77] Xia, J., Chen, H., Li, Q., Zhou, M., Chen, L., Cai, Z., ... & Zhou, H., “Ultrasound-based
differentiation of malignant and benign thyroid Nodules: An extreme learning machine
approach”, Computer methods and programs in biomedicine, vol. 147, pp. 37-49, 2017.
[78] Nguyen, D. T., Pham, T. D., Batchuluun, G., Yoon, H. S., & Park, K. R., “Artificial
80
intelligence-based thyroid nodule classification using information from spatial and
frequency domains”, Journal of clinical medicine, vol. 8(11),2019.
[79] Qin, P., Wu, K., Hu, Y., Zeng, J., & Chai, X., “Diagnosis of benign and malignant thyroid
nodules using combined conventional ultrasound and ultrasound elasticity imaging”, IEEE
journal of biomedical and health informatics, vol. 24(4), pp. 1028-1036, 2019.
[80] Xie, J., Guo, L., Zhao, C., Li, X., Luo, Y., & Jianwei, L., “A Hybrid Deep Learning and
Handcrafted Features based Approach for Thyroid Nodule Classification in Ultrasound
Images”, In Journal of Physics: Conference Series, vol. 1693, 2020.
[81] Sun, H., Yu, F., & Xu, H., “Discriminating the Nature of Thyroid Nodules Using the
Hybrid Method”, Mathematical Problems in Engineering, 2020.
[82] Hang, Y, “Thyroid nodule classification in ultrasound images by fusion of conventional
features and res-GAN deep features”, Journal of Healthcare Engineering, 2021.
[83] Srivastava, R., Kumar, P., “A hybrid model for the identification and classification of
thyroid nodules in medical ultrasound images”, International Journal of Modelling,
Identification and Control (IJMIC), 2021.
[84] Kwak, J. Y., Han, K. H., Yoon, J. H., Moon, H. J., Son, E. J., Park, S. H., & Kim, E. K,
“Thyroid imaging reporting and data system for US features of nodules: a step in
establishing better stratification of cancer risk.”, Radiology, vol. 260(3), pp. 892-899,
2011.
[85] Karanja, E. M., Masupe, S., & Jeffrey, M. G, “Analysis of internet of things malware using
image texture features and machine learning techniques”, Internet of Things, vol. 9, pp.
100-153, 2020.
[86] Garg, M., & Dhiman, G, “A novel content-based image retrieval approach for
classification using GLCM features and texture fused LBP variants”, Neural Computing
and Applications, pp. 1-18, 2020.
[87] Chapelle, O., Haffner, P., & Vapnik, V. N, “Support vector machines for histogram-based
image classification”, IEEE transactions on Neural Networks, vol. 10(5), pp. 1055-1064,
1999.
[88] Keramidas, E. G., Maroulis, D., & Iakovidis, D. K, “ΤND: a thyroid nodule detection
system for analysis of ultrasound images and videos”, Journal of medical systems, vol.
36(3), pp. 1271-1281, 2012.
[89] Zhang, B., Tian, J., Pei, S., Chen, Y., He, X., Dong, Y., & Zhang, S, “Machine learning–
assisted system for thyroid nodule diagnosis”, Thyroid, vol. 29(6), pp. 858-867, 2019.
81
[90] Jothi, A. A. J., & Rajam, M. A, “Automatic classification of thyroid histopathology images
using multi-classifier system”, Multimedia Tools and Applications, vol. 76(18), pp. 18711-
18730, 2017.
[91] Zhou, H., Zhang, J., Zhou, Y., Guo, X., & Ma, Y, “A feature selection algorithm of
decision tree based on feature weight”, Expert Systems with Applications, vol.164, 113842,
2021.
[92] Abiodun, O. I., Jantan, A., Omolara, A. E., Dada, K. V., Mohamed, N. A., & Arshad, H,
“State-of-the-art in artificial neural network applications: A survey”, Heliyon, vol. 4(11),
e00938, 2018.
[93] A. Liaw, and M. Wiener, “Classification and regression by random Forest”, R news, vol.
2, pp. 18-22, 2002.
[94] Gu, J., Wang, Z., Kuen, J., Ma, L., Shahroudy, A., Shuai, B., ... & Chen, T, “Recent
advances in convolutional neural networks”, Pattern recognition, vol. 77, pp. 354-377,
2018.
[95] Zhou, D. X, “Theory of deep convolutional neural networks: Down sampling”, Neural
Networks, vol. 124, pp. 319-327, 2020.
[96] Srivastava, R., Kumar, P, “A contemporary review on soft computing techniques for
thyroid disorder identification and detection”, International Journal of Computer
Applications in Technology (IJCAT), 2021.
[97] Kang, M., & Tian, J, “Machine Learning: Data Pre‐processing. Prognostics and Health
Management of Electronics: Fundamentals”, Machine Learning, and the Internet of
Things, pp. 111-130, 2018.
[98] Masoudi, S., Harmon, S. A., Mehralivand, S., Walker, S. M., Raviprakash, H., Bagci, U.,
... & Turkbey, B, “Quick guide on radiology image pre-processing for deep learning
applications in prostate cancer research”, Journal of Medical Imaging, vol. 8(1), 010901,
2021.
[99] Qi, P. F., Meng, Q. H., & Zeng, M, “A CNN-based simplified data processing method for
electronic noses”, In 2017 ISOCS/IEEE International Symposium on Olfaction and
Electronic Nose (ISOEN), pp. 1-3, IEEE, 2017.
[100] Pitaloka, D. A., Wulandari, A., Basaruddin, T., & Liliana, D. Y, “Enhancing CNN with
preprocessing stage in automatic emotion recognition”, Procedia computer science,
vol.116, pp. 523-529, 2017.
[101] Vimal Kurup, R., Sowmya, V., & Soman, K. P, “Effect of data pre-processing on brain
82
tumor classification using capsulenet”, In International Conference on Intelligent
Computing and Communication Technologies, pp. 110-119, 2019.
[102] Haryanto, T., Wasito, I., & Suhartanto, H, “Convolutional Neural Network (CNN) for
gland images classification”, In 2017 11th International Conference on Information &
Communication Technology and System (ICTS), pp. 55-60, IEEE, 2017.
[103] Rajan, S., Chenniappan, P., Devaraj, S., & Madian, N, “Novel deep learning model for
facial expression recognition based on maximum boosted CNN and LSTM”, IET Image
Processing, vol. 14(7), pp. 1373-1381, 2020.
[104] Soomro, T. A., Afifi, A. J., Shah, A. A., Soomro, S., Baloch, G. A., Zheng, L., & Gao, J,
“Impact of image enhancement technique on CNN model for retinal blood vessels
segmentation”, IEEE Access, vol. 7, pp. 158183-158197, 2019.
[105] Garg, G., Kumar, D., Sonker, Y., & Garg, R, “A Hybrid MLP-SVM Model for
Classification using Spatial-Spectral Features on Hyper-Spectral Images”, arXiv preprint
arXiv:2101.00214, 2021.
[106] Guo, T., Dong, J., Li, H., & Gao, Y, “Simple convolutional neural network on image
classification”, In 2017 IEEE 2nd International Conference on Big Data Analysis
(ICBDA), pp. 721-724, IEEE, 2017.
[107] Han, D., Liu, Q., & Fan, W, “A new image classification method using CNN transfer
learning and web data augmentation”, Expert Systems with Applications, vol. 95, pp. 43-
56, 2018.
[108] Lei, X., Pan, H., & Huang, X, “A dilated CNN model for image classification”, IEEE
Access, vol. 7, pp. 124087-124095, 2019.
[109] Li, Q., Cai, W., Wang, X., Zhou, Y., Feng, D. D., & Chen, M, “Medical image
classification with convolutional neural network”,In 2014 13th international conference
on control automation robotics & vision (ICARCV), pp. 844-848, IEEE, 2014.
[110] Sultana, F., Sufian, A., & Dutta, P, “Advancements in image classification using
convolutional neural network”, In 2018 Fourth International Conference on Research in
Computational Intelligence and Communication Networks (ICRCICN), pp. 122-129,
IEEE, 2018.
[111] Jmour, N., Zayen, S., & Abdelkrim, A, “Convolutional neural networks for image
classification”, In 2018 international conference on advanced systems and electric
technologies (IC_ASET), pp. 397-402, 2018.
[112] Li, W., Chen, C., Zhang, M., Li, H., & Du, Q, “Data augmentation for hyperspectral image
83
classification with deep CNN”, IEEE Geoscience and Remote Sensing Letters, vol. 16(4),
pp. 593-597, 2018.
[113] Alkhaleefah, M., & Wu, C. C, “A hybrid CNN and RBF-based SVM approach for breast
cancer classification in mammograms” In 2018 IEEE International Conference on
Systems, Man, and Cybernetics (SMC), pp. 894-899, 2018.
[114] Sun, X., Liu, L., Li, C., Yin, J., Zhao, J., & Si, W, “Classification for remote sensing data
with cimproved CNN-SVM method”, IEEE Access, vol. 7, pp. 164507-164516, 2019.
[115] Rosasco, L., De Vito, E., Caponnetto, A., Piana, M., & Verri, A, “Are loss functions all
the same?”, Neural computation, vol. 16(5), pp. 1063-1076, 2004.
[116] Ilhan, U., & Ilhan, A, “Brain tumor segmentation based on a new threshold approach”,
Procedia computer science, vol.120, pp. 580-587, 2017.
[117] Yu, S., Jia, S., & Xu, C, “Convolutional neural networks for hyperspectral image
classification”, Neurocomputing, vol. 219, pp. 88-98, 2017.
[118] Priyadarshini, I., & Cotton, C, “A novel LSTM–CNN–grid search-based deep neural
network for sentiment analysis”, The Journal of Supercomputing, vol. 77(12), pp. 13911-
13932, 2021.
[119] Louverdis, G., Vardavoulia, M. I., Andreadis, I., & Tsalides, P, “A new approach to
morphological color image processing”, Pattern recognition, vol. 35(8), pp. 1733-1741,
2002.
[120] Merzban, M. H., & Elbayoumi, M, “Efficient solution of Otsu multilevel image
thresholding: A comparative study”, Expert Systems with Applications, vol. 116, pp.299-
309, 2019.
[121] Araujo, R. L., de Araujo, F. H., & Silva, R. R, “Automatic segmentation of melanoma skin
cancer using transfer learning and fine-tuning”, Multimedia Systems, vol. 28(4), pp. 1239-
1250, 2022.
[122] Khan, M. Q., Hussain, A., Rehman, S. U., Khan, U., Maqsood, M., Mehmood, K., & Khan,
M. A, “Classification of melanoma and nevus in digital images for diagnosis of skin
cancer”, IEEE Access, vol.7, pp. 90132-90144, 2019.
[123] Kass, M., Witkin, A., & Terzopoulos, D, “Snakes: Active contour models”, International
journal of computer vision, vol.1(4), pp. 321-331, 1988.
[124] Bovik, A. C, “Basic binary image processing. In The Essential Guide to Image
Processing”, Academic Press, pp. 69-96, 2009.
[125] He, K., Zhang, X., Ren, S., & Sun, J, “Deep residual learning for image recognition”, In
84
Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770-
778, 2016.
[126] Fulton, L. V., Dolezel, D., Harrop, J., Yan, Y., & Fulton, C. P, “Classification of
Alzheimer’s disease with and without imagery using gradient boosted machines and Res-
Net-50”, Brain sciences, vol.9(9), pp. 1-16, 2019.
[127] Yao, X., Wang, X., Wang, S. H., & Zhang, Y. D, “A comprehensive survey on
convolutional neural network in medical image analysis”, Multimedia Tools and
Applications, pp. 1-45, 2020.
[128] Krizhevsky, A., Sutskever, I., & Hinton, G. E, “Imagenet classification with deep
convolutional neural networks”, Communications of the ACM, vol. 60(6), pp. 84-90, 2017.
[129] Simonyan, K., & Zisserman, A, “Very deep convolutional networks for large-scale image
recognition”, arXiv preprint arXiv:1409.1556, 2014.
[130] Liu, Y., Yin, B., Yu, J., & Wang, Z, “Image classification based on convolutional neural
networks with cross-level strategy”, Multimedia Tools and Applications, vol.76(8), pp.
11065-11079, 2017.
[131] Srivastava, S., Kumar, P., Chaudhry, V., & Singh, A, “Detection of ovarian cyst in
ultrasound images using fine-tuned VGG-16 deep learning network”, SN Computer
Science, vol.1(2), pp.1-8, 2020.
[132] Hosny, K. M., Kassem, M. A., & Fouad, M. M, “Classification of skin lesions into seven
classes using transfer learning with AlexNet”, Journal of digital imaging, vol. 33(5), pp.
1325-1334, 2020.
[133] Islam, S., Khan, S. I. A., Abedin, M. M., Habibullah, K. M., & Das, A. K, “Bird species
classification from an image using VGG-16 network”, In Proceedings of the 2019 7th
international conference on computer and communications management, pp. 38-42, 2019.
[134] Liu, X. Y., Fang, Y., Yang, L., Li, Z., & Walid, A, “High-performance tensor
decompositions for compressing and accelerating deep neural networks”, In Tensors for
Data Processing, Academic Press, pp. 293-340, 2022.
[135] Rezaee, M., Zhang, Y., Mishra, R., Tong, F., & Tong, H, “Using a vgg-16 network for
individual tree species detection with an object-based approach”, In 2018 10th IAPR
Workshop on Pattern Recognition in Remote Sensing (PRRS) pp. 1-7, 2018.
[136] Montavon, G., Samek, W., & Müller, K. R, “Methods for interpreting and understanding
deep neural networks”, Digital signal processing, vol. 73, pp. 1-15, 2018.
[137] Mahmood, A., Bennamoun, M., An, S., Sohel, F., Boussaid, F., Hovey, R., ... & Fisher, R.
85
B, “Deep learning for coral classification. In Handbook of neural computation, Academic
Press, pp. 383-401, 2017.
[138] Bingham, E., Kaski, S., Laaksonen, J., & Lampinen, J, “Advances in independent
component analysis and learning machines”, Academic Press, 2015.
[139] Shams, S., Platania, R., Zhang, J., Kim, J., Lee, K., & Park, S. J, “Deep generative breast
cancer screening and diagnosis”, In International Conference on Medical Image
Computing and Computer-Assisted Intervention, pp. 859-867, Springer, 2018.
[140] Lipkind, H. S., Vazquez-Benitez, G., DeSilva, M., Vesco, K. K., Ackerman-Banks, C.,
Zhu, J., ... & Kharbanda, E. O, “Receipt of COVID-19 vaccine during pregnancy and
preterm or small-for-gestational-age at birth—Eight Integrated Health Care
Organizations”, Morbidity and Mortality Weekly Report, United States, December 15,
2020–July 22, 2021, vol. 71(1), 26, 2022.
[141] Mahajan, H. B., Rashid, A. S., Junnarkar, A. A., Uke, N., Deshpande, S. D., Futane, P. R.,
... & Alhayani, B, "Integration of Healthcare 4.0 and blockchain into secure cloud-based
electronic health records systems", Applied Nanoscience, vol.13(3), pp.2329-2342, 2023.
[142] Chengoden, R., Victor, N., Huynh-The, T., Yenduri, G., Jhaveri, R. H., Alazab, M., ... &
Gadekallu, T. R, "Metaverse for healthcare: A survey on potential applications, challenges
and future directions", IEEE Access, 2023.
[143] Dicuonzo, G., Donofrio, F., Fusco, A., & Shini, M, "Healthcare system: Moving forward
with artificial intelligence", Technovation, vol. 120, 102510, 2023.
[144] Santana, I. R., Mason, A., Gutacker, N., Kasteridis, P., Santos, R., & Rice, N, "Need,
demand, supply in health care: working definitions, and their implications for defining
access", Health Economics, Policy and Law, vol. 18(1), pp. 1-13, 2023.
[145] Ali, O., Abdelbaki, W., Shrestha, A., Elbasi, E., Alryalat, M. A. A., & Dwivedi, Y. K, "A
systematic literature review of artificial intelligence in the healthcare sector: Benefits,
challenges, methodologies, and functionalities", Journal of Innovation & Knowledge, vol.
8(1), 100333, 2023.
[146] Rumelhart, D. E., Hinton, G. E., & Williams, R. J, “Learning representations by back-
propagating errors”, Nature, vol. 323(6088), pp. 533-536, 1986.
[147] Wang, D., Lu, Z., Xu, Y., Wang, Z., Santella, A., & Bao, Z, “Cellular structure image
classification with small targeted training samples”, IEEE Access, vol.7, pp. 148967-
148974, 2019.
[148] Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., ... &
86
Bengio, Y, “Generative adversarial nets”, Advances in neural information processing
systems, vol. 27, 2014.
[149] Aggarwal, A., Mittal, M., & Battineni, G, “Generative adversarial network: An overview
of theory and applications” International Journal of Information Management Data
Insights, vol.1(1), 100004, 2021.
[150] Alqahtani, H., Kavakli-Thorne, M., & Kumar, G, “Applications of generative adversarial
networks (gans): An updated review”, Archives of Computational Methods in Engineering,
vol.28(2), pp. 525-552, 2021.
[151] https://developers.google.com/machine-learning/gan/gan_structure
[152] Akbar, S. B., Thanupillai, K., & Sundararaj, S, “Combining the advantages of Alex-Net
convolutional deep neural network optimized with anopheles search algorithm-based
feature extraction and random forest classifier for COVID‐19 classification”, Concurrency
and Computation: Practice and Experience, pp. 1-15, 2022.
[153] Rasheed, A., Ali, N., Zafar, B., Shabbir, A., Sajid, M., & Mahmood, M. T, “Handwritten
Urdu characters and digits recognition using transfer learning and augmentation with
AlexNet”, IEEE Access, vol. 10, pp. 102629-102645, 2022.
[154] Hammad, I., & El-Sankary, K, “Impact of approximate multipliers on VGG deep learning
network”, IEEE Access, vol. 6, pp. 60438-60444, 2018.
[155] Jibhakate, P., Gole, S., Yeskar, P., Rangwani, N., Vyas, A., & Dhote, K, “Early Glaucoma
Detection Using Machine Learning Algorithms of VGG-16 and Resnet-50”, In 2022 IEEE
Region 10 Symposium (TENSYMP), pp. 1-5, 2022.
87
LIST OF PUBLICATIONS
Journal(s)
Published
1. Srivastava, R., & Kumar, P., “Deep-GAN: an improved model for thyroid nodule
identification and classification”, Neural Computing and Application, Springer.
Index: SCIE, Scopus, ACM digital library, Impact factor: 6
2. Srivastava, R., & Kumar, P., “Optimizing CNN Based Model for Thyroid Nodule
Classification Using Data Augmentation, Segmentation and Boundary Detection
Techniques”, Multimedia Tools and Applications, Springer.
Index: SCIE, Scopus, ACM digital library, Impact factor:3.6
3. Srivastava, R., & Kumar, P., “A hybrid model for the identification and classification of
benign and malignant thyroid nodules on medical USG image”, International Journal of
Modelling, Identification and Control, Inderscience., 2022.
Index: ESCI, Scopus, ACM digital library
DOI: 10.1504/IJMIC.2022.10051644
4. Srivastava, R., & Kumar, P., “A contemporary review on soft computing techniquesfor
thyroid identification and detection”, International Journal of Computer Applications in
Technology (IJCAT), Inderscience, 2022.
Index: ESCI, Scopus, ACM digital library
DOI: 10.1504/IJCAT.2022.10051647
5. Srivastava, R., & Kumar, P., “GSO-CNN based model for the identification and
classification of thyroid nodule in medical USG images”, Network Modeling Analysis in
Health Informatics and Bioinformatics (NHIB), Springer, 2022.
Index: ESCI, Scopus, ACM digital library
DOI: 10.1007/s13721-022-00388-w
6. Srivastava, R., & Kumar, P., “A CNN-SVM hybrid model for the classification of thyroid
nodules in medical ultrasound images”, International Journal of Grid and Utility
Computing (IJGUC), Inderscience, 2022.
Index: ESCI, Scopus, ACM digital library
88
DOI: 10.1504/IJGUC.2022.10052488
7. Srivastava, R., & Kumar, P., “Performance comparison of various machine learning
algorithms for the identification and classification of thyroid nodules”, International
Journal of Grid and Utility Computing (IJGUC), Inderscience.
[Index: ESCI, Scopus, ACM digital library]
Communicated
1. Srivastava, R., & Kumar, P., “Alex-GAN based synthetic ultrasound images
augmentation model for increasing the performance of thyroid nodules classification”,
Multimedia Tools and Applications.
[Submitted] [Index: SCIE, Scopus, ACM digital library, Impact factor:3.6]
Conference(s)
1. Srivastava, R., & Kumar, P, “BL_SMOTE ensemble method for prediction of thyroid
disease on imbalanced classification problem”, In Proceedings of Second International
Conference on Computing, Communications, and Cyber-Security, pp. 731-741, Springer,
Singapore.
[Index: Scopus, DBLP, Google Scholar, EI-Compendex, SCImago]
DOI: https://doi.org/10.1007/978-981-16-0733-2_52
2. Srivastava, R., & Kumar, P, “A novel model for novel model for the identification and
classification of thyroid nodules using deep neural network”, 4th International Conference
on Machine Intelligence and Signal Processing (MISP2022), NIT Raipur, Springer, 12-
14 March 2022.
[Index: Scopus, DBLP, Google Scholar, EI-Compendex, SCImago]
89