Deep Convolutional Neural Network Framework With Multi-Modal Fusion For Alzheimer's Detection
Deep Convolutional Neural Network Framework With Multi-Modal Fusion For Alzheimer's Detection
Corresponding Author:
Manoj Kumar Sharma
Amity School of Engineering and Technology, Amity University
SP-1, Kant Kalwar, RIICO Industrial Area, NH-11C, Jaipur, Rajasthan, India
Email: [email protected]
1. INTRODUCTION
Alzheimer’s disease (AD) is a neurological brain condition that permanently damages the brain cells
that are responsible for thinking and remembering. In the United States, AD affects around 5.7 million people,
making it the sixth biggest cause of mortality, according to facts and figures from 2018 [1]. The datasets
of magnetic resonance imaging (MRI) as modality-1 and positron emission tomography (PET) as modality-
2 are used to diagnose AD. These modalities are combined to produce a dataset that is significantly more
varied and trustworthy [2]. The data-fused dataset contributes to the robustness of the deep learning (DL)
models. Although there are several categorization algorithms in use today, DL has captured the attention of all
academics due to its adaptability and ability to generate the best results [3].
The study also found that the DL network was able to detect early signs of Alzheimer’s in brain
images with high accuracy. These findings suggest that DL networks could be used to develop early detection
and prediction tools for AD [4]. Accuracy increased with a hybrid architecture built on transfer learning. High-
level features like edges, patterns, and other features are easily recognized by pre-trained models [5]. The deep
neural networks (DNN) based models’ performance is highly influenced by their hyper-parameters [6].
The performance of DNNs is heavily dependent on the hyper-parameters, which are values set be-
fore the training process, including the learning rate, batch size, and number of layers. The optimal hyper-
parameters can result in a more efficient and robust model, but it is much harder to identify the optimal values
for the hyper-parameters than it seems. With the advancements in the DNN’s architectures, the need for ef-
fective optimization algorithms to search for optimal values also increased and became more important than
ever. Recently, nature-inspired optimization algorithms are one such algorithm that has helped researchers
immensely in this regard [7]. Therefore, researchers dealing with extensive and intricate datasets in DL mod-
els have come to rely on nature-inspired optimization algorithms as a crucial tool. The rest of the paper is
structured as follows; section 2 includes a review of recent and highly relevant literature; section 3 explains
the suggested approach; section 4 lists the availability of the data and materials used; section 5 describes the
results and discussion; and sections 6 discuss the conclusion, limitations, and future scope of the current work
respectively.
2. LITERATURE REVIEW
Islam and Zhang [8] created an extremely deep convolutional model and displayed the outcomes on
the open access series of imaging studies (OASIS) database. Brain MRI dataset is used to detect and classify
AD (critical neurological brain disorder) through the very deep convolutional network (DCN). The proposed
model is based on the pre-trained CNN model named Inception and the parameters i.e., weights were optimized
using a gradient-based optimization algorithm named root-mean-squared-propagation (RMSProp).
Ghoniem [9] proposed a DL approach to diagnosing liver cancer. These are two key contributions of
this method. Firstly, segNet is used to separate the liver from the abdominal scans, U-net model is used for
lesion extraction, and artificial bee colony (ABC) optimization named SegN et + U N et + ABC is used for the
proposed novel hybrid segmentation technique to extract liver lesions. Secondly, a hybrid technique proposed
named LeN et + 5 + ABC is used to extract features and classify the liver lesions. The final result shows
that the SegN et + U N et + ABC technique is better compared to other techniques regarding convergence
time, dice index, correlation coefficient, and jaccard index. The leNet-5/ABC model performs better regarding
computational time, F-1 score, accuracy, and specificity. Ismael et al. [10] proposed an enhanced approach
of residual networks to classify brain tumor types. The proposed model is evaluated on a benchmark dataset
having 3,064 MRI images of three brain tumor types (meningiomas, gliomas, and pituitary). On the same
dataset, the proposed model’s accuracy of 98% was the highest. Joo et al. [11] developed a DL method for
automatic detection and localization of intracranial aneurysms and evaluation of the performance. A three-
dimensional framework (ResNet) related to the DL algorithm is determined by the trained set. The results
gave positive predictive, sensitivity, and specificity of 91.5%, 85.7%, and 98.0% for the external testing set and
92.8%, 87.1%, and 92.0% for the internal testing set, respectively.
Kim et al. [12] developed a computer-assisted detection scheme with the help of a convolutional neu-
ral network (CNN)-based model on an image of 3D digital-subtraction angiography for smaller-size aneurysm
ruptures. A retrospective dataset comprising 368 subjects was utilized as a training cohort for CNNs with
the TensorFlow platform. Six-direction aneurysm image of each patient is attained and region-of-interest is
extracted from each image. Jnawali et al. [13] presented DNN-based to predict brain hemorrhage, based
on the CT imagery data. The presented architecture’s first three-dimensional CNN is used to extract fea-
tures and detect brain hemorrhage using logistic function as the last layer of the network. Finally, proposed
three different 3D CNN algorithms to improve the performance of machine learning (ML) algorithms. Shi
et al. [14] proposed a specific DL based method that has a good understanding of image quality and is vali-
dated with various architectures. Several experiments are conducted in cohorts, externally, and internally, in
which it achieves an improved lesion in terms of enhancement and sensitivity on the subject level. Chen et
al. [15] presented an artificial intelligence technology to improve the performance of the magnetic-induction-
tomography (MIT) inverse problem. Four DL methods, including stacked autoencoders (SAE), deep belief
networks (DBN), denoising autoencoders (DAE), and restricted boltzmann machines (RBM) are used to solve
Int J Reconfigurable & Embedded Syst, Vol. 13, No. 1, March 2024: 179–191
Int J Reconfigurable & Embedded Syst ISSN: 2089-4864 ❒ 181
the nonlinear recreation problem of MIT, and then the results of the back-projection method and DL methods
are compared. Solorio-Ramı́rez et al. [16] presented a new pattern identification algorithm based on the im-
plementation of minimalist-machine-learning (MML) and a higher relevant attribute selection technique called
dMeans. Afterward, to conduct the identification through CT brain images the proposed algorithm performance
is examined and compared with k-nearest neighbors (KNN), multilayer perceptron (MLP), Naı̈ve Bayes (NB),
AdaBoost, random forests (RF), and support vector machine (SVM) classifiers. Phan et al. [17] presented a
new method based on the DL algorithm and hounsfield unit system. The proposed method not only describes
the level and duration of hemorrhage but also classifies the brain hemorrhagic region on the MRI image. To
select the most suitable method for classification three neural network systems are compared and evaluated.
Due to its importance in medical diagnostics, computer vision, and the internet of things, multimodal medical
imaging has become a hot research area in the scientific community in recent years [18]-[20]. In order to detect
AD progression based on the late fusion of MRI, demographics, neuropsychological, and apolipoprotein E4
(APOe4) genetic data, Spasov et al. [21] suggested a multimodal single-task classification model based on
a CNN. Kumar et al. [22] integration of anatomical and functional modalities for the early identification of
malignant tissue is one of the significant clinical applications of medical imaging fusion.
3. PROPOSED METHOD
In this section, the proposed method framework for AD detection was provided by the author. The
multi-modal datasets were downloaded from the website and stored on the hard drive. These stored Alzheimer’s
databases were manually separated into two modalities MRI and PET scan on the basis of patients with and
without AD. Then these images are pre-processed with format changing, image registration, segmentation, and
resizing done through MATLAB code. After pre-processing, the fusion process was implemented and the fused
data were stored in a MATLAB drive. Using this augmented datastore of fused images, the DCNN custom and
pre-trained networks are trained, validated, and evaluated. To achieve the best outcomes, the nature-inspired
particle swarm optimization (PSO) and Bayesian algorithm are used with custom and pre-trained models for
hyper-parameter tuning. Results were eventually gathered and evaluated. The multi-modal data fusion process
and optimization workflow of the system is shown in Figure 1 and can be observed from top to bottom. In
this paper, the author has used two databases Alzheimer’s disease neuroimaging initiative (ADNI) and Kaggle.
The pre-processing was done on MRI and PET.dicom images that were converted into the .jpg format using
the MATLAB program. Images that have been converted to JPEG format can be analyzed and stored more
effectively, which raises diagnostic accuracy [23]. Relying less on specialized DICOM image viewer tools to
see medical images [24]. The workflow of the suggested technique is briefly outlined. The MRI and PET
images were initially pre-processed and converted to JPEG format before being used in the multi-modal data
fusion technique. After that, nature-inspired optimization techniques and conventional optimization techniques
were utilized to optimize the hyper-parameters. After that, the custom and pre-trained models were trained
with and without optimized hyper-parameters on the fused datasets. Finally, these trained models were tested
and the outcomes of each model can be compared and evaluated to determine the most effective approach.
This workflow has the potential to improve the accuracy and reliability of ML models in medical imaging
applications, allowing for more precise diagnoses and treatment planning.
MRI scans provide a detailed description of the brain, including gray and white matter, and PET scans
measure levels of certain metabolites in the brain. Combining these two data sources provides a powerful tool
for accurately diagnosing and predicting AD [25]. Then these two multi-modal images were undergone through
the fusion process and the fused database was created. The use of multiple modalities in data collection helps
to mitigate the impact of any inherent biases that may exist in a single modality. By merging different sources
of data, a more holistic perspective of the subject matter can be attained, resulting in a more thorough compre-
hension of it. Figure 2 is a pictorial representation of the steps followed from data collection to categorizing
the fused datasets in train and test folders for both ADNI and Kaggle fused datasets. An interactive and simple
fusion process is implemented in MATLAB, as demonstrated in Figure 3. This graphical user interface (GUI)
in MATLAB, which was made using the MATLAB app designer named data fusion, is used to achieve the
fusion process. These fused datasets were utilized to train the pre-trained deep convolutional neural networks
(DCNN) like custom CNN, AlexNet, MobileNetV2, and GoogLeNet using a DL toolbox in MATLAB. Addi-
tionally, the use of a GUI for data fusion can help reduce the time and effort required for data preprocessing,
enabling more efficient experimentation and analysis of multi-modal data.
Deep convolutional neural network framework with multi-modal fusion for ... (Manoj Kumar Sharma)
182 ❒ ISSN: 2089-4864
Multi-modal
Alzheimer
Datastore
Modality-1 Modality-2
Pre-processing
and cleaning of
the data
Fusion Process
done by the Matlab App
Designed
Multi-
Modal Data Fusion Process
Fused
Dataset
Vertical Container
Figure 1. The multi-modal data fusion process and nature-inspired hyper-parameters optimization workflow
of our proposed framework for diagnosing AD
Train data
set
MRI
AD
AD Fused
Data
PET Test data
set
Image Fusion
ADNI/Kaggle
Database Train Data
Data MRI set
Preparation NC
NC Fused
Data
PET
Test Data
set
Figure 2. Steps to achieve multi-modal fusion with ADNI and Kaggle databases
Figure 3. A fusion MATLAB app interface is shown to achieve multi-modal fusion with ADNI and Kaggle
databases
Int J Reconfigurable & Embedded Syst, Vol. 13, No. 1, March 2024: 179–191
Int J Reconfigurable & Embedded Syst ISSN: 2089-4864 ❒ 183
The procedures for training and testing the optimized pre-trained models are shown in Figure 4. Before
training the pre-trained CNN models, they are optimized using a nature-inspired algorithm. Then a test dataset
was used to test whether the trained model was performing well or not. If not, then further iterations were
required to optimize the hyper-parameters of the selected DCNN network.
The concept of the transfer learning approach is illustrated in Figure 5. In transfer learning, pre-
trained weights are transferred to predict a new, similar task with some changes in the last layers. This was
accomplished with MATLAB to speed up the training and testing process using transfer learning.
Multi-Modal
Fused Dataset
Transfer Learning
Test the
selected Model
Models
YES Results Stored
Figure 4. The transfer learning approach used on to the DCNN with hyper-parameter optimization techniques
with multi-modal fused datasets
Convolution Layers
Dense Layer
ImageNET
database Classify
>1 million Images
1000 classes
Trained Network
Transfer Learning
Multi-Modal
Fused database
Classify
2 classes
Dense Layer
s
Convolution Layers
Figure 5. The transfer learning approach used to increase the performance of selected DCNN
Deep convolutional neural network framework with multi-modal fusion for ... (Manoj Kumar Sharma)
184 ❒ ISSN: 2089-4864
3.1.2. GoogLeNet
ILSRVRC’2014 was won by Google architecture, which had fewer errors than the runner-up VGG,
and the previous winner AlexNet. The architecture of GoogLeNet consists of 22 layers [30]. The architecture is
a combination of 1×1 convolutional layers, an inception module, global average pooling layers, and auxiliary
classifiers. The concept of 1×1 was used to minimize the parameters, i.e., weights and biases, to lower the
computational cost with a much deeper network. The inception module consists of different sizes of CNN
layers, i.e., 1×1, 3×3, and 5×5, and a max pooling layer of size 3×3, working in parallel to extract deep
features from the objects of different sizes on a larger scale. The auxiliary classifiers are used by the inception
architecture to calculate the loss at different stages during the training and add them to the final loss with weights
valued at 0.3 to generate the overall loss. The auxiliary classifiers assist in overcoming the gradient vanishing
problem and in regularization. Google has been widely used in object detection and face recognition [31], [32].
3.1.3. MobileNetV2
MobileNetV2 is also known as the “lightweight” model, which has a comparatively much lower com-
plexity cost that makes it suitable for mobile devices. The architecture consists of depth-wise convolution
and point-wise convolution. In the depth-wise convolution, a single convolutional filter is applied to each in-
put signal to perform lightweight filtering, whereas, in the point-wise convolution, 1×1 convolution-based is
performed to extract deep features by computing linear combinations between the input channels. Table 1 sum-
marizes and the pre-trained CNN architectures are compared. That was already utilized for object detection
and recognition across vast numbers of classes [33].
Int J Reconfigurable & Embedded Syst, Vol. 13, No. 1, March 2024: 179–191
Int J Reconfigurable & Embedded Syst ISSN: 2089-4864 ❒ 185
Deep convolutional neural network framework with multi-modal fusion for ... (Manoj Kumar Sharma)
186 ❒ ISSN: 2089-4864
(a) (b)
Figure 8. Distribution of participants based on gender and age (a) with AD and (b) in the NC group
Int J Reconfigurable & Embedded Syst, Vol. 13, No. 1, March 2024: 179–191
Int J Reconfigurable & Embedded Syst ISSN: 2089-4864 ❒ 187
(a)
(b)
Figure 9. Montages of MRI and PET fused images for both AD and NC categories sample of
(a) PET MRI AD FUSED IMG image and (b) PET MRI NC FUSED IMG image
Deep convolutional neural network framework with multi-modal fusion for ... (Manoj Kumar Sharma)
188 ❒ ISSN: 2089-4864
The performance comparison of custom and pre-trained DL models on the ADNI fusion dataset
is shown in Figure 10. Figures 10(a) to 10(d) respectively depict the performance of the custom model,
GoogLeNet model, MobileNetV2 model, and AlexNet model. The use of the PSO optimization algorithm
improves the performance of all four models, as observed in the figures. Specifically, the PSO algorithm im-
proves the performance of the custom model in Figure 10(a), GoogLeNet model in Figure 10(b), MobileNetV2
model in Figure 10(c), and AlexNet model in Figure 10(d).
Table 4. Comparative results of using uni-modal and multi-modal datasets in diagnosing Alzheimer
DCNN models Uni-modal (MRI) Multi-modal fused (MRI+PET)
average accuracy (%) average accuracy (%)
GoogLeNet 96.3 97.19
AlexNet 94.39 99.98
ResNet-18 97.51 75.4
MobileNetV2 - 61.84
Custom model - 68.15
Int J Reconfigurable & Embedded Syst, Vol. 13, No. 1, March 2024: 179–191
Int J Reconfigurable & Embedded Syst ISSN: 2089-4864 ❒ 189
Table 6. PSO optimization on the custom model compared with ADNI and Kaggle fused datasets
Model name ADNI fused Kaggle fused Optimization
Average accuracy (%) Average accuracy (%) algorithm
Custom model 68.15 67.706 Without PSO
91.35 83.77 With PSO
(a) (b)
(c) (d)
Figure 10. A comparison of the performance of custom and pre-trained deep learning models on the ADNI
fusion dataset with and without the use of the optimization algorithm (a) the custom model,
(b) the GoogLeNet model, (c) the mobilenetv2 model, and (d) the AlexNet model
6. CONCLUSIONS
In this article, optimized DL models based on an automatic computer-aided AD detection approach
are proposed. The different pre-trained models, including AlexNet, GooLeNet, MobileNetV2, and a custom
model, are assessed using the ADNI and Kaggle datasets. Two optimization algorithms, Bayesian and PSO
are used to optimize the hyper-parameters of the models and the results before and after the optimization
are reported. The performance is evaluated in terms of training accuracy, testing accuracy, validation accuracy,
precision, and recall. It is found that the nature-inspired optimization algorithm i.e., PSO provides better results
on some of the pre-trained models. But when the PSO is applied to the very light custom model can outperform
in comparison to larger pre-trained architectures. This shows that for mobile application development, lighter
customized models should be utilized. The PSO and Bayesian are found to have improved the results by 15%
on average i.e., 2 to 5% in the case of pre-trained models and up to 22% for a custom model. Similarly, the
fused dataset of PET and MRI also contributed to the improvement of overall performance by up to 5%.
ACKNOWLEDGEMENTS
We also acknowledge the support of the ADNI data and other open-source databases for approving
the access request. We are assisted in our research effort by software and relevant MATLAB webinars. We are
grateful to MIT, a private research university in Cambridge, Massachusetts, for offering some of the research
linked to our work that is open-access.
Deep convolutional neural network framework with multi-modal fusion for ... (Manoj Kumar Sharma)
190 ❒ ISSN: 2089-4864
REFERENCES
[1] Alzheimer’s Association, “2019 Alzheimer’s disease facts and figures,” Alzheimer’s & Dementia, vol. 15, no. 3, pp. 321–387, Mar.
2019, doi: 10.1016/j.jalz.2019.01.010.
[2] C. Catana, A. R. Guimaraes, and B. R. Rosen, “PET and MR imaging: the odd couple or a match made in heaven?,” Journal of
Nuclear Medicine, vol. 54, no. 5, pp. 815–824, May 2013, doi: 10.2967/jnumed.112.112771.
[3] D. Prakash, N. Madusanka, S. Bhattacharjee, H.-G. Park, C.-H. Kim, and H.-K. Choi, “A comparative study of Alzheimer’s disease
classification using multiple transfer learning models,” Journal of Multimedia Information System, vol. 6, no. 4, pp. 209–216, Dec.
2019, doi: 10.33851/JMIS.2019.6.4.209
[4] A. Ebrahimi, S. Luo, and R. Chiong, “Introducing transfer learning to 3D ResNet-18 for Alzheimer’s disease detection on MRI
images,” in 2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ), Nov. 2020, pp. 1–6, doi:
10.1109/IVCNZ51579.2020.9290616.
[5] S. R. Chandaran, G. Muthusamy, L. R. Sevalaiappan, and N. Senthilkumaran, “Deep learning-based transfer learning model in
diagnosis of diseases with brain magnetic resonance imaging,” Acta Polytechnica Hungarica, vol. 19, no. 5, pp. 127–147, 2022, doi:
10.12700/APH.19.5.2022.5.7.
[6] H. Bertrand, R. Ardon, M. Perrot, and I. Bloch, “Hyperparameter optimization of deep neural networks: combining hyperband with
Bayesian model selection state of the art,” in Conference sur l’Apprentissage Automatique, 2017, pp. 1–5.
[7] X.-S. Yang, Nature-inspired optimization algorithms, 2nd ed. Academic Press, 2020.
[8] J. Islam and Y. Zhang, “A novel deep learning based multi-class classification method for Alzheimer’s disease detection using brain
MRI data,” in Lecture Notes in Computer Science(), Cham: Springer, 2017, pp. 213–222, doi: 10.1007/978-3-319-70772-3 20.
[9] R. M. Ghoniem, “A novel bio-inspired deep learning approach for liver cancer diagnosis,” Information, vol. 11, no. 2, p. 1-36, Jan.
2020, doi: 10.3390/info11020080.
[10] S. A. A. Ismael, A. Mohammed, and H. Hefny, “An enhanced deep learning approach for brain cancer MRI images classification
using residual networks,” Artificial Intelligence in Medicine, vol. 102, Jan. 2020, doi: 10.1016/j.artmed.2019.101779.
[11] B. Joo et al., “A deep learning algorithm may automate intracranial aneurysm detection on MR angiography with high diagnostic
performance,” European Radiology, vol. 30, no. 11, pp. 5785–5793, Nov. 2020, doi: 10.1007/s00330-020-06966-8.
[12] H. C. Kim et al., “Machine learning application for rupture risk assessment in small-sized intracranial aneurysm,” Journal of Clinical
Medicine, vol. 8, no. 5, pp. 1-10, May 2019, doi: 10.3390/jcm8050683.
[13] K. Jnawali, M. R. Arbabshirani, N. Rao, and A. A. Patel, “Deep 3D convolution neural network for CT brain hemorrhage classifi-
cation,” in Medical Imaging 2018: Computer-Aided Diagnosis, Feb. 2018, doi: 10.1117/12.2293725.
[14] Z. Shi et al., “A clinically applicable deep-learning model for detecting intracranial aneurysm in computed tomography angiography
images,” Nature Communications, vol. 11, no. 1, pp. 1-11, Nov. 2020, doi: 10.1038/s41467-020-19527-w.
[15] R. Chen, J. Huang, Y. Song, B. Li, J. Wang, and H. Wang, “Deep learning algorithms for brain disease detection with magnetic
induction tomography,” Medical Physics, vol. 48, no. 2, pp. 745–759, Feb. 2021, doi: 10.1002/mp.14558.
[16] J.-L. Solorio-Ramı́rez, M. Saldana-Perez, M. D. Lytras, M.-A. Moreno-Ibarra, and C. Yáñez-Márquez, “Brain hemorrhage clas-
sification in CT scan images using minimalist machine learning,” Diagnostics, vol. 11, no. 8, Aug. 2021, doi: 10.3390/diagnos-
tics11081449.
[17] A.-C. Phan, H.-P. Cao, T.-N. Trieu, and T.-C. Phan, “Detection and classification of brain hemorrhage using hounsfield unit and
deep learning techniques,” in Future Data and Security Engineering. Big Data, Security and Privacy, Smart City and Industry 4.0
Applications. FDSE 2020, Singapore: Springer, 2020, pp. 281–293, doi: 10.1007/978-981-33-4370-2 20.
[18] S. Bebortta et al., “Evidence of power-law behavior in cognitive IoT applications,” Neural Computing and Applications, vol. 32, no.
20, pp. 16043–16055, Oct. 2020, doi: 10.1007/s00521-020-04705-0.
[19] W. Tan, P. Tiwari, H. M. Pandey, C. Moreira, and A. K. Jaiswal, “Multimodal medical image fusion algorithm in the era of big data,”
Neural Computing and Applications, Jul. 2020, doi: 10.1007/s00521-020-05173-2.
[20] F. Amato, V. Moscato, A. Picariello, F. Piccialli, and G. Sperlı́, “Centrality in heterogeneous social networks for lurkers detection:
An approach based on hypergraphs,” Concurrency and Computation: Practice and Experience, vol. 30, no. 3, Feb. 2018, doi:
10.1002/cpe.4188.
[21] S. Spasov, L. Passamonti, A. Duggento, P. Liò, and N. Toschi, “A parameter-efficient deep learning approach to predict
conversion from mild cognitive impairment to Alzheimer’s disease,” NeuroImage, vol. 189, pp. 276–287, Apr. 2019, doi:
10.1016/j.neuroimage.2019.01.031.
[22] N. Kumar, N. Hoffmann, M. Oelschlägel, E. Koch, M. Kirsch, and S. Gumhold, “Structural similarity based anatomical and func-
tional brain imaging fusion,” in Multimodal Brain Image Analysis and Mathematical Foundations of Computational Anatomy. MBIA
MFCA 2019, Cham: Springer, 2019, pp. 121–129, doi: 10.1007/978-3-030-33226-6 14.
[23] D. R. Varma, “Managing DICOM images: tips and tricks for the radiologist,” Indian Journal of Radiology and Imaging, vol. 22, no.
01, pp. 4–13, Jan. 2012, doi: 10.4103/0971-3026.95396.
[24] S. Desai and B. S. Usha, “Medical image transcoder for telemedicine based on wireless communication devices,” in 2011 3rd
International Conference on Electronics Computer Technology, Apr. 2011, pp. 389–393, doi: 10.1109/ICECTECH.2011.5941629.
[25] C. Marcus, E. Mena, and R. M. Subramaniam, “Brain PET in the diagnosis of Alzheimer’s disease,” Clinical Nuclear Medicine,
vol. 39, no. 10, pp. e413–e426, Oct. 2014, doi: 10.1097/RLU.0000000000000547.
[26] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” Communications
of the ACM, vol. 60, no. 6, pp. 84–90, May 2017, doi: 10.1145/3065386.
[27] M. Shaha and M. Pawar, “Transfer learning for image classification,” in 2018 Second International Conference on Electronics,
Communication and Aerospace Technology (ICECA), Mar. 2018, pp. 656–660, doi: 10.1109/ICECA.2018.8474802.
[28] L. Xiao, Q. Yan, and S. Deng, “Scene classification with improved AlexNet model,” in 2017 12th International Conference on
Intelligent Systems and Knowledge Engineering (ISKE), Nov. 2017, pp. 1–6, doi: 10.1109/ISKE.2017.8258820.
[29] L. Zhu, Z. Li, C. Li, J. Wu, and J. Yue, “High performance vegetable classification from images based on AlexNet deep
learning model,” International Journal of Agricultural and Biological Engineering, vol. 11, no. 4, pp. 190–196, 2018, doi:
10.25165/j.ijabe.20181104.2690.
[30] C. Szegedy et al., “Going deeper with convolutions,” in 2015 IEEE Conference on Computer Vision and Pattern Recognition
Int J Reconfigurable & Embedded Syst, Vol. 13, No. 1, March 2024: 179–191
Int J Reconfigurable & Embedded Syst ISSN: 2089-4864 ❒ 191
BIOGRAPHIES OF AUTHORS
Dr. Kanad Ray is a professor and head of physics at the Amity School of Applied Sci-
ences Physics Amity University Rajasthan (AUR), Jaipur, India. He has obtained M.Sc. and Ph.D.
degrees in physics from Calcutta University and Jadavpur University, West Bengal, India. In an
academic career spanning over 22 years, he has published and presented research papers in several
national and international journals and conferences in India and abroad. He has authored a book
on the Electromagnetic Field Theory. He current research areas of interest include cognition, com-
munication, electromagnetic field theory, antenna and wave propagation, microwave, computational
biology, and applied physics. Presently he is guiding 8 Ph.D. scholars in various interdisciplinary
fields. He has served as editor of the Springer Book Series. Presently he is an associated editor of
the Journal of Integrative Neuroscience published by IOS Press, Netherlands. He has traveled to
more than a dozen countries on different academic missions. He has established an MOU between
his University and the University of Montreal, Canada for various joint research activities. He has
also established collaboration with National Institute for Materials Science(NIMS), Japan for joint
research activities and visits NIMS as a visiting scientist. He organizes international conference se-
ries such as SoCTA and ICOEVCI as general chair. He is currently an IEEE Executive Committee
member of the Rajasthan Subsection. He can be contacted at email: [email protected].
Deep convolutional neural network framework with multi-modal fusion for ... (Manoj Kumar Sharma)