Duong Et Al. - 2023
Duong Et Al. - 2023
article info a b s t r a c t
Article history: The world has been undergoing the most ever unprecedented circumstances caused by the coronavirus
Received 29 October 2020 pandemic, which is having a devastating global effect in different aspects of life. Since there are not
Received in revised form 2 October 2022 effective antiviral treatments for Covid-19 yet, it is crucial to early detect and monitor the progression
Accepted 14 November 2022
of the disease, thereby helping to reduce mortality. While different measures are being used to
Available online 24 November 2022
combat the virus, medical imaging techniques have been examined to support doctors in diagnosing
Keywords: the disease. In this paper, we present a practical solution for the detection of Covid-19 from chest
COVID-19 X-ray (CXR) and lung computed tomography (LCT) images, exploiting cutting-edge Machine Learning
AI Diagnosis systems techniques. As the main classification engine, we make use of EfficientNet and MixNet, two recently
Expert systems developed families of deep neural networks. Furthermore, to make the training more effective and
Chest X-ray image
efficient, we apply three transfer learning algorithms. The ultimate aim is to build a reliable expert
Lung CT images
system to detect Covid-19 from different sources of images, making it be a multi-purpose AI diagnosing
system. We validated our proposed approach using four real-world datasets. The first two are CXR
datasets consist of 15,000 and 17,905 images, respectively. The other two are LCT datasets with 2,482
and 411,528 images, respectively. The five-fold cross-validation methodology was used to evaluate the
approach, where the dataset is split into five parts, and accordingly the evaluation is conducted in five
rounds. By each evaluation, four parts are combined to form the training data, and the remaining one
is used for testing. We obtained an encouraging prediction performance for all the considered datasets.
In all the configurations, the obtained accuracy is always larger than 95.0%. Compared to various
existing studies, our approach yields a substantial performance gain. Moreover, such an improvement
is statistically significant.
© 2022 Elsevier B.V. All rights reserved.
https://doi.org/10.1016/j.asoc.2022.109851
1568-4946/© 2022 Elsevier B.V. All rights reserved.
L.T. Duong, P.T. Nguyen, L. Iovino et al. Applied Soft Computing 132 (2023) 109851
The proliferation of disruptive Machine Learning (ML) and Generally, kernels of size k(3,3) [12,13], k(5,5) [14], or k(7,7)
especially Deep Learning (DL) algorithms in recent years has [15] are used as filters for deep neural networks. However, larger
enabled a plethora of applications across several domains [3,4]. kernels can potentially improve a model’s accuracy and effi-
Such techniques work on the basis of complex artificial neural ciency. Furthermore, large kernels help to capture high-resolution
networks, which are capable of effectively learning from data patterns, while small kernels allow us to better extract low-
by means of a large number of hyper-parameters distributed in resolution ones. To maintain a balance between accuracy and
different network layers. In this way, they are able to simulate efficiency, the MixNet [8] family has been built based on the
humans’ cognitive functions, aiming to acquire real-world knowl- MobileNets architectures [13,16]. This network family also aims
edge autonomously [5]. In a nutshell, ML/DL techniques are an to reduce the number of parameters as well as FLOPs, i.e., the
advanced paradigm that brings in substantial improvement in metric used to measure the computational complexity counted
performance compared to conventional learning algorithms. In
as the number of float-point operations (in billions). The most
the Health care sector, the potential of ML/DL to allow for rapid
simple architecture of the MixNet family is MixNet-Small, which
diagnosis of diseases has also been proven by various research
consists of a large number of layers and channels. Furthermore,
work [6].
the size of the filters varies depending on the layers. Similar
Aiming to assist the clinical care, this paper presents a prac-
to the EfficientNet family, other configurations of the MixNet
tical solution for the detection of Covid-19 from CXR and LCT
images exploiting two cutting-edge deep neural network families, family, such as MixNet-Medium or MixNet-Large, are derived
i.e., EfficientNet [7] and MixNet [8]. Moreover, we empower the from MixNet-S with different scaling values.
learning process by means of three different transfer learning
strategies, namely ImageNet [9], AdvProp [10] and Noisy Stu- 2.2. Transfer learning
dent [11]. The evaluation on four large CXR and LCT images
datasets demonstrate that our proposed models obtain a superior To fine tune the hyper-parameters, i.e., the internal weights
performance compared to the existing studies that we are aware and biases, deep neural networks need a huge number of la-
of. In this respect, our work makes the following contributions: beled data. Moreover, the deeper/wider a network is, the more
• A system for detection of Covid-19 from CXR and LCT images parameters it possesses. As a result, deeper/wider networks re-
exploiting cutting-edge deep learning algorithms; quire more data to guide the learning, with the aim of avoiding
• A successful empirical evaluation on four large datasets overfitting and being effective. Thus, it is crucial to train them
consisting of CXR and LCT images; with enough data, in order to facilitate the learning. Nevertheless,
• A software prototype in the form of a mobile app ready to such a requirement is hard to come by in real-world settings,
be downloaded. since the labeling process usually is done manually, thus being
time consuming and subject to error [17]. To this end, transfer
The paper is structured as follows. Section 2 briefly introduces learning has been conceptualized as an effective way to extract
EfficientNet and MixNet as well as the transfer learning methods. and transfer the knowledge from a well-defined source domain
In Section 3, we present the dataset and metrics used for our
to a novice target domain [18,19]. In other words, transfer learn-
evaluation, while in Section 4 we analyze the experimental re-
ing facilitates the export of existing convolution weights from a
sults. The related work is reviewed in Section 5. Finally, Section 6
model trained using large datasets to create new accurate models
discusses future work and concludes the paper.
exploiting a relatively lower number of labeled images. As it has
2. Background been shown in various studies [20,21], transfer learning remains
helpful even when the target domain is quite different from the
As a base for our presentation, Section 2.1 provides a back- one in which the original weights have been obtained. In this
ground on two families of deep neural networks, i.e., EfficientNet work, we consider the following learning methods:
and MixNet, which are used as the classification engine in our
• ImageNet [9]: The ImageNet dataset has been widely ex-
work. Afterwards, a brief introduction to transfer learning is given
ploited to apply transfer learning by several studies, since
in Section 2.2.
it contains more than 14 million images, covering miscella-
2.1. EfficientNet and MixNet neous categories;
• AdvProp [10]: adversarial propagation has been proposed
Based on the observation that a better accuracy and efficiency as an improved training scheme, with the ultimate aim of
can be obtained by imposing a balance between all network avoiding overfitting. The method treats adversarial exam-
dimensions, EfficientNet [7] has been proposed by scaling in ples as additional examples, and uses a separate auxiliary
three dimensions, i.e., width, depth, and resolution, using a set batch norm for adversarial examples;
of fixed scaling coefficients that meet some specific constraints. • NS [11]: the Noisy Student learning method attempts to
By the most compact configuration, i.e., EfficientNet-B0, there are improve ImageNet classification Noisy Student Training by:
18 convolution layers in total, i.e., D = 18, and each layer is (i) enlarging the trainee/student equal to or larger than the
equipped with a kernel k(3,3) or k(5,5). The input image con- trainer/teacher, aiming to make the trainee learn better on
tains three color channels R, G, B, each of size 224 × 224. The a large dataset, and (ii) adding noise to the student, thus
next layers are scaled down in resolution to reduce the feature forcing him to learn more.
map size, but scaled up in width to increase accuracy. For in-
stance, the second convolution layer consists of W = 16 filters, To assist doctors in early detecting Covid-19 from CXR and
and the number of filters in the next convolution layer is W = LCT images, we develop an expert system that makes use of
24. The maximum number of filters is D = 1280 in correspon- EfficientNet and MixNet as the classification engine. Moreover, in
dence of the last layer, which is fed to the final fully connected order to accelerate the learning process and to achieve a higher
layer. The other configurations of the EfficientNet family are accuracy, we propose the three different learning strategies for
generated from EfficientNet-B0 by means of different scaling val- obtaining network weights mentioned above, i.e., ImageNet, Ad-
ues [7]. EfficientNet-B7 outperforms a CNN by achieving a better vProp, and NS. The succeeding section introduces the evaluation
accuracy, while considerably reducing the number of parameters. settings used to study the performance of our approach.
2
L.T. Duong, P.T. Nguyen, L. Iovino et al. Applied Soft Computing 132 (2023) 109851
Table 1 Table 2
Chest X-ray datasets. Lung CT datasets.
Dataset Type Categories Total Dataset Type Categories Total
Covid-19 Normal Pneumonia Covid-19 Normal Pneumonia
Train 98 7,966 5,447 13,511 Train 1,001 – 984 1,985
D1 Test 10 885 594 1,489 D3 [24] Test 125 – 123 248
Total 108 8,851 6,041 15,000 Validation 126 – 123 249
Train 261 8,154 5,909 14,324 Total 1,252 – 1,230 2,482
3. Evaluation
Table 3
Hardware and software configurations.
We explain in detail the datasets and methods used to study
Name Description
the performance of our proposed solution. Four real-world
RAM 24 GB
datasets have been used in the evaluation. Moreover, we make ®
CPU Intel Core™ i5-2400 CPU @ 3.10 GHz ×4
use of recent implementations2 of EfficientNet and MixNet, which
GPU GeForce GTX 1080 Ti
have been built atop the PyTorch framework.3 We imported pre- OS Ubuntu 18.04
trained weights from various sources to speed up the training. Python 3.7.5
We publish the tool developed through this paper in GitHub to Pytorch 1.5
facilitate future research.4 Torchvision 0.5.0
Numpy 1.15.4
Git 2.0
3.1. Research objectives Timm 0.1.26
Table 4
Experimental configurations.
Conf. Network Batch size # of Params Learning method Size (MB)
C1 EfficientNet-B0 110 7,919,391 ImageNet 53.1
C2 EfficientNet-B0 110 7,919,391 AdvProp 53.1
C3 EfficientNet-B0 110 7,919,391 NS 53.1
C4 EfficientNet-B3 64 14,352,075 ImageNet 106.9
C5 EfficientNet-B3 64 14,352,075 AdvProp 106.9
C6 EfficientNet-B3 64 14,352,075 NS 106.9
C7 MixNet-Small 110 6,253,449 ImageNet 41.8
C8 MixNet-Medium 90 7,133,225 ImageNet 48.9
C9 MixNet-Large 60 9,448,095 ImageNet 67.5
C10 MixNet-XL 60 14,015,611 ImageNet 104.2
three learning strategies mentioned in Section 2.2, we got 10 following reason: precision, recall and F1 are useful, given that the
experimental configurations C1 , . . . , C10 , as shown in Table 4. Batch number of positive images accounts for a very small percentage of
size corresponds to the number of items used for each training all the items in the dataset. In fact, in this case a classifier always
step; # of Params specifies the number of parameters used by providing a negative prediction would have a very high accuracy.
each network; and finally Size is the file size needed to store True positive TPi = |Gi ∩ Si |, i = 1, 2, 3 is defined as the
the parameters. It is evident that EfficientNet-B3 is the largest number of items that appear both in the results and ground-truth
network with respect to the number of parameters as well as the data of class i. The metrics are defined as follows.
file size to store them. To be concrete, i.e., C4 , C5 , and C6 have Accuracy: It is the fraction of correctly classified items with
more than 14 millions of parameters, accounting for more than respect to the total number of images in the test set.
100MB of storage space each. In the evaluation, we employed the ∑3
TPi
five-fold cross-validation methodology on the datasets. Namely, accuracy = ∑3i × 100% (1)
each dataset is split into five equal parts, and in each validation i |Gi |
round one part is used for testing and the other four ones for the
Precision and Recall: Precision is the fraction of classified
training.
images for Si being found in the ground-truth data Gi , while Recall
the fraction of true positives being found in the ground-truth
3.4. Evaluation metrics data.
TPi
Each image in all the datasets has been manually labeled, i.e., precisioni = (2)
either Normal or Pneumonia or Covid-19, resulting in three inde- |Si |
pendent groups, i.e., G = (G1 , G2 , G3 ), called ground-truth data. TPi
Using either EfficientNet or MixNet as the classifier on a test set, recalli = (3)
|Gi |
three predicted sets, i.e., S = (S1 , S2 , S3 ) of images are obtained.
We measured the classification performance by evaluating the F1 score (F-Measure): It is calculated as the average of preci-
similarity of the classified categories with the ground-truth ones. sion and recall using the following formula:
To this end, three metrics, namely accuracy, precision and recall, 2 · precisioni · recalli
F1 = (4)
and F1 score are used [21]. We selected these metrics for the precisioni + recalli
4
L.T. Duong, P.T. Nguyen, L. Iovino et al. Applied Soft Computing 132 (2023) 109851
Table 5
Experimental results on dataset D1 .
Configuration C1 C2 C3 C4 C5 C6 C7 C8 C9 C10
Accuracy (%) 95.64 95.77 95.30 96.17 96.64 95.90 95.30 95.98 96.11 96.37
Covid-19 1.000 1.000 1.000 1.000 0.875 0.857 1.000 1.000 1.000 1.000
Precision Normal 0.952 0.960 0.950 0.957 0.968 0.957 0.953 0.966 0.961 0.964
Pneu. 0.961 0.954 0.956 0.968 0.964 0.963 0.951 0.950 0.960 0.962
Covid-19 0.300 0.300 0.300 0.300 0.700 0.600 0.300 0.400 0.400 0.600
Recall Normal 0.981 0.975 0.977 0.985 0.978 0.978 0.974 0.971 0.978 0.978
Pneu. 0.929 0.942 0.927 0.937 0.952 0.936 0.932 0.951 0.944 0.947
Covid-19 0.461 0.461 0.461 0.461 0.778 0.705 0.461 0.571 0.571 0.750
F1 -score Normal 0.967 0.967 0.963 0.971 0.973 0.967 0.963 0.969 0.969 0.971
Pneu. 0.945 0.948 0.941 0.953 0.958 0.949 0.942 0.954 0.952 0.955
While accuracy is important, we are also interested in effi- Similarly, let us analyze the results on dataset D2 presented
ciency, considering the fact that the model needs to have a high in Table 6 to ascertain the best network configuration. It is ev-
recognition speed in practice. ident that C1 achieves most of the best scores with respect to
Speed: The system presented in Table 3 is used to benchmark various metrics as well as categories. For instance, C1 get 95.82%
the processing time, i.e., the average number of predicted items as accuracy, together with C8 , which is the best over all the
in a second. configurations.
In the next section, we present in detail the experimental Altogether, through Tables 5 and 6 we can see that C1 and
results by referring to the research questions introduced in Sec- C5 are the configurations among the others that bring the best
tion 3.1. prediction performance.
Compared to existing work that performs evaluation on the
4. Results same dataset [22,29], our approach achieves a better performance
with respect to accuracy, precision, recall, and F1 -score. For in-
stance, the work by Wang et al. [22], the maximum accuracy
This section reports and analyzes the results obtained from
is 93.0% with similar experimental settings. In this respect, we
the conducted experiments by referring to the research objectives
conclude that application of the two network families EfficientNet
presented in Section 3.1.
and MixNet as well as the different transfer learning techniques
brings a good prediction performance on the considered dataset.
4.1. Performance on the CXR datasets We conclude that EfficientNet and MixNet can successfully
predict Covid-19 from CXR images, obtaining a high accuracy and
The results obtained by performing the experiments on D1 are precision. Among others, EfficientNet yields the best prediction
shown in Table 5. The accuracy for all configurations is always performance.
larger than 95%, and the maximum accuracy is 96.64% obtained by
C5 , i.e., EfficientNet-B3 using pre-trained weights with AdvProp. 4.2. Performance on the LCT datasets
With respect to Precision, eight among ten configurations get
1.000 as precision for the Covid-19 category. This means that all The experimental results for the two LCT datasets, i.e., D3 and
images classified as Covid-19 by the classifiers are actually Covid- D4 are shown in Tables 7 and 8, respectively. Table 7 demon-
19. For the other two categories, i.e., Normal and Pneumonia, the strates that EfficientNet-B0 trained with ImageNet is the most ef-
maximum precision is 0.968, achieved also by C5 for Category fective configuration, i.e., it obtains 97.99% as accuracy. Moreover,
Normal, and by C4 for Category Pneumonia. Altogether, we see this configuration also gets the best F1 score for both categories.
that all the classifiers are able to recognize the testing images, Concerning D4 , the dataset consists of a large number of im-
obtaining high precision. ages (cf. Table 2). This resembles a real-world scenario where
Concerning recall, all the configurations get a considerably low images are collected and fed to the system on a daily basis,
score for the Covid-19 category: the highest recall is 0.700, ob- resulting in a big database. We see that both network families
tained by C5 . This means that the models are not able to retrieve get a high performance with respect to different metrics. All the
all the items in the ground-truth data, though they can yield good configurations obtain an accuracy larger than 98% with 99.66% as
predictions for the category. We assume that this happens due maximum accuracy.
to the limited data available in the training set. Referring back In comparison to existing studies that performed evaluation
to Table 1, for the Covid-19 category there are only 98 images on same datasets, we can see that our proposed framework
and 10 images for training and testing, respectively. Meanwhile, achieves a better performance. For instance, He et al. [30] obtain
for other two image categories, the recall scores are substantially 87.93% as maximum accuracy, while we get 97.99% and 99.66%
improved. The best performance is seen by category Normal, i.e., for D3 and D4 , respectively.
0.985; while by Pneumonia, recall is 0.952. As depicted in Table 1, On the LCT datasets, EfficientNet is also the network family
these categories consist of a larger number of training and testing that helps to obtain the best prediction performance.
images compared to the Covid-19.
For what concerns the F1 scores, for the Covid-19 category the 4.3. The benefit of transfer learning
maximum F1 is 0.778, obtained by C5 . The classifiers obtain a
low F1 in the other configurations, and this happens due to the We conducted experiments following the five-fold cross-
low recall scores as shown above. For the other two categories validation technique. Moreover, to further investigate the appli-
Normal and Pneumonia, the F1 scores are improved considerably cability of the proposed approach, we made use of D2 and D4 ,
compared to Covid-19. C5 is the most suitable configuration as it which contain more images than D1 and D3 (cf. Table 1). Figs. 3(a),
obtains the best F1 scores for all the categories, i.e., also 0.973 for 3(b), and 3(c) depict the confusion matrices for EfficientNet-B0
Normal and 0.958 for Pneumonia. using the three different transfer learning techniques mentioned
5
L.T. Duong, P.T. Nguyen, L. Iovino et al. Applied Soft Computing 132 (2023) 109851
Table 6
Experimental results on dataset D2 .
Configuration C1 C2 C3 C4 C5 C6 C7 C8 C9 C10
Accuracy (%) 95.82 94.39 93.30 95.05 95.59 95.00 95.79 95.82 95.68 95.53
Covid-19 0.968 0.948 0.889 0.950 0.968 0.978 0.982 0.966 0.983 0.967
Precision Normal 0.958 0.942 0.932 0.942 0.955 0.953 0.957 0.957 0.951 0.951
Pneu. 0.957 0.946 0.935 0.964 0.955 0.944 0.958 0.959 0.963 0.960
Covid-19 0.924 0.560 0.363 0.863 0.924 0.667 0.833 0.863 0.909 0.909
Recall Normal 0.958 0.942 0.932 0.942 0.955 0.944 0.974 0.973 0.975 0.960
Pneu. 0.942 0.922 0.906 0.912 0.939 0.937 0.941 0.941 0.933 0.932
Covid-19 0.945 0.704 0.616 0.905 0.945 0.792 0.901 0.922 0.945 0.937
F1 -score Normal 0.644 0.956 0.951 0.959 0.962 0.960 0.965 0.965 0.963 0.962
Pneu. 0.950 0.934 0.920 0.940 0.947 0.940 0.950 0.950 0.947 0.946
Table 7
Experimental results on dataset D3 .
Configuration C1 C2 C3 C4 C5 C6 C7 C8 C9 C10
Accuracy (%) 97.99 96.24 94.09 97.31 97.58 97.31 96.77 97.31 96.51 97.81
Covid-19 0.991 0.972 0.982 0.978 0.994 0.973 0.973 0.978 0.978 0.957
Prec.
Pneu. 0.968 0.952 0.905 0.967 0.991 0.973 0.962 0.967 0.958 0.957
Covid-19 0.968 0.951 0.898 0.967 0.958 0.972 0.962 0.967 0.951 0.978
Rec.
Pneu. 0.991 0.973 0.983 0.978 0.994 0.973 0.973 0.978 0.978 0.967
Covid-19 0.979 0.962 0.938 0.973 0.975 0.973 0.967 0.973 0.964 0.967
F1
Pneu. 0.979 0.962 0.943 0.973 0.975 0.973 0.967 0.973 0.965 0.967
Table 8
Experimental results on dataset D4 .
Configuration C1 C2 C3 C4 C5 C6 C7 C8 C9 C10
Accuracy (%) 99.66 99.52 98.85 99.59 99.62 99.47 99.52 98.41 99.59 99.53
Covid-19 0.997 0.995 0.991 0.997 0.996 0.995 0.994 0.990 0.996 0.996
Precision Normal 0.990 0.989 0.977 0.988 0.989 0.987 0.990 0.962 0.989 0.986
Pneu. 0.999 0.998 0.992 0.999 0.999 0.998 0.999 0.991 0.999 0.998
Covid-19 0.994 0.994 0.988 0.993 0.994 0.992 0.994 0.983 0.994 0.992
Recall Normal 0.996 0.991 0.983 0.995 0.995 0.991 0.990 0.978 0.994 0.995
Pneu. 0.992 0.998 0.991 0.999 0.998 0.998 0.998 0.988 0.998 0.999
Covid-19 0.995 0.994 0.989 0.995 0.995 0.994 0.994 0.986 0.995 0.994
F1 -score Normal 0.993 0.990 0.980 0.991 0.992 0.989 0.990 0.970 0.991 0.990
Pneu. 0.999 0.998 0.992 0.999 0.999 0.998 0.998 0.989 0.999 0.998
in Section 2.2. The metrics for all the confusion matrices are classifying 1,984 images among 2,038 images, while it suffers of
shown in Table 6. a low precision and recall for the other categories. For instance,
As we can see, each transfer learning method may have dif- with Pneumonia, only 1,377 out of 1,477 images are properly rec-
ferent effects on the different categories. For example, using ognized by MixNet-XL with weights pre-trained with ImageNet.
EfficientNet-B0 with weights pre-trained by ImageNet is benefi- The experimental results demonstrate that, depending on the
cial to Covid-19 and Pneumonia, but not to Normal. As shown in
network family, each transfer learning technique has a diverse
Fig. 3(a), 61 out of 66 images in Covid-19 are correctly classified,
influence on the final outcomes. By considering the results in
while for Pneumonia 1,392 out of 1,477. However, for the Normal
Table 6, it is evident C1 , corresponding to training EfficientNet-
category, only 1978 images are correctly classified over a total of
2,038 images, accounting for 97.05%. On the other hand, transfer B0 with weights by ImageNet, is the most effective configuration
learning with AdvProp (cf. Fig. 3(b)) induces a better performance with respect to accuracy, precision, recall, and F1 for almost all
for Normal, i.e., 1,981 among 2,038 images are classified to the categories. Moreover, together with the results obtained from
correct categories. In Fig. 3(c), we see that compared to the other Section 4.1, we conclude that ImageNet is the best transfer learning
learning methods, NS has an adverse effect on the recognition strategy for both network families on the two datasets D1 and D2 .
of all the categories. Altogether, we come to the conclusion that Next, we consider the results obtained for the LCT dataset in
training EfficientNet-B0 with weights from ImageNet yields the Figs. 4(a)–4(j). In general, training the deep neural networks with
best prediction performance. ImageNet helps to obtain a better performance compared to the
For EfficientNet-B3, we see that weights pre-trained with Im- other transfer learning techniques in all the three categories. As
ageNet are beneficial to the Normal category (cf. Fig. 3(d)). At
an example, with C1 we get 23,281 correctly classified Covid-
the same time, AdvProp is the transfer learning method that is
19 images, while the corresponding numbers for AdvProp and
suitable for recognition of Pneumonia, i.e., it helps to detect 1,388
NS are 23,270 and 23,137, respectively. For other categories,
out of 1,477 pneumonia images, which is the best among the
others. training EfficientNet-B0 with ImageNet weights also gets a good
Among the configurations, C9 is the best one for the Pneumonia performance.
category. Other MixNet configurations do not outperform the The experimental results show that weights pre-trained from
ones of EfficientNet-B0 and EfficientNet-B3. MixNet-XL obtains a the ImageNet dataset contribute to the best prediction perfor-
considerably good performance with Category Normal, correctly mance on all the CXR and LCT datasets.
6
L.T. Duong, P.T. Nguyen, L. Iovino et al. Applied Soft Computing 132 (2023) 109851
Fig. 3. Confusion matrices of EfficientNet and MixNet using different transfer learning techniques on D2 .
Fig. 4. Confusion matrices of EfficientNet-B0 and EfficientNet-B3 using different transfer learning techniques on D4 .
7
L.T. Duong, P.T. Nguyen, L. Iovino et al. Applied Soft Computing 132 (2023) 109851
for the medical image classification for Sars-CoV-2 based on Chest based on his personal data as well as other risk factors [50]. Jiang
X-ray (CXR) and Computed Tomography (CT). The new proposed et al. [51] propose an approach to the identification of clinical
model EfficientNet-B0 contains 7.92 M params, but still keeps characteristics of Covid-19, and develop an AI tool to recognize
the same number of MMac, i.e., 385.89 MMac. In summary, the patients at risk of a more severe impact of the disease.
modification of the original backbones is our key contribution for Ozturk et al. [52] proposed an approach to deliver accurate
the classification task. diagnostics for binary classification, i.e., Covid-19 vs. No-Findings
and multi-nominal classification, i.e., Covid-19 and No-Findings
4.6. Threats to validity and Pneumonia. The DarkNet model has been exploited as the
classification engine, consisting of 17 convolutional layers and a
We discuss the threats that may affect the internal, external, different filtering in each layer. The proposed model has a predic-
construct, and conclusion validity of our findings. tion accuracy of 98.08% and 87.02% for binary and multi-nominal
Internal validity. This concerns the internal factors that might classification, respectively.
have an adverse influence on the findings. A possible threat here Various studies have demonstrated the usefulness of CXR ex-
could come from the results for the Covid-19 category, since they ams in detecting Covid-19. Hall et al. [53] analyzed 135 CXR im-
are obtained with a considerably low number of items for training ages confirmed as Covid-19 and 320 images of viral and bacterial
and testing, i.e., D1 with 98 and 10 images and D2 with 327 and pneumonia. A Resnet50 DNN was trained on 102 Covid-19 and
98 images for training and testing, respectively. This threat is 102 pneumonia cases by means of the ten-fold cross-validation
mitigated by the other two categories in the datasets, as they technique. The experimental results showed an overall accuracy
contain a considerably large number of items. To the best of our of 89.2% with a Covid-19 true positive rate of 0.804 and an area
knowledge, there exists no CXR dataset with more images for under the curve (AUC) of 0.95. As a matter of fact, the dataset used
the Covid-19 category. As a matter of fact, research in medical by Hall et al. [53] is quite small, and it is necessary to investigate
imaging on Covid-19 suffers a general lack of data. For this reason, the proposed model on a larger amount of data to see if it still
unfortunately we are not able to study the models on a larger achieves such a good performance.
scale. Narin et al. [37] developed a system for the detection of
External validity. The main threat to external validity is due to coronavirus patients from CXR images. Three different CNN-
the factors that might hamper the generalizability of our results based models have been exploited, i.e., ResNet50, InceptionV3
and Inception-ResNetV2. The results show that the ResNet50
to other scenarios outside the scope of this work, e.g., in prac-
model achieves the best prediction performance with 98.0% as
tice we may encounter a limited amount of training data. The
the final accuracy. Although the approach obtains a good classi-
threat is moderated by evaluating EfficientNet and MixNet using
fication performance, it has been studied on a considerably small
the experimental settings following the five-fold cross-validation
dataset. It is our belief that performance may substantially change
methodology. In particular, the dataset is split into five parts, and
on larger datasets like the ones used in our evaluation.
accordingly the evaluation is conducted in five rounds. By each
Apostolopoulos et al. [38] evaluated their solution to auto-
round, four parts are combined to form the training data, and the
matic detection of Covid-19, making use of a dataset of CXR
remaining one is used for testing.
images from patients with common bacterial pneumonia, con-
Construction validity. This is related to the experimental settings
firmed Covid-19, and normal incidents. The datasets consists of
presented in the paper, concerning the simulation performed to
1427 CXR images including 224 images with confirmed Covid-19
evaluate the system. To mitigate the threat, the evaluation has
cases, 700 images with common bacterial pneumonia, and 504
been conducted on a training set and a test set, attempting to
images of normal situations. The experimental outcomes demon-
simulate a real usage where training data is already available for
strate that deep neural networks can be exploited to extract
feeding the system, while testing data is the part that needs to
important biomarkers related Covid-19. Nevertheless, like some
be predicted. Depending on the given settings, we ran several other existing studies et al. [37], again the approach has been
rounds of training/testing, to make sure that the final results are studied by means of a small amount of data. It is our assumption
generalizable, i.e., they are not obtained by chance. In the paper, that such a good performance might considerably change with
we reported the most stable experimental results. larger datasets.
Conclusion validity. This is related to the factors that might COVID-Net [54] is a deep convolutional neural network design
have an impact on the obtained outcome. The evaluation metrics tailored for the detection of Covid-19 cases from CXR images.
accuracy, precision, recall, F1 and execution time might induce a COVID-Net achieves an accuracy of 93.3%, with 98.9% positive
conclusion threat. To face the issue, we adopted such measures predictive values that is related to the detection of false positives.
as recommended by the previous scientific literature related to A deep learning model has been proposed [55] to detect Covid-
our setting, and employed the same metrics for evaluating all the 19 and differentiate it from common acquired pneumonia and
classifiers. other lung diseases. The analyzed dataset consists of 4356 chest
CT exams collected from 3322 patients. The per-exam sensitivity
5. Related work and specificity for detecting COVID-19 in the independent test set
was 114 of 127 (90.0%) and 294 of 307 (96.0%), respectively, with
The recent months have witnessed a large number of studies an area under the receiver operating characteristic curve (AUC) of
related to the topic Covid-19 and Machine Learning, and multiple 0.96 (p-value < 0.001). The per-exam sensitivity and specificity
Covid-19/ML applications have been proposed. We summarize for detecting community acquired pneumonia in the independent
in Table 10 some of the most notable research, providing the test set was 87% (152 of 175) and 92% (239 of 259), respectively.
number of considered images for each category as well as the pre- Abbas et al. [36] introduced Decompose, Transfer, and Com-
diction accuracy. In this work, since we support the recognition pose (DeTraC), a deep neural network for automated recognition
of Covid-19 from CXR and LCT images, in the remaining of this of Covid-19 from CXR images. An accuracy of 95% was achieved
section we concentrate on analyzing these studies. in the detection of Covid-19 CXR images from normal, and severe
Deep learning techniques have been exploited to predict which acute respiratory syndrome cases. COVID-CAPS [56] is a capsule
current antivirals might be more effective in patients infected Network-based Framework for Identification of Covid-19 cases
with coronavirus [48,49]. Similarly, a specific model has been de- from CXR Images. The approach yielded a good accuracy when
veloped to forecast if a Covid-19 patient has the chance to survive working with small datasets.
9
L.T. Duong, P.T. Nguyen, L. Iovino et al. Applied Soft Computing 132 (2023) 109851
Table 10
A summary of the related studies on CXR datasets.
Study Number of images Network Acc. (%)
Covid-19 Normal Pneumonia
Ghoshal et al. [35] 68 1583 2786 ResNet 89.82
Abbas et al. [36] 105 80 11 DeTraC based on ResNet-18 95.12
Nari et al. [37] 50 50 – ResNet-50 98.00
Apostolopous et al. [38] 224 504 700 VGG19 93.48
Luz et al. [39] 183 – – EfficientNet-B3 93.90
Zhang et al. [40] 100 1431 1531 ResNet18 96.00
Hemdan et al. [41] 25 25 – VGG19, DenseNet121 90.00
Table 11
A summary of the related studies on LCT datasets.
Study Number of images Network Acc. (%)
Covid-19 Normal Pneumonia
Rahimzadeh et al. [42] 465 7878 – ResNet50V2 98.49
Anwar et al. [43] 98 – 105 EfficientNet 89.70
Gunraj et al. [44] 4346 9450 7395 COVID-Net-CT 99.10
Mobiny et al. [45] 47 58 – DECAPS 87.60
He et al. [30] 25,442 14,471 28,160 ResNet3D34 95.90
Ardakanu et al. [46] 510 510 – ResNet-101 99.63
Bai et al. [47] 5030 – 9152 EfficientNet-B4 96.00
Concerning LCT datasets, we review the most relevant studies Declaration of competing interest
to our work in Table 11. Overall, most of the studies use a small
dataset for their evaluation. Only Gunraj et al. [44] made use of a The authors declare that they have no known competing finan-
considerably large dataset, however the obtained performance is cial interests or personal relationships that could have appeared
much lower than our proposed approach, i.e., 95.90% compared to to influence the work reported in this paper.
99.66%. In this respect, we conclude that by combining different
deep neural networks with transfer learning strategies, we can Data availability
obtain a high prediction accuracy even on a very large dataset.
To the best of our knowledge, compared to different existing The authors do not have permission to share data.
studies [36,38], our work is the first one that deals with big
datasets. In particular, for the CXR datasets, there are 15,000 and
References
17,905 images in D1 and D2 , respectively [57]. While for the
LCT datasets, especially with D4 , there are more than 400K of [1] J.M. Connors, J.H. Levy, COVID-19 and its implications for thrombosis and
images. This suggests that with the proposed model we can work anticoagulation, Blood J. Am. Soc. Hematol. 135 (23) (2020) 2033–2040.
with a huge number of images and still obtain a good timing [2] Q. Sun, H. Qiu, M. Huang, Y. Yang, Lower mortality of COVID-19 by
efficiency. More importantly, the application of the two deep early recognition and intervention: experience from Jiangsu Province, Ann.
Intensive Care 10 (1) (2020) 1–4.
neural network families, i.e., EfficientNet and MixNet, allows us
[3] L. Iovino, P.T. Nguyen, A.D. Salle, F. Gallo, M. Flammini, Unavailable transit
to build an expert system being capable of working with images feed specification: Making it available with recurrent neural networks, IEEE
coming from different sources. Trans. Intell. Transp. Syst. 22 (4) (2021) 2111–2122, http://dx.doi.org/10.
1109/TITS.2021.3053373.
6. Conclusions [4] L.T. Duong, N.H. Le, T.B. Tran, V.M. Ngo, P.T. Nguyen, Detection of
tuberculosis from chest X-ray images: Boosting the performance with
vision transformer and transfer learning, Expert Syst. Appl. 184 (2021)
We presented a workable solution for the detection of Covid- 115519, http://dx.doi.org/10.1016/j.eswa.2021.115519, URL https://www.
19 from chest X-ray and lung computed tomography images. We sciencedirect.com/science/article/pii/S0957417421009295.
designed and implemented the models based on two building [5] I. Portugal, P. Alencar, D. Cowan, The use of machine learning algorithms
blocks: EfficientNet and MixNet as the prediction engine and in recommender systems: A systematic review, Expert Syst. Appl. 97
(2018) 205–227, http://dx.doi.org/10.1016/j.eswa.2017.12.020, URL http:
effective transfer learning algorithms. The proposed models have //www.sciencedirect.com/science/article/pii/S0957417417308333.
been studied by means of four datasets which have been widely [6] X. Mei, H.-C. Lee, K.-y. Diao, M. Huang, B. Lin, C. Liu, Z. Xie, Y. Ma, P.
used in various papers. The experimental results show that our Robson, M. Chung, A. Bernheim, V. Mani, C. Calcagno, K. Li, S. Li, H. Shan, J.
proposed approach obtains a better prediction performance com- Lv, T. Zhao, J. Xia, Y. Yang, Artificial intelligence–enabled rapid diagnosis of
patients with COVID-19, Nat. Med. (2020) 1–5, http://dx.doi.org/10.1038/
pared to some relevant previous studies. To the best of our
s41591-020-0931-3.
knowledge, our work is the first one that deals with images com- [7] M. Tan, Q. Le, EfficientNet: Rethinking model scaling for convolutional
ing from different sources. As future work, we are going to refine neural networks, in: K. Chaudhuri, R. Salakhutdinov (Eds.), Proceedings of
and evaluate the approach by taking into consideration more the 36th International Conference on Machine Learning, in: Proceedings
datasets and tuning other deep neural network configurations. of Machine Learning Research, vol. 97, PMLR, Long Beach, California, USA,
2019, pp. 6105–6114, URL http://proceedings.mlr.press/v97/tan19a.html.
[8] M. Tan, Q.V. Le, MixConv: Mixed depthwise convolutional kernels, 2019,
CRediT authorship contribution statement CoRR abs/1907.09595. arXiv:1907.09595. URL http://arxiv.org/abs/1907.
09595.
Linh T. Duong: Conceptualization, Software, Writing – original [9] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A.
draft. Phuong T. Nguyen: Conceptualization, Methodology, Writ- Karpathy, A. Khosla, M. Bernstein, A.C. Berg, L. Fei-Fei, ImageNet large scale
visual recognition challenge, Int. J. Comput. Vis. 115 (3) (2015) 211–252,
ing – original draft, Writing – review & editing. Ludovico Iovino: http://dx.doi.org/10.1007/s11263-015-0816-y.
Validation, Visualization, Writing – review & editing. Michele [10] C. Xie, M. Tan, B. Gong, J. Wang, A. Yuille, Q.V. Le, Adversarial examples
Flammini: Writing – review & editing, Supervision. improve image recognition, 2019, arXiv e-prints arXiv:1911.09665.
10
L.T. Duong, P.T. Nguyen, L. Iovino et al. Applied Soft Computing 132 (2023) 109851
[11] Q. Xie, E. Hovy, M.-T. Luong, Q.V. Le, Self-training with Noisy Student [30] X. He, S. Wang, S. Shi, X. Chu, J. Tang, X. Liu, C. Yan, J. Zhang, G. Ding,
improves ImageNet classification, 2019, cite arxiv:1911.04252. URL http: Benchmarking deep learning models and automated model design for
//arxiv.org/abs/1911.04252. COVID-19 detection with chest CT scans, MedRxiv (2020) http://dx.doi.
[12] F. Chollet, Xception: Deep learning with depthwise separable convolutions, org/10.1101/2020.06.08.20125963.
2016, cite arxiv:1610.02357. URL http://arxiv.org/abs/1610.02357. [31] K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition,
in: 2016 IEEE Conference on Computer Vision and Pattern Recognition,
[13] A.G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. CVPR, 2016, pp. 770–778, http://dx.doi.org/10.1109/CVPR.2016.90.
Andreetto, H. Adam, MobileNets: Efficient convolutional neural networks [32] M. Elpeltagy, H. Sallam, Automatic prediction of COVID-19 from chest
for mobile vision applications, 2017, cite arxiv:1704.04861. URL http: images using modified ResNet50, Multimedia Tools Appl. 80 (17) (2021)
//arxiv.org/abs/1704.04861. 26451–26463, http://dx.doi.org/10.1007/s11042-021-10783-6.
[14] M. Tan, B. Chen, R. Pang, V. Vasudevan, Q.V. Le, MnasNet: Platform- [33] D. Yang, C. Martinez, L. Visuña, H. Khandhar, C. Bhatt, J. Carretero,
aware neural architecture search for mobile, 2018, CoRR abs/1807.11626. Detection and analysis of COVID-19 in medical images using deep learning
arXiv:1807.11626. URL http://arxiv.org/abs/1807.11626. techniques, 11 (1) 19638. http://dx.doi.org/10.1038/s41598-021-99015-3.
[15] H. Cai, L. Zhu, S. Han, ProxylessNAS: Direct neural architecture [34] F. Wilcoxon, Individual comparisons by ranking methods, in: S. Kotz, N.L.
search on target task and hardware, in: International Conference on Johnson (Eds.), Breakthroughs in Statistics: Methodology and Distribution,
Learning Representations, 2019, URL https://openreview.net/forum?id= Springer New York, New York, NY, 1992, pp. 196–202, http://dx.doi.org/
HylVB3AqYm. 10.1007/978-1-4612-4380-9_16.
[35] B. Ghoshal, A. Tucker, Estimating uncertainty and interpretability in deep
[16] M. Sandler, A.G. Howard, M. Zhu, A. Zhmoginov, L. Chen, Inverted residuals learning for coronavirus (COVID-19) detection, 2020, arXiv abs/2003.10769.
and linear bottlenecks: Mobile networks for classification, detection and [36] A. Abbas, M.M. Abdelsamea, M.M. Gaber, Classification of COVID-19 in
segmentation, 2018, CoRR abs/1801.04381. arXiv:1801.04381. URL http: chest X-ray images using DeTraC deep convolutional neural network, 2020,
//arxiv.org/abs/1801.04381. arXiv preprint arXiv:2003.13815.
[17] A. Kamilaris, F.X. Prenafeta-Boldú, Deep learning in agriculture: A survey, [37] A. Narin, C. Kaya, Z. Pamuk, Automatic detection of coronavirus disease
Comput. Electron. Agric. 147 (2018) 70–90, http://dx.doi.org/10.1016/j. (COVID-19) using X-ray images and deep convolutional neural networks,
compag.2018.02.016. 2020, arXiv:2003.10849.
[18] K. Weiss, T. Khoshgoftaar, D. Wang, A survey of transfer learning, J. Big [38] I.D. Apostolopoulos, T.A. Mpesiana, Covid-19: automatic detection from x-
Data 3 (2016) http://dx.doi.org/10.1186/s40537-016-0043-6. ray images utilizing transfer learning with convolutional neural networks,
Phys. Eng. Sci. Med. (2020) 1.
[19] L. Torrey, T. Walker, J. Shavlik, R. Maclin, Using advice to transfer [39] E. Luz, P.L. Silva, R. Silva, G. Moreira, Towards an efficient deep learning
knowledge acquired in one reinforcement learning task to another, in: model for covid-19 patterns detection in x-ray images, 2020, arXiv preprint
Proceedings of the 16th European Conference on Machine Learning, ECML arXiv:2004.05717.
’05, Springer-Verlag, Berlin, Heidelberg, 2005, pp. 412–424, http://dx.doi. [40] J. Zhang, Y. Xie, Y. Li, C. Shen, Y. Xia, Covid-19 screening on chest x-ray
org/10.1007/11564096_40. images using deep learning based anomaly detection, 2020, arXiv preprint
[20] Z. Huang, Z. Pan, B. Lei, Transfer learning with deep convolutional neural arXiv:2003.12338.
network for SAR target classification with limited labeled data, Remote [41] E.E.-D. Hemdan, M.A. Shouman, M.E. Karar, Covidx-net: A framework of
Sens. 9 (9) (2017) http://dx.doi.org/10.3390/rs9090907. deep learning classifiers to diagnose covid-19 in x-ray images, 2020, arXiv
[21] L.T. Duong, P.T. Nguyen, C. Di Sipio, D. Di Ruscio, Automated fruit preprint arXiv:2003.11055.
recognition using EfficientNet and MixNet, Comput. Electron. Agric. 171 [42] M. Rahimzadeh, A. Attar, S.M. Sakhaei, A fully automated deep learning-
(2020) 105326, http://dx.doi.org/10.1016/j.compag.2020.105326, URL http: based network for detecting COVID-19 from a new and large lung
//www.sciencedirect.com/science/article/pii/S0168169919319787. CT scan dataset, MedRxiv (2020) http://dx.doi.org/10.1101/2020.06.08.
20121541, arXiv:https://www.medrxiv.org/content/early/2020/09/01/2020.
[22] Z.Q.L. Linda Wang, A. Wong, COVID-Net: A tailored deep convolutional 06.08.20121541.full.pdf.
neural network design for detection of COVID-19 cases from chest [43] T. Anwar, S. Zakir, Deep learning based diagnosis of COVID-19 using chest
radiography images, 2020, arXiv:2003.09871. CT-scan images, 2020, http://dx.doi.org/10.36227/techrxiv.12328061.
[23] J.P. Cohen, P. Morrison, L. Dao, COVID-19 image data collection, 2020, [44] H. Gunraj, L. Wang, A. Wong, COVIDNet-CT: A tailored deep convolutional
arXiv:2003.11597. URL https://github.com/ieee8023/covid-chestxray- neural network design for detection of COVID-19 cases from chest CT
dataset. images, 2020, arXiv:2009.05383.
[24] E. Soares, P. Angelov, S. Biaso, M. Higa Froes, D. Kanda Abe, SARS-CoV-2 [45] A. Mobiny, P.A. Cicalese, S. Zare, P. Yuan, M. Abavisani, C.C. Wu, J. Ahuja,
CT-scan dataset: A large dataset of real patients CT scans for SARS- P.M. de Groot, H.V. Nguyen, Radiologist-level COVID-19 detection using CT
CoV-2 identification, MedRxiv (2020) http://dx.doi.org/10.1101/2020.04.24. scans with detail-oriented capsule networks, 2020, arXiv:2004.07407.
20078584, arXiv:https://www.medrxiv.org/content/early/2020/05/14/2020. [46] A.A. Ardakani, A.R. Kanafi, U.R. Acharya, N. Khadem, A. Mohammadi,
04.24.20078584.full.pdf. URL https://www.medrxiv.org/content/early/2020/ Application of deep learning technique to manage COVID-19 in routine
05/14/2020.04.24.20078584. clinical practice using CT images: Results of 10 convolutional neu-
ral networks, Comput. Biol. Med. 121 (2020) 103795, http://dx.doi.org/
[25] K. Zhang, X. Liu, J. Shen, Z. Li, Y. Sang, X. Wu, Y. Zha, W. Liang, C. Wang, K. 10.1016/j.compbiomed.2020.103795, URL http://www.sciencedirect.com/
Wang, L. Ye, M. Gao, Z. Zhou, L. Li, J. Wang, Z. Yang, H. Cai, J. Xu, L. Yang, science/article/pii/S0010482520301645.
W. Cai, W. Xu, S. Wu, W. Zhang, S. Jiang, L. Zheng, X. Zhang, L. Wang, L. [47] H.X. Bai, R. Wang, Z. Xiong, B. Hsieh, K. Chang, K. Halsey, T.M.L. Tran, J.W.
Lu, J. Li, H. Yin, W. Wang, O. Li, C. Zhang, L. Liang, T. Wu, R. Deng, K. Wei, Choi, D.-C. Wang, L.-B. Shi, J. Mei, X.-L. Jiang, I. Pan, Q.-H. Zeng, P.-F. Hu,
Y. Zhou, T. Chen, J. Lau, M. Fok, J. He, T. Lin, W. Li, G. Wang, Clinically Y.-H. Li, F.-X. Fu, R.Y. Huang, R. Sebro, Q.-Z. Yu, M.K. Atalay, W.-H. Liao,
applicable AI system for accurate diagnosis, quantitative measurements, Artificial intelligence augmentation of radiologist performance in distin-
and prognosis of COVID-19 pneumonia using computed tomography, Cell guishing COVID-19 from pneumonia of other origin at chest CT, Radiology
181 (6) (2020) http://dx.doi.org/10.1016/j.cell.2020.04.045. 296 (3) (2020) E156–E165, http://dx.doi.org/10.1148/radiol.2020201491,
[26] G. Marques, D. Agarwal, I. de la Torre Díez, Automated medical diagnosis PMID: 32339081. arXiv:https://doi.org/10.1148/radiol.2020201491.
of COVID-19 through EfficientNet convolutional neural network, Appl. Soft [48] H. Zhang, K.M. Saravanan, Y. Yang, M.T. Hossain, J. Li, X. Ren, Y. Pan, Y.
Comput. 96 (2020) 106691, http://dx.doi.org/10.1016/j.asoc.2020.106691, Wei, Deep learning based drug screening for novel coronavirus 2019-nCov,
URL http://www.sciencedirect.com/science/article/pii/S1568494620306293. Interdiscip. Sci. Comput. Life Sci. (2020) 1.
[27] A. Iqbal, M. Usman, Z. Ahmed, An efficient deep learning-based framework [49] B.R. Beck, B. Shin, Y. Choi, S. Park, K. Kang, Predicting commercially
for tuberculosis detection using chest X-ray images, Tuberculosis 136 available antiviral drugs that may act on the novel coronavirus (SARS-
(2022) 102234, http://dx.doi.org/10.1016/j.tube.2022.102234, URL https: CoV-2) through a drug-target interaction deep learning model, Comput.
//www.sciencedirect.com/science/article/pii/S1472979222000713. Struct. Biotechnol. J. (2020).
[50] L. Yan, H.-T. Zhang, Y. Xiao, M. Wang, C. Sun, J. Liang, S. Li, M. Zhang, Y.
[28] A. Saygılı, A new approach for computer-aided detection of coronavirus Guo, Y. Xiao, et al., Prediction of criticality in patients with severe Covid-19
(COVID-19) from CT and X-ray images using machine learning meth- infection using three clinical features: a machine learning-based prognostic
ods, Appl. Soft Comput. 105 (2021) 107323, http://dx.doi.org/10.1016/ model with clinical data in Wuhan, MedRxiv (2020).
j.asoc.2021.107323, URL https://www.sciencedirect.com/science/article/pii/ [51] X. Jiang, M. Coffee, A. Bari, J. Wang, X. Jiang, J. Huang, J. Shi, J. Dai, J. Cai, T.
S1568494621002465. Zhang, et al., Towards an artificial intelligence framework for data-driven
[29] T. Ozturk, M. Talo, E.A. Yildirim, U.B. Baloglu, O. Yildirim, U.R. Acharya, prediction of coronavirus clinical severity, CMC: Comput. Mater. Contin.
Automated detection of COVID-19 cases using deep neural networks with 63 (2020) 537–551.
X-ray images, Comput. Biol. Med. 121 (2020) 103792, http://dx.doi.org/ [52] T. Ozturk, M. Talo, E.A. Yildirim, U.B. Baloglu, O. Yildirim, U.R. Acharya,
10.1016/j.compbiomed.2020.103792, URL http://www.sciencedirect.com/ Automated detection of COVID-19 cases using deep neural networks with
science/article/pii/S0010482520301621. X-ray images, Comput. Biol. Med. (2020) 103792.
11
L.T. Duong, P.T. Nguyen, L. Iovino et al. Applied Soft Computing 132 (2023) 109851
[53] L.O. Hall, R. Paul, D.B. Goldgof, G.M. Goldgof, Finding covid-19 from chest [56] P. Afshar, S. Heidarian, F. Naderkhani, A. Oikonomou, K.N. Plataniotis,
x-rays using deep learning on a small dataset, 2020, arXiv preprint arXiv: A. Mohammadi, Covid-caps: A capsule network-based framework for
2004.02060. identification of covid-19 cases from x-ray images, 2020, arXiv preprint
[54] L. Wang, A. Wong, COVID-net: A tailored deep convolutional neural arXiv:2004.02696.
network design for detection of COVID-19 cases from chest X-Ray images, [57] L.T. Duong, P.T. Nguyen, L. Iovino, M. Flammini, Deep learning for
2020, arXiv preprint arXiv:2003.09871. automated recognition of Covid-19 from chest X-ray images, MedRxiv
[55] L. Li, L. Qin, Z. Xu, Y. Yin, X. Wang, B. Kong, J. Bai, Y. Lu, Z. Fang, Q. Song, et (2020) http://dx.doi.org/10.1101/2020.08.13.20173997, arXiv:https://www.
al., Artificial intelligence distinguishes COVID-19 from community acquired medrxiv.org/content/early/2020/08/14/2020.08.13.20173997.full.pdf. URL
pneumonia on chest CT, Radiology (2020) 200905. https://www.medrxiv.org/content/early/2020/08/14/2020.08.13.20173997.
12