fcvm-10-1229743
fcvm-10-1229743
RECEIVED 26 May 2023 Introduction: Cardiovascular disease remains a significant problem in modern
ACCEPTED 05 July 2023 society. Among non-invasive techniques, the electrocardiogram (ECG) is one of
PUBLISHED 31 July 2023 the most reliable methods for detecting cardiac abnormalities. However, ECG
CITATION interpretation requires expert knowledge and it is time-consuming. Developing a
Pham H, Egorov K, Kazakov A and Budennyy S novel method to detect the disease early improves the quality and efficiency of
(2023) Machine learning-based detection of
medical care.
cardiovascular disease using ECG signals:
performance vs. complexity.
Methods: The paper presents various modern approaches for classifying cardiac
Front. Cardiovasc. Med. 10:1229743. diseases from ECG recordings. The first approach suggests the Poincaré
doi: 10.3389/fcvm.2023.1229743 representation of ECG signal and deep-learning-based image classifiers.
COPYRIGHT Additionally, the raw signals were processed with the one-dimensional
© 2023 Pham, Egorov, Kazakov and Budennyy. convolutional model while the XGBoost model was facilitated to predict based
This is an open-access article distributed under
on the time-series features.
the terms of the Creative Commons Attribution
License (CC BY). The use, distribution or Results: The Poincaré-based methods showed decent performance in predicting
reproduction in other forums is permitted, AF (atrial fibrillation) but not other types of arrhythmia. XGBoost model gave an
provided the original author(s) and the
acceptable performance in long-term data but had a long inference time due to
copyright owner(s) are credited and that the
original publication in this journal is cited, in highly-consuming calculations within the pre-processing phase. Finally, the 1D
accordance with accepted academic practice. convolutional model, specifically the 1D ResNet, showed the best results in both
No use, distribution or reproduction is
studied CinC 2017 and CinC 2020 datasets, reaching the F1 score of 85% and
permitted which does not comply with these
terms. 71%, respectively, and they were superior to the first-ranking solution of each
challenge. The 1D models also presented high specificity. Additionally, our paper
investigated efficiency metrics including power consumption and equivalent
CO2 emissions, with one-dimensional models like 1D CNN and 1D ResNet being
the most energy efficient. Model interpretation analysis showed that the
DenseNet detected AF using heart rate variability while the 1D ResNet assessed
the AF patterns in raw ECG signals.
Discussion: Despite the under-performed results, the Poincaré diagrams are still
worth studying further because of the accessibility and inexpensive procedure.
In the 1D convolutional models, the residual connections are useful to keep the
model simple but not decrease the performance. Our approach in power
measurement and model interpretation helped understand the numerical
complexity and mechanism behind the model decision.
KEYWORDS
TABLE 1 Descriptive statistics and demographics information of CinC 2017 and CinC 2020 datasets.
2017 dataset, the data from CinC 2020 is also WFDB-compliant. below 10%. The detail description of each class is mentioned in
The header files embedded the demographics information and Table S1 in the Supplementary Material.
diagnosis labels.
The embedded headers in CinC 2020 reveal the distribution of
age and gender. Most of the subjects in CinC 2020 are aged 65 and 2.2. Model architecture, training pipeline,
older, followed by the age group of 50–64 years. The remaining age and experiment conditions
groups have relatively smaller proportions. Additionally, every
subset has minimal missing data for age, with less than 1% of This section describes the details of the configuration of each
entries having no specified age information. Regarding gender, model as well as the flow of data when training the model. The
the sex ratio is relatively balanced. Notably, there are nearly zero overview of the training pipeline is given in Figure 2 and
missing data points for gender in CinC 2020. interpreted in the following. There are three main flows of
In term of label distribution, while the CinC 2017 dataset was learning in our experiments. They shared the same signal loader
limited to four classes, including Normal, Atrial Fibrillation (AF), and evaluation method. Each flow had a separate way of pre-
Other, and Noisy, the CinC 2020 dataset has 28 categories of processing data and model optimization, which will be described
cardiovascular disease (Figure 1). The distribution of classes is in detail below.
very imbalanced. In CinC 2017, the AF samples are under 10% All experiments were conducted on the same computer with
of the dataset while the Normal samples are nearly 60%. 1 Intel Core i7-9700F CPU, 64 GB RAM, and 1 NVIDIA
Likewise, within the CinC 2020 dataset, the Sinus Rhythm and GeForce GTX 3060 GPU. The metrics was estimated on both
Other categories exceed 20%, whereas the remaining classes fall 5-fold cross-validation and a separated test set. In cross-
FIGURE 1
The classes distribution in datasets CinC 2017 (left side) and CinC 2020 (right side).
FIGURE 2
The high-level scheme of training pipeline, including data preparation, model optimization, and performance evaluation.
validation experiments, we combined the train and validation NNiþ1 intervals. Figure 3 shows the examples of Poincaré
subset into one dataset before performing the fivefold cross- diagram of a short and long recording.
validation. To estimate the performance on the unseen data, we To predict the cardiovascular disease over the Poincaré
trained the models on the train subset and tested on the diagrams, the default architecture of ResNet50 (16) and
separated test subset, while the validation subset was used to DenseNet121 (17) were used to train from the scratch
optimize the hyperparameters. (without pre-trained weights). The last layers of these
models were also tailored to match the number of classes
of each dataset.
FIGURE 3
The Poincaré diagrams of the short-term (A) and long-term (B) ECGs. The diagrams plot the normal R-peak intervals (or NN intervals).
0.99 for every block. The Pooling of base block is Max Pooling of biased if the dataset is heavily imbalanced among classes (19),
which the kernel size and stride size are 2. The dropout probability precision and recall are preferred in these cases. In this study, we
is set to 0.3 in every place. Before flattening the tensor and feeding used the harmonic mean of precision and recall (F1 score), which
to the last fully connected layer for the logit outputs, there is an is easy to compare methods with only one measure. Besides the
average pooling layer with a kernel size of 1 and stride size of 2. F1 score, sensitivity and specificity were also calculated to get
The structure of the base block in 1D ResNet includes a 1D more insight into each model. These scores were calculated as
Convolutional layer, 1D Batch Normalization, ReLU activation
function, a Drop-out layer, another 1D Convolutional layer, and 1D
TP TP
Batch Normalization. In a base block, the input would go through Precision ¼ , Recall=Sensitivity ¼
TP þ FP TP þ FN
these layers before adding the residual which is also the input
TN 2 Precision Recall
tensor. This summation is activated by the ReLU function after Specificity ¼ , F1 Score ¼
FP þ TN Precision þ Recall
leaving the block.
In our work, the 1D ResNet starts with a 1D Convolutional layer
with a kernel size of 15 and the number of output channels is 64 where TP, FP, TN, and FN are the number of True Positive, False
followed by a 1D Batch Normalization, ReLU Activation function, Positive, True Negative, and False Negative extracted from the
and Max Pooling layer. After that, there are four base blocks with confusion matrix, respectively. Because the dataset is multiple
kernel sizes increasing from 65 to 256. The output of the last base classes (in CinC 2017) or multiples labels (in CinC 2020), the
block goes through two pooling layers: an Average Pooling layer average of each metric was reported in our study.
and a Max Pooling layer. These outputs are concatenated before To assess the power efficiency of the proposed methods, we
feeding to the final fully connected layer to compute the output logits. measured the power consumption, the equivalent CO2 emission,
In both 1D CNN and 1D ResNet models, the signals are and the inference time of each model. The inference time was
converted to first-order difference and scaled to zero mean and calculated by the average time of predictions on the test set. The
unit variance before transferring to the models. power consumption and carbon emissions information were
monitored and calculated using the open-source library eco2AI,
which is available at https://github.com/sb-ai-lab/Eco2AI (20).
2.2.3. Learning over XGBoost feature space
To explore the mechanism behind the decision of (1D and 2D)
In the pipeline of the XGBoost model, the processed ECG signals
convolutional model, the Gradient-weighted Class Activation
need to feed to module tsfresh to extract the features before
Mapping or GradCAM (21) was constructed to visualize which
training model. The features extraction used the list of efficient
region in the input that the model focused on when predicting.
features set but filtered out the time-consuming features including:
The basic idea of GradCAM is to calculate the gradient of output
entropy-related features, matrix profile, the number of Continuous
with respect to the feature map of last convolutional layers, and
Wavelet Transform (CWT) peak, partial autocorrelation,
then these gradients are aggregated to obtain the weight for the
aggregated linear trend, and the statistics of Augmented Dickey–
feature map. An heatmap is built based on these weights to show
Fuller test. In the feature matrix, the pipeline filled the missing
the contribution of each region to the prediction score.
data with 999 and removed the low-variance features.
For the XGBoost model, the feature importance of each feature
The hyperparameters of XGBoost were optimized by searching
was calculate based on the frequency of that feature in splitting data
within the predefined space (Table 2). The optimum collection was
when training the tree. Because the number of variables is
found by Bayesian optimization implemented in the library
enormous, they were aggregated into common group.
scikit-optimize. (18) The number of search trials was
limited to 100 because of time constraints.
3. Results
2.3. Metrics and model interpretation
In this section, we summarize the classification performance
The popular metrics for the classification problem include and power consumption metrics of each classifier. After that, we
accuracy, precision, recall, and F1 score. While the accuracy is try to interpret the model decision using the mentioned methods.
3.1. Cardiovascular diseases classification model in sensitivity. Regarding the best model (1D ResNet), the
metrics for each class are presented in the Supplementary
The experiment results showed the superior performance of the Table S2. This model worked very well on classifying
1D ResNet model learned over raw data in both datasets. the class Atrial Fibrillation, Sinus tachycardia, and Left/Right
Especially, in both datasets, this model surpassed the first rank bundle branch block while there are a number of minor patterns
solution in many measurements. The comparison of F1 scores, that were unable to detect such as Q-wave abnormal or
sensitivity, specificity, and the efficiency metrics (power T-wave inversion.
consumption, equivalent CO2 ) are given in Table 3. The fivefold cross-validation results, especially the standard
For the CinC 2017 dataset, the 1D ResNet is the best-performing deviation of performance metrics, reveals the stability and
model with the highest F1 score in both fivefold cross-validation and robustness of the models. Because of lower standard deviation in
separated test dataset. The model also demonstrates high sensitivity the cross-validation metrics, the 1D convolutional models are
and specificity. The 1D CNN on the raw signal input also performs more stable than other models in both datasets, while the
well among the other models. For the CinC 2020 dataset, the 1D performance of gradient-boosting model and Poincaré-based
ResNet models also outperforms other models in F1 scores and models fluctuate among the splits.
specificity. However, the sensitivity of the model is lower than the To better illustrate the tandem “Performance vs. Complexity”
transformer-based model (12). for examined models, Figure 4 shows cross-plots on F1 score
In detail, the Poincaré-based methods have adequate and CO2 emissions for both datasets. In particular, one can
performance in the CinC 2017 challenge. However, they did not reveal that DenseNet121 and ResNet50 models learned over
perform well in the CinC 2020 challenge. Particularly, some Poincaré diagrams stand out from other models as inefficient
classes are not discriminated in the Poincaré diagrams. The while ResNet learned on raw ECG signals outperforms.
models ResNet50 and DenseNet121 only identified the types We also analyzed the performance of investigated models in
Atrial fibrillation (AF), Sinus bradycardia (SB), Sinus rhythm each source of the CinC 2020 dataset (Table 4). The ResNet50
(SNR), Sinus tachycardia (STach), and other, while the metrics was good at the short-term recordings while performing poorly
for the remaining types are close to zero. This result is in long-term data. The DenseNet121 was better than ResNet50
understandable as the information on heart rate variability is not in long-term signal classification but did not surpass the 1D
sufficient to identify several types of heart disease. Convolutional model. The XGBoost outperformed the others in
The XGBoost is under-expected because it ranked lowest in long-term ECG. However, the number of long-term signals is not
CinC 2017 and only third in CinC 2020 despite the gradient- very large, so their metrics might not be stable.
boosting family usually gaining the highest place at many
machine learning benchmarks.
The unidimensional convolutional models yielded excellent 3.2. Efficiency metrics
results in both classification challenges. The 1D CNN and 1D
ResNet shared the top two places in both datasets. In CinC 2017, The results of the efficiency measurement are summarized in
the 1D ResNet is the best model, followed by the 1D CNN. In Table 5. In terms of power consumption, 2D models are also the
CinC 2020, the 1D ResNet surpassed the top-1 solution by a most power-hungry models: ResNet50 and DenseNet121
large margin in F1 score and 2% higher specificity. However, consumed two to three times more energy than the others. On
these models were still not better than the transformer-based the other hand, the 1D models are more efficient than any other
TABLE 3 Performance benchmark results on the fivefold cross-validation and the separated test datasets.
FIGURE 4
Test F1 score vs. CO2 emissions: left side—models learned over CinC 2017 dataset; right side—models learned over CinC 2020 dataset. Dotted red ellipses
highlight relatively heavy models.
TABLE 5 Power consumption, equivalent CO2 emission, and inference time of each model.
models: the 1D ResNet needed more energy than only XGBoost in ResNet blocks). This made the former require more power while
CinC 2017 and was the most efficient model in CinC 2020. Despite performing less well than the latter.
the 1D CNN having a simpler base block than the 1D ResNet, the Equivalent CO2 emissions are directly related to the power
CNN required more layers than ResNet (12 CNN blocks vs. 4 consumption. Higher power consumption leads to higher CO2
emissions. Similar to the power consumption pattern, the 3.3.2. 1D ResNet on ECG signal classification
DenseNet121 model on the Poincaré input data exhibits the Our work also took advantage of GradCAM to explore the
highest CO2 emissions for both datasets, while the 1D ResNet mechanism of the 1D ResNet model. In medical literature, the
model on the raw signal input has the lowest emissions. ECG of Atrial fibrillation was detected by the irregular pattern in
Regarding the inference time, although XGBoost had a P- and T-waves around the QRS complex.
lightning prediction time, this model dominated the total Figure 6 shows the focusing points of the 1D ResNet when
inference time benchmark, which comes from the heavy predicting the AF signal. The yellow area is the segment that the
processing steps. This problem leads to the fact that XGBoost model attracts. These heatmaps show that the classifier focused
still inferred 24 times longer than the second place. The on the signal at the neighbor of the QRS complex. These regions
Poincaré-based method requires an approximate twofold longer are corresponding to the P-wave and T-wave of ECG recordings.
inference time than the 1D CNN or 1D ResNet. This result In fact, the absence or abnormality of P-wave and T-wave is
complies with the mathematical characteristics of the 1D and related to the fluctuation of heart rate and predicts arrhythmia
2D convolutional operators. disease (22).
FIGURE 5
Explaining CinC 2017 predictions on the Poincaré diagrams using GradCAM.
FIGURE 6
Explaining 1D ResNet decision by GradCAM methods in case of normal regimes and arrhythmia.
relatively complicated and expensive ECG procedure. Thus, it is discriminate the normal and AF data. The DenseNet detected AF
still worth studying this approach further. using the heart rate variability, which was measured by the
XGBoost’s performance is more impressive in the subset of spreading of the data cloud and the presence of data in the
long-term than the short-term recordings. This model learned upper left and lower right in the Poincaré diagram. On the other
more information in the long signal, and we can transfer that hand, the 1D ResNet assessed the AF pattern in raw ECG signal
knowledge to short signal via data augmentation. This could be similar to a medical expert: this model focused on the area
done by segmenting the long signal into many short signals, and around the QRS complex, which is also the location of P- and
then combining to the original dataset before training. Besides T-waves.
that, this gradient-boosting model has a long inference time
because of the expensive calculation in the feature engineering
step. The feature calculation could be improve by narrowing
Data availability statement
down the number of calculated features based on the feature
importance ranking.
Our study used the public repository of the challenge CinC
The one-dimensional convolutional model showed the best
2017 and CinC 2020. The code for reproducing results is
results in both studied datasets. Especially the 1D ResNet was
available at https://github.com/pnhuy/ecg-analysis. The original
superior to the first-ranking solution of each challenge. The
contributions presented in the study are included in the article/
residual connection showed its advantages in transferring
Supplementary Material, further inquiries can be directed to the
information while keeping the model not too deep. Our
corresponding author.
experiments proved the superior advantage of the 1D CNN–based
model in ECG signal classification over other deep learning
architectures. In the experiments on the CinC 2020 dataset, 1D
CNN and 1D ResNet were better than the transformer-based Author contributions
model of Natarajan et al., which is well known in sequence tasks.
To explain this behavior, we can consider the ECG classification KE, AK, and SB contributed to conception and design of the
as local pattern recognition, so the model needs to detect the study. HP organized the code repository and performed the
abnormalities in each P, Q, R, S, and T location in the signal. The experiments. All authors contributed to the article and approved
intensity, appearance, or absence of any component in that the submitted version.
complex is evidence of cardiovascular disease. This is also the way
the physician reads the ECG recordings. However, the attention
mechanism of the transformer architect is likely to collect the Conflict of interest
global information by connecting the information in many parts
of the sequence than focusing on a particular region of the series The authors declare that the research was conducted in the
(23). On the other side, when applied to the sequence processing, absence of any commercial or financial relationships that could
the CNN model used the small size filter to focus on a short and be construed as a potential conflict of interest.
fixed segment of signal, which was helpful in capturing the local
pattern of each region of the recording.
Although there were papers studying the power consumption
on ECG classification problem (24, 25), our paper is the first to
Publisher’s note
investigate the efficiency metrics while training the models,
All claims expressed in this article are solely those of the
including power consumption, equivalent CO2 emissions, as well
authors and do not necessarily represent those of their affiliated
as the inference time, in a wide range of models. The numerical
organizations, or those of the publisher, the editors and the
complexity is also represented by these metrics. Because of the
reviewers. Any product that may be evaluated in this article, or
high workload when processing 2D images, 2D ResNet and 2D
claim that may be made by its manufacturer, is not guaranteed
DenseNet are at the top in power consumption rankings. The
or endorsed by the publisher.
XGBoost is energy efficient for the short term, but the power
requirement is multiplied many times when training on long-
term signals, which came from the complexity of feature
engineering. Since the 1D convolution operator is optimized in Supplementary material
calculation, unidimensional models like 1D CNN and 1D ResNet
are the most energy efficient among the studied methods. The Supplementary Material for this article can be found
In the aspect of model interpretation, three models (DenseNet, online at: https://www.frontiersin.org/articles/10.3389/fcvm.2023.
1D ResNet, and XGBoost) were analyzed to figure out how they 1229743/full#supplementary-material
References
1. Tsao CW, Aday AW, Almarzooq ZI, Alonso A, Beaton AZ, Bittencourt 13. Carreiras C, Alves AP, Lourenço A, Canento F, Silva H, Fred A, et al.
MS, et al. Heart disease, stroke statistics—2022 update: a report from the BioSPPy: Biosignal Processing in Python [Dataset] (2015). Available at: https://
American Heart Association. Circulation. (2022) 145:e153–e639. doi: 10.1161/ biosppy.readthedocs.io/ (Accessed May 21, 2023).
CIR.0000000000001052 14. Zong W, Heldt T, Moody G, Mark R. An open-source algorithm to detect
2. Mensah GA, Roth GA, Fuster V. The global burden of cardiovascular diseases, onset of arterial blood pressure pulses. In: Murray A, editor. Computers in
risk factors: 2020, beyond. J Am Coll Cardiol. (2019) 74:2529–32. doi: 10.1016/j.jacc. cardiology, 2003. Thessaloniki Chalkidiki, Greece: IEEE (2003). p. 259–62.
2019.10.009 15. Hamilton P. Open source ECG analysis. In: Computers in cardiology. IEEE
3. Kligfield P, Gettes LS, Bailey JJ, Childers R, Deal BJ, Hancock EW, et al. (2002). p. 101–4.
Recommendations for the standardization and interpretation of the electrocardiogram. 16. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. 2016
Circulation. (2007) 115:1306–24. doi: 10.1161/CIRCULATIONAHA.106.180200 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.
4. Berkaya SK, Uysal AK, Gunal ES, Ergin S, Gunal S, Gulmezoglu MB. A survey on Piscataway, NJ: IEEE. (2016). p. 770–8.
ECG analysis. Biomed Signal Process Control. (2018) 43:216–35. doi: 10.1016/j.bspc. 17. Huang G, Liu Z, Van Der Maaten L, Weinberger KQ. Densely connected
2018.03.003 convolutional networks. 2017 IEEE Conference on Computer Vision and Pattern
5. Jun TJ, Nguyen HM, Kang D, Kim D, Kim D, Kim YH. ECG arrhythmia Recognition, Honolulu, HI, USA. Piscataway, NJ: IEEE (2017). p. 4700–8.
classification using a 2-D convolutional neural network. arXiv. (2018) 18. Head T, Kumar M, Nahrstaedt H, Louppe G, Shcherbatyi I. scikit-optimize/
arXiv:1804.06812. doi: 10.48550/arXiv.1804.06812 scikit-optimize (2021) [Dataset]. Available from: https://doi.org/10.5281/zenodo.
6. Hong S, Xu Y, Khare A, Priambada S, Maher K, Aljiffry A, et al. HOLMES: 5565057.
health online model ensemble serving for deep learning models in intensive care 19. He H, Garcia EA. Learning from imbalanced data. IEEE Trans Knowl Data Eng.
units. New York: ACM (2020). p. 1614–24. Available from: https://doi.org/10. (2009) 21:1263–84. doi: 10.1109/TKDE.2008.239
1145/3394486.3403212 (Accessed September 15, 2022).
20. Budennyy S, Lazarev V, Zakharenko N, Korovin A, Plosskaya O, Dimitrov D,
7. Ribeiro AH, Ribeiro MH, Paixão GMM, Oliveira DM, Gomes PR, Canazart JA, et al. Eco2AI: carbon emissions tracking of machine learning models as the
et al. Automatic diagnosis of the 12-lead ECG using a deep neural network. Nat first step towards sustainable AI. In: Doklady mathematics. Moscow, Russia:
Commun. (2020) 11:1760. doi: 10.1038/s41467-020-15432-4 Springer (2023). p. 1–11.
8. Zhang D, Yang S, Yuan X, Zhang P. Interpretable deep learning for automatic 21. Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-CAM:
diagnosis of 12-lead electrocardiogram. iScience. (2021) 24:102373. doi: 10.1016/j. visual explanations from deep networks via gradient-based localization. 2017 IEEE
isci.2021.102373 international conference on computer vision, Venice, Italy. Piscataway, NJ: IEEE
9. Clifford G, Liu C, Moody B, wei Lehman L, Silva I, Li Q, et al. AF classification (2017). p. 618–26. Available from: https://doi.org/10.1109/ICCV.2017.74 (Accessed
from a short single lead ECG recording: the physionet computing in cardiology September 15, 2022).
challenge 2017 (2017). Available from: https://doi.org/10.22489/CinC.2017.065-469 22. Hampton J. The ECG made easy. New York: Elsevier Health Sciences (2013).
(Accessed September 15, 2022).
23. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al.
10. Alday EAP, Gu A, Shah AJ, Robichaux C, Wong A-KI, Liu C, et al. Classification Attention is all you need. In: Guyon I, Luxburg UV, Bengio S, Wallach H, Fergus
of 12-lead ECGs: the physionet/computing in cardiology challenge 2020. Physiol Meas. R, Vishwanathan S, Garnett R, editors. Advances in neural information processing
(2020) 41:124003. doi: 10.1088/1361-6579/abc960 systems. Vol. 30. Long Beach, CA: Curran Associates, Inc. (2017).
11. Kamaleswaran R, Mahajan R, Akbilgic O. A robust deep convolutional neural 24. Amirshahi A, Hashemi M. ECG classification algorithm based on STDP and R-
network for the classification of abnormal cardiac rhythm using single lead STDP neural networks for real-time monitoring on ultra low-power personal wearable
electrocardiograms of variable length. Physiol Meas. (2018) 39:035006. doi: 10.1088/ devices. IEEE Trans Biomed Circuits Syst. (2019) 13:1483–93. doi: 10.1109/TBCAS.
1361-6579/aaaa9d 2019.2948920
12. Natarajan A, Chang Y, Mariani S, Rahman A, Boverman G, Vij SG, et al. A wide, 25. Yan Z, Zhou J, Wong W-F. Energy efficient ECG classification with spiking
deep transformer neural network for 12-lead ECG classification. 2020 Comput Cardiol. neural network. Biomed Signal Process Control. (2021) 63:102170. doi: 10.1016/j.
(2020) 12:1–4. doi: 10.1088/1361-6579/abc960 bspc.2020.102170