| Literature DB >> 35684694 |
Muhammad Zubair1, Changwoo Yoon1.
Abstract
Arrhythmia detection algorithms based on deep learning are attracting considerable interest due to their vital role in the diagnosis of cardiac abnormalities. Despite this interest, deep feature representation for ECG is still challenging and intriguing due to the inter-patient variability of the ECG's morphological characteristics. The aim of this study was to learn a balanced deep feature representation that incorporates both the short-term and long-term morphological characteristics of ECG beats. For efficient feature extraction, we designed a temporal transition module that uses convolutional layers with different kernel sizes to capture a wide range of morphological patterns. Imbalanced data are a key issue in developing an efficient and generalized model for arrhythmia detection as they cause over-fitting to minority class samples (abnormal beats) of primary interest. To mitigate the imbalanced data issue, we proposed a novel, cost-sensitive loss function that ensures a balanced deep representation of class samples by assigning effective weights to each class. The cost-sensitive loss function dynamically alters class weights for every batch based on class distribution and model performance. The proposed method acquired an overall accuracy of 99.81% for intra-patient classification and 96.36% for the inter-patient classification of heartbeats. The experimental results reveal that the proposed approach learned a balanced representation of ECG beats by mitigating the issue of imbalanced data and achieved an improved classification performance as compared to other studies.Entities:
Keywords: ECG classification; arrhythmia detection; convolutional neural networks; cost-sensitive learning; imbalanced data
Mesh:
Year: 2022 PMID: 35684694 PMCID: PMC9185309 DOI: 10.3390/s22114075
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Mapping of MIT-BIH arrhythmia database beat types to AAMI beat classes.
| AAMI Heartbeat Classes Description | |||||
|---|---|---|---|---|---|
| Normal beat (N) | Supraventricular ectopic beat (S) | Ventricular ectopic beat (V) | Fusion beat (F) | Unknown beat (Q) | |
| Normal beat (N) | Atrial premature beat (A) | Premature ventricular contraction (V) | Fusion of ventricular and normal beat (F) | Paced beat (/) | |
| Left bundle branch block beat (L) | Aberrated atrial premature beat (a) | Ventricular escape beat (E) | Fusion of paced and normal beat (f) | ||
|
| Right bundle branch block beat (R) | Nodal (junctional) premature beat (J) | Unclassified beat (Q) | ||
| Atrial escape beat (e) | Supraventricular premature beat (S) | ||||
| Nodal (junctional) escape beat (j) | |||||
Figure 1Overview of the heartbeat classification methodology.
Figure 2Pre-processing of ECG.
Summary of ECG Beats.
| Types | Number of Beats |
|---|---|
| Normal beats (N) | 89,976 |
| Supraventricular ectopic beats (S) | 2774 |
| Ventricular ectopic beats (V) | 7002 |
| Fusion beats (F) | 802 |
| Unknown beats (Q) | 15 |
| Total Beats | 100,569 |
Figure 3Architecture of the proposed deep learning model.
Figure 4Temporal transition module.
Distribution of training and test sets under the inter-patient paradigm.
| Set | Total Subjects | Subject ID | Samples Distribution | |
|---|---|---|---|---|
| DS1 | 22 | 101, 106, 108, 109, 112, 114, 115, 116, | N | 45,791 |
| S | 939 | |||
| V | 3785 | |||
| F | 414 | |||
| Q | 8 | |||
| DS2 | 22 | 100, 103, 105, 111, 113, 117, 121, 123, | N | 44,185 |
| S | 1835 | |||
| V | 3217 | |||
| F | 388 | |||
| Q | 7 | |||
Confusion matrix for the intra-patient paradigm.
| Predicted Labels | ||||||
|---|---|---|---|---|---|---|
| N | S | V | F | Q | ||
| True labels | N | 89,820 | 117 | 34 | 4 | 1 |
| S | 122 | 2636 | 16 | 0 | 0 | |
| V | 59 | 13 | 6895 | 35 | 0 | |
| F | 32 | 3 | 39 | 728 | 0 | |
| Q | 1 | 0 | 4 | 1 | 9 | |
Classification results for the intra-patient paradigm.
| Classes | Accuracy | Sensitivity | Specificity | Positive Productivity |
|---|---|---|---|---|
| N | 99.63% | 99.83% | 97.98% | 99.76 % |
| S | 99.73% | 95.03% | 99.86% | 95.20% |
| V | 99.80% | 98.47% | 99.90% | 98.67% |
| F | 99.89% | 90.77% | 99.96% | 94.79% |
| Q | 99.99% | 60.00% | 100% | 90.00% |
| Macro | 99.81% | 88.82% | 99.54% | 95.68% |
| Aggregate | 99.52% | 97.96% | 99.83% | 98.50% |
Confusion matrix for the inter-patient paradigm.
| Predicted Labels | ||||||
|---|---|---|---|---|---|---|
| N | S | V | F | Q | ||
| True labels | N | 40,629 | 1832 | 645 | 989 | 90 |
| S | 321 | 1427 | 83 | 1 | 3 | |
| V | 224 | 156 | 2788 | 29 | 20 | |
| F | 53 | 4 | 50 | 264 | 17 | |
| Q | 2 | 0 | 3 | 0 | 2 | |
Classification results for the inter-patient paradigm.
| Dataset | Classes | Accuracy | Sensitivity | Specificity | Positive Productivity |
|---|---|---|---|---|---|
| Train Set (DS1) | N | 98.47% | 98.54% | 97.86% | 99.76 % |
| S | 99.26% | 93.50% | 99.36% | 73.41% | |
| V | 99.16% | 96.94% | 99.34% | 92.16% | |
| F | 99.69% | 93.24% | 99.74% | 74.66% | |
| Q | 99.99% | 100% | 99.99% | 66.67% | |
| Macro | 99.31% | 96.44% | 99.26% | 81.33% | |
| Aggregate | 98.28% | 97.82% | 98.54% | 88.06% | |
| Test Set (DS2) | N | 91.63% | 91.95% | 88.98% | 98.54 % |
| S | 95.16% | 77.75% | 95.83% | 41.74% | |
| V | 97.56% | 86.66% | 98.32% | 78.12% | |
| F | 97.70% | 68.04% | 97.93% | 20.58% | |
| Q | 99.73% | 28.57% | 99.74% | 1.52% | |
| Macro | 96.36% | 70.60% | 96.16% | 48.10% | |
| Aggregate | 90.89% | 88.19% | 91.95% | 55.75% | |
Figure 5(a) Intra-patient beat classification performance comparison between cost-sensitive loss and conventional cross-entropy loss. (b) Inter-patient beat classification performance comparison between cost-sensitive loss and conventional cross-entropy loss.
Comparison of classifications for intra-patient and inter-patient beat classification.
| Authors | Intra-Patient Beat Classification | |||||
|---|---|---|---|---|---|---|
| Classes | Methodology | Acc | Sen | Spe | Ppr | |
| Yun-Chi et al. (2012) [ | 5 | clustering | 94.30% | 93.13% | - | 89.50% |
| Martis et al. (2013) [ | 5 | PNN | 99.63% | 99.83% | 97.92% | 99.75% |
| Acharya et al. (2017) [ | 5 | CNN+Synthetic data | 94.03% | 96.71% | 91.54% | 97.86% |
| Dang et al. (2017) [ | 5 | CNN | 95.48% | 96.53% | 87.74% | - |
| Raj et al. (2018) [ | 5 | SVM | 89.93% | 72.35% | - | 49.29% |
| Shu Lih Oh et al. (2019) [ | 5 | LSTM and CNN | 98.10% | 97.50% | 98.70% | 98.69% |
| Fujita et al. (2019) [ | 4 | CNN | 98.45% | 99.87% | 99.27% | -% |
| Fujita et al. (2019) [ | 4 | CNN + CWT | 97.78% | 99.76% | 98.82% | -% |
| Shu Lih Oh et al. (2019) [ | 5 | CNN | 98.45% | 86.02% | 98.40% | 87.02% |
| shaker et al. (2020) [ | 5 | GAN | 98.30% | 99.77% | 99.23% | 90.00% |
|
| 5 | CNN | 99.81% | 88.82% | 99.54% | 95.68% |
| Huang et al. (2014) [ | 5 | SVM + Threshold | 94.55% | 96.91% | 94.74% | 66.68% |
| Mathews et al. (2018) [ | 5 | Deep belief networks | 94.13% | 66.90% | 95.99% | 40.98% |
| A. Sellami et al.(2019) [ | 5 | CNN | 88.34% | 90.90% | 88.51% | 48.25% |
| Dechazal et al. (2004) [ | 5 | LDA | 85.88% | 92.85% | 86.86% | 42.21% |
| Wang et al. (2020) [ | 4 | Dual fully connected neural networks | 96.12% | 67.68% | 95.43% | 59.17% |
|
| 5 | CNN | 96.36% | 70.60% | 96.16% | 48.10% |
PNN: Probabilistic neural networks, CWT: Continuous wavelet transform, CNN: Convolutional neural networks, GAN: Generative adversarial networks, LDA: Linear discriminant analysis, SVM: Support vector machines.