| Literature DB >> 35978890 |
Degaga Wolde Feyisa1,2, Taye Girma Debelee1,2, Yehualashet Megersa Ayano1, Samuel Rahimeto Kebede1,3, Tariku Fekadu Assore4.
Abstract
The electrical activity produced during the heartbeat is measured and recorded by an ECG. Cardiologists can interpret the ECG machine's signals and determine the heart's health condition and related causes of ECG signal abnormalities. However, cardiologist shortage is a challenge in both developing and developed countries. Moreover, the experience of a cardiologist matters in the accurate interpretation of the ECG signal, as the interpretation of ECG is quite tricky even for experienced doctors. Therefore, developing computer-aided ECG interpretation is required for its wide-reaching effect. 12-lead ECG generates a 1D signal with 12 channels among the well-known time-series data. Classical machine learning can develop automatic detection, but deep learning is more effective in the classification task. 1D-CNN is being widely used for CVDS detection from ECG datasets. However, adopting a deep learning model designed for computer vision can be problematic because of its massive parameters and the need for many samples to train. In many detection tasks ranging from semantic segmentation of medical images to time-series data classification, multireceptive field CNN has improved performance. Notably, the nature of the ECG dataset made performance improvement possible by using a multireceptive field CNN (MRF-CNN). Using MRF-CNN, it is possible to design a model that considers semantic context information within ECG signals with different sizes. As a result, this study has designed a multireceptive field CNN architecture for ECG classification. The proposed multireceptive field CNN architecture can improve the performance of ECG signal classification. We have achieved a 0.72 F 1 score and 0.93 AUC for 5 superclasses, a 0.46 F 1 score and 0.92 AUC for 20 subclasses, and a 0.31 F 1 score and 0.92 AUC for all the diagnostic classes of the PTB-XL dataset.Entities:
Mesh:
Year: 2022 PMID: 35978890 PMCID: PMC9377844 DOI: 10.1155/2022/8413294
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1ECG waves, segments, and intervals [4].
Literature summary table.
| Author(s) | Dataset | Purpose | Method | Key findings | Limitations |
|---|---|---|---|---|---|
| Thanapatay et al. [ | MIT-BIH Arrhythmia Database | ECG beat classification | Discrete wavelet transform (DWT) and Principal Component Analysis (PCA) to extract 20 principal features and SVM for classification | The classifier achieved 99.6% in classifying beats | It is only for beat classification, and it is based on image processing techniques |
| Billeci et al. [ | MIT-BIH AF Arrhythmia Database | Classification in short ECG recordings acquired using a smartphone device | 51 features extracted and LS-SVM is used for classification | Performance for normal rhythm 0.98, AF rhythm 0.99, and global 0.98 | Atrial fibrillation classification only |
| Khatibi and Rabinezhadsadatmahaleh [ | MIT-BIH Arrhythmia Database | Arrhythmia detection from beat classification | Features extracted by pretraining ResNet50, VGGNet16, and handcrafted features (RR features). SVM, KNN, Decision Trees, and Random Forests are used for classification | It has a lower computational need for training than a deep learning model built from the ground up, and it outperforms classical machine learning | The model has millions of parameters, and it only detects arrhythmias |
| Zhao et al. [ | MIT-BIH Arrhythmia Database | Patient-specific classification of arrhythmias | ResNet34 deep learning architecture with a 1D filter | Average accuracy of 98.6% in the categorization of heartbeats | Model with millions of parameters and it overfits quickly |
| Śmigiel et al. [ | PTB-XL database | Subclasses of diagnostic categories in the PTB-XL | CNN with six layers with entropy as an additional feature | 0.698 accuracy, 0.332 | Low accuracy, |
| Śmigiel et al. [ | PTB-XL database | Classification of 2, 5, and 20 classes of heart diseases in the PTB-XL dataset | CNN to perform the encoding of a single QRS complex with the addition of entropy-based features | Adding entropy-based features and extracted QRS complexes to the raw signal is beneficial | Low accuracy, |
| Pałczyński et al. [ | PTB-XL database | Training deep CNN to recognize 2, 5, and 20 different heart disease classes from the PTB-XL dataset | Achieved better results in classifying five other disease classes than softmax-based counterparts | Determining the Few-Shot Learning (FSL) applicability for ECG signal proximity-based classification | Low accuracy, |
| Zhu et al. [ | PTB-XL database | Classifying ECG signal into abnormal and normal | CNN as feature extraction module, recursive feature elimination, and fully connected layer for final classification | 0.889 accuracy and 0.904 | It is only abnormality detection, no further effort to identify specific heart disease from the signal |
Figure 2Framework of the proposed methodology.
Hierarchical arrangement in the diagnostic class of the PTB-XL dataset.
| Superclasses | Subclasses | All diagnostic classes | Description |
|---|---|---|---|
| CD | IRBBB | IRBBB | Incomplete Right Bundle Branch Block |
| IVCD | IVCD | Nonspecific intraventricular CD | |
| CRBBB | CRBBB | Complete Right Bundle Branch Block | |
| CLBBB | CLBBB | Complete Left Bundle Branch Block | |
| LAFB/LPFB | LAFB | Left anterior fascicular block | |
| LPFB | Left posterior fascicular block | ||
| WPW | WPW | Wolff-Parkinson white syndrome | |
| ILBBB | ILBBB | Incomplete Left Bundle Branch Block | |
| _AVB | 3AVB | Third-degree AV block | |
| 2AVB | Second-degree AV block | ||
| AVB | First-degree AV block | ||
|
| |||
| HYP | LVH | LVH | Left ventricular hypertrophy |
| LAO/LAE | LAO/LAE | Left atrial overload/enlargement | |
| RVH | RVH | Right ventricular hypertrophy | |
| RAO/LAE | RAO/LAE | Right atrial overload/enlargement | |
| SEHYP | SEHYP | Septal hypertrophy | |
|
| |||
| MI | AMI | INJLA | Subendocardial injury infarction |
| ASMI | Anteroseptal myocardial infarction | ||
| INJAL | Subendocardial injury in the anterolateral leads | ||
| AMI | Anterior myocardial infarction | ||
| ALMI | Anterolateral myocardial infarction | ||
| INJAS | Subendocardial injury in anteroseptal leads | ||
| LMI | LMI | Lateral myocardial infarction | |
| IMI | IPLMI | Inferoposterolateral myocardial infarction | |
| IPMI | Inferoposterior myocardial infarction | ||
| ILMI | Inferolateral myocardial infarction | ||
| INJIL | Subendocardial injury in inferolateral leads | ||
| IMI | Inferior myocardial infarction | ||
| INJIN | Subendocardial injury in lateral leads | ||
| PMI | PMI | Posterior myocardial infarction | |
|
| |||
| NORM | NORM | NORM | Normal ECG |
|
| |||
| STTC | STTC | NDT | Nondiagnostic T abnormalities |
| DIG | Digitalis effect | ||
| ANEUR | ST-T changes compatible with ventricular aneurysm | ||
| EL | Electrolytic disturbance or drug | ||
| LNGQT | Long QT-interval | ||
| NST_ | NST_ | Nonspecific ST changes | |
| ISC_ | ISC_ | Nonspecific ischemic | |
| ISCI | ISCIN | Ischemic in inferior leads | |
| ISCIL | Ischemic in inferolateral leads | ||
| ISCA | ISCAL | Ischemic in inferior leads | |
| ISCAS | Ischemic in anteroseptal leads | ||
| ISCLA | Ischemic in lateral leads | ||
| ISCAN | Ischemic in anterior leads | ||
Figure 3Receptive field for 2-dimensional convolutional neural network [20].
Figure 4Multireceptive field (a) using two kernels with a dilation rate of 1 and (b) two kernels with a dilation rate of 2, while the dots represent data points in a given signal.
Typical amplitudes and durations of ECG signal for adult.
| Wave | Amplitudes (mV) | Durations (seconds) |
|---|---|---|
| P wave | 0.10–0.30 | 0.04–0.12 |
| PR interval | 120–200 | 0.12–0.20 |
| QRS interval | 1-2 | 0.05–0.10 |
| R wave | 0.2–1.7 | <0.07 |
| ST interval | — | 0.12–0.32 (mean) |
| ST segment | — | 0.24 (mean) |
| T wave | 0.05–0.80 | 0.10–0.25 |
| QT interval | — | 0.30–0.40 |
| PQRST | — | 0.42–0.60 |
Figure 5The proposed model architecture: the picture in (a) indicates the proposed lightweight and multireceptive field CNN architecture; the LargeBlock and SmallBlock in (b) have different kernel size and dilation rates to build the multireceptive field CNN. The model has six layers: five convolutional layers and one fully connected layer. Finally, Sigmoid is used as an activation function for class probability prediction.
Confusion matrix table for binary class classification.
| Prediction result | |||
|---|---|---|---|
| Positive | Negative | ||
| Actual result | Positive | True positive (TP) | False positive (FP) |
| Negative | False negative (FN) | True negative (TN) | |
Hyperparameter values.
| Hyperparameters | Value |
|---|---|
| Number of epochs | 100 |
| Dropout rate | 0.20 |
| Loss function | Binary cross-entropy |
| Batch size | 16, 32, 64, 128, and 256 |
| Optimizer | Adam, SGD, and RMSProp |
| Activation function | Leaky ReLu ( |
| Learning rate | 0.01, 0.001, 0.0001, and 0.00001 |
Results of the proposed model on the dataset after removing classes with sample number of less than 20 and feeding the whole signal to the model without using any sliding window approach.
| Class number | Accuracy | Precision | Recall |
| AUC |
|---|---|---|---|---|---|
| 5 (superdiagnostic) | 0.897 | 0.73 | 0.71 | 0.72 | 0.93 |
| 20 (subdiagnostic) | 0.962 | 0.42 | 0.56 | 0.46 | 0.92 |
| 41 (diagnostic) | 0.98 | 0.28 | 0.31 | 0.29 | 0.92 |
Comparison of results of the proposed model and those of the existing model for 20 class number.
| Model | Acc | Prec | Recall |
| AUC | Total param |
|---|---|---|---|---|---|---|
| [ | 0.765 | 0.355 | 0.339 | 0.332 | 0.815 | 58,664 |
| [ | 0.685 | — | — | 0.336 | 0.861 | — |
| [ | 0.671 | — | — | 0.324 | 0.844 | — |
| Proposed | 0.962 | 0.420 | 0.560 | 0.460 | 0.920 | 56,732 |
Comparison of results of the proposed model and those of the existing model for 5 class number.
| Model | Acc | Prec | Recall |
| AUC | Total param |
|---|---|---|---|---|---|---|
| [ | 0.765 | 0.714 | 0.662 | 0.680 | 0.910 | 58,259 |
| [ | 0.763 | — | — | 0.683 | 0.907 | — |
| [ | 0.790 | — | — | 0.717 | 0.936 | — |
| Proposed | 0.897 | 0.730 | 0.710 | 0.720 | 0.930 | 55,277 |
Comparison of results of the proposed model and those of the existing model for the diagnostic, subdiagnostic, and superdiagnostic classes number.
| Authors/method | Diagnostic | Subdiagnostic | Superdiagnostic | Total parameters | ||||
|---|---|---|---|---|---|---|---|---|
| Authors | Method | AUC |
| AUC |
| AUC |
| |
| [ | LSTM-BiDir | 0.932 | 0.737 | 0.923 | 0.757 | 0.921 | 0.815 | 2,332,564 |
| XResNet101 | 0.937 | 0.736 | 0.929 | 0.760 | 0.928 | 0.815 | 1,874,196 | |
| LSTM | 0.927 | 0.731 | 0.928 | 0.759 | 0.927 | 0.820 | 905,620 | |
| Inception | 0.931 | 0.737 | 0.930 | 0.752 | 0.921 | 0.810 | 509,588 | |
| ResNet-Wang | 0.936 | 0.741 | 0.928 | 0.762 | 0.930 | 0.823 | 745,284 | |
| FCN-Wang | 0.926 | 0.735 | 0.928 | 0.762 | 0.930 | 0.823 | 311,700 | |
| Proposed | 0.930 | 0.720 | 0.922 | 0.743 | 0.927 | 0.816 | 59,060 | |
Result of the proposed model without using the sliding window approach and by keeping the whole classes in the dataset.
| Diagnostic | Subdiagnostic | Superdiagnostic | Total parameters | |||
|---|---|---|---|---|---|---|
| AUC |
| AUC |
| AUC |
| 59,060 |
| 0.879 | 0.677 | 0.904 | 0.685 | 0.910 | 0.783 | |