| Literature DB >> 32062420 |
Tsai-Min Chen1, Chih-Han Huang2, Edward S C Shih3, Yu-Feng Hu4, Ming-Jing Hwang5.
Abstract
Electrocardiograms (ECGs) are widely used to clinically detect cardiac arrhythmias (CAs). They are also being used to develop computer-assisted methods for heart disease diagnosis. We have developed a convolution neural network model to detect and classify CAs, using a large 12-lead ECG dataset (6,877 recordings) provided by the China Physiological Signal Challenge (CPSC) 2018. Our model, which was ranked first in the challenge competition, achieved a median overall F1-score of 0.84 for the nine-type CA classification of CPSC2018's hidden test set of 2,954 ECG recordings. Further analysis showed that concurrent CAs were adequately predictive for 476 patients with multiple types of CA diagnoses in the dataset. Using only single-lead data yielded a performance that was only slightly worse than using the full 12-lead data, with leads aVR and V1 being the most prominent. We extensively consider these results in the context of their agreement with and relevance to clinical observations.Entities:
Keywords: Arrhythmology; Artificial Intelligence; Diagnostics
Year: 2020 PMID: 32062420 PMCID: PMC7031313 DOI: 10.1016/j.isci.2020.100886
Source DB: PubMed Journal: iScience ISSN: 2589-0042
Figure 1The Architecture of Deep Learning Artificial Neural Network for 12-Lead ECG CA Detection and Classification
Layers and blocks are specified in rectangle boxes; “X5” indicates that five CNN blocks are tandem connected before connecting to the bidirectional RNN layer, which is a gated recurrent unit layer. The output layer at the bottom contains the probabilities predicted by the model for each of the nine types of the CA classification. The type with the highest probability is the type predicted by the model for the input ECG recording.
Comparison of Model Performances on Tests
| Best Validation Models | Ensemble Model | ||||||
|---|---|---|---|---|---|---|---|
| CA Typ | Median Accuracy | Median AUC (95% CI) | Median F1-Score | Median Accuracy | Median AUC (95% CI) | Median F1-Score | Hidden Set F1-Score |
| Normal | 0.940 | 0.890 (0.810–0.942) | 0.795 | 0.949 | 0.867 (0.832–0.973) | 0.808 | |
| AF | 0.969 | 0.928 (0.902–0.985) | 0.897 | 0.983 | 0.963 (0.914–0.993) | 0.944 | |
| I-AVB | 0.972 | 0.899 (0.864–0.988) | 0.865 | 0.977 | 0.950 (0.875–0.990) | 0.899 | |
| LBBB | 0.990 | 0.914 (0.748–1.000) | 0.821 | 0.995 | 0.942 (0.763–1.000) | 0.899 | |
| RBBB | 0.955 | 0.956 (0.887–0.988) | 0.911 | 0.952 | 0.946 (0.871–0.976) | 0.903 | |
| PAC | 0.957 | 0.867 (0.749–0.955) | 0.734 | 0.963 | 0.920 (0.779–0.981) | 0.797 | |
| PVC | 0.970 | 0.928 (0.841–0.988) | 0.852 | 0.977 | 0.932 (0.864–0.996) | 0.874 | |
| STD | 0.951 | 0.878 (0.797–0.972) | 0.788 | 0.959 | 0.906 (0.815–0.970) | 0.834 | |
| STE | 0.976 | 0.707 (0.558–0.995) | 0.509 | 0.977 | 0.773 (0.603–0.993) | 0.600 | |
Results are from the best validation models and the ensemble model on the ten 10-fold tests, except for those in the last column (boldfaced), which are the ensemble model's median F1-scores for the hidden test set of CPSC2018 reported at its website http://2018.icbeb.org/Challenge.html, which did not provide accuracy or AUC results.
Figure 2Probabilities Output by the Best Validation Models in the Test Folds of the 10-Fold Tests
On the right is the color-coded probability scale.
Label Count Statistics of the 476 Multi-Labeled Subjects in the Released CPSC2018 Dataset
| AF | I-AVB | LBBB | RBBB | PAC | PVC | STD | STE | |
|---|---|---|---|---|---|---|---|---|
| AF | 0 | 0 | 29 | 4 | 8 | 33 | 2 | |
| I-AVB | 0 | 8 | 10 | 3 | 5 | 6 | 4 | |
| LBBB | 0 | 0 | 10 | 6 | 3 | 4 | ||
| RBBB | 0 | 20 | 19 | |||||
| PAC | 2 | 3 | 6 | 5 | ||||
| PVC | 0 | 18 | 2 | |||||
| STD | 0 | 2 | ||||||
| STE | 0 |
Only the upper triangle portion of the symmetrical concurrent CA label counts is shown. The three largest counts are boldfaced.
Figure 3The Ranked F1-Score Results of Single-Lead Models
The F1-scores (on the y axis) are from the single-lead models performed on the 10-fold tests (see Methods). Lead aVR is shown in red, V1 in green, I in blue, and II in orange.