| Literature DB >> 35270923 |
Kwang-Sig Lee1, Hyun-Joon Park2, Ji Eon Kim3, Hee Jung Kim3, Sangil Chon4, Sangkyu Kim4, Jaesung Jang4, Jin-Kook Kim4, Seongbin Jang4, Yeongjoon Gil4, Ho Sung Son3.
Abstract
The importance of an embedded wearable device with automatic detection and alarming cannot be overstated, given that 15-30% of patients with atrial fibrillation are reported to be asymptomatic. These asymptomatic patients do not seek medical care, hence traditional diagnostic tools including Holter are not effective for the further prevention of associated stroke or heart failure. This is likely to be more so in the era of COVID-19, in which patients become more reluctant on hospitalization and checkups. However, little literature is available on this important topic. For this reason, this study developed efficient deep learning with model compression, which is designed to use ECG data and classify arrhythmia in an embedded wearable device. ECG-signal data came from Korea University Anam Hospital in Seoul, Korea, with 28,308 unique patients (15,412 normal and 12,896 arrhythmia). Resnets and Mobilenets with model compression (TensorFlow Lite) were applied and compared for the diagnosis of arrhythmia in an embedded wearable device. The weight size of the compressed model registered a remarkable decrease from 743 MB to 76 KB (1/10000), whereas its performance was almost the same as its original counterpart. Resnet and Mobilenet were similar in terms of accuracy, i.e., Resnet-50 Hz (97.3) vs. Mo-bilenet-50 Hz (97.2), Resnet-100 Hz (98.2) vs. Mobilenet-100 Hz (97.9). Here, 50 Hz/100 Hz denotes the down-sampling rate. However, Resnets took more flash memory and longer inference time than did Mobilenets. In conclusion, Mobilenet would be a more efficient model than Resnet to classify arrhythmia in an embedded wearable device.Entities:
Keywords: Mobilenet; Resnet; arrhythmia; compressed deep learning; embedded wearable device
Mesh:
Year: 2022 PMID: 35270923 PMCID: PMC8914813 DOI: 10.3390/s22051776
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Categories: Normal vs. Arrhythmia.
| Category | Diagnosis | Count |
|---|---|---|
| 1 | Normal Sinus Rhythm | 9760 |
| 2 | Sinus Bradycardia | 1944 |
| 3 | Sinus Tachycardia | 1754 |
| 4 | First-Degree Atrioventricular Block | 1954 |
| 5 | Premature Ventricular Contraction | 1566 |
| 6 | Atrial Fibrillation | 9584 |
| 7 | Atrial Flutter | 1746 |
| Total | 28,308 |
Note: Normal [1,2,3,4] vs. Arrhythmia [5,6,7].
Figure 1Electrocardiogram Signal.
Figure 2Resnet Architecture. Note: BN Batch Normalization, CONV Convolution, PARAMS Parameters.
Figure 3Mobilenet Architecture. Note: BN Batch Normalization, CONV Convolution, PARAMS Parameters.
Model Compression for Deep Learning.
| Approach | Explanation |
|---|---|
| Pruning | We use pruning to remove some of model weights, i.e., to set their values as zeroes: suitable for both training from scratch and using a pre-trained model [ |
| Quantization | We use quantization to decrease the sizes of the weights by mapping their values in an original set to their smaller-set counterparts (e.g., 8-bit to 1-bit): suitable for both training from scratch and using a pre-trained model [ |
| Clustering | We use clustering to divide the weights into several groups, then share central values for all weights in the same group: suitable for both training from scratch and using a pre-trained model [ |
| Low-Rank Approximation | We use low-rank approximation to reduce the redundancy (or “rank”) of convolutional filters, that is, to approximate the original filters based on their lower-rank counterparts: suitable for both training from scratch and using a pre-trained model |
| Knowledge Distillation | We use knowledge distillation to condense an original model to its smaller counterpart with a similar loss function (and performance): suitable for training from scratch [ |
Confusion Matrix.
| True | |||
|---|---|---|---|
|
|
| ||
|
|
| TP (True Positive) | FP (False Positive) |
|
| FN (False Negative) | TN (True Negative) |
Original vs. Compressed Deep Learning: Model Weight Size, Accuracy and Inference Time.
| Cloud Version | Embedded Version | |
|---|---|---|
| Model Weight Size | 743 MB |
|
| Accuracy | 98.4% |
|
| Inference Time | NA |
|
Model Performance.
|
|
|
| ||||
|
| Metric | Validation Set | Test Set | Metric | Validation Set | Test Set |
|
| Acc | 0.9735 | 0.9728 | Acc | 0.9802 | 0.9823 |
| F1 | 0.9721 | 0.9706 | F1 | 0.9791 | 0.9808 | |
| Sensitivity | 0.9660 | 0.9868 | Sensitivity | 0.9850 | 0.9907 | |
| Specificity | 0.9820 | 0.9611 | Specificity | 0.9760 | 0.9753 | |
| Precision | 0.9837 | 0.9550 | Precision | 0.9733 | 0.9711 | |
| AUC | 0.9932 | 0.9937 | AUC | 0.9964 | 0.9963 | |
|
|
|
| ||||
|
| Metric | Validation Set | Test Set | Metric | Validation Set | Test Set |
|
| Acc | 0.9675 | 0.9717 | Acc | 0.9785 | 0.9792 |
| F1 | 0.9656 | 0.9694 | F1 | 0.9772 | 0.9773 | |
| Sensitivity | 0.9692 | 0.9822 | Sensitivity | 0.9812 | 0.9829 | |
| Specificity | 0.9660 | 0.9630 | Specificity | 0.9760 | 0.9760 | |
| Precision | 0.9620 | 0.9569 | Precision | 0.9732 | 0.9716 | |
| AUC | 0.9942 | 0.9908 | AUC | 0.9945 | 0.9967 | |
Note: Acc Accuracy, AUC Area Under the Receiver-Operating-Characteristic Curve.
Figure 4Area Under the Receiver-Operating-Characteristic Curve for the Test Set. Note: TPR True Positive Rate (Sensitivity), FPR False Positive Rate (1—Specificity).
Figure 5Model Size (a) Model Size (FLASH) (b) Model Arena Size (SRAM) (c) Model Build Size.
Figure 6Execution Time.
Figure 7Current Consumption.
Figure 8Lite Module. Source: [23].
Figure 9Fused Lightweight Recurrent Neural Network. Source: [25].