| Literature DB >> 33841584 |
M Fraiwan1, L Fraiwan2, M Alkhodari3, O Hassanin3.
Abstract
In this paper, a study is conducted to explore the ability of deep learning in recognizing pulmonary diseases from electronically recorded lung sounds. The selected data-set included a total of 103 patients obtained from locally recorded stethoscope lung sounds acquired at King Abdullah University Hospital, Jordan University of Science and Technology, Jordan. In addition, 110 patients data were added to the data-set from the Int. Conf. on Biomedical Health Informatics publicly available challenge database. Initially, all signals were checked to have a sampling frequency of 4 kHz and segmented into 5 s segments. Then, several preprocessing steps were undertaken to ensure smoother and less noisy signals. These steps included wavelet smoothing, displacement artifact removal, and z-score normalization. The deep learning network architecture consisted of two stages; convolutional neural networks and bidirectional long short-term memory units. The training of the model was evaluated based on a k-fold cross-validation scheme of tenfolds using several performance evaluation metrics including Cohen's kappa, accuracy, sensitivity, specificity, precision, and F1-score. The developed algorithm achieved the highest average accuracy of 99.62% with a precision of 98.85% in classifying patients based on the pulmonary disease types using CNN + BDLSTM. Furthermore, a total agreement of 98.26% was obtained between the predictions and original classes within the training scheme. This study paves the way towards implementing deep learning models in clinical settings to assist clinicians in decision making related to the recognition of pulmonary diseases. SUPPLEMENTARY INFORMATION: The online version supplementary material available at 10.1007/s12652-021-03184-y.Entities:
Keywords: Convolutional neural network; Deep learning; Long short-term memory; Lung sounds; Pulmonary diseases; Stethoscope
Year: 2021 PMID: 33841584 PMCID: PMC8019351 DOI: 10.1007/s12652-021-03184-y
Source DB: PubMed Journal: J Ambient Intell Humaniz Comput
Fig. 1The complete procedure followed in the proposed study
Demographic information of the subjects
| Data-set | Category | Normal | Asthma | Pneumonia | BRON | COPD | HF | Overall |
|---|---|---|---|---|---|---|---|---|
| Local recordings | Number of subjects | 35 (24 M, 11 F) | 32 (15 M, 17 F) | 5 (3 M, 2 F) | 3 (2 M, 1 F) | 9 (8 M, 1 F) | 19 (10 M, 9 F) | 103 (62 M, 41 F) |
| Age (mean ± SD) | 43 ± 20 | 46 ± 16 | 56 ± 10 | 37 ± 27 | 57 ± 10 | 59 ± 19 | 50 ± 17 | |
| Number of recordings | 110 | 88 | 18 | 6 | 23 | 56 | 301 | |
| ICBHI’17 | Number of subjects | 26 (13 M, 13 F) | 1 (1 F) | 6 (3 M, 2 F) | 13 (6 M, 7 F) | 64 (48 M, 16 F) | N/A | 110 (70 M, 39 F) |
| Age (mean ± SD) | 12 ± 20 | 70 | 62 ± 29 | 25 ± 21 | 69 ± 8 | N/A | 48 ± 20 | |
| Number of recordings | 135 | 4 | 148 | 116 | 779 | N/A | 1182 |
SD standard deviation
Fig. 2Examples of lung sound signals coming from normal and five types of respiratory diseases patients: a normal, b asthma, c pneumonia, d BRON, e COPD, f HF
Fig. 3Convolutional neural network and bidirectional long short-term memory (CNN + BDLSTM) model architecture
Fig. 4The structure of the convolutional neural network (CNN) designed in the proposed study
Fig. 5The structure of the bidirectional long short-term memory (BDLSTM) designed in the proposed study
Fig. 6The preprocessing of a selected lung sound signal (2.5 s segment) showing: a original signal, b MODWT wavelet smoothing, c rLOESS displacement removal, d z-score normalization
Fig. 7The confusion matrix and per-class precision percentage in respiratory diseases recognition using: a BDLSTM, b CNN, c CNN + BDLSTM
Performance comparison, expressed in , of different neural networks (in the order BDLSTM/CNN/CNN + BDLSTM) based on tenfold cross validation
| Condition | Cohen’s kappa ( | Accuracy | Sensitivity | Specificity | Precision | F1-score |
|---|---|---|---|---|---|---|
| Normal | 92.41/94.91/98.29 | 97.91/98.58/99.53 | 93.47/96.73/98.37 | 98.79/98.95/99.76 | 93.85/94.80/98.77 | 93.66/95.76/98.57 |
| Asthma | 86.23/90.13/94.76 | 98.38/98.92/99.39 | 88.04/84.78/94.57 | 99.07/99.86/99.71 | 86.17/97.50/95.60 | 87.10/90.70/95.08 |
| Pneumonia | 92.90/94.17/99.33 | 98.58/98.79/99.87 | 93.98/100.00/100.00 | 99.16/98.63/99.85 | 93.41/90.22/98.81 | 93.69/94.86/99.40 |
| BRON | 86.39/97.26/99.10 | 98.04/99.60/99.87 | 82.79/95.08/98.36 | 99.41/100.00/100.00 | 92.66/100.00/100.00 | 87.45/97.48/99.17 |
| COPD | 93.88/97.96/98.10 | 96.97/98.99/99.06 | 98.13/99.38/99.25 | 95.59/98.53/98.83 | 96.33/98.76/99.00 | 97.22/99.07/99.13 |
| HF | 86.55/89.85/100.00 | 99.06/99.33/100.00 | 83.93/82.14/100.00 | 99.65/100.00/100.00 | 90.38/100.00/100.00 | 87.04/90.20/100.00 |
| Overall (average) | 89.73/94.05/98.26 | 98.16/99.04/99.62 | 90.06/93.02/98.43 | 98.61/99.33/99.69 | 92.13/96.88/98.70 | 91.03/94.68/98.56 |
Summary table of recent studies found in literature for the use of machine/deep learning approaches in lung sounds classification
| Study | No. patients | No. recordings | No. classes | Extracted features | Models | Performance |
|---|---|---|---|---|---|---|
|
Aykanat et al. ( | 1630 | 15,328 | 3: Normal. rale, rhonchus | MFCC/spectrograms | SVM/CNN | Accuracy: 80.00 Sensitivity: 89.00 Specificity: N/A |
|
Bardou et al. ( | 15 | 2141 | 7: Normal, monophonic wheeze polyphonic wheeze, stridor squawk, fine crackle, coarse crackle | Spectrograms | CNN | Accuracy: 95.56 Sensitivity: N/A Specificity: N/A |
|
Shi et al. ( | 384 | 1152 | 3: Normal, asthma, pneumonia | Spectrograms | VGG-BDGRU | Accuracy: 87.41 Sensitivity: N/A Specificity: N/A |
|
Demir et al. ( | 126 | 6898 | 4: Normal, crackles, wheezes crackles+wheezes | Spectrograms | CNN | Accuracy: 71.15 Sensitivity: 61.00 Specificity: 86.00 |
|
García-Ordás et al. ( | 126 | 920 | 6: Normal, asthma, pneumonia BRON, COPD, respiratory tract infection | Spectrograms | CNN | Accuracy: N/A Sensitivity: 98.81 Specificity: 98.61 |
| This study | 213 | 1,483 | 6: Normal, asthma, pneumonia BRON, COPD. heart failure | Spatial and temporal (CNN + BDLSTM) | CNN + BDLSTM | Accuracy: 99.62 Sensitivity: 98.43 Specificity: 99.69 |
Fig. 8Examples of three correctly classified signals and three miss-classified signals along with the prediction probabilities using the best performing deep learning model (CNN + BDLSTM)