| Literature DB >> 35270938 |
Amith Khandakar1,2, Muhammad E H Chowdhury1, Mamun Bin Ibne Reaz2, Sawal Hamid Md Ali2, Tariq O Abbas3, Tanvir Alam4, Mohamed Arselene Ayari5, Zaid B Mahbub6, Rumana Habib7, Tawsifur Rahman1, Anas M Tahir1, Ahmad Ashrif A Bakar2, Rayaz A Malik8.
Abstract
Diabetes mellitus (DM) can lead to plantar ulcers, amputation and death. Plantar foot thermogram images acquired using an infrared camera have been shown to detect changes in temperature distribution associated with a higher risk of foot ulceration. Machine learning approaches applied to such infrared images may have utility in the early diagnosis of diabetic foot complications. In this work, a publicly available dataset was categorized into different classes, which were corroborated by domain experts, based on a temperature distribution parameter-the thermal change index (TCI). We then explored different machine-learning approaches for classifying thermograms of the TCI-labeled dataset. Classical machine learning algorithms with feature engineering and the convolutional neural network (CNN) with image enhancement techniques were extensively investigated to identify the best performing network for classifying thermograms. The multilayer perceptron (MLP) classifier along with the features extracted from thermogram images showed an accuracy of 90.1% in multi-class classification, which outperformed the literature-reported performance metrics on this dataset.Entities:
Keywords: deep learning; diabetic foot; machine learning; thermal change index; thermogram
Mesh:
Year: 2022 PMID: 35270938 PMCID: PMC8915003 DOI: 10.3390/s22051793
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Illustration of the computational workflow for this study.
Details of the dataset used for training (with and without augmentation), validation, and testing.
| Dataset | Count of Diabetic Thermograms/Cluster Identified in the Paper | Training Dataset Details | |||
|---|---|---|---|---|---|
| Training (60% of the Data) Thermogram/Fold | Augmented Train Thermogram/Fold | Validation (20% of the Data) Thermogram/Fold | Test (20% of the Data) Image/Fold | ||
| Contreras et al. [ | Class 1 (TCI < 2) | 34 | 1020 | 11 | 11 |
| Class 2 (2 < TCI < 3) | 22 | 1100 | 07 | 07 | |
| Class 3 (3 < TCI < 4) | 17 | 1020 | 05 | 06 | |
| Class 4 (4 < TCI < 5) | 22 | 1100 | 07 | 08 | |
| Class 5 (5 < TCI) | 52 | 1044 | 17 | 18 | |
Figure 2(a) Sample of thermograms from different classes [22] and (b) sample of MPA, LPA, MCA, and LCA angiosomes of the foot for the control and diabetic Group [23].
Figure 3Original thermogram versus enhanced thermogram using AHE and Gamma correction for diabetic patient’s thermogram [23].
Performance metrics and inference time for five-fold cross-validation using 2D CNNs.
| Enhancement | Network | 95% Confidence Interval Results | ||||||
|---|---|---|---|---|---|---|---|---|
| Class | Accuracy | Precision | Sensitivity | F1-Score | Specificity | Inference Time | ||
| Original | ResNet50 | Class 1 | 92.62 ± 06.85 | 81.67 ± 10.13 | 87.50 ± 08.66 | 84.48 ± 09.48 | 94.15 ± 06.15 | 6.247 |
| Class 2 | 88.93 ± 10.25 | 60.98 ± 15.93 | 69.44 ± 15.05 | 64.40 ± 15.64 | 92.31 ± 08.70 | |||
| Class 3 | 90.98 ± 10.61 | 66.67 ± 17.46 | 42.86 ± 18.33 | 52.18 ± 18.50 | 97.22 ± 06.09 | |||
| Class 4 | 86.89 ± 10.88 | 55.56 ± 16.01 | 67.57 ± 15.08 | 60.98 ± 15.72 | 90.34 ± 09.52 | |||
| Class 5 | 93.85 ± 05.05 | 95.00 ± 04.58 | 87.36 ± 06.98 | 91.02 ± 06.01 | 97.45 ± 03.31 | |||
| Overall | 91.46 ± 03.51 | 77.69 ± 05.22 | 76.64 ± 05.31 | 76.66 ± 05.31 | 94.83 ± 02.78 | |||
| AHE | MobileNetv2 | Class 1 | 94.26 ± 6.09 | 88.89 ± 08.23 | 85.71 ± 09.17 | 87.27 ± 08.73 | 96.81 ± 04.60 | 5.412 |
| Class 2 | 91.39 ± 09.16 | 70.27 ± 14.93 | 72.22 ± 14.63 | 71.23 ± 14.79 | 94.71 ± 07.31 | |||
| Class 3 | 88.11 ± 11.99 | 47.06 ± 18.49 | 28.57 ± 16.73 | 35.55 ± 17.73 | 95.83 ± 07.40 | |||
| Class 4 | 84.43 ± 11.68 | 48.94 ± 16.11 | 62.16 ± 15.63 | 54.76 ± 16.04 | 88.41 ± 10.31 | |||
| Class 5 | 94.26 ± 04.89 | 91.01 ± 06.01 | 93.10 ± 05.33 | 92.04 ± 05.69 | 94.90 ± 04.62 | |||
| Overall | 91.64 ± 03.47 | 76.04 ± 05.36 | 76.23 ± 05.34 | 75.74 ± 05.38 | 94.43 ± 02.88 | |||
| Original | ResNet18 | Class 1 | 92.21 ± 07.02 | 83.64 ± 09.69 | 82.14 ± 10.03 | 82.88 ± 09.87 | 95.21 ± 05.59 | 2.536 |
| Class 2 | 88.11 ± 10.57 | 58.14 ± 16.12 | 69.44 ± 15.05 | 63.29 ± 15.75 | 91.35 ± 09.18 | |||
| Class 3 | 90.98 ± 10.61 | 63.64 ± 17.82 | 50.00 ± 18.52 | 56.00 ± 18.39 | 96.30 ± 06.99 | |||
| Class 4 | 86.89 ± 10.88 | 56.10 ± 15.99 | 62.16 ± 15.63 | 58.97 ± 15.85 | 91.30 ± 09.08 | |||
| Class 5 | 92.62 ± 05.49 | 91.57 ± 05.84 | 87.36 ± 06.98 | 89.42 ± 06.46 | 95.54 ± 04.34 | |||
| Overall | 90.80 ± 03.63 | 76.23 ± 05.34 | 75.41 ± 05.40 | 75.61 ± 05.39 | 94.29 ± 02.91 | |||
| Gamma Correction | ResNet18 | Class 1 | 93.03 ± 06.67 | 88.24 ± 08.44 | 80.36 ± 10.41 | 84.12 ± 09.57 | 96.81 ± 04.60 | 3.347 |
| Class 2 | 89.75 ± 09.91 | 63.41 ± 15.73 | 72.22 ± 14.63 | 67.53 ± 15.30 | 92.79 ± 08.45 | |||
| Class 3 | 90.16 ± 11.03 | 59.09 ± 18.21 | 46.43 ± 18.47 | 52.00 ± 18.51 | 95.83 ± 07.40 | |||
| Class 4 | 82.79 ± 12.16 | 44.90 ± 16.03 | 59.46 ± 15.82 | 51.16 ± 16.11 | 86.96 ± 10.85 | |||
| Class 5 | 91.80 ± 05.77 | 91.36 ± 05.90 | 85.06 ± 07.49 | 88.10 ± 06.80 | 95.54 ± 04.34 | |||
| Overall | 90.23 ± 03.73 | 75.77 ± 05.38 | 73.77 ± 05.52 | 74.41 ± 05.48 | 94.16 ± 02.94 | |||
| Gamma Correction | ResNet50 | Class 1 | 92.21 ± 07.02 | 80.33 ± 10.41 | 87.50 ± 08.66 | 83.76 ± 09.66 | 93.62 ± 06.40 | 7.764 |
| Class 2 | 88.93 ± 10.25 | 63.64 ± 15.71 | 58.33 ± 16.11 | 60.87 ± 15.94 | 94.23 ± 07.62 | |||
| Class 3 | 87.30 ± 12.33 | 44.00 ± 18.39 | 39.29 ± 18.09 | 41.51 ± 18.25 | 93.52 ± 09.12 | |||
| Class 4 | 87.70 ± 10.58 | 59.46 ± 15.82 | 59.46 ± 15.82 | 59.46 ± 15.82 | 92.75 ± 08.36 | |||
| Class 5 | 93.03 ± 05.35 | 89.77 ± 06.37 | 90.80 ± 06.07 | 90.28 ± 06.22 | 94.27 ± 04.88 | |||
| Overall | 90.77 ± 03.63 | 73.90 ± 05.51 | 74.59 ± 05.46 | 74.17 ± 05.49 | 93.80 ± 03.03 | |||
Figure 4AUC for the (A) original, (B) AHE-enhanced, and (C) Gamma-enhanced thermograms in TCI-based classification.
Figure 5Top ranked 10 features from the reduced 28 features using feature selection techniques (A) XGBoost, (B) Random Forest, and (C) Extra Tree.
Performance metrics for the best-performing combinations (feature selection technique and number of features) for the 10 ML Classifiers.
| Classifier | Feature Selection | # of Feature | 95% Confidence Interval Results | Inference Time (ms) | ||||
|---|---|---|---|---|---|---|---|---|
| Accuracy | Precision | Sensitivity | F1-Score | Specificity | ||||
| MLP | XGBoost | 2 | 0.91 ± 01.19 | 0.91 ± 01.19 | 0.91 ± 01.19 | 0.91 ± 01.19 | 0.91 ± 01.19 | 0.592 |
| Extra Tree | Random Forest | 5 | 0.88 ± 01.17 | 0.88 ± 01.17 | 0.88 ± 01.17 | 0.88 ± 01.17 | 0.88 ± 01.17 | 0.406 |
| Random Forest | XGBoost | 2 | 0.87 ± 01.17 | 0.87 ± 01.17 | 0.87 ± 01.17 | 0.87 ± 01.17 | 0.87 ± 01.17 | 0.412 |
| KNN | XGBoost | 2 | 0.87 ± 01.17 | 0.87 ± 01.17 | 0.87 ± 01.17 | 0.87 ± 01.17 | 0.87 ± 01.17 | 0.464 |
| SVM | XGBoost | 2 | 0.86 ± 01.16 | 0.86 ± 01.16 | 0.86 ± 01.16 | 0.86 ± 01.16 | 0.86 ± 01.16 | 0.456 |
| Gradient Boost | XGBoost | 2 | 0.84 ± 01.15 | 0.84 ± 01.15 | 0.84 ± 01.15 | 0.85 ± 01.15 | 0.84 ± 01.15 | 0.492 |
| XGBoost | Random Forest | 5 | 0.84 ± 01.15 | 0.84 ± 01.15 | 0.84 ± 01.15 | 0.84 ± 01.15 | 0.84 ± 01.15 | 0.426 |
| Logistic Regression | Random Forest | 2 | 0.81 ± 01.13 | 0.81 ± 01.13 | 0.81 ± 01.13 | 0.81 ± 01.13 | 0.81 ± 01.13 | 0.532 |
| LDA | Random Forest | 9 | 0.78 ± 01.11 | 0.78 ± 01.11 | 0.78 ± 01.11 | 0.78 ± 01.11 | 0.78 ± 01.11 | 0.406 |
| AdaBoost | Random Forest | 3 | 0.68 ± 01.03 | 0.68 ± 01.03 | 0.68 ± 01.03 | 0.70 ± 01.05 | 0.68 ± 01.03 | 0.492 |
A detailed summary of the performance metric for the best performing combination.
| Top Combination of Classifier, | Class | Accuracy | Precision | Sensitivity | F1-Score | Specificity | Inference time (ms) |
|---|---|---|---|---|---|---|---|
| MLP Classifier | Class 1 | 0.91 ± 02.49 | 0.96 ± 02.56 | 0.95 ± 02.54 | 0.95 ± 02.55 | 0.90 ± 02.47 | 0.592 |
| Class 2 | 0.91 ± 03.10 | 0.86 ± 03.02 | 0.89 ± 03.07 | 0.88 ± 03.05 | 0.91 ± 03.11 | ||
| Class 3 | 0.91 ± 03.52 | 0.83 ± 03.36 | 0.86 ± 03.41 | 0.84 ± 03.38 | 0.92 ± 03.53 | ||
| Class 4 | 0.91 ± 03.06 | 0.80 ± 02.87 | 0.86 ± 02.98 | 0.83 ± 02.93 | 0.92 ± 03.07 | ||
| Class 5 | 0.91 ± 01.99 | 0.98 ± 02.07 | 0.93 ± 02.02 | 0.95 ± 02.04 | 0.90 ± 01.98 | ||
| Overall | 0.91 ± 01.19 | 0.91 ± 01.19 | 0.91 ± 01.19 | 0.91 ± 1.19 | 0.91 ± 01.19 |
Figure 6ROC plot for the top-performing feature combination from 10 classical ML classifiers.
Figure 7ROC plot for 10 individual features using MLP classifier (best performing model).
Comparison with similar studies.
| Studies | Reported Approach | Approach Results |
|---|---|---|
| Cruz et al. in [ | A shallow CNN model named DFTNet was developed to classify using thermogram images | 94.57% F1-score for 10 folds with an unconventional approach of taking 2 different classes in each fold and reporting the average of the 10 folds |
| Khandakar et al. [ | Transfer learning using MobileNetV2 and image enhancement to classify thermograms into control and diabetic | A comparatively shallow CNN model, MobilenetV2 achieved an F1 score of ∼95% for a two-feet thermogram image-based classification, and the AdaBoost Classifier used 10 features and achieved an F1 score of 97% |
| This study | MLP classifier using 2 features extracted from the thermogram | 91.18% F1-score for 5-fold cross-validation for 5 class-classification |