| Literature DB >> 36105626 |
Abstract
With the current health crisis caused by the COVID-19 pandemic, patients have become more anxious about infection, so they prefer not to have direct contact with doctors or clinicians. Lately, medical scientists have confirmed that several diseases exhibit corresponding specific features on the face the face. Recent studies have indicated that computer-aided facial diagnosis can be a promising tool for the automatic diagnosis and screening of diseases from facial images. However, few of these studies used deep learning (DL) techniques. Most of them focused on detecting a single disease, using handcrafted feature extraction methods and conventional machine learning techniques based on individual classifiers trained on small and private datasets using images taken from a controlled environment. This study proposes a novel computer-aided facial diagnosis system called FaceDisNet that uses a new public dataset based on images taken from an unconstrained environment and could be employed for forthcoming comparisons. It detects single and multiple diseases. FaceDisNet is constructed by integrating several spatial deep features from convolutional neural networks of various architectures. It does not depend only on spatial features but also extracts spatial-spectral features. FaceDisNet searches for the fused spatial-spectral feature set that has the greatest impact on the classification. It employs two feature selection techniques to reduce the large dimension of features resulting from feature fusion. Finally, it builds an ensemble classifier based on stacking to perform classification. The performance of FaceDisNet verifies its ability to diagnose single and multiple diseases. FaceDisNet achieved a maximum accuracy of 98.57% and 98% after the ensemble classification and feature selection steps for binary and multiclass classification categories. These results prove that FaceDisNet is a reliable tool and could be employed to avoid the difficulties and complications of manual diagnosis. Also, it can help physicians achieve accurate diagnoses without the need for physical contact with the patients.Entities:
Keywords: Computer-aided facial diagnosis; deep learning; discrete cosine transform; ensemble classification; feature selection; stacking; transfer learning
Year: 2022 PMID: 36105626 PMCID: PMC9465585 DOI: 10.1177/20552076221124432
Source DB: PubMed Journal: Digit Health ISSN: 2055-2076
A summary of related automated systems for facial disease diagnosis along with their limitations.
| Article | Dataset | Abnormality | Method | Results | Limitation |
|---|---|---|---|---|---|
| Kong et al., 2018
| 1123 Patients (private) | Acromegaly |
Open CV for face detection. Facial locations landmarks as features. Fronatization. SVM, LR, K-NN, CNN, and RF ensemble classifiers. | Precision = 96% |
Detect only one type of disease (binary classification). Used manual segmentation. Detect the face for diagnosis (controlled face diagnosis). Utilized only Facial location landmarks as features. Used only spatial features. Used only handcrafted features. Did not use DL features. |
| Schneider et al.
| 117 Patients (private) | Acromegaly |
Geometric and Gabor filter feature extraction methods. FIDA (facial image diagnostic aid) software | Accuracy = 81.9% |
Detect only one type of disease (binary classification). Used only handcrafted features. Did not use DL features. Controlled face diagnosis. Very small dataset. Private dataset. Low accuracy. Used commercial software to perform diagnosis. Large feature space. |
| Meng et al.
| 124 Patients (private) | Acromegaly |
35 Anatomical facial landmarks, 55 Angular, index, and linear features, Geometric features. LDA classifier | Accuracy = 92.86% |
Detect only one type of disease (binary classification). Used only handcrafted features. Did not use DL features. Controlled face diagnosis. Very small dataset. Private dataset. Large feature space. |
| Zhao et al.
| 48 Patients (private) | Down syndrome |
Geometric, contourlet transform, and LBP features. SVM classifier | Accuracy = 97.9% |
Detect only one type of disease (binary classification). Used only handcrafted features. Did not use DL features. Controlled face diagnosis. Very small dataset. Private dataset. Large feature space. |
| Zhao et al.
| 100 Patients (private) | Down syndrome |
HCLM, geometric, and textural features ICA SVM classifier | Accuracy = 95.6% |
Detect only one type of disease (binary classification). Used only handcrafted features. Did not use DL features. Controlled face diagnosis. Very small dataset. Private dataset. |
| Zhao et al.
| 130 | Down syndrome |
HCLM, geometric, Gabor, and LBP features ICA LDA classifier | Accuracy = 96.7% |
Detect only one type of disease (binary classification). Used only handcrafted features. Did not use DL features. Controlled face diagnosis. Very small dataset. Private dataset. |
| 24 Patients (private) | 14 Dysmorphic |
HCLM, geometric, Gabor, and LBP features ICA SVM classifier | Accuracy = 97% |
Used only handcrafted features. Did not use DL features. Controlled face diagnosis. Very small dataset. Private dataset. Large feature space. | |
| Lui et al.
| 87 Images (private) | Autism |
K means and histogram features. SVM classifier | Accuracy = 88.51% |
Detect only one type of disease (binary classification). Used only handcrafted features. Used only spatial features. Did not use DL features. Controlled face diagnosis. Very small dataset. Private dataset. Relatively low accuracy. |
| Sajid et al.
| 2000 Images | Palsy |
GAN for augmentation. VGG-16 and CNN for feature extraction. CNN for classification | Accuracy = 92.6% |
Grade only one type of disease. Used only spatial features. Controlled face diagnosis. Used individual feature extraction to perform classification. Utilized individual classifiers to perform classification. Large feature space. |
| Guo et al.
| 1840 Images (private) | UPFP |
Dlib DAN | AUC = 60.66% |
Detect only one type of disease (binary classification). Used only spatial features. Controlled face diagnosis. Private dataset. Low performance. Large feature space. |
| Kuan Wang and Jiebo Luo
| 8509 Images | 20 Diseases |
Manually segmented symptoms. Used binary features Employed color features, Hough transform, k means clustering | Accuracy = 80.2% |
Used manual segmentation to crop the abnormality. Used only handcrafted features. Did not use DL features. Controlled face diagnosis. Very small dataset. Relatively low performance. Unbalanced dataset. Large feature space. |
| Gurovich et al.
| 26,692 Images (private) | 216 Syndromes |
DeepGestalt model (holistic + local + CNN features extraction methods) CNN for classification | Top-10-accuracy = 91% |
Utilized individual classifiers to perform classification. Public only for health professionals. Private dataset. |
| Pantel et al.
| 646 Images | 17 Syndromes and normal faces (binary classification) |
DeepGestalt model | AUC = 89% |
Utilized individual classifiers to perform classification. Public only for health professionals. Private dataset. Performed binary classification. |
| Jin et al.
| 350 Images | 4 Diseases |
Open CV (HOG + SVM) AlexNet, ResNet-50, VGG-16 SVM | Accuracy = 93.3% |
Used only spatial features. Utilized individual spatial DL features. Utilized individual classifiers. Employed a large number of features to perform classification. |
| Beta-thalassemia | Accuracy = 95% |
Note. SVM: support vector machine; LR: logistic regression; K-NN: k-nearest neighbor; CNN: convolutional neural networks; LDA: linear discriminate analysis; HCLM: Hierarchical Constrained Local Model; ICA: independent component analysis; LBP: local binary pattern; GAN: generative adversarial network; DAN: deep alignment network.
Figure 1.A block diagram explaining the six stages of FaceDisNet.
Figure 2.The classification accuracy of the binary class classification category of the three classifiers trained with the spatial features obtained from the four CNNs. CNNs: convolutional neural networks.
Figure 3.The classification accuracy of the multiclass classification category of the three classifiers trained with the spatial features obtained from the four CNNs. CNNs: convolutional neural networks.
The accuracy (%) achieved for the three classifiers trained with fused spatial-spectral features of the binary class classification category.
| Feature set | SVM | DT | NB |
|---|---|---|---|
|
| |||
| DenseNet + Inception | 95 | 84.28 | 83.57 |
| DenseNet + ResNet-50 | 97.14 | 90.71 | 90.71 |
| DenseNet + ResNet-101 | 95 | 84.28 | 84.28 |
| ResNet-50 + ResNet-101 | 98.57 | 87.14 | 87.14 |
| ResNet-50 + Inception | 95 | 93.57 | 93.57 |
| Inception + ResNet-101 | 97.86 | 86.43 | 86.43 |
|
| |||
| DenseNet + Inception + ResNet-50 | 97.58 | 93.57 | 93.57 |
| DenseNet + Inception + ResNet-101 | 96.43 | 88.57 | 88.57 |
| DenseNet + ResNet-50 + ResNet-101 | 96.43 | 90.71 | 90.71 |
| Inception + ResNet-50 + ResNet-101 | 97.14 | 90.71 | 90.71 |
|
| |||
| DenseNet + Inception + ResNet-50 + ResNet-101 | 96.43 | 92.14 | 92.14 |
The accuracy (%) achieved for the three classifiers trained with fused spatial-spectral features of the multiclass classification category.
| Feature set | SVM | DT | NB |
|---|---|---|---|
|
| |||
| DenseNet + Inception | 94.86 | 93.15 | 93.14 |
| DenseNet + ResNet-50 | 94.57 | 93.43 | 93.43 |
| DenseNet + ResNet-101 | 95.14 | 91.7 | 91.7 |
| ResNet-50 + ResNet-101 | 94 | 89.14 | 88.86 |
| ResNet-50 + Inception | 96.29 | 92 | 92 |
| Inception + ResNet-101 | 94.23 | 89.7 | 89.7 |
|
| |||
| DenseNet + Inception + ResNet-50 | 96.86 | 94 | 94 |
| DenseNet + Inception + ResNet-101 | 96.29 | 93.43 | 93.43 |
| DenseNet + ResNet-50 + ResNet-101 | 96.57 | 93.71 | 93.71 |
| Inception + ResNet-50 + ResNet-101 | 95.71 | 91.4 | 91.4 |
|
| |||
| DenseNet + Inception + ResNet-50 + ResNet-101 | 96.29 | 94 | 94 |
Figure 4.The binary class classification accuracy of the three classifiers trained with spatial-spectral features of ResNet-50 + ResNet-101 versus the number of features selected using the CFS method. CFS: correlation-based feature selection.
Figure 5.The binary class classification accuracy of the three classifiers trained with spatial-spectral features of ResNet-50 + ResNet-101 versus the number of features selected using the RF method.
Figure 6.The binary class classification accuracy of the three classifiers trained with spatial-spectral features of ResNet-50 + ResNet-101 before and after the two feature selection methods.
Figure 7.The size of features of the spatial-spectral features of ResNet-50 + ResNet-101 before and after the two feature selection methods for the SVM classifier of binary class classification category. SVM: support vector machine.
Figure 8.The multi-class classification accuracy of the three classifiers trained with spatial-spectral features of DenseNet + Inception + ResNet-50 versus the number of features selected using the CFS method. CFS: correlation-based feature selection.
Figure 9.The multi-class classification accuracy of the three classifiers trained with spatial-spectral features of DenseNet + Inception + ResNet-50 versus the number of features selected using the RF method.
Figure 10.The multi-class classification accuracy of the three classifiers trained with spatial-spectral features of DenseNet + Inception + ResNet-50 before and after the two feature selection methods.
Figure 11.The size of features of the spatial-spectral features of denseNet + inception + resNet-50 before and after the two feature selection methods for the SVM classifier of the multiclass classification category. SVM: support vector machine.
Figure 12.The binary class classification accuracy of the stacking ensemble method and the three classifiers trained with spatial-spectral features of ResNet-50 + ResNet-101 after the two feature selection methods.
Figure 13.The multi-class classification accuracy of the stacking ensemble method and the three classifiers trained with spatial spectral features of DenseNet + Inception + ResNet-50 after the two feature selection methods.
The performance metrics achieved for the stacking of the ensemble classifier trained with fused spatial-spectral features of schema III after feature selection for the binary and multiclass classification categories.
| Feature selection approach | Sensitivity | Specificity | Precision | F1-Score | MCC |
|---|---|---|---|---|---|
|
| |||||
| CFS | 0.986 | 0.996 | 0.986 | 0.986 | 0.972 |
| RF | 0.986 | 0.996 | 0.986 | 0.986 | 0.972 |
|
| |||||
| CFS | 0.98 | 0.995 | 0.98 | 0.98 | 0.975 |
| RF | 0.977 | 0.994 | 0.977 | 0.977 | 0.972 |
A comparison between the performance of FaceDisNet and a related computer-aided facial diagnosis system based on the same dataset.
| Model | Sensitivity | Specificity | Precision | F1-Score | Accuracy |
|---|---|---|---|---|---|
|
| |||||
| ResNet-50 + SVM
| 0.9 | 0.932 | 0.931 | 0.915 | 91.7% |
| VGG-16 + SVM
| 1 | 0.9 | 0.909 | 0.952 | 95% |
| FaceDisNet | 0.986 | 0.996 | 0.986 | 0.986 | 98.57% |
|
| |||||
| ResNet-50 + SVM
| - | - | - | - | 92.7% |
| VGG-16 + SVM
| - | - | - | - | 93.3% |
| FaceDisNet | 0.98 | 0.995 | 0.98 | 0.98 | 98% |
A comparison between the accuracy (%) of FaceDisNet and the state-of-the-art DL models based on the same dataset.
| Model | Binary class | Multi-class |
|---|---|---|
| Inception-V3
| 78.57 | 80 |
| DenseNet-201
| 85.71 | 80 |
| ResNet-50
| 85.71 | 81 |
| ResNet-101
| 90.48 | 81 |
| FaceDisNet | 98.57 | 98 |