| Literature DB >> 35206540 |
Susrutha Babu Sukhavasi1, Suparshya Babu Sukhavasi1, Khaled Elleithy1, Ahmed El-Sayed1, Abdelrahman Elleithy2.
Abstract
Monitoring drivers' emotions is the key aspect of designing advanced driver assistance systems (ADAS) in intelligent vehicles. To ensure safety and track the possibility of vehicles' road accidents, emotional monitoring will play a key role in justifying the mental status of the driver while driving the vehicle. However, the pose variations, illumination conditions, and occlusions are the factors that affect the detection of driver emotions from proper monitoring. To overcome these challenges, two novel approaches using machine learning methods and deep neural networks are proposed to monitor various drivers' expressions in different pose variations, illuminations, and occlusions. We obtained the remarkable accuracy of 93.41%, 83.68%, 98.47%, and 98.18% for CK+, FER 2013, KDEF, and KMU-FED datasets, respectively, for the first approach and improved accuracy of 96.15%, 84.58%, 99.18%, and 99.09% for CK+, FER 2013, KDEF, and KMU-FED datasets respectively in the second approach, compared to the existing state-of-the-art methods.Entities:
Keywords: DeepNet; K.L.T.; MTCNN; advanced driver assistance systems (ADAS); deep neural networks; driver emotion detection; face detection; facial expression recognition; machine learning
Mesh:
Year: 2022 PMID: 35206540 PMCID: PMC8871818 DOI: 10.3390/ijerph19042352
Source DB: PubMed Journal: Int J Environ Res Public Health ISSN: 1660-4601 Impact factor: 3.390
Figure 1Proposed deep network approaches process flow for driver emotion detection.
Figure 2Gaussian distribution.
Figure 3Haar feature face detection.
Figure 4Cascade classifier.
Figure 5Proposed deep neural network.
Figure 6Image pyramid.
Figure 7The first stage in multi-task cascaded convolutional neural networks.
Figure 8The second stage in multi-task cascaded convolutional neural networks.
Figure 9The third stage in multi-task cascaded convolutional neural networks.
Parameter settings used to train our deep network approaches on all four databases.
| Databases | Parameters | First Approach-Values | Second Approach-Values |
|---|---|---|---|
| Image Size | 256 × 256 | 256 × 256 | |
| Optimizer | Stochastic Gradient Descent (S.G.D.) | Adam | |
| CK+ | Loss Function | Cross-Entropy | Cross-Entropy |
| FER 2013 | Activation Function | ReLU | ReLU |
| KDEF | Batch Size | 128 | 128 |
| KMU-FED | Learning Rate | 0.01 | 0.001 |
| Epochs | 100 | 100 | |
| Momentum | 0.9 | 0.9 | |
| Validation Frequency | 30 | 30 |
Figure 10Sample images from CK+ database.
Figure 11Sample images from FER 2013 database.
Figure 12Sample images from KMU-FED database.
Comparison of proposed approaches with the state-of-the-art methods on CK+ database.
| Comparison Methods | Accuracy (%) |
|---|---|
| DNN [ | 93.2 |
| Inception-Resnet and LSTM [ | 93.2 |
| Single-WRF [ | 92.2 |
| Hierarchical W.R.F. with Normal Information Gain [ | 90.9 |
| Hierarchical W.R.F. with Data Similarity [ | 92.6 |
| DCMA-CNN [ | 93.4 |
| LMRF [ | 93.4 |
| First Proposed Approach | 93.4 |
| Second Proposed Approach | 96.1 |
Performance accuracies of different methods adapted from different papers.
Comparison of proposed approaches with the state-of-the-art methods on FER 2013 database.
| Comparison Methods | Accuracy (%) |
|---|---|
| D.N.N. [ | 66.4 |
| CNN-MNF [ | 70.3 |
| Simple CNN Model [ | 65.7 |
| eXnet [ | 73.5 |
| eXnet-Resnet [ | 71.1 |
| eXnet-DeXpression [ | 68.0 |
| Deep-Emotion [ | 70.0 |
| First Proposed Approach | 83.6 |
| Second Proposed Approach | 84.5 |
Performance accuracies of different methods adapted from different papers.
Comparison of proposed approaches with the state-of-the-art methods on KDEF database.
| Comparison Methods | Accuracy (%) |
|---|---|
| TLCNN [ | 86.4 |
| TLCNN-FOS [ | 88.2 |
| MPCNN [ | 86.9 |
| DSCAE-CNN [ | 95.5 |
| STL + GRADIENT + LAPLACIAN RTCNN [ | 88.1 |
| DL-FER [ | 96.6 |
| RBFNN [ | 88.8 |
| First Proposed Approach | 98.4 |
| Second Proposed Approach | 99.1 |
Performance accuracies of different methods adapted from different papers.
Comparison of proposed approaches with the state-of-the-art methods on KMU-FED database.
| Comparison Methods | Accuracy (%) |
|---|---|
| Facial Landmarks + WRF [ | 94.0 |
| CNN [ | 97.3 |
| SqueezeNet [ | 89.7 |
| MobileNetV2 [ | 93.8 |
| MobileNetV3 [ | 94.9 |
| LMRF [ | 95.1 |
| VGG16 [ | 94.2 |
| First Proposed Approach | 98.1 |
| Second Proposed Approach | 99.0 |
Performance accuracies of different methods adapted from different papers.
Figure 13Confusion matrices with accuracy (%) of the second proposed approach using different databases (a) CK+ database, (b) F.E.R. 2013 database, (c) KDEF database, and (d) KMU-FED database.