| Literature DB >> 32867182 |
Vianney Perez-Gomez1, Homero V Rios-Figueroa1, Ericka Janet Rechy-Ramirez1, Efrén Mezura-Montes1, Antonio Marin-Hernandez1.
Abstract
An essential aspect in the interaction between people and computers is the recognition of facial expressions. A key issue in this process is to select relevant features to classify facial expressions accurately. This study examines the selection of optimal geometric features to classify six basic facial expressions: happiness, sadness, surprise, fear, anger, and disgust. Inspired by the Facial Action Coding System (FACS) and the Moving Picture Experts Group 4th standard (MPEG-4), an initial set of 89 features was proposed. These features are normalized distances and angles in 2D and 3D computed from 22 facial landmarks. To select a minimum set of features with the maximum classification accuracy, two selection methods and four classifiers were tested. The first selection method, principal component analysis (PCA), obtained 39 features. The second selection method, a genetic algorithm (GA), obtained 47 features. The experiments ran on the Bosphorus and UIVBFED data sets with 86.62% and 93.92% median accuracy, respectively. Our main finding is that the reduced feature set obtained by the GA is the smallest in comparison with other methods of comparable accuracy. This has implications in reducing the time of recognition.Entities:
Keywords: facial expression recognition; facial geometric features; feature selection
Mesh:
Year: 2020 PMID: 32867182 PMCID: PMC7506644 DOI: 10.3390/s20174847
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Methodology.
3D facial expressions datasets.
| Database | Samples | Subject | Content | Temporality |
|---|---|---|---|---|
| Bosphorus [ | 4652 | 105 | Poses with different occlusion conditions | Static |
| and the six basic expressions and the neutral state | ||||
| UIBVFED: Virtual facial | 640 | 20 | 32 expressions | Static |
| expression dataset [ |
Number of instances per facial expression.
| Facial Expression | Instances |
|---|---|
| Surprise | 63 |
| Sadness | 66 |
| Happiness | 99 |
| Fear | 62 |
| Disgust | 64 |
| Anger | 70 |
| Total | 424 |
Figure 2Landmarks in UIBVFED.
Figure 3Twenty-two landmarks adapted in the UIBVFED database.
Figure 4Measurements that define Facial Animation Parameter Units (FAPUs) based on the work in [28].
Description of the six basic facial expressions.
| Facial | Eyebrows | Inner Eyebrows | Eyes | Mouth | Jaw |
|---|---|---|---|---|---|
| Anger | - | Pulled downward | Wide open | Lips are pressed against | - |
| Disgust | Relaxed | - | Eyelids: relaxed | Upper lip: raised | - |
| Fear | Raised and | Bent upward | Tense and alert | - | |
| Happiness | Relaxed | - | - | Open. Corners | - |
| Sadness | - | Bent upward | Slightly closed | Relaxed | - |
| Surprise | Raised | - | Upper eyelids: | - | Open |
Figure 5Twenty-two landmarks used in feature extraction.
3D and 2D original features.
| 3D Features | 2D Features | ||||||||
|---|---|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
|
|
|
| 1 | 1D3 |
| 1A3 |
| 1 | 1D2 |
| 1A2 |
|
| 2 | 2D3 |
| 2A3 |
| 2 | 2D2 |
| 2A2 |
|
| 3 | 3D3 |
| 3A3 |
| 3 | 3D2 |
| 3A2 |
|
| 4 | 4D3 |
| 4A3 |
| 4 | 4D2 |
| 4A2 |
|
| 5 | 5D3 |
| 5A3 |
| 5 | 5D2 |
| 5A2 |
|
| 6 | 6D3 |
| 6A3 |
| 6 | 6D2 |
| 6A2 |
|
| 7 | 7D3 |
| 7A3 |
| 7 | 7D2 |
| 7A2 |
|
| 8 | 8D3 |
| 8A3 |
| 8 | 8D2 |
| 8A2 |
|
| 9 | 9D3 |
| 9A3 |
| 9 | 9D2 |
| 9A2 |
|
| 10 | 10D3 |
| 10A3 |
| 10 | 10D2 |
| 10A2 |
|
| 11 | 11D3 |
| 11A3 |
| 11 | 11D2 |
| 11A2 |
|
| 12 | 12D3 |
| 12A3 |
| 12 | 12D2 |
| 12A2 |
|
| 13 | 13D3 |
| 13A3 |
| 13 | 13D2 |
| 13A2 |
|
| 14 | 14D3 |
| 14A3 |
| 14 | 14D2 |
| 14A2 |
|
| 15 | 15D3 |
| 15A3 |
| 15 | 15D2 |
| 15A2 |
|
| 16 | 16D3 |
| 16A3 |
| 16 | 16D2 |
| 16A2 |
|
| 17 | 17D3 |
| 17A3 |
| 17 | 17D2 |
| 17A2 |
|
| 18 | 18D3 |
| 18A3 |
| 18 | 18D2 |
| 18A2 |
|
| 19 | 19D3 |
| 19A3 |
| 19 | 19D2 |
| 19A2 |
|
| 20 | 20A3 |
| 20 | 20D2 |
| 20A2 |
| ||
| 21 | 21A3 |
| 21 | 21D2 |
| ||||
| 22 | 22A3 |
| 22 | 22D2 |
| ||||
| 23 | 23A3 |
| 23 | 23D2 |
| ||||
| 24 | 24A3 |
| 24 | ||||||
| 25 | 25A3 |
| 25 | ||||||
| 26 | 26A3 |
| 26 | ||||||
| 27 | 27A3 |
| 27 | ||||||
Figure 6GA feature set solution representation.
Accuracy of classification (original feature set).
| Measure | Classifier | |||
|---|---|---|---|---|
| SVM3 | SVM2 | kNN | E-KNN | |
| Standard deviation | 0.50 | 0.38 | 0.55 | 0.48 |
| Median accuracy | 85.25 | 83.17 | 83.01 | 84.61 |
| Mean accuracy | 85.11 | 83.17 | 83.07 | 84.65 |
| Maximum accuracy | 85.73 | 83.81 | 84.29 | 85.41 |
| Minimum accuracy | 84.29 | 82.53 | 82.21 | 83.81 |
Confusion matrix: 89 features.
| % | SU | SA | HA | FE | DI | AN |
|---|---|---|---|---|---|---|
|
|
| 0 | 1 | 18 | 2 | 0 |
|
| 0 |
| 0 | 1 | 3 | 6 |
|
| 0 | 0 |
| 1 | 2 | 0 |
|
| 12 | 3 | 0 |
| 7 | 1 |
|
| 0 | 5 | 1 | 4 |
| 7 |
|
| 0 | 12 | 0 | 1 | 4 |
|
Feature selection using principal component analysis (PCA).
| Reduced Features Using PCA | |||
|---|---|---|---|
| % Variance | 97% | 98% | 99% |
| Accuracy | 75.48% | 77.24% | 81.25% |
| Features | 21 | 27 | 39 |
SVM3 (reduced feature set using PCA).
| Measure | |
|---|---|
| Standard deviation | 1.03 |
| Median accuracy | 81.08 |
| Mean accuracy | 81.20 |
| Maximum accuracy | 82.85 |
| Minimum accuracy | 79.16 |
Confusion matrix: 39 features.
| % | SU | SA | HA | FE | DI | AN |
|---|---|---|---|---|---|---|
|
|
| 0 | 0 | 15 | 4 | 0 |
|
| 0 |
| 0 | 1 | 9 | 8 |
|
| 0 | 0 |
| 1 | 8 | 0 |
|
| 15 | 4 | 0 |
| 5 | 1 |
|
| 2 | 11 | 2 | 4 |
| 6 |
|
| 0 | 12 | 0 | 2 | 5 |
|
Description of features and accuracy for the best and the worst fits.
| 3D Features | 2D Features | Number of | Average | |||||
|---|---|---|---|---|---|---|---|---|
|
| Dist | Total |
| Dist | Total | |||
| Worst fit | 12 | 9 | 21 | 12 | 13 | 25 | 46 | 87.82% |
| Best fit | 14 | 12 | 26 | 8 | 13 | 21 | 47 | 89.58% |
Figure 7Convergence graph.
3D and 2D features obtained through a genetic algorithm (GA).
| 3D Features | 2D Features | ||||||||
|---|---|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
|
|
|
| 1 | 2D3 |
| 2A3 |
| 1 | 1D2 |
| 5A2 |
|
| 2 | 4D3 |
| 3A3 |
| 2 | 2D2 |
| 9A2 |
|
| 3 | 6D3 |
| 4A3 |
| 3 | 4D2 |
| 10A2 |
|
| 4 | 8D3 |
| 5A3 |
| 4 | 5D2 |
| 12A2 |
|
| 5 | 9D3 |
| 7A3 |
| 5 | 7D2 |
| 14A2 |
|
| 6 | 10D3 |
| 11A3 |
| 6 | 10D2 |
| 15A2 |
|
| 7 | 11D3 |
| 12A3 |
| 7 | 12D2 |
| 18A2 |
|
| 8 | 12D3 |
| 15A3 |
| 8 | 14D2 |
| 20A2 |
|
| 9 | 13D3 |
| 19A3 |
| 9 | 16D2 |
| ||
| 10 | 14D3 |
| 21A3 |
| 10 | 17D2 |
| ||
| 11 | 16D3 |
| 23A3 |
| 11 | 19D2 |
| ||
| 12 | 18D3 |
| 24A3 |
| 12 | 20D2 |
| ||
| 13 | 26A3 |
| 13 | 21D2 |
| ||||
| 14 | 27A3 |
| 14 | ||||||
SVM3 (reduced feature set using GA).
| Measure | |
|---|---|
| Standard deviation | 0.73 |
| Median accuracy | 86.69 |
| Mean accuracy | 86.62 |
| Maximum accuracy | 87.17 |
| Minimum accuracy | 85.25 |
Confusion matrix: 47 features.
| % | SU | SA | HA | FE | DI | AN |
|---|---|---|---|---|---|---|
|
|
| 0 | 0 | 16 | 3 | 0 |
|
| 0 |
| 0 | 1 | 5 | 1 |
|
| 0 | 0 |
| 1 | 3 | 0 |
|
| 16 | 1 | 0 |
| 6 | 0 |
|
| 0 | 4 | 0 | 2 |
| 5 |
|
| 0 | 11 | 0 | 0 | 6 |
|
SVM3 (GA and database UIBVFED).
| Measure | |
|---|---|
| Standard deviation | 1.11 |
| Median accuracy | 93.75 |
| Mean accuracy | 93.92 |
| Maximum accuracy | 95.83 |
| Minimum accuracy | 92.50 |
Confusion matrix: 47 features employed UIBVFED database.
| % | SU | SA | HA | FE | DI | AN |
|---|---|---|---|---|---|---|
|
|
| 0 | 0 | 15 | 0 | 0 |
|
| 0 |
| 0 | 0 | 15 | 0 |
|
| 0 | 0 |
| 0 | 0 | 0 |
|
| 0 | 0 | 0 |
| 0 | 0 |
|
| 0 | 5 | 0 | 0 |
| 0 |
|
| 0 | 0 | 0 | 0 | 0 |
|
Mean of classification.
| Acuraccy | Features | Database | |
|---|---|---|---|
| Original feature set | 85.11% | 89 | Bosphorus |
| Best PCA | 81.2% | 39 | Bosphorus |
| GA | 86.62% | 47 | Bosphorus |
| GA | 93.92% | 47 | UIBVFED |
Comparison between the original and reduced feature sets.
| 3D Features | 2D Features | Time to Classify | ||||
|---|---|---|---|---|---|---|
|
| Dist |
| Dist | Total | New Instance (ms) | |
| Original feature set | 27 | 19 | 20 | 23 | 89 | 0.00046364 |
| Reduced feature set obtained via a GA | 14 | 12 | 8 | 13 | 47 | 0.00032031 |
| Reduction percentage | 48.15% | 36.84% | 60% | 43.48% | 47.2% | |
Figure 8Original 3D features.
Figure 93D reduced features obtained through a GA.
Figure 10Original 2D features.
Figure 112D reduced features obtained through a GA.
Comparison on Bosphorus Dataset with handcrafted features.
| Proposed GA | Salahshoor & Faez (2012) | Ujir (2013) | Zhang et al. (2015) | |
|---|---|---|---|---|
| Approach | Static | Static | Static | Dynamic |
| Classes | 6 | 6 | 6 | 6 |
| Feature | GA | None | mRMR | mRMR |
| Features | 47 | 21600 | 115 | 64 |
| Decision | SVM | Modified | Voting | Adaptive |
| Accuracy (%) | 86.62% | 85.36% | 66% | 92.2% |
Comparison on Bosphorus Dataset with deep features.
| Proposed GA | Li et al. (2017) | Kun Tin (2019) | |
|---|---|---|---|
| Data | 2D + 3D | 2D + 3D | 2D + 3D |
| Features | 47 | 32-D Deep Feature | Deep Feature Fusion |
| Classifier | SVM3 | SVM | CNN |
| Accuracy | 86.62% | 79.17% | 80.28% |
Comparison with other handcrafted feature method on different datasets.
| Proposed Median Accuracy | Proposed Median Accuracy | Goulart et al. (2019) | Oh & Kim (2020) | |
|---|---|---|---|---|
| Approach | Static | Static | Static | Dynamic |
| Data Base | Bosphorus | UIBVFED | Cohn–Kanade | Own |
| Classes | 6 | 6 | 7 | 5 |
| Feature | GA | GA | PCA + FNCA | Grid Map |
| Features | 47 | 47 | 60 | 2912 |
| Decision | SVM | SVM | SVM | ECOC-SVM |
| Accuracy (%) | 86.62% | 93.92% | 89.98% | 98.47% |