| Literature DB >> 35684825 |
Antonio A Aguileta1, Ramón F Brena2,3, Erik Molino-Minero-Re4, Carlos E Galván-Tejada5.
Abstract
Automatic identification of human facial expressions has many potential applications in today's connected world, from mental health monitoring to feedback for onscreen content or shop windows and sign-language prosodic identification. In this work we use visual information as input, namely, a dataset of face points delivered by a Kinect device. The most recent work on facial expression recognition uses Machine Learning techniques, to use a modular data-driven path of development instead of using human-invented ad hoc rules. In this paper, we present a Machine-Learning based method for automatic facial expression recognition that leverages information fusion architecture techniques from our previous work and soft voting. Our approach shows an average prediction performance clearly above the best state-of-the-art results for the dataset considered. These results provide further evidence of the usefulness of information fusion architectures rather than adopting the default ML approach of features aggregation.Entities:
Keywords: facial expressions; information fusion; machine learning
Mesh:
Year: 2022 PMID: 35684825 PMCID: PMC9185323 DOI: 10.3390/s22114206
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Figure 1Overview of the method that predicts GFEs.
F1 scores achieved by our proposed model and other state-of-art approaches using the dataset that stored grammatical facial expressions of the Brazilian sign language (Libras) made by subject one. Best results are in bold.
| GFE | Freitas | Bhuvan | Our Proposal |
|---|---|---|---|
| Wh question | 0.8942 | 0.945338 |
|
| Y/N question | 0.9412 | 0.940299 |
|
| Doubt question | 0.9607 | 0.898678 |
|
| Negative | 0.9582 | 0.910506 |
|
| Affirmative | 0.8773 | 0.890052 |
|
| Conditional | 0.9534 | 0.94964 |
|
| Relative | 0.9680 | 0.954064 |
|
| Topic | 0.9544 | 0.902439 |
|
| Focus | 0.9836 | 0.975 |
|
| Average | 0.9434 | 0.9296 |
|
Accuracy achieved by our proposed model and other state-of-art approaches using a dataset storing grammatical facial expressions of the Brazilian sign language (Libras) made by subject one. Best results are in bold.
| GFE | Gafar (FRFS-ACO and MLP) | Gafar (FRFS-ACO and FRNN) | Ours |
|---|---|---|---|
| Wh question | 0.9237 | 0.9447 |
|
| Y/N question | 0.9467 | 0.9438 |
|
| Doubt question | 0.9329 | 0.9077 |
|
| Negative | 0.9119 | 0.919 |
|
| Affirmative | 0.8635 | 0.8983 |
|
| Conditional | 0.9622 | 0.9701 |
|
| Relative | 0.9665 | 0.9656 |
|
| Topic | 0.9649 | 0.9532 |
|
| Focus | 0.9593 | 0.933 |
|
| Average | 0.936 | 0.9373 |
|
ROC-AUC scores achieved by our proposed model and other state-of-art approaches using a dataset that stores grammatical facial expressions of the Brazilian sign language (Libras) made by subject one. Best results are in bold.
| GFE | Bhuvan et al. | Our Proposal |
|---|---|---|
| Wh question | 0.9768 |
|
| Y/N question | 0.9925 |
|
| Doubt question | 0.9713 |
|
| Negative | 0.9695 |
|
| Affirmative | 0.9763 |
|
| Conditional | 0.9915 |
|
| Relative | 0.9946 |
|
| Topic | 0.9863 |
|
| Focus | 0.9948 |
|
| Average | 0.9837 |
|
F1 scores achieved by our proposed model and other state-of-art approaches using a dataset storing grammatical facial expressions of the Brazilian sign language (Libras) made by subject two. Best results are in bold.
| GFE | Freitas | Bhuvan | Ours |
|---|---|---|---|
| Wh question | 0.8988 | 0.938776 |
|
| Y/N question | 0.9129 | 0.90566 |
|
| Doubt question | 0.9700 | 0.911765 |
|
| Negative | 0.7269 | 0.905556 |
|
| Affirmative | 0.8641 | 0.854772 |
|
| Conditional | 0.8814 | 0.867384 |
|
| Relative | 0.9759 | 0.935252 |
|
| Topic | 0.9322 | 0.853448 |
|
| Focus | 0.9213 | 0.934959 |
|
| Average | 0.8982 | 0.9008 |
|
ROC-AUC scores achieved by our proposed model and other state-of-art approaches using a dataset storing grammatical facial expressions of the Brazilian sign language (Libras) made by subject two. Best results are in bold.
| GFE | Bhuvan et al. | Our Proposal |
|---|---|---|
| Wh question | 0.9872 |
|
| Y/N question | 0.9754 |
|
| Doubt question | 0.9697 |
|
| Negative | 0.9749 |
|
| Affirmative | 0.9485 |
|
| Conditional | 0.9691 |
|
| Relative | 0.9856 |
|
| Topic | 0.9732 |
|
| Focus | 0.9811 |
|
| Average | 0.9739 |
|
F1 scores achieved by our proposed model and other state-of-art approaches using a dataset storing grammatical facial expressions of the Brazilian sign language (Libras) made by subject one and subject two. Best results are in bold.
| GFE | Freitas | Bhuvan | Ours |
|---|---|---|---|
| Wh question | 0.8599 | 0.925125 |
|
| Y/N question | 0.8860 | 0.922591 |
|
| Doubt question | 0.9452 | 0.928896 |
|
| Negative | 0.7830 | 0.909091 |
|
| Affirmative | 0.8209 | 0.898734 |
|
| Conditional | 0.8776 | 0.927176 |
|
| Relative | 0.8973 | 0.946087 |
|
| Topic | 0.9164 | 0.874109 |
|
| Focus | 0.9321 | 0.932462 |
|
| Average | 0.8798 | 0.9183 |
|
ROC-AUC scores achieved by our proposed model and other state-of-art approaches using a dataset storing grammatical facial expressions of the Brazilian sign language (Libras) made by subject one and subject two. Best results are in bold.
| GFE | Uddin | Bhuvan | Acevedo | Ours |
|---|---|---|---|---|
| Wh question | 0.9853 | 0.9785 |
| 0.9995 |
| Y/N question |
| 0.9818 | 0.9594 | 0.9985 |
| Doubt question | 0.9833 | 0.9839 | 0.9500 |
|
| Negative |
| 0.9759 |
| 0.9999 |
| Affirmative |
| 0.9629 |
| 0.9989 |
| Conditional | 0.9866 | 0.9835 | 0.9915 |
|
| Relative | 0.9918 | 0.9935 |
| 0.9999 |
| Topic | 0.9770 | 0.9728 |
| 0.9999 |
| Focus | 0.9867 | 0.9874 |
| 0.9999 |
| Average | 0.9901 | 0.9800 | 0.9890 |
|
F1 scores achieved by our proposed model and other state-of-art approaches that train a RFC with the Libras GFEs made by the subject one and test with the Libras GFEs of the subject two. Best results are in bold.
| GFE | Freitas | Our Proposal |
|---|---|---|
| Wh question |
| 0.8409 |
| Y/N question |
| 0.8346 |
| Doubt question | 0.9052 |
|
| Negative |
| 0.6667 |
| Affirmative | 0.7478 |
|
| Conditional | 0.7704 |
|
| Relative | 0.8653 |
|
| Topic | 0.8953 |
|
| Focus | 0.9022 |
|
| Average | 0.8303 |
|