| Literature DB >> 22267930 |
M Hamedi1, Sh-Hussain Salleh, T S Tan, K Ismail, J Ali, C Dee-Uam, C Pavaganun, P P Yupapin.
Abstract
The authors present a new method of recognizing different human facial gestures through their neural activities and muscle movements, which can be used in machine-interfacing applications. Human-machine interface (HMI) technology utilizes human neural activities as input controllers for the machine. Recently, much work has been done on the specific application of facial electromyography (EMG)-based HMI, which have used limited and fixed numbers of facial gestures. In this work, a multipurpose interface is suggested that can support 2-11 control commands that can be applied to various HMI systems. The significance of this work is finding the most accurate facial gestures for any application with a maximum of eleven control commands. Eleven facial gesture EMGs are recorded from ten volunteers. Detected EMGs are passed through a band-pass filter and root mean square features are extracted. Various combinations of gestures with a different number of gestures in each group are made from the existing facial gestures. Finally, all combinations are trained and classified by a Fuzzy c-means classifier. In conclusion, combinations with the highest recognition accuracy in each group are chosen. An average accuracy >90% of chosen combinations proved their ability to be used as command controllers.Entities:
Keywords: electromyography; machine learning; muscle activity; neural activity; neural system
Mesh:
Year: 2011 PMID: 22267930 PMCID: PMC3260039 DOI: 10.2147/IJN.S26619
Source DB: PubMed Journal: Int J Nanomedicine ISSN: 1176-9114
Figure 1Procedure of designing the proposed interface. A shows subject preparation, site selection, and electrode placement. System setup and data acquisition, signal recording protocol from all participants, and signal filtering process are demonstrated in B. Data segmentation, windowing, and feature extraction are carried out in C. The use of threshold lines to find the active features, which are more suitable for training and classification, is shown in D. The multipurpose interface is materialized by making all possible combinations from all existing facial gestures (E). All obtained combinations are trained and classified as illustrated in F to find the most accurate group of facial gestures and a better distribution of them.
Figure 2All considered facial gestures in this work: (A) natural, (B) smiling, (C) smiling with right side, (D) smiling with left side, (E) open the mouth like saying “a” in “apple,” (F) clenching molar teeth, (G) pulling up the eyebrows, (H) closing both eyes, (I) closing right eye, (J) closing left eye, (K) frowning.
Facial gestures used in this work
| Gesture | Gesture name | Effective channel(s) | Effective muscle(s) | |
|---|---|---|---|---|
| Major | Minor | |||
| 1 | Smiling | 1, 3 | Zygomaticus major | Levator anguli oris |
| 2 | Smiling with right side (pulling up the right corner of lip) | 3 | Zygomaticus major | Levator anguli oris |
| 3 | Smiling with left side (pulling up the left corner of lip) | 2 | Zygomaticus major | Levator anguli oris |
| 4 | Rage (clenching molar teeth) | 1, 3 | Masseter | Zygomaticus major |
| 5 | Gesturing “NO” (pull up the eyebrows) | 2 | Frontalis, pars lateralis | Pars medialis, Levator palpebrae superioris |
| 6 | Opening the mouth like saying “a” in “apple” | 1, 3 | Pterygoids, digastric | Masseter |
| 7 | Closing both eyes | 1, 2, 3 | Orbicularis oculi, pars palpebralis, orbitalis | Frontalis, temporalis |
| 8 | Closing left eye | 1, 2 | Orbicularis oculi, pars palpebralis, orbitalis | Frontalis, temporalis |
| 9 | Closing right eye | 2, 3 | Orbicularis oculi, pars palpebralis, orbitalis | Frontalis, temporalis |
| 10 | Frowning | 2 | Corrugator supercilii, Depressor supercilii | Frontalis |
| 11 | Natural | 1, 2, 3 | – | – |
Figure 3Electrode and wire positions. gesture.
Figure 4Raw signal of gesturing “a” in apple.
Figure 5The bottom plot determines the root mean squares (RMS) of gesturing “a” in apple word and the threshold line and above plot indicates the active RMS.
Number of combinations in each group of facial gestures per participant
| Number of facial gestures in each combination | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | Total |
| Number of combination(s) | 45 | 120 | 210 | 252 | 210 | 120 | 45 | 10 | 1 | 1013 |
All combinations of nine facial gestures
| Combination | 1 | 2 | 3 | 4 | 5 |
| Facial gestures | 1, 2, 3, 4, 5, 6, 7, 8, 9 | 1, 2, 3, 4, 5, 6, 7, 8, 10 | 1, 2, 3, 4, 5, 6, 7, 9, 10 | 1, 2, 3, 4, 5, 6, 8, 9, 10 | 1, 2, 3, 4, 5, 7, 8, 9, 10 |
| Combination | 6 | 7 | 8 | 9 | 10 |
| Facial gestures | 1, 2, 3, 4, 6, 7, 8, 9, 10 | 1, 2, 3, 5, 6, 7, 8, 9, 10 | 1, 2, 4, 5, 6, 7, 8, 9, 10 | 1, 3, 4, 5, 6, 7, 8, 9, 10 | 2, 3, 4, 5, 6, 7, 8, 9, 10 |
Final classification results of all participants with different number of facial gestures
| Participants | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | Average | SD |
| Results (%) | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 0 |
| Participants | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | Average | SD |
| Results (%) | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 0 |
| Participants | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | Average | SD |
| Results (%) | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 0 |
| Participants | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | Average | SD |
| Results (%) | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 100 | 0 |
| Participants | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | Average | SD |
| Results (%) | 98.21 | 97.68 | 97.2 | 98.1 | 97.7 | 99.1 | 98.45 | 97.9 | 97.34 | 98.03 | 98.03 | 0.5515 |
| Participants | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | Average | SD |
| Results (%) | 97.24 | 96.39 | 96.85 | 97.24 | 96.98 | 98.43 | 97.62 | 96.81 | 96.92 | 97.14 | 97.16 | 0.5517 |
| Participants | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | Average | SD |
| Results (%) | 95.5 | 95.21 | 94.45 | 95.78 | 94.65 | 96.13 | 95.89 | 94.82 | 94.37 | 95.08 | 95.18 | 0.6217 |
| Participants | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | Average | SD |
| Results (%) | 93.1 | 93.74 | 91.95 | 92.23 | 92.7 | 94.76 | 93.15 | 92.37 | 92.87 | 93.07 | 93.02 | 0.8084 |
| Participants | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | Average | SD |
| Results (%) | 90.63 | 88.45 | 89.52 | 90.11 | 86.71 | 91.02 | 89.15 | 88.97 | 87.83 | 91.27 | 90.41 | 3.1270 |
Abbreviation: SD, standard deviation.
Figure 6Data distribution of the best combinations with different numbers of facial gestures for the first participant.
All combination results for first participant
| Combination | Results | Combination | Results |
|---|---|---|---|
| 123456789 | 88.42% | 1234678910 | 81.56% |
| 1234567810 | 90.04% | 1235678910 | 93.10% |
| 1234567910 | 76.30% | 1245678910 | 78.07% |
| 1234568910 | 83.32% | 1345678910 | 76.98% |
| 1234578910 | 78.19% | 2345678910 | 84.60% |
The average of all results of all participants
| Number of facial gestures in each combination | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
| Average of the best combinations | 100% | 100% | 100% | 100% | 98.03% | 97.16% | 95.18% | 93.02% | 90.41% |
Comparison of research in this field of study
| Reference | Participants | Classes | Feature | Classifier | Results | Application |
|---|---|---|---|---|---|---|
| Present work | 10 | 11 | RMS | FCM | 90.41% | Interface design |
| Moon et al | 1 | 5 | MAV | Thresholding | 93% | Intelligent robotic wheelchair |
| Ang et al | 2 | 3 | MSD, RMS, PDS | Minimum Distance | 94.44% | Interface design via three expressions |
| Ferreira et al | 8 | 2 | – | Thresholding | 95.71% | Mobile robot |
| Arjunan et al | – | 3 | MRMS | Linear separation | 100% | Unvoiced speech interface |
| Firoozabadi et al | 3 | 5 | MAV | SVM | 100% | Control a virtual wheelchair |
| Guillaume et al36 | 1 | 6 | Absolute value | Gaussian models | 92% | Enhancement of HCI via facial EMGs |
| Rezazadeh et al | 10 | 5 | RMS | SFCM | 93.2% | Virtual brain control |
| Egon et al | 21 | 4 | Mean | AD, SD, VAR K-NN, SVM, MLP | 61%, 60.71%, 56.19% | Facial EMG-based interface used in MMI |
Abbreviations: AD, absolute deviation; EMGs, electromyograms; FCM, fuzzy C-mean; HCI, human–computer interface; K-NN, K-nearest neighbors; MAV, mean absolute value; MLP, multilayer perception; MMI, man–machine interface; MRMS, moving root mean square; MSD, maximum scatter difference; PDS, power spectrum density; RMS, root mean squares; SD, standard deviation; SFCM, subtractive fuzzy C-mean; SVM, support vector machine; VAR, variance.