| Literature DB >> 30181532 |
Fotis P Kalaganis1,2, Elisavet Chatzilari3, Spiros Nikolopoulos3, Ioannis Kompatsiaris3, Nikos A Laskaris4,5.
Abstract
Gaze-based keyboards offer a flexible way for human-computer interaction in both disabled and able-bodied people. Besides their convenience, they still lead to error-prone human-computer interaction. Eye tracking devices may misinterpret user's gaze resulting in typesetting errors, especially when operated in fast mode. As a potential remedy, we present a novel error detection system that aggregates the decision from two distinct subsystems, each one dealing with disparate data streams. The first subsystem operates on gaze-related measurements and exploits the eye-transition pattern to flag a typo. The second, is a brain-computer interface that utilizes a neural response, known as Error-Related Potentials (ErrPs), which is inherently generated whenever the subject observes an erroneous action. Based on the experimental data gathered from 10 participants under a spontaneous typesetting scenario, we first demonstrate that ErrP-based Brain Computer Interfaces can be indeed useful in the context of gaze-based typesetting, despite the putative contamination of EEG activity from the eye-movement artefact. Then, we show that the performance of this subsystem can be further improved by considering also the error detection from the gaze-related subsystem. Finally, the proposed bimodal error detection system is shown to significantly reduce the typesetting time in a gaze-based keyboard.Entities:
Mesh:
Year: 2018 PMID: 30181532 PMCID: PMC6123473 DOI: 10.1038/s41598-018-31425-2
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1(Top) Schematic outline of the error-aware keyboard. A hybrid BCI system, relying on brain activity patterns and eye-movement features, detects and deletes characters that are mistyped. The depicted Machine Learning (ML) modules correspond to linear SVMs. (Bottom) Timeline describing the sequence of events during the typesetting experiment. Initially, the participant starts gazing at the desired letter. When he completes a 500 ms time-interval of continuous gazing, the key is registered and simultaneously the associated visual indication is presented. The physiological responses following this indication are used to detect typesetting errors. We note that the “eye” icon was not presented in the experiments and it is only shown here for presentation clarity purposes.
Figure 2Single-subject averaged brain activation traces for the correct (blue) and wrong (red) selections of buttons are shown in the top middle panel. Particular latencies are indicated on these traces (E: error; C: correct) and the corresponding topographies have been included in the top left/right panels. The traces shown in the bottom middle panel reflect eye-movement activity (derived by averaging correspondingly across the epochs of gaze-related signal). Zero time indicates the instant that the typing of the current letter has been completed and the eyes are free to move towards the next letter.
Figure 3(Left) Scatter-plot of gaze centre displacements (derived by integrating the derivatives of eye position coordinates within a time interval that includes the key registration at 0 latency). Each dot indicates the main direction of the eye after a correct typesetting of a single letter. The point swarm has been partitioned into 4 groups, and the membership of each dot is indicated by colour. The associated brain-signal and eye-movement activity traces have been grouped accordingly and their (sub)averages are indicated in top right and bottom right respectively using the colour code defined in scatter-plot.
Performance metrics for the classification task of discriminating between correct and erroneous typesetting based on EEG traces and gaze-movement patterns (used both separately and jointly).
| Subject ID | EEG | Eye Motion | Early Fusion | Late Fusion | ||||
|---|---|---|---|---|---|---|---|---|
| Sensitivity | Specificity | Sensitivity | Specificity | Sensitivity | Specificity | Sensitivity | Specificity | |
| S01 | 73.41 | 83.19 | 97.82 | 51.39 | 78.95 | 85.96 | 91.51 | 78.00 |
| S02 | 75.54 | 87.97 | 90.14 | 71.94 | 80.48 | 89.34 | 85.55 | 87.00 |
| S03 | 85.52 | 93.20 | 89.36 | 82.83 | 87.40 | 94.00 | 89.50 | 93.13 |
| S04 | 74.76 | 85.18 | 87.45 | 70.26 | 80.13 | 85.18 | 83.84 | 81.92 |
| S05 | 76.38 | 85.90 | 88.34 | 75.71 | 82.58 | 89.83 | 85.25 | 89.96 |
| S06 | 76.69 | 80.09 | 95.68 | 63.95 | 88.92 | 75.82 | 92.65 | 73.69 |
| S07 | 61.61 | 78.79 | 83.98 | 67.89 | 71.71 | 77.79 | 79.37 | 75.57 |
| S08 | 69.31 | 83.76 | 94.21 | 71.35 | 89.87 | 85.30 | 90.26 | 83.76 |
| S09 | 67.39 | 82.36 | 87.73 | 77.18 | 78.64 | 83.94 | 82.71 | 83.70 |
| S10 | 70.47 | 81.41 | 80.77 | 80.56 | 79.35 | 80.80 | 79.76 | 82.69 |
|
|
|
|
|
|
|
|
|
|
Tabulated are the results averaged from 100 repetitions of Monte-Carlo cross validation. Four implementation scenarios have been validated, for each subject independently.
Figure 4The grand average sensitivity and specificity values (a) along with the Utility gain (b), after 100 Monte-Carlo cross validation repetitions, with respect to threshold moving within the normalized SVM margins.
Columns 2–6: chance that the eye-tracker will interpret user’s intention falsely and utility gain for the four classification schemes.
| Subject ID | Typing Error Chance | Utility Gain | specificity | sensitivity | accuracy | |||
|---|---|---|---|---|---|---|---|---|
| EEG | eye motion | early fusion | late fusion | |||||
| S01 | 11.60% | 1.02 | 0.69 | 1.05 | 1.05 | 98.64 | 45.60 | 92.48 |
| S02 | 7.92% | 1.01 | 0.93 | 1.02 | 1.03 | 98.55 | 45.40 | 94.33 |
| S03 | 6.36% | 1.04 | 0.99 | 1.04 | 1.04 | 99.28 | 71.00 | 97.48 |
| S04 | 11.60% | 1.04 | 0.93 | 1.04 | 1.05 | 98.53 | 46.50 | 92.48 |
| S05 | 8.87% | 1.02 | 0.91 | 1.04 | 1.04 | 98.94 | 50.60 | 94.66 |
| S06 | 13.67% | 1.03 | 0.92 | 1.03 | 1.03 | 96.68 | 39.20 | 88.81 |
| S07 | 13.50% | 0.99 | 0.92 | 0.99 | 1.02 | 97.97 | 24.10 | 87.99 |
| S08 | 9.24% | 1.00 | 0.94 | 1.03 | 1.03 | 98.89 | 28.70 | 92.39 |
| S09 | 6.16% | 0.98 | 0.91 | 0.99 | 1.01 | 97.66 | 35.40 | 93.82 |
| S10 | 13.84% | 1.03 | 1.01 | 1.05 | 1.05 | 98.84 | 31.20 | 89.44 |
|
|
|
|
|
|
|
|
|
|
Columns 7–9: classification performance metrics at the optimal threshold for the late fusion classification scheme. Tabulated are the results averaged from 100 repetitions of Monte-Carlo cross validation.
Chance that the eye-tracker will interpret user’s intention falsely in the two typesetting procedures (T1 and T2), the average gain in time (accompanied by the respective percentage) that is obtained by the EDS system taking advantage of the late fusion classifier and the average time required to type one sentence in both tasks.
| Subject ID | Typing Error Chance | Gain (T2-T1) | Average Sentence Time | |||
|---|---|---|---|---|---|---|
| T1 | T2 | Time (seconds) | Percentage | T1 (seconds) | T2 (seconds) | |
| S01 | 11.60% | 6.48% | 1.23 | 4.51% | 26.07 | 27.30 |
| S02 | 7.92% | 4.14% | 4.17 | 13.99% | 25.64 | 29.81 |
| S03 | 6.36% | 8.74% | 7.33 | 22.96% | 24.60 | 31.93 |
| S04 | 11.60% | 8.44% | 7.53 | 21.68% | 27.21 | 34.74 |
| S05 | 8.87% | 9.05% | 4.92 | 16.12% | 25.61 | 30.53 |
| S06 | 13.67% | 3.56% | −0.78 | −2.90% | 27.72 | 26.94 |
| S07 | 13.50% | 5.83% | 0.44 | 1.52% | 28.51 | 28.95 |
| S08 | 9.24% | 3.36% | −0.11 | −0.42% | 26.20 | 26.09 |
| S09 | 6.16% | 2.55% | −1.38 | −5.41% | 26.88 | 25.50 |
| S10 | 13.84% | 9.50% | 3.74 | 11.74% | 28.11 | 31.85 |
|
|
|
|
|
|
|
|
The results are obtained according to a Leave-one-Sentence-out cross validation manner.