| Literature DB >> 24824369 |
Dario Cazzato1, Marco Leo2, Cosimo Distante3.
Abstract
This paper investigates the possibility of accurately detecting and tracking human gaze by using an unconstrained and noninvasive approach based on the head pose information extracted by an RGB-D device. The main advantages of the proposed solution are that it can operate in a totally unconstrained environment, it does not require any initial calibration and it can work in real-time. These features make it suitable for being used to assist human in everyday life (e.g., remote device control) or in specific actions (e.g., rehabilitation), and in general in all those applications where it is not possible to ask for user cooperation (e.g., when users with neurological impairments are involved). To evaluate gaze estimation accuracy, the proposed approach has been largely tested and results are then compared with the leading methods in the state of the art, which, in general, make use of strong constraints on the people movements, invasive/additional hardware and supervised pattern recognition modules. Experimental tests demonstrated that, in most cases, the errors in gaze estimation are comparable to the state of the art methods, although it works without additional constraints, calibration and supervised learning.Entities:
Mesh:
Year: 2014 PMID: 24824369 PMCID: PMC4063032 DOI: 10.3390/s140508363
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1.A block diagram of the proposed method.
Figure 2.Three different snapshots of the face tracking module.
Figure 3.A scheme of the gaze estimation solution.
Figure 4.The used grouping scheme for target points during tests.
Figure 5.A portion of the panel used in the experimental phase.
Figure 6.A picture of the monitor where gaze hits can be drawn.
Figure 7.The different users positions for the experimental phase.
Experiments with the first group of experienced persons: they knew how the system works and that already tried the system before the test session.
| P0 | 70 cm | 1.50 | 1.22 | 2.66 | 2.18 | 0.34 | 0.87 |
| 150 cm | 3.50 | 1.33 | 4.83 | 1.84 | 0.54 | 0.93 | |
| 250 cm | 6.00 | 1.37 | 8.50 | 1.94 | 0.67 | 0.85 | |
| P1 | 70 cm | n.a. | n.a. | n.a. | n.a. | n.a | n.a |
| 150 cm | 6.00 | 1.61 | 8.00 | 2.79 | 0.93 | 1.36 | |
| 250 cm | 2.60 | 0.52 | 4.00 | 0.88 | 0.20 | 0.37 | |
| P2 | 70 cm | 8.77 | 5.03 | 7.66 | 4.37 | 2.36 | 2.09 |
| 150 cm | 0.16 | 0.05 | 11.83 | 4.15 | 0.01 | 1.66 | |
| 250 cm | 4.33 | 0.95 | 5.83 | 1.29 | 0.38 | 0.52 | |
| P3 | 70 cm | 5.61 | 4.58 | 3.50 | 1.94 | 2.33 | 1.06 |
| 150 cm | 6.83 | 2.60 | 8.83 | 3.08 | 0.98 | 1.20 | |
| 250 cm | 4.66 | 1.06 | 1.83 | 0.40 | 0.44 | 1.50 | |
| P4 | 70 cm | n.a. | n.a. | n.a. | n.a. | n.a. | n.a. |
| 150 cm | 0.50 | 0.13 | 9.33 | 3.56 | 0.07 | 1.95 | |
| 250 cm | 0.66 | 0.13 | 15.66 | 3.47 | 0.49 | 1.97 | |
| P5 | 70 cm | 3.66 | 2.03 | 3.83 | 3.13 | 1.11 | 1.58 |
| 150 cm | 0.33 | 0.11 | 4.16 | 1.59 | 0.04 | 0.60 | |
| 250 cm | 4.33 | 0.95 | 8.16 | 1.87 | 0.36 | 0.97 | |
| Total Averages | 70 cm | 4.88 | 3.22 | 4.41 | 2.90 | 1.53 | 1.40 |
| 150 cm | 2.88 | 0.97 | 7.83 | 2.83 | 0.42 | 1.28 | |
| 250 cm | 3.77 | 0.83 | 7.25 | 1.64 | 0.42 | 1.03 | |
Experiments with the second group persons that were trying the system for the first time but that have been informed how it works.
| P0 | 70 cm | 2.50 | 2.04 | 2.33 | 1.90 | 0.53 | 0.57 |
| 150 cm | 6.24 | 2.38 | 7.41 | 2.83 | 0.95 | 1.09 | |
| 250 cm | 28.5 | 6.50 | 13 | 2.97 | 1.95 | 1.18 | |
| P1 | 70 cm | n.a. | n.a. | n.a. | n.a. | n.a. | n.a. |
| 150 cm | 2.00 | 0.53 | 21.58 | 7.70 | 0.30 | 3.61 | |
| 250 cm | 19.00 | 3.84 | 27.00 | 6.05 | 2.30 | 2.96 | |
| P2 | 70 cm | 10.16 | 5.89 | 17.16 | 10.40 | 2.59 | 5.30 |
| 150 cm | 5.83 | 1.89 | 19.08 | 6.78 | 0.92 | 3.45 | |
| 250 cm | 3.00 | 1.50 | 0.65 | 0.33 | 0.79 | 0.33 | |
| P3 | 70 cm | 2.83 | 2.31 | 8.33 | 4.77 | 1.01 | 2.85 |
| 150 cm | 15.83 | 6.02 | 13.33 | 4.69 | 3.99 | 2.83 | |
| 250 cm | 18.5 | 4.23 | 20.5 | 4.58 | 1.98 | 2.12 | |
| P4 | 70 cm | n.a. | n.a. | n.a. | n.a. | n.a. | n.a. |
| 150 cm | 4.75 | 1.27 | 15.08 | 5.74 | 0.77 | 2.91 | |
| 250 cm | 33.00 | 6.79 | 12.50 | 2.86 | 3.66 | 2.01 | |
| P5 | 70 cm | 11.16 | 0.83 | 6.51 | 0.68 | 0.36 | 0.29 |
| 150 cm | 7.50 | 2.53 | 3.00 | 1.14 | 1.10 | 0.59 | |
| 250 cm | 22.50 | 5.03 | 19.00 | 4.34 | 2.73 | 1.52 | |
| Total Averages | 70 cm | 6.66 | 4.19 | 7.16 | 4.44 | 1.12 | 2.25 |
| 150 cm | 6.98 | 2.44 | 13.25 | 4.81 | 1.33 | 2.41 | |
| 250 cm | 20.75 | 4.51 | 15.58 | 3.52 | 2.23 | 1.68 | |
Experiments with the third group of persons that were totally unaware of how the system works.
| P0 | 70 cm | 4.00 | 3.27 | 0.00 | 0.00 | 1.92 | 0.00 |
| 150 cm | 15.00 | 5.71 | 10.00 | 3.81 | 3.01 | 2.19 | |
| 250 cm | 71.00 | 12.00 | 15.85 | 2.74 | 8.86 | 2.03 | |
| P1 | 70 cm | n.a. | n.a. | n.a. | n.a. | n.a. | n.a. |
| 150 cm | 17.00 | 4.73 | 29.00 | 10.46 | 3.15 | 7.61 | |
| 250 cm | 62.00 | 13.15 | 34.00 | 7.64 | 7.96 | 4.19 | |
| P2 | 70 cm | 24.99 | 23.13 | 34.61 | 23.13 | 12.17 | 14.01 |
| 150 cm | 1.78 | 0.61 | 10.90 | 3.82 | 0.49 | 1.38 | |
| 250 cm | 27.00 | 6.05 | 17.00 | 3.79 | 3.42 | 1.83 | |
| P3 | 70 cm | 24.61 | 19.37 | 10.10 | 5.11 | 10.33 | 2.91 |
| 150 cm | 33.20 | 12.48 | 17.10 | 6.06 | 7.08 | 3.20 | |
| 250 cm | 90.23 | 19.84 | 12.6 | 2.80 | 10.05 | 1.18 | |
| P4 | 70 cm | n.a. | n.a. | n.a. | n.a. | n.a. | n.a. |
| 150 cm | 3.90 | 1.01 | 4.80 | 1.83 | 0.73 | 0.81 | |
| 250 cm | 12.54 | 2.52 | 24.67 | 5.63 | 1.89 | 3.21 | |
| P5 | 70 cm | 19.80 | 4.10 | 12.20 | 3.35 | 1.87 | 1.79 |
| 150 cm | 2.78 | 0.96 | 17.38 | 6.60 | 0.49 | 4.03 | |
| 250 cm | 4.44 | 0.98 | 24.24 | 5.53 | 0.41 | 3.31 | |
| Total Averages | 70 cm | 18.35 | 12.67 | 12.20 | 7.90 | 6.57 | 4.67 |
| 150 cm | 12.27 | 4.25 | 14.86 | 5.43 | 2.49 | 3.20 | |
| 250 cm | 44.61 | 9.75 | 20.75 | 4.69 | 5.43 | 2.62 | |
Head motion ranges (in world coordinate system) used in the final evaluation.
| Proposed (experienced users) | Model | 3.1° | Depth Sensor | None |
| Proposed (informed users) | Model | 3.6° | Depth Sensor | None |
| Proposed (unaware users) | Model | 6.9° | Depth Sensor | None |
| Lu | Appearance | 2–3° | 1 | Capture video |
| Sugano | Appearance | 4–5° | 1 | ≈ 103 training samples |
| Nakazawa and Nitschke [ | Model | 0.9° | 1 IR | IR LEDs & projector |
| Villanueva and Cabeza [ | Model | 1° | 1 IR | 2–4 IR LEDs |
| Zhu and Ji [ | Model | 2° | 2 IR | n IR LEDs |
| Guestrin and Eizenman [ | Model | 1–3° | 1 IR | 2 IR LEDs |
| Yoo and Chung [ | Model | 1–2.5° | 2 IR | 5 IR LEDs |
| Noureddin | Model | 1–3° | 2–4 IR | IR LEDs + Mirrors |