| Literature DB >> 33806863 |
Sebastian Kapp1, Michael Barz2,3, Sergey Mukhametov1, Daniel Sonntag2,3, Jochen Kuhn1.
Abstract
Currently an increasing number of head mounted displays (HMD) for virtual and augmented reality (VR/AR) are equipped with integrated eye trackers. Use cases of these integrated eye trackers include rendering optimization and gaze-based user interaction. In addition, visual attention in VR and AR is interesting for applied research based on eye tracking in cognitive or educational sciences for example. While some research toolkits for VR already exist, only a few target AR scenarios. In this work, we present an open-source eye tracking toolkit for reliable gaze data acquisition in AR based on Unity 3D and the Microsoft HoloLens 2, as well as an R package for seamless data analysis. Furthermore, we evaluate the spatial accuracy and precision of the integrated eye tracker for fixation targets with different distances and angles to the user (n=21). On average, we found that gaze estimates are reported with an angular accuracy of 0.83 degrees and a precision of 0.27 degrees while the user is resting, which is on par with state-of-the-art mobile eye trackers.Entities:
Keywords: accuracy; augmented reality; eye tracking; precision; toolkit
Mesh:
Year: 2021 PMID: 33806863 PMCID: PMC8004990 DOI: 10.3390/s21062234
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1A diagram visualizing the components of the toolkit and their interaction.
Overview of recorded data.
| Data Column | Description |
|---|---|
| Time data | |
| eyeDataTimestamp | Unix timestamp of the gaze data (in ms) |
| eyeDataRelativeTimestamp | Relative timestamp of the gaze data (in ms, 100 ns precision) |
| frameTimestamp | Unix timestamp of the frame in which the data was processed (in ms) |
| Gaze data | |
| isCalibrationValid | Flag if the calibration of the wearer is valid |
| gazeHasValue | Flag if valid gaze data exists (origin/direction) |
| gazeOrigin_(x/y/z) | Gaze origin in the global reference frame |
| gazeDirection_(x/y/z) | Gaze direction in the global reference frame |
| gazePointHit | Flag if the raycast hit an object and a gaze position exists |
| gazePoint_(x/y/z) | Position of the gaze point in the global reference frame |
| gazePoint_target_name | Name of the game object hit by the gaze ray |
| gazePoint_target_(x/y/z) | Position of the gaze point in the local reference frame of the hit object |
| gazePoint_target_(pos/rot/scale)_(x/y/z) | Position, rotation, and scale of the game object hit by the gaze ray |
| gazePoint(Left/Right/Mono)Screen_(x,y,z) | Position of the gaze point on the left, right and virtual mono display |
| gazePointWebcam_(x,y,z) | Position of the gaze point on the webcam image |
| AOI data | |
| gazePointAOIHit | Flag if the gaze ray hit an AOI |
| gazePointAOI_(x/y/z) | Position of the gaze point on the AOI in global coordinates |
| gazePointAOI_target_name | Name of the game object representing the AOI |
| gazePointAOI_target_(x/y/z) | Position of the gaze point in the local reference frame of the AOI |
| gazePointAOI_target_(pos/rot/scale)_(x/y/z) | Position, rotation, and scale of the game object hit by the AOI ray |
| gazePointAOIWebcam_(x,y,z) | Position of the gaze point on the AOI on the webcam image |
| Additional information | |
| gameObject_ | Position, rotation, and scale of selected game objects |
| info | Info string of a logged event |
Figure 2Screenshot of the control interface accessible over the network.
Figure 3Mixed reality photo of our HoloLens 2 applications for all three settings which are presented to the participants. The fixation grid for settings I and II is displayed at a fixed distance from the user and resized such that the angular size is identical for all distances (a). The sphere in setting III is positioned 15 cm above the table and stays fixed on top of the visual marker when the participant moves (b). These screenshots are 2D projections which do not reflect the field-of-view and depth perception of a participant in augmented reality (AR).
Figure 4Example of setting I and II in our study with the participant wearing a Microsoft HoloLens 2 and the supervisor controlling the recording using our toolkit.
Accuracy and precision for setting I—resting.
| Distance | Accuracy (SD) | Precision (SD) | ||
|---|---|---|---|---|
| in cm | in deg | in cm | in deg | |
| 0.5 m | 0.91 (0.41) | 1.00 (0.44) | 0.40 (0.16) | 0.29 (0.13) |
| 1.0 m | 1.56 (0.83) | 0.85 (0.46) | 0.67 (0.24) | 0.25 (0.11) |
| 2.0 m | 2.85 (1.31) | 0.77 (0.35) | 1.35 (0.49) | 0.24 (0.10) |
| 4.0 m | 5.03 (2.27) | 0.68 (0.31) | 3.12 (1.26) | 0.28 (0.12) |
Figure 5Plot of the mean accuracy at each distance for each target in setting I—resting. The accuracy angle for all targets is smaller than 1.5 degrees.
Figure 6Recorded gaze point of one participant in relation to the upper left target in setting I—resting. The red dot represents the mean gaze position with each cross being one recorded gaze point.
Results of the post hoc Wilcoxon signed-rank tests for setting I—resting. * the Bonferroni corrected significane level is .
| Comparison | ||||||
|---|---|---|---|---|---|---|
| Z | −2.63 | −3.57 | −3.43 | −1.68 | −2.06 | −1.44 |
| p | 0.009 | <0.001 * | 0.001 * | 0.093 | 0.039 | 0.149 |
Accuracy and precision for setting II—walking.
| Distance | Accuracy (SD) | Precision (SD) | ||
|---|---|---|---|---|
| in cm | in deg | in cm | in deg | |
| 0.5 m | 2.29 (0.64) | 2.52 (0.69) | 1.89 (0.34) | 1.31 (0.25) |
| 1.0 m | 3.35 (1.50) | 1.84 (0.81) | 3.33 (1.00) | 1.16 (0.47) |
| 2.0 m | 5.07 (1.94) | 1.39 (0.53) | 6.32 (1.52) | 1.03 (0.27) |
| 4.0 m | 9.75 (3.08) | 1.33 (0.42) | 12.58 (3.19) | 1.03 (0.32) |
Figure 7Plot of the mean accuracy at each distance for each target in setting II—walking.
Figure 8Recorded gaze point of one participant in relation to the upper left target in setting II—walking. The red dot represents the mean gaze position with each cross being one recorded gaze point.
Results of the post hoc Wilcoxon signed-rank tests for setting II—walking. * the Bonferroni corrected significance level is .
| Comparison | ||||||
|---|---|---|---|---|---|---|
| Z | −3.432 | −3.621 | −3.621 | −3.574 | −2.817 | −0.686 |
| p | 0.001 * | <0.001 * | <0.001 * | <0.001 * | 0.005 * | 0.492 |
Results of the Wilcoxon signed-rank tests for the comparison of the accuracy between setting I and II.
| Distance | 0.5 m | 1.0 m | 2.0 m | 4.0 m |
|---|---|---|---|---|
| Z | −3.62 | −3.62 | −3.57 | −3.53 |
| p | <0.001 | <0.001 | <0.001 | <0.001 |
Results of the Wilcoxon signed-rank tests for the comparison of the precision between setting I and II.
| Distance | 0.5 m | 1.0 m | 2.0 m | 4.0 m |
|---|---|---|---|---|
| Z | −3.62 | −3.62 | −3.62 | −3.62 |
| p | <0.001 | <0.001 | <0.001 | <0.001 |
Accuracy, precision, and mean distance for setting III—stationary target.
| Distance (SD) | Accuracy (SD) | Precision (SD) | ||
|---|---|---|---|---|
| in cm | in cm | in deg | in cm | in deg |
| 49.87 (13.53) | 0.34 (0.27) | 0.39 (0.31) | 0.87 (0.35) | 1.00 (0.40) |
Figure 9Recorded gaze point of one participant in setting III—stationary target. The distance angle for all gaze points is smaller than 3 degrees.
Recommended minimum target size in cm based on Feit et al. [34] and the identified accuracy and precision.
| Distance | Setting I (Resting) | Setting II (Walking) |
|---|---|---|
| 0.5 m | 3.42 cm | 12.14 cm |
| 1.0 m | 5.80 cm | 20.02 cm |
| 2.0 m | 11.10 cm | 35.42 cm |
| 4.0 m | 22.54 cm | 69.82 cm |