| Literature DB >> 33742418 |
Chiara Jongerius1, T Callemein2, T Goedemé2, K Van Beeck2, J A Romijn3, E M A Smets4, M A Hillen4.
Abstract
The assessment of gaze behaviour is essential for understanding the psychology of communication. Mobile eye-tracking glasses are useful to measure gaze behaviour during dynamic interactions. Eye-tracking data can be analysed by using manually annotated areas-of-interest. Computer vision algorithms may alternatively be used to reduce the amount of manual effort, but also the subjectivity and complexity of these analyses. Using additional re-identification (Re-ID) algorithms, different participants in the interaction can be distinguished. The aim of this study was to compare the results of manual annotation of mobile eye-tracking data with the results of a computer vision algorithm. We selected the first minute of seven randomly selected eye-tracking videos of consultations between physicians and patients in a Dutch Internal Medicine out-patient clinic. Three human annotators and a computer vision algorithm annotated mobile eye-tracking data, after which interrater reliability was assessed between the areas-of-interest annotated by the annotators and the computer vision algorithm. Additionally, we explored interrater reliability when using lengthy videos and different area-of-interest shapes. In total, we analysed more than 65 min of eye-tracking videos manually and with the algorithm. Overall, the absolute normalized difference between the manual and the algorithm annotations of face-gaze was less than 2%. Our results show high interrater agreements between human annotators and the algorithm with Cohen's kappa ranging from 0.85 to 0.98. We conclude that computer vision algorithms produce comparable results to those of human annotators. Analyses by the algorithm are not subject to annotator fatigue or subjectivity and can therefore advance eye-tracking analyses.Entities:
Keywords: Areas-of-interest; Computer vision algorithm; Eye-tracking glasses; Gaze behaviour; Person re-identification; Pose estimation
Mesh:
Year: 2021 PMID: 33742418 PMCID: PMC8516759 DOI: 10.3758/s13428-021-01544-2
Source DB: PubMed Journal: Behav Res Methods ISSN: 1554-351X
Fig. 1An illustration of the primary area-of-interest (right) and the explorative area-of-interest (left)
Fig. 2A screenshot of the eye-tracker analysis software in operation
Fig. 3A graphical example of how the aspect ratios and margin scales for frontal and profile faces were determined
Single class agreement results on the first minute of all seven eye-tracking videos between manual annotators and the computer vision algorithm
| Video | Duration (s) | Face-gaze duration (s) | Δ (s) | normalized Δ (%) | Cohen’s kappa | |
|---|---|---|---|---|---|---|
| Manual | Algorithm | |||||
| #1 | 59.84 | 30.44 | 30.60 | – 0.16 | 0.27 | 0.97 |
| #2 | 60.50 | 27.00 | 26.60 | 0.40 | 0.66 | 0.89 |
| #3 | 60.04 | 30.08 | 31.40 | – 1.32 | 0.22 | 0.94 |
| #4 | 60.46 | 35.36 | 36.56 | – 1.20 | 1.98 | 0.96 |
| #5 | 60.20 | 24.80 | 26.00 | – 1.20 | 1.99 | 0.98 |
| #6 | 60.00 | 24.64 | 26.68 | – 2.04 | 3.40 | 0.91 |
| #7 | 60.00 | 34.4 | 34.56 | – 0.16 | 0.27 | 0.90 |
Multiple class identification accuracy and agreement results for all videos between human annotators and the computer vision algorithm using the Re-ID technique
| Video | Duration of identity - manual (s) | Duration of identity – algorithm (s) | Δ (s) | Normalized Δ (%) | Cohen’s kappa | Re-ID accuracy (%) | ||
|---|---|---|---|---|---|---|---|---|
| Patient | Researchers | Patient | Researchers | |||||
| #1 | 30.16 | 0.28 | 30.44 | 0.16 | – 0.16 | 0.27 | 0.93 | 96.26 |
| #2 | 26.90 | 0.10 | 26.5 | 0.10 | 0.40 | 0.66 | 0.88 | 94.18 |
| #3 | 28.76 | 1.32 | 30.16 | 1.24 | – 1.32 | 2.20 | 0.92 | 95.94 |
| #4 | 34.20 | 1.16 | 35.72 | 0.84 | – 1.20 | 1.98 | 0.93 | 96.50 |
| #5 | 24.80 | 0.00 | 26.00 | 0.00 | – 1.20 | 1.99 | 0.96 | 98.01 |
| #6 | 24.56 | 0.08 | 26.56 | 0.12 | – 2.04 | 3.40 | 0.91 | 95.80 |
| #7 | 24.00 | 0.04 | 23.68 | 0.08 | – 0.16 | 0.27 | 0.91 | 94.53 |
Note: Duration of identity: the time (in s) a specific individual is identified. The identities are patients and researchers, in video #7 additionally a caregiver is identified, the time (in s) the caregiver is identified is shown in italics
Results of agreement analysis on videos #6 and #7 of long duration, and with a different shape of area-of-interest
| Video, condition | Face-gaze duration (s) | Δ (s) | Normalized Δ (%) | Cohen’s kappa | |
|---|---|---|---|---|---|
| Manual | Algorithm | ||||
| #6, long duration, rectangular AOI | 173.92 | 183.52 | – 9.60 | 0.87 | 0.95 |
| #6, long duration, oval AOI | 174.6 | 183.52 | – 8.92 | 0.81 | 0.95 |
| #7, long duration, rectangular AOI | 229.16 | 239.4 | – 10.24 | 1.35 | 0.89 |
| #7, long duration, oval AOI | 209.24 | 239.4 | – 30.16 | 3.97 | 0.87 |
Identification accuracy and agreement results for the explorative videos between human annotators and the computer vision algorithm using the Re-ID technique
| Video, condition | Duration of identity – manual (s) | Duration of identity – algorithm (s) | Δ (s) | Normalized Δ (%) | Cohen’s kappa | Re-ID Accuracy (%) | ||
|---|---|---|---|---|---|---|---|---|
| Patient | Researchers | Patient | Researchers | |||||
| #6 long duration, rectangular AOI | 173.84 | 0.08 | 183.4 | 0.12 | – 9.52 | 0.86 | 0.95 | 98.61 |
| #6 long duration, oval AOI | 173.56 | 0.04 | 183.4 | 0.12 | – 9.76 | 0.88 | 0.95 | 98.65 |
| #7 long duration, rectangular AOI | 195.6 | 0.04 | 189.72 | 0.52 | 22 | 2.90 | 0.87 | 93.92 |
| #7 long duration, oval AOI | 182.16 | 0.2 | 189.72 | 0.52 | 15.04 | 1.98 | 0.85 | 93.18 |
Note: Duration of identity: the time (in s) a specific individual is identified. The identities are patients and researchers, in video #7 additionally a caregiver is identified, the time (in s) the caregiver is identified is shown in italics
| To calculate the areas-of-interest we used an OpenPose base head detection based on a previously published study (Callemein et al., |
| Annotator | Video | Frames | Area-of-interest shape | No. of people in video | Physical examination |
|---|---|---|---|---|---|
| AM | 1 | 1496 | rectangular | 2 | No |
| AM | 2 | 1513 | rectangular | 2 | No |
| AM | 3 | 1501 | rectangular | 2 | No |
| AM | 4 | 1514 | rectangular | 3 | No |
| AM | 5 | 1505 | rectangular |
| No |
| LO | 6 | 27569 | rectangular |
| No |
| LO | 6 | 27569 | oval | 1 | No |
| TB | 6 | 27569 | rectangular | 1 | No |
| TB | 7 | 18973 | rectangular | 3 | Yes |
| TB | 7 | 18973 | oval | 3 | Yes |
| algorithm | ||||
|---|---|---|---|---|
| none | patient | researcher | ||
| manual | none | 6.33 | 0.11 | |
| patient | 0.98 | 0 | ||
| researcher | 0 | 0 | ||
| algorithm | ||||
|---|---|---|---|---|
| none | patient | researcher | ||
| manual | none | 3.95 | 0.14 | |
| patient | 2.92 | 0 | ||
| researcher | 0 | |||
| algorithm | ||||
|---|---|---|---|---|
| none | patient | researcher | ||
| manual | none | 4.65 | 0 | |
| patient | 7.28 | 0 | ||
| researcher | 0 | 0 | ||
| algorithm | |||||
|---|---|---|---|---|---|
| none | patient | researcher1 | researcher2 | ||
| manual | none | 0 | 0 | 6.28 | |
| patient | 16.67 | 0 | 0 | ||
| researcher1 | 3.7 | 0 | 0 | ||
| researcher2 | 1.67 | 0 | 0 | ||
| algorithm | ||||||
|---|---|---|---|---|---|---|
| none | patient | researcher1 | researcher2 | researcher3 | ||
| manual | none | 6.51 | 0 | 0 | 0 | |
| patient | 0.12 | 0 | 0.23 | 0 | ||
| researcher1 | 10 | 0 | 15 | 0 | ||
| researcher2 | 0 | 0 | 0 | 0 | 0 | |
| researcher3 | 33.33 | 0 | 0 | 0 | ||
| algorithm | |||
|---|---|---|---|
| none | patient | ||
| manual | none | 3.39 | |
| patient | 0 | ||
| algorithm | |||||
|---|---|---|---|---|---|
| none | patient | caregiver | researcher | ||
| manual | none | 2.81 | 3.44 | 0 | |
| patient | 4.67 | 0 | 0.33 | ||
| caregiver | 3.09 | 1.16 | 0 | ||
| researcher | 0 | 0 | 0 | ||