| Literature DB >> 31428665 |
Francisco M Costela1,2, Russell L Woods1,2.
Abstract
The provided database of tracked eye movements was collected using an infra-red, video-camera Eyelink 1000 system, from 95 participants as they viewed 'Hollywood' video clips. There are 206 clips of 30-s and eleven clips of 30-min for a total viewing time of about 60 hours. The database also provides the raw 30-s video clip files, a short preview of the 30-min clips, and subjective ratings of the content of the videos for each in categories: (1) genre; (2) importance of human faces; (3) importance of human figures; (4) importance of man-made objects; (5) importance of nature; (6) auditory information; (7) lighting; and (8) environment type. Precise timing of the scene cuts within the clips and the democratic gaze scanpath position (center of interest) per frame are provided. At this time, this eye-movement dataset has the widest age range (22-85 years) and is the third largest (in recorded video viewing time) of those that have been made available to the research community. The data-acquisition procedures are described, along with participant demographics, summaries of some common eye-movement statistics, and highlights of research topics in which the database was used. The dataset is freely available in the Open Science Framework repository (link in the manuscript) and can be used without restriction for educational and research purposes, providing that this paper is cited in any published work.Entities:
Keywords: Eye movements; Fixations; Gaze; Movies; Natural viewing; Saccades; Video
Year: 2019 PMID: 31428665 PMCID: PMC6693682 DOI: 10.1016/j.dib.2019.103991
Source DB: PubMed Journal: Data Brief ISSN: 2352-3409
Comparison between freely-available video-viewing eye-movement datasets. Bold indicates the largest amount in each category.
| Database | #subjects | Age range | Sampling rate | #videos | Duration (secs) | Total hours |
|---|---|---|---|---|---|---|
| Actions | 20 | 21–41 | 500 | 1707 | ∼240 | 92 |
| ASCMN | 13 | 23–35 | 30 | 24 | 30 | 2.6 |
| Coutrot Database | 72 | 20–35 | 1000 | 60 | 17 | 20.4 |
| DIEM | 31–218 | 18–36 | 1000 | 85 | 27–217 | ∼90 |
| GazeCom Video | 54 | 18–34 | 250 | 18 | 20 | 5.4 |
| Eye-2-I | 51 | Students | 60 | 4 | 600 | 34 |
| IRCCyN Video 1 | 37 | N/A | 50 | 51 | 10 | 5.2 |
| IRCCyN Video 2 | 30 | ∼23 | 50 | 100 | 6 | 5 |
| IRCCyN Stereoscopic | 40 | 19–44 | 50 | 41 | ∼70 | 32 |
| IRCCyN/HD UHD | 34 | 19–44 | 30 | 37 | ∼11 | 3.9 |
| SAVAM | 50 | 18–27 | 500 | 41 | 20 | 11.4 |
| SFU | 15 | 18–30 | 30 | 12 | 10 | 0.5 |
| TUD Task | 12 | Students | 250 | 25 | 20 | 1.7 |
| USC CRCNS Orig. | 8 | 23–32 | 240 | 50 | 30 | 3.4 |
| USC MTV | 16 | 23–32 | 240 | 50 | 30 | 6.7 |
| USC VAGBA | 14 | 23–32 | 240 | 50 | 30 | 5.8 |
| Our 30 Secs Dataset | 76 | 22–85 | 1000 | 206 | 30 | 21.5 |
| Our 30 Mins Dataset | 19 | 22–58 | 1000 | 11 | 1800 | 37.5 |
Self-reported demographic characteristics of participants.
| Gender | Male | 52 (54%) |
| Female | 43 (46%) | |
| Age (median, min-max) | 56.2 y (22-85y) | |
| Race/Ethnicity | Black | 5 (5%) |
| White | 87 (91%) | |
| Asian | 4 (4%) | |
| Hispanic | 1 (1%) | |
| Not registered | 3 (3%) | |
| Highest education | High school diploma | 4 (5%) |
| Some college | 6 (8%) | |
| Bachelor's degree | 23 (32%) | |
| Associate degree | 2 (2%) | |
| Master's degree | 18 (24%) | |
| Professional degree | 7 (9%) | |
| Doctoral degree | 11 (17%) |
Fig. 1Saccadic features for all 95 subjects and 296,249 saccades. A) Saccadic peak velocity–magnitude main sequence. The distribution is plotted on logarithmic scale where peak velocity is indicated on the y-axis and magnitude indicated on the x-axis. Colors represent the number of samples (right-hand scale). B) Polar distribution of saccade directions (bin size 4.5°). Distance from the origin represents frequency.
Fig. 2Example of democratic COI determination. (A) Gaze locations of the 24 participants during one frame of one video clip. (B) Kernel density estimate of those gaze locations in that video frame shown as a heat map, with red representing a higher density. The green rectangle represents the box used to determine the democratic COI.
Fig. 3Mean NSS score for participants. Gray filled circles represent the average NSS score for each of 61 participants who viewed the 30-s clips. The flat horizontal black line corresponds to the global average for the group.
Specifications Table
| Subject area | |
| More specific subject area | Visual Science and Psychophysics area, interested in analyzing gaze data from people watching video with a wide spectrum of ages |
| Type of data | Gaze eye movements, videoclips, trial condition data, demographics tables, timing tables |
| How data was acquired | An EyeLink 1000 eye-tracker was used |
| Data format | Gaze unfiltered data (x, y, time coordinates, trial condition data), video, Excel files |
| Experimental factors | Raw data (from EDF files) were exported as Matlab (easy to access) files preserving all coordinates and adding experimental trial data. |
| Experimental features | Subjects viewed a subset of a total of 217 clips from professionally recorded video material (“Hollywood” movies) using a high-resolution, infrared-sensing eye tracker |
| Data source location | Data were collected and stored at Schepens Eye Research Institute, Boston, Massachusetts, USA. |
| Data accessibility | Data stored in public repository Open Science Framework. Link: |
| Related research article | Costela, F. M., & Woods, R. L. (2018). When watching video, many saccades are curved and deviate from a velocity profile model. |
Comparison between freely-available video-viewing eye-movement datasets. Bold indicates the largest amount in each category.
Watching television, movies, and other video content is a major source of entertainment, relaxation, education, information (e.g., current affairs), and supports personal identity, integration, and social interaction At this time, the database is, to our knowledge, the largest video dataset with balanced age distribution among participants (up to 85 years; see Comparison between freely-available video-viewing eye-movement datasets. Bold indicates the largest amount in each category. Most publicly-available eye-movement datasets have static-images as stimuli, such as images of scenes or faces. (see also the MIT Saliency Benchmark: This database could be put to use under a variety of research circumstances, such as: gaze statistics, developing and testing models of visual salience, models of ocular motor control, reduction of band-width by restricting high resolution to the video scan path, and certainly others that were not considered. The dataset could be used as stimuli for neurons or computational neurons or networks. It has been proposed that neurons and neuronal systems may respond differently to natural stimuli as compared to manufactured stimuli. The dataset contains natural scenes that are in motion and for which there is gaze data, so may further enhance our understanding of neural mechanisms. For testing of neural arrays, a model of the peripheral degradation in the retinal image quality could be incorporated by modifying the video based on the gaze data. The database may make a significant contribution to the image processing community, for both educational and research purposes. |