Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2020 Jul 27;13(4):10.16910/jemr.13.4.5.
doi: 10.16910/jemr.13.4.5.

Two hours in Hollywood: A manually annotated ground truth data set of eye movements during movie clip watching

Affiliations

Two hours in Hollywood: A manually annotated ground truth data set of eye movements during movie clip watching

Ioannis Agtzidis et al. J Eye Mov Res. .

Abstract

In this short article we present our manual annotation of the eye movement events in a subset of the large-scale eye tracking data set Hollywood2. Our labels include fixations, saccades, and smooth pursuits, as well as a noise event type (the latter representing either blinks, loss of tracking, or physically implausible signals). In order to achieve more consistent annotations, the gaze samples were labelled by a novice rater based on rudimentary algorithmic suggestions, and subsequently corrected by an expert rater. Overall, we annotated eye movement events in the recordings corresponding to 50 randomly selected test set clips and 6 training set clips from Hollywood2, which were viewed by 16 observers and amount to a total of approximately 130 minutes of gaze data. In these labels, 62.4% of the samples were attributed to fixations, 9.1% - to saccades, and, notably, 24.2% - to pursuit (the remainder marked as noise). After evaluation of 15 published eye movement classification algorithms on our newly collected annotated data set, we found that the most recent algorithms perform very well on average, and even reach human-level labelling quality for fixations and saccades, but all have a much larger room for improvement when it comes to smooth pursuit classification. The data set is made available at https://gin.g-node.org/ioannis.agtzidis/hollywood2_em.

Keywords: Eye tracking; eye movement; eye movement classification; gaze; hand-labelling; movie viewing; smooth pursuit.

PubMed Disclaimer

Conflict of interest statement

The authors declare that the contents of the article are in agreement with the ethics described in http://biblio.unibe.ch/portale/eli-brary/BOP/jemr/ethics.html and that there is no conflict of interest regarding the publication of this paper.

Figures

Figure 1.
Figure 1.
Sample scenes from the Hollywood2 data set overlaid with 600 ms of gaze samples. The gaze pattern of each participant is visualised with a unique colour. The left and right snapshots show scenes with substantial motion and they contain 72% and 59% of smooth pursuit respectively. The high prevalence of pursuit is also visible by the elongated coloured lines (representing consecutive gaze samples) that are oriented along the direction of the motion. The middle snapshot does not contain smooth pursuit and the trace directions are more varied.
Figure 2.
Figure 2.
Example screenshot of the tool as it was used during labelling. The two panels on the right are used during the labelling procedure in order to change the borders of the colour-coded areas depending on the x and y coordinates (black lines). The still frame here (top-left panel) comes from the same video as the rightmost frame in Figure 1. For the specific participant here, most of the samples are labelled as SP (blue boxes) due to the gaze following either camera or object motion.
Figure 3.
Figure 3.
Distributions of speed (3a), duration (3b), and amplitude (3c) for the three labelled eye movement classes. Note the logarithmic scale of the x axis (chosen due to the large range of the reported statistics for the three classes). To facilitate the comparison between the distributions, we visualise the first and third quartiles of each distribution as vertical dashed lines.
Figure 4.
Figure 4.
Event speed distribution for the GazeCom data set. Note the logarithmic scale of the x axis (chosen due to the large range of the reported speed for the three classes). The figure presented here is a reproduction of Figure 4 from [9].

References

    1. Agtzidis, I., Startsev, M., & Dorr, M. In the pur-suit of (ground) truth: A hand-labelling tool for eye movements recorded during dynamic scene viewing. In: 2016 IEEE Second Work-shop on Eye Tracking and Visualization (ETVIS); 2016. p. 65–68, OI: 10.1109/ETVIS.2016.7851169 - DOI
    1. Agtzidis, I., Startsev, M., & Dorr, M.. 360-degree Video Gaze Behaviour: A Ground-Truth Da-ta Set and a Classification Algorithm for Eye Movements. In: Proceedings of the 27th ACM International Conference on Multime-dia. MM ’19; 2019. p. 1007–1015, DOI: 10.1145/3343031.3350947 - DOI
    1. Agtzidis, I., Startsev, M., & Dorr, M.. Smooth pursuit detection based on multiple observ-ers. In: Proceedings of the Ninth Biennial ACM Symposium on Eye Tracking Research & Applications. ETRA ’16; 2016. p. 303–306, DOI: 10.1145/2857491.2857521 - DOI
    1. Alers, H., Redi, J. A., & Heynderickx, I. (2012). Examin-ing the effect of task on viewing behavior in videos using saliency maps. In Human Vi-sion and Electronic Imaging XVII (Vol. 8291, pp. 1–9). International Society for Optics and Photonics., 10.1117/12.907373 - DOI
    1. Andersson, R., Larsson, L., Holmqvist, K., Stridh, M., & Nyström, M. (2017). One algorithm to rule them all? An evaluation and discussion of ten eye movement event-detection algorithms. Behavior Research Methods, 49(2), 616–637. 10.3758/s13428-016-0738-9 - DOI - PubMed

LinkOut - more resources