Two hours in Hollywood: A manually annotated ground truth data set of eye movements during movie clip watching
- PMID: 33828806
- PMCID: PMC8005322
- DOI: 10.16910/jemr.13.4.5
Two hours in Hollywood: A manually annotated ground truth data set of eye movements during movie clip watching
Abstract
In this short article we present our manual annotation of the eye movement events in a subset of the large-scale eye tracking data set Hollywood2. Our labels include fixations, saccades, and smooth pursuits, as well as a noise event type (the latter representing either blinks, loss of tracking, or physically implausible signals). In order to achieve more consistent annotations, the gaze samples were labelled by a novice rater based on rudimentary algorithmic suggestions, and subsequently corrected by an expert rater. Overall, we annotated eye movement events in the recordings corresponding to 50 randomly selected test set clips and 6 training set clips from Hollywood2, which were viewed by 16 observers and amount to a total of approximately 130 minutes of gaze data. In these labels, 62.4% of the samples were attributed to fixations, 9.1% - to saccades, and, notably, 24.2% - to pursuit (the remainder marked as noise). After evaluation of 15 published eye movement classification algorithms on our newly collected annotated data set, we found that the most recent algorithms perform very well on average, and even reach human-level labelling quality for fixations and saccades, but all have a much larger room for improvement when it comes to smooth pursuit classification. The data set is made available at https://gin.g-node.org/ioannis.agtzidis/hollywood2_em.
Keywords: Eye tracking; eye movement; eye movement classification; gaze; hand-labelling; movie viewing; smooth pursuit.
Conflict of interest statement
The authors declare that the contents of the article are in agreement with the ethics described in http://biblio.unibe.ch/portale/eli-brary/BOP/jemr/ethics.html and that there is no conflict of interest regarding the publication of this paper.
Figures




References
-
- Agtzidis, I., Startsev, M., & Dorr, M. In the pur-suit of (ground) truth: A hand-labelling tool for eye movements recorded during dynamic scene viewing. In: 2016 IEEE Second Work-shop on Eye Tracking and Visualization (ETVIS); 2016. p. 65–68, OI: 10.1109/ETVIS.2016.7851169 - DOI
-
- Agtzidis, I., Startsev, M., & Dorr, M.. 360-degree Video Gaze Behaviour: A Ground-Truth Da-ta Set and a Classification Algorithm for Eye Movements. In: Proceedings of the 27th ACM International Conference on Multime-dia. MM ’19; 2019. p. 1007–1015, DOI: 10.1145/3343031.3350947 - DOI
-
- Agtzidis, I., Startsev, M., & Dorr, M.. Smooth pursuit detection based on multiple observ-ers. In: Proceedings of the Ninth Biennial ACM Symposium on Eye Tracking Research & Applications. ETRA ’16; 2016. p. 303–306, DOI: 10.1145/2857491.2857521 - DOI
-
- Alers, H., Redi, J. A., & Heynderickx, I. (2012). Examin-ing the effect of task on viewing behavior in videos using saliency maps. In Human Vi-sion and Electronic Imaging XVII (Vol. 8291, pp. 1–9). International Society for Optics and Photonics., 10.1117/12.907373 - DOI
LinkOut - more resources
Full Text Sources