Predicting Visual Fixations
- PMID: 37419107
- DOI: 10.1146/annurev-vision-120822-072528
Predicting Visual Fixations
Abstract
As we navigate and behave in the world, we are constantly deciding, a few times per second, where to look next. The outcomes of these decisions in response to visual input are comparatively easy to measure as trajectories of eye movements, offering insight into many unconscious and conscious visual and cognitive processes. In this article, we review recent advances in predicting where we look. We focus on evaluating and comparing models: How can we consistently measure how well models predict eye movements, and how can we judge the contribution of different mechanisms? Probabilistic models facilitate a unified approach to fixation prediction that allows us to use explainable information explained to compare different models across different settings, such as static and video saliency, as well as scanpath prediction. We review how the large variety of saliency maps and scanpath models can be translated into this unifying framework, how much different factors contribute, and how we can select the most informative examples for model comparison. We conclude that the universal scale of information gain offers a powerful tool for the inspection of candidate mechanisms and experimental design that helps us understand the continual decision-making process that determines where we look.
Keywords: benchmarking; eye movements; fixations; information theory; model comparison; saliency; taxonomy; transfer learning; unifying framework.
Similar articles
-
What can saliency models predict about eye movements? Spatial and sequential aspects of fixations during encoding and recognition.J Vis. 2008 Feb 20;8(2):6.1-17. doi: 10.1167/8.2.6. J Vis. 2008. PMID: 18318632
-
Augmented saliency model using automatic 3D head pose detection and learned gaze following in natural scenes.Vision Res. 2015 Nov;116(Pt B):113-26. doi: 10.1016/j.visres.2014.10.027. Epub 2014 Nov 13. Vision Res. 2015. PMID: 25448115
-
On metrics for measuring scanpath similarity.Behav Res Methods. 2021 Apr;53(2):609-628. doi: 10.3758/s13428-020-01441-0. Behav Res Methods. 2021. PMID: 32779104 Review.
-
Gravitational Laws of Focus of Attention.IEEE Trans Pattern Anal Mach Intell. 2020 Dec;42(12):2983-2995. doi: 10.1109/TPAMI.2019.2920636. Epub 2020 Nov 3. IEEE Trans Pattern Anal Mach Intell. 2020. PMID: 31180885
-
Information-theoretic model comparison unifies saliency metrics.Proc Natl Acad Sci U S A. 2015 Dec 29;112(52):16054-9. doi: 10.1073/pnas.1510393112. Epub 2015 Dec 10. Proc Natl Acad Sci U S A. 2015. PMID: 26655340 Free PMC article.
Cited by
-
A robotics-inspired scanpath model reveals the importance of uncertainty and semantic object cues for gaze guidance in dynamic scenes.J Vis. 2025 Feb 3;25(2):6. doi: 10.1167/jov.25.2.6. J Vis. 2025. PMID: 39928323 Free PMC article.
-
Potsdam data set of eye movement on natural scenes (DAEMONS).Front Psychol. 2024 May 10;15:1389609. doi: 10.3389/fpsyg.2024.1389609. eCollection 2024. Front Psychol. 2024. PMID: 38800681 Free PMC article. No abstract available.
-
The fundamentals of eye tracking part 1: The link between theory and research question.Behav Res Methods. 2024 Dec 12;57(1):16. doi: 10.3758/s13428-024-02544-8. Behav Res Methods. 2024. PMID: 39668288 Free PMC article.
Publication types
MeSH terms
LinkOut - more resources
Full Text Sources
Research Materials