Learning a saliency map using fixated locations in natural scenes
- PMID: 21393388
- DOI: 10.1167/11.3.9
Learning a saliency map using fixated locations in natural scenes
Abstract
Inspired by the primate visual system, computational saliency models decompose visual input into a set of feature maps across spatial scales in a number of pre-specified channels. The outputs of these feature maps are summed to yield the final saliency map. Here we use a least square technique to learn the weights associated with these maps from subjects freely fixating natural scenes drawn from four recent eye-tracking data sets. Depending on the data set, the weights can be quite different, with the face and orientation channels usually more important than color and intensity channels. Inter-subject differences are negligible. We also model a bias toward fixating at the center of images and consider both time-varying and constant factors that contribute to this bias. To compensate for the inadequacy of the standard method to judge performance (area under the ROC curve), we use two other metrics to comprehensively assess performance. Although our model retains the basic structure of the standard saliency model, it outperforms several state-of-the-art saliency algorithms. Furthermore, the simple structure makes the results applicable to numerous studies in psychophysics and physiology and leads to an extremely easy implementation for real-world applications.
Similar articles
-
Learning visual saliency by combining feature maps in a nonlinear manner using AdaBoost.J Vis. 2012 Jun 15;12(6):22. doi: 10.1167/12.6.22. J Vis. 2012. PMID: 22707429
-
Saliency does not account for fixations to eyes within social scenes.Vision Res. 2009 Dec;49(24):2992-3000. doi: 10.1016/j.visres.2009.09.014. Epub 2009 Sep 24. Vision Res. 2009. PMID: 19782100
-
What can saliency models predict about eye movements? Spatial and sequential aspects of fixations during encoding and recognition.J Vis. 2008 Feb 20;8(2):6.1-17. doi: 10.1167/8.2.6. J Vis. 2008. PMID: 18318632
-
A quantitative theory of immediate visual recognition.Prog Brain Res. 2007;165:33-56. doi: 10.1016/S0079-6123(06)65004-8. Prog Brain Res. 2007. PMID: 17925239 Review.
-
Predicting Visual Fixations.Annu Rev Vis Sci. 2023 Sep 15;9:269-291. doi: 10.1146/annurev-vision-120822-072528. Epub 2023 Jul 7. Annu Rev Vis Sci. 2023. PMID: 37419107 Review.
Cited by
-
An improved saliency model of visual attention dependent on image content.Front Hum Neurosci. 2023 Feb 28;16:862588. doi: 10.3389/fnhum.2022.862588. eCollection 2022. Front Hum Neurosci. 2023. PMID: 36926377 Free PMC article.
-
Visual search as an embodied process: The effects of perspective change and external reference on search performance.J Vis. 2022 Sep 2;22(10):13. doi: 10.1167/jov.22.10.13. J Vis. 2022. PMID: 36107125 Free PMC article.
-
DeepGaze III: Modeling free-viewing human scanpaths with deep learning.J Vis. 2022 Apr 6;22(5):7. doi: 10.1167/jov.22.5.7. J Vis. 2022. PMID: 35472130 Free PMC article.
-
Weighting the factors affecting attention guidance during free viewing and visual search: The unexpected role of object recognition uncertainty.J Vis. 2022 Mar 2;22(4):13. doi: 10.1167/jov.22.4.13. J Vis. 2022. PMID: 35323870 Free PMC article.
-
Assessment of Color Perception and Preference with Eye-Tracking Analysis in a Dental Treatment Environment.Int J Environ Res Public Health. 2021 Jul 28;18(15):7981. doi: 10.3390/ijerph18157981. Int J Environ Res Public Health. 2021. PMID: 34360272 Free PMC article.
Publication types
MeSH terms
LinkOut - more resources
Full Text Sources
Other Literature Sources