Advertisement

The Rules of Guidance in Visual Search

  • Jeremy M. Wolfe
Part of the Lecture Notes in Computer Science book series (LNCS, volume 7143)

Abstract

It is impossible to identify all objects in the visual world at the same time. Accordingly, we must direct attention to specific objects in order to fully recognize them. The deployment of attention is far from random. Attention is guided toward likely targets by a limited set of stimulus attributes such as color and size (“classic guidance”). Attention is also guided by a number of scene-based properties. Thus, if we were looking for sheep, we would expect them on surfaces that could support sheep, not in mid-air. We use information about the 3D layout of a space to determine which objects could plausibly be sheep-sized in that space. This paper briefly reviews the diverse set of guiding properties and the rules that govern their use.

Keywords

Visual Search Visual Attention Target Absent Trial Selective Visual Attention Overt Visual Attention 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Tsotsos, J.: A Computational Perspective on Visual Attention. MIT Press, Cambridge (2011)CrossRefGoogle Scholar
  2. 2.
    Tsotsos, J.K.: Analyzing vision at the complexity level. Brain and Behavioral Sciences 13, 423–469 (1990)CrossRefGoogle Scholar
  3. 3.
    Treisman, A.: Feature binding, attention and object perception. Philos. Trans. R Soc. Lond. B Biol. Sci. 353, 1295–1306 (1998)CrossRefGoogle Scholar
  4. 4.
    Findlay, J.M.: Visual attention: the active vision perspective. Perception ECVP abstract, S1 (2000)Google Scholar
  5. 5.
    Treisman, A., Gelade, G.: A feature-integration theory of attention. Cognitive Psychology 12, 97–136 (1980)CrossRefGoogle Scholar
  6. 6.
    Zelinsky, G.J., Sheinberg, D.L.: Eye movements during parallel / serial visual search. J. Experimental Psychology: Human Perception and Performance 23, 244–262 (1997)Google Scholar
  7. 7.
    Wolfe, J.M.: Moving towards solutions to some enduring controversies in visual search. Trends Cogn. Sci. 7, 70–76 (2003)CrossRefGoogle Scholar
  8. 8.
    Wolfe, J.M., Cave, K.R., Franzel, S.L.: Guided Search: An alternative to the Feature Integration model for visual search. J. Exp. Psychol. - Human Perception and Perf. 15, 419–433 (1989)CrossRefGoogle Scholar
  9. 9.
    Wolfe, J.M.: Guided Search 2.0: A revised model of visual search. Psychonomic Bulletin and Review 1, 202–238 (1994)CrossRefGoogle Scholar
  10. 10.
    Wolfe, J.M.: Guided Search 4.0: Current Progress with a model of visual search. In: Gray, W. (ed.) Integrated Models of Cognitive Systems, Oxford, New York, pp. 99–119 (2007)Google Scholar
  11. 11.
    Wolfe, J.M., Palmer, E.M., Horowitz, T.S.: Reaction time distributions constrain models of visual search. Vision Res. 50, 1304–1311 (2010)CrossRefGoogle Scholar
  12. 12.
    Egeth, H.E., Virzi, R.A., Garbart, H.: Searching for conjunctively defined targets. J. Exp. Psychol: Human Perception and Performance 10, 32–39 (1984)Google Scholar
  13. 13.
    Wolfe, J.M., Horowitz, T.S.: What attributes guide the deployment of visual attention and how do they do it? Nature Reviews Neuroscience 5, 495–501 (2004)CrossRefGoogle Scholar
  14. 14.
    Zeki, S.: Localization and globalization in conscious vision. Annu. Rev. Neurosci. 24, 57–86 (2001)CrossRefGoogle Scholar
  15. 15.
    Found, A., Muller, H.J.: Searching for unknown feature targets on more than one dimension: Investigating a ’dimension weighting’ account. Perception and Psychophysics 58, 88–101 (1996)CrossRefGoogle Scholar
  16. 16.
    Koch, C., Ullman, S.: Shifts in selective visual attention: Towards the underlying neural circuitry. Human Neurobiology 4, 219–227 (1985)Google Scholar
  17. 17.
    Itti, L., Koch, C., Niebur, E.: A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal.Mach. Intell. 20, 1254–1259 (1998)CrossRefGoogle Scholar
  18. 18.
    Zehetleitner, M., Krummenacher, J., Geyer, T., Hegenloh, M., Müller, H.: Dimension intertrial and cueing effects in localization: support for pre-attentively weighted one-route models of saliency. Attention, Perception, & Psychophysics 73, 349–363 (2011)CrossRefGoogle Scholar
  19. 19.
    Bisley, J.W., Goldberg, M.E.: Attention, Intention, and Priority in the Parietal Lobe. Annual Review of Neuroscience 33, 1–21 (2010)CrossRefGoogle Scholar
  20. 20.
    Zhaoping, L., May, K.A.: Psychophysical tests of the hypothesis of a bottom-up saliency map in primary visual cortex. PLoS Comput. Biol. 3, e62 (2007)CrossRefGoogle Scholar
  21. 21.
    Thompson, K.G., Bichot, N.P.: A visual salience map in the primate frontal eye field. Prog. Brain Res. 147, 249–262 (2004)CrossRefGoogle Scholar
  22. 22.
    Parkhurst, D., Law, K., Niebur, E.: Modeling the role of salience in the allocation of overt visual attention. Vision Res. 42, 107–123 (2002)CrossRefGoogle Scholar
  23. 23.
    Serences, J.T., Yantis, S.: Selective visual attention and perceptual coherence. Trends Cogn. Sci. 10, 38–45 (2006)CrossRefGoogle Scholar
  24. 24.
    Goldsmith, M.: What’s in a location? Comparing object-based and space-based models of feature integration in visual search. J. Experimental Psychology: General 127, 189–219 (1998)CrossRefGoogle Scholar
  25. 25.
    Egly, R., Driver, J., Rafal, R.D.: Shifting attention between objects and loctions: Evidence from normal and parietal lesion subjects. J. Experimental Psychology: General 123, 161–177 (1994)CrossRefGoogle Scholar
  26. 26.
    Einhauser, W., Spain, M., Perona, P.: Objects predict fixations better than early saliency. Journal of Vision 8, 1–26 (2008)Google Scholar
  27. 27.
    Duncan, J., Humphreys, G.W.: Visual search and stimulus similarity. Psychological Review 96, 433–458 (1989)CrossRefGoogle Scholar
  28. 28.
    Foster, D.H., Ward, P.A.: Asymmetries in oriented-line detection indicate two orthogonal filters in early vision. Proc. R. Soc. Lond. B 243, 75–81 (1991)CrossRefGoogle Scholar
  29. 29.
    Wolfe, J.M., Friedman-Hill, S.R., Stewart, M.I., O’Connell, K.M.: The role of categorization in visual search for orientation. J. Exp. Psychol: Human Perception and Performance 18, 34–49 (1992)Google Scholar
  30. 30.
    Lindsey, D.T., Brown, A.M., Reijnen, E., Rich, A.N., Kuzmova, Y.I., Wolfe, J.M.: Color channels, not color appearance or color categories, guide visual search for desaturated color targets. Psychol. Sci. 21, 1208–1214 (2010)CrossRefGoogle Scholar
  31. 31.
    Wolfe, J.M.: Bound to guide: A surprising, preattentive role for conjunctions in visual search. Journal of Vision 10, 1289 (2010)CrossRefGoogle Scholar
  32. 32.
    Huang, L.: What is the unit of visual attention? Object for selection, but Boolean map for access. Journal of Experimental Psychology: General 139, 162–179 (2010)CrossRefGoogle Scholar
  33. 33.
    Wolfe, J.M., Yu, K.P., Stewart, M.I., Shorter, A.D., Friedman-Hill, S.R., Cave, K.R.: Limitations on the parallel guidance of visual search: Color X color and orientation X orientation conjunctions. J. Exp. Psychol: Human Perception and Performance 16, 879–892 (1990)Google Scholar
  34. 34.
    Wolfe, J.M., Friedman-Hill, S.R., Bilsky, A.B.: Parallel processing of part/whole information in visual search tasks. Perception and Psychophysics 55, 537–550 (1994)CrossRefGoogle Scholar
  35. 35.
    Vickery, T.J., King, L.-W., Jiang, Y.: Setting up the target template in visual search. J. of Vision 5, 81–92 (2005)CrossRefGoogle Scholar
  36. 36.
    Wolfe, J.M., Alvarez, G.A., Rosenholtz, R.E., Kuzmova, Y.I.: Visual search for arbitrary objects in real scenes. Atten Percept Psychophys 73, 1650–1671 (2011)CrossRefGoogle Scholar
  37. 37.
    Oliva, A.: Gist of the scene. In: Itti, L., Rees, G., Tsotsos, J. (eds.) Neurobiology of Attention, pp. 251–257. Academic Press / Elsevier, San Diego, CA (2005)CrossRefGoogle Scholar
  38. 38.
    Rensink, R.A.: Seeing, sensing, and scrutinizing. Vision Res. 40, 1469–1487 (2000)CrossRefGoogle Scholar
  39. 39.
    Torralba, A., Oliva, A., Castelhano, M.S., Henderson, J.M.: Contextual guidance of eye movements and attention in real-world scenes: The role of global features on object search. Psychological Review 113, 766–786 (2006)CrossRefGoogle Scholar
  40. 40.
    Greene, M.R., Oliva, A.: The briefest of glances: the time course of natural scene understanding. Psychol. Sci. 20, 464–472 (2009)CrossRefGoogle Scholar
  41. 41.
    Vo, M.L.H., Henderson, J.M.: Does gravity matter? Effects of semantic and syntactic inconsistencies on the allocation of attention during scene perception. Journal of Vision 9, 1–15 (2009)CrossRefGoogle Scholar
  42. 42.
    Vo, M.L.-H., Wolfe, J.M.: Does repeated search in scenes need memory? Looking AT versus looking FOR objects in scenes. J. Exp. Psychol: Human Perception and Performance (in press, 2011)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2012

Authors and Affiliations

  • Jeremy M. Wolfe
    • 1
  1. 1.Visual Attention LabBrigham and Women’s Hospital & Harvard Medical SchoolUSA

Personalised recommendations