Abstract
The manner in which we sample visual information from the world is constrained by the spatial and temporal sampling limits of the human eye. High acuity vision is restricted to the small central foveal region of the retina, which is limited to just a few degrees of visual angle in extent. Moreover, visual sampling is effectively limited to when the retinal image is relatively stabilised for periods of fixation (Erdmann and Dodge 1898) , which last on average around 200–400 ms when viewing text, scenes or real environments (Land and Tatler 2009; Rayner 1998) . It is clear from these severe spatiotemporal constraints on visual sampling that high acuity vision is a scarce resource and, like any scarce resource, it must be distributed carefully and appropriately for the current situation.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Ballard, D. H., Hayhoe, M. M., Li, F., & Whitehead, S. D. (1992). Hand-eye coordination during sequential tasks. Philosophical Transactions of the Royal Society of London Series B-Biological Sciences, 337(1281), 331–338.
Buswell, G. T. (1920). An experimental study of the eye-voice span in reading. Chicago: Chicago University Press.
Buswell, G. T. (1935). How people look at pictures: A Study of the Psychology of Perception in Art. Chicago: University of Chicago Press.
Cameron, E. H., & Steele, W. M. (1905). The poggendorff illusion. Psychological Monographs, 7(1), 83–111.
Carmi, R., & Itti, L. (2006). Visual causes versus correlates of attentional selection in dynamic scenes. Vision Research, 46(26), 4333–4345. doi:10.1016/j.visres.2006.08.019.
Cristino, F., & Baddeley, R. J. (2009). The nature of the visual representations involved in eye movements when walking down the street. Visual Cognition, 17(6–7), 880–903.
Dicks, M., Button, C., & Davids, K. (2010). Examination of gaze behaviors under in situ and video simulation task constraints reveals differences in information pickup for perception and action. Attention, Perception, & Psychophysics, 72(3), 706–720. doi:10.3758/APP.72.3.706.
Dorr, M., Martinetz, T., Gegenfurtner, K. R., & Barth, E. (2010). Variability of eye movements when viewing dynamic natural scenes. Journal of Vision, 10(10), 28, 1–17. doi:10.1167/10.10.28.
Ehinger, K. A., Hidalgo-Sotelo, B., Torralba, A., & Oliva, A. (2009). Modeling Search for People in 900 Scenes: A combined source model of eye guidance. Visual Cognition, 17(6–7), 945–978. doi:10.1080/13506280902834720.
Einhauser, W., Spain, M., & Perona, P. (2008). Objects predict fixations better than early saliency. Journal of Vision, 8(14):18, 1–26, http://www.journalofvision.org/content/8/14/18, doi:10.1167/8.14.18.
Elazary, L., & Itti, L. (2008). Interesting objects are visually salient. Journal of Vision, 8(3), 3.1–15. doi:10.1167/8.3.3.
Epelboim, J. L., Steinman, R. M., Kowler, E., Edwards, M., Pizlo, Z., Erkelens, C. J., & Collewijn, H. (1995). The function of visual search and memory in sequential looking tasks. Vision Research, 35(23–24), 3401–3422.
Epelboim, J. L., Steinman, R. M., Kowler, E., Pizlo, Z., Erkelens, C. J., & Collewijn, H. (1997). Gaze-shift dynamics in two kinds of sequential looking tasks. Vision Research, 37(18), 2597–2607.
Erdmann, B., & Dodge, R. (1898). Psychologische Untersuchungen über das Lesen auf experimenteller Grundlage. Halle: Niemeyer.
Foulsham, T., & Underwood, G. (2008). What can saliency models predict about eye movements? Spatial and sequential aspects of fixations during encoding and recognition. Journal of Vision, 8(2), 6.1–17. doi:10.1167/8.2.6.
Frith, C.D., Blakemore, S-J., & Wolpert, D.M. (2000). Abnormalities in the awareness of and control of action. Philosophical Transactions of the Royal Society of London B 355, 1771–1788.
Furneaux, S., & Land, M. F. (1999). The effects of skill on the eye-hand span during musical sight- reading. Proceedings of the Royal Society of London Series B-Biological Sciences, 266(1436), 2435–2440.
Hayhoe, M. M., McKinney, T., Chajka, K., & Pelz, J. B. (2011). Predictive eye movements in natural vision. Experimental Brain Research, 217(1), 125–136. doi:10.1007/s00221-011-2979-2.
Hayhoe, M., Mennie, N., Sullivan, B., & Gorgos, K. (2005) The role of internal models and prediction in catching balls. In: Proceedings of AAAI fall symposium series.
Hayhoe, M. M., Shrivastava, A., Mruczek, R., & Pelz, J. B. (2003). Visual memory and motor planning in a natural task. Journal of Vision, 3(1), 49–63. doi:10:1167/3.1.6.
Henderson, J. M. (2003). Human gaze control during real-world scene perception. Trends in Cognitive Sciences, 7(11), 498–504.
Henderson, J. M., Brockmole, J. R., Castelhano, M. S., & Mack, M. L. (2007). Visual saliency does not account for eye movements during search in real-world scenes. In R. P. G. van Gompel, M. H. Fischer, W. S. Murray, & R. L. Hill (Eds.), Eye movements: A window on mind and brain (pp. 537–562). Oxford, UK: Elsevier.
Itti, L., & Koch, C. (2000). A saliency-based search mechanism for overt and covert shifts of visual attention. Vision Research, 40(10–12), 1489–1506.
Itti, L., Koch, C., & Niebur, E. (1998). A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20, 1254–1259.
Jovancevic-Misic, J., & Hayhoe, M. (2009). Adaptive Gaze Control in Natural Environments. Journal of Neuroscience, 29(19), 6234–6238. doi:10.1523/JNEUROSCI.5570-08.2009.
Judd, C. H. (1905). The Müller-Lyer illusion. Psychological Monographs, 7(1), 55–81.
Judd, C. H., & Courten, H. C. (1905). The Zöllner illusion. Psychological Monographs, 7(1), 112–139.
Kanan, C., Tong, M. H., Zhang, L., & Cottrell, G. W. (2009). SUN: Top-down saliency using natural statistics. Visual Cognition, 17(6–7), 979–1003.
Koch, C., & Ullman, S. (1985). Shifts in selective visual-attention—towards the underlying neural circuitry. Human Neurobiology, 4(4), 219–227.
Land, M. F., & Furneaux, S. (1997). The knowledge base of the oculomotor system. Philosophical Transactions of the Royal Society of London Series B-Biological Sciences, 352(1358), 1231–1239.
Land, M. F., & Lee, D. N. (1994). Where we look when we steer. Nature, 369(6483), 742–744. doi:10.1038/369742a0.
Land, M. F., & McLeod, P. (2000). From eye movements to actions: How batsmen hit the ball. Nature Neuroscience, 3(12), 1340–1345. doi:10.1038/81887.
Land, M. F., & Tatler, B. W. (2001). Steering with the head: The visual strategy of a racing driver. Current Biology, 11(15), 1215–1220.
Land, M. F., & Tatler, B. W. (2009). Looking and acting: Vision and eye movements in natural behaviour. Oxford: OUP.
Land, M. F., Mennie, N., & Rusted, J. (1999). The roles of vision and eye movements in the control of activities of daily living. Perception, 28(11), 1311–1328.
McAllister, C. N. (1905). The fixation of points in the visual field. Psychological Monographs, 7(1), 17–53.
Mehta, B. & Schaal, S. (2002). Forward models in visuomotor control. Journal of Neurophysiology 88, 942–953.
Miall, R.C. & Wolpert, D.M. (1996). Forward models for physiological motor control. Neural Networks, 9, 1265–1279.
Montague, P., & Hyman, S. (2004). Computational roles for dopamine in behavioural control. Nature, 431, 760–767.
Navalpakkam, V., & Itti, L. (2005). Modeling the influence of task on attention. Vision Research, 45, 205–231.
Navalpakkam, V., & Itti, L. (2010). A goal oriented attention guidance model. Biologically Motivated Computer Vision, 81–118.
Nuthmann, A., & Henderson, J. M. (2010). Object-based attentional selection in scene viewing. Journal of Vision, 10(8) 20, 1–19. doi:10.1167/10.8.20.
Nyström, M., & Holmqvist, K. (2008). Semantic override of low-level features in image viewingboth initially and overall. Journal of Eye Movement Research, 2, 1–11.
Parkhurst, D., Law, K., & Niebur, E. (2002). Modeling the role of salience in the allocation of overt visual attention. Vision Research, 42(1), 107–123.
Patla, A. E., & Vickers, J. N. (1997). Where and when do we look as we approach and step over an obstacle in the travel path? Neuroreport, 8(17), 3661–3665.
Patla, A. E., & Vickers, J. N. (2003). How far ahead do we look when required to step on specific locations in the travel path during locomotion? Experimental brain research Experimentelle Hirnforschung Expérimentation cérébrale, 148(1), 133–138. doi:10.1007/s00221-002-1246-y.
Pelz, J. B., & Canosa, R. (2001). Oculomotor behavior and perceptual strategies in complex tasks. Vision Research, 41(25–26), 3587–3596.
Rayner, K. (1998). Eye Movements in Reading and Information Processing: 20 Years of Research. Psychological bulletin, 124(3), 372–422.
Reinagel, P., & Zador, A. M. (1999). Natural scene statistics at the centre of gaze. Network (Bristol, England), 10(4), 341–350.
Rothkopf, C. A., Ballard, D. H., & Hayhoe, M. M. (2007). Task and context determine where you look. Journal of Vision, 7(14), 16.1–20. doi:10.1167/7.14.16.
Sailer, U., Flanagan, J. R., & Johansson, R. S. (2005). Eye-hand coordination during learning of a novel visuomotor task. The Journal of neuroscience: The official journal of the Society for. Neuroscience, 25(39), 8833–8842. doi:10.1523/JNEUROSCI.2658-05.2005.
Schultz, W. (2000). Multiple reward signals in the brain. Nature Reviews Neuroscience, 1(3), 199–207.
Seydell, A., McCann, B. C., Trommershäuser, J., & Knill, D. C. (2008). Learning stochastic reward distributions in a speeded pointing task. Journal of Neuroscience, 28, 4356–4367.
Sprague, N., & Ballard, D. H., & Robinson, A. (2007). Modeling embodied visual behaviors. ACM Transactions on Applied Perception, 4, 11.
Steinman, R. M. (2003). Gaze control under natural conditions. In: Chalupa, L.M., & Werner, J.S. (Eds). The Visual Neurosciences. pp. 1339–1356. Cambridge: MIT Press.
Stritzke, M., Trommershäuser, J., & Gegenfurtner, K. R. (2009). Effects of salience and reward information during saccadic decisions under risk. Journal of the Optical Society of America A, 26, B1–B13.
Stratton, G. M. (1906). Symmetry, linear illusions, and the movements of the eye. Psychological Review, 13, 82–96.
Tassinari, H., Hudson, T. E., & Landy, M. S. (2006). Combining priors and noisy visual cues in a rapid pointing task. Journal of Neuroscience, 26, 10154–10163.
Tatler, B. W. (2007). The central fixation bias in scene viewing: Selecting an optimal viewing position independently of motor biases and image feature distributions. Journal of Vision, 7(14):4, 1–17, http://www.journalofvision.org/content/7/14/4, doi:10.1167/7.14.4.
Tatler, B. W., & Kuhn, G. (2007). Don’t look now: the misdirection of magic. In: van Gompel, R., Fischer, M., Murray, W., & Hill, R. (Eds). Eye Movement Research: Insights into Mind and Brain. pp. 697–714. Amsterdam: Elsevier.
Tatler, B. W., Baddeley, R. J., & Gilchrist, I. (2005). Visual correlates of fixation selection: Effects of scale and time. Vision Research, 45(5), 643–659.
Tatler, B. W., Hayhoe, M. M., Land, M. F., & Ballard, D. H. (2011). Eye guidance in natural vision: Reinterpreting salience. Journal of Vision, 11(5), 5 1–23. doi:10.1167/11.5.5.
Torralba, A., Oliva, A., Castelhano, M. S., & Henderson, J. M. (2006). Contextual guidance of eye movements and attention in real-world scenes: The role of global features in object search. Psychological Review, 113(4), 766–786.
Treisman, A. M., & Gelade, G. (1980). A feature-integration theory of attention. Cognitive Psychology, 12, 97–136.
Trommershäuser, J., Maloney, L. T., & Landy, M. S. (2003). Statistical decision theory and the selection of rapid, goal-directed movements. Journal of the Optical Society of America A, 20, 1419–1433.
Trommershäuser, J., Maloney, L. T., & Landy, M. S. (2008). Decision making, movement planning, and statistical decision theory. Trends in Cognitive Sciences, 12, 291–297.
Underwood, G., Foulsham, T., van Loon, E., Humphreys, L., & Bloyce, J. (2006). Eye movements during scene inspection: A test of the saliency map hypothesis. European Journal of. Cognitive Psychology, 18(3), 321–342. doi:10.1080/09541440500236661.
Wischnewski, M., Belardinelli, A., & Schneider, W. (2010). Where to look next? Combining static and dynamic proto-objects in a TVA-based model of visual attention. Cognitive Computation, 2(4), 326–343.
Wolfe, J. (2007). Guided Search 4.0: Current Progress with a model of visual search. In W. Gray (Ed.), Integrated models of cognitive systems (pp. 99–119). New York: OUP.
Wolfe, J. M. (1998). What can 1 million trials tell us about visual search? Psychological Science, 9(1), 33–39.
Wolpert. D.M., Ghahramani, Z., & Jordan, M.I. (1995). An internal model for sensorimotor integration. Science 269, 1880–1882.
Zago M, McIntyre J, Patrice Senot P, & Lacquaniti F (2009) Visuo-motor coordination and internal models for object interception. Experimental Brain Research 192(4):571–604.
Yarbus, A. L. (1967). Eye movements and vision. New York: Plenum Press.
Zelinsky, G. J. (2008). A theory of eye movements during target acquisition. Psychological Review, 115(4), 787–835. doi:10.1037/a0013118.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Tatler, B. (2014). Eye Movements from Laboratory to Life. In: Horsley, M., Eliot, M., Knight, B., Reilly, R. (eds) Current Trends in Eye Tracking Research. Springer, Cham. https://doi.org/10.1007/978-3-319-02868-2_2
Download citation
DOI: https://doi.org/10.1007/978-3-319-02868-2_2
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-02867-5
Online ISBN: 978-3-319-02868-2
eBook Packages: Humanities, Social Sciences and LawEducation (R0)