Abstract
Sensory systems use limited resources to mediate the perception of a great variety of objects and events. Here a normative framework is presented for exploring how the problem of efficient allocation of resources can be solved in visual perception. Starting with a basic property of every measurement, captured by Gabor’s uncertainty relation about the location and frequency content of signals, prescriptions are developed for optimal allocation of sensors for reliable perception of visual motion. This study reveals that a large-scale characteristic of human vision (the spatiotemporal contrast sensitivity function) is similar to the optimal prescription, and it suggests that some previously puzzling phenomena of visual sensitivity, adaptation, and perceptual organization have simple principled explanations.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
For brevity, here “frequency content” will sometimes be shortened to “content.”
- 2.
Different criteria of measurement and sensor shapes correspond to different magnitudes of \( C_{x} \).
- 3.
Here the sensors are characterized by intervals following the standard notion that biological motion sensors are maximally activated when the stimulus travels some distance \( \varDelta s \) over some temporal interval \( \varDelta t \) [17].
References
Marr D (1982) Vision: a computational investigation into the human representation and processing of visual information. W. H. Freeman, San Francisco
Gabor D (1946) Theory of communication. Inst Electr Eng, Part III 93:429–457
Marcelja S (1980) Mathematical description of the response by simple cortical cells. J Opt Soc Am 70:1297–1300
MacKay DM (1981) Strife over visual cortical function. Nature 289:117–118
Daugman JG (1985) Uncertainty relation for the resolution in space spatial frequency, and orientation optimized by two-dimensional visual cortex filters. J Opt Soc Am A 2(7):1160–1169
Glezer VD, Gauzel’man VE, Iakovlev VV (1986) Principle of uncertainty in vision. Neirofiziologiia [Neurophysiology] 18(3):307–312
Field DJ (1987) Relations between the statistics of natural images and the response properties of cortical cells. J Opt Soc Am A 4:2379–2394
Jones A, Palmer L (1987) An evaluation of the two-dimensional Gabor filter model of simple receptive fields in cat striate cortex. J Neurophysiol 58:1233–1258
Simoncelli EP, Olshausen B (2001) Natural image statistics and neural representation. Annu Rev Neurosci 24:1193–1216
Saremi S, Sejnowski TJ, Sharpee TO (2013) Double-Gabor filters are independent components of small translation-invariant image patches. Neural Comput 25(4):922–939
Gabor D (1952) Lectures on communication theory. Technical report 238, MIT Research Laboratory of Electronics, Cambridge, MA, USA
Resnikoff HL (1989) The illusion of reality. Springer, New York
MacLennan B (1994) Gabor representations of spatiotemporal visual images. Technical report. University of Tennessee, Knoxville, TN, USA
Gepshtein S, Tyukin I (2006) Why do moving things look as they do? vision. J Vis Soc Jpn, Supp. 18:64
von Neumann J (1928) Zur Theorie der Gesellschaftsspiele. [On the theory of games of strategy]. Mathematische Annalen 100 (1928):295–320, English translation in [57]
Luce RD, Raiffa H (1957) Games and decisions. John Wiley, New York
Watson AB, Ahumada AJ (1985) Model of human visual-motion sensing. J Opt Soc Am A 2(2):322–341
Gepshtein S (2010) Two psychologies of perception and the prospect of their synthesis. Philos Psychol 23:217–281
Gepshtein S, Tyukin I, Kubovy M (2007) The economics of motion perception and invariants of visual sensitivity. J Vis 7, 8(8):1–18
Gepshtein S, Kubovy M (2007) The lawful perception of apparent motion. J Vis 7, 8(9):1–15
Gepshtein S, Tyukin I, Kubovy M (2011) A failure of the proximity principle in the perception of motion. Humana Mente 17:21–34
Korte A (1915) Kinematoskopische Untersuchungen. Zeitschrift für Psychologie 72:194–296
Burt P, Sperling G (1981) Time, distance, and feature tradeoffs in visual apparent motion. Psychol Rev 88:171–195
Wertheimer M (1923) Untersuchungen zur Lehre von der Gestalt, II. Psychologische Forschung 4:301–350
Kubovy M, Holcombe AO, Wagemans J (1998) On the lawfulness of grouping by proximity. Cogn Psychol 35:71–98
Koffka K (1935/1963) Principles of Gestalt psychology. A Harbinger Book, Harcourt, Brace and World, Inc., New York
Nakayama K (1985) Biological image motion processing: a review. Vis Res 25(5):625–660
Weiss Y, Simoncelli EP, Adelson EH (2002) Motion illusions as optimal percepts. Nat Neurosci 5(6):598–604
Longuet-Higgins HC, Prazdny, K (1981) The interpretation of a moving retinal image. Proc Roy Soc London. Ser B, Biol Sci 208(1173):385–397
Landy MS, Maloney LT, Johnsten E, Young M (1995) Measurement and modeling of depth cue combinations: in defense of weak fusion. Vis Res 35:389–412
Kelly DH (1979) Motion and vision II. Stabilized spatio-temporal threshold surface. J Opt Soc Am 69(10):1340–1349
Kelly DH (1994) Eye movements and contrast sensitivity. In: Kelly DH (ed) Visual science and engineering (models and applications). Marcel Dekker Inc, New York, pp 93–114
Gorban A, Pokidysheva L, Smirnova E, Tyukina T (2011) Law of the minimum paradoxes. Bull Math Biol 73(9):2013–2044
van Doorn AJ, Koenderink JJ (1982) Temporal properties of the visual detectability of moving spatial white noise. Exp Brain Res 45:179–188
van Doorn AJ, Koenderink JJ (1982) Spatial properties of the visual detectability of moving spatial white noise. Exp Brain Res 45:189–195
Laddis P, Lesmes LA, Gepshtein S, Albright TD (2011) Efficient measurement of spatiotemporal contrast sensitivity in human and monkey. In: 41st annual meeting of the society for neuroscience, vol 577.20
Gepshtein S, Lesmes LA, Albright TD (2013) Sensory adaptation as optimal resource allocation. Proc Natl Acad Sci, USA 110(11):4368–4373
Lesmes LA, Gepshtein S, Lu Z-L, Albright TD (2009) Rapid estimation of the spatiotemporal contrast sensitivity surface. J Vis 9(8):696
Sakitt B, Barlow HB (1982) A model for the economical encoding of the visual image in cerebral cortex. Biol Cybern 43:97–108
Laughlin SB (1989) The role of sensory adaptation in the retina. J Exp Biol 146(1):39–62
Wainwright MJ (1999) Visual adaptation as optimal information transmission. Vis Res 39:3960–3974
Laughlin SB, Sejnowski TJ (2003) Communication in neuronal networks. Science 301(5641):1870–1874
Clifford CWG, Wenderoth P (1999) Adaptation to temporal modulation can enhance differential speed sensitivity. Vis Res 39:4324–4332
Krekelberg B, Van Wezel RJ, Albright TD (2006) Adaptation in macaque MT reduces perceived speed and improves speed discrimination. J Neurophysiol 95:255–270
Gepshtein S (2009) Closing the gap between ideal and real behavior: scientific versus engineering approaches to normativity. Philos Psychol 22:61–75
Yeshurun Y, Carrasco M (1998) Attention improves or impairs visual performance by enhancing spatial resolution. Nature 396:72–75
Yeshurun Y, Carrasco M (2000) The locus of attentional effects in texture segmentation. Nat Neurosci 3(6):622–627
Jurica P, Gepshtein S, Tyukin I, Prokhorov D, van Leeuwen C (2007) Unsupervised adaptive optimization of motion-sensitive systems guided by measurement uncertainty. In: International conference on intelligent sensors, sensor networks and information, ISSNIP 2007, vol 3, Melbourne, Qld:179–184
Jurica P, Gepshtein S, Tyukin I, van Leeuwen C (2013) Sensory optimization by stochastic tuning. Psychol Rev 120(4):798–816
Hebb DO (1949) The organization of behavior. John Wiley, New York
Bienenstock EL, Cooper LN, Munr PW (1982) Theory for the development of neuron selectivity: orientation specificity and binocular interaction in visual cortex. J Neurosci 2:32–48
Paulsen O, Sejnowski TJ (2000) Natural patterns of activity and long-term synaptic plasticity. Curr Opin Neurobiol 10(2):172–180
Bi G, Poo M (2001) Synaptic modification by correlated activity: Hebb’s postulate revisited. Annu Rev Neurosci 24:139–166
Gardiner CW (1996) Handbook of stochastic methods: for physics, chemistry and the natural sciences. Springer, New York
Vergassola M, Villermaux E, Shraiman BI (2007) `Infotaxis’ as a strategy for searching without gradients. Nature 445:406–409
Gepshtein S, Jurica P, Tyukin I, van Leeuwen C, Albright TD (2010) Optimal sensory adaptation without prior representation of the environment. In: 40th Annual Meeting of the Society for Neuroscience, vol 731.7
von Neumann J (1963) Theory of games, astrophysics, hydrodynamics and meteorology. In: Taub AH (ed) Collected works, vol VI. Pergamon Press, New York
Shannon CE (1948) A mathematical theory of communication. Bell Syst Tech J 27:379–423, 623–656
Jaynes ET (1957) Information theory and statistical mechanics. Phys Rev 106:620–630
Gorban A (2013) Maxallent: maximizers of all entropies and uncertainty of uncertainty. Comput Math Appl 65(10):1438–1456
Cover TM, Thomas JA (2006) Elements of information theory. John Wiley, New York
Acknowledgments
This work was supported by the European Regional Development Fund, National Institutes of Health Grant EY018613, and Office of Naval Research Multidisciplinary University Initiative Grant N00014-10-1-0072.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Appendices
Appendices
7.1.1 Appendix 1. Additivity of Uncertainty
For the sake of simplicity, the following derivations concern the stimuli that can be modeled by integrable functions \( I:{\mathbb{R}} \to {\mathbb{R}} \) of one variable \( x \). Generalizations to functions of more than one variable are straightforward. Consider two quantities:
-
Stimulus location on \( x \), where \( x \) can be space or time, and the “location” indicates respectively “where” or “when” the stimulus has occurred.
-
Stimulus content on \( f_{x} \), where \( f_{x} \) can be spatial or temporal frequency of stimulus modulation.
Suppose a sensory system is equipped with many measuring devices (“sensors”), each used to estimate both stimulus location and frequency content from “image” (or “input”) \( I(x) \). Assume that the outcome of measurement is a random variable with probability density function \( p(x,f) \). Let
be the (marginal) means of \( p(x,f) \) on dimensions \( x \) and \( f_{x} \) (abbreviated as \( f \)).
It is sometimes assumed that sensory systems “know” \( p(x,f) \), which is not true in general. Generally, one can only know (or guess) some properties of \( p(x,f) \), such as its mean and variance. Reducing the chance of gross error due to the incomplete information about \( p(x,f) \) is accomplished by a conservative strategy: finding the minima on the function of maximal uncertainty, i.e., using a minimax approach [15, 16].
The minimax approach is implemented in two steps. The first step is to find such \( p_{x} (x) \) and \( p_{f} (f) \), for which measurement uncertainty is maximal. (The uncertainty is characterized conservatively, in terms of variance alone [2]). The second step is to find the condition(s), at which the function of maximal uncertainty has the smallest value: the minimax point(s).
Maximal uncertainty is evaluated using the well-established definition of entropy [58] (cf. [59, 60]):
According to the independence bound on entropy (Theorem 2.6.6 in [61]),
where
Therefore, the uncertainty of measurement cannot exceed
Eq. 7.14 is the “envelope” of maximal measurement uncertainty: a “worst-case” estimate.
By the Boltzmann theorem on maximum-entropy probability distributions [61], the maximal entropy of probability densities with fixed means and variances is attained, when the functions are Gaussian. Then, the maximal entropy is a sum of their variances [61] and
where \( \sigma_{x} \) and \( \sigma_{f} \) are the standard deviations. Then maximal entropy is
That is, when \( p(x,f) \) is unknown, and all one knows about marginal distributions \( p_{x} (x) \) and \( p_{f} (f) \) is their means and variances, the maximal uncertainty of measurement is the sum of variances of the estimates of \( x \) and \( f \). The following minimax step is to find the conditions of measurement, at which the sum of variances is the smallest.
7.1.2 Appendix 2. Improving Resolution by Multiple Sampling
How does an increased allocation of resources to a specific condition of measurement affect the (spatial or temporal) resolution at that condition? Consider set \( \varPsi \) of sampling functions
where \( \sigma \) is a scaling parameter and \( \delta \) is a translation parameter. For a broad class of functions \( \psi ( \cdot ) \), any element of \( \varPsi \) can be obtained by addition of weighted and shifted \( \psi (s) \). The following argument proves that any function from a sufficiently broad class that includes \( \psi (s\sigma + \delta ) \) can be represented by a weighted sum of translated replicas of \( \psi (s) \).
Let \( \psi^{ * } (s) \) be a continuous function that can be expressed as a sum of a converging series of harmonic functions:
For example, Gaussian sampling functions of arbitrary widths can be expressed as a sum of \( \cos ( \cdot ) \) and \( \sin ( \cdot ) \). Let us show that, if \( |\psi (s)| \) is Riemann-integrable, i.e., if
and its Fourier transform \( \widehat{\psi } \) does not vanish for all \( \omega \in {\mathbb{R}} \): \( \widehat{\psi }(\omega ) \ne 0 \) (i.e., its spectrum has no “holes”), then the following expansion of \( \psi^{ * } \) is possible:
where \( \varepsilon (s) \) is a residual that can be arbitrarily small. This goal is attained by proving identities
where \( c_{i,1} \), \( c_{i,2} \) and \( d_{i,1} \), \( d_{i,2} \) are real numbers, while \( \varepsilon_{1} (s) \) and \( \varepsilon_{2} (s) \) are arbitrarily small residuals.
First, write the Fourier transform of \( \psi (s) \) as
and multiply both sides of the above expression by \( e^{{i\omega_{0} \upsilon }} \):
Change the integration variable:
such that Eq. 7.18 transforms into
Notice that \( \widehat{\psi }(\omega ) = a(\omega ) + ib(\omega ) \). Hence
and
Since \( \widehat{\psi }(\omega ) \ne 0 \) is assumed for all \( \omega \in {\mathbb{R}} \), then \( a(\omega ) + ib(\omega ) \ne 0 \). In other words, either \( a(\omega ) \ne 0 \) or \( b(\omega ) \ne 0 \) should hold. For example, suppose that \( a(\omega ) \ne 0. \) Then
Therefore,
Because function \( \psi (s) \) is Riemann-integrable, the integrals in Eq. 7.19 can be approximated as
where \( \;x_{k} \) and \( x_{p} \) are some elements of \( {\mathbb{R}} \).
From Eqs. 7.19–7.21 it follows that
Given that \( \widehat{\psi }(\omega ) \ne 0 \) for all \( \omega \) and letting \( \omega = \omega_{0} \), it follows that
where
An analogue of Eq. 7.22 for \( \sin (\omega_{0} \upsilon ) \) follows from \( \sin (\omega_{0} \upsilon ) = \cos (\omega_{0} \upsilon + \pi /2) \). This completes the proof of Eq. 7.17 and hence of Eq. 7.16.
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Gepshtein, S., Tyukin, I. (2015). Optimal Measurement of Visual Motion Across Spatial and Temporal Scales. In: Favorskaya, M., Jain, L. (eds) Computer Vision in Control Systems-1. Intelligent Systems Reference Library, vol 73. Springer, Cham. https://doi.org/10.1007/978-3-319-10653-3_7
Download citation
DOI: https://doi.org/10.1007/978-3-319-10653-3_7
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-10652-6
Online ISBN: 978-3-319-10653-3
eBook Packages: EngineeringEngineering (R0)