Abstract
We propose a biologically motivated motion analysis model using a dynamic bottom-up saliency map model and a neural network for motion analysis of which the input is an optical flow. The dynamic bottom-up saliency map model can generate a human-like visual scan path by considering dynamics of continuous input scenes as well as saliency of the primitive features of a static input scene. Neural network for motion analysis responds selectively to rotation, expansion, contraction and planar motion of the optical flow in a selected area. The experimental results show that the proposed model can generate effective motion analysis results for analyzing only an interesting area instead of considering the whole input scenes, which makes faster analysis mechanism for dynamic input scenes.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Giboson, J.J.: The perception of the visual world. Boston, Houghton Mifflin (1950)
Duffy, C.J., Wurtz, R.H.: Sensitivity of MST neurons to optic flow stimuli. I. A continuum of response selectivity to large-field stimuli. Journal of Neurophysiology 65(6), 1329–1345 (1991)
Duffy, C.J., Wurtz, R.H.: Sensitivity of MST neurons to optic flow stimuli. II. Mechanisms of response selectivity revealed by small-field stimuli. Journal of Neurophysiology 65(6), 1346–1359 (1991)
Kazuya, T., Fukushima, K.: Neural network model for extracting optic flow. Neural Networks 18(5-6), 1–8 (2005)
Taylor, N.R., Hartley, M., Taylor, J.G.: Coding of objects in low-level visual cortical area. In: Duch, W., Kacprzyk, J., Oja, E., Zadrożny, S. (eds.) ICANN 2005. LNCS, vol. 3696, pp. 57–63. Springer, Heidelberg (2005)
Park, S.J., An, K.H., Lee, M.: Saliency map model with adaptive masking based on independent component analysis. Neurocomputing 49, 417–422 (2002)
Barlow, H.B., Tolhust, D.J.: Why do you have edge detection? Optical Society of America Technical Digest 23, 172 (1992)
Bell, J., Sejnowski, T.J.: The independent components of natural scenes are edge filters. Vision Research 37, 3327–3338 (1997)
Lanyon, L.J., Denham, S.L.: A model of active visual search with object-based attention guiding scan paths. Neural Networks 17(5-6), 873–897 (2004)
Kadir, T., Brady, M.: Scale, saliency and image description. International Journal of Computer Vision 45(2), 83–105 (1996)
Horn, B.K.P., Schunck, B.G.: Determining optical flow. Artificial Intelligence 17, 185–203 (1981)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Lee, I., Ban, SW., Fukushima, K., Lee, M. (2006). Selective Motion Analysis Based on Dynamic Visual Saliency Map Model. In: Rutkowski, L., Tadeusiewicz, R., Zadeh, L.A., Żurada, J.M. (eds) Artificial Intelligence and Soft Computing – ICAISC 2006. ICAISC 2006. Lecture Notes in Computer Science(), vol 4029. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11785231_85
Download citation
DOI: https://doi.org/10.1007/11785231_85
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-35748-3
Online ISBN: 978-3-540-35750-6
eBook Packages: Computer ScienceComputer Science (R0)