Classification of Visual Attention Level During Target Gazing Using Microsaccades

  • Soichiro YokooEmail author
  • Nobuyuki Nishiuchi
  • Kimihiro Yamanaka
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 11960)


In the previous researches on microsaccades, a typical and basic experimental method involves the following: a dot is displayed as a visual target in the middle or near the middle of a monitor, and eye movements of the subject are measured, and then the number of microsaccades is extracted from the measured eye movements. However, it is difficult to determine whether or not the subject is paying visual attention while gazing at the target, and the degree of visual attention paid by the subject. This paper proposes a system that uses microsaccades to classify visual attention levels during visual target gazing. In our experiment, ten subjects performed three tasks requiring different levels of visual attention. Microsaccades were measured and the number of microsaccades was extracted for each task. Statistical analysis showed that the number of microsaccades differed among the tasks. Our results suggest that visual attention level can be classified by the number of microsaccades.


Microsaccade Visual attention Eye tracking 



We would like to thank Dr. Takao Fukui at Tokyo Metropolitan University for comments that greatly improved the manuscript.


  1. 1.
    Ukai, K.: Eye movement: characteristics and method of measurement. Japan. J. Opt. 23, 2–8 (1994)Google Scholar
  2. 2.
    Pritchard, R.M.: Stabilized images on the retina. Sci. Am. 204, 72–78 (1961)CrossRefGoogle Scholar
  3. 3.
    Kohama, T., Shinkai, K., Usui, S.: Quantitatively measuring visual attention by analyzing microsaccades. J. Inst. Image Inf. Telev. Eng. 52(4), 571–576 (1998)Google Scholar
  4. 4.
    Zaid, M., James, J.C.: Microsaccades as an overt measure of covert attention shifts. Vis. Res. 42(22), 2533–2545 (2002)CrossRefGoogle Scholar
  5. 5.
    Handa, T., Kohama, T.: Inhibition of microsaccade rate caused by focus of visual attention. In: Memoirs of the Faculty of Biology-Oriented Science and Technology of Kinki University, vol. 29, pp. 7–15 (2012)Google Scholar
  6. 6.
    Ohtani, S., Kohama, T., Yoshikawa., Yoshida., H.: Influence of target patterns on dynamic characteristics of fixation eye movements, IEICE Technical report, Institute of Electronics, Information and Communication Engineers, HIP2016-42-60, vol. 113, no. 229, pp. 61–66 (2016)Google Scholar
  7. 7.
    Kanoh, Y., Kohama, T., Kikkawa, S., Yoshida, H.: The relationship between spatial extent of focal attention and latencies of microsaccades. Trans. Japan. Soc. Med. Biol. Eng. 53(O1-05-5) (2014)Google Scholar
  8. 8.
    Engbert, R., Kliegl, R.: Microsaccades uncover the orientation of covert attention. Vision. Res. 43(9), 1035–1045 (2003)CrossRefGoogle Scholar
  9. 9.
    Laubrock, J., Engbert, R., Kliegl, R.: Microsaccade dynamics during covert attention. Vis. Res. 45(6), 721–730 (2005)CrossRefGoogle Scholar
  10. 10.
    Takahashi, M.: A study on covert attention using microsaccadeas an index: introducing anticipatory response tasks toward tennis serve, Kyushu Institute of Technology University, 2018, Ph.D. thesis (2018)Google Scholar
  11. 11.
    Kashiwara, K., Okanoya, K., Kawai, N.: Effects of emotional pictures on eye movements: IEICE technical report, Human communication science, vol. 110, no. 34, pp. 41–46, HIP2010-8 (2015)Google Scholar
  12. 12.
    Sakaino, H., Tetutani, N., Kishino, F.: Spatio-temporal analysis of eye movement patterns based on the maximum entropy method and the higher moment features. IEICE Trans. Fundam. Electron. Commun. Comput. Sci. (Jpn Ed.) 76(8), 1027–1041 (1993)Google Scholar
  13. 13.
    Carpenter, R.H.S.: Movements of the Eyes (2nd Edn. Revised and Enlarged), pp. 124–138. Pion London (1988)Google Scholar
  14. 14.
    Hoffman, J.E., Subramaniam, B.: The role of visual attention in saccadic eye movements. Percept. Psychophys. 57(6), 787–795 (1995)CrossRefGoogle Scholar
  15. 15.
    Deubel, H., Schneider, W.X.: Saccade target selection and object recognition: evidence for a common attentional mechanism. Vis. Res. 36(12), 1827–1837 (1996)CrossRefGoogle Scholar
  16. 16.
    Yoshimatsu, H.: Smoothing of miniature eye movement using wavelet analysis when gazing a fixed target. J. Inst. Telev. Eng. Jp. 50(12), 1903–1912 (1996)Google Scholar
  17. 17.
    Noguchi, D., Kohama, T., Yoshikawa, S., Yoshida, H.: Microsaccade detection with a order-statistic low-pass differentiation filter. In: ITE Winter Annual Convention 2011, 12-2-1 (2011)Google Scholar
  18. 18.
    Emoto, J., Hirata, Y.: Realtime microsaccade detection with convolutional neural network. IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 5(2), 456–467 (2018)Google Scholar
  19. 19.
    Miyatake, K., Kohama, T., Yoshida, H.: A microsaccade detection method using a recursive neural network. In: Abstracts and Proceedings of the Annual Conference of Japanese Society for Medical and Biological Engineering, vol. 57, no. O2-7-3-2 (2018)Google Scholar
  20. 20.
    Suzuki, K., Toyoda, H., Hanayama, R., Ishii, K.: Development of binocular microsaccade measurement system using intelligent vision sensor and evaluation of synchronization between left and right microsaccades. Trans. Japan. Soc. Med. Biol. Eng. 53(5), 99.247–99.254 (2015)Google Scholar
  21. 21.
    Fukuda, T.: The functional difference between central vision and peripheral vision in pattern perception. J. Inst. Telev. Eng. Jpn. 32(9), 492–498 (1978)Google Scholar
  22. 22.
    Martinez-Conde, S., Macknik, S.L., Hubel, D.H.: The role of fixational eye movements in visual perception. Nat. Rev. Neurosci. 5, 229–240 (2004)CrossRefGoogle Scholar
  23. 23.
    Zuber, B.L., Stark, L., Cook, G.: Microsaccades and the velocity-amplitude relationship for saccadic eye movements. Science 50, 1459–1460 (1965)CrossRefGoogle Scholar
  24. 24.
    Ueda, T.: Soukan ga aruka wo mitsukeru kanbenhou, Operations research as a management science Communications of the Operations Research Society of Japan, vol. 42, no. 7, pp. 99.493–496 (1997)Google Scholar

Copyright information

© Springer-Verlag GmbH Germany, part of Springer Nature 2020

Authors and Affiliations

  • Soichiro Yokoo
    • 1
    Email author
  • Nobuyuki Nishiuchi
    • 1
  • Kimihiro Yamanaka
    • 2
  1. 1.Graduate School of Systems Design, Faculty of Computer ScienceTokyo Metropolitan UniversityHinoJapan
  2. 2.Graduate School of Natural Science, Faculty of Intelligence and InformaticsKonan UniversityKobeJapan

Personalised recommendations