Skip to main content

Towards Context-Dependence Eye Movements Prediction in Smart Meeting Rooms

  • Conference paper
Artificial Neural Networks and Machine Learning – ICANN 2014 (ICANN 2014)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 8681))

Included in the following conference series:

Abstract

Being able to predict gaze locations, as compared to only measuring them, is desirable in many systems such as the design of web pages and commercials adaptive user interfaces, interactive visualization, or attention management systems. However, accurately predicting eye movements remains a challenging problem. In this paper, we present the results of experimental study to improve the prediction of saliency maps in smart meeting rooms. More specifically, we investigate meeting scenarios in terms of their context-dependence saliency based on different image features. We have recorded the center of gaze of users in meeting rooms in different scenarios (giving a talk, listening). We then used a data-driven approach to find out which features are important in each scenario. We found that the predictions differ according to the type of features we selected. Most interestingly, we found that models trained on face features perform better than the models trained on other features in the giving a talk scenario, but in the listening scenario the models trained on competing saliency features from Itti and Koch perform better than the models trained on another features. This finding points towards including context information about the scene and situation into the computation of saliency maps as important towards developing models of eye movements, which operate well under natural conditions such as those encountered in ubiquitous computing settings.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Roda, C.: Human Attention in Digital Environments. Cambridge University Press, Cambridge (2011)

    Book  Google Scholar 

  2. Wellner, P., Flynn, M., Guillemot, M.: Browsing recorded meetings with ferret. In: Bengio, S., Bourlard, H. (eds.) MLMI 2004. LNCS, vol. 3361, pp. 12–21. Springer, Heidelberg (2005)

    Chapter  Google Scholar 

  3. Ellis, C.S., Barthelmess, P.: The neem dream. In: Proceedings of the 2003 Conference on Diversity in Computing, TAPIA 2003, pp. 23–29. ACM, New York (2003)

    Google Scholar 

  4. Kleinbauer, T., Becker, S., Becker, T.: T.: Combining multiple information layers for the automatic generation of indicative meeting abstracts. In: Proc. of ENLG 2007 (2007)

    Google Scholar 

  5. McCowan, L., Gatica-Perez, D., Bengio, S., Lathoud, G., Barnard, M., Zhang, D.: Automatic analysis of multimodal group actions in meetings. IEEE Transactions on Pattern Analysis and Machine Intelligence 27(3), 305–317 (2005)

    Article  Google Scholar 

  6. Favre, S., Salamin, H., Vinciarelli, A., Hakkani Tür, D., Garg, N.P.: Role recognition for meeting participants: an approach based on lexical information and social network analysis. In: ACM International Conference on Multimedia (October 2008)

    Google Scholar 

  7. Itti, L., Koch, C., Niebur, E.: A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence 20(11), 1254–1259 (1998)

    Article  Google Scholar 

  8. Mahadevan, V., Vasconcelos, N.: Spatiotemporal saliency in dynamic scenes. IEEE Trans. Pattern Anal. Mach. Intell. 32(1), 171–177 (2010)

    Article  Google Scholar 

  9. Gao, D., Vasconcelos, N.: Discriminant saliency for visual recognition from cluttered scenes. In: NIPS (2004)

    Google Scholar 

  10. Judd, T., Ehinger, K., Durand, F., Torralba, A.: Learning to predict where humans look. In: ICCV (2009)

    Google Scholar 

  11. Yarbus, A.: Eye-movements and vision. Plenum Press, New York (1967)

    Book  Google Scholar 

  12. Simoncelli, E.P., Freeman, W.T.: The steerable pyramid: A flexible architecture for multi-scale derivative computation. In: IEEE Intl Conf. on Image Processing, pp. 444–447. IEEE Signal Processing Society (1995)

    Google Scholar 

  13. Torralba, A.: Modeling global scene factors in attention. JOSA - A 20, 1407–1418 (2003)

    Article  Google Scholar 

  14. Viola, P., Jones, M.J.: Robust real-time face detection. Int. J. Comput. Vision 57(2), 137–154 (2004)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2014 Springer International Publishing Switzerland

About this paper

Cite this paper

Mohammed, R.A.A., Schwabe, L., Staadt, O. (2014). Towards Context-Dependence Eye Movements Prediction in Smart Meeting Rooms. In: Wermter, S., et al. Artificial Neural Networks and Machine Learning – ICANN 2014. ICANN 2014. Lecture Notes in Computer Science, vol 8681. Springer, Cham. https://doi.org/10.1007/978-3-319-11179-7_32

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-11179-7_32

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-11178-0

  • Online ISBN: 978-3-319-11179-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics