Implementation of Emotional-Aware Computer Systems Using Typical Input Devices

  • Kaveh Bakhtiyari
  • Mona Taghavi
  • Hafizah Husain
Part of the Lecture Notes in Computer Science book series (LNCS, volume 8397)

Abstract

Emotions play an important role in human interactions. Human Emotions Recognition (HER - Affective Computing) is an innovative method for detecting user’s emotions to determine proper responses and recommendations in Human-Computer Interaction (HCI). This paper discusses an intelligent approach to recognize human emotions by using the usual input devices such as keyboard, mouse and touch screen displays. This research is compared with the other usual methods like processing the facial expressions, human voice, body gestures and digital signal processing in Electroencephalography (EEG) machines for an emotional-aware system. The Emotional Intelligence system is trained in a supervised mode by Artificial Neural Network (ANN) and Support Vector Machine (SVM) techniques. The result shows 93.20% in accuracy which is around 5% more than the existing methods. It is a significant contribution to show new directions of future research in this topical area of emotion recognition, which is useful in recommender systems.

Keywords

human emotion recognition keyboard keystroke dynamics mouse movement touch-screen human computer interaction affective computing 

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Cowie, R., Douglas-Cowie, E., Tsapatsoulis, N., Votsis, G., Kollias, S., Fellenz, W., Taylor, J.G.: Emotion recognition in human-computer interaction. IEEE Signal Processing Magazine 18(1), 32–80 (2001)CrossRefGoogle Scholar
  2. 2.
    Ekman, P., Friesen, W.V.: Unmasking the face: A guide to recognizing emotions from facial clues. Malor Books (2003)Google Scholar
  3. 3.
    Konar, A., Chakraborty, A., Halder, A., Mandal, R., Janarthanan, R.: Interval Type-2 Fuzzy Model for Emotion Recognition from Facial Expression. In: Kundu, M.K., Mitra, S., Mazumdar, D., Pal, S.K. (eds.) PerMIn 2012. LNCS, vol. 7143, pp. 114–121. Springer, Heidelberg (2012)CrossRefGoogle Scholar
  4. 4.
    Huang, T.: Audio-visual human computer interface. In: IEEE International Symposium on Consumer Electronics. IEEE, University of Illinois (2008)Google Scholar
  5. 5.
    Monrose, F., Rubin, A.D.: Keystroke dynamics as a biometric for authentication. Future Generation Computer Systems 16(4), 351–359 (2000)CrossRefGoogle Scholar
  6. 6.
    Chang, M.: Iowa State engineers use keyboard, mouse and mobile device ‘fingerprints’ to protect data (November 18, 2013), http://www.news.iastate.edu/news/2013/11/18/fingerprints
  7. 7.
    Schuller, B., Rigoll, G., Lang, M.: Emotion recognition in the manual interaction with graphical user interfaces. In: IEEE International Conference on Multimedia and Expo. IEEE (2004)Google Scholar
  8. 8.
    Epp, C., Lippold, M., Mandryk, R.L.: Identifying emotional states using keystroke dynamics. In: Proceedings of the 2011 Annual Conference on Human Factors in Computing Systems, pp. 715–724. ACM, Vancouver (2011)CrossRefGoogle Scholar
  9. 9.
    Kemp, F.: Applied Multiple Regression/Correlation Analysis for the Behavioral Sciences. Journal of the Royal Statistical Society: Series D (The Statistician) 52(4), 691–691 (2003)Google Scholar
  10. 10.
    Cohen, J.: Applied multiple regression/correlation analysis for the behavioral sciences. Lawrence Erlbaum Associates, Hillsdale (2003)Google Scholar
  11. 11.
    Liu, Y., Sourina, O., Nguyen, M.K.: Real-time EEG-based Human Emotion Recognition and Visualization. In: International Conference on Cyberworlds. IEEE, Alberta (2010)Google Scholar
  12. 12.
    Schaaff, K., Schultz, T.: Towards emotion recognition from electroencephalographic signals. In: Affective Computing and Intelligent Interaction and Workshops (ACII). IEEE, Memphis (2009)Google Scholar
  13. 13.
    Li, H., Pang, N., Guo, S., Wang, H.: Research on textual emotion recognition incorporating personality factor. In: IEEE International Conference on Robotics and Biomimetics. IEEE, Sanya (2008)Google Scholar
  14. 14.
    Amarakeerthi, S., Ranaweera, R., Cohen, M.: Speech-Based Emotion Characterization Using Postures and Gestures in CVEs. In: International Conference on Cyberworlds. IEEE, Alberta (2010)Google Scholar
  15. 15.
    Xiao, Z., Dellandrea, E., Dou, W., Chen, L.: Automatic hierarchical classification of emotional speech. In: Ninth IEEE International Symposium on Multimedia Workshops (2007)Google Scholar
  16. 16.
    Gunes, H., Piccardi, M.: Fusing face and body gesture for machine recognition of emotions. In: IEEE International Workshop on Robots and Human Interactive Communication. IEEE (2005)Google Scholar
  17. 17.
    Milanova, M., Sirakov, N.: Recognition of Emotional states in Natural Human-Computer Interaction. In: IEEE International Symposium on Signal Processing and Information Technology (ISSPIT). IEEE, Ho Chi Minh City (2008)Google Scholar
  18. 18.
    Kao, C.Y., Fahn, C.S.: A Design of Face Detection and Facial Expression Recognition Techniques Based on Boosting Schema. Applied Mechanics and Materials 121, 617–621 (2012)Google Scholar

Copyright information

© Springer International Publishing Switzerland 2014

Authors and Affiliations

  • Kaveh Bakhtiyari
    • 1
    • 2
  • Mona Taghavi
    • 1
  • Hafizah Husain
    • 1
  1. 1.Department of Electrical, Electronics and Systems Engineering, Faculty of Engineering and Built EnvironmentUniversiti Kebangsaan Malaysia (The National University of Malaysia)Selangor Darul EhsanMalaysia
  2. 2.Department of Computer & Cognitive Science, Faculty of EngineeringUniversity of Duisburg-EssenDuisburgGermany

Personalised recommendations