An EEG-Based Emotion Recognition Model with Rhythm and Time Characteristics

  • Jianzhuo Yan
  • Sinuo DengEmail author
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 11309)


As a senior function of human brain, emotion has a great influence on human study, work, and all aspects of life. Correctly recognizing human emotion can make artificial intelligence serve human being better. EEG-based emotion recognition (ER) has become more popular in these years, which is one of the utilizations of Brain Computer Interface (BCI). However, due to the ambiguity of human emotions and the complexity of EEG signals, the EEG-ER system which can recognize emotions with high accuracy is not easy to achieve. In this paper, based on the time scale, we choose recurrent neural network as the breakthrough point of the screening model. And according to the rhythmic characteristics and temporal memory characteristics of EEG, we propose a Rhythmic Time EEG Emotion Recognition Model (RT-ERM) based on the valance and arousal of LSTM. When using this model, the classification results of different rhythms and time scales are different. Through the results of the classification accuracy of different rhythms and different time scales, the optimal rhythm and time scale of the RT-ERM model are obtained, and the classification of emotional EEG is carried out by the best time scales corresponding to different rhythms, and we found some interesting phenomena. Finally, by comparing with other existing emotional EEG classification methods, it is found that the rhythm and time scale of the model can provide a good accuracy rate for RT-ERM.


EEG Emotion recognition Rhythm and time characteristics LSTM 



This work is supported by the CERNET Innovation Project (No. NGII20170719). We sincerely thank Hongzhi Kuai for the helpful discussion on the experiment design and the equipment support.


  1. 1.
    Khosrowabadi, R., Wahab, A., Ang, K.K., Baniasad, M.H.: Affective computation on EEG correlates of emotion from musical and vocal stimuli. In: Proceedings of the IJCNN, pp. 1168–1172. IEEE Press, Atlanta (2009)Google Scholar
  2. 2.
    Esslen, M., Pascual, M.R.D., Kochi, K., et al.: Brain areas and time course of emotional processing. Neuroimage 21(4), 1189–1203 (2014)CrossRefGoogle Scholar
  3. 3.
    Yoon, H.J., Chung, S.Y.: EEG-based emotion estimation using Bayesian weighted-log-posterior function and perceptron convergence algorithm. Comput. Biol. Med. 43(12), 2230–2237 (2013)CrossRefGoogle Scholar
  4. 4.
    Hopfield, J.J.: Neural networks and physical systems with emergent collective computational abilities. Proc. Natl. Acad. Sci. 79(8), 2554–2558 (1982)MathSciNetCrossRefGoogle Scholar
  5. 5.
    Schuster, M., Paliwal, K.K.: Bidirectional recurrent neural networks. Signal Process. 45(11), 2673–2681 (1997)Google Scholar
  6. 6.
    Jordan, M.I.: Attractor Dynamics and Parallelism in a Connectionist Sequential Machine, pp. 112–127. IEEE Press, Amherst (1990)Google Scholar
  7. 7.
    Elman, J.L.: Finding structure in time. Cogn. Sci. 14(2), 179–211 (1990)CrossRefGoogle Scholar
  8. 8.
    Lang, K.J., Waibel, A.H., Hinton, G.E.: A time-delay neural network architecture for isolated word recognition. Neural Netw. 3(1), 23–43 (1990)CrossRefGoogle Scholar
  9. 9.
    Jaeger, H.: The “Echo State” Approach to Analysing and Training Recurrent Neural Networks. überwachtes lernen (2001)Google Scholar
  10. 10.
    Hochreiter, S.: Untersuchungen zu Dynamischen Neuronalen Netzen. PhD thesis, Institut für Informatik, Technische Universität München (1991)Google Scholar
  11. 11.
    Hochreiter, S., Schmidhuber, J.: Long Short-Term Memory. Neural Comput. 9(8), 1735–1780 (1997)CrossRefGoogle Scholar
  12. 12.
    Gers, F.A., Schmidhuber, J.: LSTM Recurrent Networks Learn Simple Context Free and Context Sensitive Languages. IEEE Trans. Neural Netw. 12(6), 1333–1340 (2001)CrossRefGoogle Scholar
  13. 13.
    Gers, F., Schraudolph, N., Schmidhuber, J.: Learning Precise Timing with LSTM Recurrent Networks. J. of Mach. Learn. Res. 3, 115–143 (2002)MathSciNetzbMATHGoogle Scholar
  14. 14.
    Hochreiter, S., Heusel, M., Obermayer, K.: Fast model-based protein homology detection without alignment. Bioinformatics 23(14), 1728 (2007)CrossRefGoogle Scholar
  15. 15.
    Eck, D., Schmidhuber, J.: Finding temporal structure in music: blues improvisation with LSTM Recurrent Networks. In: Proceedings of the 12th IEEE Workshop on Neural Networks for Signal Processing, pp. 747–756. IEEE, Martigny (2002)Google Scholar
  16. 16.
    Graves, A., Schmidhuber, J.: Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. 18(5–6), 602–610 (2005b)CrossRefGoogle Scholar
  17. 17.
    Koelstra, S., et al.: Deap: a database for emotion analysis using physiological signals. IEEE Trans. Affect. Comput. 3(1), 18–31 (2012)CrossRefGoogle Scholar
  18. 18.
    Rozgić, V., Vitaladevuni, S.N., Prasad, R.: Robust EEG emotion classification using segment level decision fusion. In: IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 1286–1290. IEEE, Vancouver (2013)Google Scholar
  19. 19.
    Zhuang, X., Rozgic, V., Crystal, M.: Compact unsupervised EEG response representation for emotion recognition. In: IEEE-EMBS International Conference on Biomedical and Health Informatics (BHI), pp. 736–739. IEEE, Valencia (2014)Google Scholar
  20. 20.
    Sepideh, H., Keivan, M., Motie, N.A.: The Emotion Recognition System Based on Autoregressive Model and Sequential Forward Feature Selection of Electroencephalogram Signals. J. Med. Signals Sens. 4(3), 194–201 (2014)Google Scholar

Copyright information

© Springer Nature Switzerland AG 2018

Authors and Affiliations

  1. 1.Faculty of Information TechnologyBeijing University of TechnologyBeijingChina
  2. 2.Beijing Advanced Innovation Center for Future Internet TechnologyBeijing University of TechnologyBeijingChina
  3. 3.Engineering Research Center of Digital CommunityBeijing University of TechnologyBeijingChina

Personalised recommendations