Abstract
Affective models based on EEG signals have been proposed in recent years. However, most of these models require subject-specific training and generalize worse when they are applied to new subjects. This is mainly caused by the individual differences across subjects. While, on the other hand, it is time-consuming and high cost to collect subject-specific training data for every new user. How to eliminate the individual differences in EEG signals for implementation of affective models is one of the challenges. In this paper, we apply Deep adaptation network (DAN) to solve this problem. The performance is evaluated on two publicly available EEG emotion recognition datasets, SEED and SEED-IV, in comparison with two baseline methods without domain adaptation and several other domain adaptation methods. The experimental results indicate that the performance of DAN is significantly superior to the existing methods.
H. Li and Y.-M. Jin—The first two authors contributed equally to this work.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
The SEED and SEED-IV datasets are available at http://bcmi.sjtu.edu.cn/~seed/index.html.
References
Chai, X., et al.: A fast, efficient domain adaptation technique for cross-domain electroencephalography (EEG)-based emotion recognition. Sensors 17(5), 1014 (2017)
Chai, X., Wang, Q., Zhao, Y., Liu, X., Bai, O., Li, Y.: Unsupervised domain adaptation techniques based on auto-encoder for non-stationary EEG-based emotion recognition. Comput. Biol. Med. 79, 205–214 (2016)
Chai, X., Wang, Q., Zhao, Y., Liu, X., Liu, D., Bai, O.: Multi-subject subspace alignment for non-stationary EEG-based emotion recognition. Technol. Health Care 26, 1–9 (2018)
Daniela, S., Maren, G., Thomas, F., Stefan, K.: Music and emotion: electrophysiological correlates of the processing of pleasant and unpleasant music. Psychophysiology 44(2), 293–304 (2007)
Duan, R., Zhu, J., Lu, B.: Differential entropy feature for EEG-based emotion classification. In: International IEEE/EMBS Conference on Neural Engineering, pp. 81–84. IEEE Press, San Diego (2013)
Ganin, Y., Lempitsky, V.: Unsupervised domain adaptation by backpropagation. In: International Conference on Machine Learning, vol. 37, pp. 1180–1189. PMLR, Lille (2015)
Jin, Y.M., Luo, Y.D., Zheng, W.L., Lu, B.L.: EEG-based emotion recognition using domain adaptation network. In: International Conference on Orange Technologies, Singapore, pp. 222–225 (2017)
Knyazev, G.G., Slobodskoj-Plusnin, J.Y., Bocharov, A.V.: Gender differences in implicit and explicit processing of emotional facial expressions as revealed by event-related theta synchronization. Emotion 10(5), 678–687 (2010)
Lan, Z., Sourina, O., Wang, L., Scherer, R., Müller-Putz, G.R.: Domain adaptation techniques for EEG-based emotion recognition: a comparative study on two public datasets. IEEE Trans. Cogn. Dev. Syst. 1 (2018)
Lin, Y.P., Jung, T.P.: Improving EEG-based emotion classification using conditional transfer learning. Front. Hum. Neurosci. 11, 334 (2017)
Long, M., Cao, Y., Wang, J., Jordan, M.: Learning transferable features with deep adaptation networks. In: International Conference on Machine Learning, vol. 37, pp. 97–105. PMLR, Lille (2015)
Mathersul, D., Williams, L.M., Hopkinson, P.J., Kemp, A.H.: Investigating models of affect: relationships among EEG alpha asymmetry, depression, and anxiety. Emotion 8(4), 560–572 (2008)
Mühl, C., Allison, B., Nijholt, A., Chanel, G.: A survey of affective brain computer interfaces: principles, state-of-the-art, and challenges. Brain-Comput. Interfaces 1(2), 66–84 (2014)
Pan, S.J., Tsang, I.W., Kwok, J.T., Yang, Q.: Domain adaptation via transfer component analysis. IEEE Trans. Neural Netw. 22(2), 199–210 (2011)
Pan, S.J., Yang, Q.: A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 22(10), 1345–1359 (2010)
Sangineto, E., Zen, G., Ricci, E., Sebe, N.: We are not all equal: personalizing models for facial expression analysis with transductive parameter transfer. In: ACM International Conference on Multimedia, pp. 357–366. ACM Press, New York (2014)
Schölkopf, B., Smola, A., Müller, K.-R.: Kernel principal component analysis. In: Gerstner, W., Germond, A., Hasler, M., Nicoud, J.-D. (eds.) ICANN 1997. LNCS, vol. 1327, pp. 583–588. Springer, Heidelberg (1997). https://doi.org/10.1007/BFb0020217
Wang, X.W., Nie, D., Lu, B.L.: Emotional state classification from eeg data using machine learning approach. Neurocomputing 129, 94–106 (2014)
Zheng, W., Lu, B.: Investigating critical frequency bands and channels for eeg-based emotion recognition with deep neural networks. IEEE Trans. Auton. Ment. Dev. 7(3), 162–175 (2015)
Zheng, W.L., Liu, W., Lu, Y., Lu, B.L., Cichocki, A.: Emotionmeter: a multimodal framework for recognizing human emotions. IEEE Trans. Cybern. 99, 1–13 (2018)
Zheng, W.L., Lu, B.L.: Personalizing EEG-based affective models with transfer learning. In: International Joint Conference on Artificial Intelligence, pp. 2732–2738. AAAI Press, New York (2016)
Acknowledgement
This work was supported in part by the grants from the National Key Research and Development Program of China (Grant No. 2017YF-B1002501), the National Natural Science Foundation of China (Grant No. 6167-3266), and the Fundamental Research Funds for the Central Universities.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Li, H., Jin, YM., Zheng, WL., Lu, BL. (2018). Cross-Subject Emotion Recognition Using Deep Adaptation Networks. In: Cheng, L., Leung, A., Ozawa, S. (eds) Neural Information Processing. ICONIP 2018. Lecture Notes in Computer Science(), vol 11305. Springer, Cham. https://doi.org/10.1007/978-3-030-04221-9_36
Download citation
DOI: https://doi.org/10.1007/978-3-030-04221-9_36
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-04220-2
Online ISBN: 978-3-030-04221-9
eBook Packages: Computer ScienceComputer Science (R0)