Abstract
While music is widely understood to induce an emotional response in the listener, the exact nature of that response and its neural correlates are not yet fully explored. Furthermore, the large number of features which may be extracted from, and used to describe, neurological data, music stimuli, and emotional responses, means that the relationships between these datasets produced during music listening tasks or the operation of a brain–computer music interface (BCMI) are likely to be complex and multidimensional. As such, they may not be apparent from simple visual inspection of the data alone. Machine learning, which is a field of computer science that aims at extracting information from data, provides an attractive framework for uncovering stable relationships between datasets and has been suggested as a tool by which neural correlates of music and emotion may be revealed. In this chapter, we provide an introduction to the use of machine learning methods for identifying neural correlates of musical perception and emotion. We then provide examples of machine learning methods used to study the complex relationships between neurological activity, musical stimuli, and/or emotional responses.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsNotes
- 1.
Please refer to Chap. 2 for an introduction to EEG electrode placement systems.
References
Albrecht R, Ewing S (1989) Standardizing the administration of the profile of mood states (POMS): development of alternative word lists. J Pers Assess 53(1):31–39
Aloise F, Schettini F, Aricó P et al (2012) A comparison of classification techniques for a gaze-independent P300-based brain–computer interface. J Neural Eng 9(4):045012
Alpaydin E (2004) Introduction to machine learning. MIT Press, Cambridge
Bradley MM, Lang PJ (1994) Measuring emotion: the self-assessment manikin and the semantic differential. J Behav Ther Experim Psychiatry 25(1):49–59
Bradley MM, Lang PJ, Margaret M et al (2007) The international affective digitized sounds affective ratings of sounds and instruction manual. Technical report B-3, University of Florida, Gainesville, Fl
Christensen T (2002) The Cambridge history of western music theory. Cambridge University Press, Cambridge
Comon P (1994) Independent component analysis, a new concept? Sig Process 36(3):287–314
Cong F, Alluri V, Nandi AK et al (2013) Linking brain responses to naturalistic music through analysis of ongoing EEG and stimulus features. IEEE Trans Multimedia 15(5):1060–1069
Cornelius RR (1996) The science of emotion. Prentice Hall, Upper Saddle River
Cowie R, Douglas-Cowie E (2000) “FEELTRACE”: an instrument for recording perceived emotion in real time. In: Proceedings of the ISCA workshop on speech and emotion: a conceptual framework for research, Belfast, pp 19–24
Daly I, Nasuto SJ, Warwick K (2011) Single tap identification for fast BCI control. Cogn Neurodyn 5(1):21–30
Daly I, Nasuto SJ, Warwick K (2012) Brain computer interface control via functional connectivity dynamics. Pattern Recogn 45(6):2123–2136
Dy J (2004) Feature selection for unsupervised learning. J Mach Learn Res 5:845–889
Ekman P (2003) Emotions revealed: recognizing faces and feelings to improve communication and emotional life. Weidenfeld & Nicolson, London
Fontaine JRJ, Scherer KR, Roesch EB et al (2007) The world of emotions is not two-dimensional. Psychol Sci 18(12):1050–1057
Guyon I, Elisseeff A (2003) An introduction to variable and feature selection. J Mach Learn Res 3:1157–1182
Hevner SK (2007) EMuJoy: software for continuous measurement. Behav Res Methods 39(2):283–290
Hwang HJ, Kim S, Choi S et al (2013) EEG-based brain–computer interfaces (BCIs): a thorough literature survey. Int J Human Comput Interact 29(12):814–826
Izard CE (2007) Basic emotions, natural kinds, emotion schemas, and a new paradigm. Perspect Psychol Sci 2:260–280
Kokkinos I, Maragos P (2005) Nonlinear speech analysis using models for chaotic systems. IEEE Trans Speech Audio Process 13(6):1098–1109
Lang PJ, Bradley MM, Cuthbert BN (2008) International affective picture system (IAPS): affective ratings of pictures and instruction manual. Technical Report A-6, University of Florida, Gainesville, Fl
Larsen RJ, Diener E (1987) Affect intensity as an individual difference characteristic: a review. J Res Pers 21(1):1–39
Lee H, Seungjin C (2003) PCA + HMM + SVM for EEG pattern classification. In: Proceedings of signal processing and its applications, vol 1, pp 541–544
Lin Y, Juny T, Chen J (2009) EEG dynamics during music appreciation. In: Annual international conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society, pp 5316–5319
Lin Y-P, Wang C-H, Jung T-P et al (2010) EEG-based emotion recognition in music listening. IEEE Trans Biomed Eng 57(7):1798–1806
Lin Y-P, Chen J-H, Duann J-R et al (2011) Generalizations of the subject-independent feature set for music-induced emotion recognition. In: 33rd annual international conference of the IEEE, EMBS, Boston, Massachusetts, USA
Liu M, Wan C (2001) Feature selection for automatic classification of musical instrument sounds. In: Proceedings of the first ACM/IEEE-CS joint conference on digital libraries—JCDL ’01. ACM Press, New York, pp 247–248
Lotte F, Congedo M, Lécuyer A et al (2007) A review of classification algorithms for EEG-based brain-computer interfaces. J Neural Eng 4(2):1–13
Matthews G (1990) Refining the measurement of mood: the UWIST mood adjective checklist. Br J Psychol 81(1):17–42
Mitrović D, Zeppelzauer M, Breiteneder C (2010) Features for content-based audio retrieval. Adv Comput 78:71–150
Müller KR, Krauledat M, Dornhege G et al (2004) Machine learning techniques for brain-computer interfaces. Biomed Eng 49(1):11–22
Müller-Putz GR, Breitwieser C, Cincotti F et al (2011) Tools for brain–computer interaction: a general concept for a hybrid BCI. Front Neuroinform 5:30
Murugappan M, Nagarajan R, Yaacob S (2010) Classification of human emotion from EEG using discrete wavelet transform. J Biomed Sci Eng 3:390–396
Obermaier B, Guger C, Neuper C et al (2001) Hidden Markov models for online classification of single trial EEG data. Pattern Recogn Lett 22(12):1299–1309
Ogawa S, Ota S, Ito S et al (2005) Influence of music listening on the cerebral activity by analyzing EEG. In: Proceedings of the 9th international conference on knowledge-based intelligent information and engineering systems (KES’05), pp 657–663
Pfurtscheller G, Allison BZ, Brunner C et al (2010) The hybrid BCI. Front Neuroprosthetics 4:30
Qin J, Li Y, Cichocki A (2005) ICA and committee machine-based algorithm for cursor control in a BCI system. In: Advances in neural networks—ISNN 2005. Springer, Berlin, pp 973–978
Rezaei S, Tavakolian K, Nasrabadi AM et al (2006) Different classification techniques considering brain computer interface applications. J Neural Eng 3(2006):139–144
Roesch EB, Fontaine JB, Scherer KR (2006) The world of emotion is two-dimensional or is it? Paper presented to the HUMAINE Summer School 2006, Genoa, Italy
Ruiz VF, Nasuto SJ (2005) Biomedical image classification methods and techniques. In: Costaridou L (ed) Medical image analysis methods. Taylor & Francis, New York, p 504
Salovey P, Pizarro DA (2003) The value of emotional intelligence. In: Sternberg RJ, Lautrey J, Lubart TI (eds) Models of intelligence: international perspectives. American Psychological Association, Washington, DC, pp 263–278
Scherer KR, Schorr A, Johnstone T (2001) Appraisal processes in emotion: theory, methods, research. Oxford University Press, Oxford
Smith LI (2002) A tutorial on principal components analysis. Technical report, Cornell University
Sohaib AT, Qureshi S, Hagelbäck J et al (2013) Evaluating classifiers for emotion recognition using EEG. In: Foundations of augmented cognition. Lecture notes in computer science, vol 8027, pp 492–501
Soleymani M, Member S, Lee J (2012) DEAP: a database for emotion analysis using physiological signals. IEEE Trans Affect Comput 3(1):18–31
Tomkins SS (1962) Affect, imagery, consciousness. Springer, New York
Tzanetakis G, Cook P (2002) Musical genre classification of audio signals. IEEE Trans Speech Audio Process 10(5):293–302
Wang XW, Nie D, Lu BL (2011) EEG-based emotion recognition using frequency domain features and support vector machines. In: Lu B-L, Zhang L, Kwok J (eds) Proceedings of the 18th international conference on neural information processing (ICONIP’11), vol 7062, pp 734–743
Wang X-W, Nie D, Lu B-L (2013) Emotional state classification from EEG data using machine learning approach. Neurocomputing 129:94
Warwick K, Nasuto SJ (2006) Historical and current machine intelligence. IEEE Instrum Meas Mag 9(6):20–26
Yongjin W, Ling G (2005) Recognizing human emotion from audiovisual information. In: IEEE international conference on acoustics, speech, and signal processing (ICASSP ’05), vol 2, pp 1125–1128
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer-Verlag London
About this chapter
Cite this chapter
Daly, I., Roesch, E.B., Weaver, J., Nasuto, S.J. (2014). Machine Learning to Identify Neural Correlates of Music and Emotions. In: Miranda, E., Castet, J. (eds) Guide to Brain-Computer Music Interfacing. Springer, London. https://doi.org/10.1007/978-1-4471-6584-2_5
Download citation
DOI: https://doi.org/10.1007/978-1-4471-6584-2_5
Published:
Publisher Name: Springer, London
Print ISBN: 978-1-4471-6583-5
Online ISBN: 978-1-4471-6584-2
eBook Packages: Computer ScienceComputer Science (R0)