Abstract
Professional musicians manipulate sound properties such as pitch, timing, amplitude, and timbre in order to add expression to their performances. However, there is little quantitative information about how and in which contexts this manipulation occurs. In this chapter, we describe an approach to quantitatively model and analyze expression in popular music monophonic performances, as well as identifying interpreters from their playing styles. The approach consists of (1) applying sound analysis techniques based on spectral models to real audio performances for extracting both inter-note and intra-note expressive features, and (2) based on these features, training computational models characterizing different aspects of expressive performance using machine learning techniques. The obtained models are applied to the analysis and synthesis of expressive performances as well as to automatic performer identification. We present results, which indicate that the features extracted contain sufficient information, and the explored machine learning methods are capable of learning patterns that characterize expressive music performance.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Gabrielsson A (1999) The performance of music. In: Deutsch D (ed) The psychology of music, 2nd edn. Academic, New York, 579 pages
Gabrielsson A (2003) Music performance research at the millennium. Psychol Music 31(3):221–272
Johnson ML (1992) An expert system for the articulation of Bach fugue melodies. In: Baggi DL (ed) Readings in computer-generated music. IEEE Computer Society, Los Alamitos, pp 41–51
Bresin R (2001) Articulation rules for automatic music performance. In Proceedings of the international computer music conference. International Computer Music Association, San Francisco, pp 294–297
Friberg A, Bresin R, Fryden L (2000) Music from motion: sound level envelopes of tones expressing human locomotion. J New Music Res 29(3):199–210
Friberg A, Bresin R, Sundberg J (2006) Overview of the KTH rule system for musical performance. Adv Cogn Psychol Spec Issue Music Perform 2(2–3):145–161
Canazza S, De Poli G, Roda A, Vidolin A (1997) Analysis and synthesis of expressive intention in a clarinet performance. In Proceedings of the 1997 international computer music conference. International Computer Music Association, San Francisco, pp 113–120
Canazza S, De Poli G, Drioli C, Roda A, Vidolin A (2004) Modeling and control of expressiveness in music performance. Proc IEEE 92(4):286–701
Dannenberg RB, Derenyi I (1998) Combining instrument and performance models for high-quality music synthesis. J New Music Res 27(3):211–238
Mitchell TM (1997) Machine learning. McGraw-Hill, New York, 542 pages
de Mantaras RL, Arcos JL (2002) AI and music, from composition to expressive performance. AI Mag 23(3):43–57
Ramirez R, Hazan A (2006) A tool for generating and explaining expressive music performances of monophonic jazz melodies. Int J Artif Intell Tools 15(4):673–691
Ramirez R, Hazan A, Maestre E, Serra X (2008) A genetic rule-based expressive performance model for jazz saxophone. Comput Music J 32(1):38–50
Widmer G (2001) Discovering strong principles of expressive music performance with the PLCG rule learning strategy. In Proceedings of the 12th European conference on machine learning (ECML'01), Germany. Springer, Berlin, pp 552–563
Widmer G (2002) Machine discoveries: a few simple, robust local expression principles. J New Music Res 31(1):37–50
Tobudic A, Widmer G (2003) Relational IBL in music with a new structural similarity measure. In Proceedings of the international conference on inductive logic programming. Springer, pp 365–382
Dovey MJ (1995) Analysis of Rachmaninoff’s piano performances using inductive logic programming. In European conference on machine learning. Springer, pp 35–38
Van Baelen E, De Raedt L (1996) Analysis and prediction of piano performances using inductive logic programming. In International conference in inductive logic programming, pp 55–71
Saunders C, Hardoon D, Shawe-Taylor J, Widmer G (2004) Using string kernels to identify famous performers from their playing style. In Proceedings of the 15th European conference on machine learning (ECML'2004), Pisa, Italy, pp 384–395
Stamatatos E, Widmer G (2005) Automatic identification of music performers with learning ensembles. Artif Intell 165(1):37–56
Ramirez R, Maestre E, Serra X (2010) Automatic performer identification in commercial monophonic jazz performances. Pattern Recognit Lett 31:1514–1523
Ramirez R, Perez A, Kersten S, Maestre E (2008) Performer identification in celtic violin recordings. In International Society of Music Information Retrieval (ISMIR) conference, Philadelphia, USA, pp 483–488
Molina-Solana M, Arcos JL, Gomez E (2010) Identifying violin performers by their expressive trends. Intell Data Anal 14(5):555–571
Narmour E (1990) The analysis and cognition of basic melodic structures: the implication realization model. University of Chicago Press, Chicago, 358 pages
Gómez E, Klapuri A, Meudic B (2003) Melody description and extraction in the context of music content processing. J New Music Res 32:33–54
Klapuri A (1999) Sound onset detection by applying psychoacoustic knowledge. In Proceedings of the IEEE international conference on acoustics, speech and signal processing, ICASSP, pp 3089–3092
Maher RC, Beauchamp JW (1994) Fundamental frequency estimation of musical signals using a two-way mismatch procedure. J Acoust Soc Am 95:2254–2263
McNab RJ, Smith LA, Witten IH (1996) Signal processing for melody transcription, working paper 95/22, Hamilton, New Zealand, University of Weikato, Department of Computer Science
Maestre E, Gomez E (2005) Automatic characterization of dynamics and articulation of monophonic expressive recordings. In Proceedings of the 118th AES convention, Barcelona, Spain, pp 36–40
Narmour E (1991) The analysis and cognition of melodic complexity: the implication realization model. University of Chicago Press, Chicago, 321 pages
Blockeel H, De Raedt L, Ramon J (1998) Top-down induction of clustering trees. In Shavlik J(ed) Proceedings of the 15th international conference on machine learning, Madison, Wisconsin, USA. Morgan Kaufmann, pp 53–63
Quinlan JR (1993) C4.5: programs for machine learning. Morgan Kaufmann, San Francisco, 305 pages. ISBN 1–55860–238–0
Rosinach V, Traube C (2006) Measuring swing in Irish traditional fiddle music. In Proceedings of international conference on music perception and cognition, pp 1168–1171
Friberg A, Sundstrom J (2002) Swing ratios and ensemble timing in jazz performances: evidence for a common rhythmic pattern. Music Percept 19(3):333–349
Maestre E, Hazan A, Ramirez R, Perez A (2006) Using concatenative synthesis for expressive performance in jazz saxophone. In Proceedings of international computer music conference, New Orleans, pp 82–85
Cristianini N, Shawe-Taylor J (2000) An introduction to support vector machines. Cambridge University Press, Cambridge, 190 pages. ISBN 0–521–78019–5
Chauvin Y, Rumelhart ED (eds) (1995) Backpropagation: theory, architectures and applications. Lawrence Erlbaum Assoc, Hillsdale, 549 pages. ISBN 0–8058–1259–8
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Questions
Questions
-
1.
What are the four musicological questions that this study attempts to answer?
-
2.
Name three areas that could be helped by answers to these questions.
-
3.
How is note segmentation done on the audio stream?
-
4.
What are the two main principles recognized by Narmour in his theory?
-
5.
How many prototypical Narmour structures are there?
-
6.
How is bow direction detected in the gesture acquisition?
-
7.
What levels of metrical strength are defined in the note descriptors?
-
8.
What are some of the traditional deviations found in Irish jig music?
-
9.
Why might the results be poor for the 1-note experiments?
-
10.
What were the most successful and least successful classifiers in the results?
Rights and permissions
Copyright information
© 2013 Springer-Verlag London
About this chapter
Cite this chapter
Ramirez, R., Maestre, E., Perez, A. (2013). Modeling, Analyzing, Identifying, and Synthesizing Expressive Popular Music Performances. In: Kirke, A., Miranda, E. (eds) Guide to Computing for Expressive Music Performance. Springer, London. https://doi.org/10.1007/978-1-4471-4123-5_5
Download citation
DOI: https://doi.org/10.1007/978-1-4471-4123-5_5
Published:
Publisher Name: Springer, London
Print ISBN: 978-1-4471-4122-8
Online ISBN: 978-1-4471-4123-5
eBook Packages: Computer ScienceComputer Science (R0)