Music and Similarity Based Reasoning

  • Josep LLuís Arcos
Part of the Studies in Fuzziness and Soft Computing book series (STUDFUZZ, volume 273)


Whenever that a musician plays a musical piece, the result is never a literal interpretation of the score. These performance deviations are intentional and constitute the essence of the musical communication. Deviations are usually thought of as conveying expressiveness. Two main purposes of musical expression are generally recognized: the clarification of the the musical structure and the transmission of affective content. The challenge of the computer music field when modeling expressiveness is to grasp the performers “touch”, i.e., the musical knowledge applied when performing a score. One possible approach to tackle the problem is to try to make explicit this knowledge using musical experts. An alternative approach, much closer to the human observation-imitation process, is to directly work with the knowledge implicitly stored in musical recordings and let the system imitate these performances. This alternative approach, also called lazy learning, focus on locally approximating a complex target function when a new problem is presented to the system. Exploiting the notion of local similarity, the chapter presents how the Case-Based Reasoning methodology has been successfully applied to design different computer systems for musical expressive performance.


Trend Model Musical Piece Expressive Performance Audio Descriptor Musical Knowledge 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Aamodt, A., Plaza, E.: Case-Based Reasoning: Foundational Issues, Methodological Variations, and System Approaches. Artificial Intelligence Communications 7(1), 39–59 (1994)Google Scholar
  2. 2.
    Arcos, J.L., de Mántaras, R.L.: An interactive case-based reasoning approach for generating expressive music. Applied Intelligence 14(1), 115–129 (2001)zbMATHCrossRefGoogle Scholar
  3. 3.
    Arcos, J.L., de Mántaras, R.L.: Combining fuzzy and case-based reasoning to generate human-like music performances. In: Technologies for Constructing Intelligent Systems: Tasks, pp. 21–31. Physica-Verlag GmbH (2002)Google Scholar
  4. 4.
    Bresin, R.: Articulation rules for automatic music performance. In: Proceedings of the International Computer Music Conference - ICMC 2001, San Francisco, pp. 294–297 (2001)Google Scholar
  5. 5.
    Gabrielsson, A.: Expressive Intention and Performance. In: Steinberg, R. (ed.) Music and the Mind Machine, pp. 35–47. Springer, Heidelberg (1995)CrossRefGoogle Scholar
  6. 6.
    Grachten, M., Arcos, J.L., de Mántaras, R.L.: A case based approach to expressivity-aware tempo transformation. Machine Learning 65(2-3), 411–437 (2006)CrossRefGoogle Scholar
  7. 7.
    Honing, H.: Is expressive timing relational invariant under tempo transformation? Psychology of Music 35(2), 276–285 (2007)CrossRefGoogle Scholar
  8. 8.
    Johnson, M.L.: An Expert System for the Articulation of Bach Fugue Melodies. In: Baggi, D.L. (ed.) Readings in Computer-Generated Music, pp. 41–51. IEEE Computes Society Press (1992)Google Scholar
  9. 9.
    Juslin, P.N.: Communicating emotion in music performance: a review and a theoretical framework. In: Juslin, P.N., Sloboda, J.A. (eds.) Music and Emotion: Theory and Research, pp. 309–337. Oxford University Press (2001)Google Scholar
  10. 10.
    Lester, J.: Bach’s Works for Solo Violin: Style, Structure, Performance. Oxford University Press (1999)Google Scholar
  11. 11.
    Levenshtein, V.I.: Binary codes capable of correcting deletions, insertions and reversals. Soviet Physics Doklady 10, 707–710 (1966)MathSciNetGoogle Scholar
  12. 12.
    Maestre, E., Gómez, E.: Automatic characterization of dynamics and articulation of expressive monophonic recordings. In: AES 118th Convention (2005)Google Scholar
  13. 13.
    de Mántaras, R.L., Acros, J.L.: Arcos: AI and music, form composition to expressive performance. AI Magazine 23(3), 43–57 (2002)Google Scholar
  14. 14.
    de Mántaras, R.L., McSherry, D., Bridge, D., David, L., Smyth, B., Susan, C., Faltings, B., Maher, M.L., Cox, M.T., Forbus, K., Keane, M., Aamodt, A., Watson, I.: Retrieval, Reuse, Revision, and Retention in CBR. Knowledge Engineering Review 20(3) (2005)Google Scholar
  15. 15.
    Meyer, L.: Emotion and Meaning in Music. University of Chicago Press, Chicago (1956)Google Scholar
  16. 16.
    Molina-Solana, M., Arcos, J.L., Gómez, E.: Identifying Violin Performers by their Expressive Trends. Intelligent Data Analysis (2010)Google Scholar
  17. 17.
    Narmour, E.: The Analysis and cognition of basic melodic structures: the implication-realization model. University of Chicago Press, Chicago (1990)Google Scholar
  18. 18.
    Palmer, C.: Anatomy of a performance: Sources of musical expression. Music Perception 13(3), 433–453 (1996)MathSciNetGoogle Scholar
  19. 19.
    Serra, X.: Musical Sound Modeling with Sinusoids plus Noise. In: Roads, C., Pope, S.T., Picialli, A., De Poli, G. (eds.) Musical Signal Processing, pp. 91–122. Swets and Zeitlinger Publishers (1997)Google Scholar
  20. 20.
    Widmer, G.: Discovering simple rules in complex data: A meta-learning algorithm and some surprising musical discoveries. Artificial Intelligence 146(2), 129–148 (2003)MathSciNetzbMATHCrossRefGoogle Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2012

Authors and Affiliations

  • Josep LLuís Arcos
    • 1
  1. 1.Artificial Intelligence Research Institute of the SpanishNational Research Council (IIIA-CSIC)BarcelonaSpain

Personalised recommendations