Advertisement

A Matlab Toolbox for Music Information Retrieval

  • Olivier Lartillot
  • Petri Toiviainen
  • Tuomas Eerola
Part of the Studies in Classification, Data Analysis, and Knowledge Organization book series (STUDIES CLASS)

Abstract

We present MIRToolbox, an integrated set of functions written in Matlab, dedicated to the extraction from audio files of musical features related, among others, to timbre, tonality, rhythm or form. The objective is to offer a state of the art of computational approaches in the area of Music Information Retrieval (MIR). The design is based on a modular framework: the different algorithms are decomposed into stages, formalized using a minimal set of elementary mechanisms, and integrating different variants proposed by alternative approaches — including new strategies we have developed —, that users can select and parametrize. These functions can adapt to a large area of objects as input.

This paper offers an overview of the set of features that can be extracted with MIRToolbox, illustrated with the description of three particular musical features. The toolbox also includes functions for statistical analysis, segmentation and clustering.

One of our main motivations for the development of the toolbox is to facilitate investigation of the relation between musical features and music-induced emotion. Preliminary results show that the variance in emotion ratings can be explained by a small set of acoustic features.

Keywords

Audio Signal Acoustic Feature Emotion Rating Musical Piece Musical Genre 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. EEROLA, T. and TOIVIAINEN, P. (2004): MIR in Matlab: The MIDI Toolbox. Proceedings of 5th International Conference on Music Information Retrieval, 22-27, Barcelona.Google Scholar
  2. FOOTE, J. and COOPER, M. (2003): Media segmentation using self-similarity decomposi-tion. In Proceedings of SPIE Storage and Retrieval for Multimedia Databases, 5021, 167-75.Google Scholar
  3. GOMEZ, E. (2006): Tonal description of polyphonic audio for music content processing. IN-FORMS Journal on Computing, 18-3, 294-304.Google Scholar
  4. JUSLIN, P. N. (1997): Emotional communication in music performance: A functionalist per-spective and some data. Music Perception, 14, 383-418.Google Scholar
  5. JUSLIN, P. N. and LAUKKA, P. (2003): Communication of emotions in vocal expression and music performance: Different channels, same code? Psychological Bulletin (129), 770-814.CrossRefGoogle Scholar
  6. KRUMHANSL, C. (1990): Cognitive Foundations of Musical Pitch. Oxford University Press, New York.Google Scholar
  7. KRUMHANSL, C. and KESSLER, E. J. (1982): Tracing the dynamic changes in perceived tonal organization in a spatial representation of musical keys. Psychological Review, 89, 334-368.CrossRefGoogle Scholar
  8. NABNEY, I. (2002): NETLAB: Algorithms for Pattern Recognition. Springer Advances In Pattern Recognition Series, Springer-Verlag, New-York.Google Scholar
  9. RABINER, L. and JUANG, B. H. (1993): Fundamentals of Speech Recognition. Prentice-Hall. SCHERER, K. R. and OSHINSKY J. S. (1977): Cue utilization in emotion attribution from auditory stimuli. Motivation and Emotion, 1-4, 331-346.Google Scholar
  10. SLANEY, M. (1998): Auditory Toolbox Version 2. Technical Report 1998-010, Interval Re-search Corporation.Google Scholar
  11. TOIVIAINEN, P. and KRUMHANSL, C. (2003): Measuring and modeling real-time re-sponses to music: The dynamics of tonality induction, Perception, 32-6, 741-766.CrossRefGoogle Scholar
  12. TOIVIAINEN, P. and SNYDER J. S. (2003): Tapping to Bach: Resonance-based modeling of pulse. Music Perception, 21(1), 43-80.CrossRefGoogle Scholar
  13. TZANETAKIS, G and COOK, P. (1999): Multifeature audio segmentation for browsing and annotation. Proceedings of the 1999 IEEE Workshop on Applications of Signal Process-ing to Audio and Acoustics. New-York.Google Scholar
  14. TZANETAKIS, G. and COOK, P. (2002): Musical genre classification of audio signals. IEEE Transactions on Speech and Audio Processing, 10(5), 293Ð302.CrossRefGoogle Scholar
  15. VESANTO, J. (1999): Self-organizing map in Matlab: the SOM Toolbox. Proceedings of the Matlab DSP Conference 1999. Espoo, Finland,35-40.Google Scholar
  16. WITTEN, I. H. and FRANK, E. (2005): Data Mining: Practical Machine Learning Tools and Techniques, 2nd Edition. Morgan Kaufmann, San Francisco, 2005.zbMATHGoogle Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2008

Authors and Affiliations

  • Olivier Lartillot
    • 1
  • Petri Toiviainen
    • 1
  • Tuomas Eerola
    • 1
  1. 1.University of Jyväskylä, PL 35(M), FI-40014Finland

Personalised recommendations