Interpretable Music Categorisation Based on Fuzzy Rules and High-Level Audio Features
Music classification helps to manage song collections, recommend new music, or understand properties of genres and substyles. Until now, the corresponding approaches are mostly based on less interpretable low-level characteristics of the audio signal, or on metadata, which are not always available and require high efforts for filtering the relevant information. A listener-friendly approach may rather benefit from high-level and meaningful characteristics. Therefore, we have designed a set of high-level audio features, which is capable to replace the baseline low-level feature set without a significant decrease of classification performance. However, many common classification methods change the original feature dimensions or create complex models with lower interpretability. The advantage of the fuzzy classification is that it describes the properties of music categories in an intuitive, natural way. In this work, we explore the ability of a simple fuzzy classifier based on high-level features to predict six music genres and eight styles from our previous studies.
KeywordsClassification Model Fuzzy Controller Audio Signal Linguistic Term Fuzzy Classification
We thank the Klaus Tschira Foundation for the financial support.
- Abeßer, J., Lukashevich, H., Dittmar, C., & Schuller, G. (2009). Genre classification using bass-related high-Level features and playing styles. In Proceedings of the 10th Int’L Conference on Music Information Retrieval (ISMIR) (pp. 453–458).Google Scholar
- Geyer-Schulz, A. (1998). Fuzzy genetic algorithms. In H. T. Ngyen & M. Sugeno (Eds.), Fuzzy systems. Boston: Kluwer Academic Publishers.Google Scholar
- Friberg, A. (2005). A fuzzy analyzer of emotional expression in music performance and body motion. In J. Sundberg & B. Brunson (Eds.), Proceedings of Music and Music Science.Google Scholar
- Hu, X., & Liu, J. (2010). User-centered music information retrieval evaluation. In Proceedings of the Joint Conference on Digital Libraries (JCDL) Workshop: Music Information Retrieval for the Masses.Google Scholar
- Mauch, M., & Levy, M. (2011). Structural change on multiple time scales as a correlate of musical complexity. In Proceedings of the 12th International Society for Music Information Retrieval Conference (ISMIR) (pp. 489–494).Google Scholar
- Mckay, C., & Fujinaga, I. (2006). Musical genre classification: Is it worth pursuing and how can it be improved? In Proceedings of the 7th International Conference on Music Information Retrieval (ISMIR) (pp. 101–106).Google Scholar
- Pachet, F., & Zils, A. (2003). Evolving automatically high-level music descriptors from acoustic signals. In Proceedings of the 1st International Symposium on Computer Music Modeling and Retrieval (CMMR) (pp. 42–53).Google Scholar
- Sturm, B. (2012). A survey of evaluation in music genre recognition. In Proceedings of the 10th International Workshop on Adaptive Multimedia Retrieval (AMR).Google Scholar
- Vatolkin, I., Theimer, W., & Botteck, M. (2012). Partition based feature processing for improved music classification. In W. A. Gaul, A. Geyer-Schulz, L. Schmidt-Thieme, & J. Kunze (Eds.), Challenges at the interface of data analysis, computer science, and optimization (pp. 411–419). Berlin: Springer.CrossRefGoogle Scholar
- Vatolkin, I. (2013). Improving supervised music classification by means of multi-objective evolutionary feature selection. PhD thesis, Department of Computer Science, TU Dortmund, 2013.Google Scholar
- Yang, Y. -H., Liu, C. -C., & Chen, H. H. (2006). Music emotion classification: A fuzzy approach. In: K. Nahrstedt, M. Turk, Y. Rui, W. Klas, & K. Mayer-Patel (Eds.), Proceedings of the 14th ACM International Conference on Multimedia (pp. 81–84).Google Scholar
- Zhou, Z. -H. (2012). Ensemble methods: Foundations and algorithms. Boca Raton: CRC Press.Google Scholar