Skip to main content

A Deep Neural Network Model for Music Genre Recognition

  • Conference paper
  • First Online:
Advances in Natural Computation, Fuzzy Systems and Knowledge Discovery (ICNC-FSKD 2019)

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 1074))

  • 1340 Accesses

Abstract

Convolutional neural networks (CNNs) have become increasingly important to deal with many image processing and pattern recognition problems. In order to use CNNs in music genre recognition, spectrograms (visual representation of the spectrum of frequencies of a signal as it varies with time) are usually employed as inputs of the network. Yet some other approaches used music features for genre classification as well. In this paper we propose a new deep network model combining CNN with a simple multi-layer neural network for music genre classification. Since other features are taken into account in the multi-layer network, the combined deep neural network has shown better accuracy than each of the single models in the experiments (Code available at: https://github.com/risengnom/Music-Genre-Recognition.).

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Tzanetakis, G., Cook, P.: Musical genre classification of audio signals. In: IEEE Transactions on Speech and Audio Processing, vol. 10.5, pp. 293–302, July 2002. ISSN: 1063-6676

    Google Scholar 

  2. Popescu, A., Gavat, I., Datcu, M.: Wavelet analysis for audio signals with music classification applications. In: 2009 Proceedings of the 5th Conference on Speech Technology and Human-Computer Dialogue, pp. 1–6. IEEE (2009)

    Google Scholar 

  3. Hamel, P., Eck, D.: Université De Montréal. Learning features from music audio with deep belief networks. In: 11th International Society for Music Information Retrieval Conference, ISMIR 2010 (2010)

    Google Scholar 

  4. Dong, M.: Convolutional neural network achieves human-level accuracy in music genre classification. CoRR, abs/1802.09697 (2018)

    Google Scholar 

  5. Lee, J., Baniya, B.K., Ghimire, D.: Automatic music genre classification using timbral texture and rhythmic content features (2014)

    Google Scholar 

  6. Kittler, J., Hater, M., Duin, R.P.W.: Combining classifiers. In: Proceedings of 13th International Conference on Pattern Recognition, vol. 2, pp. 897–901, August 1996

    Google Scholar 

  7. Perrot, D.: Scanning the dial: An exploration of factors in the identification of musical style. In: Proceedings of ICMPC 1999 (1999)

    Google Scholar 

  8. Elbir, A., Bilal am, H., Emre Iyican, M., Ztrk, B., Aydin, N.: Music genre classification and recommendation by using machine learning techniques. In: 2018 Innovations in Intelligent Systems and Applications Conference (ASYU), pp. 1–5, October 2018

    Google Scholar 

  9. Logan, B.: Mel frequency cepstral coefficients for music modeling. In: Proceedings 1st International Symposium Music Information Retrieval, November 2000

    Google Scholar 

  10. Ghosal, D., Kolekar, M.H.: Music genre recognition using deep neural networks and transfer learning, pp. 2087–2091, September 2018

    Google Scholar 

  11. Fulzele, P., Singh, R., Kaushik, N., Pandey, K.: A hybrid model for music genre classification using LSTM and SVM. In: 2018 Eleventh International Conference on Contemporary Computing (IC3), pp. 1–3, August 2018

    Google Scholar 

  12. Peltonen, V., Tuomi, J., Klapuri, A., Huopaniemi, J., Sorsa, T.: Computational auditory scene recognition. In: 2002 IEEE International Conference on Acoustics, Speech, and Signal Processing, vol. 2, pp. II-1941–II-1944, May 2002

    Google Scholar 

  13. Driedger, J.: Processing Music Signals Using Audio Decomposition Techniques. PhD thesis, January 2016

    Google Scholar 

  14. LeCun, Y., et al. Generalization and network design strategies. In: Connectionism in perspective, vol. 19. Citeseer (1989)

    Google Scholar 

  15. Nielsen, M.A.: Neural Networks and Deep Learning, vol. 25. Determination press, San Francisco (2015)

    Google Scholar 

  16. Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: Pereira, F., Burges, C.J.C., Bottou, L., Weinberger, K.Q., (eds.) Advances in Neural Information Processing Systems, vol. 25, pp. 1097–1105. Curran Associates, Inc. (2012)

    Google Scholar 

  17. O’Shaughnessy, D.: Speech Communication: Human and Machine, p. 150. Addison-Wesley, Boston (1987)

    Google Scholar 

  18. Hecht-Nielsen, R.: Theory of the back propagation neural network. In: Neural Networks for Perception, pp. 65–93. Elsevier (1992)

    Google Scholar 

  19. Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. CoRR, abs/1412.6980 (2014)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to N. Xiong .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Suero, M., Gassen, C.P., Mitic, D., Xiong, N., Leon, M. (2020). A Deep Neural Network Model for Music Genre Recognition. In: Liu, Y., Wang, L., Zhao, L., Yu, Z. (eds) Advances in Natural Computation, Fuzzy Systems and Knowledge Discovery. ICNC-FSKD 2019. Advances in Intelligent Systems and Computing, vol 1074. Springer, Cham. https://doi.org/10.1007/978-3-030-32456-8_41

Download citation

Publish with us

Policies and ethics