Abstract
Musical genres are used as a way to categorize and describe different types of music. The common qualities of musical genre members are connected with the song's melody, rhythmic structure, and harmonic composition. The enormous amounts of music available on the Internet are structured using genre hierarchies. Currently, the manual annotation of musical genres is ongoing. Development of a framework is essential for automatic classification and analysis of music genre which can replace human in the process making it more valuable for musical retrieval systems. In this paper, a model is developed which categorizes audio data in to musical genre automatically. Three models such as neural network, CNN, and RNN-LSTM model were implemented and CNN model outperformed others, with training data accuracy of 72.6% and test data accuracy of 66.7%. To evaluate the performance and relative importance of the proposed features and statistical pattern recognition, classifiers are trained considering features such as timbral texture, rhythmic content, pitch content, and real-world collections. The techniques for classification are provided for both whole files and real-time frames. For ten musical genres, the recommended feature sets result in a classification rate with an accuracy of 61%.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Elbir A, Çam HB, Iyican ME, Öztürk B, Aydin N (2018) Music genre classification and recommendation by using machine learning techniques. In: 2018 Innovations in intelligent systems and applications conference (ASYU). IEEE, pp 1–5
Pandey P. Music genre classification with Python. In: Towards data science
Identifying the genre of a song with neural networks
Singh N. Identifying the genre of a song with neural networks. Medium
Scaringella N, Zoia G, Mlynek D (2006) Automatic genre classification of music content: a survey. IEEE Signal Process Mag 23(2):133–141
Zhang W, Lei W, Xu X, Xing X (2016) Improved music genre classification with convolutional neural networks. In: Interspeech, pp 3304–3308
Sigtia S, Siddharth, Dixon S (2014) Improved music feature learning with deep neural networks. In: 2014 IEEE international conference on acoustics, speech and signal processing (ICASSP). IEEE, pp 6959–6963
Wu Y, Mao H, Yi Z. Audio classification using attention-augmented convolutional neural network. Machine Intelligence Lab, College of Computer Science, Sichuan University, Chengdu 610065, China
Palanisamy K, Singhania D, Yao A. Rethinking CNN models for audio classification
Nanni L, Maguolo G, Shery B, Paci M. An ensemble of convolutional neural networks for audio classification
Pandeya YR, Kim D, Lee J (1949) Domestic cat sound classification using learned features from deep neural nets. Appl Sci 2018:8
Nanni L, Brahnam S, Maguolo G (2019) Data augmentation for building an ensemble of convolutional neural networks. In: Smart innovation systems and technologies. In: Chen Y-W (ed). Springer, Singapore
Salamon J, Bello J (2017) Deep convolutional neural networks and data augmentation for environmental sound classification. IEEE Signal Process Lett
Nanni L, Costa YM, Aguiar RL, Silla CN, Jr, Brahnam S (2018) Ensemble of deep learning visual and acoustic features for music genre classification. J New Music Res
Tran T, Lundgren J (2020) Drill fault diagnosis based on the scalogram and mel spectrogram of sound signals using artificial intelligence. IEEE Access
Bengio Y, Simard P, Frasconi P (1994) Learning long-term dependencies with gradient descent is difficult. IEEE Trans Neural Netw 5(2):157–166
Alexandridis A, Chondrodima E, Paivana G, Stogiannos M, Zois E, Sarimveis H (2014) Music genre classification using radial basis function networks and particle swarm optimization. In: 2014 6th computer science and electronic engineering conference (CEEC). IEEE, pp 35–40
Bergstra J, Casagrande N, Erhan D, Eck D, Kegl B (2006) Aggregate features and adaboost for music classification. Mach Learn 65(2–3):473–484
Xu Y, Kong Q, Wang W, Plumbley MD (2018) Large-scale weakly supervised audio classification using gated convolutional neural network. Center for Vision, Speech and Signal Processing, University of Surrey, UK, https://doi.org/10.1109/ICASSP.2018.8461975
Hershey S, Chaudhuri S, Ellis DPW, Gemmeke JF, Aren Jansen R, Moore C, Plakal M, Platt D, Saurous RA, Seybold B, Slaney M, Weiss RJ, Wilson K. CNN Architecture for large scale audio classification. Google, Inc., New York, NY, Mountain View, CA, USA
Lee J, Kim T, Park J, Nam J (2017) Raw-waveform based audio classification using sample level CNN architecture. In: NIPS, machine learning for audio signal processing workshop (ML4Audio)
Krizhevsky A, Sutskever I, Hinton GE (2012) ImageNet classification with deep convolutional neural networks. In: Pereira F (ed) Advances in neural information processing systems. Curran Associates Inc., Red Hook, NY, USA
Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov, D, Erhan D, Vanhoucke V, Rabinovich A (2015) Going deeper with convolutions. In: Proceedings of the IEEE computer society conference on computer vision and pattern recognition, Boston, MA, USA, 7–12 June 2015
Kumar A, Khadkevich M, Fügen C (2018) Knowledge transfer from weakly labeled audio using convolutional neural network for sound events and scenes. In: Proceedings of the 2018 IEEE international conference on acoustics speech and signal processing (IEEE ICASSP), Calgary, AB, Canada, 15–20 Apr 2018
Nanni L, Costa Costa YM, Alessandra L, Kim MY, Baek SR (2016) Combining visual and acoustic features for music genre classification. Expert Syst Appl
Kim J (2020) Urban sound tagging using multi-channel audio feature with convolutional neural networks. In: Proceedings of the detection and classification of acoustic scenes and events 2020, Tokyo, Japan, 2–3 Nov 2020
Acknowledgements
We thank all the members of the Artificial Intelligence Club, JSS Academy of Technical Education, Bengaluru, for their useful suggestions during this work.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Suguna, G.C., Shirahatti, S.L., Veerabhadrappa, S.T., Bangari, S.R., Jain, G.A., Bhat, C.V. (2023). Automatic Detection of Musical Note Using Deep Learning Algorithms. In: Rajakumar, G., Du, KL., Vuppalapati, C., Beligiannis, G.N. (eds) Intelligent Communication Technologies and Virtual Mobile Networks. Lecture Notes on Data Engineering and Communications Technologies, vol 131. Springer, Singapore. https://doi.org/10.1007/978-981-19-1844-5_29
Download citation
DOI: https://doi.org/10.1007/978-981-19-1844-5_29
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-19-1843-8
Online ISBN: 978-981-19-1844-5
eBook Packages: EngineeringEngineering (R0)