Abstract
In this work, effort has been made to classify audio songs based on their music pattern which helps us to retrieve the music clips based on listener’s taste. This task is helpful in indexing and accessing the music clip based on listener’s state. Seven main categories are considered for this work such as devotional, energetic, folk, happy, pleasant, sad and, sleepy. Forty music clips of each category for training phase and fifteen clips of each category for testing phase are considered; vibrato-related features such as jitter and shimmer along with the mel-frequency cepstral coefficients (MFCCs); statistical values of pitch such as min, max, mean, and standard deviation are computed and added to the MFCCs, jitter, and shimmer which results in a 19-dimensional feature vector. feedforward backpropagation neural network (BPNN) is used as a classifier due to its efficiency in mapping the nonlinear relations. The accuracy of 82 % is achieved on an average for 105 testing clips.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
References
Perrot, D., Gjerdigen, R.: Scanning the dial: an exploration of factors in the identification of musical style. In: Proceedings of the 1999 Society for Music Perception and Cognition, p. 88 (1999)
Park, H.S., Yoo, J.O., Cho, S.B.: A context-aware music recommendation system using fuzzy bayesian networks with utility theory. In: Fuzzy Systems and Knowledge Discovery, pp. 970–979. Springer (2006)
Casey, M.A., Veltkamp, R., Goto, M., Leman, M., Rhodes, C., Slaney, M.: Content-based music information retrieval: current directions and future challenges. Proc. IEEE 96, 668–696 (2008)
Freed, A.: Music metadata quality: a multiyear case study using the music of skip james. In: Audio Engineering Society Convention 121, Audio Engineering Society (2006)
Mesaros, A., Virtanen, T., Klapuri, A.: Singer Identification in Polyphonic Music Using Vocal Separation and Pattern Recognition Methods, pp. 375–378. ISMIR (2007)
Ratanpara, T., Patel, N.: Singer identification using mfcc and lpc coefficients from indian video songs. In: Emerging ICT for Bridging the Future-Proceedings of the 49th Annual Convention of the Computer Society of India (CSI), vol. 1, pp. 275–282. Springer (2015)
Cai, W., Li, Q., Guan, X.: Automatic singer identification based on auditory features. In: Seventh International Conference on Natural Computation (ICNC), IEEE, vol. 3, pp. 1624–1628 (2011)
Mesaros, A., Astola, J.: The Mel-Frequency Cepstral Coefficients in the Context of Singer Identification, pp. 610–613. ISMIR, Citeseer (2005)
Rabiner, L.R., Juang, B.H.: Fundamentals of speech recognition, vol. 14. PTR Prentice Hall Englewood Cliffs (1993)
Seddik, H., Rahmouni, A., Sayadi, M.: Text independent speaker recognition using the mel frequency cepstral coefficients and a neural network classifier. In: First International Symposium on Control, Communications and Signal Processing, IEEE, pp. 631–634 (2004)
Fredrickson, S., Tarassenko, L.: Text-Independent Speaker Recognition Using Neural Network Techniques (1995)
Mafra, A.T., Simões, M.G.: Text independent automatic speaker recognition using selforganizing maps. In: Industry Applications Conference, 2004. 39th IAS Annual Meeting. Conference Record of the 2004 IEEE, vol. 3, pp. 1503–1510 (2004)
Brown, J.C.: Computer identification of musical instruments using pattern recognition with cepstral coefficients as features. J. Acoustical Soc. Am. 105, 1933–1941 (1999)
Collier, W.G., Hubbard, T.L.: Musical scales and brightness evaluations: effects of pitch, direction, and scale mode. Musicae Scientiae 8, 151–173 (2004)
Joder, C., Essid, S., Richard, G.: Temporal integration for audio classification with application to musical instrument classification. IEEE Trans. Audio Speech Lang. Proc. 17, 174–186 (2009)
Eronen, A.: Comparison of features for musical instrument recognition. In: IEEE Workshop on the Applications of Signal Processing to Audio and Acoustics, pp. 19–22 (2001)
Eronen, A., Klapuri, A.: Musical instrument recognition using cepstral coefficients and temporal features. In: IEEE Int. Conf. Acoustics Speech Signal Proc. (ICASSP’00) 2, II753–II756 (2000)
Chui, C.K.: An Introduction to Wavelets, vol. 1. Academic Press (2014)
Hlawatsch, P.F., Boudreaux-Bartels, G.: Special Issue on Wavelets and Signal Processing. Urbana (51) 61801
Li, G., Khokhar, A.A.: Content-based indexing and retrieval of audio data using wavelets. IEEE Int. Conf. Multimedia Expo (ICME) 2, 885–888 (2000)
Li, T., Ogihara, M., Li, Q.: A comparative study on content-based music genre classification. In: Proceedings of the 26th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, ACM, pp. 282–289 (2003)
Tzanetakis, G., Essl, G.: Automatic musical genre classification of audio signals. In: Proceedings in International Symposium on Music Information Retrieval, ISMIR (Oct. 2001)
Wold, E., Blum, T., Keislar, D., Wheaten, J.: Content-based classification, search, and retrieval of audio. MultiMedia IEEE 3, 27–36 (1996)
Kim, H.G., Sikora, T.: Audio spectrum projection based on several basis decomposition algorithms applied to general sound recognition and audio segmentation. na (2004)
Lerch, A.: An Introduction to Audio Content Analysis: Applications in Signal Processing and Music Informatics. Wiley (2012)
Sun, X.: Pitch determination and voice quality analysis using subharmonic-to-harmonic ratio. IEEE Int. Conf. Acoustics Speech Signal Proc. (ICASSP) 1, I–333 (2002)
Koolagudi, S., Shivakranthi, B., Rao, K.S., Ramteke, P.B.: Contribution of telugu vowels in identifying emotions. In: Eighth International Conference on Advances in Pattern Recognition (ICAPR), IEEE, pp. 1–6 (2015)
Berenzweig, A.L., Ellis, D.P., Lawrence, S.: Using voice segments to improve artist classification of music. In: Audio Engineering Society Conference: 22nd International Conference: Virtual, Synthetic, and Entertainment Audio, Audio Engineering Society (2002)
Murthy, Y.S., Koolagudi, S.G.: Classification of vocal and non-vocal regions from audio songs using spectral features and pitch variations. In: IEEE 28th Canadian Conference on Electrical and Computer Engineering (CCECE), IEEE, pp. 1–6 (2015)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer India
About this paper
Cite this paper
Sharma, R., Srinivasa Murthy, Y.V., Koolagudi, S.G. (2016). Audio Songs Classification Based on Music Patterns. In: Satapathy, S., Raju, K., Mandal, J., Bhateja, V. (eds) Proceedings of the Second International Conference on Computer and Communication Technologies. Advances in Intelligent Systems and Computing, vol 381. Springer, New Delhi. https://doi.org/10.1007/978-81-322-2526-3_17
Download citation
DOI: https://doi.org/10.1007/978-81-322-2526-3_17
Published:
Publisher Name: Springer, New Delhi
Print ISBN: 978-81-322-2525-6
Online ISBN: 978-81-322-2526-3
eBook Packages: EngineeringEngineering (R0)