Abstract
In this paper, we explore the usage of char-RNN which is special type of recurrent neural network (RNN) in generating music pieces and propose an approach to do so. First, we train a model using existing music data. The generating model mimics the music patterns in such a way that we humans enjoy. The generated model does not replicate the training data but understands and creates patterns to generate new music. We generate honest quality music which should be good and melodious to hear. By tuning, the generated music can be beneficial for composers, film makers, artists in their tasks, and it can also be sold by companies or individuals. In our paper, we focus more on char ABC-notation because it is reliable to represent music using just sequence of characters. We use bidirectional long short-term memory (LSTM) which takes input as music sequences and observer that the proposed model has more accuracy compared with other models.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Briot JP, Pachet F (2020) Deep learning for music generation: challenges and directions. Neural Comput Appl 32(4):981–993
Dong HW, Hsiao WY, Yang LC, Yang YH (2018) Musegan: multi-track sequential generative adversarial networks for symbolic music generation and accompaniment. In: AAAI conference on artificial intelligence, vol 32, no 1
Walder C (2016) Modelling symbolic music: beyond the piano roll. In: Asian conference on machine learning. PMLR, pp 174–189
Skúli S (2017) How to generate music using a LSTM neural network in keras. www.shorturl.at/ezFO4
Wang J (2016) Deep learning in music. Stanford University
Eck D, Schmidhuber J (2002) A first look at music composition using lstm recurrent neural networks. Istituto Dalle Molle Di Studi Sull Intelligenza Artificiale 103:48
Heines JM (2016) Converting MIDI notes to ABC notes in pencil code. ACM inroads 7(2):84–84
Schuster M, Paliwal KK (1997) Bidirectional recurrent neural networks. IEEE Trans Signal Process
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Vemula, D.R., Tripathi, S.K., Sharma, N.K., Hussain, M.M., Swamy, U.R., Polavarapu, B.L. (2023). Music Generation Using Deep Learning. In: Kumar Singh, K., Bajpai, M.K., Sheikh Akbari, A. (eds) Machine Vision and Augmented Intelligence. Lecture Notes in Electrical Engineering, vol 1007. Springer, Singapore. https://doi.org/10.1007/978-981-99-0189-0_47
Download citation
DOI: https://doi.org/10.1007/978-981-99-0189-0_47
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-0188-3
Online ISBN: 978-981-99-0189-0
eBook Packages: Computer ScienceComputer Science (R0)