Theoretical Aspects of Music Definition Language and Music Manipulation Language

  • Hanchao LiEmail author
  • Xiang FeiEmail author
  • Ming YangEmail author
  • Kuo-ming ChaoEmail author
  • Chaobo HeEmail author
Conference paper
Part of the Lecture Notes on Data Engineering and Communications Technologies book series (LNDECT, volume 41)


As there is an increase in the amount of on-line music and digital albums, the way of express the music has become more and more important nowadays, so it can be used in several Music Information Retrieval tasks. For example, the music search engine and the plagiarism detection tool. From the existing symbolic approaches, there are a lot of music notation and relevant theories so the musician can understand and follow easily. Thus, we need to build a theoretical system for our proposed coding scheme, naming Music Definition Language and Music Manipulation Language. Therefore, this paper is focused on some of the important theories that derived from MDL and MML. After the theoretical analysis and practical discussion, we have showed the relationship between the proposed coding scheme and the existing audio or symbolic formats. These proved the feasibility of using the new coding scheme for those MIR tasks.


Audio Music Definition Language Music Manipulation Language Theoretical aspects Symbolic 



This paper involves the melodies from the following music track(s):

1. Mozart, W. A. (1806) Twinkle, Twinkle, Little Star (a.k.a. Twelve Variations On “Ah Vous Dirai-Je, Maman”, K.265).

2. BASTARZ (Block B) (Lee, M., Kim, Y. and Pyo, J.) (2015)  Open image in new window(Zero For Conduct).

2a. D.J.S 137 (covers Bastarz of Block B) (2015) BASTARZ (Block b) - Zero For ConductOpen image in new window(Piano Tutorial) [Sheets + MIDI].

This work was partially supported by the following projects:

The Science and Technology Support Program of Guangdong Province of China under Grant 2017A040405057, and in part by the Science and Technology Support Program of Guangzhou City of China under Grant 201807010043 and Grant 201803020033.


  1. 1.
    Casey, M., Veltkamp, R., Goto, M., Leman, M., Rhodes, C., Slaney, M.: Content-based music information retrieval: current directions and future challenges. Proc. IEEE 96(4), 668–696 (2008)CrossRefGoogle Scholar
  2. 2.
    Lamere, P.: Social tagging and music information retrieval. J. New Music Res. 37(2), 101–114 (2008)CrossRefGoogle Scholar
  3. 3.
    Typke, R., Wiering, F., Veltkamp, R.: A survey of music information retrieval systems. In: Proceedings of 6th International Conference on Music Information Retrieval. pp. 153–160 (2005)Google Scholar
  4. 4.
    Duong, N., Duong, H.: A review of audio features and statistical models exploited for voice pattern design. In: Seventh International Conferences on Pervasive Patterns and Applications (PATTERNS 2015, Nice, France) (2015)Google Scholar
  5. 5.
    Fan, Y., Feng, S.: A music identification system based on audio fingerprint. In: 2016 4th International Conference on Applied Computing and Information Technology/3rd International Conference on Computational Science/Intelligence and Applied Informatics/1st International Conference on Big Data, Cloud Computing, Data Science & Engineering (ACIT-CSII-BCD), pp. 363–367 (2016)Google Scholar
  6. 6.
    Lubiw, A., Tanur, L.: Pattern matching in polyphonic music as a weighted geometric translation problem. In: Proceedings 5th International Conference on Music Information Retrieval, pp. 289–296 (2004)Google Scholar
  7. 7.
    McKay, C., Fujinaga, I.: Jsymbolic: A feature extractor for MIDI files. In: Proceedings of International Computer Music Conference, pp. 302–305 (2006)Google Scholar
  8. 8.
    Huang, T., Xia, G., Ma, Y., Dannenberg, R., Faloutsos, C.: MidiFind: fast and effective similarity searching in large MIDI databases. In: Proceedings of the 10th International Symposium on Computer Music Multidisciplinary Research, pp. 209–224 (2013)Google Scholar
  9. 9.
    Saxena, A., Jasola, S., Barik, K., Joshi, A., Dubey, A., Jauhari, M.: Unit-9: the computer system: hardware for educational computing. In: Block-3: Introduction to Computers in Education, pp. 5–18 (2018)Google Scholar
  10. 10.
    Dixon, S.: On the computer recognition of solo piano music. In: Proceedings of Australasian computer music conference, pp. 31–37 (2000)Google Scholar
  11. 11.
    Chen, C., Jang, J., Liu, W., Weng, C.: An efficient method for polyphonic audio-to-score alignment using onset detection and constant Q transform. In: 2016 IEEE International Conference on Acoustics, Speech And Signal Processing (ICASSP), pp. 2802–2806 (2016)Google Scholar
  12. 12.
    Cogliati, A., Temperley, D., Duan, Z.: Transcribing human piano performances into music notation. In: Proceedings of 17th International Conference on Music Information Retrieval, pp. 758–764 (2016)Google Scholar
  13. 13.
    Fournier-S’niehotta, R., Rigaux, P., Travers, N.: Is there a data model in music notation? In: International Conference on Technologies for Music Notation and Representation (TENOR 2016), pp. 85–91 (2016)Google Scholar
  14. 14.
    Yang, N., Usman, M., He, X., Jan, M., Zhang, L.: Time-frequency filter bank: a simple approach for audio and music separation. IEEE Access 5, 27114–27125 (2017)CrossRefGoogle Scholar
  15. 15.
    Li, H., Fei, X., Chao, K., Yang, M., He, C.: Towards a hybrid deep-learning method for music classification and similarity measurement. In: 2016 IEEE 13th International Conference on e-Business Engineering (ICEBE), pp. 9–16 (2016)Google Scholar

Copyright information

© Springer Nature Switzerland AG 2020

Authors and Affiliations

  1. 1.The Faculty of EECCoventry UniversityCoventryUK
  2. 2.School of Information Science and TechnologyZhongKai University of Agriculture and EngineeringGuangzhouChina

Personalised recommendations