Multimedia Tools and Applications

, Volume 76, Issue 4, pp 6065–6077 | Cite as

Effective music searching approach based on tag combination by exploiting prototypical acoustic content



Within the music information retrieval community, many studies and applications have focused on tag-based music categorization. The limitation in employing music tags is the ambiguity of each tag. Thus, a single music tag covers too many sub-categories. To circumvent this, multiple tags can be used simultaneously to specify music clips more precisely. However, in conventional music recommendation systems, this might not be achieved because music clips identified by the system might not be prototypical to both or each tag. In this paper, we propose a new technique for ranking proper tag combinations based on the acoustic similarity of music clips. Based on empirical experiments, proper tag combinations are suggested by our proto-typicality analysis.


Music recommendation Music tag Acoustic feature Associative tag mining 


  1. 1.
    Agrawal R and Srikant R (1994) Fast algorithm for mining association rules, Proc. Int. Conf. Very Large Data Bases, pp.487-499, Santiago, ChileGoogle Scholar
  2. 2.
    Chen W et al (2007) Document Transformation for Multi-label Feature Selection in Text Categorization, Proc. IEEE Int. Conf. Data Mining, Omaha, USAGoogle Scholar
  3. 3.
    Deng J, Leung C (2013) Music retrieval in joint emotion space using audio features and emotional tags. Lect Notes Comput Sci 7732(1):524–534CrossRefGoogle Scholar
  4. 4.
    Font F, Serra J and Serra X (2012) Folksonomy-based Tag Recommendation for Online Audio Clip Sharing, Proc. Int. Soc. Music Information Retrieval, pp. 73–78, Porto, PortugalGoogle Scholar
  5. 5.
    Font F, Serra J, Serra X (2014) Class-based tag recommendation and user-based evaluation in online audio clip sharing. Knowl Syst 67(1):131–142CrossRefGoogle Scholar
  6. 6.
    Fu Z, Lu G, Ting K, Zhang D (2011) A survey of audio-based music classification and annotation. IEEE Trans Multimed 13(2):303–319CrossRefGoogle Scholar
  7. 7.
    Huron D (2000) Perceptual and cognitive applications in music information retrieval. Perception 10(1):83–92CrossRefGoogle Scholar
  8. 8.
    Juslin P (2000) Cue utilization in communication of emotion in music performance. J Exp Psychol Hum Percep Per 26(1):1797–1813CrossRefGoogle Scholar
  9. 9.
    Lamere P (2008) Social tagging and music information retrieval. J New Music Res 37(2):101–114CrossRefGoogle Scholar
  10. 10.
    Lartillot O and Toiviainen P (2007) MIR in Matlab (II): A Toolbox for Musical Feature Extraction from Audio, Proc. Int. Soc. Music Information Retrieval, pp. 237–244, Vienna, AustriaGoogle Scholar
  11. 11.
    Lin YC, Yang YY and Homer H (2011) Exploiting Online Music Tags for Music Emotion Classification, ACM Trans. Multimedia Comput. Comm. Appl., 78(1):Article 26Google Scholar
  12. 12.
    Liu B, Hsu W, Ma Y (1998) Integrating classification and association rule mining. Proc. Int. Conf Knowledge Discovery and Data Mining, New YorkGoogle Scholar
  13. 13.
    Miotto R and Orio N (2012) A Probabilistic Model to Combine Tags and Acoustic Similarity for Music Retrieval, ACM Trans. Information Systems, 30(2):Article 8Google Scholar
  14. 14.
    Nanopoulos A and Karydis I (2011) Know Thy Neighbor: Combining Audio Features and Social Tags for Effective Music Similarity, Proc. IEEE Int. Conf. Acoustics, Speech and Signal Processing, pp. 165–168, Prague, CzechGoogle Scholar
  15. 15.
    Ness S R, Theocharis A, Tzanetakis G, Martins L G (2009) Improving automatic music tag annotation using stacked generalization of probabilistic SVM outputs, Proc. 17th ACM Int. Conf. Multimedia, pp. 705–708, Beijing, ChinaGoogle Scholar
  16. 16.
    Saari P et al (2013) The role of audio and tags in music mood prediction: a study using semantic layer projection, Proc. 14th Int. Soc. Music Information Retrieval, pp. 201–206, Curitiba, BrazilGoogle Scholar
  17. 17.
    Song Y, Dixon S, Pearce M and Halpern A (2013) Do Online Social Tags predict Perceived or Induced Emotional Responses to Music?, Proc. 14rd Int. Soc. Music Information Retrieval, pp. 89–94, Curitiba, BrazilGoogle Scholar
  18. 18.
    Tax D (2000) Data Description in Subspaces, Proc. 15th Int. Conf. Pattern Recognition, pp. 672–675, Barcelona, SpainGoogle Scholar
  19. 19.
    Tingle D, Kim YE, Turnbull D (2010) Exploring automatic music annotation with “acoustically-objective” tags, Proc. 11th Int. ACM Conf. Multimedia Information Retrieval, pp. 55–62, Philadelphia, USAGoogle Scholar
  20. 20.
    Trohidis K, Tsoumakas K, Kalliris G, Vlahavas I (2008) Multi-label classification of music into emotions. Int. Soc Music Information Retrieval, PhiladelphiaGoogle Scholar
  21. 21.
    Turnbull D, Barrington L, Lanckriet G, and Yazdani M (2009) Combining Audio Content and Social Context for Semantic Music Discovery, Proc. 32nd Int. ACM SIGIR Conf. Research and Development in Information Retrieval, pp. 387–394, Boston, USAGoogle Scholar
  22. 22.
    Turnbull D, Barrington L, Torres D, Lanckriet G (2008) Semantic annotation and retrieval of music and sound effects. IEEE Trans Audio Speech Lang Proc 16(2):467–476CrossRefGoogle Scholar
  23. 23.
    Wang DD, Li T and Ogihara M (2010) Are Tags better than Audio Features? The Effect of Joint Use of Tags and Audio Content Features for Artistic Style Clustering, Proc. 11th Int. Soc. Music Information Retrieval, pp. 57–62, Utrecht, NetherlandsGoogle Scholar
  24. 24.
    Wang JC, Shih YC, Wu MS, Wang HM and Jeng SK (2011) Colorizing tags in tag cloud: a novel query-by-tag music search system, Proc. 19th ACM Int. Conf. Multimedia, pp. 293–302, Scottsdale, USAGoogle Scholar
  25. 25.
    Yang Y-H, Lin Y-C, Su Y-F, Chen H (2008) A regression approach to music emotion recognition. IEEE Trans Audio Speech Lang Proc 16(2):448–457CrossRefGoogle Scholar
  26. 26.
    Zhao Z, Wang X, Xiang Q, Sarroff A, Li Z, and Wang Y (2010) Large-scale Music Tag Recommendation with Explicit Multiple Attributes, Proc. 18th Int. Conf. Multimedia, pp.204-410, Firenze, ItalyGoogle Scholar

Copyright information

© Springer Science+Business Media New York 2016

Authors and Affiliations

  1. 1.School of Computer Science and EngineeringChung-Ang UniversitySeoulKorea

Personalised recommendations