Advertisement

MusicKG: Representations of Sound and Music in the Middle Ages as Linked Open Data

  • Victoria EyharabideEmail author
  • Vincent Lully
  • Florentin Morel
Open Access
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 11702)

Abstract

The World Wide Web is one of the main ways of accessing knowledge in cultural heritage. Recently, several projects in digital humanities have emerged; however only a few are specialized in musicology. In this paper, we present MusicKG, a multilingual knowledge graph about medieval musicology and musical iconography. A specific ontology has been designed to integrate data from several iconographic and musicology databases. In addition, MusicKG is connected to the Linked Open Data cloud with a significant part of its classes, properties and instances being linked to Wikidata, Getty Vocabularies, MIMO, Iconclass and GeoNames. MusicKG is accessible and reusable by three means: a downloadable RDF dump, a Virtuoso faceted browser and a public SPARQL endpoint. Some representative SPARQL query examples are given to illustrate the scope of MusicKG and to show the potential impact on the research work in medieval musicology.

Keywords

Knowledge graph Linked Open Data Ontology Musicology Cultural heritage Musical iconography 

1 Introduction

The conservation of cultural heritage is very important for humankind. Today, many cultural organizations and actors store and maintain cultural data in digital forms. Applications like virtual visit of museums and culture search portal have been developed to provide digital experiences and interactions with cultural data.

Semantic Web technologies have been used in the cultural heritage field since more than a decade. An important amount of semantic data models, vocabularies and knowledge graphs (KG) have flourished. On the data model and vocabularies side, we can mention CIDOC Conceptual Reference Model [3], Cultural-ON [8], Sampo [6] and the Getty vocabularies1. With its own data model [7], Europeana tries to facilitate the discoverability of cultural resources by collecting the resources’metadata and by centralizing them [5].

On the knowledge graph side, the Amsterdam Museum’s Linked Open Data comprises the entire collection of the Amsterdam Museum consisting of more than 70,000 object descriptions [2]. ArCo2 is a knowledge graph containing around 800.000 catalogue records of Italian cultural heritage entities (ex. archeological objects, numismatic objects). In the music domain which concerns more directly our work, we can mention the LinkedBrainz3 project that helps MusicBrainz (an open music encyclopedia that collects music metadata) publish its database as Linked Data. Last, the DOREMUS knowledge graph [1] describes classical music works and their associated events (e.g. performances in concerts). The data come from three major French cultural institutions: the French National Library, Radio France and the Philharmonie de Paris.

In this paper, we introduce MusicKG, a unique cultural heritage knowledge graph containing representations of sound and music in the Middle Ages. In Sect. 2, we describe the source data of MusicKG. In Sect. 3, we detail the ontology underlying MusicKG. In Sect. 4, we show how MusicKG is connected to the Linked Open Data cloud. In Sect. 5, we illustrate the data scope of MusicKG with several SPARQL query examples. Section 6 discusses the potential impact of MusicKG on the research in medieval musicology and concludes the paper.

2 Source Data

The data in MusicKG comes from Musiconis, a database of musical iconographies created from several partner databases4: Musicastallis, Vitrail, Metropolitan Museum (NY), Mandragore, Initiale, Sculpture, Gothic Ivories, Et Stalla, and Romane. Each of the partner databases has its own specificity, generally related to the material support of its representations. For example, the Musicastallis database catalogs musical iconographic representations presented on the carved choir stalls of religious buildings. Currently, the Musiconis database contains 2154 iconographic representations whose scenes not only contain musical but also vocal, acrobatic or choreographic performances. These scenes are deeply analyzed and each instrument is described with organological details. Figure 1 depicts a Musiconis illumination representing King David tuning his harp. In this illumination of the character “B”, it is possible to observe many details: the number of strings, the tuning key, the characteristics and the detailed form of the instrument.
Fig. 1.

King David tuning his harp - http://musiconis.huma-num.fr/en/fiche/39/x.html

3 MusicKG Data Model

We follow the W3C recommendation about “Data on the web best practices” [4]. We reuse vocabularies and resources as much as possible, including Wikidata from which the P… and Q… items listed below are taken from. In this section, we present the MusicKG data model which depicts representations of sound and music in the Middle Ages. The main class of our model is Visual artwork (Q4502142) (herein “artwork”) which represents a visual artistic work or creation (see Fig. 2). Each Visual artwork instance (see example in Fig. 3) is connected to the original sources through several predicates: official website (P856), collection (P195), inventory number (P217) and described at URL (P973). Also, each artwork instance has a title from the Musiconis database and a title from its original database described by title (P1476) and stated as (P1932) respectively.
Fig. 2.

The Visual artwork entity with its corresponding relations in our KG

Fig. 3.

Representation of the artwork instance describing the example of Fig. 1.

Images are essential for iconographic data. Generally, several images (P18) are associated with an artwork to capture all the details from different angles and with different resolutions. Regarding dates, each artwork has three different properties: start time (P580); end time (P582) and time period (P2348) that indicate the century, the date on which the artist began and finished creating the artwork respectively.

The class Visual artist (Q3391743) refers to the artist who made the artwork. An artwork is associated to its creator with the relation creator (P170). Each artist entity is portrayed with the properties birth name (P1477) and notable work (P800). In addition, we added two relations to each artwork instance: material used (P186) and fabrication method (P2079). In one hand, the relation material used describes the material an artwork is made of. This relation associates instances of artwork with Art materials (Q15303351) such as Wood (Q287) or Ivory (Q82001) for sculptures; Textile (Q28823) for embroideries and tapestry weavings; or Glass (Q11469) for stained glasses. On the other hand, the relation fabrication method relates an artwork with its Artistic technique (Q11177771), such as Sculpture technique (Q21711025) or Painting technique (Q1231896). In many cases, we have the information about the manuscript (Q87167) or the Architectural element (Q391414) to which a certain artwork belongs to. Examples of architectural elements are archivolts (Q636008), misericords (Q1938805), among many others. In those cases, we relate both entities through the relation part of (P361).

4 Linking MusicKG to the Linked Open Data Cloud

The singularity of MusicKG is its analysis of performances and the relationships between performances. This is one of the main contributions of our Knowledge Graph since, as far as we know, there are no other works that describe to this level of detail the relationships between entities within iconographic representations. Moreover, our model has been enriched with additional information coming from other popular Knowledge Graphs: Wikidata, Getty Vocabularies, Iconclass, MIMO and Geonames. Figure 4 shows an example of the interconnections between MusicKG and the aforementioned Knowledge Graphs.
Fig. 4.

Example of links between MusicKG and external KGs on the LOD cloud

We used the Wikidata entity Q35140 to represent performances. All performances are related to one or several performers (Q16010345) through the relation performer (P175) and practiced by (P3095) respectively. In our KG there are instrumental, vocal, choreographic and acrobatic performances. In the case of instrumental performances, the relation instrument (P1303) is used to associate a performance with the instrument (Q3095) played.

The property occupation (P106) establishes a relationship between a performer and a profession (Q35140) that represents their occupation or the activity they perform in the artwork. Some of the sixteenth century professions represented in our Knowledge Graph are: acrobat, singer, dancer, animal trainer, conjurer, juggler, pedagogue or partition holder. In addition, a performer may be an instance of (P31) an animal (Q729), adult (QQ9584157) or mythical entity (Q24334685); or have a sex or gender (P21) such as male (Q6581097) or female (Q6581072).

5 SPARQL Query Examples

MusicKG can be accessed by three means: a downloadable data dump, a Virtuoso faceted browser and a public SPARQL endpoint. All the information and links are available online5. MusicKG is oriented towards visual artworks, performances, performers and instruments. Users can specify techniques, materials, historical periods, etc. To illustrate the data we may retrieve from MusicKG, in Table 1, we provide two representative examples with their associated SPARQL query and result.
Table 1.

Representative competency questions, SPARQL queries and results

Example

SPARQL query

Result

Artworks using the marquetry technique

SELECT ?visualArtwork

WHERE {

?visualArtwork rdf:type wd:Q4502142 .

?visualArtwork wdt:P2079 ?technique .

?technique skos:exactMatch wd:Q1049923 .}

Musiconis100: “Two musicians playing the lute and the transverse flute”

Musiconis241: “Two bagpipe players & two dancing dogs”

Artworks depicting a rabbit playing the trumpet

SELECT ?visualArtwork

WHERE {

?visualArtwork rdf:type wd:Q4502142 .

?performance wdt:P1299 ?visualArtwork .

?performance wdt:P1303 ?trumpet .

?trumpet skos:exactMatch wd:Q8338 .

?performance wdt:P175 ?performer .

?performer wdt:P31 ?rabbit .

?rabbit skos:exactMatch wd:Q9394 .}

Musiconis299: “Rabbit playing the trumpet astride a naked man”

6 Conclusion and Future Work

In this paper, we presented MusicKG, a multilingual cultural heritage knowledge graph containing representations of sound and music in the Middle Ages. We presented respectively the source data, the ontology data model, how it is connected to external sources in the Linked Open Data Cloud and representative SPARQL queries. MusicKG may have a great impact on the research in medieval musicology, and more particularly, in musical iconography. The SPARQL endpoint allows to make more precise queries and to retrieve more accurate results. Furthermore, the connection with the LOD cloud may bring several benefits that we envisage exploiting. Wikidata items have known multilingual labels and aliases. We plan to retrieve these data more exhaustively to enable the multilingual display of the knowledge graph. We consider making the MusicKG searchable in multiple languages so that more people can access easily this unique cultural heritage database. As Wikidata is becoming the central hub for cultural heritage datasets with lots of institutions publishing their catalogue data, we will study the ingestion of MusicKG into Wikidata.

Footnotes

References

  1. 1.
    Achichi, M., Lisena, P., Todorov, K., Troncy, R., Delahousse, J.: DOREMUS: a graph of linked musical works. In: Vrandečić, D., et al. (eds.) ISWC 2018. LNCS, vol. 11137, pp. 3–19. Springer, Cham (2018).  https://doi.org/10.1007/978-3-030-00668-6_1CrossRefGoogle Scholar
  2. 2.
    de Boer, V., et al.: Amsterdam museum linked open data. Semant. Web 4(3), 237–243 (2013)Google Scholar
  3. 3.
    Doerr, M.: The CIDOC conceptual reference module: an ontological approach to semantic interoperability of metadata. AI Mag. 24(3), 75 (2003)Google Scholar
  4. 4.
    Farias Lóscio, B., Burle, C., Calegari, N. (eds.): Data on the Web Best Practices. W3C Recommendation, 31 January 2017. https://www.w3.org/TR/dwbp/
  5. 5.
    Freire, N., Meijers, E., Voorburg, R., Isaac, A.: Aggregation of cultural heritage datasets through the Web of Data. Procedia Comput. Sci. 137, 120–126 (2018)CrossRefGoogle Scholar
  6. 6.
    Hyvönen, E.: Cultural heritage linked data on the semantic web: three case studies using the sampo model. Artium, Vitoria-Gasteiz, Spain, 19–20 October 2016 (2016)Google Scholar
  7. 7.
    Isaac, A., Haslhofer, B.: Europeana linked open data–data.europeana.eu. Semant. Web 4(3), 291–297 (2013)Google Scholar
  8. 8.
    Lodi, G., et al.: Semantic web for cultural heritage valorisation. In: Vrandečić, D., et al. (eds.) Data Analytics in Digital Humanities, pp. 3–37. Springer, Cham (2018).  https://doi.org/10.1007/978-3-319-54499-1_1CrossRefGoogle Scholar

Copyright information

© The Author(s) 2019

Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license and indicate if changes were made.

The images or other third party material in this chapter are included in the chapter's Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the chapter's Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.

Authors and Affiliations

  • Victoria Eyharabide
    • 1
    Email author
  • Vincent Lully
    • 1
  • Florentin Morel
    • 2
  1. 1.STIH LaboratorySorbonne UniversityParisFrance
  2. 2.IReMus LaboratorySorbonne UniversityParisFrance

Personalised recommendations