Keywords

1 Motivation

Due to the variability of natural language, a relation can be expressed in a wide variety of ways. When counting how often a certain pattern is used to express a relation (e.g. which movie is starring which actor), the distribution has a very long tail: frequently used patterns make up only a small fraction; the majority of expressions use rare patterns (see Welty et al., [18]). While it would be possible to manually create patterns for a small set of languages, this would be a tedious task, results would not necessarily be correct, and coverage would most likely be far from optimal due to the size of the long tail. Thus, automatically extracting a set of syntactical variants of relations from text corpora would ease this task considerably.

However, there are numerous challenges associated to automating this task. It is essential to capture the context in which such a pattern applies. Typically, all of the information conveyed in a sentence is crucial to disambiguate the meaning of a relation expressed in text. Thus, a rich meaning representation is needed that goes beyond simple patterns consisting of named entity pairs and the string in-between them. Furthermore, semantically related relations need to be detected, grouped and linked to existing formalized knowledge. The latter is essential, if the meaning of the learned representations need to be related to human conceptualizations of knowledge, like questions answering over linked data. Finally, another dimension of complexity arises when we also consider the variability of natural language across different languages (e.g., English and Spanish). Then, finding patterns, aligning semantically related ones across languages, and linking them to one existing formal knowledge representations requires the learning of a cross-lingual semantic representation of relations expressed in text of different languages.

Unsupervised learning of distributional semantic representations from textual data has received increasing attention in recent years [10], since such representations have shown to be useful for solving tasks like document comparison, information retrieval and question answering. However, research has focused almost exclusively on the syntactic level and on single languages. At the same time, there has been progress in the area of cross-lingual entity disambiguation and linking, but this work is mostly confined to (named) entities and does not extend to other expressions in text, like the phrases indicating the relations between entities. What is missing so far is a representation that links linguistic variations of semantically related and contextualized textual elements across languages to their corresponding relation in a knowledge base.

In this paper, we present the first approach to unsupervised clustering of semantically related and cross-lingual relations expressed in text. This is achieved by combining multi-lingual semantic role labeling (SRL) with cross-lingual entity linking followed by spectral clustering of the resulting annotated SRL graphs. The resulting cross-lingual semantic representation of relations is, whenever possible, linked to English DBpedia properties, and enables e.g., to extend the schema with new properties, or to support cross-lingual question answering over linked data systems.

In our initial implementation we built a cross-lingual library of relation expressions from English and Spanish Wikipedia articles containing 25,000 SRL graphs with 2000 annotations to DBpedia entities. To demonstrate the usefulness of this novel language resource we show its performance on the Multi-lingual Question Answering over Linked Data challenge (QALD-4)Footnote 1. Our results show that we can clearly outperform baseline approaches in respect to correctly linking (English) DBpedia properties in the SPARQL queries, specifically in a cross-lingual setting where the question to be answered is provided in Spanish.

In summary, the main contributions of our proposed approach to extract, cluster and link contextualized relation expressions in text are the following:

  • Relation expressions can be extracted from text of different languages and are not restricted to a predefined set of relations (as defined by DBpedia).

  • Extracted expressions are embedded in a semantic graph, describing the context this expression appears in.

  • Semantically related relation expressions and their associated context are disambiguated and clustered across languages.

  • If existing, relation clusters are linked to their corresponding property in the English DBpedia.

In the remainder of this paper we first discuss related work, before introducing our approach to learning a cross-lingual semantic representation of grounded relations (Sects. 3, 4, 5 and 6). In Sect. 7 we evaluate our initial implementation on the QALD-4 benchmark and conclude in Sect. 8.

2 Related Work

Lewis and Steedman [7] present an approach to learning clusters of semantically equivalent English and French binary relations between referring expressions. Similar to us, a cluster is a language-independent semantic representation that can be applied to a variety of tasks such as translation, relation extraction, summarization, question answering, and information retrieval. The main difference is that we perform clustering on Semantic Role Label (SRL) graphs – thus operating on an abstract meaning representation - instead of binary syntactic relations. A meaning representation is more language-independent than a syntactic representation (like string patterns or dependency trees) since it abstracts from grammatical variations of different languages. This facilitates the learning of cross-lingual and language-independent semantic representations. This basic difference applies to almost all of the remaining approaches listed in this section, like Lin and Pantel (DIRT, [8]), who learn textual inference rules such as (“X wrote Y”, “X is the author of Y”) from dependency-parsed sentences by building groups of similar dependency paths.

An additional difference of related approaches like [5, 11, 14, 16, 17] is their dependency on preexisting knowledge base properties. In contrast, our approach does not start from a predefined set of knowledge base property for which we learn textual representations, but instead derives clusters of textual expressions via Semantic Role Labeling first for which we then try to find a corresponding relation in the KB. Thus, our approach is not confined to finding relations preexisting in a knowledge base. Newly identified relations could even be used for extending the ontology. This, however, would be contribution to ontology learning and is out of the scope of this paper. The approaches restricted to preexisting KB relations (and shallow parsing) are discussed in more detail now. Walter et al. (M-ATOLL, [17]) learn dependency paths as natural language expressions for KB relations. They begin with a relation from DBpedia, retrieve triples for this relation and search within a text corpus for sentences where the two arguments of the relation can be found within a sentence. The sentence is dependency-parsed and, given a set of six dependency patterns, a pattern matches the dependency tree. Mahendra et al. ([11]) learn textual expressions of DBpedia relations from Wikipedia articles. Given a relation, triples are retrieved and sentences are identified where the two arguments of the relation can be found within a sentence. The longest common substring between the entities in sentences collected for a relation is learned as the relation’s textual expression. Vila et al. (WRPA, [16]) learn English and Spanish paraphrases from Wikipedia for four pre-specified relations. Textual triples are derived using data from an article’s infobox and its name. The string between the arguments of a relation within a sentence is extracted and generalized and regular expressions are created. Gerber and Ngonga Ngomo (BOA, [5]) language-independently learn textual expressions of DBpedia relations from Wikipedia by regarding the strings between a relation’s arguments within sentences. Nakashole et al. (PATTY, [14]) learn textual expressions of KB relations from dependency-parsed or POS-tagged English sentences. Textual expressions are sequences of words, POS-tags, wildcards, and ontological types.

In contrast to the work just mentioned, there are a few approaches that leverage a semantic representation. Grounded Unsupervised Semantic Parsing by Poon (GUSP, see [15]) translates natural-language questions to database queries via a learned probabilistic grammar. However, GUSP is not cross-lingual. Similarly, Exner and Nugues [4] learn mappings from PropBank to DBpedia based on Semantic Role Labeling. Relations in Wikipedia articles are detected via SRL, named entities are identified and linked to DBpedia and use these links to ground PropBank relations to DBpedia. Again, this is not cross-lingual.

To the best of our knowledge, our approach is the only one that (i) extracts potentially novel relations and (ii) where possible, links to preexisting relation in a KB and (iii) does this across languages by exploiting a language-independent semantic representation rather than a syntactic one.

3 A Pipeline for Learning a Cross-Lingual Semantic Representations of Grounded Relations

Our pipeline, as shown in Fig. 1, consists of three major stages.

In the first stage (see Sect. 4), the multi-lingual text documents are transformed and processed by Semantic Role Labeling (SRL). In our evaluation we use Wikipedia articles as input data, but any text that produces valid SRL graphs is feasible. Please note, to construct a cross-lingual representation a multi-lingual comparable corpus covering similar topics is advisable. However, there is no need for a aligned or parallel corpus. SRL produces semantic graphs of frames with predicates and associated semantic role-argument pairs. In parallel, we apply cross-lingual entity linking to the same text documents. This detects entity mentions in multi-lingual text and annotates the corresponding mention strings with the entity URI originating exclusively from the English DBpedia. After that, we combine and align the output of both, SRL and entity linking in order to extract a cross-lingual SRL graphs. The only remaining language-dependent elements in a cross-lingual SRL graph are the predicate nodes.

Fig. 1.
figure 1

Schematic summary of the processing pipeline.

The next stage performs relational learning of cross-lingual clusters (Sect. 5) on the previously acquired annotated SRL graphs. The similarity metrics that we define in Sect. 5.1 are central to this stage of the pipeline.

In the subsequent third stage, the obtained clusters are linked to DBpedia properties. Section 6 describes this procedure in greater detail. As a result we get cross-lingual clusters of annotated SRL graphs, i.e. textual relation expressions, augmented with a ranked set of DBpedia properties. Ultimately, these grounded clusters of relation expressions are evaluated in the task of property linking on multi-lingual questions of the QALD-4 dataset.

4 Extracting and Annotating SRL Graphs

Multi-lingual Semantic Role Labeling is performed on the input text independently for every language. SRL is accomplished by means of shallow and deep linguistic processing as described in [9]. The result of this processing step is a semantic graph consisting of semantic frames with predicates and their arguments. Each semantic frame is represented as a tree with the predicate as the root and its arguments as the leaf nodes. The edges are given by the semantic roles of the predicate arguments (cmp. Fig. 2).

Fig. 2.
figure 2

Example sentence with corresponding partial XML outputs produced by SRL (frame element) and the cross-lingual entity linking tool (DetectedTopic elements).

SRL graphs are directed, node and edge labelled graphs describing the content of a whole document. Several predicates appear in one graph, so one sub-tree per predicate is extracted for clustering (the predicate being the root of the tree), resulting in a few trees per sentence and many trees per document. Trees from one document contain partially duplicated information. Formally, an SRL graph is a set of triples \(t=(p,r,v)\) where the predicate \(p\) belongs to a set of SRL predicates (\(p \in P_{SRL}\)), the role \(r\) belongs to a set of SRL roles (\(r \in R_{SRL}\)), and \(v\) is either a string value or an SRL predicate (\(v \in P_{SRL} \cup String)\). We consider a frame as valid, if it has at least two non-frame arguments. Such a constraint reduces the number of usable frames, which, in turn is compensated by the large amount of the raw textual data.

The example in Fig. 2 demonstrates the operation of the SRL pipeline, beginning with an example sentence for which the semantic frame is obtained. To achieve cross-lingual SRL graphs role labels of non-English SRL graphs are mapped to their corresponding English role labels. Whenever possible SRL predicates from all languages are linked to English wordnet synsets. That’s not always possible since not every phrase of a predicate in an extracted SRL graph is mentioned in WordNet, specifically for non-English languages.

The next step towards generating cross-lingual SRL graphs is cross-lingual entity linking to the English DBpedia. This language-independent representation of the predicate arguments provides additional cross-lingual context for the subsequent predicate cluster analysis.

We treat this step as a replaceable black-box component by using the approach described in [19]. Reference [19] relies on linkage information in different Wikipedia language versions (language links, hyper links, disambiguation pages, ...) plus a statistical cross-lingual text comparison function, trained on a comparable corpora. The cross-lingual nature of our analysis is achieved by mapping text mentions in both languages to the English-language DBpedia URIs. The bottom part of Fig. 2 is a sample of the annotation output for the above example sentence. Annotations that correspond to SRL arguments are enclosed in URL attributes of DetectedTopic elements.

Fig. 3.
figure 3

Cross-lingual SRL graphs extracted from English and Spanish sentences.

The intermediate results of both, the SRL and annotation steps finally need to be combined in order to extract the actual graphs. Figure 3 contains an example of four sentences along with the extracted cross-lingual SRL graphs from English and Spanish sentences. The graph vertices show the SRL predicate and argument mention strings along with DBpedia URIs (dbr namespace http://dbpedia.org/resource/) and Wordnet-IDs. Edge labels specify the semantic role. Obviously, the graphs on the top and on the bottom are more similar to each other compared to the graphs on the level and right, respectively. Thus, cross-lingual SRL graphs are similar regarding the content, not the language.

5 Learning a Cross-Lingual Semantic Representation of Relation Expressions

For the purpose of clustering a set of cross-lingual SRL graphs we introduce a set of metrics specifying a semantic distance of SRL graphs (see Sect. 5.1). Section 5.2 discusses the spectral clustering algorithm.

5.1 Constructing Similarity Matrices of Annotated SRL Graphs

Goal of this step is to construct a similarity matrix, specifying the pair-wise similarity of all SRL graphs. We tried three different graph-similarity metrics \(m_1, m_2, m_3\).

Formally, a cross-lingual SRL graph is an SRL graph where \(v\) is either a string value, an SRL predicate, or a unique identifier (\(v \in P_{SRL} \cup String \cup \mathcal {U})\). \(g(p)\) denotes the graph with predicate \(p\) as the root SRL predicate. \(m_{1}:G \times G \rightarrow \lbrace 1;0 \rbrace \) compares the SRL graphs’ root predicates according to their names, e.g. exist.01 vs. meet.02:

$$\begin{aligned} m_1(g_i,g_j) := \left\{ \begin{array}{l} 1\ , p(g_i) = p(g_j) \\ 0 \ , else \end{array} \right. \end{aligned}$$
(1)

\(m_{2}:G \times G \rightarrow \left[ 1;0 \right] \) compares two SRL graphs’ root predicates according to their annotated role values:

$$\begin{aligned} m_{2}(g_i,g_j) := \frac{ \vert A(g_i) \cap A(g_j) \vert }{ \vert A(g_i) \cup A(g_j) \vert } \end{aligned}$$
(2)

where \(A(g_k) := \lbrace v ~|~ \exists r \in R_{SRL} : (p(g_k),r,v)\in g_k \wedge v \in \mathcal {U} \rbrace \).

\(m_{3}:G \times G \rightarrow \left[ 1;0 \right] \) compares two SRL graphs’ root predicates according to their role labels:

$$\begin{aligned} m_{3}(g_i,g_j) := \frac{ \vert B(g_i) \cap B(g_j) \vert }{ \vert B(g_i) \cup B(g_j) \vert } \end{aligned}$$
(3)

where \(B(g_k) := \lbrace r ~|~ \exists v \in P_{SRL} \cup String \cup \mathcal {U} : (p(g_k),r,v)\in g_k \rbrace \).

Now, given the set of cross-lingual SRL graphs \(\lbrace g_1, ... g_n\rbrace \) and given the three SRL predicate similarity metrics, we can construct three SRL predicate similarity matrices. Each SRL predicate similarity metric is applied for pairwise comparison of two (annotated) SRL graphs’ root predicates. The root predicate \(p\) of an (annotated) SRL graph \(g\), denoted by \(p(g)\), is the predicate for which no triple \((p_2,r,p) \in g\) exists with \(p \ne p_2\). \(G\) denotes the set of all SRL graphs.

Based on a separate evaluation of each metric we introduce a combined similarity metric as a weighted sum of the three single metrics.

5.2 Spectral Clustering of Annotated SRL Graphs

Spectral Clustering uses the spectrum of a matrix derived from distances between different instances. Using the spectrum of a matrix has been successfully used in many computer vision applications [12] and is also applicable for similarity matrices. As input a similarity matrix \(S\) derived from one metric or a weighted combination of several metrics is given. As a first step the Laplacian matrix \(L\) is built by subtracting the similarity matrix \(S\) from the diagonal matrix \(D\) which contains the sum of each row on the diagonal (respectively column since \(S\) is symmetric) (Eq. 4).

$$\begin{aligned} L_{ij} = D_{ij} - S_{ij} = \left\{ \begin{array}{ll} {\sum }_m{S_{im}} - S_{ij} = {\sum }_m{S_{mj}} - S_{ij} &{}\text{ if } i = j \\ - S_{ij} &{} \text{ otherwise } \end{array}\right. \end{aligned}$$
(4)

For building \(k\) clusters, the second up to the \(k + 1\) smallest eigenvalue and corresponding eigenvector of the Laplacian matrix are calculated. Afterwards the actual clustering starts with running the \(k\)-means algorithm on the eigenvectors which finally results in a clustering for the instances of S.

To enforce the learning of cross-lingual clusters, we introduce the weighting matrix \(W\) which is used to weight the mono- and cross-lingual relations in the similarity matrix \(S\) (Eq. 5). While setting the monolingual weight \(w_{monolingual}\) to zero, forces the construction of only cross-lingual clusters, we received better results by setting \(w_{monolingual} > 0\). This can be intuitively understood as we get more clean clusters when we don’t force cross-lingual relations into every cluster, as there is no guarantee that a matching cross-lingual relation even exists. Finally the weighted matrix \(S^*\), the result of the product \(W\) and \(S\) (Eq. 6), is given as input to the previously described spectral clustering algorithm.

$$\begin{aligned} W_{ij} = \left\{ \begin{array}{ll} w_{monolingual} &{}\text{ if } \text{ i } \text{ and } \text{ j } \text{ are } \text{ monolingual }\\ 1 &{}\text{ if } \text{ i } \text{ and } \text{ j } \text{ are } \text{ crosslingual } \end{array}\right. \end{aligned}$$
(5)
$$\begin{aligned} S^*_{ij} = W_{ij} \cdot S_{ij} \end{aligned}$$
(6)

6 Linking Annotated SRL Graph Clusters to DBpedia Properties

In order to find potential links of the obtained clusters to DBpedia properties, we exploit the SRL graphs’ argument structure as well as the DBpedia entity URIs provided by cross-lingual entity linking. The origin of possible candidates is limited to the DBpedia ontologyFootnote 2 and infoboxFootnote 3 properties.

Acquisition of Candidate Properties. For a given annotated SRL graph we retrieve a list of candidate properties by querying DBpedia for the in- and outbound properties associated with its arguments’ entities. Consequently, the candidate properties of an entire predicate cluster are determined by the union of the individual graphs’ candidate lists. Several specific properties, such as the Wikipedia-related structural properties (e.g. wikiPageID, wikiPageRevisionID etc.) are excluded from the candidate list.

Scoring of Candidate Properties. After the construction of the candidate list, the contained properties are scored. The purpose behind this is to determine a ranking of properties by their relevance with respect to a given cluster. In principle, several different scoring approaches are applicable to the underlying problem. For example, a relative normalized frequency score of property \(p_{i}\) w.r.t. cluster \(C_{j}\) calculated as

$$S^{rnf}(p_{i}, C_{j})=\frac{relative \ frequency \ of \ p_{i} \ in \ C_{j}}{relative \ frequency \ of \ p_{i} \ over \ all \ clusters}$$

is appropriate to reflect the importance as well as the exclusiveness of property \(i\) for cluster \(j\). However, our experiments determined the absolute frequency score of a property within a cluster to be the best performing measure.

Algorithm 1 shows the structure of the complete grounding algorithm in a simplified form. This algorithm is similar to the approach by Exner and Nugues [4].

figure a

7 Evaluation on Cross-Lingual Relation Linking for Question Answering over Linked Data

We make use of the evaluation data set provided by the Multi-lingual Question Answering over Linked Data challenge (task 1 of QALD-4). The data set contains 200 questions (12 out of 200 are out-of-scope w.r.t DBpedia knowledge base) in multiple languages as well as corresponding gold-standard SPARQL queries against DBpedia.

To evaluate the quality of our results, we conducted property linking experiments. We deliberately concentrate on the sub-task of property linking to avoid distortion of the performance by various pre- and post-processing steps of a full QA-system. Linking the properties necessary for constructing the SPARQL query constitutes an important step of a question answering system such as QAKiS [1], SemSearch [6], ORAKEL [2], FREyA [3], and TcruziKB [13] which generate SPARQL queries based on user input.

7.1 Linking Properties in the QALD Challenge

First, we generated compatible data representation from the QALD-4 question sentences by sending them through stage 1 of our processing pipeline (see Sect. 3). Hereby we obtained cross-lingual SRL graphs for English and Spanish questions.

Next, using our similarity metrics and the previously learned grounded clusters, we classified each individual SRL graph of the questions set and determined its target cluster. Consequently, each SRL graph of the questions set was assigned DBpedia properties according to the groundings of its associated target cluster. This way, for each question, our approach linked properties, which were finally evaluated against the gold-standard properties of the QALD-4 training dataset.

7.2 Data Set and Baselines

We employed Wikipedia as the source of multi-lingual text documents in the English (EN, Wikipedia dump version 2013.04.03) and Spanish (ES, Wikipedia dump version 2012.05.15) language. Over 23,000,000 cross-lingual annotated SRL graphs were extracted from more than 300,000 pairs of language link-connected English and Spanish Wikipedia articles.

In order to get an initial assessment of our approach we conducted our experiments on two samples of the original data. Table 1 provides an overview of the key dataset statistics. Dataset 1 consists of a random sample of long Wikipedia article pairs, which together sum up to approximately 25,000 SRL graph instances. The second sample with a similar number of graphs was derived from randomly selected short article pairs in order to provide a wider coverage of different topics and corresponding DBpedia entities.

Table 1. Key statistics of the data sets used for our experiments.

Baseline 1: String Similarity-Based Property Linking. This first naïve baseline links properties based on string similarity between the question tokens and DBpedia property labels. Given a question from the QALD-4 training dataset, we firstly obtain the question tokens using the Penn treebank-trained tokenizer. In the next step, each token is assigned the one DBpedia property with the highest string similarity between its label and the token string. String similarity is measured by means of the normalized Damerau-Levenshtein distance. For each token, the one property with the highest label similarity enters the candidate set. Finally, the identified candidate properties are evaluated against the QALD-4 gold-standard properties. Because the vast majority of property labels are of English origin, we could not apply this baseline to Spanish QALD-4 data.

Baseline 2: Entity-Based Property Linking. Baseline 2 takes a more sophisticated approach to finding good candidate properties. For this baseline, we first use the set of entities associated with a given question for linking of candidate properties exactly the same way as we perform grounding of cross-lingual SRL graph clusters (Sect. 5.1). In the next step, the list of candidate properties is pruned by thresholding the normalized Damerau-Levenshtein similarity of their labels to the question tokens. Again, this will have negative effect on the performance for Spanish-language questions for the same reasons as discussed in Sect. 7.2. We report results for two variations of this baseline, which differ in the mode of entity retrieval for a given question: In the first case, entities are collected from the cross-lingual annotated SRL graphs, while in the second case we obtain the entities directly from the output of the entity linking tool.

Table 2. Performance of Baseline 2 without and with SRL graph extraction.

7.3 Evaluation Results

Baseline 1: Results. A naïve selection of candidate properties based solely on string similarity between the question tokens and property labels shows poor overall performance on the English-language QALD-4 questions:

  • precision: 2.15 %

  • recall: 10.68 %

  • F1-measure: 3.58 %

As discussed in Sect. 7.2, this baseline is limited to English-language questions.

Baseline 2: Results. The top part of Table 2 shows the performance of Baseline 2 in the case without SRL graph extraction.

Due to the cross-lingual nature of property linking through our grounding algorithm, there is a clear performance increase for Spanish-language questions. It is also notable that the behaviour of the performance measure is consistent over all string similarity thresholds for both languages. The bottom part of Table 2 shows Baseline 2 results with SRL graph extraction. Here, we see a small but consistent performance increase for the English language over Baseline 2 without SRL. This observation supports our assumption that the inclusion of the semantic structure of annotated arguments as provided by Semantic Role Labeling does improve performance.

Table 3. Best performing grounded clusters configurations for QALD-4 questions.
Table 4. Best performing results for “short articles” vs “long articles”.

Results with Grounded Cross-Lingual SRL Graph Clusters. The evaluation of our approach was conducted on the previously described (Table 1) experimental datasets and a variety of different clustering configurations with respect to different similarity matrices as well as different internal parameter sets of the spectral clustering algorithm.

Table 3 reports the results of several top performing configurations. It is notable that across languages and different parameter sets, the completely cross-lingual, entity-focused metric \(m_{2}\) outperforms the other configurations, which supports the basic idea of our approach. In addition to this, we observe a consistent improvement over our baselines for English, and even more so for the Spanish language.

To investigate the effect of input data and parameter choice on the quality of results, we conducted further experiments, which involved grounded clusters computed on a weighted sum of all metrics with cross-lingual constraints. In particular, we demonstrate the effect of the short- versus long-articles dataset, i.e. the impact of more diverse input data. Table 4 shows results of this comparison. Obviously, shorter and more concise articles seem to produce SRL graphs with more meaningful clusters. It would be interesting to evaluate whether co-reference resolution would improve the performance for longer articles.

Another aspect of interest is the effect of the number of Eigenvectors within the spectral clustering algorithm. This parameter greatly increases the computational resources needed to compute the clustering. But our experimental results also clearly show an advantage of a high number of Eigenvectors (Table 5).

Both experiments revealed that more input data as well as higher-dimensional clustering has the potential to further improve the performance of our approach. Another incentive for scaling those dimensions is to cover the long tail of relation expressions. Still, we would argue that this limited evaluation clearly demonstrates the benefits of our approach, since we outperform Baseline 2 by about 6 % and Baseline 2 is comparable to what is used in most of the related work. That shows a big potential to improve those QA systems.

Table 5. Best performing results in respect to number of eigenvectors.

8 Conclusion and Future Work

This paper introduces an approach to unsupervised learning of a cross-lingual semantic representation of relations expressed in text. To the best of our knowledge this is the first meaning representation induced from text that is (i) cross-lingual, (ii) builds on semantic instead of shallow syntactic features, and (iii) generalizes over relation expressions. The resulting clusters of semantically related relation graphs can be linked to DBpedia properties and thus support tasks like question answering over linked data. Our results show that we can clearly outperform baseline approaches on the sub-task of property linking.

Directions for future work include, learning the semantic representation from more documents. Our current implementation serves as a strong proof-of-concept, but does not yet cover the long tail of relation expressions sufficiently. Including all Wikipedia articles resulting in millions of graphs is merely an engineering challenge, only the clustering step would need to be adjusted. In addition, we would like to assess the potential of our approach to discover novel relation-types (and their instantiations) to the knowledge base.