Introduction

Extracting entities, relationships, or events from a vast amount of unstructured text is critically important for building large-scale, reusable knowledge [1,2,3,4,5]. It can promote many real-world tasks, including construction of knowledge base [6,7,8,9], automatic question and answering system [10, 11], and biomedical text mining [12,13,14]. The input consists of unstructured text, while the output comprises triples containing source entities, target entities, and their corresponding entity relationships.

Traditional methods for relation extraction, known as pipelined approaches, treat extraction as two independent subtasks: first, entity identification, and then, extraction of the relationships between them [15,16,17]. This method is flexible and straightforward, but its phased execution does not exploit deeper interactions between the subtasks. Consequently, the upstream and downstream subtasks cannot improve their extraction strategies through interactive execution. In contrast, joint entity and relationship extraction frameworks use a single model to extract both entities and relationships, achieving better performance by leveraging the relationship between the two subtasks [18, 19, 19, 20, 20,21,22]. Most notably, Takanobu and colleagues develops a hierarchical reinforcement learning relationship extraction framework called HRL-RE. By decomposing the total extraction procedure into a hierarchical structure with two reinforcement learning strategies dedicated to relation detection and entity extraction, this framework enhances the interaction between entity identification and relation-type extraction. It makes handling superimposed relationships more feasible and natural [23]. Nevertheless, this approach has not achieved satisfying results in figuring out overlapped entities and sentence relations. One of the leading factors is that the learning procedure is sluggish and has many futile attempts, leading to low efficiency of strategy learning.

To address the issues mentioned above, we present a novel relation extraction approach known as quantum hierarchical reinforcement learning for relationship extraction (QHRL-RE), which incorporates the quantum computing advantages of quantum entanglement and superposition into a hierarchical reinforcement learning relation extraction model. Specifically, drawing inspiration from the breakthroughs of quantum reinforcement learning in speech recognition and control domains [24,25,26,27], we employ quantum long short-term memory (QLSTM) network models [28] for encoding and decoding representations in relation extraction tasks. These QLSTM models can better capture long-term dependencies in unstructured text data than traditional methods. Then, we utilize a hybrid quantum-classical algorithm, which iteratively optimizes tasks applicable to relation extraction while harnessing the enhanced expressive power conferred by quantum superposition. As a result, our presented approach is more efficient in discovering superimposed entities and relationships from structureless text than traditional methods. The experiment results on two classical relationship extraction datasets, NTY10 and NTY11, demonstrate that our proposed method outperforms classical relation extraction methods with similar architectures and model parameter counts, showcasing improved performance.

The organization of this article is as follows: in the second section, a brief overview of previous relationship extraction methods is provided, along with introductions to hierarchical reinforcement learning, quantum reinforcement learning, and a brief overview of the HRL-RE approach. The third section introduces the technique, quantum hierarchical reinforcement learning for relationship extraction (QHRL-RE) presented in this paper. This section explains how this method addresses the challenges of entity and relationship extraction in cases of overlap, employing quantum hierarchical reinforcement learning techniques. Subsequently, the fourth section presents experimental results, including experiments conducted on two publicly accessible New York Times corpora to show the superiority of the proposed algorithm in relationship extraction tasks, particularly in the context of superimposed entity relationships. Finally, the fifth section summarizes the main findings and contributions of the entire paper.

Related work

This section presents a concise review of relation extraction methods, a brief review of quantum reinforcement learning methods and hierarchical reinforcement learning, and an introduction to hierarchical reinforcement learning for relation extraction.

Relations extraction

Relationship extraction plays a powerful role in information extraction applications [1, 29,30,31,32,33]. Javeed proposes a distant supervised relation extraction model based on the attention mechanism of a new relation representation [34]. Various joint extraction methods have been proposed [18,19,20, 35]. For example, Zheng et al. treat entity and relationship extraction as sequence tagging tasks. They use bidirectional LSTM and unidirectional LSTM for encoding and decoding, with the output layer simultaneously labeling entities and relationships, achieving entity relationship joint extraction [20]. Bjorne et al. introduces the concept of relation identifiers, explicitly representing phrases that indicate the presence of relationships in the sentences and then selecting their parameters to reduce the intrinsic complexity of tasks [36]. More recently, reinforcement learning has been effective in relationship extraction tasks [37,38,39,40]. Feng et al. uses reinforcement learning to discover entities and relationship types jointly [41]. Qin et al. proposes a deep reinforcement learning method for relationship extraction [40]. Feng et al. also suggests a relationship extraction method comprising an RL entity trigger and a CNN relationship identifier [42]. These approaches aim to improve the accuracy and robustness of relationship extraction by considering entity recognition and relationship identification as interconnected tasks, in contrast to the traditional pipeline methods.

Quantum reinforcement learning

Quantum reinforcement learning (QRL) can be dated to reference [43]. Nevertheless, this method requires quantifying the environment, which may not be feasible in most real environment scenarios. This paper focuses on the latest developments of variational quantum circuit (VQC)-based QRL for traditional domains. The first VQC-based QRL [44] is a quantum version of deep Q-learning (DQN) and adopts discrete state spaces and action spaces in experimental domains such as Frozen Lake. Subsequent advanced work in quantum deep Q-learning has been deemed continuous observation spaces, for example, in the Cart-Pole problem [45,46,47,48,49]. The work in [50] extends the VQC framework further to improve DQN into double-deep Q-learning (DDQN) and adopts QRL to address robot operation tasks. In addition to learning Q functions as value functions, recent developments have introduced QRL methods to learn policy functions. For example, [51] describes quantum policy gradient reinforcement learning using the REINFORCE algorithm. Subsequently, work [52] considers an enhanced policy gradient algorithm known as proximal policy optimization (PPO) with VQC and demonstrates that quantum models with few parameters can outperform classical models.

Hierarchical reinforcement learning

Hierarchical reinforcement learning (HRL) is a significant branch of reinforcement learning (RL) that distinguishes itself from classical RL methods [53,54,55]. HRL leverages hierarchical abstraction techniques to improve RL structurally, focusing on addressing challenges RL struggles with, such as sparse rewards, sequential decision-making, and weak transferability. This approach enhances exploration and transfer capabilities. Most representatively, the options framework may be the most common formalism that allows agents to reason regarding extended actions [56,57,58,59,60]. This framework models courses of action as options, which can accelerate learning in different ways, allowing, for example, faster credit assignment, planning, transfer learning, and better exploration.

Hierarchical reinforcement learning for relationship extraction

The HRL-RE approach breaks down the total entity and relationship extraction mission into two component tasks [23]. It first identifies sentence relationships and then discovers a pair of entities corresponding to that relation type. HRL-RE calculates a policy based on the states processed by a Bi-LSTM and obtains the relationship type in the high-level relationship detection subtask. Once the relationship type is received, this high-level strategy delegates the low-level component task of entity extraction. HRL-RE computes a strategy using the Monte Carlo (MC) gradient estimation approach to get the entity pair associated with that relationship in the low-level subtask. After the present low-level component task is completed, the high-level reinforcement learning component task searches for the subsequent relationship in the sentence. This hierarchical method aims to ameliorate the effectiveness of relationship extraction by breaking down the task into more manageable subtasks, with each level focusing on a specific aspect of the extraction process [23].

The HRL-RE method enhances the accuracy of entity and relationship extraction and, to some extent, addresses the issue of superimposed entities and relationships. However, this approach only achieves satisfactory results when handling superimposed entities and sentence relationships. The main reason behind this is that the learning process is cumbersome, with many ineffective attempts leading to inefficient policy learning. In cases involving superimposed entities and complex sentence structures, the learning process may need help to navigate the search space effectively. This inefficiency can hinder the method’s ability to accurately extract relationships and entities from such sentences. Improving the efficiency of the strategy learning process by optimizing the reinforcement learning algorithm or exploring alternative approaches could be a potential avenue for addressing this limitation.

Quantum hierarchical reinforcement learning for relationship extraction

This section introduces a novel approach to jointly extract superimposed entities and relationships, called quantum hierarchical reinforcement learning for relationship extraction (QHRL-RE, as shown in Figure 3). This method leverages the advantages of quantum computing, specifically quantum entanglement, and superposition, in combination with hierarchical reinforcement learning to address the problems outlined in “Hierarchical reinforcement learning for relationship extraction”. Specifically, drawing inspiration from the breakthroughs of quantum reinforcement learning in the speech recognition domain, we employ quantum long short-term memory (QLSTM) network models [28] for encoding and decoding representations in relation extraction tasks. These QLSTM models can better capture long-term dependencies in unstructured text data. In our proposed method, we utilize a hybrid quantum-classical method, which iteratively optimizes tasks applicable to relation extraction while harnessing the enhanced expressive power conferred by quantum superposition (as shown in Algorithm 1).

High-level QRL model for relationship detection

We employ the perspective of the HRL-RE approach to accomplish the high-level relation recognition mission [23]. In the high-level relations recognition component task, the sentence is scanned progressively, and the current high-level strategy \(\mathcal {O}\) (\(\mathcal {O}\in NR \cup \mathcal {R}\)) for the current time step is computed based on the state. Here, \(\mathcal {R}\) represents all the relationship types in the current dataset, and NR stands for "no relationship."

State: \(s_t^h \in \mathcal {S}\) of the high-level task at current step t is calculated as Eq. 1. It is calculated from the present hidden state \(h_t\) of the current time step t, the relationship type vector \({v}_{t}^{r}\) of the latest \(non-NR\) high-level strategy \(o'\) and the state \(s_{t-1}\) of the previous time step \(t-1\).

$$\begin{aligned} \mathbf {s_t^h}=f^h\left( \textbf{W}^h_s\left[ \textbf{qh}_t;{v}_{t}^{r};\mathbf {s_{t-1}}\right] \right) , \end{aligned}$$
(1)

where \(f^h(\cdot )\) denotes a non-linear transfer function, and \(\textbf{W}_{s}^h\) denotes a weighting matrix. To get the hidden layer state \(h_t\), we adopt a quantum Bi-LSTM over the present input word vectoring \(w_t\):

$$\begin{aligned}&\overrightarrow{\textbf{qh}_t}=\overrightarrow{QLSTM}(\overrightarrow{\textbf{qh}_{t-1}},\mathbf {w_t}) \nonumber \\&\overleftarrow{\textbf{qh}_t}=\overleftarrow{QLSTM}(\overleftarrow{\textbf{qh}_{t+1}},\mathbf {w_t}) \nonumber \\&\textbf{qh}_t=[\overrightarrow{\textbf{qh}_t},\overleftarrow{\textbf{qh}_t}]. \end{aligned}$$
(2)

The mathematical expression of QLSTM is as follows:

$$\begin{aligned}&g_t= \sigma (VQC_1(\mathbf {w_t})) \nonumber \\&\tau _t=\sigma (VQC_2(\mathbf {w_t})) \nonumber \\&\tilde{L}_t= tanh(VQC_3(\mathbf {w_t})) \nonumber \\&v_t=g_t*v_{t-1}+\tau _t*\tilde{L}_t \nonumber \\&k_t=\sigma (VQC_4(s_t)) \nonumber \\&z_t=VQC_5(k_t*tanh(v_t)) \nonumber \\&y_t=VQC_6(k_t*tanh(v_t)), \end{aligned}$$
(3)

where \(\mathbf {w_t}\) represents the input at time t, \(z_t\) represents the hidden layer state, \(v_t\) represents the cell state, and \(y_t\) represents the output, \(*\) represents element-wise multiplication (as shown in Fig. 1).

A random policy \(\mu : S\rightarrow \mathcal {O}\) is employed, computed based on the current time step state \(\mathbf {s_t^h}\) through a softmax layer. The output of the softmax layer is stochastically sampled to get the \(o_t\) behavior of the present time step:

$$\begin{aligned} o_t \sim \mu (o_t|\mathbf {s_t^h})=softmax(\textbf{W}_\mu \mathbf {s_t^h}), \end{aligned}$$
(4)

where \(\textbf{W}_\mu \) denotes a weighting matrix.

Reward: The experiment domain provides a return signal \(r_t^h\) to estimate the reward for performing policy \(o_t\):

$$\begin{aligned} r_t^h={\left\{ \begin{array}{ll}-1, if\ o_t\ not\ in\ S\\ 0,if\ o_t=NR\\ 1,if\ o_t\ in\ S.\end{array}\right. } \end{aligned}$$
(5)

Finally, the ultimate reward \(r^h_{fin}\) is gained to evaluate the sentence-level extraction manifestation that \(\mu \) discovers:

$$\begin{aligned} r^h_{fin}=F_\beta (\mathcal {R})=\frac{(1+\beta ^2)Prec \cdot Rec}{\beta ^2 Prec \cdot Rec}, \end{aligned}$$
(6)

where \(F_\beta \) denotes the weighted harmonic mean (WHM) of accuracy and recall rate in terms of the relationships in \(\mathcal {R}\).

Fig. 1
figure 1

The framework of quantum long short-term memory (QLSTM)

Low-level QRL model for entity extraction

We employ the perspective of the HRL-RE algorithm to construct the low-level entity identification mission [23]. In the low-level mission, the sentence is scanned line by line, and the action of the current time step is computed based on the state \(s_t^l\) and strategy \(\pi \). If the high-level RL policy forecasts the NR (\(non-NR\)) relationship type, the low-level RL will extract entity information within the relationship. The high-level strategy \(o_t\) from high-level RL is an additional input parameter to the low-level RL mission.

Action: this action assigns an entity tag to each word at every time step. The entity tags are represented as \(\mathcal {A} = (\{\mathcal {S},\mathcal {T},\mathcal {O}\}\times \{\mathcal {B},\mathcal {I}\}\cup \{\mathcal {U}\})\), where \(\mathcal {S}\) indicates the original entity, \(\mathcal {T}\) is the subjective entity, \(\mathcal {O}\) indicates the unrelated entity, \(\mathcal {N}\) denotes non-entity words, \(\mathcal {B}\) denotes the beginning of an entity and \(\mathcal {I}\) indicates internal parts of an entity.

State: the normative expression of the state \(s_t^l\) for the low-level task is as follows:

$$\begin{aligned}&\textbf{c}_t=g\left( \textbf{W}^l_h \mathbf {s_{t'}^h}\right) ,\nonumber \\&\mathbf {s_t^l}= f^ {l}\left( \textbf{W}^l_s\left[ \textbf{qh}_t;\mathbf {v_t^e}; \mathbf {s_{t-1}}; \textbf{c}_{t'}\right] \right) , \end{aligned}$$
(7)

where \(\textbf{qh}_t\) is the hidden state obtained from the Quantum Bi-LSTM module in Eq. (2), and \(g(\cdot )\), \(f^ {l}(\cdot )\) are non-linear functions implemented by MLP. Low-level strategy uses a randomized strategy \(\pi :\mathcal {S}\rightarrow \mathcal {A}\) to stochastically sample the probabilities output from the softmax layer to obtain the action \(a_t\) of the current time step t.

$$\begin{aligned} a_t \sim \pi \left( a_t|\mathbf {s_t^l;}o_{t'}\right) =softmax\left( \textbf{W}_\pi [o_{t'}] \mathbf {s_t^l}\right) , \end{aligned}$$
(8)

where \(\textbf{W}_\pi \) denotes an array of relation matrices \(\mathcal {R}\).

Reward: the reward \(r_t^l\) received by the present time step t is shown in Eq (8):

$$\begin{aligned} r_t^l=\lambda (y_t)\cdot sgn(a_t=y_t(o_{t'})), \end{aligned}$$
(9)

where the immediate reward \(r^l_t\) is provided when the action \(a_t\) is sampled by simply the prediction error gold-standard annotation. The function \(y_t(o_{t'})\) is the gold-standard entity tag conditioned on the predicted relationship type \(o_{t'}\), \(\lambda (y)\) is a weighting function for low-weight non-entity tag, denoted as follows:

$$\begin{aligned} \lambda (y)={\left\{ \begin{array}{ll}1, if\ y\ne N\\ \alpha ,if\ y=N.\end{array}\right. } \end{aligned}$$
(10)

The small \(\alpha \) results to less reward on words that are not entities.

Quantum hierarchical strategy learning models

Similar to HRL-RE, QHRL-RE optimizes the strategy by maximizing the expected decay cumulative return:

$$\begin{aligned} J(\theta _{\mu ,t})=E_{\mathbf {s^h},o,r^h \sim \mu (o|\mathbf {s^h})}\left[ \sum _{k=t}^T\gamma ^{k-t} r_k^h\right] , \end{aligned}$$
(11)

where high-level strategy \(\mu \) is parameterized by \(\theta _{\mu }\), \(\gamma \) denotes the decay factor in RL.

Unlike HRL-RE, QHRL-RE calculates the expected discounted cumulative return for the low-level model using the following formula:

$$\begin{aligned}&J(\theta _{\pi ,t};o_t')=E_{\mathbf {s^l},a,r^l \sim \pi \left( a|\mathbf {s^l};o_t'\right) } \nonumber \\&\qquad \left( r_k^l+\left( (1-\epsilon )J(\theta _{\mu ,t})-\epsilon \max _{\mu \in \mathcal {O}}J(\theta _{\mu ,t})\right) \right) , \end{aligned}$$
(12)

where low-level strategy \(\pi \) is parameterized by \(\theta _{\pi }\), \(\epsilon \) represents a hyperparameter, and \(A(\theta _{\pi ,t};o_{t'})\) represents the advantage function.

We decompose the expected decay cumulative rewards into a Bellman equation:

$$\begin{aligned}{} & {} R^\mu \left( \mathbf {s_t^h},o_t\right) =E\left[ \sum _{j=0}^{N-1} r_{t+j}^h \gamma ^N R^\mu \left( \mathbf {s_{t+N}^H},o_{t+N}\right) | \mathbf {s_t^h},o_t\right] ,\nonumber \\ \end{aligned}$$
(13)
$$\begin{aligned}{} & {} R^\pi \left( \mathbf {s_t^l},a_t;o_t\right) =E\left[ r_t^l+\gamma R^\pi \left( \mathbf {s_{t+1}^l},a_t;o_t\right) |\mathbf {s_t^l},o_t\right] , \nonumber \\ \end{aligned}$$
(14)

where N denotes the time steps of the entity identification component task started under the current high-level strategy \(o_t\).

The gradient for the high-level policy is defined as follows:

$$\begin{aligned} \nabla _{\theta _\mu }J(\theta _{\mu ,t})=E_{\mathbf {s^h},o,r^h \sim \mu (o|\mathbf {s^h})}\left[ R^\mu \left( \mathbf {s_t^h},o_t\right) \nabla _{\theta _\mu } \mu \left( o|\mathbf {s_t^h}\right) \right] . \end{aligned}$$
(15)

Unlike HRL-RE, QHRL-RE adopts the following Equation to update the gradient for the low-level policy:

Algorithm 1
figure a

Quantum Hierarchical Reinforcement Learning for Relation Extraction (QHRL-RE)

$$\begin{aligned} \nabla _{\theta _\pi }J(\theta _{\pi ,t;o_t'})&=E_{\mathbf {s^l}a,r^l\sim \pi (a|\mathbf {s^l}o_t')}\left[ R^\pi \left( \mathbf {s_t^l},a_t;o_t\right) \right. \nonumber \\&\quad \left. \nabla _{\theta _\pi }\pi \left( a|\mathbf {s_t^l};o_t\right) A(\theta _{\pi ,t};o_{t'})\right] . \end{aligned}$$
(16)

Experiments

Experimental setup

The dataset used in this article is the New York Times (NYT) corpus, sourced from distant supervision research, and it contains noisy relationship data [61, 62]. The corpus has two versions: (1) the traditional version generated by aligning the original data with Freebase relationships [61]. (2) A thin version of which the test set was manually annotated. We call the traditional version the NYT10 and the thin version the NTY11 [62].

Evaluation criterion: We evaluate the performance of this method using precision, recall, and micro F1 scores.

Baselines: We choose the following several entity extraction methods as the baselines.

FCM ([29]): A pipeline method that combines manually crafted lexicalized language context with word embeddings for entity and relationship extraction.

MultiR ([62]): A distant supervision approach that uses multiple weighted instances to handle noisy labels in training data.

CoType ([63]): A single approach that embeds entities, relationships, text features, and type labels into representations, treating the extraction task as a global embedding problem.

SPTree ([19]): A joint extraction approach that employs bidirectional sequential and bidirectional tree-structured LSTM-RNN in a single model to discover entities and relationships.

Tagging: A joint extraction approach that discovers entities and relationships using new labeling patterns.

CopyR ([64]): A Seq2Seq learning approach that utilizes multiple decoders to generate triplets for collectively extracting entities and relationships.

HRL-RE ([23]): A method based on HRL that breaks down the entire extraction task into high-level relations detection subtasks and low-level entity extraction subtasks.

Entities and relationships extraction

Table 1 presents the experimental results for relationship extraction. Since all methods were trained on noisy data, it is worth noting that there is a prominent difference in performance between the noisy dataset (NYT10) and the clean dataset (NYT11). It can be found that our algorithm QHRL-RE outperforms other entity relationship extraction approaches on both datasets. Crucially, the scores on the NYT10 dataset are much higher than those on the NYT11 dataset, indicating that the presented approach is more robust to noisy data.

Table 1 The experimental results for entity and relationship extraction

Superimposed entities and relations extraction

We showcase the effectiveness of our method in discovering superimposed entities and relationships on two test sets: NYT11-plus and NYT10-sub. Here, we categorize superimposed entities and relationships into two types.

  • Type 1: One entity participates in multiple relationships in the same sentence.

  • Type 2: The identical entity pair in a sentence is associated with disparate relations.

Table 2 Manifestation comparison for discovering superimposed entities and relationships
Table 3 Example of relation detection

Table 2 displays the manifestation of different entity and relationship extraction methods in extracting superimposed entities and relationships. The results of the experiment on the NYT10-sub dataset indicate that our method outperforms the HRL-RE approach. Furthermore, compared to our QHRL-RE and HRL-RE methods, other relationship extraction methods could improve when dealing with noisy data in handling the 2nd class of superimposed entities and relationships. This suggests that traditional joint extraction methods could be more effective in solving the problem of superimposed entity relationships. Therefore, our method is better suited to address the 2nd class of superimposed entity relationship problems in noisy data. Additionally, experimental results on the NYT11-plus dataset show that our method outperforms other entity relationship extraction algorithms in extracting Type 1 superimposed entities and relationships in clean data. In a word, our algorithm can extract both superimposed entities and relations more effectively.

To verify the result of our method, a sample, "Arthur Lee, the leader of Love, was born in Memphis and lived there until 1952." stochastically chosen from the dataset is exhibited in Table 3. There are three categories of relationships in this example, and the corresponding triplets are \(<Arthur\ Lee, /person/location /place\_birth, Memphis>\), \(<Arthur\ Lee, \) \(/person/location/place\_lived, Memphis>\) and \(<Arthur\ Lee, /person/\) \(leader\_of/organization, Love>\). Our method detects three relations triumphantly, and the results are exhibited in Table 3. Take the first of these relations as an example. When the high-level relation detection subtask scans the sentence to "born in," it detects the word as a relationship indicator and identifies the relationship as "\(/person/location/place\_birth\)." Then, the low-level subtask starts to scan the sentence. When it scans the word "Arthur Lee," it identifies this word as the source entity, and when it scans the word "Memphis," it identifies this word as the target entity.

Interaction between the two levels of component tasks

The results of the experiments in Table 4 demonstrate that our approach outperforms other relationship extraction approaches in the relationship detection task on both datasets. Particularly, the improvement in the NYT11-plus dataset is more pronounced, indicating that our approach is better suited for discovering multiple relationships from sentences. Thus, embedding entities as relationship parameters in relationship detection can better leverage relationship information in the text.

The performance on the NYT11 dataset exhibits slight variations when the low-level policy is omitted separately from models HRL-RE-Ent and QHRL-RE-Ent. This is because nearly every sentence in this test set contains almost only one relationship. In such cases, the interaction between high-level and low-level component task policies has minimal impact on relationship detection results. In contrast, there is a significant difference in the NYT11-plus dataset, indicating that QHRL-RE and the hierarchical reinforcement learning-based QHRL-RE can capture dependencies between multiple extraction tasks. Furthermore, this interaction can increase the rewards for high-level component task policies. Thus, the entity and relationship extraction methods based on HRL intensify the interaction between relationship detection and entity identification.

Table 4 Comparison of experiment results for relationship prediction

Conclusion

This paper presents a new relations extraction method, quantum hierarchical reinforcement learning for relation extraction (QHRL-RE), which incorporates the quantum computing advantages of quantum entanglement and superposition into a hierarchical reinforcement learning relation extraction model. Specifically, drawing inspiration from the breakthroughs of quantum reinforcement learning in speech recognition and control domains, we employ quantum long short-term memory (QLSTM) network models for encoding and decoding representations in relation extraction tasks. These QLSTM models can better capture long-term dependencies in unstructured text data. Our proposed method utilizes a hybrid quantum-classical approach, which iteratively optimizes tasks applicable to relation extraction while harnessing the enhanced expressive power conferred by quantum superposition. In this way, our QHRL-RE approach is more effective for discovering superimposed entities and relations from unstructured text. Experiments on the commonly used datasets show that our method performs better than the selected baselines. As future work, our QHRL-RE method can be generalized to many other pairwise or triple-wise extraction tasks, such as aspect-opinion mining or ontology induction.