Skip to main content
Log in

Joint entity and relation extraction model based on directed-relation GAT oriented to Chinese patent texts

  • Application of soft computing
  • Published:
Soft Computing Aims and scope Submit manuscript

Abstract

The joint extraction task aims to construct an entity-relation triple comprising two entities and the relation between them. Existing joint models make it difficult to process too many overlapping relations in Chinese patent texts (CPT). This article introduces a joint entity and relation extraction model based on directed-relation graph attention network (DGAT) oriented to CPT to locate this problem. First, word-character tokens are obtained from CPT using BERT as the DGAT model input. Global tokens are expanded using the BiLSTM network to enhance contextual connection from the model input. Second, the DGAT model encodes the global tokens as a fully connected graph whose nodes represent the global tokens and edges denote the relations between global tokens. The edges with directed relation in the fully connected graph are assigned weights by the DGAT model, and other edges are pruned, resulting in a directed-relation-connected graph. Finally, the entity-relation triples are decoded using conditional random fields (CRF) from the directed relation-connected graph. Experimental results show that the proposed model was highly accurate based on the CPT dataset.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Algorithm 1
Fig. 7
Fig. 8
Fig. 9
Algorithm 2
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15

Similar content being viewed by others

Data availability

Data cannot be available for privacy reasons.

References

Download references

Funding

This work was supported by the Anhui University Postgraduate Scientific Research Project (Grant No. YJS20210368), the National Natural Science Foundation of China (Grant NO.62076006), and the National Natural Science Foundation of China (Grant No. 60973050).

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Kuan-Ching Li or Shunxiang Zhang.

Ethics declarations

Conflict of interest

The author Yushan Zhao declares that she has no conflict of interest. The author Kuan-Ching Li declares that she has no conflict of interest. The author Tengke Wang declares that she has no conflict of interest. The author Shunxiang Zhang declares that she has no conflict of interest. Also this manuscript is approved by all the authors for publication. Yushan Zhao would like to declare on behalf of all the co-authors that the work described was original research that has not been published previously. All the authors listed have approved the manuscript that is enclosed.

Ethical approval

This article does not contain any studies with human participants or animals performed by any of the authors.

Informed consent

No humans or any individual participants are involved in this study.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhao, Y., Li, KC., Wang, T. et al. Joint entity and relation extraction model based on directed-relation GAT oriented to Chinese patent texts. Soft Comput (2024). https://doi.org/10.1007/s00500-024-09629-8

Download citation

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s00500-024-09629-8

Keywords

Navigation