Skip to main content
Log in

Sequence tagging with a rethinking structure for joint entity and relation extraction

  • Original Article
  • Published:
International Journal of Machine Learning and Cybernetics Aims and scope Submit manuscript

Abstract

Joint entity and relation extraction have become increasingly popular due to their knowledge graph construction advantages. Despite the promising results on this task, disregarding the multi-triple interaction poses a problem since multi-triples commonly appear in sentences and interact implicitly. This paper proposes a sequence tagging with a rethink structure method for joint extraction and builds upon a Seq2Seq framework. The model starts with a sentence encoder that vectorizes input text. Then the vectors are fed into a sequence tagging component to acquire pre-extracted triples. Afterward, a newly designed rethinking structure is utilized as a decoder, double-checking the pre-extracted triples. Specifically, the pre-extracted triples serve as the decoder’s query vector. Thus, the decoder can simultaneously generate triples in a non-autoregressive manner. Since the decoder is aware of the entire set of per-extracted triples, it can effectively model the interactions. The model was comprehensively compared to state-of-the-art (SOTA) benchmarks on widely used datasets, NYT and WebNLG. The result shows that the extraction performance of our model is superior to that of SOTA, especially when the sample contains sufficient triple interactions.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

Availability of data and materials

The data that support the findings of this study are available from the corresponding author on reasonable request.

References

  1. Huang YY, Wang WY (2017) Deep residual learning for weakly-supervised relation extraction. In: Proceedings of the 2017 conference on empirical methods in natural language processing, pp 1803–1807

  2. Nasir JA, Varlamis I, Ishfaq S (2019) A knowledge-based semantic framework for query expansion. Inf Process Manage 56(5):1605–1617

    Article  Google Scholar 

  3. Lai K, Porter JR, Amodeo M, Miller D, Marston M, Armal S (2022) A natural language processing approach to understanding context in the extraction and geocoding of historical floods, storms, and adaptation measures. Inf Process Manage 59(1):102735

    Article  Google Scholar 

  4. Shang F, Ran C (2022) An entity recognition model based on deep learning fusion of text feature. Inf Process Manage 59(2):102841

    Article  Google Scholar 

  5. Elnagar A, Al-Debsi R, Einea O (2020) Arabic text classification using deep learning models. Inf Process Manage 57(1):102121

    Article  Google Scholar 

  6. Katsimpras G, Paliouras G (2020) Class-aware tensor factorization for multi-relational classification. Inf Process Manage 57(2):102068

    Article  Google Scholar 

  7. Christopoulou F, Miwa M, Ananiadou S (2018) A walk-based model on entity graphs for relation extraction. In: Proceedings of the 56th annual meeting of the association for computational linguistics (volume 2: short papers), pp 81–88

  8. Misawa S, Taniguchi M, Miura Y, Ohkuma T (2017) Character-based bidirectional lstm-crf with words and characters for japanese named entity recognition. In: Proceedings of the first workshop on subword and character level models in NLP, pp 97–102

  9. Ensan F, Al-Obeidat F (2019) Relevance-based entity selection for ad hoc retrieval. Inf Process Manage 56(5):1645–1666

    Article  Google Scholar 

  10. Kate R, Mooney R (2010) Joint entity and relation extraction using card-pyramid parsing. In: Proceedings of the fourteenth conference on computational natural language learning, pp 203–212

  11. Bekoulis G, Deleu J, Demeester T, Develder C (2018) Joint entity recognition and relation extraction as a multi-head selection problem. Expert Syst Appl 114:34–45

    Article  Google Scholar 

  12. Li X, Yang J, Hu P, Liu H (2021) Laprel: a label-aware parallel network for relation extraction. Symmetry 13(6):961

    Article  Google Scholar 

  13. Zheng S, Wang F, Bao H, Hao Y, Zhou P, Xu B (2017) Joint extraction of entities and relations based on a novel tagging scheme. In: Proceedings of the 55th annual meeting of the association for computational linguistics (volume 1: long papers), pp 1227–1236

  14. Eberts M, Ulges A (2020) Span-based joint entity and relation extraction with transformer pre-training. In: Proceedings of the 24th European Conference on Artificial Intelligence. IOS Press, pp 2006–2013

  15. Yu B, Zhang Z, Sheng J, Liu T, Wang Y, Wang Y, Wang B (2021) Semi-open information extraction. In: Proceedings of the web conference 2021, pp 1661–1672

  16. Ranzato M, Chopra S, Auli M, Zaremba W (2016) Sequence level training with recurrent neural networks. In: 4th International conference on learning representations, ICLR 2016

  17. Kenton JDM-WC, Toutanova LK (2019) Bert: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 17th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp 4171–4186

  18. Zheng H, Wen R, Chen X, Yang Y, Zhang Y, Zhang Z, Zhang N, Qin B, Ming X, Zheng Y (2021) Prgc: Potential relation and global correspondence based joint relational triple extraction. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (volume 1: long papers), pp 6225–6235

  19. Dai D, Xiao X, Lyu Y, Dou S, She Q, Wang H (2019) Joint extraction of entities and overlapping relations using position-attentive sequence labeling. In: Proceedings of the AAAI conference on artificial intelligence, vol 33, pp 6300–6308

  20. Hang T, Feng J, Yan L, Wang Y, Lu J (2022) Joint extraction of entities and relations using multi-label tagging and relational alignment. Neural Comput Appl 34(8):6397–6412

    Article  Google Scholar 

  21. Hong Y, Liu Y, Yang S, Zhang K, Wen A, Hu J (2020) Improving graph convolutional networks based on relation-aware attention for end-to-end relation extraction. IEEE Access 8:51315–51323

    Article  Google Scholar 

  22. Wan Q, Wei L, Chen X, Liu J (2021) A region-based hypergraph network for joint entity-relation extraction. Knowl-Based Syst 228:107298

    Article  Google Scholar 

  23. Liu M, Zhang Y, Li W, Ji D (2020) Joint model of entity recognition and relation extraction with self-attention mechanism. ACM Trans Asian Low Resour Lang Inf Process (TALLIP) 19(4):1–19

    Article  Google Scholar 

  24. Gupta P, Schütze H, Andrassy B (2016) Table filling multi-task recurrent neural network for joint entity and relation extraction. In: Proceedings of the 26th international conference on computational linguistics: technical papers, pp 2537–2547

  25. Wang Y, Yu B, Zhang Y, Liu T, Zhu H, Sun, L (2020) Tplinker: single-stage joint extraction of entities and relations through token pair linking. In: Proceedings of the 28th international conference on computational linguistics, pp 1572–1582

  26. Zhang H, Boons F, Batista-Navarro R (2019) Whose story is it anyway? Automatic extraction of accounts from news articles. Inf Process Manage 56(5):1837–1848

    Article  Google Scholar 

  27. Briskilal J, Subalalitha C (2022) An ensemble model for classifying idioms and literal texts using bert and roberta. Inf Process Manage 59(1):102756

    Article  Google Scholar 

  28. Nayak T, Ng HT (2020) Effective modeling of encoder–decoder architecture for joint entity and relation extraction. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 8528–8535

  29. Currey A, Heafield K (2019) Incorporating source syntax into transformer-based neural machine translation. In: Proceedings of the fourth conference on machine translation (volume 1: research papers), pp 24–33

  30. Zeng X, Zeng D, He S, Liu K, Zhao J (2018) Extracting relational facts by an end-to-end neural model with copy mechanism. In: Proceedings of the 56th annual meeting of the association for computational linguistics (volume 1: long papers), pp 506–514

  31. Zeng X, He S, Zeng D, Liu K, Liu S, Zhao J (2019) Learning the extraction order of multiple relational facts in a sentence with reinforcement learning. In: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing, pp 367–377

  32. Zeng D, Zhang H, Liu Q (2020) Copymtl: copy mechanism for joint extraction of entities and relations with multi-task learning. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 9507–9514

  33. Laskar MTR, Huang X, Hoque E (2020) Contextualized embeddings based transformer encoder for sentence similarity modeling in answer selection task. In: Proceedings of The 12th language resources and evaluation conference, pp 5505–5514

  34. Ye H, Zhang N, Deng S, Chen M, Tan C, Huang F, Chen H (2021) Contrastive triple extraction with generative transformer. In: Proceedings of the AAAI conference on artificial intelligence, vol 35, pp 14257–14265

  35. Zhang RH, Liu Q, Fan AX, Ji H, Zeng D, Cheng F, Kawahara D, Kurohashi S (2020) Minimize exposure bias of seq2seq models in joint entity and relation extraction. In: Findings of the association for computational linguistics: the 2020 conference on empirical methods in natural language processing, pp 236–246

  36. Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S (2020) End-to-end object detection with transformers. In: European conference on computer vision. Springer, pp 213–229

  37. Gabrovšek B, Novak T, Povh J, Rupnik Poklukar D, Žerovnik J (2020) Multiple Hungarian method for k-assignment problem. Mathematics 8(11):2050

    Article  Google Scholar 

  38. Riedel S, Yao L, McCallum A (2010) Modeling relations and their mentions without labeled text. In: Proceedings of the 2010 European conference on machine learning and knowledge discovery in databases: Part III, pp 148–163. https://doi.org/10.1007/978-3-642-15939-8_10

  39. Gardent C, Shimorina A, Narayan S, Perez-Beltrachini L (2017) Creating training corpora for nlg micro-planning. In: 55th annual meeting of the association for computational linguistics, pp 179–188

  40. Yuan Y, Zhou X, Pan S, Zhu Q, Song Z, Guo L (2020) A relation-specific attention network for joint entity and relation extraction. In: Proceedings of the International Joint Conference on Artificial Intelligence, vol 2020, pp 4054–4060

  41. Li X, Luo X, Dong C, Yang D, Luan B, He Z (2021) Tdeer: An efficient translating decoding schema for joint extraction of entities and relations. In: Proceedings of the 2021 conference on empirical methods in natural language processing, pp 8055–8064

  42. Ren F, Zhang L, Zhao X, Yin S, Liu S, Li B (2022) A simple but effective bidirectional framework for relational triple extraction. In: Proceedings of the fifteenth ACM international conference on web search and data mining, pp 824–832

Download references

Acknowledgements

This work was supported by the National Natural Science Fund of China under Grant No. 62276095, the National Social Science Fund of China under Grant No. 20 &ZD047.

Funding

This work was supported by the National Natural Science Fund of China under Grant No. 62276095, the National Social Science Fund of China under Grant No. 20 &ZD047.

Author information

Authors and Affiliations

Authors

Contributions

DZ: conceptualization, supervision, methodology; LX: writing-original draft, software visualization, experiments; CJ: investigation; JLZ: investigation; HC: writing-reviewing and editing; JD: writing-reviewing and editing; LJ: writing-reviewing and editing.

Corresponding author

Correspondence to Jianhua Dai.

Ethics declarations

Conflict of interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Ethical approval

We declare that this manuscript is original, has not been published before, and is not currently being considered for publication elsewhere.

Consent to participate

We confirm that the manuscript has been read and approved by all named authors and that there are no other persons who satisfied the criteria for authorship but are not listed. We further confirm that the order of authors listed in the manuscript has been approved by all of us.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zeng, D., Xu, L., Jiang, C. et al. Sequence tagging with a rethinking structure for joint entity and relation extraction. Int. J. Mach. Learn. & Cyber. 15, 519–531 (2024). https://doi.org/10.1007/s13042-023-01923-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s13042-023-01923-5

Keywords

Navigation