Abstract
Joint entity and relation extraction have become increasingly popular due to their knowledge graph construction advantages. Despite the promising results on this task, disregarding the multi-triple interaction poses a problem since multi-triples commonly appear in sentences and interact implicitly. This paper proposes a sequence tagging with a rethink structure method for joint extraction and builds upon a Seq2Seq framework. The model starts with a sentence encoder that vectorizes input text. Then the vectors are fed into a sequence tagging component to acquire pre-extracted triples. Afterward, a newly designed rethinking structure is utilized as a decoder, double-checking the pre-extracted triples. Specifically, the pre-extracted triples serve as the decoder’s query vector. Thus, the decoder can simultaneously generate triples in a non-autoregressive manner. Since the decoder is aware of the entire set of per-extracted triples, it can effectively model the interactions. The model was comprehensively compared to state-of-the-art (SOTA) benchmarks on widely used datasets, NYT and WebNLG. The result shows that the extraction performance of our model is superior to that of SOTA, especially when the sample contains sufficient triple interactions.
Similar content being viewed by others
Availability of data and materials
The data that support the findings of this study are available from the corresponding author on reasonable request.
References
Huang YY, Wang WY (2017) Deep residual learning for weakly-supervised relation extraction. In: Proceedings of the 2017 conference on empirical methods in natural language processing, pp 1803–1807
Nasir JA, Varlamis I, Ishfaq S (2019) A knowledge-based semantic framework for query expansion. Inf Process Manage 56(5):1605–1617
Lai K, Porter JR, Amodeo M, Miller D, Marston M, Armal S (2022) A natural language processing approach to understanding context in the extraction and geocoding of historical floods, storms, and adaptation measures. Inf Process Manage 59(1):102735
Shang F, Ran C (2022) An entity recognition model based on deep learning fusion of text feature. Inf Process Manage 59(2):102841
Elnagar A, Al-Debsi R, Einea O (2020) Arabic text classification using deep learning models. Inf Process Manage 57(1):102121
Katsimpras G, Paliouras G (2020) Class-aware tensor factorization for multi-relational classification. Inf Process Manage 57(2):102068
Christopoulou F, Miwa M, Ananiadou S (2018) A walk-based model on entity graphs for relation extraction. In: Proceedings of the 56th annual meeting of the association for computational linguistics (volume 2: short papers), pp 81–88
Misawa S, Taniguchi M, Miura Y, Ohkuma T (2017) Character-based bidirectional lstm-crf with words and characters for japanese named entity recognition. In: Proceedings of the first workshop on subword and character level models in NLP, pp 97–102
Ensan F, Al-Obeidat F (2019) Relevance-based entity selection for ad hoc retrieval. Inf Process Manage 56(5):1645–1666
Kate R, Mooney R (2010) Joint entity and relation extraction using card-pyramid parsing. In: Proceedings of the fourteenth conference on computational natural language learning, pp 203–212
Bekoulis G, Deleu J, Demeester T, Develder C (2018) Joint entity recognition and relation extraction as a multi-head selection problem. Expert Syst Appl 114:34–45
Li X, Yang J, Hu P, Liu H (2021) Laprel: a label-aware parallel network for relation extraction. Symmetry 13(6):961
Zheng S, Wang F, Bao H, Hao Y, Zhou P, Xu B (2017) Joint extraction of entities and relations based on a novel tagging scheme. In: Proceedings of the 55th annual meeting of the association for computational linguistics (volume 1: long papers), pp 1227–1236
Eberts M, Ulges A (2020) Span-based joint entity and relation extraction with transformer pre-training. In: Proceedings of the 24th European Conference on Artificial Intelligence. IOS Press, pp 2006–2013
Yu B, Zhang Z, Sheng J, Liu T, Wang Y, Wang Y, Wang B (2021) Semi-open information extraction. In: Proceedings of the web conference 2021, pp 1661–1672
Ranzato M, Chopra S, Auli M, Zaremba W (2016) Sequence level training with recurrent neural networks. In: 4th International conference on learning representations, ICLR 2016
Kenton JDM-WC, Toutanova LK (2019) Bert: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 17th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp 4171–4186
Zheng H, Wen R, Chen X, Yang Y, Zhang Y, Zhang Z, Zhang N, Qin B, Ming X, Zheng Y (2021) Prgc: Potential relation and global correspondence based joint relational triple extraction. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (volume 1: long papers), pp 6225–6235
Dai D, Xiao X, Lyu Y, Dou S, She Q, Wang H (2019) Joint extraction of entities and overlapping relations using position-attentive sequence labeling. In: Proceedings of the AAAI conference on artificial intelligence, vol 33, pp 6300–6308
Hang T, Feng J, Yan L, Wang Y, Lu J (2022) Joint extraction of entities and relations using multi-label tagging and relational alignment. Neural Comput Appl 34(8):6397–6412
Hong Y, Liu Y, Yang S, Zhang K, Wen A, Hu J (2020) Improving graph convolutional networks based on relation-aware attention for end-to-end relation extraction. IEEE Access 8:51315–51323
Wan Q, Wei L, Chen X, Liu J (2021) A region-based hypergraph network for joint entity-relation extraction. Knowl-Based Syst 228:107298
Liu M, Zhang Y, Li W, Ji D (2020) Joint model of entity recognition and relation extraction with self-attention mechanism. ACM Trans Asian Low Resour Lang Inf Process (TALLIP) 19(4):1–19
Gupta P, Schütze H, Andrassy B (2016) Table filling multi-task recurrent neural network for joint entity and relation extraction. In: Proceedings of the 26th international conference on computational linguistics: technical papers, pp 2537–2547
Wang Y, Yu B, Zhang Y, Liu T, Zhu H, Sun, L (2020) Tplinker: single-stage joint extraction of entities and relations through token pair linking. In: Proceedings of the 28th international conference on computational linguistics, pp 1572–1582
Zhang H, Boons F, Batista-Navarro R (2019) Whose story is it anyway? Automatic extraction of accounts from news articles. Inf Process Manage 56(5):1837–1848
Briskilal J, Subalalitha C (2022) An ensemble model for classifying idioms and literal texts using bert and roberta. Inf Process Manage 59(1):102756
Nayak T, Ng HT (2020) Effective modeling of encoder–decoder architecture for joint entity and relation extraction. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 8528–8535
Currey A, Heafield K (2019) Incorporating source syntax into transformer-based neural machine translation. In: Proceedings of the fourth conference on machine translation (volume 1: research papers), pp 24–33
Zeng X, Zeng D, He S, Liu K, Zhao J (2018) Extracting relational facts by an end-to-end neural model with copy mechanism. In: Proceedings of the 56th annual meeting of the association for computational linguistics (volume 1: long papers), pp 506–514
Zeng X, He S, Zeng D, Liu K, Liu S, Zhao J (2019) Learning the extraction order of multiple relational facts in a sentence with reinforcement learning. In: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing, pp 367–377
Zeng D, Zhang H, Liu Q (2020) Copymtl: copy mechanism for joint extraction of entities and relations with multi-task learning. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 9507–9514
Laskar MTR, Huang X, Hoque E (2020) Contextualized embeddings based transformer encoder for sentence similarity modeling in answer selection task. In: Proceedings of The 12th language resources and evaluation conference, pp 5505–5514
Ye H, Zhang N, Deng S, Chen M, Tan C, Huang F, Chen H (2021) Contrastive triple extraction with generative transformer. In: Proceedings of the AAAI conference on artificial intelligence, vol 35, pp 14257–14265
Zhang RH, Liu Q, Fan AX, Ji H, Zeng D, Cheng F, Kawahara D, Kurohashi S (2020) Minimize exposure bias of seq2seq models in joint entity and relation extraction. In: Findings of the association for computational linguistics: the 2020 conference on empirical methods in natural language processing, pp 236–246
Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S (2020) End-to-end object detection with transformers. In: European conference on computer vision. Springer, pp 213–229
Gabrovšek B, Novak T, Povh J, Rupnik Poklukar D, Žerovnik J (2020) Multiple Hungarian method for k-assignment problem. Mathematics 8(11):2050
Riedel S, Yao L, McCallum A (2010) Modeling relations and their mentions without labeled text. In: Proceedings of the 2010 European conference on machine learning and knowledge discovery in databases: Part III, pp 148–163. https://doi.org/10.1007/978-3-642-15939-8_10
Gardent C, Shimorina A, Narayan S, Perez-Beltrachini L (2017) Creating training corpora for nlg micro-planning. In: 55th annual meeting of the association for computational linguistics, pp 179–188
Yuan Y, Zhou X, Pan S, Zhu Q, Song Z, Guo L (2020) A relation-specific attention network for joint entity and relation extraction. In: Proceedings of the International Joint Conference on Artificial Intelligence, vol 2020, pp 4054–4060
Li X, Luo X, Dong C, Yang D, Luan B, He Z (2021) Tdeer: An efficient translating decoding schema for joint extraction of entities and relations. In: Proceedings of the 2021 conference on empirical methods in natural language processing, pp 8055–8064
Ren F, Zhang L, Zhao X, Yin S, Liu S, Li B (2022) A simple but effective bidirectional framework for relational triple extraction. In: Proceedings of the fifteenth ACM international conference on web search and data mining, pp 824–832
Acknowledgements
This work was supported by the National Natural Science Fund of China under Grant No. 62276095, the National Social Science Fund of China under Grant No. 20 &ZD047.
Funding
This work was supported by the National Natural Science Fund of China under Grant No. 62276095, the National Social Science Fund of China under Grant No. 20 &ZD047.
Author information
Authors and Affiliations
Contributions
DZ: conceptualization, supervision, methodology; LX: writing-original draft, software visualization, experiments; CJ: investigation; JLZ: investigation; HC: writing-reviewing and editing; JD: writing-reviewing and editing; LJ: writing-reviewing and editing.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Ethical approval
We declare that this manuscript is original, has not been published before, and is not currently being considered for publication elsewhere.
Consent to participate
We confirm that the manuscript has been read and approved by all named authors and that there are no other persons who satisfied the criteria for authorship but are not listed. We further confirm that the order of authors listed in the manuscript has been approved by all of us.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Zeng, D., Xu, L., Jiang, C. et al. Sequence tagging with a rethinking structure for joint entity and relation extraction. Int. J. Mach. Learn. & Cyber. 15, 519–531 (2024). https://doi.org/10.1007/s13042-023-01923-5
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13042-023-01923-5