Abstract

Entity relation extraction aims to identify the semantic relation between entity pairs in a sentence and is an important technical support for downstream tasks such as question answering systems and semantic searching. Existing relation extraction models mainly rely on neural networks to extract the semantic information of sentences, ignoring the critical role of important phrase information in relation extraction. For this problem, this paper proposes a relation extraction model based on BERT gated multi-window attention network (BERT-GMAN). The model first uses BERT to extract the semantic representation features of the sentence and its constraint information. Secondly, it constructs the key phrases extraction network to obtain multi-granularity phrase information and uses element-wise max pooling to select key phrases features. Thirdly, it adopts classification feature perception network to further filter and globally perceive key phrase feature to form the overall features of relation classification. Finally, it combines with Softmax classifier to perform relation extraction. The experimental results on the Semeval-2010 Task 8 dataset show that the performance of the model in this paper is further improved compared with the existing methods, and the F1-score reaches 90.25%.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.