Abstract
As an important module in most knowledge base question answering (KBQA) systems, entity and relation linking maps proper nouns and relational phrases to corresponding semantic constructs (entities and relations, respectively) in a given KB. Because different entities/relations may have the same mentions, joint disambiguation has been proposed to identify the exact entity/relation from a list of candidates using context information. Existing joint disambiguation methods, like the method in EARL (Entity and Relation Linker), mainly focus on modeling the co-occurrence probabilities of different entities and relations in input questions, while paying little attention to other non-mention expressions (e.g., wh-words). In this paper, we propose the Extended Entity and Relation Linker (EEARL), which leverages full context information to improve linking accuracy. EEARL firstly extracts the context information for each mention and the attribute features for each entity/relation via character-level and word-level LSTMs and constructs context vectors and feature vectors, respectively, and then calculates the similarity between the two vectors to re-score all the candidates. Experimental results on two benchmark datasets (LC-QuAD and QALD) show that EEARL outperforms EARL and several baseline methods in terms of both entity linking and relation linking accuracy.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.