Abstract

Chinese knowledge base question answering (KBQA) is designed to answer the questions with the facts contained in a knowledge base. This task can be divided into two subtasks: topic entity extraction and relation selection. During the topic entity extraction stage, an entity extraction model is built to locate topic entities in questions. The Levenshtein Ratio entity linker is proposed to conduct effective entity linking. All the relevant subject-predicate-object (SPO) triples to topic entity are searched from the knowledge base as candidates. In relation selection, an attention-based multi-granularity interaction model (ABMGIM) is proposed. Two main contributions are as follows. First, a multi-granularity approach for text embedding is proposed. A nested character-level and word-level approach is used to concatenate the pre-trained embedding of a character with corresponding embedding on word-level. Second, we apply a hierarchical matching model for question representation in relation selection tasks, and attention mechanisms are imported for a fine-grained alignment between characters for relation selection. Experimental results show that our model achieves a competitive performance on the public dataset, which demonstrates its effectiveness.

Highlights

  • Open-domain question answering is a challenging task that aims at providing corresponding answers to natural language questions

  • An increasing amount of research work focuses on knowledge base question answering (KBQA) [5,6]

  • The evaluation of a KBQA system is generally considered by precision, recall, averaged F1 and accuracy@N

Read more

Summary

Introduction

Open-domain question answering is a challenging task that aims at providing corresponding answers to natural language questions. Typical examples include knowledge bases in English such as Freebase [1], DBpedia [2], and Chinese knowledge bases like zhishi.me [3], XLore [4], and CN-DBpedia (http://kw.fudan.edu.cn/cndbpedia/). Due to their structured form of knowledge, knowledge bases have become a significant resource of open-domain question answering. For KBQA, the answer to the target question is definitely extracted from knowledge bases, so the major challenge is to understand the query and pick up the best subject-predicate-object (SPO) triple from knowledge bases. Given a question “特朗普是什么时候出生的? || When was Trump born?” the task is first to locate an entity from the knowledge base that contains an entity like “唐纳德·特朗普 || Donald Trump” that describes the Information 2018, 9, 98; doi:10.3390/info9040098 www.mdpi.com/journal/information

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call