Abstract

This paper considers the problem of embedding Knowledge Graphs (KGs) consisting of entities and relations into lowdimensional vector spaces. Most of the existing methods perform this task based solely on observed facts. The only requirement is that the learned embeddings should be compatible within each individual fact. In this paper, aiming at further discovering the intrinsic geometric structure of the embedding space, we propose Semantically Smooth Embedding (SSE). The key idea of SSE is to take full advantage of additional semantic information and enforce the embedding space to be semantically smooth, i.e., entities belonging to the same semantic category will lie close to each other in the embedding space. Two manifold learning algorithms Laplacian Eigenmaps and Locally Linear Embedding are used to model the smoothness assumption. Both are formulated as geometrically based regularization terms to constrain the embedding task. We empirically evaluate SSE in two benchmark tasks of link prediction and triple classification, and achieve significant and consistent improvements over state-of-the-art methods. Furthermore, SSE is a general framework. The smoothness assumption can be imposed to a wide variety of embedding models, and it can also be constructed using other information besides entities’ semantic categories.

Highlights

  • Knowledge Graphs (KGs) like WordNet (Miller, 1995), Freebase (Bollacker et al, 2008), and DB-pedia (Lehmann et al, 2014) have become extremely useful resources for many NLP related applications, such as word sense disambiguation (Agirre et al, 2014), named entity recognition (Magnini et al, 2002), and information extraction (Hoffmann et al, 2011)

  • 2) By leveraging additional entity category information, the Semantically Smooth Embedding (SSE) models can deal with the data sparsity issue that commonly exists in most KGs

  • The metric Mean drops by about 10% to 65%, Median drops by about 5% to 75%, and Hits@10 rises by about 5% to 190%

Read more

Summary

Introduction

Knowledge Graphs (KGs) like WordNet (Miller, 1995), Freebase (Bollacker et al, 2008), and DB-pedia (Lehmann et al, 2014) have become extremely useful resources for many NLP related applications, such as word sense disambiguation (Agirre et al, 2014), named entity recognition (Magnini et al, 2002), and information extraction (Hoffmann et al, 2011). A new research direction called knowledge graph embedding has attracted much attention (Socher et al, 2013; Bordes et al, 2013; Bordes et al, 2014; Lin et al, 2015). It attempts to embed components of a KG into continuous vector spaces, so as to simplify the manipulation while preserving the inherent structure of the original graph. The learned embeddings can further be used to benefit all kinds of tasks, such as KG completion (Socher et al, 2013; Bordes et al, 2013), relation extraction (Riedel et al, 2013; Weston et al, 2013), and entity resolution (Bordes et al, 2014)

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call