Abstract

Knowledge Graphs (KGs) are gaining popularity and are being widely used in a plethora of applications. They owe their popularity to the fact that KGs are an ideal form to integrate and retrieve data originating from various sources. Using KGs as input for Machine Learning (ML) tasks allows to perform predictions on these popular graph structures. However, KGs cannot directly be used as ML input in their graph representation, they first require to be transformed to a vector space representation through an embedding technique. As ML techniques are data-driven, they can generalize over unseen input data that deviates to some extent from the data they were trained upon. To fully exploit the generalization capabilities of ML algorithms when using embedded KGs as input, small changes in the KGs should also result in small changes in the embedding space. Various embedding techniques for graphs in general exist, however, they have not been tailored towards embedding whole KGs, while KGs can be considered a special kind of graph that adheres to a certain KG schema. This paper evaluates if these existing embedding techniques that embed the whole graphs can represent the similarity between KGs in their embedding space, allowing ML algorithms to generalize over their input. We compare the similarities between KGs in terms of changes in size, entity labels, and KG schema. We found that most techniques were able to represent the similarities in terms of size and entity labels in their embedding space, however, none of the techniques were able to capture the similarities in KG schema.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call