Fully inductive knowledge graph completion (KGC) aims to predict triplets involving both unseen entities and relations. Recent several approaches transform paths between entities into descriptions and modeling semantic correlations between paths using pre-trained language models (PLMs), have emerged as a promising solution for fully inductive reasoning. However, these methods often adopt a simplistic concatenation strategy for path-to-sentence transformation, which impedes PLMs’ ability to capture subtle nuances in context, resulting in sub-optimal path context embeddings. Furthermore, they ignore the high-order semantics underlying the complete context, which can provide richer information for inductive reasoning. To address these issues, we propose a Multi-Granularity Contextual Semantic (MGCS) modeling framework, utilizing a Path Modeling Network (PMN) and a Subgraph Modeling Network (SMN) to extract two granularity levels of contextual semantics from single paths and complete subgraphs, for fully inductive KGC. The PMN extracts paths between head and tail entities and employs reasoning patterns from similar cases to filter out unreliable paths. Then two innovative path conversion strategies are designed to significantly enhance the pre-trained language model’s understanding of specific path contexts. The SMN employs a neighbor interactive graph neural network to extract high-order semantics from the complete subgraph context with a concept-enhanced relation encoding, and optimizes it through a contrastive learning method. Finally, the confidence of the triples is evaluated from the perspective of global complete context by comparing the semantics between the subgraphs surrounding the target triplet and the subgraphs surrounding similar cases. Experimental results on benchmark datasets demonstrate the effectiveness of MGCS.
Read full abstract