Abstract

Graph embedding learns low-dimensional vector representations which capture and preserve information in original graphs. Common shallow neural networks and deep autoencoder only use adjacency matrix as input, and usually ignore node attributes and features. Shallow graph neural networks cannot spread the node characteristic information on a large scale. Many deep models suffer the problem of over-smoothing. Therefore, these methods can't fully incorporate network information. In this paper, we propose a novel Semi-AttentionAE model to fully utilize node features, node labels, and network structure. More specifically, we integrate a supervised information extraction graph attention network to capture both node features and network structure, with an unsupervised feature extraction autoencoder to reduce dimension while preserving structure information. Finally, ensemble learning is introduced to jointly train the combined model to obtain final embedding. We conduct the node classification and visualization experiments on four real-world datasets, including two citation networks, one co-occurrence network, and one commodity network. The results suggest that the proposed Semi-AttentionAE model is capable of embedding both graph structure and node features. The integrated model has successfully exceeded or matched performance across four well-established baselines.

Highlights

  • Many real-world data can be represented as graphs [1], such as molecular structures, physical models, social networks [2], and traffic networks [3]

  • In order to solve the shortcomings of previous methods, and fully utilizes node features, node labels, and network structure, we proposed a scalable semi-supervised deep graph embedding model Semi-AttentionAE

  • In order to learn the hidden features generated by Graph Attention Network (GAT), Semi-AttentionAE uses undercomplete AE to embed the main features of nodes into low-dimensional vector space

Read more

Summary

INTRODUCTION

Many real-world data can be represented as graphs [1], such as molecular structures, physical models, social networks [2], and traffic networks [3]. Graph embedding algorithms mainly use factorization-based methods to learn the representations These approaches relate to the properties of the connection matrix. The shallow neural network model can be applied to giant graph and generate embeddings containing original topology, but the node features and node labels are ignored. GCN is a classical GNN for graph representation learning It defines a convolution operator on original graph and iteratively aggregates the embedding of neighbors for a node. In order to solve the shortcomings of previous methods, and fully utilizes node features, node labels, and network structure, we proposed a scalable semi-supervised deep graph embedding model Semi-AttentionAE. GCN uses convolution operation to extract features from graph, and generates embedding for each node. GCN encodes graph structure and node features by neural network model f and generates representation Y

AUTOENCODER
LAPLACIAN EIGENMAPS
SEMI-ATTENTIONAE MODEL
EXPERIMENTS
BASELINES
Findings
CONCLUSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call