Abstract

Graph neural networks (GNNs) have achieved impressive success in various applications. However, training dedicated GNNs for small-scale graphs still faces many problems such as over-fitting and deficiencies in performance improvements. Traditional methods such as data augmentation are commonly used in computer vision (CV) but are barely applied to graph structure data to solve these problems. In this paper, we propose a training framework named MTDA (Multi-Task learning with Data Augmentation)-GNN, which combines data augmentation and multi-task learning to improve the node classification performance of GNN on small-scale graph data. First, we use Graph Auto-Encoders (GAE) as a link predictor, modifying the original graphs’ topological structure by promoting intra-class edges and demoting interclass edges, in this way to denoise the original graph and realize data augmentation. Then the modified graph is used as the input of the node classification model. Besides defining the node pair classification as an auxiliary task, we introduce multi-task learning during the training process, forcing the predicted labels to conform to the observed pairwise relationships and improving the model’s classification ability. In addition, we conduct an adaptive dynamic weighting strategy to distribute the weight of different tasks automatically. Experiments on benchmark data sets demonstrate that the proposed MTDA-GNN outperforms traditional GNNs in graph-based semi-supervised node classification.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call