Abstract

In complex networks, contrastive learning has emerged as a crucial technique for acquiring discriminative representations from graph data. Maximizing the similarity among relevant sample pairs while minimizing that among irrelevant pairs is pivotal in contrastive learning. Therefore, careful consideration must be given to the design of sample pairs in contrastive learning. However, existing node-level self-supervised contrastive learning often treats the enhanced representation of a central node as a positive sample, while considering representations of all other nodes as negative samples. This approach can lead to conflicts in downstream tasks on some graph data, as nodes of the same class are treated as negatives during learning. Precision in sample pair design is essential for enhancing the performance of contrastive learning. To address this issue, this paper introduces a negative sample debiased sampling contrastive learning (NDSCL), specifically tailored for node classification tasks. In particular, this method integrates contrastive learning with semi-supervised learning. A trained classifier assigns pseudo-labels to unlabeled data, and debiased sampling is applied to negative samples. Unlike other methods that focus on negative sample selection, NDSCL also addresses the imbalance in pseudo-label distribution by employing debiasing techniques. Finally, in conjunction with diffusion augmentation, the model is provided with diverse views as inputs to maximize the retention of underlying semantic information. Experimental results demonstrate that the proposed model significantly outperforms baseline models in node-level classification tasks across multiple network datasets. Moreover, the model not only enhances accuracy but also improves computational speed and memory requirements for handling large-scale graph data structures.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call