Abstract

Deep learning models, such as convolutional neural networks, have long been applied to image and multi-media tasks, particularly those with structured data. More recently, there has been more attention to unstructured data that can be represented via graphs. These types of data are often found in health and medicine, social networks, and research data repositories. Graph convolutional neural networks have recently gained attention in the field of deep learning that takes advantage of graph-based data representation with automatic feature extraction via convolutions. Given the popularity of these methods in a wide range of applications, robust uncertainty quantification is vital. This remains a challenge for large models and unstructured datasets. Bayesian inference provides a principled approach to uncertainty quantification of model parameters for deep learning models. Although Bayesian inference has been used extensively elsewhere, its application to deep learning remains limited due to the computational requirements of the Markov Chain Monte Carlo (MCMC) methods. Recent advances in parallel computing and advanced proposal schemes in MCMC sampling methods has opened the path for Bayesian deep learning. In this paper, we present Bayesian graph convolutional neural networks that employ tempered MCMC sampling with Langevin-gradient proposal distribution implemented via parallel computing. Our results show that the proposed method can provide accuracy similar to advanced optimisers while providing uncertainty quantification for key benchmark problems.

Highlights

  • Graph neural networks are a type of artificial neural network designed for data which features graph-based representation [1]–[4]

  • EXPERIMENTS AND RESULTS We evaluate distinct features of Bayesian graph convolutional neural networks (GCNNs) in terms of computational efficiency of parallel tempering Markov Chain Monte Carlo (MCMC) sampling, effect of Langevin-gradient proposal distribution, and prediction accuracy for established benchmark datasets

  • This paper serves as a proof of concept of implementing Bayesian inference via MCMC for GCNNs, achieving comparable performance (Table 5) with traditional methods, and provides a principled approach to uncertainty quantification for deep learning models

Read more

Summary

Introduction

Graph neural networks are a type of artificial neural network designed for data which features graph-based representation [1]–[4]. We refer the reader to [24] for a further overview Deep learning methods, such as convolution neural networks [25], [26] (CNNs) and recurrent neural networks [27], [28] (RNNs) have been applied to image data and temporal sequences, but these are structured, regular, Euclidean data, though even they can be viewed as graphs (i.e., lattices). Each node has its own graph embedding via a feature vector, which summarises the properties of that particular data element. The nodes send their graph embedding to their immediate neighbours in the form of messages [39]. The message received by node v at GNN layer index t, mtv is constructed by aggregating over the set of neighbours v, N (v), the the results of a message function Mt which takes three arguments: the feature vector of v itself (htv), the

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call