Abstract

As the state-of-the-art technology of Bayesian inference, based on low-dimensional principal components analysis (PCA) subspace inference methods can provide approximately accurate predictive distribution and well calibrated uncertainty. However, the main problem of PCA method is that it is a linear subspace feature extractor, and it cannot effectively represent the nonlinearly high-dimensional parameter space of deep neural networks (DNNs). Firstly, in this paper, in order to solve the main problem of the linear characteristics of PCA in high-dimensional space, we apply kernel PCA to extract higher-order statistical information in parameter space of DNNs. Secondly, to improve the efficiency of subsequent computation, we propose a strictly ordered incremental kernel PCA (InKPCA) subspace of parameter space within stochastic gradient descent (SGD) trajectories. In the proposed InKPCA subspace, we employ two approximation inference methods: elliptical slice sampling (ESS) and variational inference (VI). Finally, to further improve the memory efficiency of computing the kernel matrix, we apply Nyström approximation to determine the suitable size of subsets in the original datasets. The novelty of this paper is that it is the first time to apply the proposed InKPCA subspace with Nyström approximation for Bayesian inference in DNNs, and the results show that it can produce more accurate predictions and well-calibrated predictive uncertainty in regression and classification tasks of deep learning.

Highlights

  • In key fields where safety is at stake, such as medical diagnoses and self-driving vehicles

  • The results demonstrate that Bayesian inference method elliptical slice sampling (ESS) or variational inference (VI) in the proposed incremental kernel PCA (InKPCA) subspace outperforms the other advanced Bayesian inference in deep neural networks (DNNs)

  • The results show that the InKPCA-based subspace inference methods outperform principal components analysis (PCA)-based subspace inference approaches and other baselines including stochastic gradient descent (SGD) and SWAG

Read more

Summary

INTRODUCTION

In key fields where safety is at stake, such as medical diagnoses and self-driving vehicles. KPCA is applied to achieve higher-order statistical information in parameter space of DNNs. Secondly, to improve calculation efficiency of KPCA, strictly ordered incremental method and Nystrom approximation are adopted to construct a low-dimensional InKPCA parameter subspace. Posterior inference over weights is performed in the proposed InKPCA subspace and Bayesian model averaging can be implemented by sampling weight parameters in InKPCA subspace to achieve the uncertainty estimates of deep learning. Our contributions are as follows: 1) We are the first to replace PCA with KPCA in extracting higher-order statistical information in parameter space of DNNs. 2) We propose an improved InKPCA approach to solve the computational efficiency problem.

BAYESIAN INFERENCE WIHTIN INKPCA SUBSPACE
BAYESIAN MODEL AVERAGING
APPROXIMATE INFERENCE APPROACHES
INKPCA SUBSPACE WITHIN SGD TRAJECTORIES
IMPLEMENTING APPROXIMATE INFERENCE METHODS WITHIN INKPCA SUBSPACE
EXPERIMENTS
VISUALIZATION REGRESSION UNCERTAINTY
Findings
CONCLUSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call