Abstract

A major challenge in many machine learning tasks is that the model expressive power depends on model size. Low-rank tensor methods are an efficient tool for handling the curse of dimensionality in many large-scale machine learning models. The major challenges in training a tensor learning model include how to process the high-volume data, how to determine the tensor rank automatically, and how to estimate the uncertainty of the results. While existing tensor learning focuses on a specific task, this paper proposes a generic Bayesian framework that can be employed to solve a broad class of tensor learning problems such as tensor completion, tensor regression, and tensorized neural networks. We develop a low-rank tensor prior for automatic rank determination in nonlinear problems. Our method is implemented with both stochastic gradient Hamiltonian Monte Carlo (SGHMC) and Stein Variational Gradient Descent (SVGD). We compare the automatic rank determination and uncertainty quantification of these two solvers. We demonstrate that our proposed method can determine the tensor rank automatically and can quantify the uncertainty of the obtained results. We validate our framework on tensor completion tasks and tensorized neural network training tasks.

Highlights

  • Tensors (Kolda and Bader, 2009) are a generalization of matrices to describe and process multidimensional data arrays

  • This paper presents a Bayesian framework that is applicable to a broad class of tensor learning problems, e.g., tensor factorization/completion, tensor regression, and tensorized neural networks

  • Afterwards, we generate T = 450 stochastic gradient Hamiltonian Monte Carlo (SGHMC) samples aftering discarding the first 50 samples, or n = 10 Stein Variational Gradient Descent (SVGD) samples.We evaluate the accuracy of this model using two criterion: the predictive log likelihood (LL) and the prediction accuracy

Read more

Summary

Introduction

Tensors (Kolda and Bader, 2009) are a generalization of matrices to describe and process multidimensional data arrays. Due to its ability to represent a huge amount of data by low-rank factorization, tensor computation has been applied in data recovery and compression (Acar et al, 2011; Jain and Oh, 2014; Austin et al, 2016), machine learning (Cichocki, 2014; Novikov et al, 2015; Sidiropoulos et al, 2017), uncertainty quantification (Zhang et al, 2014, 2016), and so forth. To overcome the rank determination challenge, Bayesian methods have been employed successfully in tensor completion tasks (Chu and Ghahramani, 2009; Xiong et al, 2010; Rai et al, 2014; Zhao et al, 2015a,c; Hawkins and Zhang, 2018; Gilbert and Wells, 2019).

Objectives
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call