High-dimensional random landscapes: From typical to large deviations

  • Abstract
  • Literature Map
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

We discuss tools and concepts that emerge when studying high-dimensional random landscapes, i.e., random functions on high-dimensional spaces. As an illustrative example, we consider an inference problem in two forms: low-rank matrix estimation (case 1) and low-rank tensor estimation (case 2). We show how to map the inference problem onto the optimization problem of a high-dimensional landscape, which exhibits distinct geometrical properties in the two cases. We discuss methods for characterizing typical realizations of these landscapes and their optimization through local dynamics. We conclude by highlighting connections between the landscape problem and large deviation theory.

Similar Papers
  • Conference Article
  • Cite Count Icon 13
  • 10.1109/isit.2018.8437628
Adaptive Path Interpolation for Sparse Systems: Application to a Simple Censored Block Model
  • Jun 1, 2018
  • Jean Barbier + 2 more

A new adaptive path interpolation method has been recently developed as a simple and versatile scheme to calculate exactly the asymptotic mutual information of Bayesian inference problems defined on dense factor graphs. These include random linear and generalized estimation, superposition codes, or low rank matrix and tensor estimation. For all these systems the method directly proves in a unified manner that the replica symmetric prediction is exact. When the underlying factor graph of the inference problem is sparse the replica prediction is considerably more complicated and rigorous results are often lacking or obtained by rather complicated methods. In this contribution we extend the adaptive path interpolation method to sparse systems. We concentrate on a Censored Block Model, where hidden variables are measured through a binary erasure channel, for which we fully prove the replica prediction.

  • Research Article
  • Cite Count Icon 2
  • 10.1016/j.procs.2015.05.420
Stable Autoencoding: A Flexible Framework for Regularized Low-rank Matrix Estimation
  • Jan 1, 2015
  • Procedia Computer Science
  • Julie Josse + 1 more

Stable Autoencoding: A Flexible Framework for Regularized Low-rank Matrix Estimation

  • Research Article
  • Cite Count Icon 19
  • 10.1109/tnnls.2018.2844242
Unified Low-Rank Matrix Estimate via Penalized Matrix Least Squares Approximation.
  • Jul 2, 2018
  • IEEE Transactions on Neural Networks and Learning Systems
  • Xiangyu Chang + 3 more

Low-rank matrix estimation arises in a number of statistical and machine learning tasks. In particular, the coefficient matrix is considered to have a low-rank structure in multivariate linear regression and multivariate quantile regression. In this paper, we propose a method called penalized matrix least squares approximation (PMLSA) toward a unified yet simple low-rank matrix estimate. Specifically, PMLSA can transform many different types of low-rank matrix estimation problems into their asymptotically equivalent least-squares forms, which can be efficiently solved by a popular matrix fast iterative shrinkage-thresholding algorithm. Furthermore, we derive analytic degrees of freedom for PMLSA, with which a Bayesian information criterion (BIC)-type criterion is developed to select the tuning parameters. The estimated rank based on the BIC-type criterion is verified to be asymptotically consistent with the true rank under mild conditions. Extensive experimental studies are performed to confirm our assertion.

  • Research Article
  • Cite Count Icon 6
  • 10.1016/j.ijar.2012.07.003
Large and moderate deviations for random sets and random upper semicontinuous functions
  • Jul 15, 2012
  • International Journal of Approximate Reasoning
  • Xia Wang

Large and moderate deviations for random sets and random upper semicontinuous functions

  • Research Article
  • Cite Count Icon 3
  • 10.1088/1361-6560/acd6d1
Low tensor train and low multilinear rank approximations of 3D tensors for compression and de-speckling of optical coherence tomography images
  • Jun 8, 2023
  • Physics in Medicine & Biology
  • Ivica Kopriva + 5 more

Objective. Many methods for compression and/or de-speckling of 3D optical coherence tomography (OCT) images operate on a slice-by-slice basis and, consequently, ignore spatial relations between the B-scans. Thus, we develop compression ratio (CR)-constrained low tensor train (TT)—and low multilinear (ML) rank approximations of 3D tensors for compression and de-speckling of 3D OCT images. Due to inherent denoising mechanism of low-rank approximation, compressed image is often even of better quality than the raw image it is based on. Approach. We formulate CR-constrained low rank approximations of 3D tensor as parallel non-convex non-smooth optimization problems implemented by alternating direction method of multipliers of unfolded tensors. In contrast to patch- and sparsity-based OCT image compression methods, proposed approach does not require clean images for dictionary learning, enables CR as high as 60:1, and it is fast. In contrast to deep networks based OCT image compression, proposed approach is training free and does not require any supervised data pre-processing. Main results. Proposed methodology is evaluated on twenty four images of a retina acquired on Topcon 3D OCT-1000 scanner, and twenty images of a retina acquired on Big Vision BV1000 3D OCT scanner. For the first dataset, statistical significance analysis shows that for CR ≤ 35, all low ML rank approximations and Schatten-0 (S 0) norm constrained low TT rank approximation can be useful for machine learning-based diagnostics by using segmented retina layers. Also for CR ≤ 35, S 0-constrained ML rank approximation and S 0-constrained low TT rank approximation can be useful for visual inspection-based diagnostics. For the second dataset, statistical significance analysis shows that for CR ≤ 60 all low ML rank approximations as well as S 0 and S 1/2 low TT ranks approximations can be useful for machine learning-based diagnostics by using segmented retina layers. Also, for CR ≤ 60, low ML rank approximations constrained with S p , p ∊ {0, 1/2, 2/3} and one surrogate of S 0 can be useful for visual inspection-based diagnostics. That is also true for low TT rank approximations constrained with S p , p ∊ {0, 1/2, 2/3} for CR ≤ 20. Significance. Studies conducted on datasets acquired by two different types of scanners confirmed capabilities of proposed framework that, for a wide range of CRs, yields de-speckled 3D OCT images suitable for clinical data archiving and remote consultation, for visual inspection-based diagnosis and for machine learning-based diagnosis by using segmented retina layers.

  • Research Article
  • Cite Count Icon 24
  • 10.1137/19m1307329
Low Rank Pure Quaternion Approximation for Pure Quaternion Matrices
  • Jan 1, 2021
  • SIAM Journal on Matrix Analysis and Applications
  • Guangjing Song + 2 more

Quaternion matrices are employed successfully in many color image processing applications. In particular, a pure quaternion matrix can be used to represent red, green, and blue channels of color images. A low-rank approximation for a pure quaternion matrix can be obtained by using the quaternion singular value decomposition. However, this approximation is not optimal in the sense that the resulting low-rank approximation matrix may not be pure quaternion, i.e., the low-rank matrix contains a real component which is not useful for the representation of a color image. The main contribution of this paper is to find an optimal rank-$r$ pure quaternion matrix approximation for a pure quaternion matrix (a color image). Our idea is to use a projection on a low-rank quaternion matrix manifold and a projection on a quaternion matrix with zero real component, and develop an alternating projections algorithm to find such optimal low-rank pure quaternion matrix approximation. The convergence of the projection algorithm can be established by showing that the low-rank quaternion matrix manifold and the zero real component quaternion matrix manifold has a nontrivial intersection point. Numerical examples on synthetic pure quaternion matrices and color images are presented to illustrate the projection algorithm can find optimal low-rank pure quaternion approximation for pure quaternion matrices or color images.

  • Conference Article
  • Cite Count Icon 72
  • 10.1109/itw.2016.7606798
Mutual information in rank-one matrix estimation
  • Sep 1, 2016
  • Florent Krzakala + 2 more

We consider the estimation of a n-dimensional vector x from the knowledge of noisy and possibility non-linear element-wise measurements of xxT , a very generic problem that contains, e.g. stochastic 2-block model, submatrix localization or the spike perturbation of random matrices. We use an interpolation method proposed by Guerra and later refined by Korada and Macris. We prove that the Bethe mutual information (related to the Bethe free energy and conjectured to be exact by Lesieur et al. on the basis of the non-rigorous cavity method) always yields an upper bound to the exact mutual information. We also provide a lower bound using a similar technique. For concreteness, we illustrate our findings on the sparse PCA problem, and observe that (a) our bounds match for a large region of parameters and (b) that it exists a phase transition in a region where the spectum remains uninformative. While we present only the case of rank-one symmetric matrix estimation, our proof technique is readily extendable to low-rank symmetric matrix or low-rank symmetric tensor estimation

  • Book Chapter
  • Cite Count Icon 6
  • 10.1007/978-3-030-43120-4_10
CUR LRA at Sublinear Cost Based on Volume Maximization
  • Jan 1, 2020
  • Qi Luan + 1 more

A matrix algorithm runs at sublinear cost if it uses much fewer memory cells and arithmetic operations than the input matrix has entries. Such algorithms are indispensable for Big Data Mining and Analysis, where input matrices are so immense that one can only access a small fraction of all their entries. Typically, however, such matrices admit their Low Rank Approximation (LRA), which one can access and process at sublinear cost. Can, however, we compute LRA at sublinear cost? Adversary argument shows that no algorithm running at sublinear cost can output accurate LRA of worst case input matrices or even of the matrices of small families of our Appendix A, but we prove that some sublinear cost algorithms output a reasonably close LRA of a matrix W if (i) this matrix is sufficiently close to a low rank matrix or (ii) it is a Symmetric Positive Semidefinite (SPSD) matrix that admits LRA. In both cases supporting algorithms are deterministic and output LRA in its special form of CUR LRA, particularly memory efficient. The design of our algorithms and the proof of their correctness rely on the results of extensive previous study of CUR LRA in Numerical Linear Algebra using volume maximization. In case (i) we apply Cross-Approximation (C-A) iterations, running at sublinear cost and computing accurate LRA worldwide for more than a decade. We provide the first formal support for this long-known empirical efficiency assuming non-degeneracy of the initial submatrix of at least one C-A iteration. We cannot ensure non-degeneracy at sublinear cost for a worst case input but prove that it holds with a high probability (whp) for any initialization in the case of a random or randomized input. Empirically we can replace randomization with sparse multiplicative preprocessing of an input matrix, performed at sublinear cost. In case (ii) we make no additional assumptions about the input class of SPSD matrices admitting LRA or about initialization of our sublinear cost algorithms for CUR LRA, which promise to be practically valuable. We hope that proper combination of our deterministic techniques with randomized LRA methods, popular among Computer Science researchers, will lead them to further progress in LRA.

  • Research Article
  • Cite Count Icon 2
  • 10.1137/20m1381046
Guarantees for Existence of a Best Canonical Polyadic Approximation of a Noisy Low-Rank Tensor
  • Mar 1, 2022
  • SIAM Journal on Matrix Analysis and Applications
  • Eric Evert + 1 more

The canonical polyadic decomposition (CPD) of a low-rank tensor plays a major role in data analysis and signal processing by allowing for unique recovery of underlying factors. However, it is well known that the low-rank CPD approximation problem is ill-posed. That is, a tensor may fail to have a best rank $R$ CPD approximation when $R>1$. This article gives deterministic bounds for the existence of best low-rank tensor approximations over ${\mathbb{K}}={\mathbb{R}}$ or ${\mathbb{K}}={\mathbb{C}}$. More precisely, given a tensor ${\mathcal T} \in {\mathbb{K}}^{I \times I \times I}$ of rank $R \leq I$, we compute the radius of a Frobenius norm ball centered at ${\mathcal T}$ in which best ${\mathbb{K}}$-rank $R$ approximations are guaranteed to exist. In addition we show that every ${\mathbb{K}}$-rank $R$ tensor inside of this ball has a unique canonical polyadic decomposition. This neighborhood may be interpreted as a neighborhood of "mathematical truth" in which CPD approximation and computation are well-posed. In pursuit of these bounds, we describe low-rank tensor decomposition as a "joint generalized eigenvalue" problem. Using this framework, we show that, under mild assumptions, a low-rank tensor which has rank strictly greater than border rank is defective in the sense of algebraic and geometric multiplicities for joint generalized eigenvalues. Bounds for existence of best low-rank approximations are then obtained by establishing perturbation theoretic results for the joint generalized eigenvalue problem. In this way we establish a connection between existence of best low-rank approximations and the tensor spectral norm. In addition we solve a "tensor Procrustes problem" which examines orthogonal compressions for pairs of tensors. The main results of the article are illustrated by a variety of numerical experiments.

  • Research Article
  • Cite Count Icon 6
  • 10.1007/s11425-016-0426-8
Degrees of freedom in low rank matrix estimation
  • Nov 16, 2016
  • Science China Mathematics
  • Ming Yuan

The objective of this paper is to quantify the complexity of rank and nuclear norm constrained methods for low rank matrix estimation problems. Specifically, we derive analytic forms of the degrees of freedom for these types of estimators in several common settings. These results provide efficient ways of comparing different estimators and eliciting tuning parameters. Moreover, our analyses reveal new insights on the behavior of these low rank matrix estimators. These observations are of great theoretical and practical importance. In particular, they suggest that, contrary to conventional wisdom, for rank constrained estimators the total number of free parameters underestimates the degrees of freedom, whereas for nuclear norm penalization, it overestimates the degrees of freedom. In addition, when using most model selection criteria to choose the tuning parameter for nuclear norm penalization, it oftentimes suffices to entertain a finite number of candidates as opposed to a continuum of choices. Numerical examples are also presented to illustrate the practical implications of our results.

  • Research Article
  • Cite Count Icon 101
  • 10.1088/1742-5468/aa7284
Constrained low-rank matrix estimation: phase transitions, approximate message passing and applications
  • Jul 1, 2017
  • Journal of Statistical Mechanics: Theory and Experiment
  • Thibault Lesieur + 2 more

This article is an extended version of previous work of Lesieur et al (2015 IEEE Int. Symp. on Information Theory Proc. pp 1635–9 and 2015 53rd Annual Allerton Conf. on Communication, Control and Computing (IEEE) pp 680–7) on low-rank matrix estimation in the presence of constraints on the factors into which the matrix is factorized. Low-rank matrix factorization is one of the basic methods used in data analysis for unsupervised learning of relevant features and other types of dimensionality reduction. We present a framework to study the constrained low-rank matrix estimation for a general prior on the factors, and a general output channel through which the matrix is observed. We draw a parallel with the study of vector-spin glass models—presenting a unifying way to study a number of problems considered previously in separate statistical physics works. We present a number of applications for the problem in data analysis. We derive in detail a general form of the low-rank approximate message passing (Low-RAMP) algorithm, that is known in statistical physics as the TAP equations. We thus unify the derivation of the TAP equations for models as different as the Sherrington–Kirkpatrick model, the restricted Boltzmann machine, the Hopfield model or vector (xy, Heisenberg and other) spin glasses. The state evolution of the Low-RAMP algorithm is also derived, and is equivalent to the replica symmetric solution for the large class of vector-spin glass models. In the section devoted to result we study in detail phase diagrams and phase transitions for the Bayes-optimal inference in low-rank matrix estimation. We present a typology of phase transitions and their relation to performance of algorithms such as the Low-RAMP or commonly used spectral methods.

  • Book Chapter
  • Cite Count Icon 316
  • 10.1007/bfb0015522
Eigenfaces vs. Fisherfaces: Recognition using class specific linear projection
  • Jan 1, 1996
  • Peter N Belhumeur + 2 more

We develop a face recognition algorithm which is insensitive to gross variation in lighting direction and facial expression. Taking a pattern classification approach, we consider each pixel in an image as a coordinate in a high-dimensional space. We take advantage of the observation that the images of a particular face under varying illumination direction lie in a 3-D linear subspace of the high dimensional feature space — if the face is a Lambertian surface without self-shadowing. However, since faces are not truly Lambertian surfaces and do indeed produce self-shadowing, images will deviate from this linear subspace. Rather than explicitly modeling this deviation, we project the image into a subspace in a manner which discounts those regions of the face with large deviation. Our projection method is based on Fisher's Linear Discriminant and produces well separated classes in a low-dimensional subspace even under severe variation in lighting and facial expressions. The Eigenface technique, another method based on linearly projecting the image space to a low dimensional subspace, has similar computational requirements. Yet, extensive experimental results demonstrate that the proposed “Fisherface” method has error rates that are significantly lower than those of the Eigenface technique when tested on the same database.

  • Research Article
  • Cite Count Icon 16
  • 10.1016/j.neucom.2018.02.057
Bayesian inference for adaptive low rank and sparse matrix estimation
  • Feb 21, 2018
  • Neurocomputing
  • Xixi Jia + 4 more

Bayesian inference for adaptive low rank and sparse matrix estimation

  • Research Article
  • Cite Count Icon 5
  • 10.1109/tsp.2007.909353
Heteroscedastic Low-Rank Matrix Approximation by the Wiberg Algorithm
  • Apr 1, 2008
  • IEEE Transactions on Signal Processing
  • Pei Chen

Low-rank matrix approximation has applications in many fields, such as 2-D filter design and 3-D reconstruction from an image sequence. In this paper, one issue with low-rank matrix approximation is investigated: heteroscedastic noise. In most of previous research, the covariance matrix of the heteroscedastic noise is assumed to be positive definite. This requirement restricts the usefulness of results derived from such research. In this paper, we extend the Wiberg algorithm, which originally deals with the missing data problem with low-rank approximation, to the cases, where the heteroscedastic noise has a singular covariance matrix. Experiments show that the proposed Wiberg algorithm converges much faster than the bilinear approach, and consequently avoids many nonconvergent cases in the bilinear approach. Experiments also show that, to some extent, the Wiberg algorithm can tolerate outliers and is not sensitive to parameter variation.

  • Research Article
  • Cite Count Icon 85
  • 10.1007/s11263-008-0135-7
Optimization Algorithms on Subspaces: Revisiting Missing Data Problem in Low-Rank Matrix
  • Apr 17, 2008
  • International Journal of Computer Vision
  • Pei Chen

Low-rank matrix approximation has applications in many fields, such as 3D reconstruction from an image sequence and 2D filter design. In this paper, one issue with low-rank matrix approximation is re-investigated: the missing data problem. Much effort was devoted to this problem, and the Wiberg algorithm or the damped Newton algorithm were recommended in previous studies. However, the Wiberg or damped Newton algorithms do not suit for large (especially "long") matrices, because one needs to solve a large linear system in every iteration. In this paper, we revitalize the usage of the Levenberg-Marquardt algorithm for solving the missing data problem, by utilizing the property that low-rank approximation is a minimization problem on subspaces. In two proposed implementations of the Levenberg-Marquardt algorithm, one only needs to solve a much smaller linear system in every iteration, especially for "long" matrices. Simulations and experiments on real data show the superiority of the proposed algorithms. Though the proposed algorithms achieve a high success rate in estimating the optimal solution by random initialization, as illustrated by real examples; it still remains an open issue how to properly do the initialization in a severe situation (that is, a large amount of data is missing and with high-level noise).

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.