Bayesian Approaches to Shrinkage and Sparse Estimation

  • Abstract
  • Literature Map
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

Bayesian Approaches to Shrinkage and Sparse Estimation

Similar Papers
  • Book Chapter
  • Cite Count Icon 1
  • 10.1007/978-3-642-32717-9_32
Sparse Point Estimation for Bayesian Regression via Simulated Annealing
  • Jan 1, 2012
  • Sudhir Raman + 1 more

In the context of variable selection in a regression model, the classical Lasso based optimization approach provides a sparse estimate with respect to regression coefficients but is unable to provide more information regarding the distribution of regression coefficients. Alternatively, using a Bayesian approach is more advantageous since it gives direct access to the distribution which is usually summarized by estimating the expectation (not sparse) and variance. Additionally, to support frequent application requirements, heuristics like thresholding are generally used to produce sparse estimates for variable selection purposes. In this paper, we provide a more principled approach for generating a sparse point estimate in a Bayesian framework. We extend an existing Bayesian framework for sparse regression to generate a MAP estimate by using simulated annealing. We then justify this extension by showing that this MAP estimate is also sparse in the regression coefficients. Experiments on real world applications like the splice site detection and diabetes progression demonstrate the usefulness of the extension.

  • Research Article
  • Cite Count Icon 2
  • 10.2139/ssrn.3971832
Bayesian Approaches to Shrinkage and Sparse Estimation
  • Jan 1, 2021
  • SSRN Electronic Journal
  • Dimitris Korobilis + 1 more

In all areas of human knowledge, datasets are increasing in both size and complexity, creating the need for richer statistical models. This trend is also true for economic data, where high-dimensional and nonlinear/nonparametric inference is the norm in several fields of applied econometric work. The purpose of this paper is to introduce the reader to the world of Bayesian model determination, by surveying modern shrinkage and variable selection algorithms and methodologies. Bayesian inference is a natural probabilistic framework for quantifying uncertainty and learning about model parameters, and this feature is particularly important for inference in modern models of high dimensions and increased complexity. We begin with a linear regression setting in order to introduce various classes of priors that lead to shrinkage/sparse estimators of comparable value to popular penalized likelihood estimators (e.g.\ ridge, lasso). We explore various methods of exact and approximate inference, and discuss their pros and cons. Finally, we explore how priors developed for the simple regression setting can be extended in a straightforward way to various classes of interesting econometric models. In particular, the following case-studies are considered, that demonstrate application of Bayesian shrinkage and variable selection strategies to popular econometric contexts: i) vector autoregressive models; ii) factor models; iii) time-varying parameter regressions; iv) confounder selection in treatment effects models; and v) quantile regression models. A MATLAB package and an accompanying technical manual allow the reader to replicate many of the algorithms described in this review.

  • Research Article
  • Cite Count Icon 12
  • 10.1016/j.spa.2017.10.007
Asymptotics for high-dimensional covariance matrices and quadratic forms with applications to the trace functional and shrinkage
  • Nov 15, 2017
  • Stochastic Processes and their Applications
  • Ansgar Steland + 1 more

Asymptotics for high-dimensional covariance matrices and quadratic forms with applications to the trace functional and shrinkage

  • Research Article
  • Cite Count Icon 72
  • 10.1016/j.frl.2018.02.026
Some improved sparse and stable portfolio optimization problems
  • Feb 26, 2018
  • Finance Research Letters
  • Zhifeng Dai + 1 more

Some improved sparse and stable portfolio optimization problems

  • Research Article
  • Cite Count Icon 2
  • 10.1080/00949655.2013.873042
Double shrinkage estimators for large sparse covariance matrices
  • Feb 5, 2014
  • Journal of Statistical Computation and Simulation
  • S.-M Chang

Covariance matrices play an important role in many multivariate techniques and hence a good covariance estimation is crucial in this kind of analysis. In many applications a sparse covariance matrix is expected due to the nature of the data or for simple interpretation. Hard thresholding, soft thresholding, and generalized thresholding were therefore developed to this end. However, these estimators do not always yield well-conditioned covariance estimates. To have sparse and well-conditioned estimates, we propose doubly shrinkage estimators: shrinking small covariances towards zero and then shrinking covariance matrix towards a diagonal matrix. Additionally, a richness index is defined to evaluate how rich a covariance matrix is. According to our simulations, the richness index serves as a good indicator to choose relevant covariance estimator.

  • Research Article
  • Cite Count Icon 311
  • 10.1093/comjnl/bxm055
Inpainting and Zooming Using Sparse Representations
  • Feb 19, 2008
  • The Computer Journal
  • M.J Fadili + 2 more

Representing the image to be inpainted in an appropriate sparse representation dictionary, and combining elements from Bayesian statistics and modern harmonic analysis, we introduce an expectation maximization (EM) algorithm for image inpainting and interpolation. From a statistical point of view, the inpainting/interpolation can be viewed as an estimation problem with missing data. Toward this goal, we propose the idea of using the EM mechanism in a Bayesian framework, where a sparsity promoting prior penalty is imposed on the reconstructed coefficients. The EM framework gives a principled way to establish formally the idea that missing samples can be recovered/interpolated based on sparse representations. We first introduce an easy and efficient sparse-representation-based iterative algorithm for image inpainting. Additionally, we derive its theoretical convergence properties. Compared to its competitors, this algorithm allows a high degree of flexibility to recover different structural components in the image (piecewise smooth, curvilinear, texture, etc.). We also suggest some guidelines to automatically tune the regularization parameter.

  • Research Article
  • Cite Count Icon 47
  • 10.1016/j.tra.2003.11.002
Economic and structural determinants of the demand for public transport: an analysis on a panel of French urban areas using shrinkage estimators
  • Jan 29, 2004
  • Transportation Research Part A: Policy and Practice
  • Georges Bresson + 3 more

Economic and structural determinants of the demand for public transport: an analysis on a panel of French urban areas using shrinkage estimators

  • Research Article
  • Cite Count Icon 33
  • 10.1080/02664760120108746
Random effects and shrinkage estimation in capture-recapture models
  • Jan 1, 2002
  • Journal of Applied Statistics
  • J Andrew Royle + 1 more

We discuss the analysis of random effects in capture-recapture models, and outline Bayesian and frequentists approaches to their analysis. Under a normal model, random effects estimators derived from Bayesian or frequentist considerations have a common form as shrinkage estimators. We discuss some of the difficulties of analysing random effects using traditional methods, and argue that a Bayesian formulation provides a rigorous framework for dealing with these difficulties. In capture-recapture models, random effects may provide a parsimonious compromise between constant and completely time-dependent models for the parameters (e.g. survival probability). We consider application of random effects to band-recovery models, although the principles apply to more general situations, such as Cormack-Jolly-Seber models. We illustrate these ideas using a commonly analysed band recovery data set.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 24
  • 10.1371/journal.pone.0131414
A Bayesian Shrinkage Approach for AMMI Models.
  • Jul 9, 2015
  • PLOS ONE
  • Carlos Pereira Da Silva + 4 more

Linear-bilinear models, especially the additive main effects and multiplicative interaction (AMMI) model, are widely applicable to genotype-by-environment interaction (GEI) studies in plant breeding programs. These models allow a parsimonious modeling of GE interactions, retaining a small number of principal components in the analysis. However, one aspect of the AMMI model that is still debated is the selection criteria for determining the number of multiplicative terms required to describe the GE interaction pattern. Shrinkage estimators have been proposed as selection criteria for the GE interaction components. In this study, a Bayesian approach was combined with the AMMI model with shrinkage estimators for the principal components. A total of 55 maize genotypes were evaluated in nine different environments using a complete blocks design with three replicates. The results show that the traditional Bayesian AMMI model produces low shrinkage of singular values but avoids the usual pitfalls in determining the credible intervals in the biplot. On the other hand, Bayesian shrinkage AMMI models have difficulty with the credible interval for model parameters, but produce stronger shrinkage of the principal components, converging to GE matrices that have more shrinkage than those obtained using mixed models. This characteristic allowed more parsimonious models to be chosen, and resulted in models being selected that were similar to those obtained by the Cornelius F-test (α = 0.05) in traditional AMMI models and cross validation based on leave-one-out. This characteristic allowed more parsimonious models to be chosen and more GEI pattern retained on the first two components. The resulting model chosen by posterior distribution of singular value was also similar to those produced by the cross-validation approach in traditional AMMI models. Our method enables the estimation of credible interval for AMMI biplot plus the choice of AMMI model based on direct posterior distribution retaining more GEI pattern in the first components and discarding noise without Gaussian assumption as requested in F-based tests or deal with parametric problems as observed in traditional AMMI shrinkage method.

  • Research Article
  • Cite Count Icon 6
  • 10.1007/bf02763003
An application of a minimax Bayes rule and shrinkage estimators to the portofolio selection problem under the Bayesian approach
  • Oct 1, 2005
  • Statistical Papers
  • Hiroyuki Kashima

This paper shows that a minimax Bayes rule and shrinkage estimators can be effectively applied to portfolio selection under the Bayesian approach. Specifically, it is shown that the portfolio selection problem can result in a statistical decision problem in some situations. Following that, we present a method for solving a problem involved in portfolio selection under the Bayesian approach.

  • Research Article
  • Cite Count Icon 47
  • 10.4310/sii.2013.v6.n2.a8
The Bayesian covariance lasso
  • Dec 30, 1899
  • Statistics and Its Interface
  • Haitao Chu + 4 more

Estimation of sparse covariance matrices and their inverse subject to positive definiteness constraints has drawn a lot of attention in recent years. The abundance of high-dimensional data, where the sample size (n) is less than the dimension (d), requires shrinkage estimation methods since the maximum likelihood estimator is not positive definite in this case. Furthermore, when n is larger than d but not sufficiently larger, shrinkage estimation is more stable than maximum likelihood as it reduces the condition number of the precision matrix. Frequentist methods have utilized penalized likelihood methods, whereas Bayesian approaches rely on matrix decompositions or Wishart priors for shrinkage. In this paper we propose a new method, called the Bayesian Covariance Lasso (BCLASSO), for the shrinkage estimation of a precision (covariance) matrix. We consider a class of priors for the precision matrix that leads to the popular frequentist penalties as special cases, develop a Bayes estimator for the precision matrix, and propose an efficient sampling scheme that does not precalculate boundaries for positive definiteness. The proposed method is permutation invariant and performs shrinkage and estimation simultaneously for non-full rank data. Simulations show that the proposed BCLASSO performs similarly as frequentist methods for non-full rank data.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 11
  • 10.1186/1687-6180-2012-13
Parameter estimation for SAR micromotion target based on sparse signal representation
  • Jan 18, 2012
  • EURASIP Journal on Advances in Signal Processing
  • Sha Zhu + 5 more

In this article, we address the parameter estimation of micromotion targets in synthetic aperture radar (SAR), where scattering parameters and micromotion parameters of targets are coupled resulting in a nonlinear parameter estimation problem. The conventional methods address this nonlinear problem by matched filter, which are computationally expensive and of lower resolutions. In contrast, we address this problem by linearizing the forward model as a linear combination of elements of an over-complete dictionary. The essential idea of sparse signal representation models comes from the fact that SAR micromotion targets are sparsely distributed in the observation scene. Accordingly, we propose to jointly estimate the target micromotion and scattering parameters via a Bayesian approach with sparsity-inducing priors. In addition, we present a variational approximation framework for Bayesian computation. Numerical simulations demonstrate the proposed sparsity-inducing reconstruction method achieves higher resolution and better performance with smaller measures compared to conventional methods.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 78
  • 10.3390/rs8100797
A Review of Image Fusion Algorithms Based on the Super-Resolution Paradigm
  • Sep 24, 2016
  • Remote Sensing
  • Andrea Garzelli

A critical analysis of remote sensing image fusion methods based on the super-resolution (SR) paradigm is presented in this paper. Very recent algorithms have been selected among the pioneering studies adopting a new methodology and the most promising solutions. After introducing the concept of super-resolution and modeling the approach as a constrained optimization problem, different SR solutions for spatio-temporal fusion and pan-sharpening are reviewed and critically discussed. Concerning pan-sharpening, the well-known, simple, yet effective, proportional additive wavelet in the luminance component (AWLP) is adopted as a benchmark to assess the performance of the new SR-based pan-sharpening methods. The widespread quality indexes computed at degraded resolution, with the original multispectral image used as the reference, i.e., SAM (Spectral Angle Mapper) and ERGAS (Erreur Relative Globale Adimensionnelle de Synthèse), are finally presented. Considering these results, sparse representation and Bayesian approaches seem far from being mature to be adopted in operational pan-sharpening scenarios.

  • Single Book
  • Cite Count Icon 18
  • 10.1575/1912/2585
Sparse Bayesian information filters for localization and mapping
  • Jan 1, 2008
  • Matthew R Walter

This thesis formulates an estimation framework for Simultaneous Localization and Mapping (SLAM) that addresses the problem of scalability in large environments. We describe an estimation-theoretic algorithm that achieves significant gains in computational efficiency while maintaining consistent estimates for the vehicle pose and the map of the environment. We specifically address the feature-based SLAM problem in which the robot represents the environment as a collection of landmarks. The thesis takes a Bayesian approach whereby we maintain a joint posterior over the vehicle pose and feature states, conditioned upon measurement data. We model the distribution as Gaussian and parametrize the posterior in the canonical form, in terms of the information (inverse covariance) matrix. When sparse, this representation is amenable to computationally efficient Bayesian SLAM filtering. However, while a large majority of the elements within the normalized information matrix are very small in magnitude, it is fully populated nonetheless. Recent feature-based SLAM filters achieve the scalability benefits of a sparse parametrization by explicitly pruning these weak links in an effort to enforce sparsity. We analyze one such algorithm, the Sparse Extended Information Filter (SEIF), which has laid much of the groundwork concerning the computational benefits of the sparse canonical form. The thesis performs a detailed analysis of the process by which the SEIF approximates the sparsity of the information matrix and reveals key insights into the consequences of different sparsification strategies. We demonstrate that the SEIF yields a sparse approximation to the posterior that is inconsistent, suffering from exaggerated confidence estimates. This overconfidence has detrimental effects on important aspects of the SLAM process and affects the higher level goal of producing accurate maps for subsequent localization and path planning. This thesis proposes an alternative scalable filter that maintains sparsity while preserving the consistency of the distribution. We leverage insights into the natural structure of the feature-based canonical parametrization and derive a method that actively maintains an exactly sparse posterior. Our algorithm exploits the structure of the parametrization to achieve gains in efficiency, with a computational cost that scales linearly with the size of the map. Unlike similar techniques that sacrifice consistency for improved scalability, our algorithm performs inference over a posterior that is conservative relative to the nominal Gaussian distribution. Consequently, we preserve the consistency of the pose and map estimates and avoid the effects of an overconfident posterior. We demonstrate our filter alongside the SEIF and the standard EKF both in simulation as well as on two real-world datasets. While we maintain the computational advantages of an exactly sparse representation, the results show convincingly that our method yields conservative estimates for the robot pose and map that are nearly identical to those of the original. Gaussian distribution as produced by the EKF, but at much less computational expense. The thesis concludes with an extension of our SLAM filter to a complex underwater environment. We describe a systems-level framework for localization and mapping relative to a ship hull with an Autonomous Underwater Vehicle (AUV) equipped with a forward-looking sonar. The approach utilizes our filter to fuse measurements of vehicle attitude and motion from onboard sensors with data from sonar images of the hull. We employ the system to perform three-dimensional, 6-DOF SLAM on a ship hull. (Copies available exclusively from MIT Libraries, Rm. 14-0551, Cambridge, MA 02139-4307. Ph. 617-253-5668; Fax 617-253-1690.)

  • Research Article
  • Cite Count Icon 61
  • 10.1016/j.probengmech.2020.103082
Sparse representations and compressive sampling approaches in engineering mechanics: A review of theoretical concepts and diverse applications
  • Jun 25, 2020
  • Probabilistic Engineering Mechanics
  • Ioannis A Kougioumtzoglou + 2 more

Sparse representations and compressive sampling approaches in engineering mechanics: A review of theoretical concepts and diverse applications

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.