Abstract

An usual problem in statistics consists in estimating the minimizer of a convex function. When we have to deal with large samples taking values in high dimensional spaces, stochastic gradient algorithms and their averaged versions are efficient candidates. Indeed, (1) they do not need too much computational efforts, (2) they do not need to store all the data, which is crucial when we deal with big data, (3) they allow to simply update the estimates, which is important when data arrive sequentially. The aim of this work is to give asymptotic and non asymptotic rates of convergence of stochastic gradient estimates as well as of their averaged versions when the function we would like to minimize is only locally strongly convex.

Highlights

  • With the development of automatic sensors, it is more and more important to think about methods able to deal with large samples of observations taking values in high dimensional spaces such as functional spaces

  • We focus here on an usual stochastic optimization problem which consists in estimating m := arg min E [g(X, h)], h∈H

  • [22] and [1] give some general conditions to get the rate of convergence in quadratic mean of averaged stochastic gradient algorithms, while [13], for instance, focus on non asymptotic rates for strongly convex stochastic composite optimization

Read more

Summary

Introduction

With the development of automatic sensors, it is more and more important to think about methods able to deal with large samples of observations taking values in high dimensional spaces such as functional spaces. [22] and [1] give some general conditions to get the rate of convergence in quadratic mean of averaged stochastic gradient algorithms, while [13], for instance, focus on non asymptotic rates for strongly convex stochastic composite optimization. Two examples of application are given in Section 3: we first focus on the estimation of geometric quantiles, which are a generalization of the real quantiles introduced by [8]. They are robust indicators which can be useful in statistical depth and outliers detection (see [30], [9] or [17]).

The algorithms and assumptions
Applications
An application in a finite dimensional space: a robust logistic regression
Rates of convergence
Simulation study
Proofs
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.