Abstract

We propose a novel stochastic gradient method—semi-stochastic coordinate descent—for the problem of minimizing a strongly convex function represented as the average of a large number of smooth convex functions: . Our method first performs a deterministic step (computation of the gradient of f at the starting point), followed by a large number of stochastic steps. The process is repeated a few times, with the last stochastic iterate becoming the new starting point where the deterministic step is taken. The novelty of our method is in how the stochastic steps are performed. In each such step, we pick a random function and a random coordinate j—both using non-uniform distributions—and update a single coordinate of the decision vector only, based on the computation of the jth partial derivative of at two different points. Each random step of the method constitutes an unbiased estimate of the gradient of f and moreover, the squared norm of the steps goes to zero in expectation, meaning that the stochastic estimate of the gradient progressively improves. The computational complexity of the method is the sum of two terms: evaluations of gradients and evaluations of partial derivatives , where is a novel condition number.

Highlights

  • In this paper we study the problem of unconstrained minimization of a strongly convex function represented as the average of a large number of smooth convex functions: min f (x)

  • We assume that the functions fi : Rd → R are differentiable and convex function, with Lipschitz continuous partial derivatives

  • This assumption was recently used in the analysis of the accelerated coordinate descent method APPROX [2]

Read more

Summary

Optimization Methods and Software

ISSN: 1055-6788 (Print) 1029-4937 (Online) Journal homepage: https://www.tandfonline.com/loi/goms. Our method first performs a deterministic step (computation of the gradient of f at the starting point), followed by a large number of stochastic steps. The novelty of our method is in how the stochastic steps are performed In each such step, we pick a random function fi and a random coordinate j—both using non-uniform distributions—and update a single coordinate of the decision vector only, based on the computation of the jth partial derivative of fi at two different points. The computational complexity of the method is the sum of two terms: O(n log(1/ )) evaluations of gradients ∇fi and O(κlog(1/ )) evaluations of partial derivatives ∇jfi, where κis a novel condition number

Introduction
Method
S2CD algorithm
Complexity result
Coordinate co-coercivity
Recursion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.