Abstract

In this paper, we study stochastic optimization of two-level composition of functions without Lipschitz continuous gradient. The smoothness property is generalized by the notion of relative smoothness which provokes the Bregman gradient method. We propose three stochastic composition Bregman gradient algorithms for the three possible relatively smooth compositional scenarios and provide their sample complexities to achieve an $$\epsilon $$ -approximate stationary point. For the smooth of relatively smooth composition, the first algorithm requires $$\mathcal {O}(\epsilon ^{-2})$$ calls to the stochastic oracles of the inner function value and gradient as well as the outer function gradient. When both functions are relatively smooth, the second algorithm requires $$\mathcal {O}(\epsilon ^{-3})$$ calls to the inner function value stochastic oracle and $$\mathcal {O}(\epsilon ^{-2})$$ calls to the inner and outer functions gradients stochastic oracles. We further improve the second algorithm by variance reduction for the setting where just the inner function is smooth. The resulting algorithm requires $$\mathcal {O}(\epsilon ^{-5/2})$$ calls to the inner function value stochastic oracle, $$\mathcal {O}(\epsilon ^{-3/2})$$ calls to the inner function gradient, and $$\mathcal {O}(\epsilon ^{-2})$$ calls to the outer function gradient stochastic oracles. Finally, we numerically evaluate the performance of these three algorithms over two different examples.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.