Abstract

This study aims to merge the well-established ideas of bundle and Gradient Sampling (GS) methods to develop an algorithm for locating a minimizer of a nonsmooth convex function. In the proposed method, with the help of the GS technique, we sample a number of differentiable auxiliary points around the current iterate. Then, by applying the standard techniques used in bundle methods, we construct a polyhedral (piecewise linear) model of the objective function. Moreover, by performing quasi-Newton updates on the set of auxiliary points, this polyhedral model is augmented with a regularization term that enjoys second-order information. If required, this initial model is improved by the techniques frequently used in GS and bundle methods. We analyse the global convergence of the proposed method. As opposed to the original GS method and some of its variants, our convergence analysis is independent of the size of the sample. In our numerical experiments, various aspects of the proposed method are examined using a variety of test problems. In particular, in contrast with many variants of bundle methods, we will see that the user can supply gradients approximately. Moreover, we compare the proposed method with some efficient variants of GS and bundle methods.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call