Abstract

Motivated by some applications in signal processing and machine learning, we consider two convex optimization problems where, given a cone $$K$$ , a norm $$\Vert \cdot \Vert $$ and a smooth convex function $$f$$ , we want either (1) to minimize the norm over the intersection of the cone and a level set of $$f$$ , or (2) to minimize over the cone the sum of $$f$$ and a multiple of the norm. We focus on the case where (a) the dimension of the problem is too large to allow for interior point algorithms, (b) $$\Vert \cdot \Vert $$ is “too complicated” to allow for computationally cheap Bregman projections required in the first-order proximal gradient algorithms. On the other hand, we assume that it is relatively easy to minimize linear forms over the intersection of $$K$$ and the unit $$\Vert \cdot \Vert $$ -ball. Motivating examples are given by the nuclear norm with $$K$$ being the entire space of matrices, or the positive semidefinite cone in the space of symmetric matrices, and the Total Variation norm on the space of 2D images. We discuss versions of the Conditional Gradient algorithm capable to handle our problems of interest, provide the related theoretical efficiency estimates and outline some applications.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.