Abstract
The classical augmented Lagrangian method (ALM) is an efficient method for solving convex optimization with linear constraints.However, the efficiency of ALM, to some extent, is hinged by the computational efforts on solving the resulting subproblems.For the convex optimization with some favorable structures, e.g., either the objective function is separable or the matrices in linear constraints are well-posed, a relaxation to the subproblems of ALM can substantially result in solutions with closed-form. Unfortunately, the relaxation skill can not be extended directly to the generic convex optimization without special structures,particularly for the case of objective function with coupled variables. In this paper, by further relaxing the resulting subproblems of ALM, we propose several novel augmented Lagrangian-based proximal point algorithms. Algorithmically, the next iterate is produced by integrating the predictor, which is obtained in either primal-dual or dual-primal order, with the current iterate. Numerical results demonstrate the promising performances of the proposed algorithms.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: Journal of Industrial & Management Optimization
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.