Abstract

We study the problem of minimizing a sum of local objective convex functions over a network of processors/agents. This problem naturally calls for distributed optimization algorithms, in which the agents cooperatively solve the problem through local computations and communications with neighbors. While many of the existing distributed algorithms with constant stepsize can only converge to a neighborhood of optimal solution, some recent methods based on augmented Lagrangian and method of multipliers can achieve exact convergence with a fixed stepsize. However, these methods either suffer from slow convergence speed or require minimization at each iteration. In this work, we develop a class of distributed first order primal-dual methods, which allows for multiple primal steps per iteration. This general framework makes it possible to control the trade off between the performance and the execution complexity in primal-dual algorithms. We show that for strongly convex and Lipschitz gradient objective functions, this class of algorithms converges linearly to the optimal solution under appropriate constant stepsize choices. Simulation results confirm the superior performance of our algorithm compared to existing methods.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.