Abstract
We consider distributed optimization in random networks where N nodes cooperatively minimize the sum Σi=1N fi(x) of their individual convex costs. Existing literature proposes distributed gradient-like methods that are computationally cheap and resilient to link failures, but have slow convergence rates. In this paper, we propose accelerated distributed gradient methods that 1) are resilient to link failures; 2) computationally cheap; and 3) improve convergence rates over other gradient methods. We model the network by a sequence of independent, identically distributed random matrices {W(k)} drawn from the set of symmetric, stochastic matrices with positive diagonals. The network is connected on average and the cost functions are convex, differentiable, with Lipschitz continuous and bounded gradients. We design two distributed Nesterov-like gradient methods that modify the D-NG and D-NC methods that we proposed for static networks. We prove their convergence rates in terms of the expected optimality gap at the cost function. Let k and K be the number of per-node gradient evaluations and per-node communications, respectively. Then the modified D-NG achieves rates O(logk/k) and O(logK/ K), and the modified D-NC rates O(1/k2) and O(1/ K2-ξ), where ξ > 0 is arbitrarily small. For comparison, the standard distributed gradient method cannot do better than Ω(1/k2/3) and Ω(1/ K2/3), on the same class of cost functions (even for static networks). Simulation examples illustrate our analytical findings.
Highlights
W ITH many distributed signal processing applications, a common research challenge is to develop distributed algorithms whereby all nodes in a generic, connected network re-Manuscript received May 11, 2013; revised August 14, 2013; accepted October 03, 2013
We design two distributed Nesterov-like gradient methods that modify the D–NG and D–NC methods that we proposed for static networks
We model the network by a sequence of random independent, identically distributed (i.i.d.) weight matrices drawn from a set of symmetric, stochastic matrices with positive diagonals, and we assume that the network is connected on average
Summary
W ITH many distributed signal processing applications, a common research challenge is to develop distributed algorithms whereby all nodes in a generic, connected network re-. Our goals are: 1) to develop distributed, iterative, gradient-based methods that solve (1), whereby nodes over iterations exchange messages only with their immediate neighbors; and 2) to provide convergence rate guarantees of the methods (on the assumed functions class) in the presence of random communication failures. We propose the mD–NG and mD–NC algorithms, which modify the D–NG and D–NC algorithms, and, beyond proving their convergence, we solve the much harder problem of establishing their convergence rate guarantees on random networks. We mean how different the solution estimates of distinct nodes are, say and for nodes and
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.