Abstract

We propose a distributed optimization algorithm for mixed L1/L2-norm optimization based on accelerated gradient methods using dual decomposition. The algorithm achieves convergence rate O(1k2), where k is the iteration number, which significantly improves the convergence rates of existing duality-based distributed optimization algorithms that achieve O(1k). The performance of the developed algorithm is evaluated on randomly generated optimization problems arising in distributed model predictive control (DMPC). The evaluation shows that, when the problem data is sparse and large-scale, our algorithm can outperform current state-of-the-art optimization software CPLEX and MOSEK.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call