Scheduled System Maintenance:
Some services will be unavailable Sunday, March 29th through Monday, March 30th. We apologize for the inconvenience.
By Topic

Zero-Gradient-Sum Algorithms for Distributed Convex Optimization: The Continuous-Time Case

Sign In

Full text access may be available.

To access full text, please use your member or institutional sign in.

The purchase and pricing options are temporarily unavailable. Please try again later.
2 Author(s)
Jie Lu ; Sch. of Electr. & Comput. Eng., Univ. of Oklahoma, Norman, OK, USA ; Choon Yik Tang

This technical note presents a set of continuous-time distributed algorithms that solve unconstrained, separable, convex optimization problems over undirected networks with fixed topologies. The algorithms are developed using a Lyapunov function candidate that exploits convexity, and are called Zero-Gradient-Sum (ZGS) algorithms as they yield nonlinear networked dynamical systems that evolve invariantly on a zero-gradient-sum manifold and converge asymptotically to the unknown optimizer. We also describe a systematic way to construct ZGS algorithms, show that a subset of them actually converge exponentially, and obtain lower and upper bounds on their convergence rates in terms of the network topologies, problem characteristics, and algorithm parameters, including the algebraic connectivity, Laplacian spectral radius, and function curvatures. The findings of this technical note may be regarded as a natural generalization of several well-known algorithms and results for distributed consensus, to distributed convex optimization.

Published in:

Automatic Control, IEEE Transactions on  (Volume:57 ,  Issue: 9 )