Professional Documents
Culture Documents
Constrained Consensus and Optimization Multiagent Networks - ParriloTAC2010
Constrained Consensus and Optimization Multiagent Networks - ParriloTAC2010
Constrained Consensus and Optimization Multiagent Networks - ParriloTAC2010
4, APRIL 2010
Abstract—We present distributed algorithms that can be used agents to reach a common decision or agreement (on a scalar or
by multiple agents to align their estimates with a particular value vector value). Recent work also studied multi-agent optimiza-
over a network with time-varying connectivity. Our framework is tion problems over networks with time-varying connectivity,
general in that this value can represent a consensus value among
multiple agents or an optimal solution of an optimization problem, where the objective function information is distributed across
where the global objective function is a combination of local agent agents (e.g., the global objective function is the sum of local ob-
objective functions. Our main focus is on constrained problems jective functions of agents). Despite much work in this area, the
where the estimates of each agent are restricted to lie in different existing literature does not consider problems where the agent
convex sets. values are constrained to given sets. Such constraints are signif-
To highlight the effects of constraints, we first consider a con-
strained consensus problem and present a distributed “projected icant in a number of applications including motion planning and
consensus algorithm” in which agents combine their local aver- alignment problems, where each agent’s position is limited to a
aging operation with projection on their individual constraint sets. certain region or range, and distributed constrained multi-agent
This algorithm can be viewed as a version of an alternating pro- optimization problems.
jection method with weights that are varying over time and across In this paper, we study cooperative control problems where
agents. We establish convergence and convergence rate results for
the projected consensus algorithm. We next study a constrained op- the values of agents are constrained to lie in closed convex
timization problem for optimizing the sum of local objective func- sets. Our main focus is on developing distributed algorithms
tions of the agents subject to the intersection of their local con- for problems where the constraint information is distributed
straint sets. We present a distributed “projected subgradient al- across agents, i.e., each agent only knows its own constraint
gorithm” which involves each agent performing a local averaging set. To highlight the effects of different local constraints, we
operation, taking a subgradient step to minimize its own objective
function, and projecting on its constraint set. We show that, with first consider a constrained consensus problem and propose
an appropriately selected stepsize rule, the agent estimates gener- a projected consensus algorithm that operates on the basis of
ated by this algorithm converge to the same optimal solution for local information. More specifically, each agent linearly com-
the cases when the weights are constant and equal, and when the bines its value with those values received from the time-varying
weights are time-varying but all agents have the same constraint neighboring agents and projects the combination on its own
set.
constraint set. We show that this update rule can be viewed as
Index Terms—Consensus, constraints, distributed optimization, a novel version of the classical alternating projection method
subgradient algorithms.
where, at each iteration, the values are combined using weights
that are varying in time and across agents, and projected on the
I. INTRODUCTION respective constraint sets.
We provide convergence and convergence rate analysis for the
projected consensus algorithm. Due to the projection operation,
HERE has been much interest in distributed cooperative
T control problems, in which several autonomous agents
collectively try to achieve a global objective. Most focus has
the resulting evolution of agent values has nonlinear dynamics,
which poses challenges for the analysis of the convergence prop-
erties of the algorithm. To deal with the nonlinear dynamics
been on the canonical consensus problem, where the goal is to
in the evolution of the agent estimates, we decompose the dy-
develop distributed algorithms that can be used by a group of
namics into two parts: a linear part involving a time-varying av-
eraging operation and a nonlinear part involving the error due to
Manuscript received March 07, 2008; revised December 13, 2008. First pub- the projection operation. This decomposition allows us to repre-
lished February 02, 2010; current version published April 02, 2010. This work
was supported in part by the National Science Foundation under CAREER sent the evolution of the estimates using linear dynamics and de-
grants CMMI 07-42538 and DMI-0545910, under grant ECCS-0621922, and couples the analysis of the effects of constraints from the conver-
by the DARPA ITMANET program. Recommended by Associate Editor M.
gence analysis of the local agent averaging. The linear dynamics
Egerstedt.
A. Nedić is with the Industrial and Enterprise Systems Engineering De- is analyzed similarly to that of the unconstrained consensus up-
partment, University of Illinois at Urbana-Champaign, Urbana IL 61801 USA date, which relies on convergence of transition matrices defined
(e-mail: angelia@uiuc.edu, angelia@illinois.edu).
A. Ozdaglar and P. A. Parrilo are with the Laboratory for Information and
as the products of the time-varying weight matrices. Using the
Decision Systems, Electrical Engineering and Computer Science Department, properties of projection and agent weights, we prove that the
Massachusetts Institute of Technology, Cambridge MA 02139 USA (e-mail: projection error diminishes to zero. This shows that the non-
asuman@mit.edu, parrilo@mit.edu). linear parts in the dynamics are vanishing with time and, there-
Color versions of one or more of the figures in this paper are available online
at http://ieeexplore.ieee.org. fore, the evolution of agent estimates is “almost linear.” We then
Digital Object Identifier 10.1109/TAC.2010.2041686 show that the agents reach consensus on a “common estimate”
0018-9286/$26.00 © 2010 IEEE
NEDIĆ et al.: CONSTRAINED CONSENSUS AND OPTIMIZATION IN MULTI-AGENT NETWORKS 923
in the limit and that the common estimate lies in the intersection this paper. Our work is also related to incremental subgradient
of the agent individual constraint sets. algorithms implemented over a network, where agents sequen-
We next consider a constrained optimization problem for op- tially update an iterate sequence in a cyclic or a random order
timizing a global objective function which is the sum of local [14]–[17]. In an incremental algorithm, there is a single iterate
agent objective functions, subject to a constraint set given by sequence and only one agent updates the iterate at a given
the intersection of the local agent constraint sets. We focus on time. Thus, while operating on the basis of local information,
distributed algorithms in which agent values are updated based incremental algorithms differ fundamentally from the algorithm
on local information given by the agent’s objective function and studied in this paper (where all agents update simultaneously).
constraint set. In particular, we propose a distributed projected Furthermore, the work in [14]–[17] assumes that the constraint
subgradient algorithm, which for each agent involves a local set is known by all agents in the system, which is in a sharp
averaging operation, a step along the subgradient of the local contrast with the algorithms studied in this paper (our primary
objective function, and a projection on the local constraint set. interest is in the case where the information about the constraint
We study the convergence behavior of this algorithm for two set is distributed across the agents).
cases: when the constraint sets are the same, but the agent con- The paper is organized as follows. In Section II, we intro-
nectivity is time-varying; and when the constraint sets are duce our notation and terminology, and establish some basic
different, but the agents use uniform and constant weights in results related to projection on closed convex sets that will be
each step, i.e., the communication graph is fully connected. We used in the subsequent analysis. In Section III, we present the
show that with an appropriately selected stepsize rule, the agent constrained consensus problem and the projected consensus al-
estimates generated by this algorithm converge to the same op- gorithm. We describe our multi-agent model and provide a basic
timal solution of the constrained optimization problem. Similar result on the convergence behavior of the transition matrices that
to the analysis of the projected consensus algorithm, our con- govern the evolution of agent estimates generated by the algo-
vergence analysis relies on showing that the projection errors rithms. We study the convergence of the agent estimates and
converge to zero, thus effectively reducing the problem into an establish convergence rate results for constant uniform weights.
unconstrained one. However, in this case, establishing the con- Section IV introduces the constrained multi-agent optimization
vergence of the projection error to zero requires understanding problem and presents the projected subgradient algorithm. We
the effects of the subgradient steps, which complicates the anal- provide convergence analysis for the estimates generated by this
ysis. In particular, for the case with different constraint sets but algorithm. Section V contains concluding remarks and some fu-
uniform weights, the analysis uses an error bound which relates ture directions.
the distances of the iterates to individual constraint sets with the
distances of the iterates to the intersection set. NOTATION, TERMINOLOGY, AND BASICS
Related literature on parallel and distributed computation A vector is viewed as a column, unless clearly stated other-
is vast. Most literature builds on the seminal work of Tsit- wise. We denote by or the th component of a vector
siklis [1] and Tsitsiklis et al. [2] (see also [3]), which focused . When for all components of a vector , we write
on distributing the computations involved with optimizing a . We write to denote the transpose of a vector . The
global objective function among different processors (assuming scalar product of two vectors and is denoted by . We use
complete information about the global objective function at to denote the standard Euclidean norm, .
each processor). More recent literature focused on multi-agent A vector is said to be a stochastic vector when its
environments and studied consensus algorithms for achieving components are nonnegative and their sum is equal to 1, i.e.,
cooperative behavior in a distributed manner (see [4]–[10]). . A set of vectors , with
These works assume that the agent values can be processed
for all , is said to be doubly stochastic when each is a sto-
arbitrarily and are unconstrained. Another recent approach
chastic vector and for all . A square
for distributed cooperative control problems involve using
matrix is doubly stochastic when its rows are stochastic vec-
game-theoretic models. In this approach, the agents are en-
tors, and its columns are also stochastic vectors.
dowed with local utility functions that lead to a game form
We write to denote the standard Euclidean dis-
with a Nash equilibrium which is the same as or close to a
tance of a vector from a set , i.e.,
global optimum. Various learning algorithms can then be used
as distributed control schemes that will reach the equilibrium.
In a recent paper, Marden et al. [11] used this approach for
the consensus problem where agents have constraints on their We use to denote the projection of a vector on a closed
values. Our projected consensus algorithm provides an alterna- convex set , i.e.
tive approach for this problem.
Most closely related to our work are the recent papers
[12], [13], which proposed distributed subgradient methods
for solving unconstrained multi-agent optimization problems. In the subsequent development, the properties of the projection
These methods use consensus algorithms as a mechanism for operation on a closed convex set play an important role. In par-
distributing computations among the agents. The presence of ticular, we use the projection inequality, i.e., for any vector
different local constraints significantly changes the operation
and the analysis of the algorithms, which is our main focus in (1)
924 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 55, NO. 4, APRIL 2010
(2)
In addition, we use the properties given in the following lemma. We provide an error bound relation in the following lemma,
Lemma 1: Let be a nonempty closed convex set in . which is illustrated in Fig. 2.
Then, we have for any , Lemma 2: Let , , be nonempty
(a) , for all . closed convex sets that satisfy Assumption 1. Let ,
(b) , for all . , be arbitrary vectors and define their average as
Proof: . Consider the vector defined by
(a) Let be arbitrary. Then, for any , we have
Proof:
(a) We first show that the vector belongs to the intersection
Part (b) of the preceding lemma establishes a relation be- . To see this, let be arbitrary
tween the projection error vector and the feasible directions of and note that we can write as
the convex set at the projection vector, as illustrated in Fig. 1.
We next consider nonempty closed convex sets , for
, and an averaged-vector obtained by taking an
average of vectors , i.e., for some
. We provide an “error bound” that relates the distance By the definition of , it follows that ,
of the averaged-vector from the intersection set implying by the interior point assumption (cf. Assumption
to the distance of from the individual sets . This relation, 1) that the vector belongs to the
which is also of independent interest, will play a key role in set , and therefore to the set . Since the vector is
our analysis of the convergence of projection errors associated the convex combination of two vectors in the set , it
with various distributed algorithms introduced in this paper. We follows by the convexity of that . The preceding
establish the relation under an interior point assumption on the argument is valid for an arbitrary , thus implying that
intersection set stated in the following: .
NEDIĆ et al.: CONSTRAINED CONSENSUS AND OPTIMIZATION IN MULTI-AGENT NETWORKS 925
(b) Using the definition of the vector and the vector , we an unconstrained convex optimization problem of the following
have form:
(4)
Substituting the definition of yields the desired relation.
In view of this optimization problem, the method can be inter-
preted as a distributed gradient algorithm where each agent is
assigned an objective function .
II. CONSTRAINED CONSENSUS At each time , an agent incorporates new information
received from some of the other agents and generates a weighted
In this section, we describe the constrained consensus sum . Then, the agent updates its estimate by
problem. In particular, we introduce our multi-agent model taking a step (with stepsize equal to 1) along the negative gra-
and the projected consensus algorithm that is locally executed dient of its own objective function at
by each agent. We provide some insights about the algorithm . In particular, since the gradient of
and we discuss its connection to the alternating projections is (see Theorem 1.5.5 in Facchinei and
method. We also introduce the assumptions on the multi-agent Pang [18]), the update rule in (3) is equivalent to the following
model and present key elementary results that we use in our gradient descent method for minimizing :
subsequent analysis of the projected consensus algorithm.
In particular, we define the transition matrices governing the
linear dynamics of the agent estimate evolution and give a basic
convergence result for these matrices. The model assumptions
and the transition matrix convergence properties will also
be used for studying the constrained optimization problem
and the projected subgradient algorithm that we introduce in
Section IV.
Motivated by the objective function of problem (4), we use
with as a Lyapunov function
A. Multi-Agent Model and Algorithm measuring the progress of the algorithm (see Section III-F).1
We consider a set of agents denoted by We
assume a slotted-time system, and we denote by the esti-
B. Relation to Alternating Projections Method
mate generated and stored by agent at time slot . The agent
estimate is a vector in that is constrained to lie in a The method of (3) is related to the classical alternating or
nonempty closed convex set known only to agent . cyclic projection method. Given a finite collection of closed
The agents’ objective is to cooperatively reach a consensus on convex sets with a nonempty intersection (i.e.,
a common vector through a sequence of local estimate updates ), the alternating projection method finds a vector
(subject to the local constraint set) and local information ex- in the intersection . In other words, the algorithm solves
changes (with neighboring agents only). the unconstrained problem (4). Alternating projection methods
We study a model where the agents exchange and update their generate a sequence of vectors by projecting iteratively on the
estimates as follows: To generate the estimate at time , sets, either cyclically or with some given order; see Fig. 3(a),
agent forms a convex combination of its estimate with where the alternating projection algorithm generates a se-
the estimates received from other agents at time , and takes the quence by iteratively projecting onto sets and ,
projection of this vector on its constraint set . More specifi- i.e., , . The
cally, agent at time generates its new estimate according convergence behavior of these methods has been established by
to the following relation: Von Neumann [19] and Aronszajn [20] for the case when the
sets are affine; and by Gubin et al. [21] when the sets are
(3) closed and convex. Gubin et al. [21] also have provided conver-
gence rate results for a particular form of alternating projection
method. Similar rate results under different assumptions have
also been provided by Deutsch [22], and Deutsch and Hundal
where is a vector of nonnegative weights.
[23].
The relation in (3) defines the projected consensus algorithm.
We note here an interesting connection between the projected 1We focus throughout the paper on the case when the intersection set \ X
consensus algorithm and a multi-agent algorithm for finding a is nonempty. If the intersection set is empty, it follows from the definition of
the algorithm that the agent estimates will not reach a consensus. In this case,
point in common to the given closed convex sets . x k
the estimate sequences f ( )g may exhibit oscillatory behavior or may all be
The problem of finding a common point can be formulated as unbounded.
926 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 55, NO. 4, APRIL 2010
(a) The entries of the transition matrices converge In the following lemma, we show some relations for the
to as with a geometric rate uniformly with sums and , and
respect to and , i.e., for all , and all and for an arbitrary
and with vector in the intersection of the agent constraint sets. Also, we
prove that the errors converge to zero as for all
. The projection properties given in Lemma 1 and the doubly
stochasticity of the weights play crucial roles in establishing
(b) In the absence of Assumption 3(b) [i.e., the weights these relations. The proof is provided in Appendix.
are stochastic but not doubly stochastic], the Lemma 3: Let the intersection set be nonempty,
columns of the transition matrices converge and let Doubly Stochasticity assumption hold (cf. Assumption
to a stochastic vector as with a geo- 3). Let , , and be defined by (7)–(9). Then, we
metric rate uniformly with respect to and , i.e., for all have the following.
, and all and with (a) For all and all , we have
(i)
for all ;
(ii) ;
(iii) .
Here, is the lower bound of Assumption 2, , (b) For all , the sequences and
is the number of agents, and is the intercommunication are monotonically nonincreasing
interval bound of Assumption 5. with .
(c) For all , the sequences and
D. Convergence
In this section, we study the convergence behavior of the are monotonically nonincreasing
agent estimates generated by the projected consensus with .
algorithm (3) under Assumptions 2–5. We write the update rule (d) The errors converge to zero as , i.e.
in (3) as
(5)
We next consider the evolution of the estimates
generated by method (3) over a period of time. In particular,
where represents the error due to projection given by we relate the estimates to the estimates gen-
erated earlier in time with by exploiting the de-
(6) composition of the estimate evolution in (5)–(6). In this, we use
the transition matrices from time to time (see Sec-
tion III-D). As we will shortly see, the linear part of the dy-
As indicated by the preceding two relations, the evolution dy- namics is given in terms of the transition matrices, while the
namics of the estimates for each agent is decomposed into nonlinear part involves combinations of the transition matrices
a sum of a linear (time-varying) term and a and the error terms from time to time .
nonlinear term . The linear term captures the effects of Recall that the transition matrices are defined as follows:
mixing the agent estimates, while the nonlinear term captures
the nonlinear effects of the projection operation. This decom-
position plays a crucial role in our analysis. As we will shortly
see [cf. Lemma 3 (d)], under the doubly stochasticity assump- for all and with , where for all ,
tion on the weights, the nonlinear terms are diminishing in and each is a matrix whose th column is the vector .
time for each , and therefore, the evolution of agent estimates Using these transition matrices and the decomposition of the
is “almost linear”. Thus, the nonlinear term can be viewed as estimate evolution of (5)–(6), the relation between
a non-persistent disturbance in the linear evolution of the esti- and the estimates at time is given by
mates.
For notational convenience, let denote
(7)
We use this relation to study the “steady-state” behavior of a Taking the limit as in the preceding relation and using
related process. In particular, we define an auxiliary sequence Lemma 4, we conclude
, where is given by
(14)
(11)
For a given , using Lemma 3(c), we have
Proof: Let denote the closed ball centered at a for all . Defining the constant
vector with radius , i.e., . and substituting in the preceding
For each , consider the sets relation, we obtain
We now establish a convergence rate result for constant uni- Combining this relation with (15), we further obtain
form weights. In particular, we show that the projected con-
sensus algorithm converges with a geometric rate under the In-
terior Point assumption.
Proposition 3: Let Interior Point, Weights Rule, Doubly Taking the square of both sides and using the convexity of the
Stochasticity, Connectivity, and Information Exchange As- square function , we have
sumptions hold (cf. Assumptions 1, 2, 3, 4, and 5). Let
the weight vectors in algorithm (3) be given by
(16)
for all and . For all , let
the sequence be generated by the algorithm (3). We
then have for all Since for all and , using Lemma 3(a)
with the substitutions and
for all , we see that for all
where the vector and the scalar are given in Assumption III. CONSTRAINED OPTIMIZATION
1. Since , the sequence is non- We next consider the problem of optimizing the sum of
increasing by Lemma 3(c). Therefore, we have convex objective functions corresponding to agents con-
930 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 55, NO. 4, APRIL 2010
nected over a time-varying topology. The goal of the agents is dient steps and projecting. In particular, we re-write the relations
to cooperatively solve the constrained optimization problem (19)–(20) as follows:
(21)
(17)
(22)
where each is a convex function, representing The evolution of the iterates is complicated due to the non-
the local objective function of agent , and each is a linear effects of the projection operation, and even more compli-
closed convex set, representing the local constraint set of agent cated due to the projections on different sets. In our subsequent
. We assume that the local objective function and the local analysis, we study two special cases: 1) when the constraint sets
constraint set are known to agent only. We denote the op- are the same [i.e., for all ], but the agent connectivity is
timal value of this problem by . time-varying; and 2) when the constraint sets are different,
To keep our discussion general, we do not assume differentia- but the agent communication graph is fully connected. In the
bility of any of the functions . Since each is convex over the analysis of both cases, we use a basic relation for the iterates
entire , the function is differentiable almost everywhere (see generated by the method in (22). This relation stems from
[28] or [29]). At the points where the function fails to be dif- the properties of subgradients and the projection error and is es-
ferentiable, a subgradient exists and can be used in the role of a tablished in the following lemma.
gradient. In particular, for a given convex function Lemma 6: Let Assumptions Weights Rule and Doubly
and a point , a subgradient of the function at is a vector Stochasticity hold (cf. Assumptions 2 and 3). Let be
such that the iterates generated by the algorithm (19)–(20). We have for
any and all
(18)
(19)
(23)
By summing the preceding relation over and
using the doubly stochasticity of the weights, i.e.
Under Assumption 6, the assumptions on the agent weights and
connectivity stated in Section III-C, and some conditions on the
stepsize , the next lemma studies the convergence properties
of the sequences for all (see Appendix for
the proof).
Lemma 8: Let Weights Rule, Doubly Stochasticity, Connec-
tivity, Information Exchange, and Same Constraint Set Assump-
tions hold (cf. Assumptions 2, 3, 4, 5, and 6). Let be the
iterates generated by the algorithm (19)–(20) and consider the
we obtain the desired relation. auxiliary sequence defined in (23).
1) Convergence When for all : We first study the (a) If the stepsize satisfies , then
case when all constraint sets are the same, i.e., for all
. The next assumption formally states the conditions we adopt
in the convergence analysis.
Assumption 6: (Same Constraint Set) (b) If the stepsize satisfies , then
(a) The constraint sets are the same, i.e, for a
closed convex set .
(b) The subgradient sets of each are bounded over the set
, i.e., there is a scalar such that for all
The next proposition presents our main convergence result
for the same constraint set case. In particular, we show that the
iterates of the projected subgradient algorithm converge
The subgradient boundedness assumption in part (b) holds to an optimal solution when we use a stepsize converging to zero
for example when the set is compact (see [28]). fast enough. The proof uses Lemmas 6 and 8.
In proving our convergence results, we use a property of the Proposition 4: Let Weights Rule, Doubly Stochasticity, Con-
infinite sum of products of the components of two sequences. In nectivity, Information Exchange, and Same Constraint Set As-
particular, for a scalar and a scalar sequence , we sumptions hold (cf. Assumptions 2, 3, 4, 5, and 6). Let
consider the “convolution” sequence be the iterates generated by the algorithm (19)–(20) with the
We have the following result, stepsize satisfying and In addi-
presented here without proof for space reasons (see [27]). tion, assume that the optimal solution set is nonempty. Then,
Lemma 7: Let and let be a positive scalar there exists an optimal point such that
sequence. Assume that . Then
(26)
Since and ,
it follows that the sequence is bounded for each and
are uniformly bounded, i.e., there exists a scalar such Using the subgradient definition and the subgradient bounded-
that ness assumption, we further have for all and
(27)
The next proposition presents our convergence result for the Combining these relations with the preceding and using the no-
projected subgradient algorithm in the uniform weight case. The tation , we obtain
proof uses the error bound relation established in Lemma 2
under the interior point assumption on the intersection set
(cf. Assumption 1) together with the basic subgradient
relation of Lemma 6.
Proposition 5: Let Interior Point and Compactness Assump-
tions hold (cf. Assumptions 1 and 7). Let be the iterates
generated by the algorithm (19)–(20) with the weight vectors
for all and , and the stepsize sat-
isfying and . Then, the sequences
, converge to the same optimal point, i.e. (29)
Since the weights are all equal, from relation (19) we have
for all and . Using Lemma 2(b) with the
Proof: By Assumption 7, each set is compact, which substitution and , we
implies that the intersection set is compact. Since obtain for all and
each function is continuous (due to being convex over ),
it follows from Weierstrass’ Theorem that problem (17) has an
optimal solution, denoted by . By using Lemma 6 with
, we have for all and
Since , we have
for all and , Furthermore, since for all ,
using Assumption 7, we obtain . Therefore,
(28) for all and
(30)
where is the scalar given in Assumption 1, Moreover, we have for all and , implying
, and (cf. Lemma
2). By using the subgradient boundedness [see (27)] and adding
and subtracting the term in (28), we obtain
In view of the definition of the error term in (22) and the
subgradient boundedness, it follows:
(31)
We now substitute the estimate (31) in (29) and obtain for all
934 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 55, NO. 4, APRIL 2010
(34)
Note that for each , we can write
Since for all and [in view of
], from (31) it follows that:
which when substituted in (32) yields We now show that the sequences , con-
verge to the same limit point, which lies in the optimal solution
set . By taking limsup as in relation (33) and then
liminf as , (while dropping the nonnegative terms on
the right hand side there), since , we obtain for any
where . By re-ar-
ranging the terms and summing the preceding relations over
implying that the scalar sequence is con-
for for some arbitrary and with ,
vergent for every . Since for all
we obtain
, it follows that the scalar sequence is also con-
vergent for every . In view of
[cf. (34)], it follows that one of the limit points of must
belong to ; denote this limit by . Since is
convergent for , it follows that .
This and for all imply that each of the
sequences converges to a vector , with .
(33)
IV. CONCLUSION
We studied constrained consensus and optimization problems
By setting and letting , in view of ,
where agent ’s estimate is constrained to lie in a closed convex
we see that
set . For the constrained consensus problem, we presented
a distributed projected consensus algorithm and studied its
convergence properties. Under some assumptions on the agent
weights and the connectivity of the network, we proved that
Since by Lemma 2(a) we have , the relation each of the estimates converge to the same limit, which belongs
holds for all , thus implying to the intersection of the constraint sets . We also showed
that that the convergence rate is geometric under an interior point
assumption for the case when agent weights are time-invariant
and uniform. For the constrained optimization problem, we
presented a distributed projected subgradient algorithm. We
showed that with a stepsize converging to zero fast enough, the
estimates generated by the subgradient algorithm converges to
an optimal solution for the case when all agent constraint sets
NEDIĆ et al.: CONSTRAINED CONSENSUS AND OPTIMIZATION IN MULTI-AGENT NETWORKS 935
are the same and when agent weights are time-invariant and a convex function. By summing the preceding relations
uniform. over , we obtain
The framework and algorithms studied in this paper motivate
a number of interesting research directions. One interesting fu-
ture direction is to extend the constrained optimization problem
to include both local and global constraints, i.e., constraints
known by all the agents. While global constraints can also be ad-
dressed using the “primal projection” algorithms of this paper,
an interesting alternative would be to use “primal-dual” subgra- Using the doubly stochasticity of the weight vectors
dient algorithms, in which dual variables (or prices) are used to , i.e., for all and [cf. Assump-
ensure feasibility of agent estimates with respect to global con- tion 3(b)], we obtain the relation in part (a)(ii), i.e., for all
straints. Such algorithms have been studied in recent work [30] and
for general convex constrained optimization problems (without
a multi-agent network structure).
Moreover, we presented convergence results for the dis-
tributed subgradient algorithm for two cases: agents have
time-varying weights but the same constraint set; and agents Similarly, from relation (35) and the doubly stochasticity
have time-invariant uniform weights and different constraint of the weights, we obtain for all and all
sets. When agents have different constraint sets, the conver-
gence analysis relies on an error bound that relates the distances
of the iterates (generated with constant uniform weights) to
each with the distance of the iterates to the intersection set
under an interior point condition (cf. Lemma 2). This error
bound is also used in establishing the geometric convergence
rate of the projected consensus algorithm with constant uniform thus showing the relation in part (a)(iii).
weights. These results can be extended using a similar analysis (b) For any , the nonincreasing properties
once an error bound is established for the general case with of the sequences and
time-varying weights. We leave this for future work. follow by combining the relations
in parts (a)(i)–(ii).
APPENDIX
(c) Since for all and , using
In this Appendix, we provide the missing proofs for some of the nonexpansiveness property of the projection operation
the lemmas presented in the text. [cf. (2)], we have for all and
Proof of Lemma 3:
(a) For any and , we consider the term
. Since for all , it follows that for all
. Since we also have , we have Summing the preceding relations over all
from Lemma 1(b) that for all and all yields for all
(36)
which yields the relation in part (a)(i) in view of relation
(9). The nonincreasing property of the sequences
By the definition of in (7) and the stochasticity of and follows
the weight vector [cf. Assumption 3(a)], we have from the preceding relation and the relation in part
for every agent and any (a)(iii).
(d) By summing the relations in part (a)(i) over ,
(35) we obtain for any and all
with and
Hence, using this relation
By letting , we obtain
and the subgradient boundedness, we obtain for all and
(37)
In view of the preceding two relations, it follows that [4] T. Vicsek, A. Czirók, E. Ben-Jacob, I. Cohen, and O. Shochet, “Novel
for all . type of phase transitions in a system of self-driven particles,” Phys. Rev.
Lett., vol. 75, no. 6, pp. 1226–1229, 1995.
(b) By multiplying the relation in (38) with , we obtain [5] A. Jadbabaie, J. Lin, and S. Morse, “Coordination of groups of mobile
autonomous agents using nearest neighbor rules,” IEEE Trans. Autom.
Control, vol. 48, no. 6, pp. 988–1001, Jun. 2003.
[6] S. Boyd, A. Ghosh, B. Prabhakar, and D. Shah, “Gossip algorithms:
Design, analysis, and applications,” in Proc. IEEE INFOCOM, 2005,
pp. 1653–1664.
[7] R. Olfati-Saber and R. M. Murray, “Consensus problems in networks of
agents with switching topology and time-delays,” IEEE Trans. Autom.
Control, vol. 49, no. 9, pp. 1520–1533, Sep. 2004.
[8] M. Cao, D. A. Spielman, and A. S. Morse, “A lower bound on con-
vergence of a distributed network consensus algorithm,” in Proc. IEEE
CDC, 2005, pp. 2356–2361.
By using and [9] A. Olshevsky and J. N. Tsitsiklis, “Convergence rates in distributed
for any and , we have consensus and averaging,” in Proc. IEEE CDC, 2006, pp. 3387–3392.
[10] A. Olshevsky and J. N. Tsitsiklis, “Convergence speed in distributed
consensus and averaging,” SIAM J. Control Optim., vol. 48, no. 1, pp.
33–56, 2009.
[11] J. R. Marden, G. Arslan, and J. S. Shamma, “Connections between
cooperative control and potential games illustrated on the consensus
problem,” in Proc. Eur. Control Conf., 2007, pp. 4604–4611.
[12] A. Nedić and A. Ozdaglar, “Distributed subgradient methods for multi-
agent optimization,” IEEE Trans. Autom. Control, vol. 54, no. 1, pp.
48–61, Jan. 2009.
[13] A. Nedić and A. Ozdaglar, “On the rate of convergence of distributed
subgradient methods for multi-agent optimization,” in Proc. IEEE
CDC, 2007, pp. 4711–4716.
[14] D. Blatt, A. O. Hero, and H. Gauchman, “A convergent incremental
where . Therefore, by gradient method with constant stepsize,” SIAM J. Optim., vol. 18, no.
1, pp. 29–51, 2007.
summing and grouping some of the terms, we obtain [15] A. Nedić and D. P. Bertsekas, “Incremental subgradient method for
nondifferentiable optimization,” SIAM J. Optim., vol. 12, pp. 109–138,
2001.
[16] S. S. Ram, A. Nedić, and V. V. Veeravalli, Incremental Stochastic Sub-
Gradient Algorithms for Convex Optimization 2008 [Online]. Avail-
able: http://arxiv.org/abs/0806.1092
[17] B. Johansson, M. Rabi, and M. Johansson, “A simple peer-to-peer al-
gorithm for distributed optimization in sensor networks,” in Proc. 46th
IEEE Conf. Decision Control, 2007, pp. 4705–4710.
[18] F. Facchinei and J.-S. Pang, Finite-Dimensional Variational Inequal-
ities and Complementarity Problems. New York: Springer-Verlag,
2003.
[19] J. von Neumann, Functional Operators. Princeton, NJ: Princeton
Univ. Press, 1950.
[20] N. Aronszajn, “Theory of reproducing kernels,” Trans. Amer. Math.
Soc., vol. 68, no. 3, pp. 337–404, 1950.
[21] L. G. Gubin, B. T. Polyak, and E. V. Raik, “The method of projections
for finding the common point of convex sets,” U.S.S.R Comput. Math.
In the preceding relation, the first term is summable Math. Phys., vol. 7, no. 6, pp. 1211–1228, 1967.
since . The second term is summable since [22] F. Deutsch, “Rate of convergence of the method of alternating projec-
. The third term is also summable by tions,” in Parametric Optimization and Approximation, B. Brosowski
and F. Deutsch, Eds. Basel, Switzerland: Birkhäuser, 1983, vol. 76,
Lemma 7. Hence, pp. 96–107.
[23] F. Deutsch and H. Hundal, “The rate of convergence for the cyclic pro-
jections algorithm I: Angles between convex sets,” J. Approx. Theory,
ACKNOWLEDGMENT vol. 142, pp. 36–55, 2006.
[24] V. D. Blondel, J. M. Hendrickx, A. Olshevsky, and J. N. Tsitsiklis,
The authors would like to thank the associate editor, three “Convergence in multiagent coordination, consensus, and flocking,” in
Proc. IEEE CDC, 2005, pp. 2996–3000.
anonymous referees, and various seminar participants for useful [25] A. Nedić, A. Olshevsky, A. Ozdaglar, and J. N. Tsitsiklis, “On dis-
comments and discussions. tributed averaging algorithms and quantization effects,” IEEE Transa.
Autom. Control, vol. 54, no. 11, pp. 2506–2517, Nov. 2009.
[26] J. Wolfowitz, “Products of indecomposable, aperiodic, stochastic ma-
REFERENCES trices,” Proc. Amer. Math. Soc., vol. 14, no. 4, pp. 733–737, 1963.
[27] A. Nedić, A. Ozdaglar, and P. A. Parrilo, Constrained consensus and
[1] J. N. Tsitsiklis, “Problems in Decentralized Decision Making and Com- optimization in multi-agent networks Tech. Rep., 2008 [Online]. Avail-
putation,” Ph.D. dissertation, MIT, Cambridge, MA, 1984. able: http://arxiv.org/abs/0802.3922v2
[2] J. N. Tsitsiklis, D. P. Bertsekas, and M. Athans, “Distributed [28] D. P. Bertsekas, A. Nedić, and A. E. Ozdaglar, Convex Analysis and
asynchronous deterministic and stochastic gradient optimization algo- Optimization. Cambridge, MA: Athena Scientific, 2003.
rithms,” IEEE Trans. Autom. Control, vol. AC-31, no. 9, pp. 803–812, [29] R. T. Rockafellar, Convex Analysis. Princeton, NJ: Princeton Univ.
Sep. 1986. Press, 1970.
[3] D. P. Bertsekas and J. N. Tsitsiklis, Parallel and Distributed Compu- [30] A. Nedić and A. Ozdaglar, “Subgradient methods for saddle-point
tation: Numerical Methods. Belmont, MA: Athena Scientific, 1989. problems,” J. Optim. Theory Appl., 2008.
938 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 55, NO. 4, APRIL 2010
Angelia Nedić (M’06) received the B.S. degree in Pablo A. Parrilo (S’92–A’95–SM’07) received
mathematics from the University of Montenegro, the electronics engineering undergraduate degree
Podgorica, in 1987, the M.S. degree in mathematics from the University of Buenos Aires, Buenos Aires,
from the University of Belgrade, Belgrade, Serbia, Argentina, in 1995 and the Ph.D. degree in control
in 1990, the Ph.D. degree in mathematics and math- and dynamical systems from the California Institute
ematical physics from Moscow State University, of Technology, Pasadena, in 2000.
Moscow, Russia, in 1994, and the Ph.D. degree in He has held short-term visiting appointments at
electrical engineering and computer science from the UC Santa Barbara, Lund Institute of Technology, and
Massachusetts Institute of Technology, Cambridge, UC Berkeley. From 2001 to 2004, he was Assistant
in 2002. Professor of Analysis and Control Systems at the
She was with the BAE Systems Advanced Infor- Swiss Federal Institute of Technology (ETH Zurich).
mation Technology from 2002 to 2006. Since 2006, she has been a member of He is currently a Professor at the Department of Electrical Engineering and
the faculty of the Department of Industrial and Enterprise Systems Engineering Computer Science, Massachusetts Institute of Technology, Cambridge, where
(IESE), University of Illinois at Urbana-Champaign. Her general interest is in he holds the Finmeccanica Career Development chair. He is affiliated with the
optimization including fundamental theory, models, algorithms, and applica- Laboratory for Information and Decision Systems (LIDS) and the Operations
tions. Her current research interest is focused on large-scale convex optimiza- Research Center (ORC). His research interests include optimization methods
tion, distributed multi-agent optimization, and duality theory with applications for engineering applications, control and identification of uncertain complex
in decentralized optimization. systems, robustness analysis and synthesis, and the development and appli-
Dr. Nedić received the NSF Faculty Early Career Development (CAREER) cation of computational tools based on convex optimization and algorithmic
Award in 2008 in Operations Research. algebra to practically relevant engineering problems.
Dr. Parrilo received the 2005 Donald P. Eckman Award of the American Au-
tomatic Control Council, and the SIAM Activity Group on Control and Sys-
tems Theory (SIAG/CST) Prize. He is currently on the Board of Directors of
Asuman Ozdaglar (S’96–M’02) received the B.S. the Foundations of Computational Mathematics (FoCM) society and on the Ed-
degree in electrical engineering from the Middle itorial Board of the MPS/SIAM Book Series on Optimization.
East Technical University, Ankara, Turkey, in 1996,
and the S.M. and the Ph.D. degrees in electrical
engineering and computer science from the Massa-
chusetts Institute of Technology (MIT), Cambridge,
in 1998 and 2003, respectively.
Since 2003, she has been a member of the faculty
of the Electrical Engineering and Computer Science
Department, MIT, where she is currently the Class of
1943 Associate Professor. She is also a member of the
Laboratory for Information and Decision Systems and the Operations Research
Center. Her research interests include optimization theory, with emphasis on
nonlinear programming and convex analysis, game theory, with applications in
communication, social, and economic networks, and distributed optimization
methods. She is the co-author of the book Convex Analysis and Optimization
(Boston, MA: Athena Scientific, 2003).
Dr. Ozdaglar received the Microsoft Fellowship, the MIT Graduate Student
Council Teaching Award, the NSF CAREER Award, and the 2008 Donald P.
Eckman Award of the American Automatic Control Council.