Professional Documents
Culture Documents
A Deterministic Annealing Neural Network For Convex Programming
A Deterministic Annealing Neural Network For Convex Programming
629-641, 1994
Copyright (c; 1994 Elsevier Science Ltd
Pergamon Printed in the USA. All rights reserved
0893-6080/94 $6.00 + .00
0893-6080(93)E0021-X
CONTRIBUTED ARTICLE
JUN WANG
University of North Dakota
Abstract--A recurrent neural network, called a deterministic' annealing neural network, is proposed jot solving
convex programming problems. The proposed deterministic annealing neural network is shown to be capable of
generating optima/solutions to convex programming problems. The conditions jot asymptotic stabilit); solution
feasibilit3; and solution optimality are derived. The design methodologyfi)r determining design parameters is discussed
Three detailed illustrative examples are also presented to demonstrate the Jimctional and operational characteristics
o[the deterministic annealing neural network in solving linear and quadratic programs.
described. Then its convergence properties and the de- For a constrained optimization problem to be solved
sign principles are discussed. Finally, simulation results through neural computing, it is usually necessary to
are given. characterize the t~asible set V using a penalty function
This paper is organized in eight sections. In Section p ( v ) to penalize a constraint violation. A penalty func-
2, the reformulation of the convex program under study tion characterizes the feasible region of an optimization
is delineated. In Section 3, the dynamics of the proposed problem with a thnctional equality. The basic require-
deterministic annealing neural network are discussed. ment tbr the penalty function is nonnegativity; precisely
In Section 4, the architectures of the deterministic an-
t 0. if r:- '~
nealing neural network are described. In Section 5, the
l)(t')[> O. ~)ther~s~
analytical results on the convergence properties of the
proposed deterministic annealing neural network for Additional requirements tbr the penalty function p(t) )
convex programming are presented. In Section 6, design are convexity and differentiability, l f p ( v ) is convex and
principles for selection of parameters in the determin- differentiable, then min~,p(v) := 0 and Vv E V, Vp(~ }
istic annealing neural network are addressed. In Section =0.
7, the performance of the proposed deterministic an- There are basically two ways to construct the penalty
nealing neural network is demonstrated via illustrative function based on the functional constraints: functional
examples. Finally, the paper is concluded in Section 8. transformanon and state augmentation.
(i) F u n c t i o n a l transformation: Functional translor-
2. PROBLEM REFORMULATION mation involves a mapping of an inequality constraint
or an equality constraint to an equality. Specifically,
In this study, we consider a general constrained convex
programming problem described as follows: 1~(, /r,[',- ~!: 5'
l)e(12) 1.1. e(r) - '
minimize I ( v ). ~I ) = O. h(c, (~
pMt;) (("~
s u b j e c t to e( r ) -c- O. ,2, ;> 0. t l ( !: ~ -" t;
h(v) - O. ~3,
There are several logical choices olp.e(v ) and Ph( V}.
For example, for g ( v ) _< 0. the integral of the Heaviside
where v E R N is a column vector of decision variables,
function can be used: tbr h ( v ) = O. ]th(v)[l~ can be
f:R u ~ R is a convex objective function to be mini-
used where II • 112 is the Euclidean norm. Namely,
mized, g:R N --~ R m is an m-tuple convex function used
to define m inequality constraints, and h : R u ~ R ~ is (7)
p~tv)- x£ I H~v~dx.
an n-tuple linear function used to define n equality t "
s ~ R m, then a surplus variable vector v- E R m can The major difference between the existing recurrent
also be applied as follows: neural networks for optimization and the proposed de-
terministic annealing neural network lies in essentially
r(v) - S o v - - O, (11)
the treatment of their penalization procedures for vi-
where 0 _< v) _< I (i = 1, 2 . . . . . rn) and o denotes the olation of constraints. In the Hopfield network for op-
operator for array multiplication (element-by-element timization, the penalization is realized by constant
product). Once all the inequality constraints are con- penalty parameters (Hopfield & Tank, 1985, Tagliarini,
verted to equality constraints, a penalty function can Christ, & Page, 1992). As evidenced by many empirical
be constructed in a way similar to that of eqn (8); that studies, the steady state of the Hopfield network may
is. not represent a feasible solution or an optimal solution
to the optimization problem. In the recurrent neural
}[L,(v)+v÷]r[g(v)+v+], (12)
networks proposed earlier (Wang, 1991: Wang &
p~2~(v) - ½[r(v)-SoV ]T[r(v) SoY ], (13) Chankong, 1992), the penalization is realized by a
monotonically increasing penalty parameter. Because
p~,(v) - ½h(t') rh (v). (14)
the penalty parameter is supposed to be unbounded
Obviously, pt,(v) and Ph (v) are also differentiable and above, it is not easy to implement physically. In the
convex with respect to v. proposed deterministic annealing neural network, the
The penalty function can be constructed based on penalization is realized by using a time-varying tem-
p~,(v) and pt,(v); for example, perature parameter. The role of the time-varying tem-
perature parameter will be more apparent in Sec-
p(v) p~,(v) + Ph(V). (15)
tion 5.
Because p~(v) and ph(V) are convex and differentiable,
p(v) is also convex and differentiable. The convex pro-
gramming problem described in eqns ( 1 ), (2), and ( 3 ) 4. N E T W O R K ARCHITECTURE
thus can be reformulated as the following equivalent
convex programming problem: To solve an optimization problem via neural compu-
tation, the key is to cast the problem into the architec-
minimize f(v), (16)
ture of a recurrent neural network for which the stable
subject to p ( v ) = 0. (17) state represents the solution to the optimization prob-
lem. As shown in eqn (18), the architecture of the de-
terministic annealing neural network depends on spe-
3. N E T W O R K DYNAMICS cific forms o f f ( v ) and p ( v ) / i . e . , f ( v ) , g ( v ) , h ( v ) , and
the method for constructing p ( v ) ] .
The dynamical equations of the deterministic annealing If the functional transformation method [e.g., eqns
neural network can be described as follows: (7) and (8)] is used to construct a penalty function,
du(t) then
dt -T(t)~'f[v(t)] - 7p[v(t)], (18) m
tivity constraint, v > 0, can be realized by using a non- ' I ": ~ e x }------:,. . . . . . , / 2~( - i "'°
negative activation function F l u ( t ) ] >- O, only h ( v ) ~-~-- r:,. LS/ '\ \ ( .. .'J ~'
needs to be used in the penalty function. Let p ( v ) = } [ .... ,.\ x ~,
ph(V) = h ( v ) r h ( v ) / 2 = (Av - b)r(Av -b)/2, thus
Vp(v) = Vph(v) =ArAv - A'rb. Equation (18 be-
comes >-____$ .......
du(t) _ i /" i
A ~ A v ( t ) + Arb - T(/)c. (24)
dt 1 •
i w; • / ~. . . . ~,. ,
In the standard form of QP, f ( v ) = v r Q v / 2 + c ~ v , t ---- ~ / / 2q ~)n~ \\" '"
( t 8) becomes
FIGURE 2. Two-layer architecture of the deterministic annealing
du( t) _ neural network.
[ArA + T ( t ) Q ] v ( t ) + A~b T(t)c. (25)
dt
There are two ways to map the dynamical equations, column of A and Q respectively, 0, and c, are the ith
eqn (24) or (25), to a neural network architecture. elements of b and c. respectively. Because Q is sym-
The first method results in a single-layer recurrent metric, the connection weight matrices are symmetric
neural network. The single-layer architecture o f t b e de- for both LP and QP (i.e.. v i , j , wij - wii). I f Q is positive
terministic annealing neural network for convex pro- semidefinite and T(t) -> 0. then the eigenvalues of the
gramming consists of N laterally connected neurons, weight matrices are always real and nonpositive for both
as shown in Figure 1. Each neuron represents one de- LP and QP.
cision variable. The lateral connection weight matrix Letting e(t) = Av(t) - b. eqns ~24 ~ and (25) can
is defined as - A r A for an LP and - A r A - T ( t ) Q for be decomposed, respectively, as
a QR The biasing threshold vector of the neurons is
defined as A ~b - T ( t ) c for both LP and QP. In other duct) ~261
- ,4re(t} - T(l)c.
words, the connection weight w o from neuron j to dt
the coefficient matrices and vectors without involving Because J{ F[ u(t)] } is positive semidefinite, du(t) v~
matrix multiplication. dtJ { F[ u( t ) ] } du( t ) / dt >- O. Furthermore, because
f [ v ( t ) ] - m > 0 and d T ( t ) / d t < O, d E [ v ( t ) , T(t)]/
dt < 0. Therefore, E[v(t), T(t)] is positive definite
5. CONVERGENCE ANALYSIS and radially unbounded, d E [ v ( t ), T(t) ] /dt is negative
definite. According to the Lyapunov's theorem, the de-
In this section, analytical results on the stability of the terministic annealing neural network is asymptotically
proposed deterministic annealing neural network and stable in the large. •
feasibility and optimality of the steady-state solutions
to convex programming problems are presented.
5.2. Solution Feasibility
For the deterministic annealing neural network to re-
5.1. Asymptotic Stability
alize a solution procedure for a convex programming
THEOREM 1. Assume that the Jacobian matrix of problem, its steady state must represent, at least, a fea-
F[ u (t) ] exists and is positive semidefinite. I f the tem- sible solution to the problem. Toward this end, the fol-
perature parameter T( t ) is nonnegative, strictly mono- lowing theorem is derived.
tone decreasingJbr t >_ O, and approaches zero as time THEOREM 2. Assume that J{ F[u(t)] } exists and is
approaches infinit): then the deterministic annealing positive semidefinite. I f T(t) >_ O, d T ( t ) / d t < 0 and
neural network is asymptotically stable in the large," l i m , ~ T( t) = O, then the steady state of the determin-
that is, Vt >_ O, J{ F[u(t)] } is p.s.d., T(t) >_ O, d T ( t ) / istic annealing neural network represents a feasible so-
dt < O, and lim,~.~ T(t) - 0 imply Vv(0) ~ V, 3~ E lution to the convex programming problem described
V such that l i m , ~ v ( / ) = ~. where J { F [ u ( t ) ] } is" the in eqns ( 16 ) and ( 17 ), that is, F)E 9.
Jacobian ofF[ u(t)] and F) is a steady-state of v( t ).
Proof The proof of Theorem 1 shows that the energy
Proof. Becausef(v) is bounded from below, there exists function E [ v (t), T(t) ] is positive definite and strictly
M such that - o c < M < inf~,evf(V). Let monotone decreasing with respect to time t, which im-
plies lim~.... E [ v ( t ) , T(t)] = 0. Because l i m t ~ T(t)
E[v(t), T(t)] : T ( l ) I f [ v ( r ) ] - M} + p[v(t)]. (28)
= 0, l i m , ~ E[v(t), T(t)] = lira, ..... T ( t ) { f [ v ( t ) ] -
Because Vt >_ 0, T(I) >_ O,f[v(t)] - M > O, andp[v(t)] M} + p[v(t)] = limt .... p[v(t)]. Therefore, l i m t ~
>- O. Vv ~ V, E[ v( t ), T(t)] >_ 0. Furthermore, because ElY(t), T(t)] = l i m , ~ plY(t)] = 0. Because p ( v ) is
l i m , ~ T(t) = 0 and min~p(v) = 0, min~,rE[v, T] = continuous, l i m , ~ plY(t)] = p[limt .... v(t)] = p(~)
0. To facilitate the stability analysis, let us consider T(t) =0.
as a dummy-state variable for the time being; that is,
let f)(t) r = [ v(t) r, T(t)] be the augmented-state vector. REMARK. Theorems 1 and 2 show that the asymptotic
In view of the fact that .f(v) and p ( v ) are convex stability of the deterministic annealing neural network
and T(t) ~ [0, oc), E --* oo as [Iv[[ --~ oc; that is, implies the solution feasibility of the steady state.
E is radially unbounded. Because - d u ( t ) / d t =
T(t)Vf[v(t)] + 7p[v(t)] and v(t) = F[u(t)] according 5.3. Solution Optimality
toeqns (18) and (19), d v ( t ) / d t = J { F [ u ( t ) ] } d u ( t ) /
dt and Theorems 1 and 2 ensure the asymptotic stability and
solution feasibility. The optimality of the steady-state
dE[v(t). 7"(t)] dT(r)
dl ~f[v(t)] - M} dt solutions, however, is not guaranteed and deserves fur-
ther exploration.
df[v(r)] dp[v(r)] There are two possibilities for the locality of the op-
+ T(t) - - +
dt dt timal solution v*: an interior point or a boundary point
dT(t) of the feasible region 9 . If v* is an interior point of V,
: If[v(r)I-m} Jr then v* is the same as the global minimum of the ob-
jective function ,f(v), hence Vf ( v* ) = Vp (v*) = 0,
+ ~ T(r)Vf[v(t)] + Vp[v(/)] }r dr(t) because both f ( v ) and p ( v ) are convex and differen-
dt tiable. In this case, according to eqn (18), a steady
: { J [ v ( r ) ] - M} dT(t) du(t)rdv(r) state of the deterministic annealing neural network is
dt dt dt an optimal solution if the temperature parameter T(t)
dT(l) sustains sufficiently long. The solution process for this
: {f[v(r)] - M} dr case is trivial. In most convex programming problems,
such as linear programs, v* is a boundary point of V.
du(t) 7 du(t) In this case, Vv ~ 9, V f ( v ) ~ O. Theorem 3 provides
dt J { F [ u ( t ) ] } - ~ -
a sufficient condition for the deterministic annealing
634 J. !l km:z
neural network to solve convex programming problems Substituting eqn (18) and dv(t)/d~ - - J{ F l u / I ) ] :
in which the optimal solutions are boundary points. X dull)/dr into the above inequality, we have
and positive semidefinite requirements on J{ F[u(t)] } explicit bound on decision variable vector v, then a
can be replaced by strictly monotone increasing F~ for linear activation function, F[ u ( t ) ] = ~u (t), can be used
all i. The activation function F determines the domain where ( > 0 is also a scaling constant. In this case, V
of the state variables v(t) (i.e., the state space V) and =R N '
1t I+=10(/+=,
<
(~)
QP, if v(0) and J{F[u(t)] } is symmetric, as in most
of neural network models, then eqn (29) implies
max T ( t ) - T(O)
> 0.5 ~ = 0.1 t-0
brAJ{F[u(l)]]( c + Arb) l
>_max 0, c T y { g [ u ( t ) ] l i c + ~ i ~ j. (35)
,
010 -5 0 5 10
In addition, if F ( u ) = [Fj( ul ), F2(u2) . . . . . Fx( ux)]
1.1
and Ft = F2 = . . . = F~, then eqn (35) becomes
(u)
loc I,~:4(c + :i'/~)I
//" Vrnax = 10 max,,07'(t) = T(0)>_ max 0, c ~ - + - ~ 4 ~ j . (36)
/
= 1 ]" v~,: = 5
>" 5 // As discussed in Section 3, the terms - T(t)V/'[ v (t) ]
and - V p [ v (t) ] in the right-hand side of the dynamical
/f Omax 1
eqn (18) represent, respectively, the objective mini-
010 -5 0 i 10 mization effect and the constraint satisfaction effect.
Because a feasible solution is not necessarily an optimal
U
one, it is essential for the first term not to vanish until
FIGURE 3. Effects of ~ and Vm.xon sigmoid activation function. v(t) reaches an optimal solution. Theorem 3 suggests
636 J tlatl~
essentially that T ( t ) sustain sufficiently long to obtain terministic annealing neural network with different
an optimal solution. In other words, similar to the tem- initial states, Example 2 shows the effects of the design
perature parameter in the stochastic counterpart, the parameters/3. ~, and ~ on the performance of the de-
value of the temperature parameter in the deterministic terministic annealing neural network, and Example 3
annealing neural network should decrease slowly. A shows the performance of the neural network in solving
simple way to implement this requirement is to select a quadratic program.
a sufficiently small value of rl in eqns (31 ), ( 32 ), or
EXAMPLE I. Consider the following linear program with
(33). Because eqn (29) is not a necessary condition,
two decision variables and three functional constraints,
violation of this condition could still result in an optimal
solution. As will be shown in the illustrative examples min --2v~ - 3c~.
in the ensuing section, optimal solutions can still be
obtained when d f [ v ( t ) ] / d t ¢ d p [ v ( t ) ] / d t .
An exponential function, fl e x p ( - 0 t ) is used im- --v~ i- 2r: - _",
plicitly as the temperature parameter in the recurrent
3v~ + ~, - f,
neural networks for solving the assignment problem
and linear programming problems (Wang, 1992,
1993 ); that is, a = e in eqn (32) where e = 2.71838 . . . . . The optimal solution of this problem is v* = [ 1.4, 0.8] r
In this case, if the first term of the right-hand side of with the minimal objective v a l u e r ( v * ) -- -5.2. In this
dynamical eqn ( 18 ) is dominant as time elapses, then example, the inequality constraints are converted to
the convergence rate of the recurrent neural network equality constraints by adding threv slack-state van-
is exclusively dictated by n and the deterministic an- ables v3, v4, and v 5.
nealing neural network needs approximately 5 / n units
of time to converge, because ~/is the time constant of t'l ~ 2L'- t V.~ .~.
the temperature parameter. This feature enables one - - U I ~- 21.~2 4- F 4 - 2,
to control or determine the convergence rate of the de-
terministic annealing neural network.
Because fl = T ( 0 ) in eqns ( 3 1 ) - ( 3 3 ) , parameter fl where v~. v2, v3, V4, 1~5 ~> O,
defines the maximal and initial value of temperature Numerical simulations have been performed using
parameter T(t). According to eqn (34), the developed simulator based on vi(t) ~ vma~/[l ~-
Vp[v(O)lJS{Flu(O)l}Vptv(O) l ] e x p ( - ~ u i ( t ) ) ] , Y ( t ) =/3(1 -~ t)- ~, Vma~ = 2, ~ = 10 S.
~ J { F[ u( 0)1 }Vp[ v(0)_]~ /3 ~ 1. r/ = 103. and At -- 10 ~'. Yhe results of the
fl>_max O'vf[v(O)]rJ~F[u(O)]lVf[v(O)]t, [. (37) simulations show that the steady slates of the deter-
- Vp[v(O)]TJ{ F [ u ( 0 ) ] / V f [ v ( 0 ) ] J mmistie annealing neural network can always accu-
rately generate the optimal solution starting from four
In addition, fl is a design parameter related to conver- different comers of the 2 × 2 square on the v~ - v,
gence rate of the deterministic annealing neural net- plane; that is. ~ - [1.400000. 0.800001]'rand f(v) =
work and can be used to balance the penalization effect -5.200001. Figures 4 and 5 illustrate, respectively, the
on constraint violation and minimization effect on the
objective function. In general, a small fl results in an
underpenalization and a large fl results in an overpen-
alization of constraint violations. The convergence is
\
slow in either case. An excessively small fl could even \
\
result in convergence to a suboptimal solution, as will
be shown in Example 2. In other words, a properly
selected parameter/3 can make the deterministic an-
nealing neural network converge rapidly, Ifp,d v), Ph (V),
a n d f ( v ) are normalized in the same scale, then a design . 1,0--
rule is to set/3 ~ 1.
7. I L L U S T R A T I V E E X A M P L E S
Simulations have been performed using a simulator for
G
a number of sample problems. The following illustrative
examples are to demonstrate the operating character-
) 0 -X+-- . . . . . . . . . . . . . . . . . . . .
istics of the proposed deterministic annealing neural
network in solving linear and quadratic programs. Spe-
cifically, Example 1 shows the performance of the de- FIGURE 4. Feasible reg~a and state trajectodes in Example 1.
A Deterministic A nnealing Network 63 7
2,0 2.0
i
1.5 2) t 1.5 ~ 2)1
1.o 1.0
2)2 2)2
0.5 0.5
Initial S t a t e v(O) = ( 0 . 0 0 0 0 0 1 , 1.999999) T Initial S t a t e v(O) - ( 1 . 9 9 9 9 9 9 , 1.999999) 7
0.0 , = i i n i i r L ] n ] i = n i ¢ ~ L j , , n ~ n I = ¢ n I r T I I I r i n i I
0 2000 4000 6000 8000 2000 4000 6000 8000
Time Inferval Time Inferval
2.0 2.0
1.0 1.0
2) 2 122
0.5 0.5
Initial S t a t e v(O) = ( 0 . 0 0 0 0 0 1 , 0 . 0 0 0 0 0 1 ) 7
I n i t i a l S t a t e v(O) = ( 1 . 9 9 9 9 9 9 , 0 . 0 0 0 0 0 1 ) 1
0.8
-6- /
0.6
0.4
-7-
0.2
0.0 i in I I , , I I ~ , t I ~ I -8
........ • " ........ I . . . . . . . . . . . . . . . . . . . I
2000 4000 6000 8000 0 2000 4000 6000 8000
Time Interval Time Interval
1.0 4- E[.(t), t]
0.8
3-
0.6
2-
0.4
1
0.2
401
5.5
2)3
of the values of temperature parameter, objective func-,
tion, penalty function, and energy function versus the
number of time intervals with fbu~ different imtiai
states, which shows that the values of objective, penalty.
and energy functions with different starting points make
little difference after a few iterations.
2.o ~/' " ~ 2 ~
EXAMPLE 2. Consider the tbllowing linear program with
five variables and three constraints,
rain 2.4v, + i.6v2 .+ 4 . 2 r , + 5.2v~ i . ! . 4 v ,
FIGURE 7. Transient states of the deterministicannealing neural 1.3t,, i.2rL, + 2.5r:~-+ 4. it~, 2.?r< - 6 . 3 .
network in Example 2.
time units. Figure 6 illustrates the convergence patterns At = 10 -6 unless otherwise indicated. Figure 7 iltus-
~.o /[v(t)]
0.8 23
- \
0.6 21
/3=0.1
o.4 ~9 ~1~..~...____.___L--0.5
0.2
0.0
0 1000 2000 30•0 4000 5000 o 10o0 2000 3000 4000. ~00
Time Interval Time Intervol
0.8
is\
0.6
\
0.4
5-
0,2
FIGURE 8. Transient behaviors of the temperature parameter, objective function, penalty function, and energy f ~ based on
different B in Example 2.
A Deterministic .4nnealing Network 639
trates the transient behavior of the activation states of 24.157710. Simulation using the simulator shows
the deterministic annealing neural network. Because that the steady state of the deterministic annealing
= 103 and Zt = l 0 - 6 , 5 / r / = 0.005 and 5/(~2~t) = neural network ~ = [2.967033, 0.000000, 1.736263,
5000. Figure 7 clearly shows that the deterministic an- 0.417583] r and .f(~) = 24.157701, where the initial
nealing neural network can converge in approximately state is v(0) - [2.5, 2.5, 2.5, 2.5] r, vi(t) = V m a x / [ 1 +
5000 time intervals or 0.005 time units. Figures 8, 9, e x p ( - ( u i ( / ) ) ] , T ( t ) = fl[ln(t + e)] ', Vmax = 5, ( -
and 10 illustrate the transient behaviors of the values 10 4, fl = 1, r/ = 10 4, /~t = 1 0 6. Figure 11 illustrates
of the temperature parameter, objective function, pen- the transient behavior of the activation states, temper-
alty function, and energy function based on different ature parameter, objective function, penalty function,
values of design parameters/3, rt, and/~, respectively. and energy function in Example 3.
EXAMPLE 3. Consider the following convex quadratic
program with four variables and three constraints. 8. C O N C L U S I O N S
rain 1.5c~ - V,t'2 + V[V3 -- 2vlr'4 + 2v¢_ In this papen a deterministic annealing neural network
-- 2•2U3 + 2.5V 2 + V3U4 -- 3v4 for solving convex programming problems has been
proposed. The convergence properties of the determin-
6Vl + 15£ 2 + 9V 3 + 4134, istic annealing neural network have been analyzed, the
s.t. v~ + 2v2 + 4 v 3 + 5v4- 12, design principles have been discussed, and three illus-
trative examples have also been detailed. It has been
3v~ - 2v2 v3 + 2V4 = 8, substantiated that the proposed deterministic annealing
2131 3V2 + V3 -- 4V4 - 6, neural network is capable of generating optimal solu-
tions to convex programming problems. Having the
U I ~ 0 , U2 ~ 0 , ~33 ~ 0 , I)4 2> 0.
same convergence properties of its predecessor (Wang,
The optimal solution to this problem v* - 1991; Wang & Chankong, 1992), the proposed deter-
[2.967033, 0.000000, 1.736264, 0.417582] T and the ministic annealing neural network possesses more de-
corresponding value of objective function f ( v * ) sirable features. From a theoretical point of view, the
0.8
i9
0.6
18 = i000
0.4
~ = 500
17 ¸
0.2
= 100
1.0 p[~(t)]
0.8
0.6
0.4 1 0 " 7 5
0.2
FIGURE 9. Transient behaviors of the lemperature parameter, objective function, penalty function, and e n e r g y function based on
different
n in E x a m p l e 2.
640
,/ ~ f Llil/.:
,o..tic(t)]
35
3O
25 = 100
1.0
0.8
0.6
0,4
! (t)
20 ¸ = 1000 0.2
~ = 10000
\
15 o.o
0 1000 2000 3000 4000 5000 500 1000 ',500 2000
Time Interval Time Interval
~.o ~,b,(O1
0.8
20
0.6 /~ \ 15-
0.4
0.2 ~'00~ ~ ~
O I l l I I t 1 1 1 1 [ I n l t U l l l l T l / l l I [ 1 1 ] 1 , I I11 I ~
0.0 luuu ii , ul ii n i Li , , uu l, 17 i ni i in iii i~ it i ~ I - - [ T 7 q 0 500 1000 1500 2000
1o0o 2000 3000 4000 5000
Time Interval
Time Interval
20- [ (O,t ]
FIGURE 11. T r a n s i e n t ~ , ~ ~ , objective
function, penalty function, and en~'gy ~ in Example 3.
Kennedy, M. E, & Chua, L. O. (1988a). Unifying the Tank and Because[gi(v')-gi(v')]2>_O, Ml M[g,(v')2+gi(v")2]>-2Ml
Hopfield linear programming circuit and the canonical nonlinear - Mgi(v')g,(v"); that is, )`gi(v') 2 + (1 - X)g,(v"): >_)`2g~(v')2 +
programming circuit of Chua and Lin. IEEE Transactions on (l - M2gj(v") 2 + 2)`(1 - Mg~(v')g,(v"). Summing the above in-
Circuits and Systems, 34 ( 2 ), 210-214. equality over i and multiplying both sides with ½. we have
Kennedy, M. P., & Chua, L. O. (1988b). Neural networks for nonlinear l m
programming. IEEE Transactions on Circuits and Systems, 35( 5 ), 2 ,=~ [)`gi(v')2+(l )`)g'(v")2l
554-562.
Maa, C. Y., & Shanblatt, M. A. (1992). Linear and quadratic pro- l m
gramming neural network analysis. IEEE Transactions on Neural >--7 ~ [X2gi(/)')2 + (1 - X)2gi(V") 2 + 2X(I Mg,(v')g,(v")].
~M'tworks, 3, 580-594.
Tagliarini, G. A., Christ, J. E, & Page, E. W. (1991). Optimization
Because
using neural networks. 1EEE Transactions on Computers, 40( 12 ),
1347-1358. l m
Tank, D. W., & Hopfield, J. J. (1986). Simple neural optimization z-2i=lZ[X2gi(v'): + 2),(1 - Mg,(v')gi(v") + ( 1 )~)2gi(v") 2]
networks: An A / D converter, signal decision circuit, and a linear
programming circuit. IEEE Transactions on Circuits and Systems, l m
33(5), 533-541. = ~ ~ [)`gi(v') + (1 - )`)gi(v")] 2
2
Wang, J. ( 1991 ). On the asymptotic properties of recurrent neural
networks for optimization. International Journal of Pattern Rec- andXgi(v')+(l X)g(v")>g,(Xv'+(l - X ) v " ) d u e t o t h e c o n v e x i t y
~nition and A rtificial Intelligence, 5 (4), 581-601. ofg,(v) for i = 1, 2 . . . . . m,
Wang J. ( 1992 ). Analog neural network for solving the assignment
problem. Eh,ctronics Letters, 28( 1i ), 1047-1050. l S [)`gi(v')+(l - )`)gi(v")]2 -> ~1 ~ g,(;~v' + (I - )`)v") 2
Wang, J. (1993). Analysis and design of a recurrent neural network -~
i=1 -i=
for linear programming. IEEE Transactions on Circuits and Sys-
tems: Fundamental Theory and Applications, 40( 9 ), 613-618. = p~[xv' + (1 )`)v"].
Wang, J., & Chankong, V. (1992). Recurrent neural networks for
linear programming: Analysis and design principles. Computers Thus, the convexity property ofpg(v ) is proven. Note that the convexity
& Operations Research. 19(3/4), 297-311. Ofpg(V) does require the convexity ofg(v). The proof of the convexity
ofph(v) is similar.
APPENDIX NOMENCLATURE
Proof of the convexity of pg(v) in eqn ( 7 ): N the number of elements in a decision variable
pg(v)isconvexifXpg(v')+(l -Mpg(v")>_pg[)`v'+(l MY"],
Vv', v" ~ R N, )` E [0, 1 ]. According to the definition of the Heaviside
m the number of inequality constraints
function eqn (9), f o r / - 1, 2 . . . . . m 17 the number of equality constraints
I) the N-dimensional activation state vector and
e,(~)tt(x)dr={O,(V)2/2 if gi(v)>O, vector of decision variables
f " if gi(1)) <_ O. the N-dimensional net input vector to neurons
Because 0 < )` _< 1. ifgi (v') _< 0 and / or gi (v") _< 0, then the convexity
f(v) the real-valued objective function
is straightforward. Let us consider gz(v') > 0 and gi(v") > O.
g(v) the m-tuple function for inequality constraints
h(v) the n-tuple function for equality contraints
)`pdv')+(l
1 m
-Mpe(v")=3-~_,= [ ) ` g ' ( v ' ) 2 + ( l Mg~(v")2]. p(v) the real-valued penalty function
T(t) the temperature parameter