Professional Documents
Culture Documents
A Systematic Search Method For Obtaining Multiple Local Optimal Solutions of Nonlinear Programming Problems
A Systematic Search Method For Obtaining Multiple Local Optimal Solutions of Nonlinear Programming Problems
A Systematic Search Method For Obtaining Multiple Local Optimal Solutions of Nonlinear Programming Problems
2, FEBRUARY 19% 99
Abstract- We propose, in this paper, a systematic method to methods and stochastic methods [lo], [20], [25]. Deterministic
find several local optimal solutions for general nonlinear opti- methods include the branch and bound, cutting planes,
mization problems. We have developed some analyticalresults for decomposition-based, enumerative techniques, homotopy,
quasi-gradient systems and reflected gradient systems, applying
these results to derive topological and geometric properties of the interval analysis, and interior point methods. Stochastic meth-
critical points of the underlying objective function. A mechanism ods include random search, simulated annealing, simulated
has also been devised to escape from a local optimal solution evolution, genetic algorithms, and simulated diffusion. In
and proceed into another local optimal solution via decomposition general, stochastic methods can find global optimal solutions
points. By properly switching between quasi-gradientsystems and in an asymptotical way, while deterministic methods are not
reflected gradient systems, our proposed method can attain a set
of local optimal solutions. The proposed method is applied to two guaranteed, except in special cases, to find global optimal
test examples with promising results. solutions for general problems.
In this paper, we will present a systematic method, which
is deterministic in nature, to find several local optimal solu-
I. INTRODUCTION tions for general nonlinear optimization problems. Given an
METHOD
V. A SYSTEMATIC
As shown in the previous sections, a conceptual method to
systematically find several local minima includes the following
steps:
Step 1: Approach a local minimum.
Step 2: Escape from the local minimum and pass through
a decomposition point.
(b) Step 3: Approach another local minimum by moving along
Fig. 4. (a) The unstable manifold Wu(p) of the decomposition point z d the opposite direction of the unstable manifold of
converges to two stable equilibrium points and zz. (b) A mechanism to the decomposition point.
escape from a local minima and move on toward another local minimum can
be accomplished by the following procedure: staring from the local minima, The first step can be implemented by the descent method, or
one can move along the unstable manifold of a decomposition point in reverse by numerically integrating the negative quasi-gradient system
time as indicated in the dashed line.
(5). Various R ( x )can be chosen. For example, if R ( x )= I ,
it is called the steepest descent method; if R ( x ) = v2c(z),
One important issue in global optimization (instead of local it is called the Newton method; if R ( x )= u d j ( v 2 c ( ~ ) ) - 'it,
optimization) is how to escape from a local optimal solution +
is Branin's method [1]; if R ( z ) = (1 11 v2c(z)ll)I,it is
and move on toward another local optimal solution. This prob- Treccani's method [24]. Different R(x)'sin the quasi-gradient
lem has been studied by several researchers (see, for examples, system will have different numerical convergent properties.
[lo], [22], [25]). In terms of our proposed formulation, we For a given objective function, it is not yet clear how to choose
can see this problem as a question of how to escape from the an optimal R ( x ) which provides the best convergent result.
stability region of a stable equilibrium point and enter stability The second step is the most challenging task. A heuristic
regions of other stable equilibrium points so as to search for way is to move along the direction of the eigenvector of
other stable equilibrium points (correspondingly, local optimal the Jacobian matrix evaluated at the stable equilibrium point
solutions). The following results will offer some insight into associated with the largest eigenvalue. Then, continue the
this issue. movement until the matrix becomes semi-definite or until an
Proposition 6 (Decompositionpoints and the stable equilib- unbounded increase of the objective function is confirmed.
rium points): Let X Q be a stable equilibrium point of quasi- However, this method of ascent does not consistently produce
gradient system (5) and Xd be a decomposition point on the satisfactory results. A theory-based approach for escaping from
practical stability boundary dA(z:).If every equilibrium point a local minimum will be proposed in the next section. The
of quasi-gradient system (5) lying on aA(z;) is hyperbolic third step involves direct numerical integration of the negative
and its stable and unstable manifolds satisfy the transversal quasi-gradient system.
condition, then there exists another stable equilibrium point From a theoretical viewpoint, the above method builds a
~f to which the 1-D unstable manifold of X d converges. graph G describing the connections between the stable equi-
Prooj See Appendix. librium points and decomposition points with the following
Proposition 6 suggests that if one starts from a stable elements (see Fig. 5):
equilibrium point, and moves along the unstable manifold of the vertices of G are stable equilibrium points x:, . . ' , x;
a decomposition point in reverse time, then one can escape and decomposition points xi,. . , xi of the negative
e
from a local optimal solution and move on toward another quasi-gradient vector field -gradRc(z), where r and s
local optimal solution by moving along the other "direction" are the total number of stable equilibrium points and
of the unstable manifold in positive time. (See Fig. 4) decomposition points, respectively.
Locating the 1-D unstable manifold of the decomposition the edge of G can only,connect vertices of different types;
point can be accomplished by the following procedure [6] xi is connected with X: if, and only if, W"(x;)nA(x%,) #
[21, p. 1501: 0 and W U ( x ; )n (A(x",)" # 0.
104 IEEE TRANSACIlONS ON CIRcuTfS AND SYSTEMS-1: FUNDAMENTAL. THEDRY AND APPLICATIONS, VOL. 43, NO. 2, FEBRUARY 1%
else
compute the unstable eigenvector E" of
J-gradRc(z) ;
numerically integrate the negative quasi-
gradient vector field -gradRc(z) with initial
points f €23" until it approach a stable
Fig. 5. A graph showing the connections between stable equilibrium points
and decomposition points. equilibrium point y, where y V.
set V = V U ( ( y > \ Vmzn);
set Vmzn = Vmzn U{Y} ;
According to Proposition 6, each vertex x; is connected set E = E U{(z, Y)};
by two edges, which connects two different stable equilibrium
points. Moreover, s = r - 1. The total number of edges is endif;
equal to 2(7- - 1) and the total number of vertices is equal to
end;
21- - 1. The graph G is a connected tree without cycling if the
critical points of the objective function have distinct function end;
values [18, p. 4331. At the end of the algorithm, two partitions of vertices V,
We are now in a position to propose an algorithm to and Vd and an edge-list E will be obtained. Since the graph
systematically find several local minima. Let variables V,,,
G is connected, any two local minima is connected by one
path which is composed of a sequence of edges in E . Such a
and Vd record the location of local minima and decomposition
points, respectively; and let E list the edges of graph G graph G is a by-product of the proposed algorithm. It make
the algorithm noniterative among existing local minimum, i.e.,
with the element of the form ( x ~ , ~the ~ )set
; V denote
it prevents the deadlock of the algorithm. The global optimum
vertices of graph G. Each vertex in V will be examined to
can thus be obtained by comparing the values of all the local
determine whether it is adjacent to a new local minimum or
minima. It is clear that the key step in the proposed solution
to a decomposition point. At each iteration of the algorithm,
algorithm is to find all the adjacent decomposition points.
a point II: in V will be chosen: If z is a local minimum,
determine all new adjacent decomposition points; otherwise
z is a decomposition point which will be used to find another VI. IDENTIFYING POINTS
DECOMPOSITION
local minimum. The element of V will be added if some new As we have seen in the previous section, the search for
vertices appear. When the set V is empty, this algorithm will various local minima of an objective function can be ac-
be terminated. complished with the aid of decomposition points. Hence, it
is essential to determine such points. A straightforward way
An Algorithm to Find Multiple Local Minima to formulate this problem is via a nonlinear optimization
Step 1: [Initialize the data] problem: through which we may find the local minima over the
practical stability boundary. However, since we cannot derive
Step 1.1: set V,,, = Vd = V = 0; an explicit expression to characterize the practical stability
Step 1.2: set an initial guess zo; boundary, this method seems very difficult. An alternative way
Step 1.3: set a sufficient small number E; is to build a new vector field such that,in the new vector field,
Step 2: [Find the first local minimum] there exists a trajectory which starts from a stable equilibrium
point and converges to a decomposition point. To this end, we
Step 2.1: numerically integrate the negative will employ a method called reJEectedgradient [ 181, [ 191. This
quasi-gradient vector field -gradRc(z) with method can be used to change the quasi-gradient vector field
initial condition zo until the trajectory approaches in a systematic way such that the location of any equilibrium
one of a local minimum zQof e(.) (cfi Proposition point is invariant while its type will be altered in a more
3); explicit way.
Step 2.2: set V, = {zt} and V = {z:}; Let x be an equilibrium point of (5) and Al(z) 5
X,(x) . . . 5 Xn(z) be the eigenvalues of the Hessian
Step 3: [Build the corresponding graph G]
matrix v2c( z) , and wuz(z) be the normalized eigenvector
w ~ i ~V e# 0 do
corresponding to AJ(z),i.e., IivJ(z)lj= 1, for J = l , . . . , n.
begin
Let ~ , ( z )= w,(z)wT(z), j = I , . . . , % . Since v2c(z) i s
choose z E V, set V = V \ {z}; symmetric, P3(x)is the orthogonal projection matrix on the
if z E V, eigenspace associated with A, (z).By the spectral factorization
theorem [15, p. 1751, the Hessian matrix v2c(z) can be
find all the adjacent decomposition points decomposed into the direct sum of orthogonal projection
xi,.. . , xpd; matrices P J ( z ) ,i.e.,
set V = VU({z,!j,...,xpd}\Vd);
set Vd = Vd U{zi,. . . , zpd};
set E = E U{(z, xi),. . . , (2, zpd)};
CHIANG AND CHU: MULTIPLE LOCAL OPTIMAL SOLUTIONS 105
Now, we define a family of matrices Mk(x), k = 1,. . ,n, boundary a A ( x s )based on the reflected gradient method. The
variable D record all the adjacent decomposition points of ji..
K n
Step l:[Initialize the data]
Step 1.1:choosep points P j , j = l , . . . p , from a
neighborhood of a local minimum, say 2 , of f ;
and the k-th order reflected gradient vector field Fh(x),
Step 1.2: set a sufficient small number E ;
332 76
x = -x
3
+ 68xy2 - --IC
3
+
128x7 Iteration 1: Step 2 Iteration #2: Step 3 Iteration #3: Step 3
lnitlal z = $,0.2) initial z = z: initial z =
+
- 192x5 128x3y4 - 64xy4 - 128x3y2 (1 1)
332 76
y = 68x2y + -y3
3
- -y
3
+
128x4y3 - 64x4y
+
- 128x2y3 128y7 - 192y5
TABLE I11 V, we can verify that the total number of the edges in the
THE ITERATION
PROCESS 2. 0:
IN EXAMPLE graph G is 14 = 2(8 - 1).Since the corresponding graph G2
LOCALMINIMUM: 0 : DECOMPOSITION
POINT.
contains three loops, it can be concluded, from the discussions
in Section V, that some critical points of the objective function
Iteration f.1, Step 2 Iteration # Z Step 3 Iteration #3: Step 3 must have the same function value. The function values of the
initial z = 0 5 0 5 ) initial L = z! initial z = z:
local minima in Table IV confirm this result.
mFm
v = {z!)
Iteration # 4 Step 3
initial z = 2:
2: =:
v = {z:,zf,z:)
Iteration # 5 Step 3
initial L = zi
=:
2,
2
:
v = {z:,.;,.:}
Iteration #6: Step 3
initial z = z:
z:
4
2, *:
4
VIII. CONCLUDING
REMARKS
In this paper, we have proposed a systematic method to
find several local optimal solutions for general nonlinear
optimization problems. Instead of using the concept of the
stability region which is commonly used in existing literature,
we employed the concept of quasi-stability regions to develop
a new mechanism to escape from a local minimum via
v = {z:,z$,.:) v = (z:,.:,.:) v = {& 4 6) the unstable manifold of a decomposition point [l], [2],
Iteration #I: Step 3
initial z = zi
Iteration #S: Step 3
initial t = z:
Iteration #9: Step 3
initial z = z:
[lo], [13], [20], [22], [24], [25]. We have also developed
o-g-&-g 4 2:
some analytical results for quasi-gradient systems and applied
them, in combination with a theory of reflected gradient
systems to derive the topological properties of the critical
=t 4 points of the underlying objective function. Moreover, we
v = {z:,z:,s:,z:} v = {z:,z:,z:) have shown that the unstable manifold of any decomposition
Iteration # 1 0 Step 3 Iteration #11: Step 3 Iteration #12: Step 3
initial z = z: initial z = zf initial z = s: point converges to two stable equilibrium points (i.e., two
local minima solutions). This result helped us to reduce
computational burden. By properly switching between quasi-
gradient systems and reflected gradient systems, our proposed
method can obtain a set of local minima and also construct the
Iteration #13: Step 3 corresponding graph describing the connection configuration
initial t = z:
among the local optimal solutions and the decomposition
points. The global optimal solution can be obtained if the
associated graph contains all the decomposition points. The
proposed method is comprised of two levels: the lower level is
V=B continuous while the upper level is discrete in nature. Further
improvements are desirable in locating all decomposition
points.
TABLE IV APPENDIX
local minima e(.) decomposition points
z: = (0.08984,0.71266) -1.0316 z1= (1.10921,0.76827)
zf= (1.70361,0.79608) -0.2236 z! = (0,O)
Proof of Proposition I
ZS = (-0.08984, -0.71266) -1.0316 Z? = (1.29607,-0.60508) This proof consists of two parts. First, we will show
= (-i.60710,0.56s65). 2.0985 = (i.63807,-0.2~867)
2: = (1.60710,-0.56865) 2.0985 ~ f =; (1.29607,-0.60508) that every bounded trajectory will converge to one of the
2: = (-1.70361, -0.79608) -0.2236 zs (-1.10921,-0.76827) equilibrium points. Second, we will show that every trajectory
+d - (-1.29607,0.60508)
is bounded.
Let @(x,t) denote the bounded trajectory starting at
minimum, then 20 initial points P~ = 2 + €(COS sin 2, g), x. Consider the time derivative & c ( @ ( s , t ) )along the
j = 0,. f . ,19 are used to locate the decomposition points on trajectory. It is clear that
the practical stability boundary dA, (2); otherwise, we just use d
2 f €EU,where E" is the unstable eigenvector of J-vc(z), -c(@(x,t)) =
dt
as the initial conditions to compute other local minima. Table - (VC(@(Z, t)))TR(@(x,
t))-lVc(qz, t ) )5 0
111 describes the value of variables at each iteration of the
algorithm. Moreover, & c ( @ ( x , t ) )= 0 if, and only if, 2 E E .
From the result of Table 111, we find that this objective Therefore, c ( x ) is a Liapunov function of the negative
function has 6 local minima and 7 decomposition points, i.e., quasi-gradient system (5) and the w-limit point of any
T = 7 and s = T - 1 = 6. The complete results are shown bounded trajectory consists of equilibrium points only,
in Table IV. i.e., any bounded trajectory will approach one of the
By comparing the values of local minima, we can conclude equilibrium point [17, p. 2031.
that the global minimum is -1.0316, which occurs at x: Since e(.) is bounded below, we want to show that every
and 2:. Recalling the graph structure described in Section trajectory @(z,t ) is also bounded. Suppose @(x,t ) is not
10s IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS--I: FUND- THEORY AM3 APPLICmONS, VOL. 43, NO. 2, FEBRUARY 1W6
= /I: .(a(., t ) ) d t 5 -k
L*
IIVc(x)112dt 5 -kc2(t - T*)
metric matrix.
[7] T. F. Coleman, “Large-scale numerical optimization: Introduction and 1241 G. Treccani, “On the convergence of Brain’s method,” in Toward Global
overview,” Tech. Rep., TR91-1236, Dept. Computer Sciences, Comell Optimization. Amsterdam: North-Holland, 1975, pp. 107-116.
Univ., NY, 1991. [25] A. Tom and A. Zilinskas, Global Optimization, Lecture Notes in
[8] C. R. Corles, “The use of regions of attraction to identify global Computer Science, vol. 350. New York Springer-Verlag, 1987.
minima,” in Toward Global Optimization, I. C. W. Dixon and G. P. [26] S. Vavasis, Nonlinear Optimization: Complexity Issues. New York
Szego, Eds. Amsterdam: North-Holland, 1975, pp. 55-93. Springer-Verlag, 1990.
[9] C. L. De Marco, “Approximating power system dynamics and en- 1271 J. Zaborszky, G. Huang, B. Zheng, and T-C. Leung, “On the phase
ergy functions by quasi-gradient models,” in Int. Symp. Circuits Syst. portrait of a class of large nonlinear dynamic systems such as the power
(ISCAS’89), 1989, pp. 1966-1969. system,” IEEE Trans. Automat. Contr., vol. 33, pp. 4-15, 1988. ~