Professional Documents
Culture Documents
Computer Program Optimization of Gas Hydrates
Computer Program Optimization of Gas Hydrates
Computer Program Optimization of Gas Hydrates
1. Introduction
Among the techniques used for unconstrained optimization of a nonlinear
objective function in the least squares sense, two of the most popular have been
Gauss' method (also called the Gauss-Newton method) [1, 14] and the Levenberg-
Marquardt algorithm [t, 1t, t2, t4]. In this paper we give derivative free (finite
difference) analogues of these methods and prove local convergence for the algo-
rithms. We show how to select parameters in such a way that quadratic conver-
gence is guaranteed for any function which is zero at the minimum. Results of
computer experiments show how the methods compare to other currently used
techniques.
(F (x) is often the residual vector obtained when approximating a data set by
a nonlinear form.) The minimum, x*, is contained among the zeros of t/2 Vr (x) =
J (x)TF (x), where Vr (x) denotes the gradient of r and J(x) is the Jacobian
matrix of F.
* On leave 1970-71 at Yale University
** The work of this author was supported in part by the National Science Founda-
tion under Grant GJ-844.
20 Numer. Math., Bd. t8
290 K.M. Brown and J. E. Dennis, Jr. :
3. Convergence Results
It is well known (see [t2] and t4]) that the nonnegative sequence {/z,} can be
chosen so that $ (xn) forms a monotone decreasing sequence. If we make some
standard assumption, for example, that the level sets of $ are bounded and we
are careful in our choice of {#~} then some subsequence of {x,} is convergent to
an extreme point, say x*, of $. If the method is locally convergent, as soon as
some term of the convergent subsequence gets sufficiently close to x*, the entire
sequence will converge to it.
Lemma 1. Let D O be an open convex set and let K > 0 . If, for all x, yED o,
[[J (x) -J(y)[Is ~K[Ix --Y[[~,
then the following inequalities hold for every x, y EDo:
i) [[J (x) r - - J ( y ) T [I, <__K[[x--y[],
ii) there exist nonnegative constants C and C' such that
[[J(x) -J(y)[]l <=C. K [ [ x - y [ h
Derivative Free Nonlinear Least Squares 29t
and
][A[[~--C'][A[[1
/or any real rectangular matrix A.
iii) [ ] F ( x ) - - F ( y ) - J ( y ) (x--y)][l <=(CK/2)[[x--y[] ~.
Proo/. A proof of (i) follows from the fact that IIA II* =llA~ll, for any rectan-
gular matrix A. This in turn follows from page 54, [22].
Statement (ii) follows from a straightforward application of the equivalence
of the [1 and t~ vector norms.
Statement (iii) can be found in [7].
The following result is inspired by and closely related to the work in [3].
Theorem 1. Let J(x*)TF(x *) = 0 for some x*ED o and let
]]J(x)-J(y)][2<=K]]x-y[] , for x, yeD o.
If KHF(x* ) ]]2 is a strict lower bound for the spectrum of J(x*)Tj(x *) and {/,,}
is any bounded nonnegative sequence of real numbers, then there is an ~ > 0 such
that if every [h,~[< ~ and h, ,--> 0, the f.d.L.M, method converges locally to x*.
Proo/. For each xED o, let ~(x, h) denote the least eigenvalue of
h-*AF(x, h)TAF(x, h).
For ]hi sufficiently small and x in some smaller neighborhood (than Do) of x*, )Lis
a nonnegative jointly continuous function of x and h. Since by hypothesis
A(x*, 0) > K[IF(x*)II, > o ,
we can find 7' > 0 and D 1, an open convex neighborhood of x*, such that D1 m D o
and for xED 1, Ihl <=~', Z(x, h) > Z' =KI[F(x* ) II, + (Z(x*, 0) --KI]F(x*)11,)/2.
Let B denote the uniform bound on
[[h-XAF(x,h)Z][,. for xr 1, he[--~',V'].
(B exists by continuity.) Now set
and
L.~.I + hT~*AF(x,,, h.)r AF (x,,, It.)
and notice that the smallest eigenvalue of L . i s / t . +X(x., h.) ~ # . +X' > 0. Hence
L;~I exists and its [, norm is bounded by (/~. + $,)-i; moreover x.+ I (obtained
from the f.d.L.M, iteration) exists and, for some 0 E (0, I) we can write:
where xE [x,, x*] ~ N (x*, r). Hence, using the above and the order equivalence
of norms, Ih~l and/z, are 0(e,) and so from (t'), e,+ 1 =0(e~).
R e m a r k 2. Theorems I and 2 and Corollaries t and 2 hold for the f.d.G, and
Gauss methods respectively by merely setting/,~ = 0 in each statement and proof.
4. Numerical Results
In some twenty numerical experiments run, the f.d.L.M, algorithm behaved
almost identical to its analytic counterpart (the L.M. method). When convergence
occurred, both algorithms converged at the same rate; when one of the algorithms
diverged (or failed to converge in the prescribed number of iterations, t00) its
analogue usually did likewise. The same behavior was observed for the f.d.G.
algorithm and Gauss' method. These results answer in the affirmative sense one
of the questions posed by Bard [t]. We give a portion of the computational
investigations below and show a comparison with other methods based on the
Box survey paper [4]. All computations were done in APL\360 on an IBM 360/50.
(The APL\360 system performs all calculations in double precision.)
Although Theorem 2 dictates that hn and/~n be chosen as 0(llF(x,)ll,n~),
these choices must be tempered in the actual implementation of the algorithm to
prevent an absurd choice of hn relative to xn; e.g., suppose Ilxoll~ =0.00a but
IIF (~0) 11~ =to00. In the computer program the algorithm was implemented as
follows:
x.+l = x. - [ ~ z + J (x., h.) r J (x., h~)]-' J (x., h.) r F (x.),
~=c. IIF(x~)ll~,
t0 whenever t0 <IIF(~,)II=
c= whenever t <llF(x.)ll~ <to
[ 0.04. whenever I IF(~.)II= ~ a ;
294 K . M . Brown and J. E. Dennis, Jr. :
Table A. Number of function evaluations required to reduce ~ to less than 10-5 (Case A :
2 dimensions)
Table B. Number o] ]unction evaluations required to reduce q~ to less than 10-5 (Case B:
3 dimensions)
R e m a r k 4. T h e c o n v e r g e n c e o b t a i n e d b y t h e G a u s s a n d f . d . G , m e t h o d s t o t h e
e x t r e m u m (t 0, 1, - - 1 ) is i n t e r e s t i n g b e c a u s e B o x s t a t e s [4, p. 69] t h a t t h i s o p t i m u m
h a s n e v e r b e e n f o u n d w i t h a n y c o m b i n a t i o n of t h e e i g h t m e t h o d s a n d all s t a r t i n g
guesses which he tried.
R e m a r k a. I n o r d e r t o see j u s t h o w close t h e f . d . L . M , a l g o r i t h m a p p r o x i m a t e s
t h e L . M . m e t h o d w h e n o n e u s e s t h e c r i t e r i a f o r s e l e c t i n g h~ a n d / ~ , g i v e n a b o v e ,
:296 K . M . Brown and J. E. Dennis, Jr. :
F.D.L.M. L~
t.858t02805E0t 1.858t0 2806E0t
1.8t 716 4122E01 t.817t64391E0t
1.77806 1116E01 1.77806 2546E0t
1.74218 6983 E 01 1.742t9 t 53t E01
1.71019 2030E01 1.710199933E01
1.68023 4391 E01 1.68024 2170E0t
1.65076 3651 E01 1.65076 9860E01
1.62143 6158E01 1.62t 44 0580E01
1.592190436E0t 1.59219 3053E0t
1.56301 4585E01 t.56301 5403E01
1.3t 709 9176E01 1.31708 9027E 01
1.08275 9847E01 t .08273 83t2E01
8.62326 5064E00 8.62295 3369E00
6.58882 6077E00 6.58844 2804 E 00
4.76392 5553E00 4.76350 3998E00
3.19750 9939E00 3.19709 3467 E 00
1.94335 0326E 00 I. 94299 0431 E 00
1.15671 7605E00 1.15668 7275E00
7.70363 2711E--01 7.71008 07t8E--01
7.O7093 2946E--01 7.07718 6346E--0t
3.161890351E--0t 3.16772 2772E--0t
2.75553 9832E--02 2.77907 5637E--02
2.03121 t656E--04 2.10392 4422 E - - 04
1.265326821 E--08 1.43025 5381E--08
N o t e how n i c e l y t h e a s y m p t o t i c q u a d r a t i c convergence p r e d i c t e d b y T h e o r e m 2
is borne o u t b y this e x a m p l e .
Conclusions.
F . d . G . m a y be used i n s t e a d of Gauss' m e t h o d a n d f.d.L.M, i n s t e a d of t h e L.M.
m e t h o d w i t h little, if any, difference in t h e q u a l i t a t i v e results. T h e L.M. a n d
f.d.L.M, m e t h o d s a p p e a r m o r e s t a b l e t h a n t h e Gauss a n d f.d.G, m e t h o d s ; however,
w h e n t h e l a t t e r do converge t h e y seem to be m u c h f a s t e r t h a n t h e L.M. a n d
f.d.L.M, a l g o r i t h m s .
References
1. Bard, Y. : Comparison of gradient methods for the solution of nonlinear p a r a m e t e r
estimation problems. SIAM J. Numer. Anal. 7, t 5 7 - t 8 6 (t970).
2. Barnes, J. G. P. : An algorithm for solving nonlinear equations based on the
secant method. The Computer Journal 8, 66-72 (1965).
3. Boggs, P . T . , Dennis, J . E . , Jr. : Function minimization b y descent along a
difference a p p r o x i m a t i o n to the gradient. To appear.
4. Box, M. J. : A comparison of several current optimization methods, and the use
of transformations in constrained problems. The Computer Journal 9, 67-77
(t966).
5. Broyden, C. G. : A class of methods for solving nonlinear simultaneous equations.
Math. Comp. 19, 577-593 (1965).
6. Davidon, W. C. : Variable metric methods for minimization. A. E. C. Research
and Development Report, ANL-5990. (t 959) (Rev.).
Derivative Free Nonlinear Least Squares 297