Professional Documents
Culture Documents
Structural and Multidisciplinary Optimization
Structural and Multidisciplinary Optimization
Optimization
3. Initial Points...................................................................................................4
4. Convergence Criteria......................................................................................5
Optimization Methods
1. Steepest Descent
2. Conjugate Gradients
3. Newton Basic
4. Newton-like
5. Quasi-Newton method - Boyden-Fletcher-Goldfard-Shanno (BFGS)
2.1 Function 1
Hessian ¿ ( 43 34)
For this function, no line search method is needed but they can still be
used.
2.2 Function 2
Unlike the first function, the second one is not a strictly convex quadratic
function. The Hessian matrix is not a symmetric positive definite. Below can be
seen the matrix for the gradient and the Hessian matrix (H m).
π π
( ( ) (( ) ))
( )
0.8∗x− 6∗ ∗sin ∗x +1
2 2
g=
π π
( (( )) (( ) ))
0.8∗y + 3∗
2
∗sin
2
∗y −1
π2 π
( ( ))
( )
0.8− 6 cos ∗x 0
4 2
H m=
π2 π
0 (
0.8+ 3
4
sin ∗y
2 ( ))
For this function, the optimisational alghoritms need a search line method. The
main used search line method in this report is the Quadratic Interpolation
method. However for the Newton-like optimization method, I used both the
Dichotomy and the Quadratic Interpolation.
In order to define the initial point which will give us the minimum value
for the objective function, I made the following assumption: x=y.
Therefore, the studied functions become:
2
f 1 ( x , y ) =7 x2 +8 x−1 f 2 ( x , y ) =0.8 x +3 cos ( π2 x) The next step, we calculate
the second derivatives of the functions, which leads us to the following results.
∂2 f ∂2 f π2 π
∂x 2
=14> 0
∂x 2
=1.6−3
4 2 ( )
cos x >0 , for x ≤−1
Now, in order to find out what is the x of the initial point, we impose that
the first derivatives of f1 and f2 must be equal to 0.
∂f1
=14 x +8=0=¿ x=−0.5714 , y=−0.5714
∂x
∂f2 π π
∂x 2 2 ( )
=1.6 x−3 sin x =0=¿ x=−1.75265 , y=−1.75265
The solutions must be close to these initial points. That means that the
Global Minimum values of the objective functions are obtained for these points.
For the sake of a better understanding of how the methods work, several
initial points have been chosen (as it will be seen in the Annexes). In this report,
the results of the following initial points will be analysed:
(0,0) for both functions – for the first function, it is a feasible point
but for the second one, the Hessian matrix will not be positive
definite. Some methods do not give good result because of this, but
it was a good way to check if the code was correct and to see how
the program will behave.
(10,10) for both functions – to see the difference in the number of
iterations and to show the convergence approaching by right
(-0.5714, -0.5714) for the first function – the point for which the
first derivative of f1=0
(-1.75265, -1.75265) for the second function – the point for which
the first derivative of f2=0
4. Convergence Criteria
Convergence Criteria:
for the first function
∂ f k +1
max
|
i=1 ,… n ∂ xi
(x ) < ε
|
It is a strict convex quadratic function and this criteria gives the best
computational time.
For this method, first take an initial guess for X initial, which in my case is
(0,0). Using the iterative method, the program moves along the downwards
direction, until it reaches the optimum value of the objective function.
Sk =−∇ f k =−∇ f ( X k )
Where:
Sk = search direction
∇ f k = gradient of the function at point Xk
Xk = the analyzed point
Xk+1 = Xk + αSk = Xk – α∇ f k
Where:
α = the optimal step length in the defined direction
Line Search:
f1 – since it is a SCQF, no line search was needed
2
‖g‖
α= '
s ∗Hm∗s
Where:
s’ = the direction matrix transposed
Hm = the Hessian matrix
s = the direction matrix
Algorithm steps
1. Initialization
a) Choose initial point, x0, which needs to have real coordinates
b) Set k=0
In the table above can be seen the results for the 3 initial points, in the
case of the first function. It is easy to see (and just as expected) that for the
second point, which is the closest to X*, the program needed only 2 iterations to
reach the Global Minimum. The big number of iterations needed for the first
initial point ([0,0]) is caused by the asymptotic behaviour of the function at the
named point. The higher the value of the derivative in a point, the faster the
convergence.
Since the second function is not a strictly convex quadratic function, the
function will not converge only to one stationary point. It presents local
minimizers as well, thing that is noticed in the table above as well. As it can be
seen, for X0 = [0,0], the method converges to a local minimizer while for
X0 = [10,10] it reaches the Global minimum. However, for the x closest to X*, it
gets remarkably close to the global minimizer, yet not there.
The cosine creates waves in the x direction which makes it difficult to
figure out how fast the convergence will be. Therefore, for X0= [0,0] the
program needed 21 iterations while for the other 2 initial points, the global
minimum was reached in only 6 iterations.
The zig-zagging behaviour is noticed again.
Algorithm Steps
1. Initialization
a) Choose initial point, X0, which needs to have real coordinates
b) Set k=0
3. Iteration k
a ¿ X k+1= X k + α k d k
k −d k' g k
b) Compute α =
dk ' H m dk
This method takes a more direct path than the Conjugated Gradients
method. However, it requires the Hessian matrix for each iteration, to find the
extrema. It is perfect for SCQF and for functions with positive definite Hessian
matrix, otherwise it will not reach the global minimizer. As it will be seen
further in this report, for the second function, it found only local minimizers.
This is because this method does not use a line search.
In this method, α is not computed for the search direction. Instead, the
inverse of the Hessian matrix is multiplied by the gradient. The descent
direction for the Basic Newton method is given by the formula shown below:
−1
X k+1= X k −[∇2 f ( X k ) ] ∇ f ( X k )
It is amazingly easy to notice that this method, for the first function,
reaches the global minimizer in just 1 iteration. The convergence to a global
minimizer from both sides (left and right) of X ¿ was verified, even if X 0 is not
close to the said global minimizer. (As it will be shown in the annexes,
X =[100 100] is convergent.)
0
Even from the path of the method, it is seen that the method is going
straight for X*.
For the second function, the Newton Basic method reaches only local
minimizers. The local minimizers are found depending on the initial point
chosen. By running the method only for one initial point, it cannot be said if it
reached the global minimum or not.
5.4 Newton-like
The Newton-like method is the first adaptation of the Newton Basic
Method. Compared to the basic one, this method converges for the non-
SCQF too. In order for the non-SCQF to converge, in this method it is used
following adaptation:
−1
X k+1= X k −α k [∇ 2 f ( X k ) ] ∇ f ( X k )
Algorithm Steps
f ( X k +α k s k ) <f ( X k )
In the case of the first function, the results are similar to the Newton
Basic method. The global minimizer is reached in one single iteration. The
convergence is verified no matter how close or how far the chosen initial point
is from X*.
However, for the second function, the method reaches the minimum
only if the chosen X0 is close to X*. In the table above it can be seen that, just as
the Newton basic, the method stops at the first local minimizer met on the path.
The Dichotomy doesn’t give a good result for the initial point [10, 10] because
of the Hessian matrix. Therefore, I used the Quadratic Interpolation for that
point.
For the first function, it is noticeable that the method reached X* but it
took 2 iterations to get to it.
Unlike the Newton-like method, this one reached the global minimum
when the chosen initial point is close to X *. However, the results suggest that
the minimizer point that the method finds is dependent on the initial point.
25. Newton Like, function 1, X0 = [10, 10], X* = [5.4286, -6.5714], Obj. F =-39.2857
26. Quasi-Newton, function 1, X0 = [10, 10], X* = [5.4286, -6.5714], Obj. F =-39.2857
29. Newton Basic, function 2, X0 = [10, 10], X* = [9.2943, 0.12234], Obj. F = 38.1108
30. Newton Like, Quadratic Interpolation, function 2, X0 = [10, 10], X* = [1.8315,
3.7241], Obj. F = -3.517