Professional Documents
Culture Documents
1 Optimal Control: 1.1 Problem Definition
1 Optimal Control: 1.1 Problem Definition
Given the system x& (t ) = f (x (t ),u (t ), t ) an admissible control u has to be found so that
x follows an admissible trajectory and minimizes the cost function
tf
J = h(x (t f ), t f ) + ∫ g (x (t ), u (t ), t )dt
t0
2 Dynamic programming
2.1 Principle of dynamic programming
Though dynamic programming, the optimal control is fund reverse, starting from the
final state and back to the initial state, while de optimal control is recorded.
Principle of optimality: A control strategy, which is optimal in the interval [0,N], is
also optimal in the interval [k,N], with 0 ≤ k ≤ N . Presuming that J *k + 1 (x k + 1 ) is the
optimal cost from the state k+1 to the final state, only L(x k ,u k ) + J *k +1 (x k +1 ) has to be
optimized. In this case uk can be calculated from
*
uk
( *
) uk
( *
)
J k (xk ) = min L(xk ,uk ) + J k +1 (xk +1 ) = min L(xk ,uk ) + J k +1 (f (xk ,uk ))
This relation can be used for computing the optimal control, starting with
( x N ) = h (x N ) .
J *N
The result of the above equation is an optimal control found by trying all the
admissible control values for all the admissible states.
2.2 The algorithm (discrete dynamic programming)