Professional Documents
Culture Documents
Chap5 Optimal Control
Chap5 Optimal Control
Chap5 Optimal Control
B(q) C(q)
G(q) = , H(q) =
A(q) A(q)
with
103
104CHAPTER 5. CONTROL THEORY FOR DISCRETE-TIME SYSTEMS
Transfer Functions
The process is descried by the time invariant dynamical system with one input
with a time delay and one output, subject to disturbances
Bp∗ (q −1 )
y(t) = u(t − k) + v(t)
A∗p (q −1 )
If it is further assume that the the disturbance v(t) is a stationary normal process
with rational strictly positive spectral density, it can be represented as
Cd∗ (q −1 )
v(t) = λ
A∗d (q −1 )
where {e(t), t ∈ T } is a sequence of independent and identically distributed normal
(0, 1) random variables and Cd and Ad are polynomials. The polynomials Ad and
Cd can always be chosen to ave their zeros inside the unit circle.
The system can be represented by the equation
Bp∗ (q −1 ) Cd∗ (q −1 )
y(t) = u(t − k) + λ e(t) (5.3)
A∗p (q −1 ) A∗d (q −1 )
To simplify the analysis, the model (5.3) is rewritten to a slightly different form,
namely,
A∗ (q −1 )y(t) = B ∗ (q −1 )u(t − k) + λC ∗ (q −1 )e(t)
where A, B and C are polynomials defined by A = Ap Ad , B = Bp Ad and C =
Cd Ap . Put
A(z) = z n + a1 z n−1 + · · · + an
B(z) = b0 z n + b1 z n−1 + · · · + bn
C(z) = z n + c1 z n−1 + · · · + cn
Solutions
Theorem 5.1 Consider the process described by
A∗ (q −1 )y(t) = B ∗ (q −1 )u(t − k) + λC ∗ (q −1 )e(t) (5.4)
where {e(t), t ∈ T } is a sequence of independent normal (0, 1) random variables.
Let the polynomial C(z) has all its roots inside the unit circle. The minimum
variance control law is then given by
B ∗ (q −1 )F ∗ (q −1 )u(t) = −G∗ (q −1 )y(t)
5.1. MINIMAL VARIANCE CONTROL 105
C ∗ (q −1 ) = A∗ (q −1 )F ∗ (q −1 ) + q −k G∗ (q −1 ) (5.5)
y(t) = λF ∗ (q −1 )e(t)
= λ [e(t) + f1 e(t − 1) + · · · + fk−1 e(t − k + 1)]
Proof.
G∗ (q −1 ) B ∗ (q −1 )F ∗ (q −1 )
y(t + k) = λF ∗ (q −1 )e(t + k) + ∗ −1
y(t) + u(t) (5.6)
C (q ) C ∗ (q −1 )
Remark 5.2 A comparion with the solution of the prediction problem shows that
the term
G∗ (q −1 ) B ∗ (q −1 )F ∗ (q −1 )
y(t) + u(t)
C ∗ (q −1 ) C ∗ (q −1 )
of (5.6) can be interpreted as the k-step prediction of y(t + k) based on y(t), y(t −
1), . . . and that the control error equals the k-step prediction error.
where B1 is of degree n1 and has all roots inside the unit circle and B2 is of degree
n2 and has all roots outside the unit circle.
The minimal variance control strategy is
G∗ (q −1 ) q k G(q)
u(t) = − y(t) = − y(t) (5.8)
B ∗ (q −1 )F ∗ (q −1 ) B(q)F (q)
106CHAPTER 5. CONTROL THEORY FOR DISCRETE-TIME SYSTEMS
Let q n+k−1 operate on the identity (5.5) and use the definition of reciprocal poly-
nomials. We find
where there are k − 1 modes with poles at the origin, n modes with poles at the
roots of B, and n modes with poles at the roots of C. Furthermore, the factor BC
cancels in (5.11). This implies that the modes associated with BC are unobservable
from the output y, or that the corresponding state variables are not controllable
from the input e. Hence, the input might excite all the modes associated with the
solutions of the characteristic equation
z k−1 B(z)C(z) = 0.
This is not a serious matter if the modes are stable. However, if some modes are
unstable, it is possible to get infinitely large errors. This situation will occur if
the polynomial BC has roots outside or on the unit circle. It follows from the
representation theorems for stationary random processes that C can always be
chosen to have zeros inside or on the unit circle. As far as C is concerned the only
critical case would be if C has a zero on the unit circle. The polynomial B will
have zero outside the unit circle if the dynamics of the system is non-minimum
phase.
There are many ways to obtain control laws for non-minimum phase systems.
One possibility is to proceed as follows.
When solving the identity (5.5), the following identity is used
that is,
instead of (5.10). Going through the arguments used when deriving Theorem 5.1,
the control law is given by
G∗ (q −1 ) q k G(q)
u(t) = − =− y(t) (5.13)
B1∗ (q −1 )F ∗ (q −1 ) B1 (q)F (q)
The control law (5.13), which is not optimal, gives an error with the variance
2
var[y] = min var[y] + λ2 fk2 + · · · + fk+n
2 −1
.
must be within the unit circle. The variance of the output is then given by
I
1 dz
E y2 = H(z)H(z −1 )
2πj z
If the optimal system is sensitive to parameter variations, the control law (5.13)
can be used to make the closed-loop system less sensitive to parameter variations.
Assume that the system is governed by the model (Ao , Bo , Co , λo ) but the con-
trol law is calculated from the model (A, B, C, λ) with slightly different parameters.
The equation describing the controlled system then becomes
with ao = −0.7 and co = 0.95. The minimum variance control strategy for this
system is obtained from a = ao , b = 0.99, c = co . Use the identity (5.5)
1 + cq −1 = 1 + aq −1 + q −1 g.
It is assumed that v(t) isindependent of x(t) and the initial state x(t0 ) is normal
with
E [x(t0 )] = m
cov [x(t0 ), x(t0 )] = R0
To complete the probem statement, we must also specify the admissible control
strategy. For a stochastic control problem, it is very important to specify the data
which is available for determining the control action.
y(t) = x(t)
5.2.2 Preliminaries
Lemma 5.5 Let x be normal with mean m and covariance R. Then
E xT Sx = mT Sm + trSR
5.2. LINEAR QUADRATIC CONTROL 111
S(N ) = QN (5.23)
Then
N −1
T T T
X
x (N )QN x(N ) + x (t)Qx x(t) + u (t)Qu u(t)
t=t0
N −1 h i
T T T
X
= x (t0 )S(t0 )x(t0 ) + [u(t) + L(t)x(t)] Γ S(t + 1)Γ + Qu [u(t) + L(t)x(t)]
t=t0
N −1 n o
T T
X
+ v (t) S(t + 1) [Φx(t) + Γu(t)] + [Φx(t) + Γu(t)] S(t + 1) + S(t + 1) v(t)
t=t0
where the matrix L is defined in (5.24), which minimizes the expected loss. The
minimal value of the expected loss is
N
X −1
T
min E [J] = m S(t0 )m + trS(t0 )R0 + trS(t + 1)Rv (t)
t=t0
Theorem 5.9 Consider the system (5.21). Let the admissible control strategies
be such that u(t) is a function of Y(t). Assume that (5.22) with initial condition
(5.23) has a solution S such that S is positive semidefinite and Qu + ΓT SΓ is
positive definite for all t. Then, there exists a unique admissible control strategy
where the matrix L is defined in (5.24), which minimizes the expected loss. The
minimal value of the expected loss is
N
X −1
min E [J] = mT S(t0 )m + trS(t0 )R0 + trS(t + 1)Rv (t)
t=t0
N
X −1
trP (t)LT (t) ΓT S(t + 1)Γ + Qu L(t)
+
t=t0