Professional Documents
Culture Documents
Adaptive Control: Enough Environment
Adaptive Control: Enough Environment
1. Introduction
Adaptive Control 2. Self-oscillating Adaptive Control
3. Model Reference Adaptive Control
K. J. Åström 4. Estimation and Excitation
5. Minimum Variance Control
Department of Automatic Control, LTH
Lund University 6. Self-Tuning Regulators
7. Learning and Dual Control
October 23, 2020
8. Applications
9. Related Fields
10. Summary
4
10
problem was two-fold; one, it was taking a great deal of time to develop a
flight control system; and two, the system in existence were not capable of
2
fulfilling future Air Force requirements. These systems lacked the ability to
10
1
Yakov Z Tsypkin 1919 - 1997 Richard Bellman 1920 - 1984
◮ BS, Moscow Electrical Engineering Institute 1941 ◮ BA math Brooklyn College 1941
◮ BS, Moscow State University 1943 ◮ MA University of Wisconsin
◮ MS Engineering, Moscow State University 1945 ◮ Los Alamos Theoretical Physics
◮ PhD, Moscow State University 1948 ◮ PhD Princeton Lefschetz 1946
◮ Engineer, senior engineer, Chief of Department, ◮ RAND Corporation
Research Institute Aircraft Equipment 1941-1949
◮ Senior researcher, Institute Control Sciences,
◮ Founding editor Math Biosciences
Moscow 1950-1957 ◮ Brain tumor 1973
◮ Head of laboratory, Institute Control Sciences, ◮ 619 papers 39 Books
Moscow, since 1957
◮ Dynamic Programming
◮ Yakov Z. Tsypkin and C. Constanda Relay Control ◮ Lenin Prize 1960
Systems. ◮ Bellman Equation HJB
◮ Quazza Medal IFAC
◮ Sampling Systems Theory and Its Application 1984
◮ Curse of dimensionality
Volume 1 and 2 (NY 1964 Macmillan. Translated
◮ Hartley Medal IMC 1985 ◮ Bellman-Ford algorithm
from Russian by A. Allen an...)
◮ Foundations of the Theory of Learning Systems by
◮ Rufus Oldenburger ◮ John von Neumann Theory Prize (1976)
Medal ASME 1989
Tsypkin, Ya. Z. (1973) Paperback ◮ IEEE Medal of Honor (1979)
It’s is rather hard to tell when we at Honeywell first became interested in adaptive control.
... In retorspect, it seems that we first conciously articulated the need in connection with our
early work in automatic approach and landing. ...
1. Introduction
Let us look at the adaptive flight control system that was proven in the flight tests on the
2. Self-oscillating Adaptive Systems F-94C. Conceptually it is simple, deceptively so. The input is applied to a model whose
dynamic performance is what we whish the dynamic performance of the aircraft to be. The
3. Model Reference Adaptive Control
actual response is compared with the reponse of the model and the difference is used as
4. Estimation and Excitation an input to the servo. If the gain of the servo is sufficiently high, the response of the aircraft
will be identical to that of the modle, no matter what the elevator effectiveness, so long as it
5. Minimum Variance Control
is finite and has the right direction.
6. Self-Tuning Regulators Design of the model is fairly simple. ... The big problem comes inconnection with the need
to make the gain of the servo sufficiently high. An ordinary linear servo loop will not do. It
7. Dual Control
simply cannot be given a sufficiently high gain and still be stable. So we go in for
8. Applications non-linearity, the most extreme form of non-linearity, in fact - the bang-bang type. Full
available power is applied one way, or the other, depending on the the direction of the
9. Related Fields swithching order.
10. Summary Now it is well known that a simple bang-bang sytem is oscillatorye. And we don’t want an
oscillatory aircraft. .. So we look for ways to tame it down, keeping its high-gain
characteristic while reducing it oscillatory activity. ...
k
The Self-Oscillating Adaptive System SOAS P (s) =
s(s + 1)2
Filter Gain 1
changer y
ym
−1
Filter Process
ym y 0 10 20 30 40 50
Model Σ G f (s) Σ G p (s) Time
e 1 u
Dither −1
0 10 20 30 40 50
Time
−1
0.5 e
◮ Shape behavior by the block called Model
◮ Make the inner loop as fast as possible −0.5
SOAS Simulation - Adding Lead Network SOAS Simulation - Adding Gain Changer
1 1
y y
ym ym
−1 −1
0 10 20 30 40 50 0 10 20 30 40 50
Time Time
1 u 1 u
−1 −1
0 10 20 30 40 50 0 10 20 30 40 50
Time Time
0.5 0.5
e e
−0.5 −0.5
0 10 20 30 40 50 0 10 20 30 40 50
Time Time
2
The X-15 with MH-96 Autopilot Crash Nov 11 1967 Adaptive Control
K. J. Åström
1. Introduction
2. Self-oscillating Adaptive Control
3. Model Reference Adaptive Control
The MIT rule -sensitivity derivatives
Direct MARS - update parameters of a process model
Indirect MRAS - update controller parameters directly
L1 adaptive control - avoid dividing with estimates
4. Estimation and Excitation
5. Self-Tuning Regulators
6. Learning and Dual Control
7. Applications
Logic for gain change did not increase gain fast enough
8. Related Fields
Dydek, Zachary, Anuradha Annaswamy, and Eugene Lavretsky. “Adaptive Control and the
NASA X-15-3 Flight Revisited.” IEEE Control Systems Magazine 30.3 (2010): 32–48.
9. Summary
Process
dy
ym = −ay + bu
Model dt
Model
Controller parameters dym
Adjustment = −am ym + bm uc
mechanism dt
Controller
uc
u y
u ( t ) = θ 1 uc ( t ) − θ 2 y ( t )
Controller Plant
Ideal controller parameters
bm
Linear feedback from e = y − ym is not adequate for parameter θ 1 = θ 10 =
b
adjustment! am − a
θ 2 = θ 20 =
The MIT rule b
dθ e
= −γ e Find a feedback that changes the controller parameters so that the closed
dt θ
loop response is equal to the desired model
Many other versions
p + a + bθ 2
G(s)
−
θ1
e b Π
= uc θ2
θ 1 p + a + bθ 2 γ γ
−
e b2θ 1 b
s s
=− uc = − y
θ 2 (p + a + bθ 2 )2 p + a + bθ 2
am
s + am Π Π
am
s + am
Approximate
p + a + bθ 2 ( p + am
dθ 1 am
3 4
= −γ uc e
Hence dt p + am
dθ 1 am
3 4
= −γ uc e dθ 2 am
3 4
dt p + am =γ y e
dt p + am
dθ 2 am
3 4
=γ y e Example a = 1, b = 0.5, am = bm = 2.
dt p + am
3
Simulation a = 1, b = 0.5, am = bm = 2. MRAS - The MIT Rule
ym The error
1
y bθ 1 dx
−1 e = y − ym , y= uc p=
p + a + bθ 2 dt
0 20 40 60 80 100
Time
5 u
e b
0 = uc
−5
θ 1 p + a + bθ 2
2
0 20 40 60 80 100
e b θ1 b
Time
=− uc = − y
θ1 γ =5 θ 2 (p + a + bθ 2 )2 p + a + bθ 2
4
γ =1
2 Approximate
0
γ = 0.2 p + a + bθ 2 ( p + am
0 20 40 60 80 100
θ2
Time
γ =5 The MIT rule: Minimize e2 (t )
2
γ =1
γ = 0.2 dθ 1 am dθ 2 am
3 4 3 4
0
= −γ uc e, =γ y e
0 20 40 60 80 100 dt p + am dt p + am
Time
1 1 1
3 4
V ( e, θ 1 , θ 2 ) = e2 + (bθ 2 + a − am )2 + (bθ 1 − bm )2
2 bγ bγ
Lyapunov MIT
Derivative of error and Lyapunov function
de G m (s)
= −am e − (bθ 2 + a − am )y + (bθ 1 − bm ) uc −
dt
e G m (s)
Σ −
e
dV de 1 dθ 2 1 dθ 1 uc
Π
+
Σ
u y + Σ
=e (bθ 2 + a − am ) (bθ 1 − bm )
G(s) +
+ + uc u y +
γ γ Π Σ
−
dt dt dt dt
G(s)
θ1 −
Π θ1
Π
1 dθ 2 θ2
3 4
= − am e 2 + (bθ 2 + a − am ) − γ ye
θ2
γ γ
−
γ
γ γ
dt s s −
s s
1 dθ 1
3 4 am am
Π Π Π Π
+ (bθ 1 − bm ) + γ uc e s + am s + am
γ dt
Adaptation law
dθ 1 dθ 2 de
= −γ uc e, = γ ye [ = − e2
dt dt dt
Error will always go to zero, what about parameters, Barbara’s lemma!
4
Indirect MRAS - Estimate Process Model Indirect MRAS
Process and estimator
dx d x̂
= ax + bu, = âx̂ + b̂u
dt dt
Nominal controller gains: kx = kx0 = (a − am )/b, kr = kr0 = bm /b. Process and estimator
Estimation error e = x̂ − x has the derivative dx d x̂
= ax + bu, = âx̂ + b̂u
de dt dt
= âx + b̂u − ax − bu = ae + (â − a)x̂ + (b̂ − b)u = ae + ãx̂ + b̃u,
dt Control law
â − am bm
where ã = â − a and b̃ = b̂ − a. Lyapunov function u=− x+ r
b̂ b̂
11 2 Very bad to divide by b̂!
2V = e2 + ã + b̃2 .
2
γ
dV de 1 1 d â d b̂ 2 1 d ã 2 1 d b̃ 2
= ae2 + ex̂ +
1 1
=e + ã + b̃ ã + eu + b̃
dt dt γ dt dt γ dt γ dt
â − am bm
u=− x+ r 1. Introduction
b̂ b̂
b̂u + (â − am )x − bm r = 0 2. Self-oscillating Adaptive Control
3. Model Reference Adaptive Control
with the differential equation
4. Estimation and Excitation
du 5. Minimum Variance Control
= K bm r − (â − am )x − b̂u
! "
dt 6. Self-Tuning Regulators
Avoid division by b̂, can be interpreted as sending the signal 7. Learning and Dual Control
b̂m r + (am − â)x through a first order filter with the pole s = −K b̂. 8. Applications
9. Related Fields
L1 has been tested on several airplanes
10. Summary
I. M. Gregory, C. Cao, E. Xargay, N. Hovakimyan and X. Zou L1 Adaptive Control Design
for NASA AirSTAR Flight Test Vehicle. AIAA Guidance, Navigation, and Control
Conference Portland Oregon August 2011
5
Persistent Excitation - Examples Lack of Identifiability due to Feedback
a aˆ
Parameters
−1
a = −0.9, b = 0.5, σ = 0.5, θˆ(0) = 0, P (0) = 100I
P (s)C(s) P (s)
Y (s) = R(s) + V (s) Excitation
1 + P (s)C(s) 1 + P (s)C(s)
◮ Unit pulse at t = 50
C(s) C(s)P (s)
U (s) = R(s) − V (s) ◮ Square wave of unit amplitude and period 100
1 + P (s)C(s) 1 + P (s)C(s)
◮ Two cases
1
Y (s) = P (s)U (s) if v = 0, and Y (s) = − U (s) if r = 0. a) u(t ) = −0.2y (t )
C(s) b) u(t ) = −0.32y (t − 1)
Identification will then give the negative inverse of controller transfer
function! Any signal entering between u and v will influence closed loop
identification severely A good model can only be obtained if v = 0, or if v
is much smaller than r!
6
Steady State Regulation of Basis Weight and Moisture Content The Scene of 1960
Process model
7
Practical Issues Minimum Variance Control - Example
KJÅ, P Hagander, J Sternby Zeros of sampled systems. Automatica 20 (1), 31-38, 1984
2
uk = k (ysp − yk ) + uk −1 , k<
g(∞)
Ts t
g(∞)
Stable if g(Ts ) > kjå: Automatica 16 1980, pp 313–315. KJÅ, Computer control of a paper machine—An application of linear stochastic control
2
theory. IBM Journal of research and development 11 (4), 389-405, 1967
8
IBM Scientific Symposium Control Theory and Applications 1964 Summary
1. Introduction
◮ Born in Budapest 1930
2. Self-oscillating Adaptive Control
◮ BS MIT 1953
3. Model Reference Adaptive Control
◮ MS MIT 1954
4. Estimation and Excitation
◮ PhD Columbia University NY 1957
5. Minimum Variance Control
◮ IBM Research Yorktown Heights 1957-58
6. Self-Tuning Regulators
◮ RIAS Baltimore 1958-1964
The self-tuning regulator
Properties ◮ Professor Stanford 1964-1971
7. Learning and Dual Control ◮ Professor University of Florida 1971-1992
8. Applications ◮ Professor 1973 Professor ETH 1973-2016
9. Related Fields
10. Summary
R. E. Kalman, Design of a self optimizing control system. Trans. ASME R. E. Kalman, Design of a self optimizing control system. Trans. ASME
80,468– 478 (1958) 80,468– 478 (1958)
9
The Self-Tuning Regulator STR ... Test at Billerud 1973
Estimates of parameters in
Proven in
KJÅ and Wittenmark, Björn", On Self Tuning Regulators, April 1972, Technical Report
TFRT-7017, Department of Automatic Control, Lund University, Sweden. Presented at 5th U. Borisson and B. Wittenmark. An Industrial Application of a Self-Tuning Regulator, 4th
IFAC World Congress Paris March 1972. IFAC/IFIP Symposium on Digital Computer Applications to Process Control 1974.
KJÅ and B. Wittenmark. On Self-Tuning Regulators, Automatica 9 (1973),185-199. U. Borisson. Self-Tuning Regulators - Industrial Application and Multivariable Theory. PhD
thesis LTH 1975.
yt + a1 yt −1 + · · · + an yt −n = b0 ut −k + · · · bm ut −n p p 1 fp 2 1 2 2
dx = f (x )dt + g(x )dw , =− + g f =0
+ et + c1 et −1 + · · · + cn et −n t x x 2 x2
Estimation model dθ
θ t + 1 = θ t + γ t φ e, = f (θ ) = Eφ e
dτ
yt +k = s0 yt + s1 yt −1 + · · · + sm yt −m + r0 (ut + r1 ut −1 + · · · rn ut −{ )
Method for convergence of recursive algorithms. Global stability of STR
Theorem: Assume that (Ay = Bu + Ce) if G(z ) = 1/C(z ) − 0.5 is SPR
◮ Time delay k of the sampled system is known L. Ljung, Analysis of Recursive Stochastic Algorithms IEEE Trans AC-22 (1967) 551–575.
◮ Upper bounds of the degrees of A, B and C are known
Converges locally if ℜC(zk ) > 0 for all zk such that B(zk ) = 0
◮ Polynomial B has all its zeros inside the unit disc
◮ Sign of b0 is known Jan Holst, Local Convergence of Some Recursive Stochastic Algorithms. 5th IFAC
Symposium on Identification and System Parameter Estimation, 1979
The the sequences ut and yt are bounded and the parameters converge to
the minimum variance controller General convergence conditions
G. C. Goodwin, P. J. Ramage, P. E. Caines, Discrete-time multivariable adaptive control. Lei Gui and Han-Fu Chen, The Åström-Wittenmark Self-tuning Regulator Revisited and
IEEE AC-25 1980, 449–456 ELS-Based Adaptive Trackers. IEEE Trans AC36:7 802–812.
Hyperstate Xt = yt , yt −1 , . . . , y0 , ut −1 , ut −2 , . . . , u0 ,
10
The Hamilton-Jakobi-Bellman Equation A Sufficient Statistic - Hyperstate
The solution to the problem is given by the Bellman equation It can be shown that a sufficient statistic for estimating future outputs is yt
2 and the conditional distribution of b given Xt . In our setting the conditional
Vt (Xt ) = EXt min E yt2+1 + Vt +1 (Xt +1 )Xt
1
distribution is gaussian N b̂t , Pt
! "
ut
How to solve the optimization problem? In our particular case the conditional distrubution depens only on by y, b̂
The curse of dimensionality: Xt has high dimension and P - a significant reduction of dimensionality!
b̂ 1
Vt (y , b̂, P ) = min E yt2 + Vt +1 yt +1 , b̂t +1 , Pt +1 y , b̂t , Pt
1 2
u=− y u=− y as P → 0
! "
[
u b̂2 + P b̂
These control laws are called cautious control and certainty equivalence
control (Herbert Simon).
b̂ y b̂
u=− y = Ky , η = . β = √ ,
b̂2 + P σ P
2 2 2
1 2
Vt (y , b̂, P ) = min (y + b̂u) + σ + u P + Vt +1 yt +1 , b̂t +1 , Pt +1 )
!
u
Py 2
VT (y , b̂, P ) = σ 2 +
b̂2 + P
Ut (η, β , µ) = (η + β µ)2 + 1 + µ2
Z∞1 p p 2
+ Wt + 1 ( η + β µ + ε 1 + µ2 , β 1 + µ2 + µε φ (ε )dε
−∞
√
Solving minimization gives control law µ = Π(η, β ), µ = u σ P ,
u= √ σ
P Π(η,β )
Numerics:
◮ Transform to the interval (0 1), quantize U function 128 $ 128
◮ Store the a gridded version of the function U (η, β , mu)
◮ Evaluate the function W (η, β , µ) by extrapolation, and numeric
integration
◮ Minimize W (η, β , µ) with respet to µ K (η, β ) larger than 3 not shown
11
Understanding Probing Controller gain for 30 Steps
Notice jump!!
Comparison Summary
◮ Use dynamic programming (move optimization inside the integral)
T T
X 1X
2
Cautious Control Dual Control min E yt2 = min EXt E yk2 Xt
Ut Ut
k =t +1 k =t +1
T
1X 2 T
1X 2
= EXt min E yk2 Xt = EXt min min E yk2 Xt
Ut ut Ut +1
k =t +1 k =t +1
T 2
1 X
= EXt min E yt2+1 + min yk2 Xt
ut Ut +1
k =t +2
√
◮ State reduction y , b̂, P → η = y /σ , β = b̂/ P
◮ Certainty equivalence, cautious and dual
◮ K. J. Åström Control of Markov Chains with Incomplete State Information
JMAA, 10, pp. 174–205, 1965.
◮ K. J. Åström and A. Helmersson. Dual Control of an Integrator with Unkown
Gain, Computers and Mathematics with Applications 12:6A, pp 653–662,
1986.
1. Introduction
2. Self-oscillating Adaptive Control
3. Model Reference Adaptive Control
4. Estimation and Excitation
5. Minimum Variance Control
6. Self-Tuning Regulators
7. Learning and Dual Control
8. Applications
9. Related Fields Forget Physics! - Hope an STR can work!
10. Summary Power increased from 170 kW to 200 kW
R. Syding, Undersökning av Honeywells adaptiva reglersystem. MS Thesis LTH 1975
U. Borisson, and R. Syding, Self-Tuning Control of an Ore Crusher, Automatica 1976, 12:1,
1–7
12
Control over Long Distance 1973 Results
Plant in Kiruna, computer in Lund. Distance Lund-Kiruna 1400 km,
home-made modem (Leif Andersson), supervision over phone, sampling
period 20s.
◮ Ship dynamics
◮ SSPA Kockums C. Källström Identification and Adaptive Control Applied to Ship Steering PhD Thesis
◮ Full scale tests on ships Department of Automatic Control, Lund University, Sweden, April 1979.
C. Källström, KJÅ, N. E. Thorell, J. Eriksson, L. Sten, Adaptive Autopilots for Tankers,
in operation
Automatica, 15 1979, 241-254
1. Introduction
◮ One-button tuning
2. Self-oscillating Adaptive Control
◮ Automatic generation of
3. Model Reference Adaptive Control
gain schedules
◮ Adaptation of feedback and 4. Estimation and Excitation
13
The Perceptron - Rosenblatt Addaline - Bernard Widrow & Ted Hoff (Intel 4004)
◮ PhD Experimental Psychology, Cornell
1956: broad interests mathematics ...
◮ Cornell Aeronautical Lab 1957-59
◮ Probabilistic model for information storage
in the brain
◮ One layer neural network classifier
14