Professional Documents
Culture Documents
Feedback
Feedback
Feedback
Feedback and control theory are important ideas that should form part of the education of a physicist
but rarely do. This tutorial essay aims to give enough of the formal elements of control theory to
satisfy the experimentalist designing or running a typical physics experiment and enough to satisfy the
theorist wishing to understand its broader intellectual context. The level is generally simple, although
more advanced methods are also introduced. Several types of applications are discussed, as the
practical uses of feedback extend far beyond the simple regulation problems where it is most often
employed. Sketches are then provided of some of the broader implications and applications of control
theory, especially in biology, which are topics of active research.
ticularly deep consequences for biological systems, and comes when they are faced with having to use or modify
all physicists should have some understanding of such a a PID !proportional-integral-derivative" control loop
basic concept. Indeed, further progress in areas of cur- that regulates some experimental quantity, such as tem-
rent interest such as systems biology is likely to rely on perature or pressure. Lacking the time to delve more
concepts from control theory. deeply into the subject, they turn to the semiqualitative
This article is a tutorial essay on feedback and control discussions found in the appendixes of manuals for com-
theory. It is a tutorial in that I give enough detail about mercial regulators and the like. While these can be
basic methods to meet most of the needs of experimen- enough to get by, they rarely give optimal performance
talists wishing to use feedback in their experiments. It is and certainly do not give a full idea of the range of
an essay in that, at the same time, I hope to convince the possible approaches to a particular experimental prob-
reader that control theory is useful not only for the en- lem. Naturally, they also do not give an appreciation for
gineering aspects of experiments but also for the con- the broader uses of control theory.
ceptual development of physics. Indeed, we shall see Thirty years ago, E. M. Forgan wrote an excellent in-
that feedback and control theory have recently found troduction to control theory for experimental physicists,
applications in areas as diverse as chaos and nonlinear “On the use of temperature controllers in cryogenics”
dynamics, statistical mechanics, optics, quantum com- !Forgan, 1974", which, despite its seemingly narrow title,
puting, and biological physics. This essay supplies the addresses many of the needs of the experimentalist dis-
background in control theory necessary to appreciate cussed above. However, it predates the widespread use
many of these developments. of digital control and time-domain methods, as well as
The article is written for physics graduate students important advances in control theory that have taken
and interested professional physicists, although most of place over the past three decades. In one sense, this es-
it should also be understandable to undergraduate stu- say is an updated, slightly more accessible version of
dents. Beyond the overall importance of the topic, this Forgan’s article, but the strategies for control and the
article is motivated by a lack of adequate alternatives. range of applications that are relevant to the physicist
The obvious places to learn about control theory— are much broader than what is implied in that earlier
introductory engineering textbooks !Dutton et al., 1997; work. I have tried to give some feeling for this breadth,
Franklin et al., 1998, 2002; Goodwin et al., 2001"—are presenting simple ideas and simple cases in some detail
not very satisfactory places for a physicist to start. They while sketching generalizations and advanced topics
are long—800 pages is typical—with the relevant infor- more briefly.
mation often scattered in different sections. Their ex- The plan of this essay, then, is as follows: In Sec. II, we
amples are understandably geared more to the engineer review some elementary features of dynamical systems
than to the physicist. They often cloak concepts familiar at the level of an intermediate undergraduate mechanics
to the physicist in unfamiliar language and notation. course. In Sec. III, we introduce the simplest tools of
And they do not make connections to similar concepts control theory, feedback and feedforward. As case stud-
that physicists will have likely encountered in standard ies, we discuss how adding a control loop can increase
courses. The main alternative, more mathematical texts the usefulness of a Michelson interferometer as a dis-
#e.g., in order of increasing sophistication, the books by placement sensor and how feedback plays an essential
Özbay !2000", Morris !2001", Doyle et al. !1992", and role in modern analog electronic circuits. In Sec. IV, we
Sontag !1998"$, are terse but assume the reader already discuss the relationship between feedback and stability,
has an intuitive understanding of the subject. focusing on how time delays generically arise and can
At the other end of the intellectual spectrum, the first limit the amount of feedback gain that may be applied.
real exposure of many experimentalists to control theory In Sec. V, we discuss various practical issues of imple-
mentation: the identification of system dynamics, the
choice of control algorithm and the tuning of any param-
control took place entirely in England, at the beginning of the
industrial revolution of the 18th century. Mayr speculates that
eters, the translation of continuous-time designs to dis-
the concept of feedback arose there and not elsewhere because crete difference equations suitable for programming on
it fit in more naturally with the prevailing empiricist philoso- a computer, the use of commercial software packages to
phy of England and Scotland !e.g., Hume and Locke". On the simulate the effects of control on given systems, and the
Continent, the rationalist philosophy of Descartes and Leibniz problems posed by noisy sensors and other types of un-
postulated preconceived goals that were to be determined and certainty. We include a case study of an active vibration-
achieved via a priori planning and not by comparison with isolation system. This section includes introductions to a
experience. While Mayr’s thesis might at first glance seem far- number of advanced topics, including model reduction,
fetched, the philosophical split between empiricism and ratio- optimal control, Kalman filtering, and robust methods.
nalism was in fact reflected in various social institutions, such
We deliberately intersperse these discussions in a section
as government !absolute vs limited monarchy", law !Napole-
on practical implementation, for the need to tune many
onic code vs case law", and economics !mercantilism vs liber-
alism". Engineering !feedback control vs top-down design" is parameters, to deal with noise and uncertainties in the
perhaps another such case. Mayr devoted much of his subse- form of the system model itself—all require more ad-
quent career to developing this thesis !Bennett, 2002". Else- vanced methods. Indeed, historically it has been the fail-
where, Bennett !1996" gives a more detailed history of control ure of simpler ideas that has motivated more complex
theory and practice in the 20th century. methods. Theorists should not skip this section!
1 1
II. BRIEF REVIEW OF DYNAMICAL SYSTEMS V̇out!t" = − Vout!t" + Vin!t". !2.3"
RC RC
In one way or another, feedback implies the modifica-
tion of a dynamical system. The modification may be Here, n = m = p = 1, x = y = Vout, u = Vin, A = −1 / RC, B
done for a number of reasons: to regulate a physical = 1 / RC, C = 1, and D = 0.
variable, such as temperature or pressure; to linearize A slightly more complicated example is a driven,
the response of a sensor; to speed up a sluggish system damped harmonic oscillator, which models the typical
or slow down a jumpy one; to stabilize an otherwise un- behavior of many systems when slightly perturbed from
stable dynamics. Whatever the reason, one always starts equilibrium and is depicted in Fig. 2, where
from a given dynamical system and creates a “better”
one. mq̈ + 2"q̇ + kq = kqo!t". !2.4"
We begin by reviewing a few ideas concerning dy-
We can simplify Eq. !2.4" by scaling time by #20 = k / m,
namical systems. A good reference here and for other
the undamped resonant frequency, and by defining $
issues discussed below !bifurcations, stability, chaos, etc."
is the book by Strogatz !1994". The general dynamical = !" / m"%m / k = " / %mk as a dimensionless damping pa-
system can be written rameter, with 0 % $ % 1 for an underdamped oscillator
and $ & 1 for an overdamped system. #In the physics lit-
x!˙ = !f!x! ,u! ", !2.1a" erature, one usually defines the quality factor Q = 1 / $.$
Then, we have
y! = g! !x! ,u! ", !2.1b"
2
where the vector x! represents n independent “states” of One finds second-order behavior in other types of sensors,
a system, u! represents m independent inputs !driving too, including thermal and hydraulic systems. For thermal sys-
terms", and y! represents p independent outputs. The tems !Forgan, 1974", mass plays the role of electrical capaci-
tance and thermal resistivity !inverse of thermal conductivity"
vector-valued function !f represents the !nonlinear" dy- the role of electrical resistance. For hydraulic systems, the cor-
namics and g! translates the state x! and “feeds through” responding quantities are mass and flow resistance !propor-
the input u! directly to the output y! . The role of Eq. tional to fluid viscosity". At a deeper level, such analogies arise
!2.1b" is to translate the perhaps-unobservable state if two conditions are met: !1" the system is near enough ther-
modynamic equilibrium that the standard “flux' gradient”
variables x! into output variables y! . The number of inter-
rule of irreversible thermodynamics applies; !2" any time de-
nal state variables !n" is usually greater than the number pendence must be at frequencies low enough that each physi-
of accessible output variables !p". Equations !2.1" differ cal element behaves as a single object !“lumped-parameter”
from the dynamical systems that physicists often study in approximation". In Sec. IV.C, we consider a situation where
that the inputs u! and outputs y! are explicitly identified. the second condition is violated.
y = !1 0"&'
x1
x2
+ 0 · u!t", !2.7"
L#y!t"$ ( y!s" = )
0
)
y!t"e−stdt. !2.8"
Note that many physics texts use the notation + Both the low-pass filter and the second-order system
= 1 / #0, so that the Fourier transform of Eq. !2.12" takes have transfer functions G!s" that are rational functions;
the form 1 / !1 + i#+". i.e., they can be written as M!s" / N!s", where M!s" and
In control-theory books, log-log graphs of +G!i#"+ and N!s" are polynomials. Not every system can be written in
linear-log graphs of arg„G!i#"… are known as Bode plots, this form. For example, consider a sensor that faithfully
and we shall see that they are very useful for under- records a signal, but with a time delay -t, i.e., v!t" = y!t
standing qualitative system dynamics. #In the physics lit- − -t". From the shift theorem for Laplace transforms,
erature, ,!#" ( G!i#" is known as the dynamical linear the transfer function for such a sensor is G!s" = e−s-t,
response function !Chaikin and Lubensky, 1995".$ In Fig. which is equivalent to an infinite-order system. Note that
3, we show the Bode plots corresponding to Eqs. !2.13" the magnitude of G is always 1 and that the phase in-
and !2.14". Note that the asymptotes in Fig. 3!a" inter- creases linearly with frequency. In contrast, in the earlier
cept at the cutoff frequency #0 and that in Fig. 3!b", the two examples, the phase tended to an asymptotic value.
phase lag is −90° asymptotically, crossing −45° at #0. The convolution theorem allows important manipula-
Note, too, that we break partly from engineering nota- tions of transfer functions. If we define G ! H
tion by using amplitude ratios in Fig. 3!a" rather than
( *)0 G!+"H!t − +"d+, then L#G ! H$ = G!s"H!s". Convolu-
decibels !1 dB= 20 log10+G+". tion is just the tool one needs to describe compound
The transfer function G!s" goes to infinity when s systems where the output of one element is fed into the
= −#0. Such a point in the complex s plane is called a input of the next element. Consider, for example, a first-
pole. Here, we see that poles on the negative real s axis order sensor element that reports the position of a
correspond to exponential decay of impulses, with the second-order mechanical system. We would have
decay rate fixed by the pole position. The closer the pole
is to the imaginary s axis, the slower the decay. Poles in ÿ + 2$ẏ + y = u!t", !2.16"
the right-hand side correspond to exponentially increas- and
ing amplitudes.
Similarly, using the Laplace !Fourier" transform, the v̇ + v = y!t", !2.17"
transfer function for the second-order system is
where u!t" drives the oscillator position y!t", which then
drives the measured output v!t". Laplace transforming,
1 1
G!s" = , or G!i#" = , 1
1 + 2$s + s2 1 + 2i$# − #2 y!s" = G!s"u!s" = · u!s",
!2.15" 1 + 2$s + s2
1
and Bode plots for various damping ratios $ are shown v!s" = H!s"y!s" = · y!s", !2.18"
in Fig. 4. !Recall that we have scaled #0 = 1." Here, there 1+s
are two poles. In the underdamped case !$ % 1", they and thus v!s" = H!s"G!s"u!s", implying an overall loop
form a complex-conjugate pair s = −$ ± i%1 − $2. In the transfer function F!s" = H!s"G!s". Having two elements
overdamped case !$ & 1", the two poles are both on the in series leads to a transfer function that is the product
real s axis. of the transfer functions of the individual elements. In
G0
A. Basic ideas: Frequency domain G!s" = . !3.3"
1 + s/#0
Consider a system whose dynamics are described by Here, we have added a dc gain G0. We apply the sim-
G!s". The goal is to have the system’s output y!t" follow plest of control laws, K!s" = Kp, a constant. This is known
a control signal r!t" as faithfully as possible. The general as proportional feedback, since the feedback signal u!t"
strategy consists of two parts: First, we measure the ac- is proportional to the error signal, u!t" = Kpe!t". Then
tual output y!t" and determine the difference between it
and the desired control signal r!t", i.e., we define e!t" K pG 0
= r!t" − y!t", which is known as the error signal. Then we T!s" =
KpG0 + 1 + s/#0
apply some “control law” K to the error signal to try to
minimize its magnitude !or square magnitude". K pG 0 1
= . !3.4"
In terms of block diagrams, we make the connections KpG0 + 1 1 + s/#0!1 + KpG0"
shown in Fig. 6, where a control law K!s" has been
added. Manipulating the block diagrams, we have This is again just a low-pass filter, with modified dc gain
and cutoff frequency. The new dc gain is deduced by
y!s" = K!s"G!s"e!s", !3.1" taking the s → 0 !# → 0" limit in T!s" and gives
KpG0 / !KpG0 + 1". The new cutoff frequency is #! = #0!1
K!s"G!s" L!s" + KpG0". This can also be seen in the time domain,
y!s" = r!s" = r!s", !3.2"
1 + K!s"G!s" 1 + L!s" where
where the loop gain L!s" ( K!s"G!s". Starting from the ẏ = − #0y + #0KpG0!r − y", !3.5"
system dynamics G!s", we have modified the “open-
loop” dynamics to be L!s" = K!s"G!s", and then we have or
transformed the dynamics a second time by “closing the
loop,” which leads to closed-loop dynamics given by T 4
( KG / !1 + KG" = L / !1 + L". One hopes, of course, that Physicists often informally use “positive feedback” to denote
the situation where −e = y − r is fed back into the controller. If
the control law is u = Kp!−e", then the feedback will tend to
drive the system away from an equilibrium fixed point. Using
the terminology defined above, however, one would better de-
scribe this as a situation with negative feedback and negative
gain.
5
Negative feedback can also lead to oscillation, although typi-
FIG. 6. Block diagram illustrating closed-loop control of a sys- cally more gain—and hence more control energy—is required
tem G!s". Controller dynamics are given by K!s". than if positive feedback is used.
e = r − 0 − y, !3.12"
which implies
KG 1
y!s" = #r!s" − 0!s"$ + d!s". !3.13"
1 + KG 1 + KG
In the previous example, disturbances d!t" would have
FIG. 7. Block diagram illustrating closed-loop control of a sys- been rejected up to the cutoff frequency #! = #0!1 + Kp".
tem G!s" subject to disturbances d!s" and measurement noise This is usually desirable. On the other hand, the control
0!s". The control signal r!s" is assumed to be noise free. If signal effectively becomes r − 0: the system has no way to
present, the block F!s" adds feedforward dynamics. distinguish the control signal r!t" from measurement
noise 0!t". Thus the higher the gain is, the noisier the
output. !Putting a “prefilter” between the reference and
ẏ = − #0!1 + Kp"y!t" + KpG0r!t". !3.6" the feedback node a is a partial solution. The controller
In effect, the driving signal is u!t" = #0KpG0!r − y". If the then has two degrees of freedom, as described above."
control signal r!t" = r) = const, then the output settles to The frequency #! is known as the feedback band-
y) = #KpG0 / !KpG0 + 1"$r). width. We can thus state that a high feedback bandwidth
If we had instead the open-loop dynamics, is good, in that it allows a system to track rapidly varying
control, and bad, in that it also allows noise to feed
ẏ!t" = − #0y!t" + #0KpG0r!t", !3.7" through and contaminate the output. This tradeoff may
also be expressed by rewriting Eq. !3.13" as
we would have the “bare” cutoff frequency #0 and a
final value y) ( y!t → )" = KpG0r). e0!s" ( r!s" − y!s" = S!s"#r!s" − d!s"$ + T!s"0!s", !3.14"
It is interesting to compare the open- and closed-loop where e0 is the tracking error, i.e., the difference be-
systems. The closed-loop system is faster for Kp & 1. tween the desired tracking signal r and the actual output
Thus, if the system is a sensor, sluggish dynamics can be y. This is distinguished from e = r − y − 0, the difference
speeded up, which might be an advantage for a measur- between the tracking signal and the measured output. In
ing instrument. Furthermore, the steady-state solution, Eq. !3.14", S!s" is the sensitivity function discussed above
and T!s" = KG / !1 + KG" = 1 − S is the complementary sen-
K pG 0
y) = r) , !3.8" sitivity function. The general performance goal is to
K pG 0 + 1 have e0 be small given various “inputs” to the system.
Here, we regard the tracking signal r, the disturbance d,
tracks r) closely if Kp . 1 / G0. One might counter that in
and the sensor noise 0 as inputs of differing kinds. A
the open-loop system, one could set Kp = 1 / G0 and have
fundamental obstacle is that S + T = 1 at all frequencies;
y) = r) exactly. However, if G0 varies over time—for ex-
ample, amplifier gains often drift greatly with thus if S is small and disturbances are rejected, then T is
temperature—this tracking will not be good. By con- large and sensor noise feeds through, and vice versa. We
discuss possible ways around this tradeoff, below.
trast, for Kp . 1 / G0, the closed-loop system tracks r)
without much sensitivity to the dc gain G0.
We can sharpen our appreciation of this second ad- B. Basic ideas: Feedforward
vantage by introducing the notion of the parametric sen-
sitivity S of a quantity Q with respect to variations of a Another basic idea of control theory is the notion of
parameter P. One defines S ( !P / Q"!dQ / dP". We can feedforward, which is a useful complement to feedback.
then compare the open- and closed-loop sensitivities of Say that one wants a step change in the reference func-
the transfer function with respect to the dc gain, G0: tion. For example, in a loop controlling the temperature
of a room, one can suddenly change the desired setpoint
G0 d from 20 to 25° C. The feedback loop may work satisfac-
Sopen = !KpG0" = 1, !3.9"
K0G0 dG0 torily in response to room perturbations, but if one
knows ahead of time that one is making a sudden
1 change, one can do better than to just let the feedback
Sclosed = / 1. !3.10" loop respond. The usual way is to try to apply a prefilter
1 + K pG 0 F!s" to the control signal r!s". In the absence of any
Thus, if KpG0 . 1, the closed-loop dynamics are much feedback, the system response is just y!s" = G!s"F!s"r!s".
less sensitive to variations in the static gain G0 than are If we can choose F = G−1, then y will just follow r exactly.
the open-loop dynamics. Because the actuator has finite power and bandwidth,
We next consider the effects of an output disturbance one usually cannot simply invert G. Still, if one can ap-
d!t" and sensor noise 0!t". From the block diagram in ply an approximate inverse to r, the dynamical response
Fig. 7 #ignoring the F!s" block for the moment$, will often be significantly improved. Often, this amounts
to inverting the system G!s" at low frequencies and leav-
y = KGe + d, !3.11" ing it unchanged above a cutoff frequency set by actua-
tor limits or by uncertainty in the system’s high- tice, one puts a “gain” on the feedforward term that
frequency dynamics. See Devasia !2002" and Sec. V.E. reflects the correlation one expects between correspond-
Similarly, if one has an independent way of measuring ing vertical pixels. Unity gain—unity correlation—
disturbances one can apply a compensator element that implies that one expects the new line to be exactly the
feeds forward the filtered disturbance to the actuator. same as the old line, justifying the full use of feedfor-
!For example, a thermometer outside a building allows ward. If there is no correlation—i.e., if statistically,
one to anticipate disturbances to an interior room." knowing the height of the pixel in a given column of the
Again, the idea is to try to create a signal that will coun- previous scan tells you nothing about the expected
teract the effect of the disturbance. height in the next scan—then one should not use feed-
Another general motivation for using feedforward is forward at all. The actual gain that should be applied
to control the movement of flexible structures. For ex- then should reflect the expected correlation from line to
ample, translation stages have mechanical resonances. line. Note the relationship here between information
Imagine that one wants a response that approximates a and feedback/feedforward, in that the appropriate
step displacement as closely as possible. If one uses as amount of feedback and feedforward is determined by
the control signal a step displacement, the high frequen- the amount of information one scan line gives regarding
cies of the step function will excite the system’s mechani- the next; see Sec. IX, below.
cal resonances. Adding a feedforward element allows Finally, we can now better appreciate the distinction
one to “shape” the input so that resonances are not ex- between “control” and “feedback.” The former refers to
cited; the actual response can then be closer to the de- the general problem of how to make a system behave as
sired step than it would be were the “naive” input used desired. The latter is one technique for doing so. What
instead !Singhose, 1997; Croft and Devasia, 1999; Schit- we have seen so far is that feedback, or closed-loop con-
ter and Stemmer, 2004; Zou et al., 2005". trol, is useful for dealing with uncertainty, in particular
Because it is usually impossible to implement perfect by reducing the effects of unknown disturbances. On the
feedforward and because disturbances are usually un- other hand, feedforward, or open-loop control, is useful
known, one generally combines feedforward with feed- for making desired !i.e., known" changes. As Eq. !3.15"
back. In Fig. 7, one includes the prefilter F!s" after the shows, one usually will want to combine both control
time varying setpoint r!s" before the loop node. The out- techniques.
put y!s" #Eq. !3.13"$ then becomes
defines x! ! = T̃x! ", the transfer function deduced in Eq. As a quick example !Franklin et al., 2002", consider an
!3.17" will not change. The state vector x! is an internal undamped pendulum of unit frequency #Eq. !2.6", with
representation of the dynamics, while G!s" represents $ = 0$. We desire to move the eigenvalues from ±i to −2
the physical input-output relation. and −2. In other words, we want to double the natural
Because the internal state x! usually has more ele- frequency and change the damping $ from 0 to 1 !critical
ments than either the number of inputs or outputs !here, damping". Let u = −!k1x1 + k2x2" in Eq. !2.6", which leads
just one each", a few subtleties arise. To understand to a new dynamical matrix,
these, let us rewrite Eqs. !3.16" for the special case
where à is a 2 ! 2 diagonal matrix, with diagonal entries Ã! = & 0 1
− 1 − k1 − k2
' . !3.20"
11 and 12. Then
Computing the characteristic equation for Ã! and
ẋ1 = 11x1 + b1u, matching coefficients with the desired equation #!1 + 2"2
= 0$, one easily finds k! T = !3 4". More systematically,
ẋ2 = 12x2 + b2u, there are general algorithms for choosing k! so that Ã!
has desired eigenvalues, such as “Ackermann’s method”
y = c 1x 1 + c 2x 2 . !3.18" for pole !eigenvalue" placement !Dutton et al., 1997; Sec.
5.4.7".
If b1 = 0, then clearly, there is no way that u!t" can influ- In the above discussion of a SISO system, the control
ence the state x1!t". More generally, if any element of b! law used the full state vector x! !t" but there was only a
is zero, the system will not be “controllable.” Likewise, single observable y!t". This is a typical situation. One
if any element of c! is zero, then y!t" will not be influ- thus needs some way to go from the single dynamical
enced at all by the corresponding element of x! !t", and variable y to the full state x! . But since y is a linear func-
we say that that state is not “observable.” More for- tion of all the x’s and since one knows the system dy-
mally, a system is controllable if it is always possible to namics, one can, in fact, do such an inversion. In the
choose a control sequence u!t" that takes the system control-theory literature, the basic strategy is to make an
from an arbitrary initial state to an arbitrary final state observer that takes the partial information as an input
in finite time. One can show that a formal test of con- and tends to the true state.7 Since the observed system is
trollability is given by examining a matrix made out of being modeled by computer, one has access to its inter-
nal state, which one can use to estimate the true dynami-
the vectors Ãib! , for i = 0 , . . . , n − 1. If the controllability
cal state. In the absence of measurement noise, observ-
matrix Ũb is invertible, the system is controllable. The ers are straightforward to implement !Dutton et al.,
n ! n matrix Ũb is explicitly 1997; Özbay, 2000". The basic idea is to update an exist-
ing estimate x̂!t" of the state using the dynamical equa-
Ũb ( !b! Ãb! Ã2b! ¯ Ãn−1b! ". !3.19" tions #Eqs. !2.2"$ with the observed output y!t". The aug-
mented dynamics are
In the more general MIMO case, b! will be a matrix B̃,
x̂˙!t" = Ax̂!t" + Bu!t" + F#y!t" − Cx̂!t"$. !3.21"
and the controllability condition is that ŨB have full
rank. One can also show that a similar test exists for For simplicity, we have dropped tildes !A, B, etc." and
observability and, indeed, that controllability and ob- vector symbols !x, y, etc.". The new matrix F is chosen so
servability are dual properties !corresponding to how that the error in the estimate of x!t" converges to zero:
one gets information into and out of the dynamical sys-
tem" and that any property pertaining to one has a coun- ė!t" = ẋ!t" − x̂˙!t" = !A − FC"e!t". !3.22"
terpart !Lewis, 1992". Note that we use “observable” To converge quickly, the real parts of the eigenvalues of
here in its classical sense. Section VIII contains a brief
A − FC should be large and negative; on the other hand,
discussion of quantum control.
large eigenvalues will lead to noisy estimates. A reason-
One can show that if the technical conditions of con-
able compromise is to choose F so that the estimator
trollability and observability are met, then one can
converges several times faster than the fastest pole in
choose a k! that will place the eigenvalues anywhere.6 the system dynamics. The estimator described here is a
The catch is that the farther one moves an eigenvalue,
the larger the elements of k! must be, which will quickly
7
lead to unattainable values for the input u. Thus one There are actually two distinct control strategies. In the
should move only the eigenvalues that are “bad” for “state-vector-feedback” !SVF" approach, one uses a feedback
system response and move them as little as possible law of the form u! = −K̃x! along with an observer to estimate x!
!Franklin et al., 2002". from the measurements y! . In the “output-feedback” approach,
one uses directly u! = −K̃!y! . The SVF approach is conceptually
cleaner and more systematic, and it allows one to prove gen-
6
This is true even if the system is unobservable, but then the eral results on feedback performance. The output-feedback
internal state x! cannot be deduced from the observed output y. approach is more direct—when it works !Lewis, 1992".
y!t − +", one has information about ẏ!t". One can show
that y! + obeys dynamics like that of the true system !more
precisely, its phase space is topologically similar". Note
that one usually must choose the embedding dimension
n. If n is too small, trajectories will cross in the
n-dimensional phase space, while if n is too large, one
will merely be embedding the system in a space that has
a larger dimension than that of the actual state vectors.
Measurement noise effects are magnified when n is cho-
sen too large. The method of time delays works even if
one does not know the correct system. If the system is
linear, then there is little excuse for not figuring it out
!see Sec. V.A", but if the system is nonlinear, then it may
not be easy to derive a good model and the method of
time delays may be the best one can do.
The direct, “state-space” formulation of feedback
with its “pole-placement” methods is easier for simulat-
FIG. 8. !Color in online edition" Phase-space plot of harmonic ing on a computer and has other mathematical advan-
oscillator and its observer. The harmonic oscillator trajectory tages, while the frequency domain is often more intui-
starts from the dot and traces a circle. The observer starts from tive. One should know both approaches.
the square with incorrect initial velocity but converges to track To summarize !and anticipate, slightly", the possible
the oscillator. For simplicity, the control signal u = 0. advantages of adding a feedback loop include the fol-
lowing:
!i" altered dynamics !e.g., faster or slower time con-
“full-order” estimator, in that it estimates all the states x stants";
whether they are observed or not. Intuitively, it is clear !ii" the ability to linearize a nonlinear sensor by hold-
that if one defines each observable y! to be one of the ing the sensor value constant !discussed below";
elements of x! , then it should be possible to define a !iii" reduced sensitivity to environmental disturbances.
“partial-order” estimator, which uses the observations Possible disadvantages include the following:
where they exist and estimates only the remaining un- !i" sensor noise may contaminate the output;
known elements of x! . This is more efficient to calculate !ii" a stable system may be driven unstable.
but more complicated to set up !Dutton et al., 1997". We shall deal with the potential disadvantages below.
Finally, if there is significant measurement noise, one We now give an example that highlights the use of feed-
generally uses a rather different strategy, the Kalman back to linearize a nonlinear sensor signal.
filter, to make the observer; see Sec. V.D.
To continue the example of the harmonic oscillator,
above, choosing F ! T = !4 3" gives Ã! repeated eigenvalues D. Two case studies
at −2. The phase-space plot !x1 vs x0" in Fig. 8 shows 1. Michelson interferometer
how the observer tracks the physical system. The oscil-
lator, starting from x0 = 0, x1 = −1 traces a circle of unit In Sec. III, we looked at controlling a device that was
radius in phase space. The observer dynamical system merely a low-pass filter. Such an example might seem
starts from different initial conditions but eventually academic in that one rarely encounters such a simple
converges to track the physical system. Here, the ob- system in the real world. Yet, as the following case study
server time scale is only half the natural period; in a real shows, simplicity can often be forced on a system, pur-
application, it should be faster. After the observer has chased at the price of possible performance.
We consider the Michelson interferometer designed
converged, one simply uses its values !x!ˆ ", fed by the ob- by Gray et al. for use in the next-generation Laser Inter-
servations y, in the feedback law. Obviously, in a real ferometer Gravitational Wave Observatory !LIGO"
application, the computer simulation of the observer gravity-wave project !Gray et al., 1999". The gravity-
must run faster than the dynamics of the physical sys- wave detector !itself an interferometer" must have all its
tem. elements !which are several kilometers long!" isolated
The above methods are satisfactory for linear equa- from Earth-induced vibrations, so that any gravity-wave-
tions. In the physics literature, one often uses the induced distortions may be detected. In order to isolate
“method of time delays” !Strogatz, 1994". Here, an the large masses, one can measure their position relative
n-element state vector is made by taking the vector y! + to the earth—hence the need for accurate displacement
( „,y!t" , y!t − +" , . . . , y#t − !n − 1"+$-…, where the delay + is measurement. Of course, the interferometer may be
chosen to be the “dominant” time constant !the choice used in many other applications, too. We refer the
often requires a bit of playing around to optimize". Mea- reader to the article of Gray et al. for details about the
suring y! + is roughly equivalent to measuring the first n project and interferometer. Here, we consider a simpli-
− 1 time derivatives of y!t". For example, from y!t" and fied version that highlights their use of feedback.
/ 0
KG 1 + !Kp/Ki"+s 11
T!s" = = , !3.32"
1 + KG 1 + !1 + Kp"+s Ki + +2s2 Ki
5 5 12
D̃ = . !4.3"
!
which is 1 for # → 0 and is asymptotically first order, with
1n
time constant + / Kp, for # → ).
We have seen that a system with integral control al- The solution y! = 0 is stable against infinitesimally small
ways tends to the setpoint, whatever the value of Ki. perturbations if Re 1i % 0 , ∀ i. Generally, the eigenvalues
This sounds like a trivial statement, but it is our first of the stability matrix à change when a control
example of how a feedback loop can lead to robust be- parameter—such as a feedback gain—is changed. If one
havior that does not depend on details of the original of the eigenvalues has positive real part, the associated
system itself. In other words, it is only the loop itself and eigenmode will grow exponentially in time. !This is lin-
the fact that integration is employed that leads to the ear stability; a solution may be stable to infinitesimal
tracking properties of feedback. We return to this point perturbations but unstable to finite perturbations of
in Sec. VII, which discusses biological applications of large enough amplitude. Such issues of nonlinear stabil-
feedback. ity are relevant to the hysteretic systems discussed be-
low." The qualitative change in behavior that occurs as
the eigenvalue crosses zero is known as a bifurcation.
IV. FEEDBACK AND STABILITY Once the system is unstable, the growth of the unstable
mode means that nonlinear terms will quickly become
Feedback can also be used to change the stability of a important. The generic behavior of the bifurcation is
dynamical system. Usually, this is undesired. For ex- then determined by the lowest-order nonlinear terms.
ample, as we shall see below, many systems will go into !The general nonlinear term is assumed to have a Taylor
spontaneous oscillation when the proportional gain Kp is expansion about the solution, y! = 0." For example, the
too high. Occasionally, feedback is used to stabilize an unstable mode !indexed by i" may behave either as
initially unstable system. A topical example is the
Stealth Fighter.9 The aircraft’s body is made of flat sur- ẏi = 1iyi − ay2i or ẏi = 1iyi − by3i . !4.4"
faces assembled into an overall polygonal hull. The flat
surfaces deflect radar beams away from their source but In Eq. !4.4", 1i is a “control parameter,” i.e., one that
make the aircraft unstable. Using active control, one can may be controlled by the experimenter. The parameters
stabilize the flight motion. A more prosaic example is a and b are assumed to remain constant. The first rela-
the problem of balancing a broom upside down in the tion in Eq. !4.4" describes a “transcritical bifurcation”
palm of your hand. and the second a “pitchfork bifurcation” !Strogatz,
Before discussing stability in feedback systems, we 1994". !If there is a symmetry y → −y, one has a pitchfork
briefly review some notions of stability in general in dy- bifurcation; otherwise, the transcritical bifurcation is the
namical systems in Sec. IV.A. In Sec. IV.B, we apply generic description of two solutions that exchange sta-
those ideas to feedback systems. bilities." In both cases, the linearly unstable mode satu-
rates with a finite amplitude, in a way determined by the
lowest-order nonlinear term.
A. Stability in dynamical systems If the eigenvalue 1i is complex and à real, then there
will be a second eigenvalue 1j = 1*i . The eigenvalues be-
Here, we review a few key ideas from stability theory
!Strogatz, 1994". We return to the time domain and write come unstable in pairs when Re 1i = 0. The system then
a simple nth-order equation in matrix form as begins to oscillate with angular frequency Im 1i. The in-
stability is known as a Hopf bifurcation.
One situation that is seldom discussed in introductory
y!˙ = Ãy! . !4.1" feedback texts but is familiar to physicists studying non-
linear dynamics is the distinction between subcritical
Assuming à to be diagonalizable, one can write the so- and supercritical bifurcations !analogous to first- and
lution as second-order phase transitions". Supercritical bifurca-
tions refer to the case described above, where the sys-
y! !t" = eÃty! !0" = R̃eD̃tR̃−1y! !0", !4.2" tem is stable until 1 = 0, where the mode spontaneously
begins to grow. But if the lowest nonlinear term is posi-
where à = R̃eD̃tR̃−1 and tive, it will reinforce the instability rather than saturating
it. Then the lowest-order negative nonlinear term will
saturate the instability. For example, a subcritical pitch-
9
The instability of planes such as the F-117 fighter !“Night- fork bifurcation would be described by
hawk”" is fairly obvious just looking at pictures of them. I have
seen anecdotal mention of this but no serious discussion of ẏi = 1iyi + by3i − cy5i , !4.5"
how active control stabilizes its flight. See, for example, http://
www.danshistory.com/f117.shtml and a subcritical Hopf bifurcation by
i# 2+#2
FIG. 11. !Color in online edition" Two scenarios for a pitch- !1 − r2#2" − + 1 = 0. !4.7"
Ki Ki
fork bifurcation, showing steady-state solution y as a function
of the control parameter 1: !a" supercritical; !b" subcritical,
where the arrows show the maximum observable hysteresis.
Both real and imaginary parts of this complex equation
must vanish. The imaginary part implies #* = 1 / +, while
ẏi = 1iyi + b+yi+2yi − c+yi+4yi . !4.6" the real part implies K*i = 2 / + ; i.e., when Ki is increased
to the critical value 2 / +, the system goes unstable and
In both these cases, the yi = 0 solution will be metastable
begins oscillating at # = 1 / +.
when Re 1i is slightly positive. At some finite value of Bode plots are useful tools in seeing whether a system
Re 1i, the system will jump to a finite-amplitude solu- will go unstable. Figure 12 shows the Bode magnitude
tion. On reducing Re 1i, the eigenvalue will be slightly and phase plots for the second-order system described in
negative before the system spontaneously drops down to the paragraph above, with + = 1. The magnitude response
the yi = 0 solution. For a Hopf bifurcation, the discussion is plotted for two different values of the integral gain Ki.
is the same, except that y is now the amplitude of the Note how changing a multiplicative gain simply shifts
spontaneous oscillation. Thus subcritical instabilities are the response curve up on the log-log plot. In this simple
associated with hysteresis with respect to control- case, the phase response is independent of Ki, which
parameter variations. In the context of control theory, a would not in general be true. For Ki = 1, we also show
system with a subcritical bifurcation might suddenly go explicitly the gain margin, defined as the factor by which
unstable when the gain was increased beyond a critical the gain must be increased to have instability, at the fre-
value. One would then have to lower the gain a finite quency where the phase lag is 180°. !We assume an
amount below this value to restabilize the system. Super- open-loop-stable system." Similarly, the phase margin is
critical and subcritical pitchfork bifurcations are illus- defined as the phase lag at the frequency where the mag-
trated in Fig. 11. nitude response is unity. In Fig. 12, the gain margin is
about a factor of 2, and the phase margin is about 20°. A
good rule of thumb is that one usually wants to limit the
B. Stability ideas in feedback systems
gain so that the gain margin is at least a factor of 2 and
As mentioned before, in a closed-loop system, varying the phase margin at least 45°. For small gain margins,
parameters such as Kp or Ki in the feedback law will perturbations decay slowly. Similarly, a 90° phase margin
continuously vary the system’s eigenvalues, raising the would correspond to critically damped response to per-
possibility that the system will become unstable. As turbations, and smaller phase margins give under-
usual, one can also look at the situation in frequency damped dynamics. The curve drawn for Ki = 2 shows that
space. If one has a closed-loop transfer function T the system is at the threshold of instability !transfer-
= KG / !1 + KG", one can see that if KG ever equals −1, function response= 1, phase lag= 180°".
the response will be infinite. This situation occurs at the
bifurcation points discussed above. !Exponential growth 10
Here, we are implicitly assuming the !common" scenario
implies an infinite steady-state amplitude for the linear where instability arises dues to time delay !two poles with con-
system." In other words, in a feedback system, instability jugate imaginary parts cross the imaginary s axis together".
will occur when +KG+ = 1 with a simultaneous phase lag Another case arises when a real pole crosses the imaginary s
of 180°. axis from the left. Still other cases arise if the system is intrin-
The need for a phase lag of 180° implies that the sically unstable: then, increasing a feedback gain can actually
open-loop dynamical system !combination of the origi- stabilize the system. For example, the unstable inverted pen-
nal system and feedback compensation" must be at least dulum discussed below in Sec. IV.D can be stabilized by in-
third order to have an instability. Since the transfer func- creasing the gains of proportional and derivative feedback
terms !Sontag, 1998". These cases can all be diagnosed and
tion of an nth-order system has terms in the denomina-
classified using the Nyquist criterion, which involves examining
tor of order sn, the frequency response will asymptoti- a polar plot of #+K!s"G!s"+s=i#$ for 0 % # % ) in the complex
cally be !i#"−n, which implies a phase lag of n* / 2. !This plane !Dutton et al., 1997"; see Sec. V.E for examples of such a
is just from the i−n term." In other words, as Figs. 3 and polar plot. In any case, it is usually clear which case is relevant
4 show, a first-order system lags 90° at high frequencies, in a given problem.
one wants integral control, the effective system is third takes heat to propagate a distance L in the bar. In con-
order. A more subtle point is that the ODE models of trolling the temperature of such a system, the controller
dynamical systems that we have been discussing are al- gain will thus have to be low enough that at #*, the
most always reductions from continuum equations. Such overall transfer function of system #Eq. !4.14" multiplied
a system will be described by partial differential equa- by the controller gain$ is less than 1. The calculation is
tions that have an infinity of normal modes. The larger similar to the one describing digitization lag.
the system, the more closely spaced in frequency are the The overall lesson is that response lags are generic to
modes. Thus if one measures G!i#" at higher frequen- almost any physical system. They end up limiting the
cies, we expect to see a series of first- and second-order feedback gain that can be applied before spontaneous
frequency responses, each one of which adds its 90° or oscillation sets in. In many cases, a bit of care in the
180° phase lag to G. experimental design can minimize the pernicious effects
In the limit of large systems and/or high frequencies, of such lags. For example, in the temperature-control
the modes will be approximately continuous, and we will example discussed above, an immediate lesson is that
be guaranteed that the system will lag by 180° for high putting the thermometer close to the heater will allow
enough frequencies. We can consider an example #sim- larger gains and tighter control !Forgan, 1974". Of
plified from a discussion in Forgan !1974"$ of controlling course, the thermometer should also be near the part of
the temperature of a thermometer embedded in a long the sample that matters, leading to a possible design
!semi-infinite" bar of metal, with a heater at one end; see tradeoff.
Fig. 15. The temperature field in the bar obeys, in a
one-dimensional approximation,
$ 2T $ T
D = , !4.11" D. Nonminimum-phase systems
$x2 $t
where D = 1 / 6Cp is the thermal diffusivity of the bar, Delays, then, are one generic source of instability. An-
with 1 the thermal conductivity, 6 the density, and Cp the other arises when the system transfer function G!s" has a
heat capacity at constant pressure. The boundary condi- zero in the right-hand plane. To understand what hap-
tions are pens, consider the transfer functions G1!s" = 1 and
G2!s" = !1 − s" / !1 + s". Both have unit amplitude response
T!x → )" = T) , !4.12" for all frequencies #they are “all-pass” transfer functions
1 1
!Doyle et al., 1992"$, but G1 has no phase lag while G2
$T has a phase lag that tends to 180° at high frequencies.
−1 = J0e−i#t , !4.13"
$x !0,t" Thus all of our statements about leads and lags and
about gain and phase margins must be revised when
with J0 the magnitude of the power input.11 In order to there are zeros in the right-hand side of the s plane.
construct the transfer function, we assume a sinusoidal Such a transfer function describes a nonminimum-phase
power input at frequency #. The solution to Eqs. system in control-theory jargon.12 In addition to arising
!4.11"–!4.13" is given by from delays, they can arise, for example, in controlling
T!x,t" − T) 1 i!kx+*/4" −kx the position of floppy structures—e.g., the tip of a fishing
= e e , !4.14" rod or, to use a time-honored control example, the bob
J 0e −i#t % 21k of a pendulum that is attached to a movable support.
where k = %# / 2D and where we have written the solu- !Think of balancing a vertical stick in your hand."
tion in the form of a transfer function, evaluated at s
= i#. If, as shown in Fig. 15, the thermometer is placed at 12
The term “nonminimum-phase” refers to Bode’s gain-phase
a distance L from the heater, there will be a phase lag relationship, which states that for any transfer function L!s"
given by kL + * / 4. This will equal 180° at a critical fre- with no zeros or poles in the right-hand plane, if L!0" & 0, the
quency phase lag 7 is given by
9 D
#* = *2 2 .
8 L
!4.15" 7 ! # 0" =
1
*
)
−)
)
d
d8
2 & '3
ln+L!i#"+ln coth
+8+
2
d8 , !4.16"
In other words, there will always be a mode whose with 8 = ln!# / #0" the normalized frequency. #See Özbay !2000",
frequency #* gives the critical phase lag. Here, D / L2 but note the misprint.$ The phase at #0 thus depends on +L+,
may be thought of as the characteristic lag, or time it over all frequencies. However, if +L+ 4 #−n over at least a de-
cade centered on #0, then Eq. !4.16" is well approximated by
7!#0" . −n* / 2. Transfer functions that have more than this
11
One usually imposes an ac voltage or current across a resis- minimum lag are nonminimum phase systems. See Franklin et
tive heating element. Because power is the square of either of al. !1998", pp. 254–256. One can show that a general transfer
these, the heater will inject a dc plus 2# signal in the experi- function G!s" can always be written as the product G1!s"G2!s",
ment. One can compensate for this nonlinearity by taking, in where G1!s" is minimum phase and G2!s" is an all-pass filter
software, the square root of the control signal before sending it with unit amplitude response; see Doyle et al. !1992", Sec.
to the heater. VI.B.
+s . − 1 + & k-+
2+
± '% k-+
+
& '
+
k-+
2+
2
, !4.24"
/ 0
cult to control is that it has rank 1 at zero frequency. !The
1 1 determinant vanishes at dc." The system is thus nearly
1 + + 1s 1 + + 2s uncontrollable—in the technical and intuitive senses of the
G̃!s" = . !4.22" word—at low frequencies.
1 1 15
The eigenvalue is not the right quantity to characterize the
1 + + 2s 1 + + 1s gain, in general, for two reasons: First, whenever the number
If one prefers a physical picture, think about a shower of inputs differs from the number of outputs, the matrix S̃ is
where pipes carrying hot and cold water are mixed to not square. Second, eigenvalues give only the gain along cer-
tain directions and can miss subtle types of “cross gains.” For
form a single stream. The shower output is characterized
example, consider the matrix
by an overall flow and temperature. Assuming there are
separate hot and cold shower knobs, one has two control
inputs and two outputs. If the +’s in Eq. !4.22" were all
S̃ = & 1
100 1
0
' , !4.25"
different, they would reflect the individual time con- which implies that the unit input !1 0"T gives rise to an output
stants to change flow and temperature. The rather unre- !1 100"T, which has a gain of roughly 100, even though the two
alistic choice of +’s in Eq. !4.22" simplifies the algebra. eigenvalues of S̃ are unity. The appropriate generalization of
Let us now try to regulate the shower’s temperature the notion of eigenvalue is that of a singular value. The
and flow by adjusting the hot and cold flows with pro-
singular-value decomposition of an m ! n matrix S̃ is given by
portional gains k that are assumed identical for each.
S̃ = Ũ9̃ṼT, where Ũ is an m ! m unitary matrix, ṼT is the trans-
The controller matrix K̃!s" is then k1̃, with 1̃ the 2 ! 2 pose of an n ! n unitary matrix, and where 9̃ is an m ! n matrix
identity matrix. The loop stability is determined by the that contains k = min!,m , n-" non-negative singular values :i
matrix analog of Eq. !3.2", S̃ = !1̃ + G̃K̃"−1, where the !Skogestad and Postlethwaite, 1996". For example, one can
closed-loop system is unstable when one of the eigenval- show that
ues of S̃ has positive real part for any value of s = i#, sup
5S̃w! 52
generalizing the ideas of Sec. IV.B. !Recall that the or- :max!S̃" = w
! , !4.26"
5w! 52
der of matrix multiplication matters." After some ma-
where 5 · 52 denotes the Euclidean norm. Equation !4.26" states
nipulation, one can easily show that the eigenvalues of S̃ that :max is the largest gain for all inputs w! . This quantity is
are given by
then the largest singular value of S̃. One can also show that the
1 :’s are the square roots of the eigenvalues of S̃TS̃. In the ex-
1±!s" = . !4.23" ample of Eq. !4.25", one finds the singular values : are .1 and
1 + k#1/!1 + +1s" ± 1/!1 + +2s"$
.100, showing that the singular values capture better the
The negative root 1− has poles at “size” of S̃ than do the eigenvalues.
quency #max. At a minimum, one should have #max system modes discussed above. The singular values of
& #0, the desired feedback bandwidth. Ideally, one the matrix describing controllability rank the effect of an
should keep enough terms that the gain !relative to the input on a particular system mode, while the singular
zero-frequency, or dc, gain" at #max is /1. Note that the values of the matrix describing observability rank the
usual structure of transfer functions means that higher- effect of the dynamics on a particular output mode. The
order terms are multiplied, not added, onto lower-order product of the relevant two matrices !“gramians”" then
terms, i.e., that we truncate a system gives the strength of “input-output” relationships of
each mode of the system. One can show that it is pos-
)
s−z sible to use the freedom in defining a state-space repre-
G!s" = 8
i,j=1 s − p
i
j
!5.1" sentation of a system to make these gramian matrices
equal—the “balanced representation.” Ordering these
to balanced modes by their singular values and retaining
only the largest ones gives a systematic approach to
N
s − z! model reduction !Skogestad and Postlethwaite, 1996".
Gt!s" = 8 s − p
i
!
, !5.2" The kind of systematic model reduction discussed
i,j=1 j above has recently attracted the attention of statistical
physicists, who are faced with a similar task when
where the last !Nth" pole or zero occurs at a frequency
“coarse-graining” dynamics in a thermodynamic system.
.#max. Note that the zeros and poles of the reduced
For example, if one considers a small number of objects
system—found by refitting the data to the truncated
in contact with many other elements of a “heat bath,”
form for G!s"—will in general differ from the “exact then a common strategy is to integrate out the degrees
values” of lower-order zeros/poles of the full system.18 of freedom corresponding to the bath. One derives thus
While the above method seems straightforward, it is a reduced dynamics for the remaining degrees of free-
not always the best thing to do. For example, if there dom. A similar task arises when one uses the renormal-
happens to be a high-frequency mode !or modes" with a ization group !Goldenfeld, 1992". Coarse graining is thus
large amplitude, neglecting it may not wise. A more a kind of model reduction. The interest in control strat-
subtle situation is that there may be so many high- egies is that they give a systematic way of handling a
frequency modes that even though the amplitude of any general system, while the usual coarse-graining strategy
one is small, they may have a collective effect. To deal is more ad hoc and fails, for example, when naively ap-
with such cases, one strategy is to try to order the sys- plied to spatially inhomogeneous systems. A recent pa-
tem’s modes not by frequency but by some measure of per by Reynolds !2003" explores these issues.
their “importance.” One definition of importance is in
terms of the Hankel singular values of the system trans-
fer function G!s". The detailed discussion is beyond the 4. Revisiting the system
scope of this tutorial, but the rough idea is to appeal to
the notions of controllability and observability of the We close by noting that determination of the transfer
function has two meanings: Whatever the system dy-
namics are, the experimentalist should measure them;
18
Note that my definition of truncation differs from that used but the experimentalist also has, in most cases, the
in standard control texts, such as Skogestad and Postle- power to influence greatly, or determine, the system it-
thwaite !1996", Chap. 11. There, the authors first define a trun- self. If a system is hard to control, think about ways of
cation where the poles and zeros are held to be the same. If changing it to make control easier. !This is a basic differ-
one then looks at the difference between the exact transfer ence in philosophy from most engineering texts, where
function and its truncated approximation, one sees that they the system—“plant” in their jargon—is usually a given."
differ most at low frequencies and the difference goes to zero
Often, the experimentalist simultaneously designs both
only at infinite frequencies. But the whole point of doing a
truncation is almost always to capture the low-frequency dy- the physical !mechanical, electrical" aspects and the con-
namics while disregarding high frequencies! Thus this kind of trol aspects of an experiment. We have already seen two
truncation seems unlikely to be useful in practice. Skogestad examples: in temperature control, where minimizing the
and Postlethwaite !1996" then go on to introduce “residualiza- physical separation between heater and thermometer
tion” !physicists would use the terms “slaving” or adiabatic makes a system much easier to control and allows higher
elimination of fast variables", where, in the time domain, the performance; and in balancing the rod, where one
fast mode derivatives are set to zero and the steady states of should look at the top of the rod, not the bottom. More
the fast modes are solved. These instantaneous or adiabatic generally, one should minimize the delay and/or separa-
steady states are then substituted wherever fast variables ap- tion between the actuator and its sensor. Another area
pear in the equations for the remaining slow modes. The equa-
in which good design plays a role is in the separation of
tions for the slow modes then form a smaller, closed set. If the
modes. As we have discussed above, closely spaced
modes are well behaved—i.e., well separated in frequency
from each other—then the procedure described in the text, modes are difficult to control, and it usually pays to
fitting to a form up to some maximum frequency, will give make different modes as widely spaced in frequency as
essentially equivalent results. Both methods share the property possible. For example, if one is trying to isolate vibra-
of agreeing well at low frequencies up to some limit, above tions, the apparatus should be as small and cube shaped,
which the approximation begins to fail. to maximize the frequency separation between the soft
isolating spring and the lumped experimental mass and approach, this means choosing the dynamics K!s". This
the internal modes of the experimental apparatus. In the task is commonly broken down into two parts: choosing
next section, we shall see that systems that are up to a general form for K!s" and choosing, or “tuning,” the
second order over the frequency range of interest may free parameters.
be perfectly controlled by the elementary “PID” law
commonly found in commercial controllers but higher-
order systems in general cannot be adequately con- 1. PID controllers
trolled by such laws.
Another example of how the design of an experimen- Probably the most common form for K!s" is the PID
tal system determines the quality of control comes from controller, which is a combination of proportional, inte-
my own laboratory, where we routinely regulate tem- gral, and differential control:
perature to 50 ;K rms near room temperature, i.e., to
fractional variations of 2 ! 10−7 !Metzger, 2002; Yethiraj Ki
K!s" = Kp + + Kds, !5.3"
et al., 2002". In order to obtain such performance, the s
simple proportional-integral !PI" law described above
sufficed. #A more sophisticated control algorithm would where Kp, Ki, and Kd are parameters that would be
have further improved the control, but we did not need tuned for a particular application. We have already dis-
a better performance. As an example, Barune et al. cussed the general motivations for proportional and in-
!1995" use loop-shaping techniques—see Sec. V.B.2 tegral control. The intuitive justification for derivative
below—to achieve 20-;K control of the temperature of control is that if one sees the system moving at high
a laser diode.$ The crucial steps all involved the physical “velocity,” one knows that the system state will be
and electrical design of the system itself. One important changing rapidly. One can thus speed the feedback re-
idea is to use a bridge circuit so that the error signal is sponse greatly by anticipating this state excursion and
centered on zero rather than about a finite level and taking counteraction immediately. For example, in con-
may thereafter be amplified. We minimized sensor noise trolling a temperature, if the temperature starts falling,
by using a watch battery to power the bridge.19 The one can increase the heat even before the objects has
overall performance turned out to be set by the tem- cooled, in order to counteract the presumably large per-
perature variations in the nonsensor resistors in the turbation that has occurred. The word “presumably”
bridge. The best place to have put these would have highlights a difficulty of derivative control. One infers a
been next to the sensor resistor itself, as that is where rapid temperature change by measuring the derivative
the temperature is most stable. In our case, that was of the system state. If the sensor is noisy, random fluc-
inconvenient, and we put them inside a thick styrofoam tuations can lead to large spurious rates of change and
box. Slow variations in the bridge temperature then to inappropriate controller response. Thus many experi-
show up as setpoint drifts that cannot be corrected by mentalists try derivative control, find out that it makes
the simple feedback loop; however, the temperature was the system noisier, and then give up. Since there are
very stable on the 1-min time scale of our particular many benefits to derivative control and since spurious
experiment. To stabilize over longer times, we could response to noise can be avoided, this is a shame, and
have added a second feedback loop to regulate the tem- hopefully this article will motivate experimentalists to
perature of the external components of the bridge. The work around the potential problems.
time scales of the two controllers should be well sepa- In order to better understand derivative control, we
rated to decouple their dynamics. The idea of using can look at it in frequency space !Kds". The linear s
nested stages of dynamically decoupled feedback loops dependence means that the response increases with fre-
can lead to outstanding performance. Perhaps the ulti- quency, explaining why high-frequency sensor noise can
mate example is a four-stage controller developed by the have such a great effect. One obvious response is to limit
group of Lipa, which achieved stabilities on the order of the action of the derivative term by adding one or more
10−10 K at low temperatures !2 K" !Day et al. 1997". low-pass elements the control law, which becomes
Kp + Ki/s + Kds
K!s" = , !5.4"
B. Choosing the controller !1 + s/#0"n
Having identified the system !and, perhaps, having al- where we have added n low-pass filters with cutoff fre-
ready made the system easier to control", we need to quencies all at #0. Indeed, since no actuator can respond
choose the control algorithm. In the frequency-domain with large amplitude at arbitrarily large frequencies,
some kind of low-pass filter will be present at high fre-
19 quencies, whether added deliberately or not. As long as
One would be tempted to use a lock-in amplifier to supply
an ac voltage to the bridge. The lock-in technique, by centering #0 is higher than the feedback bandwidth, it will have
the bandwidth about a finite carrier frequency, can lower sen- little effect on the system’s dynamics, while limiting the
sor noise by moving the passband to a high enough frequency effects of sensor noise. A more sophisticated way to
that 1 / f noise is unimportant. In our own case, this was not the minimize sensor-noise feedthrough, the Kalman filter,
limiting factor for performance. will be discussed in Sec. V.D below.
2. Loop shaping
Many experimentalists limit their attention to PID
controllers !and often just PI controllers if they have no
luck with the derivative term". While PID controllers
can give good results and have the advantage that each
term has an intuitive justification, they are by no means
the only possibility. Indeed, the frequency-domain ap-
proach we have developed suggests that one can think of
choosing K!s" to “sculpt” or “shape” the closed-loop re-
sponse T!s". For example, given a system G!s", one can
invert Eq. !3.2" and write
T!s"
K!s" = G−1!s". !5.5"
1 − T!s"
Klag!i#" =
a!1 + i#/#0"
1 + ai#/#0
, !5.6" J= )0
)
V#x!t",u!t"$dt = )
0
)
#e2!t"Q + u2!t"R$dt, !5.8"
where a is a constant !typically of order 10" and #0 is the where the general function V!x , u" commonly has a qua-
scale frequency. Figure 19 shows the Bode response of dratic form and where Q and R are positive constants
Eq. !5.6". The gain goes from a at low frequencies !# that balance the relative “costs” of errors e!t" and con-
/ #0" to 1 at high frequencies !# . #0". The phase re- trol efforts u!t". Large Q and small R penalize control
sponse has a transient lag in the crossover frequency errors with little regard to the control effort, while small
region but goes to zero at both low and high frequencies. Q and large R penalize control effort with little regard
One can think of a lag compensator as either an ap- for control error. Equation !5.8" has an obvious vector-
proximation to integral control !a boost of a at zero fre- matrix generalization to higher-order systems. Implicit
quency rather than infinite gain" but without the often in Eq. !5.8" is the choice of a control signal r!t" and a
disturbance d!t". For example, one often assumes a step minimizing the functional L = *)0 Ldt, where L!x , ẋ , u , 1"
function input r!t" = (!t", with d = 0. One can, alterna- = 10V!x , u" + 1!f − ẋ" by free variation of x, u, and the
tively, keep r constant and add a step disturbance for Lagrange multiplier 1.20 For nth-order systems, there
d!t". are n Lagrange multiplier functions 1i!t". Setting the
As a simple example, we consider the one- variation of J with respect to x, u, and 1 equal to zero
dimensional system ẋ = −<x!t" + u!t" with proportional leads to three sets of Euler-Lagrange equations:
control u!t" = −Kpe!t" and reference signal r!t" = 0
!Skogestad and Postlethwaite, 1996". The proportional d $L $L
− = 0, !5.11a"
control gives a motion x!t" = x0e−!<+Kp"t, which, when in- dt $ẋ $x
serted into the cost function, Eq. !5.8", gives !with Q
= 1" $L
= 0, !5.11b"
$u
x20
J!Kp" = !1 + RKp2 " . !5.9"
2!< + Kp" $L
= 0. !5.11c"
Minimizing J with respect to Kp gives an “optimal” $1
Equations !5.11a" and !5.11b" give the equations obeyed
Kp* = − < + %<2 + 1/R. !5.10"
by the performance index V; Eq. !5.11c" gives the equa-
The decay rate of the optimal system is <! = %<2 + 1 / R. tions of motion.
From this simple example, one can see the good and bad For reasons to be discussed below, the Hamiltonian
points of optimal control. First of all, optimal control formulation is usually preferred. There, one defines a
does not eliminate the problem of tuning a parameter. Hamiltonian H = L + 1ẋ = V + 1f. The Euler-Lagrange
Rather, in this case, it has replaced the problem of equations can then be transformed into the control-
choosing the proportional gain Kp with that of choosing theory version of Hamilton’s equations:
the weight R. What one gains, however, is a way of mak- $H
ing any tradeoffs in the design process more transparent. ẋ = , !5.12a"
$1
Here, for example, one is balancing the desire to have
small tracking error e!t" with that of wanting minimal
$H
control effort u!t". The coefficient R expresses that 1̇ = − , !5.12b"
tradeoff explicitly. For example, in the expression for the $x
optimal Kp in Eq. !5.10", a small R !“cheap control”"
leads to a large Kp while a large R !“expensive control”" $H
= 0. !5.12c"
leads to a small Kp. In such a trivial example, the result $u
could have been easily foreseen and there would be little
Equations !5.12a" and !5.12b" describe the evolution of
point in going through this exercise. When the system is
the “co-states” 1!t", which are the equivalent of the con-
more complex, one often has more intuition into how to
jugate momenta in the classical-mechanics formulation.
set the matrices Q̃ and R̃ of the generalization of Eq. Note that, in the general case, the state and co-state
!5.8" than to tune the parameters. At any rate, do not vectors have n components while the control vector u
confuse “optimal” with “good,” for a poor choice of has r components. In the simple example discussed
weights will lead to a poor controller. above, one has H = Qx2 + Ru2 + 1!−<x + u".
The above discussion gives short shrift to a rich and Optimal-control problems often lead to two types of
beautiful branch of applied mathematics. One can be generalizations that are less likely to be familiar to
more ambitious and ask for more than the optimal set of physicists: In classical mechanics, one typically assumes
parameters, given a previously chosen control law. In- that the starting and ending states and times of the dy-
deed, why not look for the best of all possible control namical system are fixed. Only variations satisfying
laws? Formulated this way, there is a close connection these boundary conditions are considered. Variational
between optimal control and variational methods. In the problems in control theory are less restrictive. For ex-
simplest cases, these variational methods are just those ample, imagine that one wants to move a dynamical sys-
familiar from classical mechanics. #For an overview of tem from an initial state x0 to a final state x1 as fast as
the history of optimal control and its relations with clas-
possible #given the dynamical equations ẋ = f!x , u"$. One
sical problems, see Sussmann and Willems !1997".$ For
can formulate this as a problem whose goal is to mini-
example, we can view the minimization of the perfor-
mize the unit performance index V = 1 over times from t0
mance index J in Eq. !5.8" as a problem belonging to the
!which can be set to be 0 in an autonomous system" to t1,
calculus of variations, where one needs to find the opti-
where t1 is to be determined. A recent book by Naidu
mal control u!t" that minimizes the functional J subject
to the constraint that the equations of motion ẋ = f!x , u"
are obeyed. #In the example above, there is only one 20
Usually, 10 = 1; however, an “abnormal” case with 10 = 0 can
equation of motion, and f!x , u" = −<x + u.$ One can solve arise when both the constraint and its derivative vanish simul-
this problem by the method of Lagrange multipliers, taneously !Sontag, 1998".
&'&
ẋ1
ẋ2
=
0
− #21 − "#1
1
'& ' & '
x1
x2
+
#21/#0
#21
u!t − +", !5.18"
nTs, for n = 0,1, . . .. As above, the sampled signal is then
)
ẋ .
xn+1 − xn
, !5.19" L#fs!t"$ = 9 f!nTs"e−s!nTs" . !5.22"
Ts n=0
for each component of x !or use higher-order approxi- If we introduce z ( esTs and define fn ( f!nTs", then Eq.
mations for the higher derivatives of y". Usually, such a !5.22" leads to the “z transform,” defined by
& '
terpretation of delay by Ts, we may rewrite Eq. !5.29" as
2 1 − z−1 a discrete difference relation !known as an “infinite im-
s→ . !5.25"
Ts 1 + z−1 pulse response,” or IIR filter in the signal-processing lit-
erature" !Oppenheim et al., 1992",
Equation !5.25" is equivalent to using the trapezoidal
rule for integrating forward the system dynamics, while un = A1un−1 + A2un−2 + ¯ + B0en + B1en−1 + B2en−2
Eq. !5.24" corresponds to using the rectangular rule +¯ !5.30"
!Lewis, 1992, Chap. 5". One advantage of Tustin’s trans-
formation is that it, too, maps Re!s" % 0 to the interior of which generalizes the result of Eq. !5.20".
the unit circle, implying that if the Laplace transform of One minor pitfall to avoid is that discrete systems can
a continuous function is stable #Re!s % 0"$, then its z introduce “ringing poles.” For example, consider the
transform will also be stable. !Iterating a linear function simple difference equation
with magnitude greater than one leads to instability." yn+1 = !1 − 1"yn . !5.31"
The direct, first-order substitution, Eq. !5.24", does not
have this property. Analogously, in the numerical inte- This, obviously, is the discrete analog of a first-order sys-
gration of differential equations, too coarse a time step tem ẏ!t" = −1y!t" and gives similar stable behavior for 0
can lead to numerical instability if one uses an explicit % 1 % 1. Now consider the range 1 % 1 % 2, where the sys-
representation of time derivatives #Eq. !5.24"$. If, on the tem continues to be stable but oscillates violently with
of families, with varying logic-gate densities and varying that the eigenvalues of Ã! have magnitude less than one.
ease of programmability. One popular family is the field- Other aspects carry over as well !e.g., Laplace transform
programmable gate array !FPGA". In physics, these to z transform". But when faced with a choice between
have been used to make trigger units in high-energy ex- reading up on fancy direct-digital-design techniques and
periments that must deal with a tremendous amount of buying a better digitizer, buy the board. The time you
data in very short times. For example, in the D0 detector save will almost certainly be worth more than the price
at Fermilab, a set of boards containing 100 field- of the more expensive hardware.
programmable gate array chips evaluates
.107 events/ sec, identifying about 1000 events/ sec as
potentially “interesting.” Subsequent trigger circuits 1. Case study: Vibration isolation of an atom
then reduce the event rate to about 50 Hz !Borcherding interferometer
et al., 1999". Until recently, the “learning curve” in pro- We mention briefly a nice example of an implementa-
gramming such devices was steep enough that they tion of a feedback loop that illustrates most of the topics
made sense only in projects large enough to justify the discussed in this section. The application is the control
services of a specialist programmer. !The language and system of a vibration isolation stage for an atom inter-
logic of field-programmable gate array devices was dif- ferometer !Hensley et al., 1999". The goal is to decouple
ferent enough from ordinary programs to require exten- the instrument from random low-frequency vibrations
sive training for such devices." Recently, field- due to machinery, micro-seismic disturbances, etc. The
programmable gate arrays that are programmable in conventional passive strategy is to mount the instrument
ordinary, high-level languages !e.g., LABVIEW" have be- on a mechanical spring system !often an air spring, but
come available.24 Field-programmable gate arrays are so here a conventional wire spring". The idea is that if the
powerful that they can implement almost any conven- driving vibrations have frequencies much higher than
tional digital or !low-power" analog electronic circuit, the resonant frequency of the mass-spring system, their
and many commercial devices are now based on such amplitude will be damped. The lower the frequency of
circuits. Despite the D0 example above, the full impact the external vibrations, the lower the required resonant
of having “hardware you can program” has not yet been frequency of the mass-spring system. But practical mass-
appreciated by the physics community, even as commer- spring systems have resonant frequencies of roughly
cial applications seem to multiply exponentially. 1 Hz for vertical vibrations. Lowering the resonant fre-
In looking over the above “recipes” for making a digi- quency requires softer springs, which stretch large dis-
tal controller, one might be tempted to think that it is tances under the influence of gravity. In order to lower
simpler to bypass the first two steps and work in the the resonant frequency still further, one can use feed-
discrete domain from the beginning. For example, any back to alter the dynamics. !Note that in the first case
discrete, linear control law will have the form of Eq. study, we decreased characteristic times; here we will
!5.20", with different values for the A, B, and C coeffi- increase them."
cients !and with the coefficient in front of un not neces- The system has an equation of motion
sarily 1 and with perhaps other terms un−1 , . . . and
en−2 , . . . as well". Indeed, our approach to digital feed- ÿ + 2$0#0!ẏ − ẏg" + #20!y − yg" = u!t", !5.33"
back is known as “design by emulation,” and it works where #20 = k / m is the natural resonance frequency of the
well as long as the sampling rate is high enough relative
undamped system, yg!t" and ẏg!t" are the position and
to system dynamics. Emulation has the virtue that it uses
velocity of the ground, u!t" is the actuator signal, and $0
the intuitive loop-shaping ideas discussed above. Still,
is the dimensionless damping, as in Eq. !2.5". Indeed
direct digital design has fewer steps and potentially
Eqs. !5.33" and !2.5" differ only in that the damping is
higher performance. We do not have space to discuss
now proportional to the difference between the mass
direct methods, except to say that virtually every
and ground velocity, and similarly for the restoring
continuous-time technique has a discrete-time analog.
force.
For example, the discrete analog of the linear system
The control-loop design uses an accelerometer to
given by Eq. !2.2" is
sense unwanted vibrations and applies a PI control
#u!t" = −Kpÿ!t" − 2#0Kiẏ!t"$ to lower the effective reso-
xn+1 = Ã!xn + B̃!un , nant frequency of the mass-spring design. Notice that
!5.32" measuring the acceleration is better than measuring the
yn+1 = C̃xn+1 , position for noise rejection. !You integrate once, as op-
posed to differentiating twice." The closed-loop transfer
function is
with Ã! = eÃTs and B̃! = Ã−1!Ã! − Ĩ"B̃. In the continuous-
time case, the eigenvalues of à needed to have negative 2s$1#1 + #21
T!s" = , !5.34"
real part for stability. Here, the analogous condition is s2 + 2s!$1 + K1!"#1 + #21
where #1 = #0 / %Kp + 1, $1 = $0 / %Kp + 1, and Ki!
24
LABVIEW, FPGA module, National Instruments = Ki / %Kp + 1. The closed-loop resonance frequency #1 is
!www.ni.com". set nearly 50 times lower than #0, and the damping,
between the environmental variable being controlled We begin by noting that there are three related quan-
!for example, a temperature" x and the sensor reading of tities that describe the internal state of a dynamical sys-
that variable !for example, a thermistor impedance" y. tem at time n + 1:
For simplicity, we have taken y = x, but in general the true state, xn+1 = 7xn + un + dn ,
units are different !°C vs ohms", and the sensor may
have its own dynamics. The discrete dynamics are evalu- predicted state, x̂n+1 = 7x̂n+n + un ,
ated at times nTs. The proportional feedback gain is Kp
!dimensionless". The environmental noise is given by the best estimate, x̂n+1+n+1 = !1 − K"x̂n+1 + Kyn+1 ,
stochastic variable dn and the sensor noise by 0n. Both of
these are Gaussian random variables, with zero mean and two quantities that describe the measurements:
and variances 6d27 ( d2 and 6027 ( 02, respectively.28 actual measurement, yn+1 = xn+1 + 0n+1 ,
In Fig. 24!a", we plot a time series xn for the open-
loop system without feedback !Kp = rn = 0". The time con- predicted measurement, ŷn+1 = x̂n+1 .
stant of the system low-pass filters the white noise due to
the environment !d2 = 0.01". In Fig. 24!b", we add feed- Keeping these straight is half the battle. In the last state-
vector item, the “Kalman gain” K is chosen to minimize
back !Kp = 10". The plot of xn shows how the propor-
tional feedback increases the cutoff frequency of the the overall uncertainty in x̂n+1 by blending, with proper
weight, the two pieces of information we have available:
low-pass filtering !#! . #0". The reduced standard devia-
the best estimate based on a knowledge of previous
tion illustrates how feedback controls the excursions of x measurements and of the system dynamics, and the ac-
from the setpoint.
tual measurement. The notation x̂n+1+n+1 means that the
In Fig. 24!c", we add sensor noise !02 = 0.1". Because estimate of xn+1 uses observations up to the time n + 1.
the control loop cannot distinguish between the desired
By contrast, the prediction x̂n+1 uses observations only
control signal rn and the undesired sensor noise 0n, the up to time n. Note that the true state x is forever un-
regulation is noticeably worse. This degradation in per- known and that usually there would be fewer measure-
formance would be aggravated were one to use deriva- ments y than state-vector components for x.
tive control, which amplifies high-frequency noise. Our goal is to derive the optimal value for the Kalman
Sensor noise can thus strongly degrade the perfor- gain K. To proceed, we write x̂n+1+n+1 as
mance of a control loop. One might try to limit the sen-
sor noise by reducing the sensor’s bandwidth—for ex- x̂n+1+n+1 = 7x̂n+n + un + K!yn+1 − ŷn+1", !5.37"
ample, by adding a low-pass filter to the sensor circuit.
in the standard, observer form #Eq. !3.21"$ using Eq.
This works as long as the sensor bandwidth remains
!5.36". Defining the estimation error en = xn − x̂n+n, we can
much greater than that of the feedback. If not, the extra
easily show using Eqs. !5.36" and !5.37" that
low-pass filter adds a phase lag that degrades the stabil-
ity margins, forcing one to lower the feedback gain. If en+1 = !1 − K"#7en + dn$ − K0n+1 . !5.38"
the sensor has significant noise within the feedback Note how Eq. !5.38", in the absence of the noise terms dn
bandwidth, straightforward frequency filtering will not and 0n+1, essentially reduces to our previous equation for
be sufficient. estimator error, Eq. !3.22".
About 40 years ago, Kalman !1960" suggested a clever The crucial step is then to choose the Kalman gain K
strategy that is a variant of the observer discussed in Sec. in the “best” way. We do this by minimizing the ex-
III.C.29 There, one used the dynamics to evolve an esti- pected value of the error variance at time n + 1. This
mate of the internal state forward in time, corrected by a error variance is just
term proportional to the difference between the actual 2
observed variable and the prediction. Here, the strategy 6en+1 7 = !1 − K"2#726en2 7 + d2$ + K202 . !5.39"
is similar, but because both the dynamics and the obser- In Eq. !5.39", the cross terms are zero because the dif-
vations are subject to noise !disturbances and sensor ferent noise and error signals are uncorrelated with each
noise", one wants to blend the two in a way that mini- other: 6d07 = 6ed7 = 6e07 = 0. Differentiating Eq. !5.39" with
mizes the overall uncertainty. Like the observer, though, respect to K, we minimize the error at time n + 1 by tak-
the key idea is to use the system dynamics to supple- ing
ment what one observes about the system, in order to
726en2 7 + d2
estimate the complete internal state of the dynamics. Kn+1 = . !5.40"
726en2 7 + d2 + 02
28 We put the index n + 1 on K because, in general, the
Caution: the variance of the noise terms dn and 0n increases
linearly in time. In other words, the associated standard devia-
minimization must be done at each time step, and the
tion, which would be used by a random-number generator in a dynamics, as well as the noise statistics, may have ex-
numerical simulation of Eq. !5.36", is proportional to %Ts. plicit time dependence. Equation !5.40" implies that if d2
29
Norbert Wiener introduced the idea that control systems is big and 02 small, Kn+1 . 1: one should trust the mea-
should be analyzed stochastically !Wiener, 1961". surement. Alternatively, if sensor noise dominates, one
should weight the dynamical predictions more heavily equation discussed above. The algebra becomes more
by taking Kn+1 . 0. Equation !5.40" gives the optimal complicated, but all the basic reasoning remains the
balance between the two terms. If the coefficients of the same !Dutton et al., 1997". #The generalization of Eq.
dynamical equations do not depend on time !n", there !5.39" leads to a “matrix Riccati equation,” which re-
will be a time-independent optimum mix K*. !Even with quires some care in its solution.$
stationary dynamics, poorly known initial conditions will The Kalman filter is similar to the Wiener filter #com-
tend to make the Kalman filter initially put more weight pare Eq. !5.40" with Eq. !13.3.6" of Press et al. !1993"$,
on the observations. Subsequently, the optimal K’s will which solves the problem of extracting a signal that is
decrease to K*." convoluted with an instrumental response and by sensor
In Fig. 24!d", the optimal K* is found to be .0.2 after noise. One difference, which is one of the most attrac-
a transient lasting 10–20 time steps, implying that after- tive features of the Kalman filter, is its recursive struc-
wards relatively little weight is placed on new measure- ture: data at time n + 1 are calculated in terms of data at
ments. This is not surprising, given that the standard time n. This means that one only has to keep track of the
deviation of the sensor noise is more than three times most recent data, rather than an entire measurement set.
that of the disturbances. Still, no matter how noisy the To understand this advantage better, consider the com-
sensor, it always pays to have K* & 0 in that one needs at putation of the average of n + 1 measurements,
least some contact with the measured system state. Note 1
n+1
E. Robust control
1. The internal model control parametrization
Control theory underwent something of an identity As a useful starting point, we introduce an alternative
crisis during the 1970s. Although the sophisticated meth- way of parametrizing control systems, known as “inter-
ods described in the above sections can give feedback nal model control” !IMC" !Morari and Zafirioiu, 1989;
schemes that work very well on paper, the results, in Goodwin et al., 2001". The basic idea is to explicitly in-
practice, were often disappointing. This led many prac-
clude a model G̃!s" of the system’s actual dynamics,
tical engineers !and physicists" to conclude that it was
G!s". This leads to the block diagram of Fig. 25. !For
not worth learning fancy techniques and that the tried-
simplicity, we omit external disturbances, sensor noise,
and-true PID controller was just about all one needed to
sensor dynamics, etc." Solving the block dynamics in Fig.
know. Indeed, the consensus is that the academic re-
25, one finds
search on control theory from 1960 to about 1980 had
“negligible effect” on industrial control practice #see the GQ
introduction to Morari and Zafirioiu !1989", as well as y!s" = r. !5.43"
1 + !G − G̃"Q
Leigh !2004"$.
The root of the problem was that the schemes pre- Notice that the feedback signal is v = !G − G̃"u. This
sented so far implicitly assume that the system itself and shows explicitly that if we were to have a perfect model
the types of inputs and disturbances it is subject to are !and no disturbances", there would be no need for feed-
well known. But in practice, models of a system have back. Feedback is required only because of our always
uncertainties—parameters differ from setup to setup, imperfect knowledge of the model system and its distur-
high-frequency modes may be neglected, components bances. As we discuss in Sec. IX below, there is a deep
age, and so on. Controllers that are optimized !e.g., in connection between feedback and information.
the sense of Sec. V.B.3" for one system may fail miser- There are a number of other advantages to the inter-
ably on the slightly different systems encountered in nal model control formulation. It is easy to relate the
practice. internal model control controller Q!s" to the “classic”
The challenge of finding effective control laws in the controller K!s" discussed previously:
face of such uncertainties led to two approaches, begin-
ning in earnest the late 1970s and early 1980s. One of Q
these, adaptive control, tries to reduce uncertainty by K!s" = . !5.44"
1 − G̃Q
learning more about the system while it is under control.
We discuss this briefly in Sec. VIII, below. The other Because the denominator in Eq. !5.43" is 1 + !G − G̃"Q,
approach is that of “robust control,” where one tries to the feedback system will become unstable only if Q
come up with control laws that take into account this
uncertainty. Here we give some of the basic ideas and = −1 / !G − G̃", which will be large when G . G̃. In other
flavor of this approach, following mainly Doyle et al. words, as long as the model G̃ is a reasonable approxi-
!1992", Skogestad and Postlethwaite !1996", and Özbay mation to the real system G!s", a stable controller Q!s"
!2000". In Sec. V.E.1, we show that in any control system will lead to a stable closed-loop system. #Here we as-
one implicitly assumes a model of the system being con- sume that G!s" itself is stable.$ This is in contrast to the
classic controller K!s", where stable controllers can systems G!s", defined by an arbitrary multiplicative un-
nonetheless destabilize the system. A more positive way certainty:
of saying this is to note that for the system G!s", the set
G!s" ( #1 + -!s"W!s"$G0!s", !5.45"
of all stable Q!s"’s generates all stable controllers K!s".
Another advantage of the IMC structure is that the where -!s" is an arbitrary transfer function that satisfies
above remarks about stability carry forward to cases 5-5) = 1 and where W!s" is a transfer function that gives
where the system G and model G̃ are nonlinear. If G the uncertainty limits. The ) subscript on the norm re-
. G̃, then their difference may be well approximated by fers to the “H)” norm, which is defined to be
sup
a linear function, even when G and G̃ are themselves
strongly nonlinear. 5-5) ( # +-!i#"+. !5.46"
Finally, in Eq. !5.43", we see that if G = G̃, and we set In words, the H) norm is computed by taking that maxi-
Q = 1 / G̃, then we will have perfect control #y!s" = r!s", mum of the magnitude of the function, in contrast to the
exactly$. This means that even when the system is more common Euclidean, or H2 norm, defined in fre-
known exactly, it is necessary to invert the model trans- quency space by
fer function G̃!s" in order to have perfect control. If G̃!s"
is nonminimum phase and has a zero in the right-hand
plane, say at s0, then the controller Q!s" will have a pole
5-52 ( 2 1
2*
)
−)
)
+-!i#"+2d# 3 1/2
. !5.47"
at s0 as well and will be impossible to implement at the The reasons for using the H) norm will become clear
corresponding frequency. !Intuitively, one needs infinite shortly. The best way to picture multiplicative uncer-
energy to move a system that has zero response." This tainty is in the complex s plane, Fig. 27 where we plot
means that the bandwidth of the closed-loop system— G0!i#", the nominal system, and the band formed by
whatever the feedback law chosen—will always be lim- superposition of the frequency-dependent multiplicative
ited by the lowest-frequency right-hand plane zero of perturbations. At each frequency, the system is located
the system G!s". Thus we see another reason to watch within a circle of radius of radius W!i#", as illustrated.
out for nonminimum phase dynamics and to eliminate Superposing all the circles gives the two bands shown.
them by redesigning the system whenever possible. The multiplicative bound means that the uncertainties
are expressed relative to G0!i#".
There is no deep reason for using multiplicative un-
2. Quantifying model uncertainty certainty. For example, one could use additive uncer-
The internal model control structure highlights the tainty, defined by G!s" ( G!s" + -!s"W!s". But an additive
role that an internal model of the system plays. Since uncertainty is easily redefined in terms of a multiplica-
models are never perfect, one must learn to deal with tive one; in addition, multiplicative uncertainties tend to
the consequences of uncertainty. The first step is to arise naturally in control problems. For example, if the
quantify the uncertainty of the system model. The most actuator !considered as part of the system" has a
obvious way is to allow for uncertainties in any model temperature-dependent gain and the equipment is to be
parameters. For example, in a second-order system such used in rooms of differing temperatures, then the set of
as Eq. !2.5", one could estimate uncertainties in the systems ranges from K0 − -KG0!s" to K0 + -KG0!s",
damping $ and natural frequency #0. There are two limi- meaning that the fractional uncertainty W = -K / K0. In
tations to this approach: First, it assumes that the form general, both K0 and -K could be functions of fre-
of the model is correct. But one may not know the phys- quency.
ics of the model well, and, even if one does, one may A more interesting example is a system that includes
choose to neglect certain parts of the dynamics !such as an unknown time delay +, ranging from 0 to +max. If the
higher-order modes", which translate into errors in the nominal system is G0!s", then, at each frequency #, the
model. Second, one would, in principle, have to come up uncertainty bound +W!i#"+ must be greater than
1 1
with a special way of dealing with the effects of different
kinds of parameters. For example, an uncertainty in e−i#+G0
− 1 = +e−i#+ − 1+ !0 = + = +max". !5.48"
natural frequency just amounts to a rescaling of the time G0
axis, whereas an uncertainty in damping could mean that
In Fig. 26, we show that a suitable bound is
sometimes a system is underdamped and sometimes
overdamped, which could have very different conse- 2.1+maxs
quences for the control law. W!s" = , !5.49"
1 + +maxs
In the control literature, the uncertainty that enters
via imperfectly known parameters is called “structured which can be derived by noting the asymptotic behavior
uncertainty,” since a certain model !“structure”" is as- for # → 0 and # → ) of Eq. !5.48" and increasing the am-
sumed. To get around the problems discussed above, plitude slightly to avoid “cutting the corner” in Fig. 26.
one can consider instead “unstructured uncertainty.” In In practice, one way to construct a bounding function
this approach, we start with a nominal system G0!s". The W!s" is to measure a system transfer function several
actual system, which is unknown, is a member of a set of times, under a variety of conditions. Let the kth transfer-
1 *
M*j ei7j
*
Mjkei7jk − M*j ei7j
1 = +W!i#j"+. !5.50"
be written
+1 + L!i#"+ = +1 + L0!i#" + -!i#"W!i#"L0!i#"+ & 0 ∀ #.
!5.51"
It is important to realize that M* and 7* need not be
the averages of the measurements. The main usefulness This can be illustrated by a diagram analogous to Fig.
of the robust approach is in dealing with the effects of 27, where instead of plotting the family of systems G one
systematic variations. If used for random variations fol- plots the family of loop transfer functions L. Equation
lowing known statistics, it will likely be over- !5.51" then states that the light shaded area cannot touch
conservative. Typically, systematic variations arise be- the point −1. Because, at each #, - is any complex num-
cause only part of a complex system is being modeled. ber with magnitude =1, we can always choose the worst
As discussed above, the effects of higher-order modes case, i.e., the - that minimizes the left-hand side of Eq.
may be neglected or projected away, and there may not !5.51". This implies
be enough separation from the lower-order term to
+1 + L0!i#"+ − +W!i#"L0!i#"+ & 0 ∀ #, !5.52"
model the effect of those neglected modes by a white-
noise term in a Langevin equation, which is the usual or
1 1
physics approach !Chaikin and Lubensky, 1995". An-
other example is nonlinear dynamics, which will be dis- W!i#"L0!i#"
%1 ∀ #. !5.53"
cussed below in Sec. VI. Because the methods we have 1 + L0!i#"
been discussing are based heavily on linear techniques, Since the complementary sensitivity function of the
one must assume that the system dynamics are linear nominal system is T = L0 / !1 + L0", we can write this as
about a given operating point. Although this is often
true, the parameters and even the form of the linear 5WT5) % 1, !5.54"
system can vary with the operating point chosen. A final
where we have used the H) norm as shorthand for
example is that one’s system is almost always embedded
+W!i#"T!i#"+ % 1 ∀ #. The use of the H) norm arises from
within a larger system, whose dynamics are not mod-
the desire to be conservative, to be stable for the worst
eled. Thus an experiment in a room may show different
behavior as a function of temperature. A biochemical
network !see Sec. VII.C below" may function in a cell 30
If the controller is implemented digitally, this will be strictly
whose environment changes significantly in different true. Almost all practical robust controllers are implemented
conditions, and so on. In all of these situations, the right digitally, since the algorithms generate fairly complex control-
thing to do is to choose typical conditions, which may or lers, which would be complicated to implement with analog
may not involve an average and then to identify the ex- circuitry.
31
pected maximum deviation at each frequency #j. +W!s"+ As discussed in Sec. IV.B above, we are simplifying some-
is then an analytic function bounding all of these devia- what. For a given controller K!s", one would invoke the Ny-
tions. quist criterion to see whether any system in the Nyquist plot of
L!i#" circles −1 an appropriate number of times to be un-
3. Robust stability stable. In practice, one is almost always worried about comput-
ing the limit where a system !or set of systems" crosses over
In the previous section, we saw one way to quantify from being stable to unstable, in which case the relevant crite-
the uncertainty in a model of system dynamics. Given rion is 1 + L = 0.
Closely related to chaos control is the notion of syn- With these limitations in mind, we can trace two ad-
chronization, by weak coupling, of two nonlinear sys- ditional approaches to understanding the role of feed-
tems !generally oscillators, perhaps chaotic" !Pikovsky et back in biology. The first course is to find simple enough
al., 2001". The classic example, discovered by Huygens settings where standard feedback ideas may safely be
in the 17th century, is the tendency for two pendulum employed. These are largely found in the biochemistry
clocks hanging on the same wall to synchronize !in an- of enzyme-catalyzed reactions and, more recently, in
tiphase". One can think of this situation as the nonlinear gene-expression models. The second course is to begin
generalization of the observers discussed above in Sec. to tackle more complex situations. Here, the main inno-
III.C. There, the goal was to construct a model of an vation is that instead of a simple feedback loop, one has
observed !linear" dynamical system where the observa- a complex network of interactions, and one has to con-
tions were used to couple !synchronize" the model to the sider both the geometry and topology of that network.
natural system. One then used the deduced internal In the following sections, we will first discuss a phenom-
state vector as the basis for feedback. This is clearly the enological example, then some simple biochemical feed-
same situation discussed by Pikovsky et al. !2001", where back loops, and finally give a brief overview of current
one dynamical system exists in nature, the other as a efforts to understand networks of interacting genes and
computer model, and the observation provides the weak proteins.
coupling. Again, we have a case where there are exten- The reader will notice that the biological examples to
sive, nonoverlapping discussions of essentially the same be discussed lack any man-made element. Our previous
phenomena in the physics and engineering literatures. discussions all started from a natural physical system
The seeming dichotomy between chaotic control and and added a deliberate coupling between variables, cre-
classical control algorithms is not as sharp as I have ating a feedback loop. In the natural system, a variable
made it seem in the last paragraph. The Ott-Grebogi- Y depends on X; then one creates an additional depen-
Yorke algorithm was the first in a long series of algo- dence of X on Y. !For example, in the vibration-
rithms for chaotic control, and many of them blend ele- isolation device described in Sec. V.C.1, the accelerom-
ments of classical algorithms !Schuster, 1999". On the eter signal depends on the forces exerted on it by the
other side, some of the classical control algorithms have Earth. The feedback apparatus then creates a way for
at least some of the flavor of the Ott-Grebogi-Yorke al- the accelerator signal to influence those forces, via the
gorithm. For example, in the example of the unstable piezoelement actuator." In the biological systems below,
Stealth Fighter plane referred to above, instability is ac- the couplings between Y and X and then X and Y are
tually desirable feature as it allows the plane to respond usually both “natural.” But insofar as the two couplings
much faster to control signals than a stable system may be separately identified, feedback will be a useful
would. Again, relatively small control signals !wing flap way of looking at the dynamics, whether the couplings
movements, etc." can produce large effects in the planes are created by man or by nature.
motion, by working about unstable equilibria.
A. Physiological example: The pupil light reflex
VII. APPLICATIONS TO BIOLOGICAL SYSTEMS
The eye is an amazing organism. It can respond with
From the beginnings of control theory, there have reasonable sensitivity over a wide range of light levels,
been attempts to make connections to biological systems from single photons to bright sunlight. The eye uses a
!Wiener, 1961". Indeed, one does not have to look far to mix of feedback methods, the most important of which
find numerous examples of regulation, or “homeosta- is adaptation, as summarized by the empirical response
sis.” Body temperature is regulated to the point where law of Weber: the eye’s sensitivity is inversely propor-
variations of one degree imply sickness and of ten de- tional to the background light level !Keener and Sneyd,
grees, death. When we are hot, we sweat and cool by 1998".
evaporation. When we are cold, we shiver and increase Another mechanism that is particularly important at
the circulation of warm blood to cool areas. Similarly, the highest light levels is the pupil light reflex. When
osmotic pressure, pH, the size of the eye’s pupils, the light levels are high, the pupil contracts, reducing the
vibration of hair cells in the inner ear—all are tightly light flux onto the retina. The size of the pupil is con-
controlled. Over the years, there have been numerous trolled by circularly arranged constricting muscles,
attempts to model such processes at the “system” or which are activated and inhibited !left to relax" by con-
physiological level !Keener and Sneyd, 1998". In many trol signals from the brain. More light causes activation
cases, it has been possible to come up with models that of the constricting muscles, which shrinks the pupil area
mimic, at least partly, the observed behavior. Because and limits the light flux at the retina.
the models describe high-level phenomena, they are The pupil light reflex is a particularly attractive ex-
largely phenomenological. That is, they involve some- ample of physiological feedback. First, the state vari-
what arbitrarily chosen elements that have the right be- able, the retinal light flux, is an intuitive quantity, and
havior but may be only loosely tied to the underlying the actuator mechanism, pupil size, is easy to under-
physiological elements. In addition, one commonly finds stand. Second, using a trick, one can “break open” the
that the more closely one studies a phenomenon, the feedback loop. Ordinarily, the incident light flux covers a
more baroquely complex become the models needed. larger area than the pupil, so that adjustments to the
dx
+x + x = E!t". !7.2"
dt
Here, we have taken simple first-order dynamics with a
time constant +x. The driving E!t" is the rate of arriving
action potentials:
&2 3'
FIG. 32. Light beams and pupil sizes. Light beam is shaded
light; pupil is shaded dark. !a" The ordinary case: the beam is A!t − -t"
larger than the pupil. !b" Breaking the feedback loop: the E!t" = "F ln , !7.3"
Ā
beam is smaller than the pupil. !c" Inducing pupil oscillations: a
small beam falls on the edge of the pupil. where " is a phenomenological rate constant and the
response delay is modeled by evaluating the retinal light
flux a time -t in the past #i.e., by A!t − -t"$. The function
pupil size adjust the retinal light flux #Fig. 32!a"$. If, how- F!x" = x for x ? 0 and 0 for x % 0, so that Ā acts as a
ever, one uses a beam of light that is narrower than the threshold retinal light level. In other words, muscles are
minimum pupil size, adjusting the area will not change activated only when there is sufficient light. The loga-
the retinal light flux #Fig. 32!b"$. The feedback loop is rithm incorporates Weber’s law, mentioned above.
broken, and one can then study how variations in the Equation !7.2" illustrates the “phenomenological” char-
light intensity change the pupil size. For example, one acter of such large-scale models, where the form of the
could impose sinusoidal intensity variations in order to equation, in particular of E!t", is chosen to be in quali-
measure the transfer function of the pupil response. tative agreement with empirical observations.
While we do not have space to describe that transfer In order to close the model, we need to relate the
function fully, its most important feature is that the pupil pupil area A to the muscular activity x. High activity
response to a changing light stimulus is delayed by about should give a small pupil area. Again, one uses an em-
300 ms. This leads to an amusing consequence. If a nar- pirical, phenomenological form:
row beam of light shines on the edge of the pupil #Fig.
32!c"$, the pupil will begin to contract. Because of the (n
delay, it will continue to contract after it is small enough A!x" = Amin + !Amax − Amin" , !7.4"
xn + (n
that the beam no longer enters the pupil. After the de-
lay, the eye realizes there is no light and starts to enlarge which smoothly interpolates between Amax at zero activ-
the pupil. This continues, because of the delay, some- ity !x = 0" to Amin at infinite activity. The parameters (
what past the moment when the pupil is large enough to and n must be fit to experiment. Putting Eqs. !7.1"–!7.4"
admit light to the retina. The system then begins to con- together, we have
&2 3'
tract, and thus continues on, in steady oscillation. This
dx I!t − -t"A#x!t − -t"$
really happens! +x + x = "F ln !7.5"
The ability to open the feedback loop by using a nar- dt Ā
row stimulus light beam also allows one to substitute an
electronic feedback for the natural one. One measures (g„x!t − -t",I!t − -t"…. !7.6"
the pupil-size variations and adjusts the light intensity
Because A!x" is a decreasing function of x, one can
electronically, according to an algorithm chosen by the
always find a steady-state solution to Eq. !7.6" when the
experimentalist. In effect, one creates a primitive “cy-
light intensity I is constant. Let us call this solution x*,
borg,” melding man and machine. If one implements or-
which satisfies x* = g!x* , I". We linearize the equations
dinary proportional feedback, one regulates artificially
about x*, defining x!t" = x* + X!t". This gives
the light intensity to a constant value. The system goes
unstable and the pupil oscillates when the gain is too dX
high. +x + X = − KpX!t − -t", !7.7"
dt
Here, we shall account qualitatively for these observa-
tions using a model describing the pupil light reflex that where Kp = −gx!x* , I" can be viewed as a proportional
is due originally to Longtin and Milton !1989a, 1989b". feedback gain. Equation !7.7" is nothing more than the
Our discussion is a simplified version of that of Keener first-order system with sensor lag and proportional feed-
and Sneyd !1998". back that we considered previously in Sec. IV.C #see
We start by relating the retinal light flux A to the light Eqs. !4.8" and !4.9"$. The stability may be analyzed
intensity I and pupil area A: analogously; one again finds that for high-enough gain
Kp, the system begins to oscillate spontaneously. This is
in accordance with the cyborg experiments using artifi-
A = IA. !7.1" cial gain and with the use of light that falls on the pupil
edge to excite oscillations, too. In the latter case, the
feedback becomes nonlinear, since the coupling changes
We next relate the muscular activity x to the rate of discontinuously when the area reaches a critical value
arriving action potentials !the signals that stimulate the A*, which divides the region where changes in A do or
muscles": do not affect the retinal light flux. Crudely, this functions
& % '
The first, due to Becksei and Serrano !2000" #cf. Gardner
and Collins !2000"$, illustrates how negative feedback 1 1 1 4</1
R* = − + 2 + , !7.9"
can limit variability in gene expression, providing ro- 2 K K K
bustness against changing external conditions. The sec-
ond shows how positive feedback can be used to switch which decreases from < / 1 → 0 when K goes from 0 → ).
a gene on and off !Gardner et al., 2000; Kaern et al., Small perturbations to the cell then lead to small varia-
2003". tions in R!t" = R* + r!t", which obeys the linear equation
<K
ṙ = − 1*r, 1* = 1 + . !7.10"
!1 + KR*"2
The decay rate 1* goes from 1 → 21 for K going from
0 → ). To model fluctuation effects, one adds a stochas-
tic term 0!t" to Eq. !7.10", with 607 = 0 and 60!t"0!t!"7
= B2>!t − t!" !white noise". Then the fluctuations of r obey
what is essentially the equipartition theorem:
1
:r ( 6r27 = !7.11"
%1 * .
In Eq. !7.11", we see explicitly that the negative feed-
back loop, which increases the value of 1*, reduces the
fluctuation of the system in response to a fixed level of
noise. This is what Becksei and Serrano observed in
their experiments. In the same system, Rosenfeld et al.
!2002" later showed directly that the negative feedback
indeed leads to faster response times.
bistability. Examples from real biological systems are webpages and their links to each other", social interac-
more complicated. For example, the lac operon is the tions !e.g., collaborations among individuals, whether as
prototype of a genetic switch, having been studied for actors in films, sexual partners, or scientific co-authors of
some 50 years. !The details are not relevant here: briefly, papers", and in biological networks, as we will discuss
E. coli can use glucose or lactose as a food source. Nor- more below !Albert and Barabási, 2002". Because of the
mally, it does not produce the enzyme to digest lactose, power-law distribution of connections, a few important
but the presence of a small amount of lactose switches nodes !“hubs”" will have many more than the average
on a gene that produces the necessary enzyme." Even number of interconnections and play a central role in
this prototypical example can be complicated. For ex- the network. They serve, first of all, to create the “small-
ample, Keener and Sneyd !1998" cite studies that begin world” phenomenon, where the number of steps needed
with simplified dynamics of six coupled equations and to go between any two nodes increases only logarithmi-
then argue that these equations can be approximated by cally with the number of nodes. They also give robust-
three others. Vilar et al. !2003" have argued that such ness to the structure of the networks: removing nodes
simplified models need to incorporate stochastic fluctua- other than one of the rare hubs will not affect substan-
tions !because of the small number of molecules of the tially the connectivity of the network. But if scale-free
relevant species in each cell" to agree with observations. networks are robust to the destruction of random nodes,
Very recently, Ozbudak et al. !2004" have explored the they are fragile and sensitive to the destruction of hubs.
phase diagram of the lac operon in E. coli !modified by Much work in biology has been devoted to the iden-
fusing fluorescent “reporter” proteins to the genome". tification of different types of networks, most notably
They make quantitative contact to the kinds of genetic- metabolic and protein interaction networks !Jeong et al.,
switch models discussed here. 2000". !Both have proteins as nodes. In the former, the
Rather than entering into the details of how best to interactions are chemical; in the latter, physical." An
model a particular feedback mechanism, we want to em- equally important type of network is that of gene regu-
phasize merely that many, if not all, basic cell functions latory networks, which govern the production of the
depend on interconnected positive and negative feed- proteins involved in the metabolic and interaction net-
back loops. Indeed, it seems likely that such feedbacks works !Maslov and Sneppen, 2002". This is in addition to
are necessary in living organisms. For catalogs of such more physical networks such as the networks of neurons
mechanisms, see, e.g., Freeman !2000" and Keener and in the brain and nervous system and the network of
Sneyd !1998"; for an analysis of elementary biochemical blood vessels. #For a review of all of these, see Newman
mechanisms !amplifications, etc." that adopts an explicit !2003".$
engineering perspective, see Detwiler et al. !2000". The statistical point of view is not the only way to
understand network structure. A number of authors
C. Network example have focused on the “modular” aspects of complex bio-
logical structures, with the goal of identifying the struc-
In the previous section, we saw that both positive and tures and interactions between relatively independent
negative feedback loops are present in basic biochemical elements !Hartwell et al., 1999". This has led to the
systems. In the cell, however, vast numbers of chemical search for “network motifs” !Shen-Orr et al., 2002; Man-
species are present, constantly being synthesized, de- gan and Alon, 2003", which are relatively simple clusters
graded, and otherwise interacting with each other. Thus of nodes that behave as individual elements in a larger
instead of a simple, isolated system modified by a single network.
feedback loop, one has many interacting systems, con- If complex networks are generically present in bio-
nected to each other in a network. In such a network, logical systems, one might suspect that they confer some
the notion of a loop can be generalized to be the set of overall benefit to the host organisms, and one such hy-
interconnections between nodes !the individual pro- pothesized benefit that has lately received much atten-
teins", with positive and negative values assigned to each tion is the notion of robustness. In influential work,
interconnection, depending on whether the presence of Leibler and collaborators have looked at the relatively
protein 1 increases or decreases the concentration of simple network involved in chemotaxis in the bacterium
protein 2. !Here, “1” and “2” represent arbitrary pro- E. coli and shown, both in a model !Barkai and Leibler,
teins in the network." The structure of such networks is a 1997" and in experiment !Alon et al., 1999", that certain
topic of intense current interest !Albert and Barabási, properties show a remarkable robustness in the face of
2002; Newman, 2003". Much attention has been focused large concentration variations of elements within the
on the statistical properties of large networks, for ex- cell. Chemotaxis in E. coli arises by controlling the time
ample, on the distribution of the number of connections the bacterium spends in two states, “smooth runs” and
k a node has with its neighbors. Random networks have “tumbling.” During smooth runs, the cell swims in a
a distribution P!k" peaked about an average number of relatively straight line. It then stops and begins tum-
interconnections 6k7 !Poisson distribution", while for bling, a motion that leads to a random reorientation.
scale-free networks, there is a power-law distribution Then the cell swims for another period of time. The cell
P!k" 4 k−", with " typically have a value of 2–3. Scale- carries receptors for various chemical attractants. If the
free networks have interesting properties and are found level of an attractant is rising, the cell will tend to swim
in many places, including communications settings !e.g., longer before tumbling. In other words, by reducing the
tumbling frequency !rate of tumbling events", the cell !IMP", which states that the controller must contain a
will tend to swim up the spatial gradient of attractor. model of the external signal in order to track robustly.33
Thus, chemotaxis occurs via a modulation of tumbling This is another motivation for the internal model control
frequency. The kind of robustness explored by Leibler discussed in Sec. V.E. Finally, we note that the restric-
and collaborators looks at the adaptation of tumbling tion to linear systems is not necessary !the biochemical
frequency to various changes. For example, the cell re- models are all nonlinear". For example, consider the
sponds to gradients of chemical attractants in a way that nonlinear dynamical system ẋ = f!x , u", with x again an
is nearly independent of their absolute concentration. n-dimensional state vector and f a nonlinear function of
More precisely, the tumbling frequency should, after a x and the input u. If we want the output y = x !for sim-
transient, return to its original value after a sudden in- plicity" to track a constant state r, we can set u = *t!r
crease in the overall concentration of an attractant. − x"dt! !integral feedback". Then differentiating ẋ = f
As Barkai and Leibler !1997" have emphasized, one shows that x = r is a steady-state solution to the modified
can imagine two ways in which perfect adaptation can be dynamics. One caveat about integral feedback is that the
obtained. One way involves a model that has fine-tuned modified dynamical system must be stable. This must be
parameter values that happen to lead to a canceling out verified case by case.
of the effects of a concentration increase. The problem The chemotaxis example discussed above is just one
with such a model is that it implies that adaptation of many instances where a complicated network of feed-
would be a fragile phenomenon, easily disrupted by back interactions plays an important biological role. An-
changes in any of the parameters of the system, which other case considers neural networks, where the firings
does not seem to be the case experimentally !Alon et al., of one neuron stimulate or inhibit other connected neu-
1999". Alternatively, the robustness could be a property rons. Doiron et al. !2003" have recently shown that nega-
of the network itself. We have already seen examples tive feedback plays an essential role in the neural re-
where the level of an output in the face of a step-change sponse of electric fish to communication and prey
disturbance. These all trace back to the use of integral signals. While the organism communicates with others,
feedback, and, indeed, Yi et al. !2000" have shown not neurons in the sensory system studied switch to a glo-
only that integral feedback is present implicitly in the bally synchronized oscillatory state that is maintained by
model of Barkai and Leibler but also that such feedback negative feedback. At other times, the firings are not
must be present in the dynamics. Rather than enter into synchronized.
the details of the chemotaxis model !even the “simpli-
fied” version of Yi et al. has 14 coupled equations", we
sketch the proof that integral feedback must be present
VIII. OTHER APPLICATIONS, OTHER APPROACHES
in order to have robust adaptation. We follow the ap-
pendix of Yi et al. !2000". At this point, it is perhaps proper to note some of the
Consider first a linear SISO model many areas that, for reasons of space but not for lack of
interest, we have decided not to explore. As our discus-
x!˙ = Ãx! + b! u,
sion of biological networks suggests, complex systems
can often be modeled as an interacting network, and
y = c! Tx! + du, !7.13" complex networks often have a modular structure whose
where x! is an n-element internal state vector, y is the function can be understood by appeal to elementary
feedback motifs. Thus, one finds feedback to be a rel-
single output, and u is the single input. At steady state,
evant concept in understanding the weather, climate
x!˙ = 0, so that y = !d − cA−1b"u. !We drop tildes and vector change, in economics, etc. For example, many of the
symbols for simplicity." Then for constant input u, the arguments in the field of global warming amount to de-
output y = 0 if and only if either c = d = 0 !trivial case" or bates about the magnitudes and sizes of feedback ef-
det 2 3
A b
c d
= 0, !7.14"
fects. While the link between industrial and other hu-
man activity and the increased amount of CO2 and other
greenhouse gasses is clear, the climatic effects are much
where the matrix in Eq. !7.14" has n + 1 by n + 1 elements.
If the determinant is zero, then there is vector k! such 33
A more precise statement is that if all disturbances d!t" sat-
that k#A b$ = #c d$. Defining z = k! Tx! , we have ż = kẋ isfy a known differential equation of the form 9ni=0pid!i" = 0,
= k!Ax + bu" = cx + du = y. Thus, if y = 0 for all parameter then one can design a control system that perfectly tracks the
variations !here, these would be variations in the ele- disturbances. Here, d!i" is the ith time derivative of d!t", and
the pi are known, constant coefficients. For example, all step
ments of A, b, c, or d", then ż = y, a condition that is
disturbances satisfy d!1" = 0, while oscillatory disturbances sat-
equivalent to having integral feedback be part of the
isfy d!2" + #d2 d!0" = 0. The internal model principle states that dis-
structure of the system itself. turbances or reference signals are canceled in steady state if
As Yi et al. !2000" comment and we have implicitly their transfer function is contained in the denominator of the
seen in this paper, the requirement of having integral controller K!s"; see Lewis !1992", Chap. 4. The nonlinear gen-
feedback in order to reject step disturbances of param- eralization of the IMP is discussed by Isidori and Byrnes
eters is a special case of the “internal model principle” !1990".
harder to predict, because of the complexity of feedback equately, even under a worst-case scenario. In essence,
effects. There is a tremendous flux of carbon between this is a conservative approach, which works best for
the atmosphere and various “sinks,” such as the oceans smaller sets of transfer functions. !If the variations are
and forests, and the dependence of these fluxes on too large, trying to satisfy all the constraints all the time
greenhouse gasses must be evaluated accurately to know will lead to very weak control." Alternatively, one can
the cumulative effect on climate !Sarmiento and Gruber, try to “learn” which of the possible set of transfer func-
2002". Even more disturbing, it has become clear that tions best describes the system at the present time and
small changes have in the past led to rapid climate to design a control law for the current best estimate of
shifts—on the time scale of decades, or even less. Mod- the dynamical system. This is the overall approach of
els of thermohaline circulation in the oceans show adaptive control, which in its simplest forms treats topics
bistable behavior analogous to the genetic switch dis- such as autotuning of parameters for simple control
cussed above, where positive feedback effects can toggle loops !e.g., PID loops" !Dutton et al., 1997; Franklin et
the climate between warm and icy states !Weart, 2003". al., 1998". The basic algorithms are akin to the Kalman
One rather distinct area left out is the application of filter, in that model parameters are estimated using a
control principles to quantum phenomena !Rice and recursive version of least-squares fitting that updates the
Zhao, 2000; Shapiro and Brumer, 2003; Walmsley and parameter estimates at each time step. In more sophis-
Rabitz, 2003". The basic insight is that adding a coherent ticated analyses, the controller’s structure can be varied
perturbation to a system can enhance or suppress the as well !Isermann et al., 1992". Adaptation and “learn-
amplitude of desired “product” states. The famous two- ing” are used also in approaches from computer science
slit experiment of Young, where two beams of light in- that include the genetic algorithms mentioned above
terfere to produce light and dark fringes, is an elemen- !Jamshidi et al., 2003", neural networks !Norgaard et al.,
tary example. Much of the practical work has used 2000", fuzzy logic !Verbruggen et al., 1999", and “intelli-
coherent light beams interacting with matter to enhance gent” control systems !Hangos et al., 2001". Broadly
or suppress phenomena such as dissociation and chemi- speaking, all of these are attempts to mimic the judg-
cal reaction. Feedback often enters here in only a weak ment of an experienced human operator manually con-
way, where one conducts repeated trials, using the re- trolling a system. Of course, they have applications far
sults to adaptively tune experimental control parameters beyond problems of control. For most physicists, the
such as the amplitude and phase of different frequency more straightforward control techniques described in
components in a shaped pulse; however, some recent this review will be more than adequate.
work has emphasized real-time corrections, where the
ideas discussed here, including state estimation !Kalman IX. FEEDBACK AND INFORMATION THEORY
filtering" and robustness, are beginning to be explored
!Wiseman et al., 2002". Many of these approaches are In various places in the discussion above, we have
based on adaptive optics, a technique that has many noted the informal connection between information and
other applications—for example, compensating for tur- control. For example, in Sec. III.A, we saw how the
bulence in astronomy, where one changes rapidly the strategy of feedforward relied on prior information
shape of a mirror to remove phase disturbances added about the nature of a disturbance. In Sec. V.E.1, we saw
by atmospheric fluctuations !Roggemann et al., 1999". that feedback is required only when our knowledge of
Finally, most of the work to date on quantum systems the system and its disturbances is incomplete. Also, in
has been “semiclassical,” in that sensors perform mea- Sec. V.D, we saw that the Kalman filter gives the optimal
surements on a system, classical computers process the way to blend the two pieces of information one has
results, and semiclassical fields are used to influence the about the current state of a linear dynamical system sub-
future evolution of the system. Lloyd has emphasized ject to white noise perturbations and to white noise in
that one can imagine a fully quantum feedback scheme the measurements. That way best blends the actual mea-
wherein one adds an element to the quantum system surement and the prediction based on the system dy-
that interacts !without making a measurement" with it in namics.
such a way that the dynamics are altered in a desired In all these discussions, the use of the term “informa-
way !Lloyd, 2000". Because information is not destroyed tion” has been informal, but it is natural to wonder
!as it is in the measurement stage of the semiclassical whether there are links to the technical subject of “in-
scheme", higher performance is in principle possible. But formation theory.” Indeed, the influential book Cyber-
this field is still young. netics by Wiener !first published in 1948, the year of
There are also many valuable topics in control theory Shannon’s fundamental papers on information theory"
that we do not have space to discuss. Foremost among explored some connections !Wiener, 1961". Still, there
these is adaptive control, which can be considered as a has been very little development of these ideas. Recent
complement to the approach of robust control, discussed papers by Touchette and Lloyd !2000, 2004" begin to
in Sec. V.E. In both cases, the system to be controlled is explore more formally these links and derive a funda-
at least partially unknown. In robust control, one first mental relationship between the amount of control
tries to describe the set of possible transfer functions of achievable !“decrease of entropy” in their formulation"
close linear systems and then tries to find a controller and the “mutual information” !Cover and Thomas,
that is stable in all situations and still performs ad- 1991" between the dynamical system and the controller
created by an initial interaction. They show that if one sensors close to actuators!" We have also emphasized
measures the accuracy of control by the statistical reduc- that it often pays to spend a bit more on equipment and
tion of uncertainty about the state of the controlled ob- less on fancy control-loop design. We have argued that a
ject, then that reduction is limited by the information mix of feedforward and feedback can work much better
that measurements can extract about the system, in con- than either technique alone. !To anticipate is better than
junction with any improvements that open-loop control to react, but cautious anticipation is better still." Taken
can offer. In their formulation, information becomes the together, the techniques outlined here are probably
common language for both measurement and control. more sophisticated, and more systematic, than what is
The full implications of their results and how they con- commonly practiced among physicists. I certainly hope
nect to the “optimal estimation theory” of Kalman dis- that this article “raises the consciousness” and perhaps
cussed here !and to similar “minimum-variance” con- even the level of practice of physicists as regards feed-
cepts related to Wiener filters !Harris et al., 1999" have back loops. And if higher performance is needed in an
yet to be worked out.34 application, one can, and should, consult the profes-
As an example of the kinds of issues to consider, in sional control-theory books, which I hope will now be
the temperature-control example of Day et al. !1997" al- more accessible.
ready discussed, the temperature control is actually bet- At the same time, we have explored some of the
ter than the sensor-noise limit, because the sample is deeper implications of control theory. We have seen how
surrounded by a regulated shield and is not directly feedback loops and other complicated networks play a
regulated. The last stage then acts like a passive filter. fundamental role in complex systems encountered in na-
Since all external disturbances pass through the shield, ture, particular in biology. We have seen how informa-
the interior has a lower level of fluctuations. This setup tion is the “common coin” of measurements and that
allowed the authors to measure the internal thermal feedback can be thought of in terms of information
fluctuations between a thermometer and a reservoir and flows, from system to controller and back again in the
between two thermometers. simplest case, or from node to node for more compli-
It is interesting that while the link between feedback cated networks. What is interesting is how the reality of
and information theory has been developing almost the physical system itself begins to fade from the picture
since the beginning of the subject itself, the connection once control loops are implemented. In simple cases,
has been slow to sink in and difficult to grasp. As Ben- this occurs when we use feedback to speed up !Sec.
nett !1993" has noted, pre-20th-century controllers were III.A" or slow down dynamics !Sec. V.C.1", implying that
almost invariably “direct” acting, with the force required one can use feedback to compensate for the physical
to effect a change on the system developed by the mea- limitations of the particular elements one may have at
suring device itself. For example, the buoyant force on a hand. Certainly, some of the simpler applications of
ball directly controls the water valve in a toilet. It took a feedback depend on such abilities.
long time for engineers to recognize that the sensor and Other, deeper applications of control depend on the
the actuator were logically distinct devices and that the robustness that feedback can lead to. We saw this in
function of the controller was just to process the infor- integral control, where tracking is achieved over a wide
mation gained through the sensor, converting it to a re- range of control parameters, and it is the loop structure
sponse by the actuator. that ensures the desired result. Such feedback is, of
course, at the heart of the PID loops beloved in indus-
trial applications. But we saw that it also exists in math-
X. CONCLUSIONS ematical and physical models of biological networks
where feedback leads to a robustness that begins to ap-
One of the main goals of this review has been to give proximate the behavior of real biological systems.
a pragmatic introduction to control theory and the basic What is perhaps most interesting to a physicist is the
techniques of feedback and feedforward. We have dis- way new kinds of behavior arise from the structure of
cussed a broad range of applications, including details of control loops. The tracking property of integral feed-
practical implementation of feedback loops. We have back comes from the structure of the feedback loop, not
emphasized basic understanding and outlined the start- the nature of the individual elements. In this sense, it is
ing points for more advanced methods. We have tried in a kind of “emergent phenomenon,” but one that differs
many places to show how a bit of thought about the from the examples familiar in physics, such as the soft
design of the physical system can reduce the demands on modes and phase rigidity that accompany symmetry-
the controller. !Remember to limit time lags by putting breaking phase transitions. Thus, engineering provides
an alternate set of archetypes for emergent phenomena,
34 which—as Carlson and Doyle !2002", Csete and Doyle
One link between the Kalman filter and information theory
!2002", and Kitano !2002" have argued—will perhaps be
is via the use of Bayes’ theorem in constructing the optimal
estimate !Maybeck, 1979"; however, the work of Touchette and more fruitful for understanding biological mechanisms
Lloyd !2000" implies deeper connections. Links to areas such than physics-inspired archetypes. My own view is that
as the robust methods of Sec. V.E are even less clear; see one should not be too dogmatic about which discipline
Kåhre !2002", Chap. 12, for other ideas on how information provides the better model, for in the past physicists have
theory can be related to control issues. often been successful precisely because they were willing
to try to solve the problem at hand “by any means nec- RHP right-hand plane
essary.” rms root-mean square
Still, the easy integration of certain engineering con- RNA ribonucleic acid
cepts into common knowledge is striking. We physicists RNAP RNA polymerase
are justly proud of having developed concepts such as SISO single input, single output
“work,” “energy,” “momentum,” “resonance,” and even SPM scanning probe microscope
“entropy” that have entered the popular language as STM scanning tunneling microscope
metaphors and are used by most people with confidence SVD singular-value decomposition
even when the technical meaning is obscure. More re- SVF state-vector feedback
cently, engineering terms as “bits,” “information,” and TetR tetracycline repressor
“feedback,” which also treat topics that physicists deal ZOH zero-order hold
with, have become equally familiar. If we take up the
last of these terms, the word “feedback” is less than a
century old and was coined to describe an advance in REFERENCES
radio technology !Simpson and Weiner, 1989". In popu-
lar language, it quickly went from a novelty to a fad to a Albert, R., and A.-L. Barabási, 2002, “Statistical mechanics of
word so often used that one can scarcely imagine its complex networks,” Rev. Mod. Phys. 74, 47–97.
recent origin. Feedback is a useful technical tool that Alon, U., M. G. Surette, N. Barkai, and S. Leibler, 1999, “Ro-
underlies much of modern technology, but it is also an bustness in bacterial chemotaxis,” Nature !London" 397, 168–
essential ingredient of biological systems and even of life 171.
Barkai, N., and S. Leibler, 1997, “Robustness in simple bio-
itself. Is it then merely by chance that “feedback,” de-
chemical networks,” Nature !London" 387, 913–917.
spite its narrow technical origins, has found such ready
Barone, F., E. Calloni, A. Grado, R. D. Rosa, L. D. Fiore, L.
acceptance and wide use in popular culture? Milano, and G. Russo, 1995, “High accuracy digital tempera-
ture control for a laser diode,” Rev. Sci. Instrum. 66, 4051–
ACKNOWLEDGMENTS 4054.
Becksei, A., and L. Serrano, 2000, “Engineering stability in
I thank Vincent Fourmond, Russ Greenall, Christoph gene networks by autoregulation,” Nature !London" 405,
Hebeisen, Bloen Metzger, Laurent Talon, and Anand 590–593.
Yethiraj for collaboration on some of the experimental Bennett, S., 1993, “Development of the PID controller,” IEEE
work cited here. I am grateful for comments by Ping Ao, Control Syst. Mag. 13 !6", 58–65.
Adrienne Drobnies, Ted Forgan, Bruce Francis, Jean- Bennett, S., 1996, “A brief history of automatic control,” IEEE
Christophe Géminard, Andrew Lewis, Albert Libch- Control Syst. Mag. 16 !3", 17–25.
aber, and Juan Restrepo, and for support by NSERC Bennett, S., 2002, “Otto Mayr: Contributions to the history of
!Canada". feedback control,” IEEE Control Syst. Mag. 22 !2", 29–33.
Black, H. S., 1934, “Stabilized feedback amplifiers,” Bell Syst.
Tech. J. 13, 1–18.
LIST OF ABBREVIATIONS Bode, H. W., 1945, Network Analysis and Feedback Amplifier
Design !van Nostrand, New York".
A/D analog to digital Borcherding, F., S. Grünendahl, M. Johnson, M. Martin, J.
AFM atomic force microscope Olsen, and K. Yip, 1999, “A first-level tracking trigger for the
D/A digital to analog upgraded D0 detector,” IEEE Trans. Nucl. Sci. 46, 359–364.
DNA deoxyribonucleic acid Carlson, J. M., and J. Doyle, 2002, “Complexity and robust-
DSP digital signal processor ness,” Proc. Natl. Acad. Sci. U.S.A. 99, 2538–2545.
EKF extended Kalman filter Chaikin, P. M., and T. C. Lubensky, 1995, Principles of Con-
FPGA field-programmable gate array densed Matter Physics !Cambridge University Press, Cam-
IIR infinite impulse response bridge, UK".
IMC internal model control Cherry, J. L., and F. R. Adler, 2000, “How to make a biological
IMP internal model principle switch,” J. Theor. Biol. 203, 117–133.
LHP left-hand plane Cohen, A. E., 2005, “Control of nanoparticles with arbitrary
LIGO Laser Interferometer Gravitational Wave two-dimensional force fields,” Phys. Rev. Lett. 94, 118102.
Cooper, W. S., 1986, “Use of optimal estimation theory, in par-
Observatory
ticular the Kalman filter, in data processing and signal analy-
LQG linear quadratic Gaussian
sis,” Rev. Sci. Instrum. 57, 2862–2869.
LSB least-significant bit Cover, T., and J. Thomas, 1991, Elements of Information
MIMO multiple input, multiple output Theory !Wiley, New York".
mRNA messenger RNA Croft, D., and S. Devasia, 1999, “Vibration compensation for
NMP nonminimum phase high speed scanning tunneling microscopy,” Rev. Sci.
ODE ordinary differential equation Instrum. 70, 4600–4605.
PI proportional-integral Csete, M. E., and J. C. Doyle, 2002, “Reverse engineering of
PID proportional-integral-derivative control law biological complexity,” Science 295, 1664–1669.
PLD programmable logic device Day, P., I. Hahn, T. C. P. Chui, A. W. Harter, D. Rowe, and J.
PRBS pseudorandom binary sequence A. Lipa, 1997, “The fluctuation-imposed limit for tempera-
ture measurement,” J. Low Temp. Phys. 107, 359–370. Harris, T. J., C. T. Seppala, and L. D. Desborough, 1999. “A
Detwiler, P. B., S. Ramanathan, A. Sengupta, and B. I. review of performance monitoring and assessment techniques
Shraiman, 2000, “Engineering aspects of enzymatic signal for univariate and multivariate control systems,” J. Process
transduction: Photoreceptors in the retina,” Biophys. J. 79, Control 9, 1–17.
2801–2817. Hartwell, L. H., J. J. Hopfield, S. Leibler, and A. W. Murray,
Devasia, S., 2002, “Should model-based inverse inputs be used 1999, “From molecular to modular biology,” Nature
as feedforward under plant uncertainty?,” IEEE Trans. Au- !London" 402, C47–C52.
tom. Control 47, 1865–1871. Hensley, J. M., A. Peters, and S. Chu, 1999, “Active low fre-
Doiron, B., M. J. Chacron, L. Maler, A. Longtin, and J. Bas- quency vertical vibration isolation,” Rev. Sci. Instrum. 70,
tian, 2003, “Inhibitory feedback required for network oscilla- 2735–2741.
tory responses to communication but not prey stimuli,” Na- Isermann, R., K.-H. Lachmann, and D. Matko, 1992, Adaptive
ture !London" 421, 539–543. Control Systems !Prentice-Hall, New York".
Doyle, J. C., B. A. Francis, and A. R. Tannenbaum, 1992, Ishida, F., and Y. E. Sawada, 2004, “Human hand moves pro-
Feedback Control Theory !Macmillan, New York", URL actively to the external stimulus: An evolutional strategy for
http://www.control.utoronto.ca/people/profs/francis/dft.html minimizing transient error,” Phys. Rev. Lett. 93, 168105.
Dutton, K., S. Thompson, and B. Barraclough, 1997, The Art Isidori, A., and C. I. Byrnes, 1990, “Output regulation of non-
of Control Engineering !Addison-Wesley, Harlow, England". linear systems,” IEEE Trans. Autom. Control 35, 131–140.
Etchenique, R., and J. Aliaga, 2004, “Resolution enhancement Jamshidi, M., L. dos Santos Coelho, R. A. Krohling, and P. J.
by dithering,” Am. J. Phys. 72, 159–163. Fleming, 2003, Robust Control Systems with Genetic Algo-
Evensen, G., 2003, “The ensemble kalman filter: Theoretical rithms !CRC, Boca Raton, FL".
formulation and practical implementation,” Ocean Dynamics Jeong, H., B. Tombor, R. Albert, Z. N. Oltvai, and A.-L.
53, 343–367. Barabási, 2000, “The large-scale organization of metabolic
Eyink, G. L., J. M. Restrepo, and F. J. Alexander, 2004, “A networks,” Nature !London" 407, 651–654.
mean field approximation in data assimilation for nonlinear Kaern, M., W. J. Blake, and J. J. Collins, 2003, “The engineer-
dynamics,” Physica D 195, 347–368. ing of gene regulatory networks,” Annu. Rev. Biomed. Eng.
Forgan, E. M., 1974, “On the use of temperature controllers in 5, 179–206.
cryogenics,” Cryogenics 14, 207–214. Kåhre, J., 2002, The Mathematical Theory of Information !Klu-
Franklin, G. F., J. D. Powell, and A. Emami-Naeini, 2002, wer Academic, Boston".
Feedback Control of Dynamical Systems, 4th ed. !Prentice- Kalman, R. E., 1960, “A new approach to linear filtering and
Hall, Upper Saddle River, NJ". prediction problems,” J. Basic Eng. 82, 35–45.
Franklin, G. F., J. D. Powell, and M. Workman, 1998, Digital Keener, J., and J. Sneyd, 1998, Mathematical Physiology
Control of Dynamical Systems !Addison-Wesley, Reading, !Springer-Verlag, New York".
MA". Kitano, H., 2002, “Looking beyond the details: A rise in
Freeman, M., 2000, “Feedback control of intercellular signal- systems-oriented approaches in genetics and molecular biol-
ling in development,” Nature !London" 408, 313–319. ogy,” Curr. Genet. 41, 1–10.
Gamkrelidze, R. V., 1999, “Discovery of the maximum prin- Leigh, J. R., 2004, Control Theory, 2nd ed. !The Institution of
ciple,” J. Dyn. Control Syst. 5, 437–451. Electrical Engineers, London".
Gardner, T. S., C. R. Cantor, and J. J. Collins, 2000, “Construc- Lewis, F. L., 1992, Applied Optimal Control and Estimation
tion of a genetic toggle switch in Escherichia coli,” Nature !Prentice-Hall, Englewood Cliffs, NJ".
!London" 405, 520–521. Ljung, L., 1999, System Identification: Theory for the User, 2nd
Gardner, T. S., and J. J. Collins, 2000, “Neutralizing noise in ed. !Prentice-Hall, Upper Saddle River, NJ".
gene networks,” Nature !London" 403, 339–342. Lloyd, S., 2000, “Coherent quantum feedback,” Phys. Rev. A
Gershenfeld, N., 1999, The Nature of Mathematical Modeling 62, 022108.
!Cambridge University Press, Cambridge, UK". Longtin, A., and J. Milton, 1989a, “Insight into the transfer
Gershenfeld, N., 2000, The Physics of Information Technology function, gain, and oscillation onset for the pupil light reflex
!Cambridge University Press, Cambridge, UK". using nonlinear delay-differential equations,” Biol. Cybern.
Goldenfeld, N., 1992, Lectures on Phase Transitions and the 61, 51–58.
Renormalization Group !Perseus, Reading, MA". Longtin, A., and J. Milton, 1989b, “Modelling autonomous os-
Goodwin, G. C., S. F. Graebe, and M. E. Salgado, 2001, Con- cillations in the human pupil light reflex using non-linear
trol System Design !Prentice Hall, Upper Saddle River, NJ". delay-differential equations,” Bull. Math. Biol. 51, 605–624.
Gosse, C., and V. Croquette, 2002, “Magnetic tweezers: micro- Mancini, R., 2002, Op Amps for Everyone: Texas Instruments
manipulation and force measurement at the molecular level,” Guide, Technical Report, Texas Instruments, URL http://
Biophys. J. 82, 3314–3329. focus.ti.com/lit/an/slod006b/slod006b.pdf
Gray, M., D. McClelland, M. Barton, and S. Kawamura, 1999, Mangan, S., and U. Alon, 2003, “Structure and function of the
“A simple high-sensitivity interferometric position sensor for feed-forward loop network motif,” Proc. Natl. Acad. Sci.
test mass control on an advanced LIGO interferometer,” U.S.A. 100, 11 980–11 985.
Opt. Quantum Electron. 31, 571–582. Maslov, S., and K. Sneppen, 2002, “Specificity and stability in
Grebogi, C., and Y.-C. Lai, 1999, in Handbook of Chaos Con- topology of protein networks,” Science 296, 910–913.
trol, edited by H. G. Schuster !Wiley-VCH, New York", pp. Maxwell, J. C., 1868, “On governors,” Proc. R. Soc. London
1–20. 16, 270–283.
Hangos, K. M., R. Lakner, and M. Gerzson, 2001, Intelligent Maybeck, P. S., 1979, Stochastic Models, Estimation, and Con-
Control Systems: An Introduction with Examples !Kluwer trol !Academic, New York", Vol. 1.
Academic, Dordrecht". Mayr, O., 1970, “The origins of feedback control,” Sci. Am.
Phys. Rev. A 66, 013807. Yi, T.-M., Y. Huang, M. I. Simon, and J. Doyle, 2000, “Robust
Wolf, D. M., and F. H. Eeckman, 1998, “On the relationship perfect adaptation in bacterial chemotaxis through integral
between genomic regulatory element organization and gene feedback control,” Proc. Natl. Acad. Sci. U.S.A. 97, 4649–
regulatory dynamics,” J. Theor. Biol. 195, 167–186. 4653.
Yethiraj, A., R. Mukhopadhyay, and J. Bechhoefer, 2002, Zou, Q., C. V. Giessen, J. Garbini, and S. Devasia, 2005, “Pre-
“Two experimental tests of a fluctuation-induced first-order
cision tracking of driving wave forms for inertial reaction de-
phase transition: Intensity fluctuation microscopy at the
vices,” Rev. Sci. Instrum. 76, 1–9.
nematic–smectic-A transition,” Phys. Rev. E 65, 021702.