Professional Documents
Culture Documents
Guo 2005
Guo 2005
9, SEPTEMBER 2005
REFERENCES I. INTRODUCTION
[1] B. D. O. Anderson, M. Deistler, L. Farina, and L. Benvenuti, “Nonneg- Consider an investor who holds a share of stock whose price at time
is X (t). The investor’s goal is to find the “best” selling time to
ative realization of a linear system with nonnegative impulse response,”
IEEE Trans. Circuits Syst. I, Fundam. Theory Appl., vol. 43, no. 2, pp. t
134–142, Feb. 1996. maximize the discounted expected payoff, i.e., the value function is
[2] B. D. O. Anderson, “New developments in the theory of positive sys- max 1 Ex [e0r (X 0 K )], where K is the amount to be paid back
tems,” in Systems and Control in the Twenty-First Century. Boston, when the investor sells the stock, X (0) = x, and r is the discount factor
MA: Birkhäuser, 1997.
(e.g., due to inflation). Assume also that the investor is not clairvoyant.
This is an optimal stopping time problem with the stopping time t
[3] L. Benvenuti and L. Farina, “A tutorial on the positive realization
problem,” IEEE Trans. Autom. Control, vol. 49, no. 5, pp. 651–664,
May 2004. measurable to the “information” available up to time t (characterized
[4] , “The design of fiber-optic filters,” J. Lightw. Technol., vol. 19, no. by the -algebra generated by X (s); 0 s t). Clearly, the choice
9, pp. 1366–1375, Sep. 2001. of an optimal stopping rule will be dictated by the underlying model
for X ( 1 ).
[5] , “An example of how positivity may force realizations of “large”
dimensions,” Syst. Control Lett., vol. 36, pp. 261–266, 1999.
[6] , “A note on minimality of positive realizations,” IEEE Trans. Cir- This problem was first studied by McKean [6] in the 1960s. As-
cuits Syst. I, Fundam Theory Appl., vol. 45, no. 6, pp. 676–677, Jun. suming that X (t) follows a geometric Brownian motion (now well
1998. known as the Black–Scholes model) such that dX (t) = X (t)dt +
X (t)dW (t), and with (X 0 K ) modified equivalently by (X 0
[7] L. Benvenuti, L. Farina, and B. D. O. Anderson, “Filtering through a
K ) , he solved the problem explicitly. He showed that if > r , then
combination of positive filters,” IEEE Trans. Circuits Syst. I, Fundam +
Theory Appl., vol. 46, no. 12, pp. 1431–1440, Dec. 1999.
[8] L. Benvenuti, L. Farina, B. D. O. Anderson, and F. De Bruyne, “Min- one can patiently “wait and see” and the value function is infinite; if
imal positive realizations of transfer functions with positive real poles,” 3 3
< r , then there exists an x = x (; r; ; K ) so that the optimal
IEEE Trans. Circuits Syst. I, Fundam Theory Appl., vol. 47, no. 9, pp. stopping time = inf ft > 0; X (t) x3 g; = r is a degenerate
3
case for which the value function is x, the initial value of X (t), with
1370–1377, Sept. 2000.
II. MAIN RESULTS Solving this second-order ODE with the form of Mi t ( ) = exp(Bt)
gives us
Consider the following optimal stopping problem with a state–space
(x; ): Mi t ( ) = Ai1 exp(B1 t) + Ai2 exp(B2 t):
3
V (x ) = sup E [e
0r (X ( ) 0 K ) j X (0) = x; (0) = i] (2)
i
0 1 Here, the coefficients are uniquely determined by the boundary condi-
tions.
= ( ( ) ( ))
where is an Ft f W s ; s j s tg-stopping time. The degenerate case when 1 2 =0
can be solved in a similar
0
Here, we assume x > and exclude the case of x =0
for which fashion. This completes the proof.
V3 =0 and = 1.
[ ]
Intuitively, if r is very small so that ert < E Xt , then it does not A. Case 1: r B1
hurt to wait and the payoff is infinite. This intuition can be formalized.
Lemma 1: When i > ; i 0 ( = 1 2)
;
()
First, note that X t in (1) can be written as
2
0 2(s)
t t t
Mi t( ) = E exp ( ) j (0) = i
s ds ( ) = X (0)exp
X t (s) ds + ()
(s) dW t :
0 0 0
= Bi1 00 BB22 eB t + BB11 00 Bi2 eB t : exp( 2
( ) (1 2) )
0 (s) dW s 0 = 0 (s) 0dsrt defines a
(3) t t
In view of this,
martingale. Moreover, it is clear from Lemma 1 that fe Xt gt0 is
If i = 0, then Mi (t) = exp(i t); and for i 6= j; j = 1; 2 submartingale if B1 > r , and supermartingale if B1 < r . Furthermore
M j ( t) =
(i 0 j )exp(j 0 j )t + j exp i t : lim 0rt (B 0r)t X0
i 0 j + j i 0 j + j t!1 E [e Xt ] = tlim
!1 e
1; if B1 > r
Here, B1 > B2 are the roots = x; if B1 r = (5)
2 0 (1 0 1 + 2 0 2 )x + (1 0 1 )(2 0 2 ) 0 1 2 = 0 0; if B1 < r
x
and for T < 1 and any stopping time
such that
E e
0r( ^T ) X ^T 0 e0r( ^T ) K E e0r( ^T ) X ^T : (6)
B1;2
with the initial conditions Mi (0) = 1; Mi0 (0) = i . This inequality becomes an equality if the strategy be optimal.
1452 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 50, NO. 9, SEPTEMBER 2005
Now, assuming the value functions are sufficiently smooth, Ito’s rule Smooth fit along the boundaries (i.e., at x = x1 and x = x2 ) yields
yields, for x 2 [x1 ; x2 ] V1 (x+) = V1 (x0) and V10 (x+) = V10 (x0) so that
A1 x1 + A2 x1 = x1 0 K;
x2 V20 (x) + x2 22 V200 (x) + 2 (x 0 K );
1
(r + 2 )V2 (x) = (16)
2 1 A1 x1 + 2 A2 x1 = x1 :
V 1 (x ) = x 0 K
(7) Smoothness of V2 (x) at x1 and x2 (i.e., the “modified smooth fit” as
in [3]) suggests
for x 2 [0; x1 )
l1 A1 x1 + l2 A2 x1 = C1 x1
+ C2 x1
+ (x1 )
0
x1 V10 (x) + x2 12 V100 (x) + 1 V2 (x) l1 1 A1 x1 + l2 2 A2 x1
1
(r + 1 )V 1 (x ) = =
1 C 1 x 1 +
2 C 2 x 1 + x 1 (x 1 )
2 (8)
0 1 2 2 00
(r + 2 )V2 (x) = x2 V2 (x) + x 2 V2 (x) + 2 V1 (x)
(17)
2
hence
1
+ 12
(
0 1) = r + 1 :
1
x2 V 0 (x) + x2 22 V 00 (x) + 2 (x 0 K ) (12)
1 (23)
(r + 2 )V (x ) = 2
2
Accordingly F1 and F2 are replaced, respectively, by
to a linear ODE via a change of variables x = ey leads to
1 1
01 l1
~ ~l2 1 1
01
V2 (x) = C1 x
+ C2 x
+ (x ) (13) F~1 (x; g(x)) =
1
2 l1 1
~ l2 2
~ 1 2
where (x) is a special solution to (12), and
i (i = 1; 2) are the real x0K g (x )
roots of
2 x
0 xg0 (x)
(24)
2
+ 22
(
0 1) = r + 2 :
1
(14) li = 1=li and
with ~
2
01 x 0 K 0 g(x)
In particular, if r + 2 0 2 6= 0, one can take F~2 (x; g(x)) =
1 1
:
x 0 xg0 (x)
(25)
~1
~2
2 K 2 x
(x ) = 0 : 3) Case x1 = x2 = x3 : It is not difficult to show that the value
r + 2 0 2
+ (15)
r + 2
function reduces to the McKean’s problem in this case. Recall that
In order to uniquely determine V1 (x) and V2 (x), we must solve for
A1 ; A2 ; C1 ; C2 ; x1 , and x2 . To this end, we need apply the smooth fit V 1 (x ) = A 1 x + A 2 x
principle. V 2 (x ) = l 1 A 1 x + l2 A 2 x
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 50, NO. 9, SEPTEMBER 2005 1453
for x 2 [0; x3 ] and V1 (x) = V2 (x) = x 0 K for x x3 . Smooth fit In particular, if 1 = 2 ; 1 = 2 , then x1 = x2 = x3 , and
conditions lead to
3 10
(x )
A 1 (x 3 ) + A 2 (x 3 ) = x 3 0 K V 3 (x ) = V 1 (x ) = V 2 (x ) = x ; if x < x3
i
1 A 1 (x 3 ) + 2 A 2 (x 3 ) = x 3 x 0 K; if x x3
(26)
and
l1 A 1 (x 3 ) + l2 A 2 (x 3 ) = x 3 0 K where > 0 satisfies r 0 (i 0 (1=2)i2) 0 (1=2)i2 2 = 0 and
l1 1 A 1 (x 3 ) + l2 2 A 2 (x 3 ) = x 3 :
(27) x3 = K=( 0 1))(>rK=(r 0 i )), and
i ;
~; ;
~ , are given by (14),
(23), (15), and (22), respectively.
Necessarily, we have A1 = l1 A1 and A2 = l2 A2 , meaning V1 = V2 One can prove the optimality of the value function by modifying the
and 1 = 2 ; 1 = 2 from (8). The value function can then be easily verification theorem argument in [3]. Here, we adopt a slightly different
rederived via the smooth fit, which is exactly the solution of McKean. approach.
We summarize our results as follows. For ease of exposition, we use Proof of Theorem 2: It is easy to see that Vi (1) = 1; i = 1; 2,
Y1 ; Y2 to denote column vectors and define and
x0
x0
D = f(x; 1) j x 2 (0; x1 )g [ f(x; 2) j x 2 (0; x2 )g:
Y1 0
0 0
H (x 1 ; x 2 ; Y 1 ; Y 2 ) = Y2
x0
x0
1 2
0 0
1 2
(28) For any v (x; i) 2C 2
, define
Y1 0
0 0
~ (x 1 ; x 2 ; Y 1 ; Y 2 ) =
H Y2 : @x2
x0 x0
1 2 2
~
~
0 1 0 2 + i (v (x; 3 0 i) 0 v (x; i)) 0 rv (x; i):
(29)
Take v (x; i) = Vi (x). Then it is easy to check that the following dy-
Theorem 2: Assume r > max(1 ; 2 ), and 1 6= 2 or 1 6= 2 . namic programming principle holds:
Case 1: If there exists an x1 < x2 such that
H (x1 ; x2 ; F1 (x1 ; (x1 )); F2 (x2 ; (x2 ))) = 0 and max( Lv; x 0 K 0 v) = 0:
x2 rK=(r 0 max(1 ; 2 )), then Vi3 (x) = Vi (x), if
Vi (x) > x 0 K with Note that Vi (x) is C 1 and not necessarily C 2 at x1 —this is remedied
by the smooth approximation. Let
A1 x + A2 x ; if x < x1
V 1 (x ) =
x 0 K; if x x1 0 (x ) = e01=(10x ) ; if jxj < 1
l1 A 1 x + l2 A 2 x ; if x < x1 0; if jxj 1:
V 2 (x ) = C1 x
+ C2 x
+ (x); if x2 > x x1 (30) 1
Define (x) = 0 (x)= 01 0 (x) dx. Then, it is easy to check that
x 0 K; if x x2 (x) 2 C01 and 01 1 (x) dx = 1. Denote k (x) to be the kernel
where 1 > 2 > 0 are roots of (9) function k (x) = (1= )(x= ) and let v (x; i) = (k 3 v )(x; i) with
v(x; i) = Vi (x) be the corresponding convolution. Then, v (x; i) 2
A1 x1 x1 01 x1 0 K C 1 , > 0. Moreover, as ! 0, we can show as in Øksendal [7, Th.
A2
=
1 x1 2 x1 x1 D.1] that: 1) v ! v uniformly on compact sets in R; 2) Lv ! Lv
01 x2 0 K 0 (x2 ) uniformly on compact sets in (@D)c ; and 3) fLv g is locally bounded
C1 x
2 x
2 on R. With these properties, one can show as in [7] that, for any stop-
:
x 2 0 x 2 0 (x 2 )
=
C2
1 x
2
2 x
2 ping time ;
Case 2: If there exists an x1 > x2 such that v(x; i) E e0r(t^ ) v(X (t ^ ); (t ^ ))
~ (x 1 ; x 2 ; F
H ~2 (x1 ;
~(x1 )); F
~1 (x2 ;
~(x2 ))) = 0 and
x1 rK=(r 0 max(1 ; 2 )), then Vi3 (x) = Vi (x) if E e0r t^ (X (t ^ ) 0 K )
( )
: (32)
Vi (x) x 0 k with
Therefore
A~1 x + A~2 x ; if x < x2
V 1 (x ) = C~1 x
~ + C~2 x
~ + ~(x); if x1 > x x2 v(x; i) E [e0r v(X ( ); ( ))] E [e0r (X ( ) 0 K )] (33)
x 0 K; if x x1
l1 A~1 x + l2 A~2 x ; if x < x2 from the uniform integrability of e0rt v (X (t); (t)). To show the
V 2 (x ) = optimality of 3 , note that if 3 < 1, a.s., then v (X ( 3 ); ( 3 )) =
x 0 K; if x x2
(31)
X ( 3 ) 0 K . In this case, Dynkin’s formula yields v(x; i) =
where E [e0r (X ( 3 ) 0 K )]. Otherwise, let Dk = D \ f(x; i) :
1=k < x < k; i = 1; 2g, for k = 1; 2; . . ., and let k = inf ft
A~1 l1 x2 l2 x2 01 x2 0 K 3
0 j (X (t); (t)) 62 Dk g. Then clearly k ! a.s. Moreover,
as in [9, Ths. 4.5 and 4.6], we see that, for each k; k < 1
=
A~2 l1 1 x2 l2 2 x2 x2
and a.s. Next, by the definition of k , we have v (X (k ); (k )) =
C~1 x1
~ x1
~ 01 x1 0 K 0 ~(x1 ) v(X (k ); (k ))IfX ( )=kg + v(X (k ); (k ))IfX ( )<kg . Note that
C~2
=
~1 x
1~
~2 x
1~ x1 0 x1 ~0 (x1 )
: v(X (k ); (k ))IfX ( )<kg = (X (k ) 0 K )IfX ( )<kg X (k ) 0
K . Note also that e0r IfX ( )=kg ! 0, as k ! 1, a.s. It follows
Furthermore, let D = f(x; i) j Vi (x) > x 0 K g. The optimal stopping that E [e0r v (X (k ); (k ))IfX ( )=kg ] ! 0. Therefore, we have
rule for both cases is given by 3 = inf ft > 0 j (X (t); (t)) 62 Dg. v(x; i) E [e0r v(X (k ); (k ))] ! E [e0r (X ( 3 ) 0 K )], as
1454 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 50, NO. 9, SEPTEMBER 2005