Professional Documents
Culture Documents
On Initialization of The Kalman Filter : Shunyi Zhao and Biao Huang
On Initialization of The Kalman Filter : Shunyi Zhao and Biao Huang
On Initialization of The Kalman Filter : Shunyi Zhao and Biao Huang
Abstract— As a recursive algorithm, the Kalman filter (KF) KF and how much performance is sacrificed with imprecise
assumes the initial state distribution is known a priori, while initial guesses have not been well answered.
the initial distributions used in practice are commonly treated In this paper, an batch form of the KF is first derived, from
as design parameters. In this paper, the influences of initial
states are analyzed under the KF framework. That is, we which we search for the answers for the above questions.
address the questions about how the initial mean and variance Then, two initialization strategies are given for the cases
affect the subsequent estimates and how much performance is with insufficient a priori knowledge about the initial state
sacrificed if incorrect values are used. Based upon this, two distribution.
initialization methods are developed for the cases with large
initial uncertainties. A drafting stochastic resonator model is II. PRELIMINARIES AND PROBLEM
employed to verify the theoretical analysis result as well as the FORMULATION
proposed initialization approach.
Consider a general linear state-space model
I. INTRODUCTION
xn = Fn xn−1 + Gn wn , (1)
The Kalman filter (KF) is known as optimal for the linear yn = Hn xn + vn , (2)
Gaussian state-space model in the minimum mean square
error (MMSE) sense [1]. Due to its global optimality and where xn ∈ RK denotes the state vector at time n, yn ∈ RP
simple recursive structure, a large number of methods are is the measurement vector, Gn ∈ RK×L is the noise matrix,
proposed based on the KF to modify or extend the original Fn ∈ RK×K and Hn ∈ RP ×K denote the state transition
algorithm to suit certain “unsatisfied” conditions [2]–[5]. matrix and measurement matrix respectively, and wn ∈ RL
According to the Kalman filtering theory, the distribution and vn ∈ RP are the white Gaussian noises with zero means
of initial state x0 is assumed to be known, and the filtering and known variances, i.e., wn ∼ N (0, Qn ) and vn ∼
procedure generally starts from the mean and variance of x0 . N (0, Rn ). It is assumed that the initial state x0 , system noise
When applying the KF, however, we often face a dilemma wn , and measurement noise vn are mutually uncorrelated at
that the available information about p(x0 ) is insufficient or every time step. A control input un can be included in this
even not available. Then, one will eventually end up with model, but it is considered as a known variable and can be
a “guess” of the mean value and increasing the variance omitted for the sake of clarity. To have a stable KF, [Fn Hn ]
1/2
artificially to accommodate for the uncertainties. Obviously, is assumed to be uniformly detectable and [Fn Hn Qn ] is
1/2 1/2 T
this is not a systematic solution. When the initial guess uniformly stabilizable with Qn (Qn ) = Qn [8].
is poor, the errors will propagate through the recursions With model (1) and (2), the KF gives the optimal estimates
as an undiscovered bias [6], resulting in some long lasting in the MMSE sense through the following three steps:
transients. In the applications where state estimate is required Initialization: With a known initial distribution p(x0 ) ∼
to be fast, this phenomenon is barely tolerated [7]. N (x̄0 , P0 ), start the recursions from x̂0|0 = x̄0 and P0|0 =
In view of this, some improved methods have been pro- P0 .
posed. For example, a strategy that using the conditional Predication: Use the transition equation and the posterior
mean value of state on the first measurement was suggested estimate associated with the previous state to compute the
in [6]. By tracking the inverse of error covariance, the infor- prediction probability density function (pdf):
mation filter can handel the cases with infinite covariance [1],
x̂n|n−1 = Fn x̂n−1|n−1 , (3)
[8]. A neural network and a fuzzy logic were combined in
[9] to accelerate the convergence rate in presence of incorrect Pn|n−1 = Fn Pn−1|n−1 FnT + Gn Qn GTn , (4)
initial guessing. In [10], a method based on state partitioning Updating: Update the prediction pdf with the likelihood:
was given, and the approach for a specific problem can be
found in [11]. Although these methods alleviate the poor x̂n|n = x̂n|n−1 + Kn (yn − Hn x̂n|n−1 ) , (5)
initialization problem to some extend, none of them really Pn|n = Pn|n−1 − Kn Hn Pn|n−1 , (6)
eliminates the negative effect of incorrect initial guessing in ( )−1
Kn = Pn|n−1 HnT Hn Pn|n−1 HnT + Rn . (7)
a rigorous way [12]. The main problem we believe is how
the initial states affect the estimate at each time step in the Here, x̂n|n−1 and Pn|n−1 are the predicted (or prior)
state estimation and variance given observations up to and
*This work was supported by NSERC and AITF. including time n − 1, Kn is the Kalman gain, and x̂n|n
S. Zhao and B. Huang are with Department of Chemical and Materials
Engineering, University of Alberta, Edmonton, Alberta, AB, Canada T6G and Pn|n are the updated (or posterior) state estimation and
2G6, shunyi@ualberta.ca, bhuang@ualberta.ca variance given observations up to and including time n.
Substituting xi with the equation associated with xi−1 , where Fn,1 E{x0 } = Kn,1 Hn,1 E{x0 } + Sn,1 x̄0
i changes from 2 to n−1, and following a similar trick for the = (Kn,1 Hn,1 + Sn,1 )E{x0 } ,
measurement equation, it is not difficult to find the extended
state-space model on the time interval [0 , n] given by where the facts that E{Wn,1 } = 0, E{Vn,1 } = 0, and x̄0 =
E{x0 } are used. Since E{x0 } cannot always be zero, one
Xn,1 = Fn,1 x0 + Gn,1 Wn,1 , (8)
consequently gets Sn,1 = Fn,1 − Kn,1 Hn,1 , which enables
Yn,1 = Hn,1 x0 + Ln,1 Wn,1 + Vn,1 , (9) us to represent (10) equivalently by
where Xn,1 ∈ RnK , Wn,1 ∈ RnL , Yn,1 ∈ RnP , and Vn,1 ∈
x̂n|n = Kn,1 Yn,1 + (Fn,1 − Kn,1 Hn,1 )x̄0 . (14)
RnP are specified by Xn,1 = [xTn , xTn−1 , · · · , xT1 ]T , Wn,1 =
[wnT , wn−1
T
, · · · , w1T ]T , Yn,1 = [ynT , yn−1
T
, · · · , y1T ]T , and Now, we consider the second condition (12) that implies
Vn,1 = [vn , vn−1 , · · · , v1 ] . The extended system ma-
T T T T
the estimation error xn − x̂n|n should be orthogonal to the
trices Fn,1 ∈ RnK×K , Gn,1 ∈ RnK×nL , Hn,1 ∈ extended measurement vector Yn,1 . Replacing xn with (13),
RnP ×K , and Ln,1 ∈ RnP ×nL are given as Fn,1 = x̂n|n with (14), and Yn,1 with (9), yields
T
[Fn,1 , Fn−1,1
T
, · · · , F1,1
T T
] , Hn,1 = H̄n,1 Fn,1 , and Ln,1 =
{[ ]
H̄n,1 Gn,1 with E Fn,1 x0 + Ḡn,1 Wn,1 − Kn,1 Yn,1 − Sn,1 x̄0
[ ]T }
Gn,1 = × Hn,1 x0 + Ln,1 Wn,1 + Vn,1 = 0,
Gn Fn Gn−1 · · · Fn,3 G2 Fn,2 G1
0 Gn−1 · · · Fn−1,3 G2 Fn−1,2 G1 which further leads to
.. .. .. .. .. {[
. . . . . , 0 = E (Kn,1 Hn,1 − Fn,1 )(x0 − x̄0 )
]
0 0 ··· G2 F2,2 G1 +(Kn,1 Ln,1 − Ḡn,1 )Wn,1 + Kn,1 Vn,1
0 0 ··· 0 G1 [ ]T }
( ) × Hn,1 x0 + Ln,1 Wn,1 + Vn,1
with H̄n,1 = diag Hn , Hn−1 , · · · , H1 and Fj,i =
= (Kn,1 Hn,1 − Fn,1 )P0 Hn,1
T
Fj Fj−1 · · · Fi , i 6 j.
Note that the dimensions of all the vectors and matrices +(Kn,1 Ln,1 − Ḡn,1 )Qn,1 LTn,1
in (8) and (9) grow linearly with time n. +Kn,1 Rn,1 , (15)
566
where where the equality
{ }
P0 , E (x0 − x̄0 )xT0 Ḡn,1 Qn,1 ḠTn,1 = Gn Qn GTn + Fn Ḡn−1,1 Qn−1,1 ḠTn−1,1 FnT
{ } { } { }
= E x0 xT0 − x̄0 E xT0 − E x0 x̄T0 + x̄0 x̄T0 is used. Substituting (19) and (20) into (17), and considering
{ }
= E (x0 − x̄0 )(x0 − x̄0 )T , Rn,1 is a diagonal matrix, Σn,1 can be rewritten by
{ } ( )
Qn,1 , E Wn,1 Wn,1T
= diag Qn , Qn−1 , · · · , Qn , Σn,1 = ∆n,1 + Ψn,1 ,
{ } ( )
Rn,1 , E Vn,1 Vn,1
T
= diag Rn , Rn−1 , · · · , Rn .
where
[ ]
After some arrangements, (15) becomes Rn 0
∆n,1 = ,
Kn,1 Σn,1 = Fn,1 P0 Hn,1
T
+ Ḡn,1 Qn,1 LTn,1 , 0 Σn−1,1
[ ]
Hn Un,1 HnT Hn Fn Dn−1,1
from which we get the required gain Kn,1 by dividing it Ψn,1 = T ,
Dn−1,1 FnT HnT 0
with Σn,1 from both-hand sides as
Un,1 = Fn,1 P0 Fn,1
T
+ Ḡn,1 Qn,1 ḠTn,1 . (21)
T
Kn,1 = (Fn,1 P0 Hn,1 + Ḡn,1 Qn,1 LTn,1 )Σ−1
n,1 , (16)
Using the inverse matrix lemma [14], we get the inverse of
where Σn,1 required in Kn,1 by
T
Σn,1 = Hn,1 P0 Hn,1 + Ln,1 Qn,1 LTn,1 + Rn,1 . (17) ( )
−1 −1
Σ−1 −1 −1
n,1 = ∆n,1 − ∆n,1 I + Ψn,1 ∆n,1 Ψn,1 ∆−1
n,1 , (22)
At this point, we can see that (14) with Kn,1 specified by
as ∆n,1 is invertible.
(16) provides us the MMSE estimate, where the coefficient
Another recursion needed is for the matrix Dn,1 , whose
Sn,1 = Fn,1 − Kn,1 Hn,1 governs the effect of x̄0 on the (a) (b)
estimate x̂n|n , and the counterpart of P0 is reflected by components Dn,1 and Dn,1 satisfy the following equalities:
Kn,1 . However, we still cannot answer the target equations (a) [ ]
Dn,1 = Fn,1 P0 Fn,1 T
HnT Fn,1 P0 Hn−1,1
T
,
yet, as the relationship between (14) and the KF (3)-(7) is
(b) [ ]
not clear. Considering the fact that both methods minimize Dn,1 = Ḡn,1 Qn,1 ḠTn,1 HnT Fn Ḡn−1,1 Qn−1,1 LTn−1,1 .
the estimation error variance, one naturally expects the | {z }
(b)
equivalence (or transformation) between them, which will Dn−1,1
567
where Kn,1 Hn,1 )x̄0 implies the effect of x̄0 on the estimate x̂n|n ,
whose recursion form is provided by
= (Λn,1 − Λn,1 Λn,1 )−1
(1,1) (1,1) (1,2) (2,1)
Λ̃n,1 ( )
(b) (b)
= (I + Hn Nn HnT Rn−1 )−1 , x̂n|n = I − Kn Hn Fn x̂n−1|n−1 , (32)
Nn = Un,1 − Fn Kn−1,1 Dn−1,1
T
FnT . (26) where Kn is the Kalman gain calculated at time n.
Proof 1: The proof has been given from section III-A to
Substituting (25) into (24), performing the multiplication, section III-C.
and making some arrangements, we get
Considering (14), it shows that the influence of P0 is
[ ]
Kn,1 = Jn Fn Kn−1,1 − Jn Hn Fn Kn−1,1 , reflected through the gain Kn,1 in an indirect way, while x̄0
affects x̂n|n directly. This is why the KF is more sensitive
where to the initial mean used, compared with the variance. In
( )−1
Jn = Nn HnT Rn + Hn Nn HnT . (27) the cases with imprecise initial distributions N (x̃0 , P̃0 ),
accordingly, the corrupted estimate x̂′n|n can be obtained by
(a)
Using the definition of x̂n|n , the corresponding recursion ( )
turns out to be x̂′n|n = K̃n,1 Yn,1 + Fn,1 − K̃n,1 Hn,1 x̃0 ,
(a) (a) ( (a) )
x̂n|n = Fn x̂n−1|n−1 + Jn yn − Hn Fn x̂n−1|n−1 . (28) where K̃n,1 is specified as (14) by replacing P0 with P̃0 .
Then, the the price in term of accuracy we pay for the
(b)
2) Recursion of x̂n|n : Following a similar line, the second incorrect initialization can be evaluated by
(b)
term x̂n|n can be computed by ∆xn|n = x̂′n|n − x̂n|n .
(b) Up to now, the required solutions are all found. A step
x̂n|n = Fn (Fn−1,1 − Kn−1,1 Hn−1,1 )x̄0
( ) forward is to eliminate the negative effect of x̃0 and P̃0 to get
−Jn Hn Fn Fn−1,1 − Kn−1,1 Hn−1,1 x̄0 satisfied estimates as quick as possible after filtering starts.
(b)
= (I − Jn Hn )Fn x̂n−1|n−1 . (29)
IV. I NITIALIZATION S TRATEGIES
3) Recursion of x̂n|n : Combining (28) and (29), the From Theorem 1, we see that the sensitiveness of x̂n|n
recursive implementation of x̂n|n is to x̄0 is due to the term Sn,1 x̄0 , which acts as a bias with
(a) (b)
a direct impact. To overcome this problem, we suggest two
x̂n|n = Fn (x̂n−1|n−1 + x̂n−1|n−1 ) different strategies below.
(a) (b)
+Jn (yn − Hn Fn (x̂n−1|n−1 + x̂n−1|n−1 )) A. Strategy One
= Fn x̂n−1|n−1 + Jn (yn − Hn Fn x̂n−1|n−1 ),(30) The first strategy is to minimize the MMSE with the
unbiasedness constraint (11). That is, solve the optimization
which has the same structure as the KF. That is, the term
problem
Fn x̂n−1|n−1 predicts the estimate from the previous estima- { }
tion, and the filter gain Jn corrects the predicted values. By arg min tr E [xn − x̂n|n ][xn − x̂n|n ]T
Kn,1
setting n = n+1 and using (21) and (23) again, the recursive
expression of Nn involved in Jn can be found as subject to Fn,1 = Kn,1 Hn,1 , where tr means the trace
operation and x̂n|n is specified by (14) that can be further
Nn = Fn Nn−1 FnT + Gn Qn GTn − Fn Jn−1 Hn−1 Nn−1 FnT , rewritten as x̂n|n = Kn,1 Yn,1 due to the constraint. This
(31) problem has been solved in our recent paper [16], leading to
with initial value N1 = F1 P0 F1T + G1 Q1 GT1 . a linear optimal unbiased filter.
Comparing (27), (30) and (31) with the KF, we complete
our proof that the batch estimate (14) is equivalent to the KF B. Strategy Two
with the following notational transformations: The second strategy is that we remove Sn,1 x̄0 artificially,
and get a new linear estimator as
Jn ⇔ Kn , Nn ⇔ Pn|n−1 .
x̂n|n = K̄n,1 Yn,1 . (33)
D. Effect of Initial States
Here, the compensation used for the neglected term is to
Now, we summarize the main results and address the
recompute a new gain K̄n,1 by solving
questions by the following theorem. { }
Thorem 1: Given the linear state-space model (1) and (2) arg min tr E [xn − x̂n|n ][xn − x̂n|n ]T , (34)
K̄n,1
with known initial distribution p(x0 ) ∼ N (x̄0 , P0 ), and
uncorrelated white Gaussian noises wn ∼ N (0, Qn ) and Substituting xn with (13) and x̂n|n with (33), we get
{[
vn ∼ N (0, Rn ), the traditional KF ( can be equivalently
) arg min tr E (K̄n,1 Hn,1 − Fn,1 )x0
transformed to x̂n|n = Kn,1 Yn,1 + Fn,1 − Kn,1 Hn,1 x̄0 , K̄n,1
(b) ][ ]T }
where Kn,1 is specified by (16). The term x̂n|n = (Fn,1 − +(K̄n,1 Ln,1 − Ḡn,1 )Wn,1 + K̄n,1 Vn,1 ··· .
568
Comparing with the strategy one, the first term K̄n,1 Hn,1 − 180 100
Fn,1 cannot be omitted since we do not have the constraint (a) Ideal KF (b) Ideal KF
160 90
K̄n,1 Hn,1 = Fn,1 . On the other hand, comparing it with the KF−W KF−W
SO 80 SO
former minimization problem associated with the estimator 140
(14), 70
120
{[ 60
RMSEs
arg min tr E (Kn,1 Hn,1 − Fn,1 )(x0 − x̄0 )
RMSEs
100
Kn,1 50
][ ]T } 80
+(Kn,1 Ln,1 − Ḡn,1 )Wn,1 + K̄n,1 Vn,1 ··· , 40
60
30
one can see that in the KF, the gain Kn,1 is designed for 40 20
x0 − x̄0 as x̄0 is assumed to be available, while K̄n,1 is
20 10
designed for x0 itself in the proposed method (33). With
this difference, the direct influence of x̄0 will be changed 0
0 20 40 60
0
0 20 40 60
to be indirect (same as P0 ), and better robustness is thus Time Time
achieved.
Using the lemma given in [16], the solution to (34) can Fig. 1. RMSEs of different algorithms with x̃0 and P̃0 based on 100
Monte-Carlo runs: (a) the second state and (b) the third state, where “SO”
be obtained by is the abbreviation of strategy one.
T
K̄n,1 = (Fn,1 Θx0 Hn,1 + Ḡn,1 Qn,1 LTn,1 )Σ̄−1
n,1 , (35)
where [P0,1 , P0,2 , P0,3 ] with P0,1 = [5 , 0 , 0]T , P0,2 = [0 , 5 , 0]T ,
T and P0,3 = [0 , 0 , 5]T .
Σ̄n,1 = Hn,1 Θx0 Hn,1 + Ln,1 Qn,1 LTn,1 + Rn,1 , (36)
As can be seen, the first state is uncorrelated with the other
Θx0 = E{x0 xT0 } = x̄0 x̄T0 + P0 . (37) two states, and thus is generally considered as a drafting
Since K̄n,1 has the same structure as Kn,1 , we can get its source in a resonator model. Therefore, we mainly focus on
recursion form by (28) easily. Here, however, more attentions the estimation of the second and third states. Considering that
should be paid on the initial values of K̄n,1 . Using the x̄0 and P0 are not precisely known and we use a poor initial
definitions of Hn,1 , Ln,1 and Ḡn,1 with n = 1, we have guess x̃0 = [8 , 100 , 80] and P̃0 = [P̃0,1 , P̃0,2 , P̃0,3 ] with
( P̃0,1 = [5 0 0]T , P̃0,2 = [0 10 0]T , and P̃0,3 = [0 0 10]T to
K̄1,1 = (F1 Θx0 F1T + G1 Q1 GT1 )H1T H1 (F1 Θx0 F1T test both the strategy one (SO) and KF (denoted as KF-W)
)−1 to demonstrate the trade-off.
+G1 Q1 GT1 )H1T + R1 , (38)
The average root mean square errors (RMSEs) of different
where Θx0 = x̃0 x̃T0 + P̃0 as the true value (37) is not known. methods based on 100 Monte-Carlo runs are provided in
Then, the first estimate x̂1|1 can be calculated by (33) using Fig. 1, where the ideal KF with correct initial distribution is
Yn,1 = y1 . Once x̂1|1 is obtained, the subsequent estimates also employed as a benchmark. It shows that, considerable
can be recursively computed with (28) starting from N1 = improvements on accuracy are achieved by the SO, especially
F1 Θx0 F1T + G1 Q1 GT1 . in the initial estimation phase, while the Kalman estimates
converge to the true states at a slower rate, due to the long
V. SIMULATIONS lasting negative effect of poor x̃0 . One also immediately
In this section, we test the proposed initialization methods notices that the difference between SO and the ideal KF is
(mainly on the strategy two since the performance of strategy minor, which can be neglected completely in practical ap-
one has been demonstrated in [16]) using a randomly drifting plications. This provides us another proof from the opposite
stochastic resonator [17] as an example, which is specified angle that the SO can be used in place of the KF when
by (1) with Gn = I and initialization is uncertain.
By introducing an coefficient as q P̃0 , we next compare SO
1 0 0
with KF using different initial covariances, where q takes
Fn = 0 cos(αn ∆n ) sin(αn ∆n )/αn ,
value from {5 , 30 , 80}. The main reason for this design
0 −αn sin(αn ∆n ) cos(αn ∆n )
is that the initial performance of KF with incorrect initial
where ∆n = 0.5s is the sampling interval and αn takes states can be improved by amplifying the corresponding
value from {0.05 , 0.06} according to a Markov chain with covariance, and we would like to see whether this rule is
a transition probability matrix whose diagonal elements are valid for the proposed algorithm or not. Based on 100 Mento-
0.9. The noise in resonator is zero mean white Gaussian with Carlo runs, the average RMSEs with respect to different
2 2 2
variance δw1 = δw2 = δw3 = 1. q are shown in Fig. 2. As expected, the KF is sensitive
The measurement equation is given by (2) with Hn = to the q used, and the negative effect of x̃0 in the KF
[1 1 0], where the measurement noise vn is white Gaussian decreases with the increase of q considerably. For the SO,
with zero mean and variance δv2 = 100. The process starts although similar phenomenon can be observed in Fig. 2(b),
from p(x0 ) ∼ N (x̄0 , P0 ) with x̄0 = [8 , −2 , −2]T and P0 = it is not significant. Comparing it with the KF, we can see
569
100 70
[10] M. Linderoth, K. Soltesz, A. Robwetsson, and R. Johansson, “Initial-
KF, q=5 KF, q=5 ization of the Kalman filter without assumputations on the initial state,
90
KF, q=30 KF, q=30 IEEE Int. Conf. on Robotics and Automation, 2011, pp. 4992-4997.
60
KF, q=80 KF, q=80
[11] J. Hyland, “An iterated-extended Kalman filter algorithm for tracking
80
SO, q=80
surface and sub-surface targets,OCEANS 02 MTS/IEEE, vol. 3, pp.
SO, q=80
70 50 1283C1290, 2002.
SO, q=30 SO, q=30
[12] D. Liu and X. Liu, “Optimal state estimation without the requirement
60 SO, q=5 SO, q=5
40 of a prior statistics information of the initial state, IEEE Trans.
RMSEs
50 RMSEs Automat. Contr., vol. 39, no. 10, pp. 2087-2091, Oct. 1994.
30 [13] H. Stark and J. W. Woods, Probability, Random Processes, and
40 Estimation Theory for Engineers, Prentice Hall, 1994.
30 [14] G. H. Golub and G. F. Van Loan, Matrix Computation, 3rd Ed. ed.
20
Baltimore, MD: Johns Hopkins Univ. Press, 1996.
20 (a) (b) [15] F. L. Lewis, Optimal Control, John Wiley & Sons, Inc., 1986.
10 [16] S.Y. Zhao, B. Huang, and F. Liu, “Linear optimal unbiased filter for
10
time-variant systems without apriori information on initial conditions,”
0 0 IEEE Trans. Autom. Control, doi: 10.1109/ TAC.2016.2557999, 2016.
0 5 10 15 20 0 5 10 15 20
[17] S. Särkkä and A. Nummenmaa, “Recursive noise adaptive Kalman
Time Time
filtering by variational Bayesian approximations,” IEEE Trans. Auto.
Contr., vol. 54, no. 3, pp. 596-600, Mar. 2009.
Fig. 2. Average RMSEs of different algorithms with respect to q: (a) the
second state and (b) the third state.
VI. CONCLUSIONS
In this paper, the effect of initial states used in the
KF is analyzed. It is shown that comparing to the initial
variance P0 , the initial mean x̄0 introduces a bias in the state
estimation in a more direct way. Two initialization strategies
are introduced. One is to minimize the error variance with
the unbiasedness constraint, and the other is to remove the
term with respect to the initial mean artificially and modify
the gain optimally. Using a drafting stochastic resonator as
an example, we demonstrate that the proposed approach can
be used as an alternative to the KF when there is uncertainty
in the initial conditions.
R EFERENCES
[1] D. Simon, Optimal State Estimation: Kalman, H∞ , and Nonlinear
Approaches, Hoboken, NJ: J. Wiley & Sons, 2006.
[2] H. Cox, “On the estimation of state variables and parameters for noisy
dynamic systems, IEEE Trans. Autom. Contr., vol. 9, no. 1, pp. 5C12,
Jan. 1964.
[3] S. J. Julier and J. K. Uhlmann, “A new extension of the Kalman filter
to nonlinear systems, in Proc. of SPIE, vol. 3068, pp. 182-193, 1997.
[4] B. Ait-El-Fquih and F. Desbouvries, “Kalman Filtering in Triplet
Markov Chains, IEEE Trans. Signal Process., vol. 54, no. 8, pp.
2957C2963, Aug. 2006.
[5] Y. Becis-Aubry, M. Boutayeb, and M. Darouach, “State estimation in
the presence of bounded disturbances, Automatica, vol. 44, no. 7, pp.
1867C1873, Jul. 2008.
[6] L. B. Weiner, “Kalman filter initialization with large initial uncertainty
and strong measurement nonlinearity, in Proceedings of the Region 3
Conference and Exhibit, Huntsville, AL; United States, pp. 150-151,
1981.
[7] U. Frese, B. Baeuml, G. Schreiber, and I. Schaefer, et al, “Off-the-shelf
vision for a robotic ball catcher, In Proc. of the IEEE/RSJ Int. Conf. on
Intelligent Robots and Systems, Oktober, Maui, pp. 1623-1629, 2011.
[8] B.D.O. Anderson and J.B. Moore, Optimal Filtering, Prenrice-Hall,
1979.
[9] J. M. Roberts, D. J. Mills, D. Charnley, and C. J. Harris, “Improved
Kalman filter initialisation using neurifuzzy estimation, Fourth Int.
Conf. on Artificial Neural Networks, pp. 26C28, 1995.
570