Professional Documents
Culture Documents
Nonlinear Time-Series Analysis: Ulrich Parlitz
Nonlinear Time-Series Analysis: Ulrich Parlitz
Nonlinear Time-Series Analysis: Ulrich Parlitz
ANALYSIS
Ulrich Parlitz
8.1 INTRODUCTION
Nonlinearity introduces a wealth of genuine physical phenomena. Among them,
chaotic dynamics is the most puzzling and intriguing one. This type of irregular
dynamics calls for new methods to cope with and has already set out for appli-
cations. The investigations on chaotic dynamics have shed a new light on the
notions of predictability and determinism, because the deterministic dynamics
of even simple systems may be notoriously difficult to predict and even impossi-
ble in the long run. Such systems abound in nature, and examples can be drawn
from almost any scientific discipline (see, for instance, [1, 2, 3, 4, 5, 6, 7, 8, 9)).
In mathematical models of dynamical systems the dynamics is described
in their state space, whose (integer) dimension is given by the number of the
dependent variables of the model. In experiments, however, usually just one
variable is measured as a function of time, and the state space is not normally
known. How, then, to arrive at the attractor that may characterize the system?
This gap between the theoretical notions and observable quantities was filled
in 1980 when Packard, Crutchfield, Farmer, and Shaw [10) published their fun-
damental paper "Geometry from a time series" where for the first time state
209
J. A. K. Suykens et al. (eds.), Nonlinear Modeling
© Springer Science+Business Media Dordrecht 1998
210 NONLINEAR MODELING
F: M --+ Rd
X t--+ Y = F(x) = (s(t), s(t + tl), ... , s(t + (d - l)tl)
that maps a state x from the original state space M to a point y in a recon-
structed state space R d , where d is the embedding dimension and tl gives the
delay time (or lag) used. Figure 8.1 shows a visualization ofthis construction.
Takens [11] proved that for d ~ 2m + 1 it is a generic property of F to be an
embedding of Min Rd, i.e., F: M --+ F(M) C Rd is a (C 2 -) diffeomorphism.
Generic means that the subset of pairs (h, tl) which yield an embedding is
an open and dense subset in the set of all pairs (h, tl). This theorem was
generalized by Sauer, Yorke and Casdagli [16, 17] in two ways. Firstly, they
replaced the condition d ~ 2m + 1 by d > 2do(A) where do(A) denotes the
capacity (or: box-counting) dimension of the attractor A eM. This is a great
progress for experimental systems that possess a low-dimensional attractor
(e.g., do(A) < 5) in a very high-dimensional space (e.g., m = 100). In this case,
the theorem of Takens guarantees only for very large embedding dimensions d
(e.g., d ~ 201) the existence of a diffeomorphic equivalence, whereas with the
condition of Sauer et al. a much smaller d will suffice (e.g., d> 10). The second
improvement of Takens' embedding theorem by Sauer et al. refers to the term
"generic" that was replaced by "prevalent" which essentially means "almost
NONLINEAR TIME-SERIES ANALYSIS 211
measurement \ h
h:M IR ~
ob ervable =h(x )
all (h, tl)" will yield an embedding. This modification of the original theorem
was necessary, because examples of open and dense (=generic) sets have been
found that are rather "thin" [16, 17]. Furthermore, Sauer et al. showed that
for dimension estimation an embedding dimension d > do(A) suffices. In this
case the delay coordinates map F is, in general, not one-to-one, but the points
where trajectories intersect are negligible for dimension calculations. More
details about the reconstruction of states, in particular in the presence of noise,
may be found in Refs. [18, 19].
If the data are measured with a high sampling rate so-called Broomhead-
King-coordinates [20, 21] may be advantageous. With this method a very
high dimensional reconstruction is used and then a new coordinate system
is introduced where the origin is shifted to the center of mass of the recon-
structed states and the axes are given by the (dominant) principal components
of the distribution of points (states). This new coordinate system is based on
a Karhunen-Loeve transformation that may be computed by a singular-value
decomposition. A discussion of the advantages (e.g., noise reduction) and disad-
vantages of this "post-processing" of the reconstructed states may, for example,
be found in Ref. [22].
For time series that consist of a sequence of sharp spikes (e.g. from firing
neurons) delay embedding may lead to very inhomogeneous sets of points in the
reconstructed state space that are difficult to analyse. As an alternative one
may use in this case the time intervals between the spikes as components for (re)
212 NONLINEAR MODELING
constructed state vectors. More details about this state space reconstruction
from interspike intervals may be found in [23, 24, 25].
Another topic that is worth mentioning although we will not discuss it in
more detail in this review are embedding theorems for driven systems [26].
These remarkable results pave the way for extending the methods to be dis-
cussed in the following to input-output systems with deterministic or stochastic
input.
In the following we will assume that the scalar signal s(t) = h(x(t)) is
sampled with a sampling time ts. The resulting time series {sn} with sn =
s(nts) is used to reconstruct the states
(8.1)
for n = 1, ... , N. The symbol l denotes the delay time or lag in units of the
sampling time (i.e., tl = its). As an example we will use experimental data
from a chaotic Colpitts Oscillator [27].
If the internal dynamics of the transistor is neglected this electronic oscillator
can be described by three ordinary differential equation:
C dVCl h - j3g(VC2 )
l~
C dVC2 VE + VC2
2~ - R - h - g(VC2) (8.2)
Ldh
dt
NONLINEAR TIME-SERIES ANALYSIS 213
where
Is V
g(V) = t3 exp( Vr )'
. . ... IR.d... .
.-... .......
...
. :.-. .. ..
.. :... ............ ..-.::-.. .
.-
Figure 8.3 Sketch of three reconstructions with different delay times: (a) delay too small,
(b) delay o. k., (c) delay too large.
Figure 8.4 (a) Chaotic attractor of the Colpitts oscillator (8.2) in the original state space
(VCl' VCl, h). (b )-( d) Three reconstructions of the attractor shown in (a) with different
delay times tl' The two black dots indicate the locations of the same states on the different
attractors.
with a = 1.4 and b = 0.3. Figure 8.5a shows a reconstruction with a time delay
of tl = 1 that provides a perfect unfolding of the underlying attractor.1 The
reconstruction with delay ~ = 2, however , that is shown in Fig. 8.5b was not
successful, because due to additional folding there are "crossings" on the set
of points where the induced flow in the reconstruction space is not one-to-one
anymore. As an example, one of these crossing points is encircled and the
locations of (pre-) images of both branches are given by the + (-) signs.
(a)
o o
- 1
\
Figure 8.5 Reconstruction of data from the Henon map with (a) delay tl = 1 and (b)
delay tl = 2. In the latter case , the reconstruction is not one-to-one anymore .
These examples show that a proper choice of tl is crucial for any further
investigation of the data.
1 Notethat the Henon map possesses the unusual feature that its "original" state space
{(Xn,Yn)} coincides with its delay reconstruction space {(xn , xn-d}.
216 NONLINEAR MODELING
FIR filter. For finite impulse response (FIR) filters one can show that those
filters which destroy the embedding constitute linear subspaces in the space of
all FIR-filters. Therefore, with probability one, FIR-filters preserve the embed-
ding. Nevertheless, one may find examples where the application of a FIR-filter
makes subsequent estimations of Lyapunov exponents or dimensions more dif-
ficult as compared with the original "noisy" time series. In these cases the
influence of the FIR filter on the geometry of the reconstructed is essentially
twofold: (i) Due to the filter elements of the time series are connected that
are not close neighbours in time. This leads to (over) folding of the attractor
(similar to a choice of too large a delay time). (ii) On the other hand, the
FIR filter averages the values of the time series and thus the components of
subsequently generated state vectors are more or less similar. This leads to
2 As an example the reader may verify that the embedding of a T-periodic sinusoidal signal
in a d-dimensional reconstruction space is optimal (Le. a circle) for dtl = kT/2 where k may
be any integer that is not a multiple of d.
NONLINEAR TIME-SERIES ANALYSIS 217
IIR filter. Infinite impulse response (IIR) filters may also destroy the dif-
feomorphic equivalence between the original state space and the reconstructed
state space. In this case, the original system and the filter may be considered
as a single, extended system, and the output of this system depends very much
on its filter subsystem. To illustrate this problem consider a simple discrete
low-pass filter
U n +l = QUn + Xn (0 < Q < 1)
that is driven by some other discrete dynamical system (e.g. the Henon map
(8.3)). The Lyapunov spectrum of the combined system consists of the Lya-
punov exponents {AI, ... , Am} of the driving system and the exponent A, = In Q
of the filter. By increasing Q -t 1 one may now shift the exponent A, into the
range of the spectrum {AI, ... , Am} and modify in this way the Lyapunov di-
mension 3 of the complete system. If the filter is not sufficiently dissipative it
may thus increase the dimension of the time series. A rigorous mathematical
treatment of this problem may be found in Ref. [47].
3For a definition of the Lyapunov dimension based of the Kaplan-Yorke conjecture see for
example [1, 3].
218 NONLINEAR MODELING
example we will use the widely used hypothesis: "The data are (just) coloured
noise generated by a linear stochastic process". In the first step, the data are
modified in such a way that all structure except for the assumed properties will
be destroyed or "randomized". In the case of the above cited nullhypothesis
this may be done, for example, by Fourier transforming the data and substi-
tuting the phases of the complex Fourier transform by random numbers. The
power spectrum (or equivalently, the autocorrelation function) is not affected
by this modification. After transforming back into the time domain we thus
obtain a new time series with the same power spectrum. If the original data
are just coloured noise we do not have destroyed any underlying structure, and
subsequent estimations of dimensions, Lyapunov exponents, prediction errors,
etc. should give the same results for the original time series and the surrogate
data. If, however, the analysis yields significant differences, then our original
time series was more than "just noise". To improve the statistical robustness
of this test one usually generates several surrogate data sets and compares the
mean values of the quantities obtained in the course of the subsequent data
analysis with the corresponding value from the original time series. Further-
more, it is necessary to take into account a possibly performed static nonlinear
transformation of the data that would distort the Gaussian distribution of the
assumed coloured noise (Algorithm II in Ref. [52], see also Ref. [62]). Exam-
ples for applications of the surrogate data method will be given in the following
sections.
thermore, good (local) models are a prerequisite for many other methods of
nonlinear time series analysis and nonlinear dynamics.
(8) I (b)
0.1
3
..
.
Q.
~ .5! 2
tn
J'10 - 4 III
10- & .s'" I
10- 6
10- 7 o
0 .01 01 0 .01 0.1
r r
Figure 8.6 (a) Correlation sum C3 (r) vs. the distance r of neighbouring points for ex-
perimental data from a chaotic Colpitts Oscillator. (b) Local slope of the curve shown in
(a). The plateau indicates a dimension D2 between 1.8 and 2. The dotted curve is from
surrogate data to detect nonlinearity.
of each reference point yn and the radius rn = r(m) of this cloud of m points
are determined. For the limit m/N --+ 0 one obtains an approximation of the
logN
information dimension Dl = - N .
-it Ln=l log rn
To investigate the scaling in the limit m/N --+ 0 one can decrease the number
of neighbours m or increase the number of data points N.
Using a fast search algorithm [80] for the nearest neighbour ynn of the re-
constructed state yn one may estimate the information dimension in terms of
the scaling behaviour
of the quantity
(8.4)
for m = 1 and N --+ 00. Figure 8.7 shows an application of this method to the
experimental data from the Colpitts oscillator.
Another very suitable method for dimension analysis is the determination of
the pointwise dimensions and their average [81, 82]. A pointwise dimension D
at a point of an attractor can be calculated by exploring the scaling behaviour
of the "mass" of the attractor in the vicinity of the point, i.e., the number of
NONLINEAR TIME-SERIES ANALYSIS 221
I·. I.
.. I · • ·1
-1
q
- 1.5
points within a certain distance r. The number N(r) of points in this ball of
radius r should scale like N(r) ex: rD for r --+ o. When this is done for all points
of the attractor, a histogram of pointwise dimensions is obtained. The average
dimension and the standard deviation are numbers then to characterize the
attractor. However, the histogram itself can be considered as a characteristic
of the attractor.
When the attractor is sufficiently low-dimensional, its (fractal) dimension start-
ing from low-dimensional embedding spaces (d = 1,2,3, ... ) will level off at
some value of d and does not increase further. This value then is said to deter-
mine the relevant number of (nonlinear) degrees of freedom, i.e., the number
of dependent variables, of the dynamical system investigated.
More details about dimension estimation methods, their possible pitfalls, ex-
tensions, and many references are given in the review articles [15, 83, 84] and
in [12] .
• using neighbouring states from a shell (rmin < r < rmax) instead of a ball
(r < rmax) to reduce the influence of noise (see, e.g., Refs. [92,96,97,99])
• using the embedding dimension d for selecting the neighbours and a small
local dimension dL for performing the fit and computing the resulting dL
Lyapunovexponents (see, e.g., [92, 101, 102]).
The latter two methods have been devised to avoid so-called spurious Lyapunov
exponents, that occur when the dimension d of the reconstructed state space
is larger than the dimension m of the original state space. Spurious Lyapunov
exponents do not depend on the dynamics but on the approximation scheme,
the noise contained in the data and in some cases on the local curvature of
the attractor. In principle they can take any values. To identify spurious Lya-
punov exponents one can estimate the local thickness of the attractor along the
directions associated with the different Lyapunovexponents [101, 102] or com-
pare the exponents obtained with those computed for the time reversed series
[98,105], because spurious exponents correspond to directions where the attrac-
tor is very thin and because they do not change their signs upon time reversal
(in contrast to the true exponents). The latter method, however, works only
for data of very high quality that enable also a correct estimation of negative
Lyapunovexponents. An investigation of the data requirements for Jacobian-
based methods may be found in Refs. [108, 109].
until it exceeds the range where a linear approximation of the flow at yn holds.
Now the neighbouring orbit {zn} has to be replaced by another neighbouring
orbit that is closer to the reference orbit and whose initial value lies on or
near the line from the current reference state to the last point of the previ-
ous neighbouring orbit in order to preserve the direction corresponding to the
largest Lyapunov exponent. Criteria for the replacement threshold and other
details of the algorithm are given in Ref. [90], including a FORTRAN program.
In principle, it is possible to use this strategy also for computing the second
largest Lyapunov exponent [90], but this turns out to be quite difficult. On the
other hand, it suffices in many cases to establish the existence of at least one
positive Lyapunov exponent. Due to its robustness the Wolf-algorithm is often
used for the analysis of experimental data (see, for example, Refs. [110, 111]).
Another direct method, which is even simpler in terms of less free parameters,
was proposed by Sato et al. [112] and Kurths and Herzel [113]. Very similar to
the Wolf-algorithm, the average exponential growth of the distance of neigh-
bouring orbits is studied on a logarithmic scale, this time via the prediction
error
(8.5)
where ynn is the nearest neighbour of yn. The dependence of the prediction
error p(k) on the number of time steps k may be divided into three phases.
Phase I is the transient where the neighbouring orbit converges to the direction
corresponding to the largest Lyapunov exponent. During phase II the distance
growths exponentially with exp(AItsk) until it exceeds the range of validity of
the linear approximation of the flow around the reference orbit {yn+k}. Then
phase III begins where the distance increases slower than exponentially until it
decreases again due to foldings in the state space. If phase II is sufficiently long,
a linear segment with slope Al appears in the p(k) vs. k diagram. This not only
allows an estimation of the largest Lyapunov exponent Al but also provides a
direct verification of exponential growth of distances to distinguish determin-
istic chaos from stochastic processes where a non-exponential separation of
trajectories occurs [114]. Figure 8.8 gives an example for the determination of
the largest Lyapunov exponent Al of the Colpitts data by this method. The
dotted curve is from surrogate data to detect determinism. The error bars are
for ±5 standard deviations.
Several authors have investigated modifications of this basic algorithm [33, 115,
116]. In particular, those states must be excluded from the search for the near-
est neighbour that are close together in time. Such states belong to the same
segment of the trajectory and thus cannot converge or diverge exponentially. A
possible criterion for a sufficiently large separation in time is the first minimum
or zero of the autocorrelation function of the data set.
224 NONLINEAR MODELING
1.5 l I I 1 I I I I I I I {I I I I I I I , I
0.5
II III
10 20 30
k
Figure8.8 Prediction error p for experimental Colpitts data vs. the number of time steps
k. The slope of the solid line in the intermediate range of k gives the largest Lyapunov
exponent Al = 0.16.
those of the driven system [130, 131, 132, 133, 134]. To illustrate the relevance
of this phenomenon for time series analysis we shall consider an example where
a Rossler system
Xl = 2 + Xl(X2 - 4)
X2 -Xl - x3
X3 = x2 + 0.45x3
Yl = -a(Yl - Y2)
Y2 = ru - Y2 - UY3
Y3 UY2 - bY3
1 N
rnm = snm - -N.i....J
'"""' snm (m -- 1, ... , M) (8.6)
n=l
The N x M -matrix U and the M x M matrix V are both orthogonal and the
columns of V constitute the EOFs or basic modes. The singular values Wi are
ordered with respect to their size starting with the largest. The squares of
the Wi give the energy (power) of the corresponding mode and are a measure
of the statistical significance of the corresponding basis vector (pattern) for
describing the STTS. If only a few singular values Wi are significantly larger
4Note that this statement is valid not only for modes that are computed using the KL-
transformation, but for any linear decomposition of the STTS.
NONLINEAR TIME-SERIES ANALYSIS 227
than the remaining Wi, the STTS can be described by the projection coefficients
with respect to these dominant spatial modes. In this case the images {sn}
constitute N vectors in JRM that span only a low dimensional linear subspace.
The columns of V corresponding to the large singular values constitute a basis
of this subspace. In order to show that such a low dimensional subspace does
(not) exists for a STTS it suffices therefore to plot the distribution of singular
values.
x
nm
= (n n n n-JL n-JL
sm-IK,···,sm,···,sm+IK,···,sm_IK,···,sm
n-JL )
, ... ,sm+IK
where I is the number of spatial neighbours, J is the number of temporal
neighbours (in the past), K is the spatial shift which has a similar meaning
as the time delay L (time lag) known from the delay embedding of scalar
time series. The dimension d of the state vector xnm E R d is given by d =
(J + 1)(1 + 21). This dimension can be reduced if the STTS stems from a
dynamical system that possesses additional spatial symmetries. For the one
228 NONLINEAR MODELING
dimensional case we may for example assume that the system is invariant with
respect to spatial reflection. The general form of the state vector may then be
written as
X nm = (sn m' g(sn
m-K, snm+K ) , ... , g(sn m+IK ) , ... , Sn-JL
m-IK' sn m , ... ,
( n-JL n-JL))
... , 9 Sm-IK' Sm+IK
with
= g(V, u).
g(u, v)
Possible choices for 9 are for example g(u,v) = .;uv or g(u,v) = ut vthat is
used in following.
The dimension d = (J + 1)(1 + 1) of the symmetrically reconstructed states
xnm E Rd is approximately about a factor of two smaller. For time series
with two or three spatial degrees of freedom this dimension reduction due to
symmetries is even larger (and thus more important).
Spatial boundaries of the experiment or of the STTS can be taken into
account by constructing their states separately but in a way analogous to that
described above. In principle all states xnm containing boundary values of the
STTS have to be divided into different classes where each class is characterized
by the distance of the center index m from the boundary. For predicting or
modelling only states of the relevant class have then to be compared. This
decomposition into classes can easily be implemented by extending the STTS
in its spatial direction with numbers -c, -2c, -3c, ... to the left and c, 2c, 3c
to the right. The parameter c has to be chosen much larger than the largest
value of the STTS. Using this construction all states of a given class are located
in different subspaces of the reconstruction space Rd. Of course, this method
is not the most efficient implementation but it allows to introduce boundaries
without changing the main part of the algorithm. The effective dimension of the
reconstructed boundary states is lower than d. Therefore, in principle different
values of the reconstruction parameters I, J, K and L could (and should) be
used for the boundary to improve the reconstruction (and thus the results of
any subsequent application of it). However, to keep the presentation of methods
and results as comprehensible as possible we will use in the following the same
reconstruction parameters inside the STTS and at its boundaries where the
reconstruction method explained above is used.
2d
F(x) =L Ck<Pk (x) (B.B)
k=l
where the functions <Pk are radial basis functions <Pk(X) = Jr 2 + Ilx nkmk - xll 2
centered at the 2d neigbouring states x nkmk • The coefficients Ck are computed
from the known future values S~kk+T (corresponding to the neighboring states)
by solving a linear set of equations. Since F is a local approximation of the
dynamics f the prediction of the desired future value S~+T of the STTS is given
by p~+T = F(xnm).
For monitoring the accuracy of this forecasting scheme the prediction error
En(r) at time n
(B.9)
N
E(r) = ~L En(r) (B.I0)
n=l
may be computed. Of course, locally constant or locally linear maps may also
be used to approximate the dynamics f. Furthermore, predictions over longer
periods of time (r > 1) can be computed iteratively by concatenating steps
with r = 1 or as a single step.
(B.11)
with
f(s) = as(1 - s).
and m = 2, ... , M - 1 and n = 1, ... , N. The boundary conditions were S1 =
0.5 = sM for all times n = 1, ... , N.
Since the structure of the dynamical equation (B.11) is exactly the same as
that of the modelling ansatz it may be no surprise that for this example very
accurate predictions are possible. Table B.l shows the average prediction error
E(r) vs. the number of temporal (I) and spatial (J) neighbours using the
symmetrical reconstruction scheme with L = 1, K = 1 and r = 1. As expected
there is a pronounced minimum for I = 0 and J = 1.
230 NONLINEAR MODELING
Table 8.1 Average prediction error E(r) (Eq. (8.10)) as a function of the number of
spatial neighbours I (rows) and temporal neighbours J in the past (columns) for K = 1 =
L.
0 1 2 3
0 1.219443 0.000001 0.000014 0.000053
1 0.218676 0.000003 0.000031 0.000102
2 0.048164 0.000007 0.000050 0.000136
3 0.028297 0.000015 0.000083 0.000206
4 0.024331 0.000025 0.000128 0.000353
5 0.024256 0.000049 0.000211 0.000654
The spatio-temporal evolution of the training set is shown in Fig. 8.9a. Fig-
ure 8.9b shows the following 30 steps of the STTS that have been forecasted
using iterated predictions. The resulting predicted STTS is given in Fig. 8.9c
and Fig. 8.9d shows the difference between the original data (Fig. 8.9b) and
the predicted values (Fig. 8.9d). In this case we used the predicted values p~+l
to replace the original elements s~+1 after each step and the prediction error
En was found to grow exponentially.
Although this example may be viewed as trivial it may serve as a motiva-
tion for the approach and demonstrates dearly the power of the local method.
For describing and predicting this STTS using global mode decomposition like
EOFs many modes would be necessary as can be seen from the distribution of
squared singular values shown in Fig. 8.10.
If the general (nonsymmetrical) reconstruction is used instead of the sym-
metric embedding all prediction errors are slightly larger. The difference, how-
ever, is not large due to the low dimension of the reconstruction in both cases.
The numerical examples show that the local reconstruction of states is a pow-
erful method for predicting spatio-temporal time series. It may also serve as
a starting point for deriving a mathematical model of the underlying dynam-
ics and subsequent bifurcation analysis. The scheme discussed in this paper
may be generalized in different directions. If the process generating the STTS
is not spatially homogeneous one may simply add to the dimension d of the
reconstruction space the number ds of spatial dimensions of the problem (Le.
ds = 1,2 or 3) and work then in the enlarged d + ds dimensional space. Of
course, in that case longer training sets will be necessary in general. Another
modification of the scheme proposed concerns the selection of spatio-temporal
neighbours of the center element s~ for reconstructing the local state xnm. At
this point one may take into account that any physical information spreads
with some maximum speed. Instead of using a rectangular region of the matrix
S a triangle (Le. a "light cone") may be more efficient for reconstructing local
states. This feature of locality may also be viewed as a motivation for the
concept of local states and is probably important for a rigorous mathematical
NONLINEAR TIME-SERIES ANALYSIS 231
(a) 100
1.0
09
60 06
07
60 0.6
QI
E 0.6
E=
40 04
0.3
20 0.2
0.1
0 0.0
0 20 40 60 60 100
Space
(b) 130
1.0
125 0.6
Q) 120 0.6
E 115 0.4
E= 110
0.2
105
100 00
0 20 40 60 60 100
Space
(C) 130
10
125 0.6
Q) 120 0.6
E 115 04
E= 110
02
105
100 00
0 20 40 60 60 100
Space
(d) 130
10
125 0.6
Q) 120 0.6
E 115 04
E= 110
0.2
105
100 0.0
0 20 40 60 60 100
Space
Figure 8.9 Spatio-temporal time series generated by the coupled logistic maps with M=
100. (a) Training set of length N = 100. (b) Original time series to be predicted. (c)
232 NONLINEAR MODELING
200
180
160
140
120
100
80
60
40
20
0
5 10 15 20 25 30 35 40 45
Mode
Figure 8.10 Squared singular value spectrum of the STTS shown in Fig. a.ga.
Acknowledgments
The author thanks the organizers of the "International Workshop on Ad-
vanced Black-Box Techniques for Nonlinear Modelling", Leuven, July 8-10,
1998 for their kind invitation and support. Thanks go also to my coworkers
for fruitful discussion and exchange of ideas, in particular to the Nonlinear
Dynamics Group of the Drittes Physikalisches Institut, Gottingen. Support by
the German Ministry for Science and Education (BMBF) is gratefully acknowl-
edged.
References
[3] Schuster, H. G. (1988). Deterministic Chaos. 2nd ed., Weinheim, VHC Pub-
lishers.
[4] Berge, P., Y. Pomeau, C. Vidal (1984). Order within Chaos: Towards a Deter-
ministic Approach to Turbulence. New York, John Wiley and Sons.
[5] Moon, F. C. (1992). Chaotic and Fractal Dynamics. New York, John Wiley and
Sons.
[6] Gaponov-Grekhov, A. V., M. I. Rabinovich (1992). Nonlinearities in Action -
Oscillations, Chaos, Order, Fractals. Berlin, Springer.
[7] Lauterborn, W., J. Holzfuss (1991). Acoustic chaos. Int. J. Bifurcation and
Chaos, 1, pp.13-26.
[8] Lauterborn, W., T. Kurz, U. Parlitz (1997). Experimental Nonlinear Physics.
Int. J. Bifurcation and Chaos, 7, pp.2003-2033.
[9] Lauterborn, W., U. Parlitz (1988). Methods of chaos physics and their appli-
cation to acoustics. J. Acoust. Soc. Am., 84, pp.1975-1993.
[10] Packard, N.H., J.P. Crutchfield, J.D. Farmer, R.S. Shaw (1980). Geometry from
a time series. Phys. Rev. Lett., 45, pp.712-716.
[11] Takens, F. (1981). Detecting strange attractors in turbulence. in Dynamical
Systems and Turbulence, eds. Rand, D.A. & Young, L.-S. , Berlin, Springer,
pp.366-381.
[12] Kantz, H., & T. Schreiber (1997). Nonlinear Time Series Analysis. Cambridge
University Press, Cambridge.
[13] Abarbanel, H.D.1. (1996). Analysis of Observed Chaotic Data. Springer, New
York.
[14] Abarbanel, H.D.I., Brown, R., Sidorowich, J.J., & Tsimring, L.S. (1993). The
analysis of observed chaotic data in physical systems. Rev. Mod. Phys., 65(4),
pp.1331-1392.
[15] Grassberger, P., Schreiber, T. & Schaffrath, C. (1991). Nonlinear time sequence
analysis. Int. J. Bif. Chaos, 1(3), pp.521-547.
[16] Sauer, T., Y. Yorke, M. Casdagli (1991). Embedology. J. Stat. Phys., 65, pp.579-
616.
[17] Sauer, T., J.A. Yorke (1993). How many delay coordinates do you need?
Int. J. Bifurcation and Chaos, 3, pp. 737-744.
[18] Casdagli, M., S. Eubank, J.D. Farmer, J. Gibson (1991). State space recon-
struction in the presence of noise. Physica D, 51, pp.52-98.
[19] Gibson, J.F., J.D. Farmer, M. Casdagli, S. Eubank (1992). An analytic ap-
proach to practical state space reconstruction. Physica D, 57, pp.I-30.
[20] Broomhead, D.S., G.P. King (1986). Extracting qualitative dynamics from ex-
perimental data. Physica D, 20, pp.217-236.
[21] Landa, P.S., M.G. Rosenblum (1991). Time series analysis for system identifi-
cation and diagnostics. Physica D, 48, pp.232-254.
[22] Palus, M., I. Dvorak (1992). Singular-value decomposition in attractor recon-
struction: pitfalls and precautions. Physica D, 55, pp.221-234.
[23] Sauer, T.(1994). Reconstruction of dynamical systems from interspike intervals.
Phys. Rev. Lett., 72, pp.3811-3814.
234 NONLINEAR MODELING
[24] Castro, R., T. Sauer (1997). Correlation dimension of attractors through inter-
spike intervals. Phys. Rev. E, 55(1), pp.287-290.
[25] Racicot, D.M., A. Longtin (1997). Interspike interval attractors from chaotically
driven neuron models. Physica D, 104, pp.184-204.
[26] Stark, J., D.S. Broomhead, M.E. Davies, J. Huke (1996). Takens embedding
theorems for forced and stochastic systems. Proceedings of the 2nd World
Congress of Nonlinear Analysts, Athens, Greece.
[27] Kennedy, M.P. (1994). Chaos in the Colpitts oscilator. IEEE 7rans. Circuits
Syst., 41(11), pp.771-774.
[28] Cenys, A., K. Pyragas (1988). Estimation of the number of degrees of freedom
from chaotic time series. Phys. Lett. A, 129, pp.227-230.
[29] Buzug, Th., T. Reimers, G. Pfister (1990). Optimal reconstruction of strange
Attractors from purely geometrical arguments. Europhys. Lett., 13, pp.605-610.
[30] Alecsic, Z. (1991). Estimating the embedding dimension. Physica D, 52, pp.362-
368.
[31] Buzug, Th., G. Pfister (1992). Optimal delay time and embedding dimension
for delay-time coordinates by analysis of the global static and local dynamical
behavior of strange attractors. Phys. Rev. A, 45, pp.7073-7084.
[32] Gao, J., Z. Zheng (1993). Local exponential divergence plot and optimal em-
bedding of a chaotic time series. Phys. Lett. A, 181, pp.153-158.
[33] Gao, J., Z. Zheng (1994). Direct dynamical test for deterministic chaos and
optimal embedding of a chaotic time series. Phys. Rev. E, 49, pp.3807-3814.
[34] Huerta, R., C. Santa Cruz, J.R. Dorronsore, V. Lopez (1995). Local state-space
reconstruction using averaged scalar products of dynamical-system flow vectors.
Europhys. Lett., 29, pp.13-18.
[35] Liebert, W., K. Pawelzik, H.G. Schuster (1991). Optimal embeddings of chaotic
attractors from topological considerations. Europhys. Lett., 14, pp.521-526.
[36] Kennel, M.B., R. Brown, H.D.1. Abarbanel (1992). Determining embedding di-
mension for phase-space reconstruction using a geometrical construction. Phys.
Rev. A, 45, pp.3403-3411.
[37] Fredkin, D.R., J.A. Rice (1995). Method of false nearest neigbors: a cautionary
note. Phys. Rev. E, 51(4), pp. 2950-2954.
[38] Cao, L. (1997). Practical method for determining the minimum embedding
dimension of a scalar time series. Physica D, 110, pp. 43-50.
[39] Kember, G., A.C. Fowler (1993). A correlation function for choosing time delays
in phase portrait reconstructions. Phys. Lett. A, 179, pp.72-80.
[40] Rosenstein, M.T., J.J. Collins, C.J. De Luca (1994). Reconstruction expansion
as a geometry-based framework for choosing proper delay times. Physica D, 73,
pp.82-98.
[41] Frazer, A.M., H.L. Swinney (1986). Independent coordinates in strange attrac-
tors from mutual information. Phys. Rev. A, 33, pp.1134-1140.
[42] Frazer, A.M. (1989). Reconstructing attractors from scalar time series: a com-
parison of singular system and redundancy criteria. Physica D, 34, pp.391-404.
[43] Frazer, A.M. (1989). Information and entropy in strange attractors. IEEE
Trans. Info. Theory, 35, pp.245-262.
NONLINEAR TIME-SERIES ANALYSIS 235
[44] Liebert, W., H.G. Schuster (1989). Proper choice of the time delay for the
analysis of chaotic time series. Phys. Lett. A, 142, pp.l07-111.
[45] Martinerie, J.M., A.M. Albano, A.1. Mees, P.E. Rapp (1992). Mutual informa-
tion, strange attractors, and the optimal estimation of dimension. Phys. Rev.
A, 45, pp.7058-7064.
[46] Broomhead, D.S., J.P. Huke, M.R Muldoon (1992). Linear filters and nonlinear
systems. J. Roy. Stat. Soc., B54, pp.373-382.
[47] Davies, M.E., & K.M. Campbell (1996). Linear recursive filters and nonlinear
dynamics. Nonlinearity, 9, pp.487-499.
[48] Kaplan, D., T. Schreiber (1996). Signal separation by nonlinear projections:
The fetal electrocardiogram. Phys. Rev. E, 53(5), pp.R4326-R4329.
[49] Grassberger, P., R. Hegger, H. Kantz, C. Schaffrath, T. Schreiber (1993). On
noise reduction methods for chaotic data. CHAOS, 3, pp.127-141.
[50] Kantz, H., T. Schreiber, I. Hoffmann, T. Buzug, G. Pfister, C.G. Flepp, J. Si-
monet, R Badii, E. Brun (1993). Nonlinear noise reduction: A case study on
experimental data. Phys. Rev. E, 48, pp.1529-1538.
[51] Kostelich, E.J.,T. Schreiber (1993). Noise reduction in chaotic time-series data:
A survey of common methods. Phys. Rev. E, 48, pp.1752-1763.
[52] Theiler, J., B. Galdrikian, A. Longtin, S. Eubank, J.D. Farmer (1992). Using
surrogate data to detect nonlinearity in time series. In: Nonlinear Modeling
and Forecasting, eds. M. Casdagli and S. Eubank, SFI Studies in the Sciences
of Complexity, Vol.XII (Reading, MA,Addison-Wesley), pp.163-188.
[53] Theiler, J., S. Eubank, A. Longtin, B. Galdrikian, J.D. Farmer (1992). Testing
for nonlinearity in time series: the method of surrogate data. Physica D, 58,
pp.77-94.
[54] Provenzale, A., L.A. Smith, R Vio, G. Murante (1992). Distinguishing between
low-dimensional dynamics and randomness in measured time series. Physica D,
58, pp.31-49.
[55] Smith, L. (1992). Identification and prediction of low dimensional dynamics.
Physica D, 58, pp.50-76.
[56] Takens, F. (1993). Detecting nonlinearities in stationary time series. Int. J. of
Bifurcation and Chaos, 3, pp.241-256.
[57] Wayland, R, D. Bromley, D. Pickett, A. Passamante (1993). Recognizing de-
terminism in a time series. Phys. Rev. Lett., 70, pp.580-582.
[58] Palus, M., V. Albrecht,1. Dvorak (1993). Information theoretic test for nonlin-
earity in time series. Phys. Lett. A, 175, pp.203-209.
[59] Kaplan, D. (1994). Exceptional events as evidence for determinism. Physica D,
73, pp.38-48.
[60] Salvino, L.W., R Cawley (1994). Smoothness implies determinism: a method
to detect it in time series. Phys. Rev. Lett., 73, pp.l091-1094.
[61] Savit, RM. Green (1991). Time series and dependent variables. Physica D, 50,
pp.95-116.
[62] Rapp, P.E., A.M. Albano, I.D. Zimmerman, M.A. Jimenez-Moltaiio (1994).
Phase-randomized surrogates can produce spurious identifications of non-
random structure. Phys. Lett. A, 192, pp.27-33.
236 NONLINEAR MODELING