Download as pdf or txt
Download as pdf or txt
You are on page 1of 6

11th

11th IFAC
IFAC Symposium
Symposium onon Dynamics
Dynamics and
and Control
Control of
of
Process
11th
Process Systems,
IFACSystems, including
Symposium Biosystems
on Dynamics
including and Control of
Biosystems
11th
June IFAC Symposium on Dynamics and Control of
Available online at www.sciencedirect.com
June 6-8,
Process 2016.
2016. NTNU,
6-8,Systems,
Process
Trondheim,
including
NTNU, Norway
Biosystems
Trondheim, Norway
June 6-8,Systems, including
2016. NTNU, Biosystems
Trondheim, Norway
June 6-8, 2016. NTNU, Trondheim, Norway
ScienceDirect
IFAC-PapersOnLine 49-7 (2016) 1014–1019
Identification
Identification of
of Linear
Linear Dynamic
Dynamic Systems
Systems
Identification
Identification
using of
of
Dynamic Linear
Linear Dynamic
Dynamic
Iterative Systems
Systems
Principal
using
using Dynamic
Dynamic Iterative
Iterative Principal
Principal
using Component
Dynamic Iterative
Analysis 

Principal
Component
Component Analysis
Analysis
Component ∗
Analysis  ∗∗
Deepak
Deepak Maurya Maurya ∗∗ Arun Arun K. K. Tangirala
Tangirala ∗∗
∗∗ ∗∗
Deepak Shankar Maurya Arun
Narasimhan K. Tangirala
∗∗
Deepak Shankar Maurya Narasimhan
Shankar

Arun K. Tangirala
Narasimhan ∗∗
∗∗
∗∗
∗ Shankar Narasimhan
∗ Department of Electrical Engineering, IIT Madras, Chennai, India
∗ Department of Electrical Engineering, IIT Madras, Chennai, India
∗ Department of Electrical (e-mail: Engineering, IIT Madras, Chennai, India
(e-mail: ee11b109@ee.iitm.ac.in).
ee11b109@ee.iitm.ac.in).
∗∗Department of Electrical Engineering,
(e-mail: ee11b109@ee.iitm.ac.in). IIT Madras, Chennai, India
∗∗ Department of Chemical Engineering,
∗∗ Department of Chemical
(e-mail: Engineering, IIT IIT Madras,
ee11b109@ee.iitm.ac.in). Madras, Chennai,
Chennai, India India
∗∗ Department (e-mail:
(e-mail:of Chemical Engineering,
arunkt@iitm.ac.in,
arunkt@iitm.ac.in, IIT Madras,
naras@iitm.ac.in).
naras@iitm.ac.in). Chennai, India
Department of Chemical Engineering,
(e-mail: arunkt@iitm.ac.in, naras@iitm.ac.in). IIT Madras, Chennai, India
(e-mail: arunkt@iitm.ac.in, naras@iitm.ac.in).
Abstract:
Abstract: The The paperpaper is is concerned
concerned with with identifying
identifying models
models from from datadata that
that havehave errors
errors in in both
both
Abstract:
outputs
outputs and
and The paper
inputs,
inputs, is concerned
popularly
popularly known
known withas
as identifying
the
the models from(EIV)
errors-in-variables
errors-in-variables data problem.
(EIV) that haveThe
problem. errors
The total
total in both
least
least
Abstract:
outputs
squares and The
formulation paper
inputs, of is
popularly
the concernedknown
problem with
is as identifying
the
known models
errors-in-variables
to offer a few from data
(EIV)
well-known that
problem. have
solutions. errors
The In total
this in both
least
work,
squares
outputs formulation
and inputs, of the problem
popularly known is as
known
the to offer a few well-known
errors-in-variables (EIV) solutions.
problem. The In total
this work,least
squares
we
we present
presentformulation
a
a novel
novel of the
and
and problemapproach
systematic
systematic is known to
approach to the
to offeridentification
the a few well-known
identification of
of solutions.
linear
linear dynamic
dynamic In models
this work,
models for
for
squares
we
the present
EIV formulation
casea novel
in the of
and the
principal problem
systematic is known
approach
component to
to
analysis offer
the a few well-known
identification
(PCA) framework.of solutions.
linear
A dynamic
methodology In this
models forwork,for
the
the
we EIV case
present a in theand
novel principal
systematic component
approach analysis
to the (PCA) framework.
identification of A methodology
linear dynamic models for the for
the EIV case
systematic
systematic in theof
recovery
recovery ofprincipal
the
the process
processcomponent
model,
model, analysis (PCA)
including
including the
the framework.of
determination
determination Aorder
of methodology
order and
and delay,
delay, forusing
the
using
the EIV
systematic case in
recovery the principal
of the process component
model, analysis (PCA) framework. A methodology for the
what
what we
we term
systematic term
recovery
as
as dynamic,
dynamic,
ofconstraint
iterative
iterative
the process PCAincluding
PCA
model,
is the determination
is presented.
presented. The
The core stepof
core step order and
consists
consists of delay, using
of determining
determining
whatstructure
the
the we term
structure ofasthe
of dynamic,
the constraint iterative
matrix
matrix PCA byincluding
by a
the determination
ais systematic
presented.
systematic The core step
exploitation
exploitation ofofconsists
of
order
the
the
and
stacking
stacking
delay,
and using
of determining
and PCA
PCA
what we
the structure term as dynamic,
of the partitioning iterative
constraint matrix PCA is presented.
by a systematic The core step consists of determining
order,
order,
the
input-output
input-output
structure of the partitioning
constraint
of
of the
the constraint
matrix constraint
by a matrixexploitation
matrix
systematic
and
and an
exploitation
of the stacking
an appropriate
appropriate
of the rotation.and
rotation.
stacking and
PCA
Optimal
Optimal
PCA
order,
estimates
estimates input-output
of the partitioning
(input-output)
of the (input-output) of the
noise
noise constraint
covariance matrix
covariancematrixmatrices
matrices and an
are
are appropriate
also obtained.
also obtained. rotation.
The Optimal
proposed
The proposed
order, input-output
estimatescan of be theapplied partitioning
(input-output) of the
noise constraint and an appropriate rotation. Optimal
method
method
estimates canof be the applied to
to a
(input-output) a broad
broad classcovariance
class
noise
of
of linear
covariance
matrices including
linear processes
processes
matrices
are also obtained.
including
are
the
the case
also obtained. case of ofThe
The
proposed
unequal
unequal
proposed
and
and
method
unknown
unknown can
error
errorbe applied
variances.
variances. to a broad
Simulation
Simulation class of
results
results linear
are processes
presented
are presented to including
demonstrate
toincluding
demonstrate the case
the
the of unequal
effectiveness
effectiveness and
method
unknown errorcan be applied
variances. to a broad
Simulation class of linear processes the case of unequal and
consistency
consistency
unknown
of
of the
error the proposed
proposed
variances. method. results
method.
Simulation
are presented to demonstrate the effectiveness and
consistency of the proposed method. results are presented to demonstrate the effectiveness and
© 2016, IFACof(International
consistency the proposed Federation
method.of Automatic Control) Hosting by Elsevier Ltd. All rights reserved.
Keywords:
Keywords: identification;
identification; principal principal component
component analysis;
analysis; errors-in-variables;
errors-in-variables; multivariable
multivariable
Keywords: identification; principal component analysis; errors-in-variables; multivariable
Keywords: identification; principal component analysis; errors-in-variables; multivariable
1.
1. INTRODUCTION
INTRODUCTION under
under certain
certain assumptions.
assumptions. A A key
key factor
factor attributed
attributed to to the
the
1. INTRODUCTION under
success
success certain
of
of PCA
PCA assumptions.
(and
(and its
its A keymethods
variants)
variants) factor attributed
methods is
is its
its to the
simplicity,
simplicity,
1. INTRODUCTION under certain assumptions. A key factor attributed to the
System identification
System identification of of linear
linear dynamic
dynamic systems systems has been aa success
has been numerical
numerical
success
of PCA (and its
robustness
robustness
of PCA (and
and
and
its
variants)
versatility.
versatility.
variants)
methods
methods
is its simplicity,
Applications
Applications of
of PCA
isdimensionality PCA
its simplicity,
System identification of linear dynamic systems numerical
span a robustness
diverse set and
including versatility. Applications
identification, of PCA
subject
subject
System
of
of study
study for
identificationfor several
several
of linear
decades
decades
dynamic
with
with a
systems richhas
a rich been a span
literature
literature a numerical a diverse set including
robustness identification,
and versatility. Applicationsdimensionality
of PCA
subject
at
at the of study
the user’s
user’s for several
disposition
disposition decades
(Ljung,
(Ljung, with
1999),
1999), a richhas
naturally
naturally
been
literature
owing
owing
span a diverse
reduction,
reduction,
span a
set including
process
process
diverse set
monitoring
monitoring
including
identification,
(fault
(fault
identification,
detection
detection dimensionality
and
and diag-
dimensionality diag-
atsubject
to of study
the crucial
its for
user’s disposition
role several decades
(Ljung, 1999), with a rich literature
naturallysystemsowing nosis),reduction,
nosis), process
feature monitoring
extraction and as(fault
a detection
pre-processing andtooldiag-
for
to
at its
the crucial
user’s role inin all
disposition all applications
applications
(Ljung, 1999),
of
of process
process
naturally systems
owing reduction,
nosis),
other
feature
feature
identification
extraction
process monitoring
extraction
and as
and
methodologies. as
a pre-processing
(fault
a detection
pre-processing
The use of andtool
tool
dynamic
for
diag-
for
to its crucial
engineering.
engineering. A role
Arole in
larger
larger all applications
proportion
proportion of
of of of
the process
literature
theprocess
literature systems
has
has otherother
nosis), identification
feature extractionmethodologies.
and as a The use
pre-processing of dynamic
tool for
to its crucial
engineering.
concerned withA in
larger
the allproportion
traditional applications
case of
of the literature
deterministic systems
has
(error- identification
PCA methods
methods methodologies.subspace
in errors-in-variables
errors-in-variables The useidentification
subspace of dynamic
identification
concerned
engineering. with the
A the traditional
larger proportion case of deterministic
ofis,deterministic
the literature (error-
has PCA PCA
other in
identification methodologies. The use of dynamic
concerned with traditional (Li
(Li and methods
Qin,
Qin, 2001)in errors-in-variables
is
is an
an example
example of of subspace
the blendidentification
the blend of
of PCA
PCA and
free)
free) inputs.
inputs.
concerned
A
A realistic
with realistic problemcase
problem
the traditional thatof
that
case ofis,
however,
however, relatively
is,deterministic (error-
relatively
(error- PCA
(Li
and
and
standard methods
Qin,
2001)
in errors-in-variables
2001)
subspace is an
methods examplein of subspace
the
identification. blend identification
of PCA and
and
free) inputs.
less-evolved
less-evolved A
is
is realistic
that
that of
of problem
the
the that
errors-in-variables
errors-in-variables however, relatively
(EIV)
(EIV) case
case standard
(Li and subspace
Qin, 2001) methods
is an examplein identification.
of the blend of PCA and
free) inputs.
less-evolved A
is realistic
that problem that is, however, relatively
case standard subspace methods in identification.
where
where the
the inputs
less-evolved inputs
is thatareof
are
of
theknown
also
also
the
errors-in-variables
known with
errors-in-variables errors. (EIV)
with errors. (EIV) case standard
A subspace
distinguishing methods
factor in in
the identification.
A distinguishing factor in the modelling philosophy of modelling philosophy of
where the inputs are also known with errors. A distinguishing
PCA-based approach factor in thestandard
vis-a-vis modelling philosophy
approaches of
(e.g.,
where
Among the
theinputs are
different also known
approaches with
that
Among the different approaches that exist for the EIV PCA-basederrors.
exist for the EIV A distinguishing approach factor vis-a-vis
in the standard
modelling approaches
philosophy (e.g.,
of
Among
problem, the
the different
total least approaches
squares that
(TLS) exist
formulationfor the EIV
offers a PCA-based
regression, approach
subspace)
regression, subspace) vis-a-vis
is that standard
it
is that standard does
it does not approaches
not a priori (e.g.,
a priori (e.g.,dis-
dis-
problem,
Among the different
the total leastapproaches
squares (TLS) that formulation
exist for theoffers
EIV a PCA-based
regression,
tinguish approach
subspace)
between vis-a-vis
is
dependent that it
and does approaches
not
independent a priori dis-
variables,
problem,
few thesolutions.
efficient total leastOne squaresof the(TLS)
of the early formulation
early techniques
techniques involved offers a tinguish between dependent
is that and independent variables,
few
few
efficient
problem,
efficient
finding
solutions.
thesolutions.
totalto
solutions least One
squares
One
the EIV of the (TLS)
early formulation
(linear) techniques
regression offers a regression,
involved
involved
problem tinguish
whereas the subspace)
betweenlatter dependent
approaches it observe
and does notthis
independent a priori dis-
variables,
distinction
finding
few solutions
efficient to
solutions. the
One EIV of (linear)
the early regression
techniques problem
involved whereas
tinguish the
betweenlatter approaches
dependent and observe
independent this distinction
variables,
finding
in the solutions to the EIV (linear) regression problem whereas
right the
upfront. latter
Further, approaches
linear modelsobserve
are this
expressed distinction
as con-
in the TLS
finding TLS framework
framework
solutions to the
(Huffel
(Huffel
EIV
and
and Vandewalle,
(linear) Vandewalle,
regression
1991).
1991). right
problem whereas upfront. Further,
the latter linear models
approaches are expressed
observe as con-
this distinction
in the TLS
Instrumental
Instrumental framework
variable
variable (Huffel
techniques
techniques and their
and Vandewalle,
their 1991). right
generalizations
generalizations
straints
straints
right
upfront.
(among
(among
upfront.
Further,
variables)
variables)
Further,
linear
linear
in
in models
the
the
models
PCA
PCA areframework.
are
expressed
framework.
expressed
asThese
con-
asThese
con-
in thea different
Instrumental
offer TLS framework
variable
approach (Huffel
techniques
to the and
EIV Vandewalle,
their 1991). straints
generalizations
regression problem constraints (amongare variables) by
identified in the
eitherPCA framework.
aa singular valueThesede-
offer a different
Instrumental approach
variable to the and
techniques EIVtheir regression problem straints
generalizations constraints (amongare identified
variables) byin either
the PCA singular
framework. valueThesede-
offer
through a different
the use approach
of instruments to the EIV
(Soderstrom, regression2007). problem
Multi- constraints
composition
composition are
(SVD)
(SVD)identified
of
of the
the bydata
data either
matrix
matrix a singular
or
or the
the value
eigenvalue
eigenvalue de-
through the use of
offer a multi-output
different instruments
approach to the (Soderstrom,
EIV are regression2007). problem constraints
Multi- composition
analysis of of theare
the(SVD)identified
sample by
of covariance either
the data matrix
covariance a singular
matrix.orThe value
the distinction
The eigenvalue
distinction de-
through
input,
input, the use
multi-output of instruments
(MIMO)
(MIMO) (Soderstrom,
systems
systems are 2007).
effectively
effectively Multi-
han-
han- analysis
composition sample
(SVD) of the data matrix.
matrix or the eigenvalue
through the use of instruments (Soderstrom, 2007). Multi- analysis
input,
dled
dled
input, bymulti-output
by the
the state-space
state-space
multi-output
(MIMO)
(MIMO)
modelling
modelling systems
systems
are effectively
paradigm
paradigm and
and the
are effectively thehan-
use between of
between
use analysis
han-
the sample
dependent
dependent
of the sample andcovariance
and independent
independent
covariance
matrix.
matrix.
The distinction
variables
variables is
is invoked
invoked
The distinction
dled
of by
subspacethe state-space
identification modelling
techniques. paradigm
The and
subspace the use
error- between
in
in the
the dependent
post-analysis
post-analysis and
stage,
stage, independent
but
but only
only when
whenvariables
required
required is invoked
by
by the
the
of subspace
dled identification
by the identification
state-space techniques.
modelling The
paradigm subspace
and the error-
use between
in dependent
the post-analysis
application. For and independent
stage, fault
instance, but only variables
when required
detection is
applications invoked
by the do
of subspace
in-variables
in-variables (Chou and
(Chou and Verhaegen, techniques.
Verhaegen, The
1997)
1997) subspace
formulation
formulation error-is
is in application.
the For
post-analysis instance,
stage, fault
but detection
only when applications
required by do
the
of subspace
in-variables
a member identification
of(Chou
this and
class techniques.
Verhaegen,
of methods. The
1997)
On subspace
theformulation
other error-
hand, is application.
not require For
this instance,
distinction fault
since detection
they applications
largely work do
with
ain-variables
member of(Chou this class
anddata of methods.
Verhaegen, On
1997) the other
formulation hand, is notnot require
application. this
For distinction
instance, since
fault they
detection largely work
applications with
do
amultivariate
member of this class
statistical of methods.
analysis On the
techniques other
have hand,
been requirefrom
deviations
deviations this constraints.
from distinction since
constraints. In
In they largely
contrast,
contrast, fault
fault work with
diagnosis,
diagnosis,
multivariate
a member ofstatistical
this class data
of analysis techniques
methods. On the been not
havehand,
other require
deviations this
from distinction since they largely work with
multivariate
proved to bestatistical
to be highly data analysis
highly successful in techniqueslinear
in identifying have been
rela- identification
identification
rela- deviations from or constraints.
or prediction In contrast, invoke
prediction applications
applications fault diagnosis,
invoke this
this dis-
dis-
proved
multivariate
proved
tionships to be statistical
betweenhighly
successful
datafrom
successful
variables analysis
in
identifying
techniqueslinear
identifying
measurements,
linear
have been
rela-
especially identification
tinction
tinction and
and or constraints.
prediction
accordingly
accordingly
In contrast,
applications
partition
partition the
the
fault diagnosis,
invoke
constraint
constraint this dis-
matrix
matrix
tionships
proved tobetween
be highly variables
successful frominmeasurements,
identifying especially
linear rela- identification
tinction
to extract andthe or prediction
accordingly
(causal) applications
partition
input-output the
model. invoke
constraint this dis-
matrix
tionships
when the
when between
the causal variables
causal variables
variables are from measurements,
are collinear.
collinear. Principal especially
Principal compo-compo- tinctionto extractand theaccordingly
(causal) input-output
partition the model. constraint matrix
tionships
when between
the causal variables
variables from measurements, especially to extract the (causal) input-output model.
nent
nent analysis
analysis
when analysis
(PCA)
(PCA)
thetechnique
causal
is
variables oneareof
is one of collinear.
the most Principal
the most widely
widely used compo-
used and
and to The extract
generic theformulation
(causal) input-output
of
of PCA
PCA is model.for
is suited for identifying
nent
effective
effective (PCA)
technique for oneareof collinear.
foris solving
solving the
the TLSPrincipal
the most
TLS widely
problem,
problem,
compo-
used and The
albeit
albeit The
generic formulation
generic formulation
steady-state or instantaneous of PCA
suited
is suited
linear
identifying
for identifying
relationships among
nent analysis (PCA) is one of the most
effective technique for solving the TLS problem, albeit The widely used and steady-state
generic or instantaneous
formulation of PCA linear
is relationships
suited for among
identifying
effective technique for solving the TLS problem, albeit steady-state
variables.
variables. or
Dynamic
Dynamic instantaneous
PCA
PCA (DPCA)
(DPCA) linear (Ku
(Ku relationships
et
et al.,
al., 1995)
1995)amongwas
was

 Corresponding author: Arun K. Tangirala, email: steady-state
variables.
introducedDynamic or instantaneous
as aa natural PCAextension
natural linear
(DPCA) of
extension (Ku
of relationships
thiset formulation
al., 1995)
formulation amongwas
to
 Corresponding author: Arun K. Tangirala, email: introduced
variables. as
Dynamic PCA (DPCA) (Ku thiset al., 1995) to
was
arunkt@iitm.ac.in
Corresponding
arunkt@iitm.ac.in
 author: Arun K. Tangirala, email: introduced as a natural extension of this formulation to
Corresponding
arunkt@iitm.ac.in author: Arun K. Tangirala, email: introduced as a natural extension of this formulation to
arunkt@iitm.ac.in
Copyright
2405-8963 ©
Copyright © 2016,
© 2016 IFAC
2016 IFAC 1014Hosting by Elsevier Ltd. All rights reserved.
IFAC (International Federation of Automatic Control)
1014
Copyright
Peer review©under
2016 responsibility
IFAC 1014Control.
of International Federation of Automatic
Copyright © 2016 IFAC
10.1016/j.ifacol.2016.07.335 1014
IFAC DYCOPS-CAB, 2016
Deepak Maurya et al. / IFAC-PapersOnLine 49-7 (2016) 1014–1019
June 6-8, 2016. NTNU, Trondheim, Norway 1015

suit identification of dynamic relationships by way of ap- a review of the DPCA with an example that highlights
plying PCA on a matrix of lagged variables. While this the limitations of technique is also included. In Section 3
idea has been extensively applied in process monitoring we present the main contribution of this work, which is
and to a certain extent in regression, a rigorous method of a systematic method for identifying the dynamic model
recovering the dynamic model for the error-free variables and its accompaniments from data. Simulation results are
from the results of DPCA is still elusive. Specifically, a discussed in Section 4, wherein Monte-Carlo simulations
theoretically sound way of determining the order, dynamic are presented to study the goodness of estimates and to
relationships and the noise covariance matrix begs atten- illustrate the impact of maximum lag on model quality.
tion. Two related works merit mention in this regard. The The paper ends with a few concluding remarks in 5.
first one pertains to the last principal component analysis
(LPCA). technique (Huang, 2001), which is based on the 2. FOUNDATIONS
DPCA idea. It recovers the dynamic model from the last
eigenvector of the sample covariance matrix. However, We begin with a review of the use of PCA in the identi-
it is severely limited by the assumption of exactly one fication of linear static models for the EIV case, followed
constraint among the lagged variables, which arises only by a brief exposition of the iterative PCA for the same
when the user has an accurate knowledge of the process purpose, however for a broader class of problems.
order (of dynamics) - an assumption that is largely un-
realistic. Further it assumes that the errors in the input It is useful to first introduce the generic problem that is
and output have equal variances, which is again quite addressed in the PCA and IPCA literature. The set up
restrictive. The second work is that of Vijaysai et al. (2005) and statement is as follows.
who develop what they term as the generalized augmented
PC-based regression, largely motivated by the limitation A matrix X ∈ RN ×M , constructed by arranging N obser-
of multivariate statistical techniques such as partial least vations of M variables 1 , i.e.,
squares and principal component regressors in solving the T T
X = [x[0] x[1] · · · x[N − 1]] , x[k] = [x1 [k] · · · xM [k]]
EIV problem as well as LPCA in handling multiple linear (1)
relationships and unequal variances. The idea therein is to
scale the data with the square root of the inverse of noise is constrained by d < M linear relations, i.e., {xi [k]}M i=1
covariance matrix that is assumed to be known. However, fall out of a deterministic (noise-free) linear process. Math-
in almost all realistic situations, the availability of this ematically,
information is largely confined to academic presentations. A0 x[k] = 0, where A0 ∈ Rd×M (2a)
In this work, we present a systematic identification of The goal is to identify the row dimension d and the
linear dynamic processes from input-output data using constraint matrix A0 from X.
dynamic, iterative PCA (IPCA). The steady-state IPCA
was proposed by Narasimhan and Shah (2008) to overcome The EIV identification problem is that of identifying the
the shortcomings of PCA in identifying the number of constraints in (2) and thereby the regression model later
steady-state relationships when the errors in variables have in (A.4) from noisy measurements of X,
unequal variances, and are possibly spatially correlated. z[k] = x[k] + e[k] (3)
The main strategy in IPCA is the scaling of data with the Z=X+E (4)
square root of inverse of noise covariance matrix followed
by an application of the ordinary PCA. In this respect, where e[k] is a vector of white-noise errors with noise
IPCA is to PCA as to what weighted LS method is covariance Σe .
to ordinary LS in standard linear regression. The noise In the following sections we review the mechanics of PCA
covariance matrix and the linear relations are estimated and IPCA in solving the above problems, particularly the
iteratively, and hence the name. Section 2.2 reviews the EIV problem.
technical details of IPCA.
The success of the proposed method rests on two key steps: 2.1 PCA for steady-state identification
(i) the relation between the process order, the number of
constraints identified by the PCA and the maximum lag Principal component analysis is more than a century-old
that is used for stacking the input and output variables, multivariate statistical analysis tool (Pearson, 1902) that
and (ii) the rotation of constraint matrices for the de- searches for correlations among the columns of a matrix
pendent (output) and independent (input) variables. The through a search for the zero singular values or the zero
first one is easy to derive, as we show later in this work. eigenvalues of the sample covariance matrix. For the noise-
Rotation of constraint matrices are derived by optimally free case, the number of zero singular values of X /
solving a set of overdetermined equations. Finally, the 1 T
noise covariance matrix estimate is obtained naturally as eigenvalues of Sx  X X and the associated singular
an outcome of the IPCA algorithm. A minor contribution N
vectors / eigenvectors provide the dimensionality of and
of this work is also a modified IPCA algorithm that relaxes a basis for the constraint matrix A, respectively. A short
the identifiability constraints in the original version. In review of the use of PCA is identification is provided in
general, the proposed method also offers provision for Appendix A.
incorporating prior knowledge of any or all of delay, order,
input dynamics and noise covariance matrix, if available. In the EIV case, the PCA of Z, i.e., eigenvalue analysis of
It must be noted that the development in this work is 1
confined to the class of open-loop systems. the sample covariance matrix Sz  ZT Z does not result
N
in any zero eigenvalue since there exists no pair of columns
The rest of the paper is organized as follows. Section 2 re- of Z that are linearly related. Moreover, the impact of noise
views the basic ideas underpinning PCA and IPCA in solv-
ing the EIV identification or regression problem. Further, 1 Another convention to set up X is as an M ×N matrix, for example
in chemometrics,

1015
IFAC DYCOPS-CAB, 2016
1016 Deepak Maurya et al. / IFAC-PapersOnLine 49-7 (2016) 1014–1019
June 6-8, 2016. NTNU, Trondheim, Norway

variance on the eigenvalues of Sz is too complicated to be the constrained residuals are independent and normally
understood. Consequently, in general, it is not possible to distributed with zero mean and covariance matrix Σri .
accurately determine the size of A in (2) and hence the
matrix as well. Σ = Aˆ(i) Σ (Aˆ(i) )T
ri e (9)
Consequently, an estimate Σe is obtained by solving the
Furthermore, even when the dimension d is known, PCA likelihood problem
cannot theoretically recover the constraint matrix consis-
tently in general. Despite this fact, one finds in literature min N log |Aˆ(i) Σ (Aˆ(i) )T |
e
Σe
several heuristic methods of identifying the dimension d
N

and the constraint matrix A. The only exception is the
case of spatially uncorrelated errors in variables with equal + (rTi [k](Aˆ(i) Σe (Aˆ(i) )T )−1 ri [k]) (10)
variances, i.e., Σe = σe2 IM ×M . In such a situation, the i=1
last d eigenvalues (assuming that the signal-to-noise ratio Identifiability constraints exist since we are essentially
(SNR) > 1) are all equal to σe2 , i.e., λM −d+1 = · · · = λM = recovering an M × M error covariance matrix from d × d
σe2 and PCA recovers an optimal and consistent estimate residual covariance matrix using the relation in (9). Since
of A0 , still differing by a non-singular factor. Σri is symmetric, we have d(d + 1)/2 unique equations.
For a diagonal Σe , the constraint is therefore
For the general scenario, i.e., non-diagonal Σe , the iterative d(d + 1)
PCA, briefly reviewed in the following section, overcomes ≥P (11)
several of the above shortcomings in a theoretically sound 2
manner. where P is the number of elements of Σe that need to be
estimated. For a diagonal Σe , P = M .
2.2 IPCA for linear identification
The idea is to iterate between estimates of A and Σe .
The iterative PCA (IPCA) introduced by Narasimhan and An initial estimate of constraint matrix Â0 is usually
Shah (2008) not only produces a consistent estimate of the generated by PCA for a guess of d, the number of linear
constraint matrix in (2) for a given order, but also correctly relations. If the guess d is correct, the converged estimate
identifies d when it is not known a priori. The key idea is of Σe should result in exactly d unity-valued eigenvalues. A
−1/2 mismatch indicates an incorrect guess. Usually one chooses
to scale the data matrix with Σe so that the problem a guess of d that satisfies the identifiability constraint in
is transformed to the ordinary PCA on the scaled data. (11) and proceeds as required.
There is, however, a difference. Unlike in the standard
case, the dimension of A0 is determined by the number of 3. PROPOSED METHODOLOGY
unity-valued eigenvalues of the sample covariance matrix
of the scaled data. The theoretical basis for the foregoing We open this section with a short review of dynamic
properties of IPCA is briefly reviewed below. For full PCA as proposed by Ku et al. (1995) by means of a
technical details, the reader is referred to Narasimhan and motivating example that also highlights the shortcomings
Shah (2008). of this method. Subsequently, the proposed method is
Starting from (3) and assuming “open-loop” conditions, elucidated.
i.e., errors are uncorrelated with the variables in x, The generic problem of interest is that of identifying a
Σz = Σx + Σe (5) MIMO system from measured data. We restrict ourselves,
where Σx = limN →∞ Sx exists under the quasi-stationarity however, in this article to a single-input single-output
assumption (Ljung, 1999) on z[k]. (SISO) system.
−1/2
Introduce the scaled data zs = Σe z[k] and similarly Consider the class of parametric deterministic SISO lin-
for the deterministic component. Then, the covariance ear time-invariant dynamic input(u )-output(y  ) systems
equation above in (5) for the scaled data takes the form described by
ny nu
Σzs = Σxs + I (6)  
y  [k] + ai y  [k − i] = bj u [k − i] (12)
Eigenvalue analysis of the LHS matrix, by virtue of the
i=1 j=D
eigenvalue shift theorem, results in
λ(Σzs ) = λ(Σxs ) + 1 (7) where ny and nu are output and input order, respectively
and D is the input-output delay. The EIV identification
Consequently, all zero-valued eigenvalues of Σxs map to ny
unity-valued eigenvalues of Σzs , the size of which deter- problem is that of estimating the coefficients {ai }i=1 ,
nu  
mines the number of linear relations d This applies to {bj }j=D from measurements of y [k] and u [k], denoted
sample covariance matrices X and Z as well. Observe by y[k] and u[k], respectively.
that scaling preserves the rank of Σxs . Thus, IPCA offers
a theoretically sound way of determining the number of 3.1 Dynamic PCA and its shortcomings
linear relations, d, and hence a basis for A0 as well from
noisy data. Dynamic PCA attempts to solve the above identification
problem by constructing a matrix of lagged measurements.
Estimation of Error Covariance Matrix Denote the es- To illustrate the basic idea, consider a unit-delay, second-
timate of A0 at the ith iteration by Â(i) (the subscript order, deterministic SISO system
is dropped since only a basis can be determined). Subse- y  [k] + 0.4y  [k − 1] + 0.6y  [k − 2] = 1.2u [k − 1] (13)
quently, one generates constrained residuals as Assume that N measurements {y[k]} and {u[k]} are avail-
ri [k] = Âi z[k] = Âi x[k] + Âi e[k] (8) able.
If the estimate Â(i) is indeed a basis for A0 , then Â(i) = DPCA performs a PCA of the lagged data matrix,
TA0 . Then, the first term on the RHS vanishes and i.e., whose columns are the stacked measurements {y[k]}

1016
IFAC DYCOPS-CAB, 2016
Deepak Maurya et al. / IFAC-PapersOnLine 49-7 (2016) 1014–1019
June 6-8, 2016. NTNU, Trondheim, Norway 1017

through {y[k − Ly ]} and {u[k]} through {u[k − Lu ]}, Assume that the order is unknown and that the maximum
where Ly and Lu are user-specified maximum lags for the lag is set to L = 2. Two linear relations would exist for the
output and input, respectively. A general practice is to let error-free variables with such a choice, i.e., d = 2. Ideally
Ly = Lu = L, a sufficiently large value. the objective is to recover the true constraint matrix,
 
For the system in (13), we assume that the order is known, 1 a 0 0 b 0
A0 = 0 11 a 0 01 b (17)
i.e., L = 2 and construct the k th row of Z as 1 1
T
z[k] = [y[k] y[k − 1] y[k − 2] u[k] u[k − 1] u[k − 2]] assuming that the error-free inputs are not bound by any
T linear relation, an assumption that is fairly unrestrictive.
Z = [z[3] z[4] · · · z[N ]] (14)
Employing the IPCA for this problem will result in two
From (13), there exists only one relation between the un- unity eigenvalues, which correspond to the two relations
derlying deterministic variables, with the true constraint: that exist in the deterministic part of Z, one for y[k] in
A0 = [1 0.4 0.6 0 − 1.2 0] (15) terms of y[k − 1] and u[k − 1] and the other for y[k − 1]
in terms of y[k − 2] and u[k − 2]. However, the constraint
When the data is generated with errors of equal variances, matrix obtained from DIPCA is a rotated version of the
i.e., σe2y = σe2u = 0.09032, DPCA of Z in (14) identifies the true A0 . The key result in this paper is the development
of a method to rotate back the identified constraint matrix
constraint correctly. Firstly, the eigenvalues of Sz are so as to get the true A0 .
Λ = diag([4.1028 3.4908 1.6919 1.1134 0.5024 0.1035])
The first step is that of process order determination. From
The last eigenvalue is, theoretically, an estimate of σe2 the number of identified linear relations and the fact that
and in close agreement with the value used in simulation. only one relation should be expected (since the system is
The corresponding eigenvector, after normalizing the first a single-output and no input relations exist), it is easy to
coefficient to unity, is derive the process order as ny = L−(d−1) = 2−(2−1) = 1.
 = [1 0.4204 0.6052 − 0.0028 − 1.2025 − 0.0377] This is the key step.
which is in close agreement with the true value in (15). Once the order is determined, the remaining step is to
The truly zero-valued coefficients on u[k] and u[k − 3] are reconstruct the true constraint. For this purpose, one could
non-zero, but negligibly small, due to finite-sample errors. go back and re-assemble Z with L = 1. However, it is
not recommended to do so due to the unequal variances
The realistic situation is that (i) the order is seldom known of errors, which calls for an iterative estimation of Σ
accurately and (ii) the noise variances are in general un- and A . From Section 2.2, it is clear that estimation ofe
0
equal and unknown. Unfortunately, DPCA fails to identify Σ is bound by the identifiability constraint in (11) that
e
the model in these situations and there exist no known calls for a “large enough” d, which translates to a large
single method that overcomes both these shortcomings. enough L. For instance, in this example, it is required to
To illustrate the effect of the second factor, which is estimate 2(L + 1) = 6 diagonal unknowns in Σ , of which
e
more severe among the two, we generate data by setting
2 2
σey = 0.2365 and σeu = 0.098 (such that the SNR is 10). strictly speaking there are only two distinct elements, σy2
2
Assuming once again that the order is known, eigenvalue and σu . Therefore, L should be large enough so that d(d +
of Sz is performed to yield, 1)/2 ≥ 2, yielding a minimum of d = 2. In fact, this is the
requirement for all SISO systems, i.e., L should be chosen
Λ = diag([4.3376 3.706 1.6914 1.0846 0.3956 0.1583]) sufficiently large that there are at least two linear relations
Since it is known that a single constraint exists, we turn in the deterministic part of Z. This is not necessarily a
to the eigenvector corresponding to the last eigenvalue limitation of the proposed method, but in general any
 = [1 0.2916 0.5542 0.0570 − 1.2913 0.1787] other PCA-based technique since at least two equations
are required to estimate two unknowns.
which is, evidently, a biased estimate of the true constraint
in (15). An important remark and a modification of the IPCA
algorithm with respect to the estimation of Σe for the
Difficulties compound, of course, when the order is un- dynamic EIV identification problem is discussed in the
known, in which case the maximum lag L is set to a suffi- generic approach shortly.
ciently high value. Consequently, multiple linear relations
exist among the corresponding deterministic columns of With an accurate estimate of the order, the next step is
Z. For instance, it is easy to verify that choosing L = 3 that of partitioning the constraint matrices corresponding
results in d = 2 constraints. There exists no systematic to dependent (lagged outputs) and independent (lagged in-
way of recovering the model from these excess constraints, puts) variables as ÂD and ÂI , respectively. Subsequently,
even if variances were equal. the structure of ÂD is constructed, i.e., the locations of
The proposed methodology addresses both the foregoing zero- and unity-valued elements are identified. For the
issues in the IPCA framework. example under discussion, the system has already been
determined to be first-order. Therefore, ÂD should possess
3.2 Dynamic IPCA for EIV identification the structure
 
 1 α1 0
To begin with, the basic idea underlying the proposed A D = 0 1 α2 (18)
method is explained with a symbolic example of a first-
order, unit-delay system. where we have used the fact that leading coefficients of
Assume that we are interested in identifying a first-order, difference equation models are always unity and exploited
unit delay SISO system: the shift property of the rows containing shifted difference
   equations. Furthermore, the constants α1 , α2 ∈ R should
y [k] + a1 y [k − 1] = b1 u[k − 1] (16) also be identical, i.e., α1 = α2 .

1017
IFAC DYCOPS-CAB, 2016
1018 Deepak Maurya et al. / IFAC-PapersOnLine 49-7 (2016) 1014–1019
June 6-8, 2016. NTNU, Trondheim, Norway

As a penultimate step, a rotation of the matrix ÂD to two. Furthermore, the optimization problem for estimating
yield this structure is determined so that Σe in (10) is modified to reflect this constraint as follows:
structure(RD ÂD ) = structure(AD ) (19) min N log |Aˆ(i) Σ (Aˆ(i) )T |
e
Σe
This rotation matrix may be determined by solving an N
exact set of equations so that only the zero- and unity- 
valued entries of the rotated and AD match, which we term + (rTi [k](Aˆ(i) Σe (Aˆ(i) )T )−1 ri [k]) (22)
as the first approach. Alternatively, an overdetermined set i=1
 
of equations may be solved by also forcing the equality of subject to diag(Σe ) = σy2 IL+1 σu2 IL+1 (23)
non-zero coefficients in the shifted rows. Intuitively, the where IL+1 is the (L + 1) × (L + 1) identity matrix.
second approach can be expected to yield more efficient
estimates of the model coefficients. Regardless of the The next section presents results from simulation studies
approach, the same rotation matrix is applied to both to demonstrate the effectiveness of the proposed algorithm.
the sub-blocks, ÂD and ÂI of the constraint matrix. For
the example under discussion, the rotation matrix contains 4. SIMULATION RESULTS
2× 2 = 4 unknowns. With the first approach, we would set
up four equations to obtain an exact solution, while with The case study pertains to the second-order example in
the second approach, five overdetermined equations would (13). N = 1000 observations of the data were generated
be used in estimating these elements. with unequal error variances of σy2 = 0.2406 and σu2 =
Finally, the difference equation model is obtained by 0.091 so as to achieve an SNR of 10.
averaging the rows of the respective rotated matrices, AD The data is run through the proposed algorithm in Table
and ÂI . Table 1 summarizes the algorithm to recover the 1 with L = 3. Two unity eigenvalues were identified by the
true matrix A0 IPCA algorithm, as reported below.
Λ = diag([11.6 11 8.9 8.09 4.7 1.92 1.0037 0.9963])
Remark: Essentially, we have solved a structured TLS
problem, where the structure (for the constraint matrix) Further, Σ̂e = diag([0.2363 0.1087]), which closely
is also derived from the data using IPCA. matches with the true value.
Table 1. Proposed DIPCA algorithm From the relation given in Step 3, the order is estimated to
be ny = 3 − 2 + 1 = 2, thus identifying the order correctly.

(1) Construct the data matrix Z by choosing a sufficiently large L. The 2 × 8 constraint matrix is estimated and partitioned
(2) Identify the number of linear relations d and the constraint into dependent and independent matrices. The structure
matrix  underpinning the error-free part of Z through IPCA. of AD is subsequently determined to be,
 
Also obtain the estimate of Σe . 1 × × 0
(3) Determine the process order using the relation d = L − AD = 0 1 × × (24)
max(ny , nu ) + 1
(4) Partition the constraint matrix into sub-matrices correspond- To determine the rotation matrix, both approaches are
ing to dependent (output) and independent (input) variables implemented. The first (exact) approach yields
as follows    
  −3.240 −22.7 1 0.420 0.572 0
 = ÂD ÂI ÂD ∈ Rd×(L+1) , ÂI ∈ Rd×(L+1) (20) RD =
−22.721 −3.564
RD AD =
0 1 0.404 0.595
so that
 
    0.054 1.231 −0.019 −0.056
RD AI =
y[k] u[k] 0.046 0.053 1.2321 −0.038
ÂD  ..  = −ÂI  .. 
. .
(21) From the above, the estimated difference equation model
y[k − L] u[k − L] is obtained using the last step of Table 1:
(5) Using information in Step 3 and the shift property of lagged y[k] + 0.412y[k − 1] + 0.583y[k − 2] =
constraints, determine the location of unity- and zero-valued 0.053u[k] + 1.232u[k − 1] − 0.029u[k − 2] − 0.056u[k − 3]
entries in A0 . Denote this structured matrix as AD
(6) Determine the rotation matrix RD so that RD ÂD structurally
Using the second approach (over-determined LS),
   
matches AD , preferably using the overdetermined approach. −3.068 −22.798 1 0.412 0.567 −0.005
RD = R D AD =
(7) Finally, obtain the difference equation model by averaging the −22.529 −3.716 0.008 1 0.407 0.592
rows of RD ÂD and RD ÂI .  
0.053 1.231 −0.029 −0.056
R D AI =
0.047 0.063 1.228 −0.038
yielding the difference equation as
Remark: The structure and rotation matrix determination
in Steps 5 and 6, respectively, of the proposed algorithm y[k] + 0.409y[k − 1] + 0.581y[k − 2] − 0.005y[k − 3] =
are equipped to accommodate any prior information on 0.058u[k] + 1.229u[k − 1] − 0.034u[k − 2] + 0.056u[k − 3]
the input order, delay, etc. The estimates obtained with both approaches are satisfac-
torily close to the true values used in simulation.
Estimating Σe using constrained optimization The generic
formulation of IPCA, as presented in Section 2.2, assumes In order to demonstrate that the extra coefficients ap-
that all elements of Σe are distinct. However, when applied pearing in both model estimates are negligibly small and
to the dynamic SISO EIV identification problem, it is also to show that the remaining coefficients are unbiased
known a priori that for the matrix Z constructed from and significant, Monte Carlo simulations were carried out.
L lagged input and output measurements, only two of the Results from 200 runs for lag order 3 are presented in
2(L + 1) elements of Σe are distinct. Consequently, the Table 2. Results suggest that both approaches yield
lower limit of the identifiability constraint in (11) is always estimates with similar error characteristics, although it

1018
IFAC DYCOPS-CAB, 2016
Deepak Maurya et al. / IFAC-PapersOnLine 49-7 (2016) 1014–1019
June 6-8, 2016. NTNU, Trondheim, Norway 1019

puts in the EIV case. The method is based on an extension


of the iterative PCA, which was originally formulated for
static systems. Using the ideas of dynamic PCA and a
systematic exploitation of the information given by IPCA,
we developed a method for determining the structure of
the true constraint matrices. The result is a consistent and
efficient method for discovering the dynamic model purely
from data with minimal user intervention. Future work
consists of extending the proposed method to closed-loop
multivariable systems.
REFERENCES
Chou, C. and Verhaegen, M. (1997). Subspace algorithms for the
Fig. 1. Plot of log2 λ with error bounds identification of multivariable dynamic errors-in-variables models.
Automatica, 33(10), 1857–1869.
must be said that Approach 2 is capable of accommodating Huang, B. (2001). Process identification based on last principal
component analysis. Journal of Process Control, 11, 19–33.
prior information (of input order, delay) in a simple way. Huffel, S.V. and Vandewalle, J. (1991). The Total Least Squares
The results for estimates of noise variances are reported Problem: Computational Aspects and Analysis. SIAM, Philadel-
in Table 3. From the results, it can be safely said that the phia, PA, USA.
Ku, W., Storer, R., and Georgakis, C. (1995). Disturbance detection
noise variance estimates are unbiased and of low error. and isolation by dynamic principal component analysis. Chemo-
Monte Carlo Simulation results of 200 runs for lag order metrics and Intelligent Laboratory Systems, 30, 179–196.
L = 10 are reported in Table 4 and Table 5. From these Li, W. and Qin, S.J. (2001). Consistent dynamic PCA based on
errors-in-variables subspace identification. Journal of Process
results, it is clear that stacking in excess yields improved Control, 11(6), 661–676.
estimates of parameters. Ljung, L. (1999). System Identification - A Theory for the User.
Figure 1 shows the eigenvalues obtained from IPCA with Prentice Hall International, Upper Saddle River, NJ, USA.
Narasimhan, S. and Shah, S. (2008). Model identification and error
the error bounds computed from MC simulations. The covariance matrix estimation from noisy data using PCA. Control
expected number of unit eigenvalues, nine in this case, Engineering Practice, 16, 146–155.
remain within bounds. Pearson, K. (1902). On the systematic fitting of curves to observa-
tions and measurements, part 1. Biometrika, 1, 265–303.
5. CONCLUSIONS Soderstrom, T. (2007). Errors-in-variables methods in system iden-
tification. Automatica, 43(6), 939–958.
Vijaysai, P., Gudi, R., and Lakshminarayanan, S. (2005). Errors-in-
In this paper, we have presented a systematic method for variables-based modeling using augmented principal components.
recovering the dynamic model of a linear multivariable Industrial & Engineering Chemistry Research, 44, 368–380.
open-loop process from measurements of inputs and out-
Table 2. MC simulation results for L = 3 Appendix A. PCA FOR STATIC IDENTIFICATION
A basis for the constraint matrix A0 , call it A, and hence its row
Approach 1 Approach 2
dimension d are given by the basis and dimension of the null space
Para. True Value mean std. dev. mean std. dev. of X, respectively. These can be obtained either through an SVD
a1 0.4 0.4029 0.0192 0.4028 0.0193 of X or via an eigenvalue analysis of the sample covariance matrix
a2 0.6 0.6001 0.0175 0.60 0.0187 1 T
a3 0 0∗ 0∗ -0.0002 0.0026 Sx  X X as follows 2 :
N
b0 0 0.0018 0.0228 0.0021 0.0230 λ 
b1 1.2 1.2031 0.0390 1.2030 0.0386 1
b2 0 0.0041 0.0374 0.0037 0.0378
Sx V = VΛ, Λ = 
 ... ,λ > ··· > λ
b3 0 -0.0033 0.0446 -0.0033 0.0446  1 M −d (A.1)
λM −d
0
Table 3. µ and σ of σ̂e2i for lag L = 3    T
V = v 1 v 2 · · · vM . A = vM −d+1 · · · vM (A.2)
Parameters True Value mean std. deviation
The projections of columns of X onto the basis vectors V collected in
σy2 0.2406 0.2456 0.0405 P = XV are termed as the principal components of X. They satisfy
2
σu 0.0910 0.0979 0.0280 pT
i pj = 0, ∀i = j = 1, · · · , M .

Table 4. µ and σ of σ̂e2i for lag L = 10 Note that A and A0 differ by a similarity transformation T s.t.
TT T = I. In general, this is not an issue and is completely
Parameters True Value mean std. deviation alleviated in regression since the variable set can be partitioned into
σy2 0.2406 0.2402 0.0273 d dependent and (M − d) independent variables and the regression
2
σu 0.0910 0.0979 0.0154 coefficients can be uniquely recovered as follows:
 T  
xI [k]
Table 5. µ and σ of estimates for L = 10 x[k] = , A = AD A I (A.3)
xD [k]
=⇒ xD [k] = −A−1xI [k],
D AI AD ∈ R d×d
(A.4)
Approach 1 Approach 2
  
B
Parameters True Value mean std. dev. mean std. dev.
a1 0.4 0.3979 0.0456 0.3990 0.0251 The coefficient matrix B is invariant to a similarity transformation
a2 0.6 0.6049 0.0356 0.6015 0.0221 of A, i.e., both A and TA yield the same B.
b0 0 0.0014 0.0225 0.0015 0.0225 2 It is a common practice to normalize columns of X to account for
b1 1.2 1.2022 0.0295 1.2022 0.0294 disparities in the units of variables.

1019

You might also like