Download as ppt, pdf, or txt
Download as ppt, pdf, or txt
You are on page 1of 44

Linear System Theory

Dr. Vali Uddin


Hamdard University

Vali.uddin@hamdard.edu.pk

Lecture 4 1
Norms
 Want to add more structures to a linear space
 Norms: Generalization of the idea of length
– Key points?
– From here, can define distance, x2 x
convergence, derivative, etc.
x1
 Norm. ||x||: (X, R)  R (or (X, C)  R) such
that
– ||x||  0 and ||x|| = 0 iff x = 0
– ||x|| = || ||x|| for all   R (or C)
– ||x1 + x2||  ||x1|| + ||x2|| ~ Triangular inequality
Lecture 4 2
 Give some norms for (Rn, R)
– ||x||1  |xi|
 [|xi|2]1/2
– ||x||2
 [|xi|p]1/p
– ||x||p
 maxi |xi|
– ||x||
• Do they satisfy the conditions in the definition?
• Find the various norms for x = (2, -3, 1)T
 |xi| = 2 + 3 + 1 = 6
– ||x||1
 [|xi|2]1/2 = (4 + 9 + 1)1/2 = 141/2 = 3.74
– ||x||2
 [|xi|3]1/3 = (8 + 27 + 1)1/3 = 361/3 = 3.302
– ||x||3
 maxi |xi| = 3
– ||x||
Lecture 4 3
 Consider the set of real-valued, piece-wise
continuous functions over [a, b]
 What are their norms?
– ||x||1, ||x||2, ||x||p, ||x||
b 1
x 1   x ( t ) dt b 2  2
x 2    x ( t ) dt 
a a 
1
b p  p
x p    x ( t ) dt  x   max x ( t )
a  a tb

Lecture 4 4
Example. x(t) = e-2t over [0, 1]
1
2 t  1 2t 1
x 1  e dt  e  0.432
0 2 0
1 1
 
1
1   2  1
 1  2 1 4  2  0.495
x 2    e dt    e
4 t 4 t
   1 e 
0   4 0 4 
x   max x( t )  1
a t b

 A normed vector space is a vector space (X,


F) with a norm ||x|| defined on X
 It has a sense of length or distance

Lecture 4 5
Inner Product
 More structure: A sense of orientation ~ X 
X F
– Suppose x = (1 0)T, y = (4 3)T, and z = (0 2)T in R2
z y

x
– What is the inner product of x and y, i.e., <x, y>? What
is <x, z>?
• <x, y> = xyT = x1y1 + x2y2 = |x| |y| cos 
– <x, y> = (1 0) (4 3)T = 4 ~ Projection of y onto x
– <x, z> = (1 0) (0 2)T = 0, as x and z are "orthogonal"
• The above is for R2. How to generalize it?
Lecture 4 6
Example. x = (1 - j, 2)T, y = (3, 1 + 2j)T
 <x, y> = ? <x, x> = ?
*
n

x y  x y   x i yi for Cn , C  *: Complex conjugate transpose
i 1
n  3 
x y   x i yi  1  j 2     5 7j
i 1 1  2 j
n 1  j
x x   xixi  1  j 2   6 ~ Always real
i 1  2 
Q. Why defined in this way?
– Want <x, x>  R and <x, x> > 0 when x  0
Q. General definition?
Lecture 4 7
Inner Product. <x, y>: X  X F such
that
 x, y   y, x  Bar ~ Complex conjugate
1x1   2 x 2 , y  1 x1, y   2 x 2 , y

x, x  0 for all x  0

• Note that from the first two conditions, we have


x, 1y1   2 y 2  1y1   2 y 2 , x

 1 y1, x  2 y 2 , x  1 x, y1  2 x, y 2
Also, x  y, x  y  x, x   x, y   y, x   y, y

x Ay  x*Ay
*
 
* *
A x y  A x y  x*Ay  x Ay A*x y  x Ay
Lecture 4 8
• It can be easily shown that the previous inner
product definitions for (Rn, R) and (Cn, C) satisfy the
above 3 conditions
Example. Real-valued, piece-wise continuous
function over [a, b]
b b
*
x y   x ( t ) y( t )dt   x( t ) y( t )dt
a a
 A pre-Hilbert space is a vector space (X, F) with
a inner product <x, y> defined on X  X
 What properties does inner product have?
The Cauchy-Schwarz Inequality
12 12
x, y  x, x  y, y
Lecture 4 9
12 12
x, y  x, x  y, y

Proof. If y = 0, the above is obviously satisfied. Now


assume y  0. Then for an arbitrary scalar  C,
0  x  y, x  y  x, x   x, y   y, x   y, y

x y x y y x
Set   then   
y y y y y y
2 2 2
x y y x x y
and 0  x, x   
y y y y y y
2
or 0  x, x y y  x y • What is the relationship
between norm and inner
12 12
or x, y  x, x  y, y product?
Lecture 4 10
Theorem: ||x||  <x, x>1/2 is a norm
Proof. Recall that norm should satisfy
– ||x||  0 and ||x|| = 0 iff x = 0
– ||x|| = || ||x|| for all   R (or C)
– ||x1 + x2||  ||x1|| + ||x2|| ~ Triangular inequality
Now 12 12
x x  0, and x x  0 iff x  0
   1 2  x x
12 12 12
x x   x x
x1  x 2 x1  x 2  x1 x1  x1 x 2  x 2 x1  x 2 x2

 x1 x1  x 2 x 2  2 x1 x 2 x1 x 2
12 12
 x1 x1  x 2 x 2  2 x1 x1 x2 x2


 x1 x1
12
 x2 x2 
12 2  ||x1 + x2||  ||x1|| + ||x2||
 ||x||  <x, x>1/2 is a norm
Lecture 4 11
Example. Consider (R2[t], R) ~ Polynomials with
real coefficients of degrees less than 2, 0  t  1
– How to define the inner product? Norm?
12
1 1 2 
x y   x ( t ) y( t )dt x    x ( t )dt 
0 0 
– Let x(t) = t + 3, y(t) = 2t - 1. Compute <x, y>, ||x||,
and ||y||

 
1 1
2  2 3 5 2 
x y   2 t  5t  3 dt   t  t  3t  
1
0 3 2 0 6

Lecture 4 12
12
 
12
1  1 3 1
x    t  6t  9 dt 
2
   t  3t 2  9 t  
0   3  
 0 
12
   3  9 
1
 3.512
3 
12
 
12
1   4 3 
1
y    4 t  4 t  1 dt 
2
   t  2t  t  
2
0   3  
 0 
12

   2  1
4
 0.577
3 

1 12 12
x y   x x  y y  2.028
6

Lecture 4 13
Orthogonality
 The concept of perpendicularity
 In a pre-Hilbert space, x and y are orthogonal (written as
x  y) iff <x, y> = 0
– {x1, x2, .., xn} is an orthogonal set iff xi  xj  i  j
– x is orthogonal to a set S  X if x  s  s  S
– Inner product extends the  concept of R2 and R3 to
general pre-Hilbert spaces
– If several nonzero vectors are orthogonal to each
other, then are they linearly independent?
– An orthogonal set of nonzero vectors is a linearly
independent set. How to show this?
Lecture 4 14
 Suppose that {x1, x2, .., xm} is an orthogonal set of
nonzero vectors and iixi = 0, then
<xk, iixi> = <xk, 0> = 0
= ii <xk, xi> = k <xk, xk>
 k = 0  k
 {x1, x2, .., xm} are linearly independent
– If m = n, then {x1, x2, .., xn} is good candidate for a
basis in view of its orthogonality
 {x1, x2, .., xm} is an orthonormal set iff xi  xj  i  j
and ||xi|| = 1  i
x2

– An even better candidate for a basis x1


x3
Lecture 4 15
– Basis, Representation, and Orthonormalization
 Relationship among a set of vectors: Linear
dependence and linear independence
 Dimension of a linear space

 The base of a linear space: Basis

 Representations of a vector in term of a basis

 Relationship among representations for different


bases
 Generalization of the idea of length: Norms

 A sense of orientation: Inner Product

 The concept of perpendicularity: Orthogonality

• Gram-Schmidt Process to obtain orthonormal vectors


 Projection and Orthogonal Projection Theorem

– Linear Operators and Representations


Lecture 4 16
Gram-Schmidt Process
 The problem: Given a set of LI vectors which are
not orthogonal, derive an orthonormal set
– How?
– Example: (R2, R) T
x2 e2 = (1, 2)
v2 T
e1 = (2, 1)
x1
v1 = e1
– The component of e2 that is  to v1:
v1 e 2 1 2  2  2    0.6 
v 2  e2  v1       
v1 v1  2 4  1  1   1.2 
Lecture 4 17
 {v1 , v2} is an orthogonal set:
<v1 , v2> = -1.2 + 1.2 = 0
– What is next?
– Normalization. How?
 2 
v1  
e1   5
v1  1 

 5 x2
 1 _ _
v2  5  e2 e1
e2   
v2  2  x1
 
 5

Lecture 4 18
 General Procedure:
– Let {e1, e2, .., en} be a linearly independent set
– Form an orthogonal set by subtracting from
each vector the components that align with
v1 e 2
previous
v1  e1 vectorsv 2  e2  v1
v1 v1
v1 e3 v2 e3
v 3  e3  v1  v2
v1 v1 v2 v2
n 1
v k en
v n  en   vk
k 1 v k v k

– Normalize the new set


vi
ei  i
vi

Lecture 4 19
Orthogonal Projection Theorem
 Extension of Pythagorean theorem to linear
spaces
Lemma. If x  y, then
2 2
xy  x  y
2 y

~ Pythagorean theorem x
• How to prove it?
2
xy  xy xy

 x x  x y  y x  y y

2 2
 x  y
Lecture 4 20
Projection
– What is a projection? What properties does
it have?
– Consider projecting x  X onto a subspace M
x
m2
m0
M
m1

– m0 is the point in M that is closest to x


– (x - m0)  M
Lecture 4 21
Orthogonal Projection Theorem
Let (X, F) be a pre-Hilbert space, M be a
subspace, and x be an arbitrary element in
X. Then x
m2
m0  arg  
 min x  m  m0
mM  M
iff (x - m0)  M m1
m0: Projection of x on M
m0 is unique
– If we want to find the best m0 to minimize the
distance, find the projection which is characterized
by (x - m0)  M
– How to prove it?
Lecture 4 22
 If m0 is the minimizing vector, show that (x - m0)  M
 Prove by contradiction
– Suppose that there  m  M s.t. <x - m0, m> =   0
– Without loss of generality, let ||m|| = 1
– m1  m0 +m, where is the complex conjugate of 
– Then want to show that ||x - m1||2 < ||x - m0||2
2
x  m1  x  m0  m, x  m0  m
 x  m0 , x  m0  x  m0 , m  m, x  m0  m, m
2 2 2 2 2
 x  m0    m x
2 2 2
 x  m0    x  m0 , contradict ion
m0 m1
Lecture 4 23
 If (x - m0)  M, show that m0 is the
minimizing vector
2 2
x  m  x  m0  m0  m , m  m0

 x  m0 , x  m0  m0  m, m0  m
 x  m0 , m0  m  m0  m, x  m0

 x  m0
2 = 0 in view that (x - m0)  M

 We can also see the uniqueness:


2 2
x  m  x  m0  m  m0

Lecture 4 24
Example t 2 f(t) = a + bt

1
 
2
Error :  t  a  bt  dt
2

t 1
-1 1

• Best approximate t2 by a linear function f(t) = a + bt for t  [-1,


1] to minimize the square of error
– How to look at this problem?
– Find the projection of t2 onto (R2[t], R)
– What is (X, F)? What is M? What is <x, y>?
 X: R3[t] (the set of polynomials with n < 3),
-1  t  1
 F: R
1
 M: R2[t], -1  t  1
x y   x( t )  y( t )dt
1
Lecture 4 25
• The problem: Find m0 such that (x - m0)  M
1
 ( x  m0 )  mdt  0  m  M
1
1
  
 t  bt  a  c1t  c0  dt  0  c0 , c1  R
2
1
1
 
  c1t 3  c0  bc1 t 2  ac1  bc 0 t  ac 0 dt
1
1
c  bc1 3 ac1  bc 0 2
  1 t 4  0 t  ac 0 t 
c
t 
4 3 2  1

 c0  bc1   2  2b
  2  ac 0  2    2a c0  c1 = 0
 3  3  3

Lecture 4 26
1 1
a ,b0 Or , m 0 
3 3

t2

-1 1 t

– There are definitely better ways to solve this


problem. The approach just presented,
however, illustrated several key concepts on
linear spaces
Lecture 4 27
Linear Operators and Representations
(X, F) (Y, F)
• Functions L
– What is a "function"?
x
– Which one below is a function? y
y y

x x
– A function f is a mapping from domain X to codomain Y
that assigns each x  X one and only one element of Y
– Range: {y  Y|  x  X, s.t. f(x) = y}  Y
– What is a "linear function"?
Lecture 4 28
– A function L that maps from (X, F) to (Y, F)
is said to be a linear operator (linear
function, linear mapping, or linear
transformation) iff
L(1x1 + 2x2) = 1L(x1) + 2L(x2)
 1, 2  F, and  x1, x2  X
– Which of the following is a linear function?
y f0

x
f1
– The is relatively simple. Consider a more
complicated case of a linear time-invariant system:
Lecture 4 29
u y
g(t)

– Assume that the initial conditions are zero, and u() is a


real-valued, piece-wise continuous function
• What is the mapping from u() to y()?
• Is it a linear mapping?
t
y( t )   g( t  )u( )d, t  0, T
0
– L: (U, R)  (Y, R), where U and Y are the set of real-
valued, piece-wise continuous functions
– L satisfies the linearity property
• Representation of a mapping in terms of bases in X/Y?
Lecture 4 30
Introduction to System Theory and Linear
Algebra
 A set of LI vectors {e1, e2, .., en} of (X, F) is said to be a
basis of X if every vector in X can be expressed as a
unique linear combination of them   1
 
x  e1 e 2 ... e n   2 
: 
  ~  
 n
e1 e2 ... en    e1 e2 ... en   
 p11   p1i 
   
e1  e1 e2 ... en   p12 
ei  e1 e2 ... en   p1i 
 :   : 
   
 p1n   p1i 
~ p1 ~ pi
Lecture 4 31
e1 e2 ... en   e1 e2 ... en  P
x  e1 e2 ... e n    e1 e2 ... en  P

 e1 e2 ... en      P

ith column of P: Representation of ei


w.r.t. the set of new basis
– Conversely, ith column of Q: Representation ofei
  Q w.r.t. the set of existing basis
P  Q 1
Inner Product. <x, y>: X  X F such
that
 x, y   y, x  Bar ~ Complex conjugate
1x1   2 x 2 , y  1 x1, y   2 x 2 , y
x, x  0 for all x  0
Lecture 4 32
Matrix Representation of Linear
Operators
 How can we represent a linear operator?
– Every L with finite dimensional X and Y has a
matrix representation with coefficients in F
Theorem. Suppose that (X, F) (Y, F)
L

– Dim X = n, Dim Y = m x
y
– {x1, x2, .., xn} is a basis of X
– {w1, w2, .., wm} is a basis of Y
Then L: (X, F)  (Y, F) is uniquely determined
by n pairs of mapping
yi  Lxi, i = 1, 2, .., n
Lecture 4 33
 What is the representation?
– With yi  Lxi, i = 1, 2, .., n, let
 ai be the representation of yi w.r.t. {w1, w2, .., wm}
 A be the matrix formed as [a1, a2, .., an]

– Then for any x  X represented by  w.r.t {x1,


x2, .., xn}, the representation  of y = Lx w.r.t.
{w1, w2, .., wm} is given by  = A
Proof. (i) Uniqueness
 n  n n
Lx  L  i x i    i Lx i    i yi
 i 1  i 1 i 1

– L is uniquely determined by yi
Lecture 4 34
– To show that  = A, let a1i 
a 2i
Lx i  yi  w1 w 2 .. w m   
 : 
– Then a  ai
 mi 
y  Lx  Lx1 x 2 .. x n  

 Lx1 x 2 .. x n  

 y1 y2 .. y n  

 w1 w 2 .. w m a1 a 2 .. a n  

 w1 w 2 .. w m  A  A
=
– ith column of A: Representation of yi  Axi
w.r.t. {w1, w2, .., wm}
Lecture 4 35
Example. Rotating counter-clock-wise in
R2 by  e2 1   0 1  0
Le e1   , e 2    w1   , w 2   
2 Le1 0 1   0 1

e1 How to proceed?
x  1e1  2e2 Lx  L1e1  2e2   1Le1  2 Le 2  1y1  2 y2

– What is y1? What is y2? a1 cos   sin 


 cos   A 
y1  cos  w1  sin  w 2  w1 w 2    sin  cos  
 sin     sin  
y 2   sin  w1  cos  w 2  w1 w 2  
 cos   a
– If  = (1, 1)T, and  = 90, then 2
e2
0  1 1   1 y x
  A       
1 0  1  1 
e1
Lecture 4 36
Change of Basis
– L: x  y ~ The mapping is independent of bases
–  = A ~ The ith column of A is the representation of
Lxi (= yi) w.r.t. {w1, w2, .., wm}
 The representation A depends on the bases for X
and Y
– Consider a special case where L: (X, F)  (X, F)
– With {e1, e2, .., en} as a basis, L:    ( = A, and
A is n  n)
– Supposed the basis is changed to {e1,e2, ..,en}.
We now have =A
– How to get?
– How are A andA related?
Lecture 4 37
• Recall that = P and = P, where the ith
column of P is the representation of ei w.r.t.
{e1,e2, ..,en}

  P  PA  A   AP

 AP  PA, or A  PAP 1 or A  P 1AP  QAQ 1


– This is called the similar transformation, and A
andA are similar matrices
– Of course, the ith column ofA is the
representation of Lei w.r.t. {ei}i from 1 to n

Lecture 4 38
 0 1
Example. Consider x  Ax , A   
  2 3
– What are the natural modes?
– What is the dynamics for a new representation with
 2  1
P   x  Ax , with A  PAP 1
 1 1 

1  2  1  0 1 1 1 1 0
A  PAP        
  1 1   2 3 1 2   0 2 
1 0 
x    x  Natural modes?
0 2 
 x1   c1e t   c11e t  c12e2 t 
1
x    2t  xP x t 2t 
 2   c2e  c21e  c22e 
Lecture 4 39
Norm of a Linear Operator
(X, F), || ||x (Y, F), || ||y
A

x y = Ax

• Want to define the norm for the linear operator A


– How?
– ||A|| should be based on ||||x and ||||y , and how A
"magnifies" x during the transformation
Ax y
A  sup , or , A  sup Ax y
x 0 x x x x 1

Lecture 4 40
 3 2 4
Example. A   4 0 6
 
1 3 2
– With ||x||1 and ||y||1, what is ||A||1?
– Recall that ||x||1 = i|xi| and A  sup Ax y
x x 1
– Try x = e1, e2, and e3
 3  2  4
Ae1  4, Ae 2  0, Ae 3  6
     
1  3 2
y1  8, y2  5, y3  12,  ||A||1 = 12
m 
• In general, A 1  max   a ij 
j  i 1  ~ Add up column-wise

Lecture 4 41
• Now with ||x|| and ||y||, what is ||A||?
– Recall that ||x|| = maxi|xi|
 3 2 4
A   4 0 6 , A  sup Ax y
  x x 1
1 3 2

 3 2 4 1  9  ||x|| = 1, ||y|| = 10


4 0 6 1  10,
    
1 3 2 1  6   ||A|| = 10
 n 
• In general, 
A   max   a ij 
i  j1  ~ Add up row-wise
• With ||x||2 and ||y||2, what is ||A||2?

Lecture 4 42
 
12
n 2 2
x 2    xi  , Ax  Ax Ax  x*A*Ax  x* A*A x
 i 1 
– It can be shown that ||A||2 is the largest eigenvalue of
(A*A)
– A is said to be bounded iff ||A|| < 
• The concept of ||A|| is needed in Chap. 5

Lecture 4 43
THE END

Lecture 4 44

You might also like