Professional Documents
Culture Documents
Linear Algebra Course Pack PDF
Linear Algebra Course Pack PDF
Linear Algebra Course Pack PDF
Bases, co-ordinates
Orthogonal bases =⇒ projections
Nearest point map: good approximations
Linear regression
Exotic examples:
R1
A(f ) := 0 f (t) dt
Usually, we shall take the set of scalars to be the set R of all real
numbers: then we say that V is a real vector space. Sometimes
we shall take the set of scalars to be the set C of all complex
numbers: then V is a complex vector space.
z1
C3 :=
z 2
: z1 , z2 , z3 ∈ C
z3
√
Example: with i = −1
2i
v = 1+i
,
3 − 5i
(1 − i)2i
(1 − i)v = (1 − i)(1 + i)
(1 − i)(3 − 5i)
2 + 2i
= 2
1−i =1+1
3 − 5 − i(3 + 5)
2 + 2i
= 2 .
−2 − 8i
1 0
0
2 , 4 + i , 0 .
i −1 3−i
1 0 0 0
a1 2 + a2 4 + i + a3 0
= 0 .
i −1 3−i 0
a1 + 0 + 0 = 0, =⇒ a1 = 0,
2a1 + (4 + i)a2 + 0a3 = 0, =⇒ a2 = 0,
ia1 − a2 + a3 (3 − i) = 0 =⇒ a3 = 0.
1 2 0
v1 = 3 , v2 =
0
, v3 = 2 .
1 −1 1
By observation
(−2)v1 + 1v2 + 3v3 = 0.
Indeed
1 2 0 0
(−2)
3 + 0 + 3 2 = 0 = 0 .
1 −1 1 0
(−2) × 1 + 2 + 3 × 0 = 0,
(−2) × 3 + 0 + 3 × 2 = 0,
(−2) × 1 + (−1) + 3 × 1 = 0.
1 2 0
v1 =
3
, v2 =
0
, v3 = 2
1 −1 1
v1T 1 3 1
=⇒ v2 = 2 0 −1
T
.
v3T 0 2 1
This is the transpose and stack part. The next step is the
row-reduce part.
On completion :
1 3 1 v1T
2 0 −1 = v T
2
0 2 1 v3T
R2 −2R1 R2 −2R1
1 3 1 v1T
0 −6 −3 = [vT − 2vT ]
2 1
0 2 1 v3T
R2 +3R3 R2 +3R3
1 3 1 v1T
0 0 0 = T T T
[v2 − 2v1 ] + 3v3
0 2 1 v3T
From
1 3 1 v1T
0 0 0 = [vT − 2vT ] + 3vT
2 1 3
0 2 1 v3T
we have
(finally).
Equivalently:
a subset U of a vector space V is a subspace of V
iff U is closed under linear combinations:
x
W = y
: 2x − y + z = 0
z
2x − y + z = 0,
2u − v + w = 0.
2[x + u] − [y + v] + [z + w] = 0,
Xk
Lin(S) = { ai vi | ai ∈ R}
i=1
Ax = b.
(A|b)
Looking just at the first two columns, which are the columns
containing leading 1s, we can read off one solution: x1 = 2 ,
x2 = 7 and x3 = 0 .
MA212 – Lecture 37 – Fri 8 December 2017 page 9
Simple example cont’d
1
1
...
1
Ea =
a
←− j th row × a 6= 0
1
...
1
1 0 0 a b a b
0 −2 0 c d = −2c −2d
0 0 1 e f e f
E1/a Ea = I,
1
...
0 1
←− ith row
1
exch
E =
...
1 0
←− j th row
...
1
0 0 1 a b e f
0 1 0 c d = c d
1 0 0 e f a b
Another example:
E exch E exch = I,
Fix i, j .
1
...
1
←− ith
row
1
Eaadd =
...
a 1
Rj −→ Rj + aRi
...
1
1 0 0 a b a b
0 1 2 c d = c + 2e d + 2f
0 0 1 e f e f
Another example:
add add
E−a Ea = I,
add
since E−a takes away a× ith row of I from the j th row of Eaadd ,
so cancels the earlier addition of that same row to the j -th row.
So Eaadd is invertible.
Ek · ... · E2 · E1 A = I.
So A is invertible and
A−1 = Ek · ... · E2 · E1 .
If E is an elementary matrix
Why? Because
B = H1 · ... · Hk ,
AB = F1 · ... · Fk · H1 · ... · Hk
Differentiate again
a1 f1′′ (t) + a2 f2′′ (t) + a3 f3′′ (t) = o(t) = 0 (0 ≤ t ≤ 1).
W(t)a = 0 (0 ≤ t ≤ 1),
W (t0 ) = det(W(t0 )) 6= 0.
Then f1 (t), f2 (t), ..., fk (t) are linearly independent if for some t0
W (t0 ) = det(W(t0 )) 6= 0.
T x : = Ax,
and
y =Ax.
Here U = R3 and V = R1 = R
x
T y
= 2x − y + z.
x u
Consider x: =
y , u : = v , then
z w
x+u
T (x + u) = T y + v
= 2[x + u] − [y + v] + [z + w]
MA212 – Lecture 23 – Tuesday 9 January 2018 page 5
z+w
Example 1 cont’d
So
x+u
T (x + u) = T
y + v = 2[x + u] − [y + v] + [z + w] =
z+w
= 2x − y + z + [2u − v + w]
= T (x)+T (u)
represents T as we have
x
T (x) =2x − y + z = 2 −1 1 y
= Ax.
Here U = R3 and V = R3
x 2x − y
T y
= y+z
z 2x + z
2 −1 0 x
=
0 1 1 y
2 0 1 z
Homogeneity
Z 1 Z 1
T (af ) = af (t) dt = a f (t) dt = aT (f ).
0 0
Here
Note the round or square brackets (but NOT the curly variety
{...} ). The notation emphasizes that the vectors ui in the list
above are presented in a fixed order.
u =x1 u1 + x2 u2 + ... + xn un ,
i.e. if also
u =x′1 u1 + x2 u2 + ... + x′n un ,
then
x1 = x′1 , ..., xn = x′n .
so
4 4
u = 4 = 4 .
6 6
E
6 6
E
so
4 4 5
u = 4 = 4 = 6
.
6 6 −1
E B
Alternatively
4 4 5
[u]E =
4 = 4 whereas [u]B = 6 .
6 6 −1
E
x1
x2
For u ∈ U, knowing its B -representation x =
..
, we
.
xn
B
need to find y with
y1
y2
v =T (u) =
.
..
.
ym
C
a a a
11 12 1n
a21 a22 a2n
T (u1 ), T (u2 ), ..., T (un ) =
..
,
..
, ...
..
. . .
am1 am2 amn
C C C
Put
a a12 ... a1n
11
a21 a22 a2n
Am×n
= .
. ..
. .
am1 am2 ... amn
y = Ax.
Indeed
So
a a a
11 12 1n
a21 a a
y = x1 + x2 22 + ... + xn 2n
.. .. ..
. . .
am1 am2 amn
= Ax.
y =AC,B
T x, or even y = C B
AT x.
since uE = u
We simplify the notation here to
A = AT
For T : R3 → R2 given by
x1
2x1 − x2
T x2 = ,
x1 + x2 + x3
x3
we can write
x1 x1
2 −1 0
T
x2 =
x2
1 1 1
x3 x3
and so
2 −1 0
AT = .
1 1 1
p ∈ Q iff p(t) = a0 + a1 t + a2 t2
Take
B = (u1 , u2 , u3 ) = (1, t, t2 )
then
a0
pB = a 1
a2
Consider T : Q → Q
Here
T (u1 ) = T (1) = 1 = u1
T (u2 ) = T (t) = t + 1 = u2 + u1
T (u3 ) = T (t2 ) = (t + 1)2 = t2 + 2t + 1 = u3 + 2u2 + u1
So
1 1 1
T (u1 )B =
0 , T (u2 )B = 1 , T (u3 )B = 2
0 0 1
...
1 1 1
A= AB,B
T = 0 1 2
.
0 0 1
i.e.
1 1 1
A= AB,B
T = 0 1 2
0 0 1
So, as B = C
a0 + a1 + a2 1 1 1 a0
qC = qB =
a1 + 2a2 = 0 1 2 a = ApB .
1
a2 0 0 1 a2
[u]B = u = xB ?
By definition of representation
x1
x2
u =x1 u1 + x2 u2 + ... + xn un , i.e. [u]B =
..
.
xn
B
but
x1
x2
u =x1 u1 + x2 u2 + ... + xn un = (u1 , u2 , ..., un )
..
.
.
xn
MA212 – Lecture 23 – Tuesday 9 January 2018 page 42
So
x =B −1 u.
B = AE,B
T = T (B) so T [u1 , u2 , ..., un ] = B = [u1 , u2 , ..., un ].
so by linearity
So
B −1 = AB,E
I
y =Ax.
yC = M xB ,
M = AC,B
T = C −1
AB.
Story so far:
1. T : U → V with bases B = (b1 , b2 , ..., bn ) in U and C in V
is represented by:
For T : R3 → R2 given by
x1 x1
2x1 − x2 2 −1 0
T x2 = = x
2
x1 + x2 + x3 1 1 1
x3 x3
1 0 1 1 0 1
B = 0 , 1 , 2 7→ 0 1 2
1 0 −1 1 0 −1
and
2 −1 2 −1
C = , 7→
4 3 4 3
−1 1 3 1
C = (swop/re-sign/divide)
6+4 −4 2
A = AEE
T ;
this uses E the natural base twice: in the domain and in the
range of T . Now switch from the natural base E to an arbitrary
base C in Rn , to be used both in the domain and the range of
T. Then the transformation T is represented by
−1 EE −1
ACC
T = C AT C = C AC.
Think of
P = (p1 , p2 , ..., pn )
B = P AP −1 and C = QBQ−1 ;
so
C = QP AP −1 Q−1 = (QP )A(QP )−1
and so C is similar to A.
Ax = λx for some x 6= 0,
so since x 6= 0 :
λ is an eigenvalue iff A − λI is singular iff det(A − λI) = 0 .
A real matrix A may have complex eigenvalues, in which case
the eigenvector x is in Cn .
pA (x) = pB (x).
and so
For
1 2
A=
3 4
x − 1 −2
pA (x) = = (x − 1)(x − 4) − (−2)(−3)
−3 x − 4
= x2 − 5x − 2
Notice that
5 = (1 + 4) = trace(A)
−2 = (4 − 6) = det(A).
pA (x) = (x − λ1 )...(x − λn ),
(Put x = 0 )
(−1)n det(A) = (−1)n λ1 λ2 ...λn
and so
det(A) = λ1 λ2 ...λn .
T
U V
࣬(T)
࣬(T)
ࣨ(T)
and
v =T (u) iff y =ACB
T x.
N (T ) = N (AT ) = N (A)
Consider
x x−y 1 −1 x
T = = .
y 2y − 2x −2 2 y
1 −1
AT =
−2 2
so we must solve
x − y = 0 and 2y − 2x = 0. So
x 1 1
N (T ) = : x ∈ R = x : x ∈ R = Lin
x 1 1
x−y x−y
R(T ) = : x, y ∈ R = : x, y ∈
2y − 2x −2(x − y)
1 1
= (x − y) : x, y ∈ R = z :z∈R
−2 −2
1
= Lin ,
−2
again a line.
MA212 – Lecture 24 – Friday 12 January 2018 page 25
Range and column space
2 4
2 1 = 2
3 6
2 4
2 1 = 2
3 6
2 4
2
1 =
2
3−2 6−4
R3 −R1
1 2 0 1 2 leader
0 0 1 1 −1 leader
B =
0 0 0 0 0 OK
0 0 0 0 0 OK
0 0 0 0 0 OK
↑ × ↑ × ×
=
x1 x2 x3 x4 x5
i.e.
x1 + 2x2 + x4 + 2x5 = 0,
x3 + x4 − x5 = 0.
x1 = −2x2 − x4 − 2x5 ,
x3 = −x4 + x5 .
−2x2 − x4 − 2x5 −2 −1 −2
x2 1 0
0
−x4 + x5 = x2 0 + x4
−1 + x5 1
x4 0 1 0
x5 0 0 1
B = M5 M4 M3 M2 M1 A
So
−2 −1 −2
1 0
0
N (A) = Lin
0 , −1
,
1 .
0
1 0
0 0 1
col(B)1 = e1
col(B)2 = 2e1 = 2col(B)1
col(B)3 = e2
col(B)4 = e1 + e2 = col(B)1 + col(B)3
col(B)5 = 2e1 − e2 = 2col(B)1 − col(B)3
So
col-space(B) = Lin{col(B)1 , col(B)3 }.
row-rank(B) = col-rank(B) = r.
row-rank(A) = r = col-rank(A).
then
b = a + (b − a) = a + c,
u · v := u1 v1 + u2 v2 ,
and
and
u·v =v·u
||c||2 = c · c
= (b − a) · (b − a) = b · (b − a) − a · (b − a)
= b · b + a · a−2a · b
So
a · b =||a|| · ||b|| cos θ
u · v = u1 v1 + u2 v2 + ... + un vn
u · v = 0.
1. Linearity property:
(αu+βv) · w = αu · w+βv · w.
Tv (u) := u · v.
2. Symmetry property
u · v = v · u.
3. Positivity property
u · u >0 for u 6= 0.
Inner products
Orthogonality
Pythagoras’s Theorem
Angles ...leading to
Orthonormal bases (if time allows)
Recapitulation
u · v = u1 v1 + ... + un vn
Now apply the old scalar product to the images. This gives a
new inner product, so we use angular brackets to denote this
new operation:
hu, vi = uT Av =
= u1 v1 + 2u2 v2 + u3 v3 + u1 v2 + u2 v1 .
Consider that
2. Symmetry
hv1 , v2 i = hv2 , v1 i,
3. Positivity
hv, vi > 0 for v 6= 0.
and in particular
h0, 0i = 0.
Indeed
then
But, likewise,
Similarly
etc .
X
hv, wi = xi yj hvi , vj i = xT W y,
i,j
E[X 2 ] < ∞.
hX, Y i = E(XY ).
hX, Xi = E[X 2 ] ≥ 0
E(XY ) = E(X)E(Y ) = 0
||αv|| = |α|.||v||
MA212 – Lecture 25 – Tuesday 16 January 2018 page 19
An application
||u + v||2 = hu + v, u + vi
= hu, ui + hv, vi + hu, vi + hv, ui
= ||u||2 +||v||2 ,
The corollary in the next slide (in the form of a Theorem) allows
us to regard an arbitrary inner product as satisfying
α := ||u|| cos θ.
Proof. We do this in two steps. MA212 – Lecture 25 – Tuesday 16 January 2018 page 24
Step 1.
||v|| =1.
α := hu, vi
w := u − αv.
So also
hw, αvi = αhw, vi = 0.
So
||u|| ≥ |hu, vi|,
as required.
Since
1
hu, v/||v||i = hu, vi
||v||
we get
1
|hu, vi| = |hu, v/||v||i| ≤ ||u||,
||v||
and so on cross-multiplying
as required.
As
−||u||.||v|| ≤hu, vi ≤ ||u||.||v||,
hu, vi
−1≤ ≤ 1,
||u||.||v||
hu, vi
cos θ = .
||u||.||v||
||u + v||2 = hu + v, u + vi
= hu, ui + hv, vi + hu, vi + hv, ui
= ||u||2 + ||v||2 + 2hu, vi
≤ ||u||2 + ||v||2 + 2||u||.||v|| by Cauchy-Schwarz
= (||u|| + ||v||)2 .
That is
||u + v||2 ≤ (||u|| + ||v||)2 .
But since both ||u + v|| and ||u|| + ||v|| are non-negative
as required.
Applications
In C[0, 1] put
Z 1
hf, gi = f (t)g(t) dt.
0
gives
|hX̃, Ỹ i|2 ≤ ||X̃||22 .||Ỹ ||22
and in particular
So
cov(X, Y )2 ≤ var(X)var(Y ).
Orthogonality
Orthogonal bases,
Orthogonal matrices: isometry & conformality properties
Rotation and reflection in R2 and R3
Recapitulation: story so far
v v hu, viv
hu, i =
||v|| ||v|| ||v||2
In an n -dimensional space
1 2
W =
2 1
1
... if u = then hu, ui = −2 < 0.
−1
Then
ha1 v1 +...+am vm , v1 i = h0, v1 i = 0.
Using linearity,
Step 1. We take
u1
v1 :=
||u1 ||
so that ||v1 || = 1.
We choose α so that
w2 = u2 − αv1 ⊥v1
This needs
So
w2 = u2 − hu2 , v1 iv1 , and w2 ⊥v1
Yes, otherwise
u2 = hu2 , v1 iv1
w3
so that w3 ⊥v1 and w3 ⊥v2 ; then take v3 = ||w3 || .
We will use
Z 1
hu, vi = u(t)v(t) dt
0
u1 (t) ≡ 1,
u2 (t) ≡ t,
u3 (t) ≡ t2 .
Z 1 1/2
||u1 || = 1 dt = 1.
0
We take v1 := u1 = 1. Now
1
w2 = u2 − hu2 , v1 iv1 = t − hu2 , v1 i1 = t − ,
2
as here
Z 1
1
hu2 , v1 i = t dt =
0 2
" 3 #1 " 3 #
Z 1 2
1 1 1 1 1 1
||w2 ||2 = t− dt = t− = − −
0 2 3 2 3 8 2
0
1
= ,
3·4
so
1
||w2 || = √ ,
2 3
and so
w2 (t) √ 1 √
v2 (t) = =2 3 t− = 3 (2t − 1) .
||w2 || 2
Z 1 √ √ Z 1
2 3 2
hu3 , v2 i = t · 3 (2t − 1) dt = 3 2t − t dt
0 0
4 1
√ t t3 √ 1 1 √ 1 1
= 3 2 − = 3 − = 3 = √
4 3 0 2 3 6 2· 3
2 1 1 √
w3 (t) = t − · 1 − √ 3 (2t − 1)
3 2· 3
1
= t2 − t + .
6
In the next slide we use the observation that
1 1 2−3
− =
6 4 12
Z 1 2
2 2 1
||w3 || = t −t+ dt
0 6
Z 1 2
1 2 1
= (t − ) − dt
0 2 12
Z 1/2 2 Z 1/2 2
1 1
= s2 − ds = 2 s2 − ds
−1/2 12 0 12
Z 1/2
4 1 2 1
= 2 s − s + ds
0 6 144
4 2
1/2
s 1s 1
= 2 s − +
5 6 3 144 0
So
√
v3 := 6 5w3
√ 2
v3 = 5 6t − 6t + 1 .
||vi || = 1 ∀i
hvi , vj i = 0 ∀i 6= j
−− v1T −− | | ...
MBT MB = −− v2 −− v1 v2 ...
T
... | |
v1T v1 v1T v2 ... 1 0 ...
= T T
v2 v1 v2 v2 ... = 0 1 ... = In×n
... ... ...
1. A is orthogonal
Ax = x1 a1 + .... + xn an
If AT A = I
So
(1) =⇒ (3): Take u = v , then
(1) =⇒ (4): Since ||Av|| =||v|| for all v, then with θ the angle
between u and v and ϕ the angle between Au and Av :
So
ϕ = θ (or 2π − θ, same thing really).
(4) =⇒ (2):
1, if i = j,
ai · aj = (Aei ) · (Aej ) = ei · ej =
0, if i 6= j.
1 2 1
(Au) · (Av) = ||Au + Av|| − ||Au − Av||2
4 4
1 1
= ||A(u + v)||2 − ||A(u − v)||2
4 4
1 2 1
= ||(u + v)|| − ||(u − v)||2
4 4
... by isometry
=u·v
A, B orthogonal =⇒ AB orthogonal
So
(AB)−1 = (AB)T .
Here A is 2 × 2. Write
cos θ
a1 = Ae1 = (1 col. of A) =
st ,
sin θ
or:
cos(θ − π2 ) sin θ
Ae2 = =
sin(θ − π2 ) − cos θ
e2
a1
+
2 e1
a2
e2
a1 ! a2
2
e1
cos θ − sin θ
A = [a1 , a2 ] = and det A = cos2 θ + sin2 θ = 1.
sin θ cos θ
Ae1 e2
Ae2
e1
cos θ sin θ
A = [a1 , a2 ] = and det A = − cos2 θ−sin2 θ = −1
sin θ − cos θ
e2
Ae1 Ae2
ɽ/2
x
e1
So
2 det(A − I) = 0 : det(A − I) = 0.
So
2 det(A + I) = 0 : det(A + I) = 0.
Orthogonal means M −1 = M T
Preservation of geometry
Group property: AB and A−1 are orthogonal if A, B are.
det(M ) = ±1 and |λ| = 1 for λ any eigenvalue (could be
complex!)
For vectors in R2 an orthogonal matrix transformation is a
rotation or a reflection
If det = +1, then there is an eigenvector v with Av = v
If det = −1, then there is an eigenvector v with Av = −v
Two 3 × 3 examples
Complex scalars and vectors
Complex Orthogonality
Story so far ...
MBT MB = I, or
MB−1 = MBT .
M −1 = M T .
Av = v.
Consider u ∈ P with u ⊥ v.
Then by conformality Au ⊥ Av = v
So Au ⊥ v.
A maps P into P
Au
(x)E = MB (x)B .
Then
Do we take
B := [a, b, v3 ] or B := [b, a, v3 ]?
AEE
T = MB ABB T EE
T MB = MB AR MB
T
√
0 1 1 1 −1 0 0 0 2
1
= √ 0 −1 1 1 1 0 1 −1 0
2 2 √ √
2 0 0 0 0 2 1 1 0
= ...
√ √ √
1 + 2 −1 + 2 2
1 √ √ √
= √ −1 + 2 1 + 2 − 2
.
2 2 √ √
− 2 2 2
MA212 – Lecture 27 – Tuesday 23 January 2018 page 15
Illustration: We want preservation of orientation.
det = +1 =⇒ (∃v) Av = v
det = −1 =⇒ (∃v) Av = −v
8−9 −1 4 x1 0
−4
1 −8 − 9 x2 = 0
4 4 −7 − 9 x3 0
−1 −1 4 x1 0
1
−17 −4 x2 = 0
4 4 −16 x3 0
−18x2 = 0 : x2 = 0.
1
MA212 – Lecture 27 – Tuesday 23 January 2018 page 22
Cont’d
Take
4
v3 = 0
.
Then
−1
1
Au =Ae2 = 2 nd
col = −8
.
9
4
−1 8
θ = cos − .
9
So
T r(B) = 1 + 2 cos θ
= T r(A) ... (by similarity)
1 7
= (8 − 8 − 7) = −
9 9
7 16 8
2 cos θ = − − 1 = − : cos θ = − .
9 9 9
" ! ! "
0 "
2
8/9
!"
"
MA212 – Lecture 27 – Tuesday 23 January 2018 page 28
2 !"
Question: is this clockwise or anticlockwise?
The upshot is that we know the angle, but not its direction.
To settle this question we need another Recap from MA100.
For vectors a, b, c
e e e e
1 2 3
c = a × b = a = a1 a2 a3
b b1 b2 b3
has
||c|| = ||a|| · ||b|| · | sin θ|.
v3
v2
c
b
v1
a
v2
b
b sin !
!
a v1
1
...because 9 > 0. So
√
v 17 v
c = a × b = sin θ =
||v|| 9 ||v||
The conjugate is
z̄ = a − ib.
So, taking c = a, d = −b :
z z̄ = (a + ib)(a − ib) = a2 + b2
1 z̄ z̄ a − ib
= = 2 = 2 2
.
z z z̄ |z| a +b
Example
z 7→ z̄ (a + ib) 7→ (a − ib)
(−i)2 = i2 = −1,
(z + w) = z̄ + w̄,
(zw) = z̄ · w̄,
z 1 2 1 3
e = 1 + z + z + z + ...,
2 3!
which mathematicians usually take as the definition of ez , a
point we revisit later.
yields
iθ 1 2 1 3
e = 1 + iθ − θ − iθ + ...
2 3!
1 2 1 3
= (1 − θ + ...) + i(θ − θ + ...)
2 3!
= cos θ + i sin θ.
(cos θ, sin θ)
and likewise
z̄ = re−iθ .
|z + w| ≤ |z| + |w|.
We define
z1
..
Cn = . : z1 , ..., zn ∈ C .
zn
det(Ā) = det A
What can be done with real scalars can be done with complex
scalars. Watch an example.
Solve Az = b where
1 0 i z1 2+i
−i 1 1 + i z2 = 3 − 2i .
0 i −1 z3 2i
1 0 i 2+i
0 1 i 2
0 i −1 2i R3 → R3 − iR2
1 0 i 2+i
0 1 i 2
0 0 0 0 R3 → R3 − iR2
z1 2 + i − iz3
z1 + iz3 = 2 + i
=⇒ z2 =
2 − iz3
z2 + iz3 = 2
z3 z3
2+i −i
=
2 + z3 −i
for z3 ∈ C
0 1
Referring to
1 0 i
B :=
0 1 i = [b1 , b2 , b3 ]
0 0 0
1 0 i
1 1+i −i 1
det(A) = −i 1 1 + i = 1
+ 0 + i
i −1 0 i
0 i −1
= −1 − i(1 + i) + i(−i2 ) = −1 − i − i2 + i = 0.
Recall that
|z|2 = z z̄ = a2 + b2 .
h·, ·i : V × V → C
Hermitian property:
Positivity:
hu, ui > 0 all u 6= 0 ∈V.
MA212 – Lecture 27 – Tuesday 23 January 2018 page 62
Comment
Note the Hermitian property implies that hu, ui =hu, ui for all
u ∈V ; so hu, ui ∈ R ; but we want more than just being real.
u · v = uT v
= u1 v1 + u2 v2 + · · · + un vn
so
P
u·u= |ui |2 ,
u · u > 0 for u 6= 0 .
For h·, ·i : V × V −→ C
Linearity:
hαu1 + βu2 , vi = αhu1 , vi + βhu2 , vi , ∀u1 , u2 , v ∈ V ,
∀α, β ∈ C
Hermitian property: hv, ui = hu, vi , ∀v, u
Positivity:
hu, ui > 0 for all u 6= 0
N.B.
hu, ui = hu, ui so hu, ui is real, but we want more (positivity).
1 1
v1 = √2
1
i
, v2 = √1
3
−i what are their norms?
0 1
Since 1 · 1 + i · i + 0 = 1 + (i)(−i) = 1 + 1 = 2 ,
||v1 || = 1 .
Since 1 · 1 + (−i) · −i + 1 · 1 = 1 − (i)(i) + 1 = 1 + 1 + 1 = 3 ,
||v2 || = 1 .
Moreover, v1 · v2 = √1 √1 (1 · 1 + i(−i) + 0 · 1) = 0 ,
2 3
so v1 ⊥ v2 .
0 1 0
0,
So hu2 , v1 i − αhv1 , v1 i = 0 ,
yielding:
α = hu2 , v1 i.
1
hu2 , v1 i = − √ {1 + (−i)(−i) + 0} = 0.
2
So
1 1 1
1 1
w2 =
−i − √2 {1 + (−i)(−i) + 0} i √2 = −i .
1 0 1
So u2 ⊥ v1 , as wanted!
Hence kw2 k2 = 1 + |(−i)|2 + 1 = 3 . Rescaling yields
1
v2 = √3
1
−i
.
1
0
Now work on u3 = 1
.
0 1
1 1 1
hu3 , v1 i = 1 · √ i = √ (0 + i + 0) = − √ i
2 2 2
0 0
and
0 1
1 1 i
hu3 , v2 i = 1 · √ −i = √ (0 + 1(−i) + 0) = √ .
3 3 3
0 0
0 1 1
i 1 i 1
w3 = 1 − (− √ ) √ i − ( √ ) √ −i
2 2 3 3
0 0 1
0 i i
1 1
=
1 + 2 −1 − 3 1
0 0 i
i/6
1 1 3−2
= 1/6 as − = .
2 3 6
−i/3
i/6 i
1
= 1/6 = 1
.
6
−i/3 −2i
1 6
Now kw3 k2 = 36 (|i|2 + 12 + |(−2i)|2 ) = 36 = 16 , so
i
w3 1
v3 = =√ 1
.
kw3 k 6
−2i
αv3 ⊥ v1 , v2 .
The multiple:
−1
iv3 = √6
1
i
Define also:
R1
hf1 , f2 i = 0 f1 (t)f2 (t)dt
Z 1
hf, gi = f (t)g(t)dt
0
Z 1
= f (t)g(t)dt
0
Z 1
= g(t)f (t)dt = hg, f i
0
Z 1 Z 1
hf, f i = f (t)f (t)dt = |f (t)|2 dt > 0 unless f (t) ≡ 0
0 0
Then
T
M B MB = I so MB∗ MB = I.
vT1 vT1 v1 vT1 v2 ...
T T T
T v 2 v2 v1 v2 v2 ...
M B MB = . [v1 , v2 , . . . , vn ] = .
. . .
.
. . . ...
vTn ... ... vTn vn
hv , v i hv2 , v1 i ...
1 1
hv1 , v2 i ... ...
= =I
.. ..
. . ...
... ... hvn , vn i
A−1 = A∗ .
√1 √1 −1 √1 −i
√ √ 0
2 3 6 2 2
√i √1
A= −i
√ √i , then A∗ = √i √1 .
2 3 6 3 3 3
√1 √2 −1 −i √2
0 3 6
√
6
√
6 6
1 1 −1
1 1 1
det(A) = √ √ √ det i −i i
2 3 6
0 1 2
1
= ((−2i − i) − i(2 + 1))
6
= −i.
So | det(A)| = 1 .
1. (A∗ )∗ = A
2. (AB)∗ = B ∗ A∗
3. det(A∗ ) = det(A)
4. The Star property: x · (A∗ y) = (Ax) · y
Indeed
T T
LHS = x (A y)
T
T
=x A y
= (xT AT )y
= (Ax)T y
= (Ax) · y = RHS
Theorem:
Equivalent characterizations for an n × n matrix:
1. A is unitary;
2. columns of A are an orthonormal basis for Cn ;
3. isometry: kAuk = kuk ;
4. conformality,
and in fact (Au) · (Av) = u · v .
T
I = A A = AT A, i.e. AT A = I.
Hence (with v = u ),
Proof:
Let Ax = λx , with x 6= 0 . Then
So λλ = 1 , as x · x = kxk2 > 0 .
AA∗ = A∗ A
As U −1 = U ∗ .
Now we meet a very significant member of the normal family:
A is Hermitian if A∗ = A
An example:
1 i −1
−i 0 3 + 2i
−1 3 − 2i 2
If A = A∗ then A∗ A = A2 = AA∗ .
So if A is Hermitian, then A is normal.
Similarly for A skew-hermitian: i.e. when
A∗ = −A (1)
0 2 0 −2
A= , then ∗
A = = −A
−2 i 2 −i
4 −2i
∗
Check: AA =
2i 5
Theorem (group-property)
Theorem
Let An×n be Hermitian. Then:
1. All eigenvalues are real
2. Eigenvectors to distinct eigenvalues are orthogonal.
x · (Ax) = x · (A∗ x)
x · (λx) = (Ax) · x
Thus
λx · x = λx · x
(Ax) · y = x · (A∗ y) = x · Ay
λx · y = x · (µy) = µx · y = µx · y
Normal if M M ∗ = M ∗ M
Unitary if M ∗ M = I i.e. M −1 = M ∗
Hermitian if M ∗ = M
Example of diagonalization
Upper triangular form
Background readings
If D is a diagonal matrix:
d1
D=
d2 0
..
.
0 dn
A is diagonalizable iff
P −1 AP = D ‘Similar to D ’
Avi = λi vi
Then A is similar to
λ1
D=
λ2 0
..
.
0 λn
MA212 – Lecture 29 – Tuesday 30 January 2018 page 6
Cont’d
B = (v1 , . . . , vn ).
so
D = MB−1 A MB
Check:
λj ej = D ej = MB−1 A MB ej
= MB−1 Avj (as vj = MB ej )
= MB−1 λj vj
= λj MB−1 vj
= λj ej (again as vj = MB ej ).
if A v = λv v 6= 0
then A v = λv but A is real so
Av = Av = λv and v 6= 0.
Consider
−1 0 −2
A=
0 3 2
1 −3 0
pA (x) = det(x I − A)
x + 1 0 2
= 0 x − 3 −2
−1 3 x
x − 3 −2 0 x − 3
= (x + 1)
+ 2
3 x −1 3
= (x + 1)(x2 − 3x + 6) + 2(x − 3)
= x3 − 2x2 + 5x
= x(x2 − 2x + 5)
Roots: x=0
√
2 ± 4 − 20
x= = 1 ± 2i
2
λ1 = 0, λ2 = 1 + 2i λ3 = λ2 = 1 − 2i.
λ1 = 0 solve Av = 0
6t
v1 = 2t
−3t
6
=t2
−3
Say t = 1 ...
6
v1 =
2
−3
MA212 – Lecture 29 – Tuesday 30 January 2018 page 14
Getting v2
λ2 = 1 + 2i solve A v = λ2 v
(λ2 I − A)v = 0
2 + 2i 0 2 x1 0
0 −2 + 2i −2 x2 = 0
−1 3 1 + 2i x3 0
(1 + i)x1 + x3 = 0 (1)
(−1 + i)x2 − x3 = 0 (2)
Put x3 = −t
−t −t(−1 − i)
(2) ⇒ x2 = =
−1 + i 2
1+i
= t (−1 + i)(−1 − i) = 2
2
t t(1 − i)
x1 = = (1 + i)(1 − i) = 2
1+i 2
yielding
1−i 1−i
t
v2 = 1 + i
say (t = 2) 1 + i
2
−2 −2
MA212 – Lecture 29 – Tuesday 30 January 2018 page 16
v3 for free from v2
Take v3 = v2
1−i 1−i
t
v2 = 1 + i
say (t = 2) 1 + i
2
−2 −2
so take
1+i
v3 = v2 =
1 − i
−2
Take
P = v1 , v2 , v3
6 1−i 1+i
=
2 1 + i 1 − i
−3 −2 −2
Then
A P = (A v1 , A v2 , Av3 )
= (λ1 v1 , λ2 v2 , λ3 v3 )
So that
λ1 0 0
A P = P D = (v1 , v2 , v3 )
0 λ2 0
0 0 λ3
λ1 0 0
∴ P −1 A P =
0 λ2 0
.
0 0 λ3
0 0 0
−1
P A P = 0 1 + 2i 0 .
0 0 1 − 2i
4 4 4
−1 1
P = −3 + i −3 − 9i −8 − 4i
20
−3 − i −3 + 9i −8 + 4i
(Here A is diagonalizable.)
Question: In general, when can this be done?...
E(A, λ) = {v : A v = λv}
This includes 0 .
But v ∈ E(A, λ) and v 6= 0 ⇒ v is an eigenvector.
FACT
If pA (x) = (x − λ1 )m1 . . . (x − λk )mk , then
1 ≤ dimE(A, λj ) ≤ mj
| {z } |{z}
geometric multiplicity algebraic multiplicity
E(A, λ) = {v : (A − λI)v = 0}
= N (A − λI)
= Null space of A − λI.
4 0 6
A = −2 1 −5
−3 0 −5
x − 4 0 −6
x − 4 −6
pA (x) = 2 x−1 5 = (x − 1)
3 x + 5
3 0 x + 5
= (x − 1)(x2 + x − 2)
= (x − 1)(x − 1)(x + 2)
= (x − 1)2 (x + 2)
solve(A − I)v = 0
3 0 6
−2 0 −5 v = 0
−3 0 −6
dimE(A, λ) ≥ dimE(B, λ)
... why?
Answer: Write A for B and B for A
dimE(B, λ) ≥ dimE(A, λ)
−1
Aui = P B P P −1 vi
= P −1 B vi
= P −1 λvi
= λui .
So, ui ∈ E(A, λ) .
We now show:
ui are linearly independent
If α1 ui + · · · + αk uk = 0 , then, as v 7→ P v is linear,
α1 P u1 + · · · + αk P uk = 0
iff α1 v1 + αk vk = 0.
1 1
A=
0 1
x − 1 −1
pA (x) =
0 x − 1
= (x − 1)2
0 1
A−I =
0 0
Av = v ⇒ x2 = 0
x1
v=
0
1
E(A, 1) = Lin of dimension 1
0
1 1
A=
0 1
not diagonalizable?
If it were λ1 = λ2 = 1
1 1 1 0
= P −1 P
0 1 0 1
= P −1 I P = I
1 0
=
0 1
T =
d2
U
..
.
0 dn
N.B.
pT (x) = (x − d1 )(x − d2 ) · · · (x − dn )
etc.
Extend to a basis of Rn :
B = (u1 , . . . , un )
a
‘Proof by mathematical induction’
MA212 – Lecture 29 – Tuesday 30 January 2018 page 39
Cont’d
Put
A′ = MB−1 A MB
A′ ej = MB−1 A MB ej
= MB−1 A uj
= MB−1 λuj
= λMB−1 uj
= λ ej
1≤ k
|{z} ≤ m
|{z}
Geometric multiplicity algebraic multiplicity
C −1 Q C = U
Put
1
..
. 0
1
N =
0 C
Ik 0 λIk P Ik 0
0 C −1
0 Q
0 C
λIk P
Ik 0
=
−1
0 C Q
0 C
λ
..
. PC
λ
=
0 U
is upper triangular.
A is similar to A′
A′ is similar to upper triangular.
∴ A is similiar to upper triangular.
The matrix
2 2 −1
A = −1 −1 1
−1 −2 2
has
pA (x) = det(x I − A) = (x − 1)3 .
1
v= 0
is an eigenvector
using v for R3
say,
1 0 0
v= 0
, e2 = 1 , e3 =
0
,
1 0 1
2 2 −1 1 0 0
A MB = −1 −1 1 0 1 0
−1 −2 2 1 0 1
1 2 −1
= 0 −1 1
1 −2 2
1 0 0
M B = 0 1 0
1 0 1
−1 0 1
(So, you get this by the opposite inverse action, i.e., subtract row
1 from row 3 )
1 0 0 1 2 −1
MB−1 A MB =
0 1 0 0 −1 1
−1 0 1 1 −2 2
1 2 −1
=
0 −1 1 ( = A′
)
0 −4 3
1
A1 v = v ⇒ v =
2
1 0
2 1
1 0 0 1 2 −1 1 0 0
0 1 0 0 −1 1 0 1 0
( = N −1 ′
AN )
0 −2 1 0 −4 3 0 2 1
1 0 0 1 0 −1
=
0 1 0 0 1 1
MA212 – Lecture 29 – Tuesday 30 January 2018 page 56
0 −2 1 0 2 3
So finally
1 0 0 1 0 −1
=
0 1 0 0 1 1
0 −2 1 0 2 3
1 0 −1
=
0 1 1
0 0 1
P −1 AP = diag{1, 1}
1 0
= =I
0 1
1 1 1 0
= A = P IP −1 = I = ,
0 1 0 1
oops, a contradiction.
1 1 x1 x1 + x2 x1 x2
= =I + .
0 1 x2 x2 x2 0
Here
x2 0 1 x1
=
0 0 0 x2
and
0 1
S=
0 0
ݔ2
ݔ Aݔ
ݔ2
Sݔ
ݔ1 ݔ2 ݔ1
‘most stretched’
‘least stretched’
MA212 – Lecture 30 – Friday 2 February 2018 page 6
More generally
λ 1 x1 λ 0 x2
Ax = = x+ = λIx + Sx
0 λ x2 0 λ 0
λ 1
J2 = is called a 2 × 2 Jordan block.
0 λ
ʄIݔ
ݔ
ݔ2
Sݔ
ݔ1
ݔ2
and
λ 1 0 λx1 + x2 x2 0
0 λ 1 x = λx2 + x3 = λIx+ 0 + x3
0 0 λ λx3 0 0
Ix
3 x
2
3
2
1
λ 1 0
0 λ 0
←− shearing absent here
0 0 λ
λ 0 0
←− shearing absent here
0 λ 1
0 0 λ
λ 0 0
←− shearing absent here
0 λ 0
←− shearing absent here
0 0 λ
MA212 – Lecture 30 – Friday 2 February 2018 page 11
Decomposition
f ′ = Af = λf
d λt
f (t) = e as e = λeλt .
λt
dt
d λt
te = eλt + λteλt : Ae2 = λe2 + e1
dt
The second term e1 here acts as a shear on e2 .
more generally:
x1 y1 ...
x2 y2 ...
A[x, y, ...] = [a1 , a2 , ..., an ]
.. .. ..
. . .
xn yn ...
λ 1 Je1 = λe1
λ 1 Je2 = e1 +λe2
λ 1 ...
J = .. =⇒
.
λ 1 Jem−1 = em−2 +λem−1
λ Jem = em−1 +λem
m×m
then
(A − λ1 I)(A − λ2 I)...(A − λn I) = O.
0 −1 x 1
A=
=⇒ pA (x) = = x2 + 2x + 1.
1 −2 −1 x + 2
0 −2 0 −1 0 −1 −1 2
2A = , A2 = =
2 −4 1 −2 1 −2 −2 3
−1 0
A + 2A =
2 = −I.
0 −1
Yesss! A2 + 2A + I = O.
U = P −1 AP.
λ1 − λj × × × ... ×
λ2 − λj × × ... ×
... × ... ×
U − λj I =
0 = λj − λj ... ...
... ×
λn − λj
So
and
(U −λn−2 I)(U −λn−1 I)(U −λn I)Cn = (U −λn−2 I)Lin{e1 , ..., en−2 }
(A − λ1 I)(A − λ2 I)...(A − λn I) = O
Substituting U = P −1 AP and I = P −1 P ,
O = (U − λ1 I)(U − λ2 I)...(U − λn I)
after cancellation.
Now pre- and post-multiply resp. by P and P −1
Finally
O = (A − λ1 I)(A − λ2 I)...(A − λn I).
(A − λI)u= 0 ∀u ∈ R2
(A − λI) ≡ O
A = λI,
and write v1 = v.
Av = λv Av1 =λv1
Au = λu + v Av2 =λv2 +v1
P = [v1 , v2 ]
Then
λ 1
P −1
AP = .
0 λ
Recall that
0 −1 x 1
A=
=⇒ pA (x) = = x2 +2x+1 = (x+1)2 .
1 −2 −1 x + 2
So λ = −1 twice.
Find u ∈N
/ (A − λI) = N (A + I), i.e. u with (A + I)u 6= 0,
i.e. u = (u1 , u2 )T such that
1 −1
u1 + u2 6= 0.
1 −1
Take
1 1 1 0 −1
P = [v1 , v2 ] = and so P −1 =
1 0 det P −1 1
working in R3 ....
Main focus here is on λ repeated three times pA (x) = (x − λ)3 .
NB The case pA (x) = (x − λ)2 (x − µ) reduces to
λ 1 0 λ 0 0
0 λ 0 or 0 λ 0 .
0 0 µ 0 0 µ
Case: (A − λI)2 6= O
2
Pick v3 = u with u ∈
/ N (A − λI)
Then
v1 := (A − λI)v2 = (A − λI)2 v3 6= 0.
Then
Av1 = λv1
Av2 = λv2 + v1
Av3 = λv3 + v2
Then
Then
λ 1 0
P −1
AP =
0 λ 1 .
0 0 λ
(A − λI)2 = O
Possibilities:
i) dim N (A − λI) = 3 =⇒ A = λI, as before (see below)
ii) dim N (A − λI) = 2... see below
(A − λI)u = 0 ∀u : Au = λu ∀u : A = λI.
(A − λI)[(A − λI)u]= 0 ∀u
(A − λI)v= 0 ∀v ∈ R(A − λI)
So
R(A − λI) ⊆ N (A − λI) =⇒ dim N (A − λI) ≥ 2,
a contradiction.
Here
λ 0 0 λ 0 0
P −1
AP =
0
= 0
λ 1
J2
0 0 0 λ
1 2 1
A = 0 2 0
−1 2 3
x − 1 −2 −1
(expand by middle row) 0 x−2 0
1 −2 x − 3
Here
(A − 2I)2 = O.
Pick u = v3 = e1 with
1
v2 := (A − 2I)e1 ∈Lin 0 6= 0 :
Av3 = v2 + λv3
1
(A − 2I)x = 0
i.e.
−1 2 1 x1 0
0 0 0 x2 = 0
−1 2 1 x3 0
−x1 + 2x2 + x3 = 0.
So
2x2 + x3 2x2 x3
x =
x2 = x2 + 0
x3 0 x3
2 1
N (A − 2I) = Lin 1 , 0
0 1
Here the obvious choice is the first of the two (not the second)
2
v1 :=
1
0
2 −1 1
P = [v1 , v2 , v3 ] = 1 0 0
0 −1 0
Then
2 0 0
P −1
AP = 0 2 1
0 0 2
Briefly also in :
Adam Ostaszewski
§6.5
Bottom-up method:
λ eigenvalue: Algebraic multiplicity m = mA .
Geometric multiplicity mG ... with mG ≤ mA .
i.e. we can get only mG lin. indep. eigenvectors to value λ .
For each of these eigenvectors we arrange a stretch-and-shear
scheme:
Av1 = λv1
Av2 = v1 + λv2
...
Avk = vk−1 + λvk
(A − λI)k = O
(A − λI)k−1 6= O
(A − λI)k−1 vk 6= 0,
vk−1 = (A − λI)vk
vk−2 = (A − λI)vk−1
...
v1 = (A − λI)v2 = ... = (A − λI)k−1 vk 6= 0
yields:
(A − λI)v1 = (A − λI)k vk = 0.
P −1 AP = J
Block A1
Block A2
Block A3
J =
...
...
Block Am
λ 1
λ
Ai :=
..
. 1
λ
ẋ1 b (t)
1
ẋ2 d b2 (t)
ẋ =
..
where ẋi = xi = xi (t) and b =
′
..
.
. dt .
ẋn bn (t)
ẍ(t) = f (t).
x1 = x, and x2 = ẋ
then we get
ẋ1 = x2 and ẍ(t) = ẋ2 = f (t)
ẋ1 0 1 x1 0
= +
ẋ2 0 0 x2 f (t)
ẋ = Ax + b.
for
0 1
A = ANewton :=
0 0
and
0
b = bNewton :=
f (t)
i.e.
ẏ = Dy + c
So
d n −λi t o
yi e = ẏi e−λi t − λi yi e−λi t = ci (t)e−λi t
dt Z
yi e−λi t = ci (t)e−λi t dt
1 −1
ẋ = x
4 1
Here
x−1 1
pA (x) = = (x − 1)2 + 4
−4 x − 1
so λ = 1 ± 2i.
as
1
v1 = ,
−2i
1 1
P = [v1 , v2 ] =
−2i +2i
we get
1 + 2i 0
P −1
AP =
0 1 − 2i
we solve
1 + 2i 0
ẏ = Dy = y.
0 1 − 2i
x1 1 1 αe(1+2i)t
x = = Py =
x2 −2i +2i βe(1−2i)t
αet (cos 2t + i sin 2t) + βet (cos 2t − i sin 2t)
=
−2iαet (cos 2t + i sin 2t) + 2iβet (cos 2t − i sin 2t)
t (α + β) cos 2t + i(α − β) sin 2t
x=e ,
2i(β − α) cos 2t + 2(α + β) sin 2t
We obtain
t K cos 2t + L sin 2t
=e
−2L cos 2t + 2K sin 2t
by taking
α+β =K ∈R 2β = K + iL
=⇒
i(α − β) = L ∈ R 2α = K − iL
gives
x2 = K2 eλt
d n −λt o
x1 e = K2 so x1 e−λt = K2 t + K1
dt
x1 = (K2 t + K1 )eλt .
x1 K2 t + K1 1 t K1
= eλt = eλt .
x2 K2 0 1 K2
that
(P −1 AP )k = P −1 Ak P.
For instance,
(P −1 AP )3 = (P −1 AP )2 (P −1 AP ) = P −1 A2 (P P −1 )AP
= P −1 A3 P
t 1 2 1 3
e := 1 + t + t + t + ...
2! 3!
and the series converges for every t.
we define:
A 1 2 1 3
e = I + A + A + A + ...
2! 3!
1 1
(I)ij + (A)ij + (A)ij + (A)3ij + ...;
2
2! 3!
one needs to verify that for a fixed A this sum converges; we
explain why so later.
λk1
λk2
P A P = (P AP ) = D =
−1 k −1 k k
..
.
.
λkn
So
−1 A −1 −1 1 −1 2 1 −1 3
P e P =P P +P AP + P A P + P A P + ...
2! 3!
−1 A 1 2 1 3
P e P = I + D + D + D + ... = eD
2! 3!
1 λ
1
1 λ2
=
.. + ..
. .
1 λn
λ21
1 λ22
+ + etc.
2!
..
.
λ2n MA212 – Lecture 31 – Tuesday 6 February 2018 page 33
Summing up:
1 2
1 + λ1 + 2! λ1 + ...
1 2
1 + λ2 + 2! λ2 + ...
..
.
1 2
1 + λn + 2! λn + ...
eλ1
eλ2
=
..
.
.
eλn
−1 A 1 2 1 3
P e P = I + J + J + J + ... = eJ
2! 3!
This is easy:
Ak1
Ak2
J =
..
.
.
Akn
Taking P −1 AP = J, we solve
ẋ = Ax
ẏ = Jy.
tJ 1 2 2 1 3 3
e = I + tJ + t J + t J + ...
2! 3!
d tJ 1 1
{e } = J + tJ 2 + t2 J 3 + ...
dt 1! 2!
1 1 2 2
= J I + tJ + t J + ... = JetJ
1! 2!
1 1 2 2
or = I + tJ + t J + ... J = etJ J.
1! 2!
d d tJ
ẏ(t) = y(t) = {e y(0)} = JetJ y(0) = Jy(t),
dt dt
as before.
by substituting x = P y into
ẋ = Ax + b
gives
ẏ =Jy+P −1 b =Jy + c ẏ−Jy = c
d −tJ
e y(t) = e−tJ ẏ−e−tJ Jy =e−tJ c
dt Z
e−tJ y(t) = e−tJ c(t) dt.
We will soon see that this integral also has a nice formula!
λ 1 λ 1
λ 1 λ 1
J2 =
..
..
. 1 . 1
λ λ
λ2 2λ 1 0 0...
λ2 2λ 1 ...
= λ2 ...
..
.
λ2
λ2 2λ 1 0 0...
λ 1
λ2 2λ 1 ...
λ 1
=
J3 = λ2 ... ..
. 1
..
.
λ
λ 2
λ3 3λ2 3λ 1 0...
λ3 3λ2 3λ 1...
= λ3 ...
..
.
λ3
For good reason: these repeated items are exactly as the terms
in: (λ + 1)3 = λ3 + 3λ2 + 3λ + 1
Starting from its diagonal position, each row exhibits the terms:
λ3 , 3λ2 , 3λ, 1 in that order, for as long as there is space in the
row available for all the terms.
In any row each inner element ‘folds in’ (sums) the two elements
above.
Compute
tJ t2 2 t3 3
e = I + tJ + J + J + .
2! 3!
X tk
k−m k(k − 1)...(k − m + 1)
λ
k! m!
k≥m
tm X k−m 1
= (tλ) [put: ℓ = k − m]
m! (k − m)!
k≥m
tm X ℓ1
= (tλ)
m! ℓ!
ℓ=0
m
λt t
= e
m!
t2 t3
1, t, 2! , 3! , ...
t2 λt t3 λt
1eλt , teλt , 2! e , 3! e , ...
λ 1 eλt teλt 1 t
J = , e tJ
= = eλt
0 λ 0 eλt 0 1
λ 1 0
J = 0 λ 1
,
0 0 λ
λt λt t2 λt 2
e te 2! e 1 t t /2!
e tJ
=
0 e λt
te λt = eλt 0 1
t
0 0 eλt 0 0 1
λ 1
λ 1
etJ = exp t
..
. 1
λ
2 3
1 t t /2! t /3!
1 t 2 3
t /2! t /3!
1 t 2 3
t /2! t /3!
= eλt
1 t 2
t /2!
..
. ... ...
1
MA212 – Lecture 31 – Tuesday 6 February 2018 page 58
Summary
To solve
ẋ = Ax
Dominant eigenvalue
Long-term forecasts
Recap: differential equations
For
λ 1
λ 1
J =
..
. 1
λ
ẏ = Jy
tJ
with y(0) initial data
=⇒ y(t) = e y(0)
ẋ = Ax, x = P y
y = P −1 x
ẏ = P −1 AP y
MA212 – Lecture 32 – Tuesday 9 February 2018 page 2
Cont’d
2
t
1 t 2 ···
etJ = eλt 1 t · · ·
..
.
−1 1 0
ẋ = 0 −1 1
x
0 0 −1
0 −1
Solve ẋ = x
1 −2
We saw in Lecture 30 slide 39 that for
1 1
P = MB =
1 0
−1 1
−1
P AP = =J
0 −1
Take
x = Py
P ẏ = AP y
−1 1
ẏ = P −1
AP y = y
0 −1
y(t) = etJ y(0)
−t 1 t A
1
y(t) = e
0 1 A2
A1
y(0) =
A2
MA212 – Lecture 32 – Tuesday 9 February 2018 page 7
so
x(t) = P y(t)
−t 1 1 1 t A1
=e
1 0 0 1 A2
x x (k)
1 1
. .
For x = .. , xk = .. , with k = 0, 1, 2, · · · . Consider
xn xn (k)
xk+1 = An×n xk
x0 given
Solution:
xk = Axk−1 = A(Axk−2 )
= · · · = Ak x0
λ1
..
P −1 AP = D = .
λn
Put
P yk = xk
P yk = Axk−1 = AP yk−1
yk = P −1 AP yk−1 = Dyk−1
yk = Dk y0
λk1
(1)
..
= . y0
λkn
But P yk = xk , y0 = P −1 x0
xk = P yk
λk1
(2)
.. −1
=P . P x0
λkn
more later on xk .
λk2 λ2 k
k
= ( 1 ) −→ 0 as k −→ ∞,
λ1 λ
λ3
and similarly for λ1 etc.
xk = P Dk y0
λk1 α
k
λ2 β
=P ..
.
λkn γ
α
λ2 k
( ) β
k λ1
= λ1 P .
..
( λλn1 )k γ
Here
y0 = P −1 x0
α
β
=
..
.
γ
xk = P Dk y0
α
∗
k
= λ1 P .
..
∗
= αλk1 v1 , approximately, as P e1 = v1 (nb Av1 = λ1 v1 ) and
α 1
∗ 0
≈ α = αe1
.. ..
. .
∗ 0 MA212 – Lecture 32 – Tuesday 9 February 2018 page 18
Subdominant eigenvalue:
x −7 −6
PA (x) = det(xI − A) = − 41 x 0
0 −2 x
1
1 6 7
= (x · 0 − ) + x(x2 − )
2 4 4
3 7 3 2 3
= x − x − = (x + 1)(x − x − )
4 4 4
3 1
= (x + 1)(x − )(x + )
2 2
3
2 −7 −6 x1
− 1 3
0 x 2 = 0 (3)
4 2
0 − 12 23 x3
−x1 + 6x2 = 0
−x2 + 3x3 = 0
18
v1 =
3
1
P = v1 , · · ·
x1 (0) α
P x2 (0)
−1
= β = y0
x3 (0) γ
γk 18
s ≈ ( 3 )k α 3
k 2
tk 1
Here
P −1 AP = J
As before sub
P yk = xk
yk = J k y0
A
1
A2
J =
..
.
An
Ak1
Ak2
J =
k
..
.
Akn
so xk = P J k P −1 x0
2 1 −2
A = 1 1 −1
1 0 0
x − 2 −1 2
pA (x) = −1 x − 1 1 = (x − 1)3
−1 0 x
So (A − I)3 = O .
1 1 −2
A − λI = A − I =
1 0 −1
1 0 −1
0 1 −1
(A − λI) = (A − I) =
2 2
0 1 −1
6= 0
0 1 −1
0
v3 := e2 = 1
“picks up” 2nd col.
0
v3 = 1
( = e2 )
0
1
v2 = (A − I)v3 =
0 ( = e1 ) = 2nd col. “pick up” from (A − I)
0
1 1 −2
as A − λI = 1 0 −1 .
1 0 −1
1
v1 = (A − I)v2 = 1
,
1
1 1 −2 1
as A − λI = 1 0 −1 and v2 = 0
( = e1 ).
1 0 −1 0
1 1 0
P = v1 v2 v3 = 1 0 1
1 0 0
1 1 0
J = P AP = 0 1 1
−1
0 0 1
Recall
k−2
k k−1 k
λ kλ 2 λ
k
Jλ = 0 λ k
kλ k−1 (4)
0 0 λk
k k−1 k k−2
(λ + 1)k = λk + λ + λ + ···
1 2
so
1
1 k 2 k(k − 1)
J =
k
0 1 k
(5)
0 0 1
Back to the xk
xk = P J k P −1 x0
1
1 1 0 1 k 2 k(k − 1) 0 0 1 a
(6)
= 1 0 1
0 1 k
1 0 −1 b
1 0 0 0 0 1 0 1 −1 c
0 0 1 a c
y0 = 1 0 −1 b = a − c
(7)
0 1 −1 c b−c
As λ1 = λ2 = λ3 = 1 here.
here, of course, of
1 1 0 c
k
xk = 1 0 1 J a − c
(8)
1 0 0 b−c
for ’large’ k
1
1 k 2 k(k − 1)
J =
k
0 1 k
(9)
0 0 1
1 1 2 1
k(k − 1) = k (1 − ) (10)
2 2 k
Influences xk ... see this by factorizing out k 2 :
1
1 1 0 ∗ ∗ 2 c
xk = k 1 0 1 0 ∗ ∗ a − c
2
1 0 0 0 0 ∗ b−c
(11)
1
2 (b − c)
=k P ∗
2
∗
1
21
xk = k (b − c)P
0+∗
2
0
(12)
1
21
= k (b − c)
1 +∗
2
1
Note that
1
v1 = 1
1
MA212 – Lecture 32 – Tuesday 9 February 2018 page 39
Subcase 2: b = c and so k 2 has no influence
If b = c , then
1
2 (b − c) 0
∗ = ∗ (13)
∗ ∗
1 0 0 b−c
for ’large’ k
1
1 k 2 k(k − 1)
J =
k
0 1 k
(15)
0 0 1
0 0 ∗ 0
a−c
= kP ∗
+∗
0
1 1 0 1
= k 1 0 1
0 (a − c) + ∗
1 0 0 0
MA212 – Lecture 32 – Tuesday 9 February 2018 page 42
That is
1 1 0 1
xk = k 1 0 1
0 (a − c) + ∗
1 0 0 0
1
= k(a − c)
1 + ∗
1
x = P y =⇒ yt+1 = Jyt , t = 0, 1, 2 · · ·
=⇒ y1 = Jy0
k
λk kλk−1 2 λk−2 · · ·
λk kλk−1 · · ·
J =
k
..
. · · ·
..
.
We are given
a
b
x0 =
c
.
..
∼ 1 2
xt = t (b − c) × eigenvector for the λ block
2
xt ∼
= t(a − c) × eigenvector
& Applications
About normality
Inner products and positive definiteness
Singular values (for non-square matrices)
Background reading
Adam Ostaszewski
Chap. 5 (§5.1-5.3, 5.5 5.6,)
Hermitian property:
Positivity:
hu, ui > 0 all u 6= 0 ∈V.
S ∗ AS = D
i.e. S −1 AS = D with S −1 = S ∗ .
When this happens the columns of S form an orthonormal basis
of Cn (as S ∗ S = I) and are eigenvectors of A as AS = SD, i.e.
λ1
A[s1 , ..., sn ] = [s1 , ..., sn ]
... = [λ1 s1 , ...,λn sn ].
λn
λ1 0
λ1 0
λ2 λ2
But, D =
..
and D∗ =
..
. .
0 λn 0 λn
A is normal if A∗ A = AA∗
A is unitary if A∗ = A−1
A is Hermitian if A∗ = A
hx, yi := y∗ Ax
yT AT x = yT ĀT x = yT A∗ x = yT Ax.
Ax = λx
Ax = A(α1 x1 + α2 x2 + · · · + αn xn )
= α1 Ax1 + α2 Ax2 + · · · + αn Axn
= α1 λ1 x1 + α2 λ2 x2 + · · · + αn λn xn .
Then x∗ Ax =
= (α1 x1 + α2 x2 + · · · + αn xn )∗ (α1 λ1 x1 + α2 λ2 x2 + · · · + αn λn xn )
= (α1 x∗1 + α2 x∗2 + · · · + αn x∗n )(α1 λ1 x1 + α2 λ2 x2 + · · · + αn λn xn )
= α1 α1 λ1 x∗1 x1 + α2 α1 λ1 x∗2 x1 + · · · + αn α1 λ1 x∗n x1
+ α1 α2 λ2 x∗1 x2 + α2 α2 λ2 x∗2 x2 + · · · + αn α2 λ2 x∗n x2
+ α1 αn λn x∗1 xn + α2 αn λn x∗2 xn + · · · + αn αn λn x∗n xn
= |α1 |2 λ1 + |α2 |2 λ2 + · · · + |αn |2 λn
MA212 – Lecture 33 – Tuesday 13 February 2018 page 17
>0
Determinantal test
a11 > 0
a11 a12
>0
a21 a22
a a a
11 12 13
a
21 a22 a23 > 0
a31 a32 a33
a11 a12 · · · a1n
a21 a22 · · · a2n
the full determinant of A . .. ..
.. = |A| > 0.
.. . . .
an1 an2 · · · ann
2 1 1
Consider the matrix A = 1 1 0
, its characteristic
1 0 3
polynomial is
x − 2 −1 −1
pA (x) = det(xI − A) = −1 x − 1 0
−1 0 x − 3
2 i 0
Consider the 2 × 3 complex matrix A =
0 −i 2
2 0
so A = −i i
∗
and
0 2
2 0 4 2i 0
2 i 0
∗
A A = −i i = −2i 2 2i
0 −i 2
0 2 0 −2i 4
is
x − 4 −2i 0
pA∗ A (x) = 2i x − 2 −2i
0 2i x − 4
So
x − 2 −2i 2i −2i
pA∗ A (x) = (x − 4) + 2i
2i x − 4 0 x − 4
= x(x − 4)(x − 6) so λ0 = 0, λ1 = 4, λ2 = 6.
A∗ Av = λv
Using
2 i 0
A=
0 −i 2
again, we have
2 0
2 i 0 5 −1
∗
AA =
−i i =
0 −i 2 −1 5
0 2
(5 − λ)2 − 1 = 0
(5 − λ)2 = 1 =⇒ 5 − λ = ±1
so λ = 5 ± 1 = 6, 4
√
Indeed, the singular values of A are 6 and 2 .
and then
MA212 – Lecture 33 – Tuesday 13 February 2018 page 32
and then
∗ ∗1 1 ∗ 1
AA ui = AA ( √ Avi ) = √ A(A Avi ) = √ A(λi vi )
λi λi λi
1
= λi ( √ Avi ) = λi ui
λi
1 1 1
ui · uj = ( √ Avi ) · ( √ Avj ) = p (Avi ) · (Avj )
λi λi λi λj
1 ∗ 1
=p (Avj ) Avi = p vj∗ A∗ Avi
λi λj λi λj
s
λi ∗ λi
= p vj vi = (vi · vj )
λi λj λ j
∴ ui · uj = 0 when i 6= j
unit length as
s r
λi λi
ui · uj = (vi · vj ) =⇒ ui · ui = (vi · vi )
λj λi
=⇒ kui k2 = kvi k2 = 1
2 i 0
Using A = again, we have
0 −i 2
5 −1
∗
AA =
−1 5
with eigenvalues 4, 6
1 1
λ1 = 4 =⇒ e-vector ui = √
2 1
1 −1
λ2 = 6 =⇒ e-vector ui = √
2 1
4 2i 0
A A = −2i 2 2i
∗
0 −2i 4
with e-values 0, 4, 6
1
1
λ1 = 4 =⇒ e-vector √ 0
2
1
−1
1
λ2 = 6 =⇒ e-vector √ i
3
1
−i
1
λ3 = 0 =⇒ e-vector √ 2
6
i
1
1 1 2 i 0
1 1 2 1 1
u1 = √ Av1 = √ 0 = √ = √
4 2 0 −i 2 2 2 2 2 2 1
1
−1
1 1 2 i 0 1
u2 = √ Av2 = √ √ i
6 6 0 −i 2 3
1
1 −3 1 −1
= √ =√
3 2 +3 2 +1
1 ∗
vi = √ A ui for 1 ≤ i ≤ k
λi
2 0 2 1
1 ∗ 1 1 1 1 1
v1 = √ A u1 = −i i
√ = √ 0
=√ 0
4 2 2 1 2 2 2
0 2 2 1
2 0 −2
1 ∗ 1
1
−1 1
v2 = √ A u2 = √ −i i √ = √ 2i
6 6 2 1 2 3
0 2 2
−1
1
=√ i
3
1
Adam Ostaszewski
Chap. 5
we also have
kA xk
kAk = max : x 6= 0 ,
kxk
Av = λv (v 6= 0)
and so
kAvk kλvk kvk
= = |λ| = |λ|
kvk kvk kvk
which means that the norm of A , i.e. kAk , must be at least as
large as the largest modulus than we can get from its
eigenvalues...
... But in general, it is larger! (see, for example, Exercise 8.2 :
√
1+ 5
Norm = 2 > 1 = modulus of largest eigenvalue).
0 1
If A = , then the eigenvalues are 0; but Ae1 = 0 and
0 0
Ae2 = e1 , so kAk = 1 > 0 .
kA xk2 = (A x) · (A x) = (A x)∗ (A x) = x∗ A∗ A x.
λ1 , λ2 , . . . , λn
x = α1 v1 + α2 v2 + · · · + αn vn
so kxk2 = hx, xi = · · · = |α1 |2 + |α2 |2 + · · · + |αn |2
2
P P
since ||x|| = hx, xi = i,j αi α¯j hvi , vj i = i αi ᾱi .
A∗ Ax = A∗ A(α1 v1 + α2 v2 + · · · + αn vn )
= λ1 α1 v1 + λ2 α2 v2 + · · · + λn αn vn
kA xk2
∴ 2
≤ λ1 with equality if x = α1 v1
kxk
kA xk √
∴ ≤ λ1 with equality if x
kxk
is a corresponding eigenvector of A∗ A.
So, for any matrix A , its norm equals its largest singular value!
Warning
A square matrix A has both eigenvalues and singular values.
Distinguish these carefully!
Denoting eigenvalues by λA i we see that
q
|λA
∗
max | ≤ ||A|| = λA A
max
MA212 – Lecture 34 – Friday 16 February 2018 page 10
EXAMPLE
kA xk
kxk
√
is 6 and this occurs when x = (−1, i, 1)⊤ (an eigenvector of
A∗ A corresponding to its eigenvalue 6 . We can, of course, take
any non-zero scalar multiple of this!)
−1
2 i 0 −3 √
Ax = ⇒ kA xk = 18
i =
0 −i 2 3
1
√
and kxk = 3
√
kA xk 18 √
so = √ = 6.
kxk 3
1 1 ∗
ui = √ A vi and vi = √ A ui
λi λi
MA212 – Lecture 34 – Friday 16 February 2018 page 13
We now derive a direct link of A to its singular values.
v1 , v2 , . . . , vk , vk+1 , vk+2 , . . . , vn
Now,
α1 β1
c1 , c2 = α1 c1 + α2 c2 , β1 c1 + β2 c2
α2 β2
= α1 c1 , β1 c1 + α2 c2 , β2 c2
= c1 (α1 , β1 ) + c2 (α2 , β2 )
r1
c1 , c2 = c1 r1 + c2 r2 ,
r2
for
r1 = (α1 , β1 ) r2 = (α2 , β2 ).
A∗ A vi = λi vi with λi 6= 0 for 1 ≤ i ≤ k
and λi = 0 for k + 1 ≤ i ≤ k
1
with ui = √ A vi for the corresponding eigenvector for A A∗ .
λi
This is only for 1 ≤ i ≤ k ;
Indeed, for k + 1 ≤ i ≤ n , we have λi = 0 so that
A∗ A vi = 0 ⇒ vi∗ A∗ A vi = 0
⇒ (A vi )∗ A vi = 0
⇒ (A vi ) · (A vi ) = 0
⇒ kA vi k = 0
MA212 – Lecture 34 – Friday 16 February 2018 page 18
⇒ Avi = 0
Conclusion: the SVD.
√ √ √
∴ A = λ1 u1 v1 + λ2 u2 v2 + · · · + λk uk vk∗
∗ ∗
√ √ √
∴ A = λ1 u1 v1 + λ2 u2 v2 + · · · + λk uk vk∗ =
∗ ∗
2 i 0
With A = from the last lecture (Lecture 33) we had
0 −i 2
∴
1
0 1 √ √1 √i − √16
A= 2 2 2
+ 6 6 6
1 1
2 0 2 − √16 − √i6 √1
6
A vi = λi vi ( and vi 6= 0).
So, as before
3 2
A=
2 0
= −1
Eigenvalues: λ1 = 4, λ2 .
1 2 1 −1
Eigenvectors: v1 = √ , v2 = √
5 1 5 2
√2 − √15
∴ A= 4 5 √2 √1 + (−1) −1
√ √2
√1 5 5 √2 5 5
5 5
4 2 1
− 25
∴ A= 4 5 5
+ (−1) 5
2 1
5 5 − 25 4
5
...to our next topic, we need to think a little more about vector
spaces..., specifically Real vector spaces cont’d.
Sums and intersections
Definition. Suppose that U and W are two subspaces of a real
vector space V . We define the sum of U and W to be
U + W = {u + w : u ∈ U, w ∈ W },
U ∩ W = {x : x ∈ U and x ∈ W }.
0 ∈ U + W as 0 = 0 + 0 and so U + W 6= ∅ .
If x1 , x2 ∈ U + W we have x1 = u1 + w1 and x2 = u2 + w2
with ui ∈ U, wi ∈ W , so that
⇒ x1 + x2 ∈ U + W.
αx = α(u + w) = αu + αw ⇒ αx ∈ U + W
For U ∩ W we have
U W (point)
U (line)
U+W (plane)
0
W (line)
U W (line)
U+W (plane)
0
U (plane)
W (line)
W (plane)
U W (line)
0
MA212 – Lecture 34 – Friday 16 February 2018 page 30
U (plane)
EXAMPLE
and so
1 0 0
0 1 0
x ∈ Lin , , .
0 0 1
0 0 1
1 0 0
0 1 0
U + W = Lin , ,
0 0 1
0 0 1
so dim(U + W ) = 3.
a
1
a
1
and so, x = ∈ Lin .
0
0
0 0
1
1
... any vector x ∈ Lin can be seen as
0
0
a a
a a
x= ∈U and x= ∈W
0 0
0 0
and so x ∈ U ∩ W .
1
1
U ∩ W = Lin . So dim(U ∩ W ) = 1 .
0
0
Indeed, we observe that
Direct sums
Orthogonal complements
Orthogonal projections
A duality result:
R(A⊤ ) = N (A)⊥
Background readings
Adam Ostaszewski
Chapter 4 (§4.1 & §4.6)
0 Px U
subspace
Concepts needed - I:
U1 + U2 = {u1 + u2 | u1 ∈ U1 , and u2 ∈ U2 }
U2
u2 u1+u2
0 u1 U1
MA212 – Lecture 35 – Tuesday 20 February 2018 page 4
Recap 2
{u1 . . . , ur } a basis of U
We write
V = U1 ⊕ U2
v = u1 + u2 with u1 ∈ U1 , u2 ∈ U2
V = R3
Suppose v1 , v2 , v3 are linearly independent
Take
U1 = Lin({v1 , v2 })
U1 = Lin({v3 })
Theorem.
V = U1 ⊕ U2 (Direct sum)
⇐⇒
(i) V = U1 + U2 (sum)
(ii) U1 ∩ U2 = {0}
( ⇒ ) Suppose V = U1 ⊕ U2 .
v = u1 + u2 ∈ U1 + U2 .
(ii) If u ∈ U1 ∩ U2 , then
u = | {z
u } + | {z
0 }
in U1 in U2
= | {z
0 } + | {z
u }
in U1 in U2
Corollary.
If V = U1 ⊕ U2 ,
then
dim(V ) = dim(U1 ) + dim(U2 ).
Proof: Here
dim(U1 ∩ U2 ) = dim({0}) = 0.
So
dim(V ) = dim(U1 + U2 )
= dim(U1 ) + dim(U2 ) − dim(U1 ∩ U2 ) = dim(U1 ) + dim(U2 ).
| {z }
=0
V =U ⊕W
Task.Finding a complement:
Solution. First we find a base for U , say it is
{u1 . . . , ur }.
dim(V ) = r + s, s > 0.
W
0
as U + W = R2
and U ∩ W = {0}
U ⊥ = {v ∈ V : v ⊥ u for all u ∈ U }
v ∈ U⊥
⇐⇒
hu, vi = 0 for all u ∈ U.
\
⊥
U = {v ∈ V : hu, vi = 0}
u∈U
Proof.
(i) Exercise for you!
(Intersection of hyperplanes through the origin?)
{u1 , . . . , ur }
for U .
Extend to an orthonormal basis of V
{u1 , . . . , ur , ur+1 , un }
For v ∈ V write
v = α1 u1 + . . . + αr ur + αr+1 ur+1 + . . . + αn un
| {z } | {z }
∈U ∈U ⊥
So V = U + U ⊥ .
Also
u ∈ U ∩ U⊥ ⇒
|{z} hu, ui = 0 ⇒ u = 0
(as u∈U ⊥ )
so U ∩ U ⊥ = {0}
and
dim(V ) = dim(U ⊥ ) + dim (U ⊥ )⊥ .
Subtracting equations
0 = dim(U ) − dim (U ⊥ )⊥
⊥
U ⊆ U⊥
hu, wi = hw, ui = 0
V=Թ3
u
W
0 w
U
W
w
0 u U
U
Application: a Duality result
a
⊤
Take A = [a, b]1×2 , so A = . Notice that, first,
b
2×1
x x x a
N (A) = { : ax + by = 0} = { :h , i = 0}
y y y b
a ⊥
= Lin{ }
b
But
* +
a x
0= ,
b y
says that
a x
⊥ .
b y
So
R(A⊤ ) = N (A)⊥
i.e.
T2 = T
When second power yields just the same as one, one says
“Idempotent”!
V = R(T ) ⊕ N (T )
!(T)
(I T)v v
(T)
Tv
v = T v + (v − T v)
T v + (I − T ) v
= |{z}
| {z }
∈R(T ) ∈N (T )
=0 as T = T 2 .
For v ∈ V
v=Iv (identity)
= T v + (I − T )v
T (I − T )v) = T v − T T (v)
= T v − T2 v = T v − T v = 0
T v = 0,
and
So v = 0 , as T v = 0 , proving (ii) .
So we conclude V = R(T ) ⊕ N (T ) .
If V =U ⊕W
v =u+w with u ∈ U, w ∈ W.
As u=u+0
Tu=u for u ∈ U
and so T2 = T
Suppose
v v2
z }| { z }| {
1
1 0
V = R3 U = Lin 0 , 1
1 −1
2
W = Lin 1
0
| {z }
v3
S = (v1 , v2 , v3 )
1 0 2
= 0 1 1
1 −1 0
S −1 P S = D, so that P = S D S −1 .
−1 2 2
S −1
= −1 2 1
1 −1 −1
Now
1
D=
1 = (e1 , e2 , 0) .
0
P = SDS −1
1 0 2
=0 1 1
D S −1
( recall SD = (Se1 , Se2 , S0) )
1 −1 0
1 0 0 −1 2 2
=0 1 0
−1 2 1
(use row operations!)
1 −1 0 1 −1 1
−1 2 2
=−1 2 1
0 0 1 MA212 – Lecture 35 – Tuesday 20 February 2018 page 36
Orthogonal Projection onto a subspace U
Theorem.
P is an orthogonal projection ⇐⇒ P 2 = P and P ⊤ = P .
As x=Ix
= P x + (I − P )x
0 = x⊤ P ⊤ (I − P )y
= x⊤ (P ⊤ − P ⊤ P )y
P ⊤ = P ⊤P
transposing
P = P ⊤ P ⊤⊤ = P ⊤ P = P ⊤
So P = P ⊤ symmetric
Suppose P2 = P
and P⊤ = P
⇒ P ⊤ = P = P P = P ⊤ P.
Then, as before,
hP x, (I − P ) yi = x⊤ (P ⊤ − P ⊤ P ) y = 0.
| {z }
0
So P is orthogonal!
1 1 2
A=
5 2 4
2 1 1 2 1 2
A =
25 2 4 2 4
1 5 10
= =A
25 10 20
and also A⊤ = A.
2
Let V = R , v1 =
3
1
and U = Lin {v1 }
1
1 0
Then U is spanned by: v2 = −2 , v3 =
⊥ 1
.
0 −1
Then
T v1 = v1 , T v2 = 0, T v3 = 0 so...
1
−1
P = R
0 R
0
2 0 0 2 1 1
1
=
1 0 0
6 2 −2 −2
1 0 0 2 −1 −5
4 2 2
1
= 2 1 1
6
2 1 1
PT = P here, of course.
MA212 – Lecture 35 – Tuesday 20 February 2018 page 45
To span U ⊥ ...(if not by inspection)
2 x1
x ∈ U⊥ ⇐⇒ 1 • x 2 = 0
1 x3
2x1 + x2 + x3 = 0
put x1 = t
x3 = u
so x2 = −u − 2 t
t 1 0
x = −u − 2 t = t −2 +u
−1
u+ 0 1
| {z } | {z }
v2 −v3
are orthogonal ( u1 ⊥ u2 ).
1
to u := −2
(got by taking u = 0 above):
0
Noting that u ⊥ u1 ,
1 * 1 0 + 0
1 1
w = −2 − √ √ −2, 1 1 − 0
2 2
0 0 −1 −1
1 0
= −2 + 1
0 −1
1
= −1
−1
−1
1
1
say u3 = √ 1 = √ (−w)
3 3
1
An×m
−1
⊤
A A A A⊤
| {z }
m×n×n×m
Why such a ‘magical’ (‘out of nowhere’) formula?
We’ll see in a later lecture!
V = Rn (Reals!)
U = Lin ({v1 . . . , vm })
| {z }
linearly independent
Put
A := [v1 | . . . | vm ]n×m
Consider
P = A(A⊤ A)−1 A⊤
A⊤ A invertible
P2 = P
P ⊤ = P (obvious)
R(P ) = U = R(A)
Find N (A⊤ A) :
A⊤ A x = 0 ⇒ x⊤ A⊤ A x = 0
⇒ hA x, A xi = 0
⇒ Ax = 0
P idempotent P 2 = P
P symmetric P ⊤ = P
⊤
A(A⊤ A)−1 A⊤
h i
P x = A (A⊤ A)−1 A⊤ x
∈ R(A).
we ask
?
A x = P applied to some vector ?
Trick of insertion!
−1
A x = A A⊤ A A⊤ A x
= A(A⊤ A)−1 A⊤ (A x)
= P (A x) ∈ R(P ).
So indeed
R(A) ⊆ R(P )
Put
1 0
A = −2 1
0 −1
3×2
1 0
1 −2 0
⊤
(A A)2×2 = −2 1
0 1 −1
0 −1
5 −2
= (now swop and sign-switch)
−2 2
1 2 2
(A⊤ A)−1 =
6 2 5 MA212 – Lecture 35 – Tuesday 20 February 2018 page 59
Using P = A(A⊤ A)A⊤
1 0
1 2 2 1 −2 0
P = −2 1
6 2 5 0 1 −1
0 −1
2 2
1 1 −2 0
= −2 1
6 0 1 −1
−2 −5
2 −2 −2
1
= −2 5 −1
6
−2 −1 5
Adam Ostaszewski
Chap. 4 (§4.7)
with v1 , ..., vm ∈ Rn
i.e. here A is built from columns in Rn
A(AT A)−1 AT
with AT A being m × m
LA = I as = (AT A)−1 AT .A = I
If
An×k = [v1 , ..., vk ]
then
−− v1T −−
A A=
T
... [v1 , ..., vk ]
−− vkT −−
The ij element is
hvi , vj i
LA = I
Ax = 0 =⇒ x =Ix = LAx = 0
P = AL and LA = I.
Of course
A |{z}
L.A L = AL.
AR = I
then
R |{z}
A.R A = RA.
So P = RA is also a projection.
Then
RT AT = I
In summary Al×r
has a left inverse if rank = r
has a right inverse if rank = l
To
PB := A(Bm×n An×m )−1 B
Suppose that
Rn = U ⊕ W
W
(n m) dim
m dim U
ni
0 x
W = N (Bm×n ) (Wanted)
W ⊥ = N (Bm×n )⊥ (Equivalently)
= R((Bm×n )T ) (Duality: R(B T ) = N (B)⊥ )
= column space of B T
Rn = U ⊕ W = Rn = R(A) ⊕ N (B)
So same dimension, so
W = N (B) = N (P ).
1
1
U = Lin 0 so A =
0
1 1
1 0
W = Lin 1 , 1
0 3
x ∈ W⊥ ⇔
1
1 · x = 0, i.e. x1 + x2 = 0, and
0
0
1 · x = 0, i.e. x2 + 3x3 = 0,
3
3x3 3
x = −3x3 = x3 −3
x3 1 MA212 – Lecture 36 – Friday 23 February 2018 page 20
Example cont’d
3
3
W ⊥
= Lin −3 = R
−3
1 1
⊥ T
3 3
∴W = R −3 = N
−3 = N ([3, −3, 1])
1 1
by duality.
1
BA = [3, −3, 1] 0
= (4)1×1
1
(BA)−1 = 1/4
P = A(BA)−1 B
1
= 0
(1/4)[3, −3, 1] = ...
1 3 −3 1
1 1
P = 0 [3, −3, 1] = 0 0 0
.
4 4
1 3 −3 1
(Ã)
= (A)
R(BA) ⊆ R(T ).
So
R(BA) = R(T )
rank(BA) + nullity(T ) = m,
rank(BA) + 0 = m.
v − P v ∈ U ⊥ & u − P v ∈ U =⇒ v − P v ⊥ P v − u
By Pythagoras’s Theorem
So
||v − u|| ≥ ||v − P v||
hypoteneuse
v
u Pv
0 Pv
U u
True value x 0 3 5 8 10
Observation y 2 5 6 9 11
y = ax + b
True value x 0 3 5 8 10
Observation y 2 5 6 9 11
0 1
3
1
y ∈ Lin
∗
5 ,
1
8
1
10 1
and
y∗ = y − e
0 1
3
1
U = Lin 5 ,
1 ⊆ R5
8
1
10 1
y∗ = P y.
0 1
3 1
∗
a
A=
5 1
, wanted: y ∗
= A
b∗
8 1
10 1
P = A(AT A)−1 AT = AL
0 1
3 1
0 3 5 8 10
AT A = 5 1
1 1 1 1 1
8 1
10 1
198 26
=
26 5
1 5 −26
T
(A A)−1
=
314 −26 189
α
A = y∗ = P y = ALy
β
α α
Pre-multiply by L: LA
|{z}
= = L A Ly = Ly
β β | {z }
=I =P y
α
= (AT A)−1 AT y
β
2
5
0 3 5 8 10 227
T
A y= 6 =
1 1 1 1 1 33
0
11
α 5 −26 227
= (AT A)−1 AT y = 1
β 314 −26 189 33
1 277 .88 1
= = not quite .
314 632 2.01 2
11
9
8
y*(x)
5 y(x)
3 5 8 10
Adam Ostaszewski:
Inverses Chap. 8 (§8.3,8.4)
L := (AT A)−1 AT
is a left inverse
LA = I.
Others?
If Bm×n An×m is of rank m for some B then
| {z }
m×m
L := (BA)−1 B gives LA = I.
Lm×n = L′m×n
AL = AL′ ⇒ L = IL = |{z}
LA L = L |{z} AL′ = |{z}
AL = L |{z} I L′ = L′ .
=I =P P =P ′ =LA
0 + x2 − x3 = 0
x1 x1 1 0
x2 = x2 = x1 0 + x2 1 .
x3 x2 0 1
1 0
W ⊥: T
= R 0 1 = R B , say.
0 1
1 0 0
W = N = N (B)
0 1 1
1 0
1 0 0 1 0
BA =
1 1 =
0 1 1 2 2
1 1 det=2
−1 1 2 0
(BA) = (after a ... swap/re-sign/divide)
2 −2 1
−1 1 2 0 1 0 0 1 2 0 0
L = (BA) B= = .
2 −2 1 0 1 1 2 −2 1 1
LA = (BA)−1 B .A = I.
| {z }
so RT is a left inverse.
Take AT to be n × m , then Am×n must have rank m.
So ...
′
Rn×m = Rn×m .
If Am×n has a right inverse: Am×n Rn×m = I, then one can solve
Ax = b ∈Rm
AR b = b.
x = Rb, because Ax = |{z}
=I
General solution of
Ax = b
because if Ax = b
Now by “duality”
and z ∈ N (A) .
MA212 – Lecture 37 – Tuesday 27 February 2018 page 15
So ...
So Rb ⊥ z.
So, by Pythagoras,
if
L = (AT A)−1 AT ,
Pv
U= (A)
0 v*
An×m = [v1 , ..., vm ] (i.e. with m columns) has a left inverse iff
rank (A) = m
If (BA)m×m has rank m then (BA)−1 B is a left inverse
Choose B with N (B) = a specified subspace = W .
How?
Method:
Motivation:
AR A = A,
|{z} R |{z}
AR = R; A |{z}
LA = A, LA L = L.
|{z}
LB B = I and CRC = I.
Take
G = RC LB
and
GAG
G
z }| { z }| {
= RC LB |{z}
BC RC LB = RC [LB B][CRC ]LB = RC LB
A
= G
and
=A
z}|{
AGA = BC RC LB BC = B[CRC ][LB B]C = BC
| {z }
= A
dim T [W ] ≤ dim W ;
A = BC.
We select a set of columns that are lin. indep. and span the
remaining columns: a1 , a2 does the trick. (These two columns
are obviously lin. indep.: look at their first component.)
So B = (a1 , a2 ) is a basis for R(A) .
4
4
a5 = 6 = 4a1 + 6a2 =
6
10 B
1 0
1 0 2 −1 4
A= 0 1
.
0 1 −1 0 6
1 1 | {z }
co-ordinate cols. rel. B
1 1 2 4
−1 2 −1 −1 R + R
2 1
A= →
3 0 5
9 R3 − 3R1
1 4 3 7 R4 − R1
1 1 2 4
0 3 1 3
→
0 −3 −1 −3 R3 + R2
0 3 1 3 R4 − R2
1 1 2 4
0 3 1 3 R1 − 2R2
→ (to put to work the 1 in R2 )
0 0 0 0
0 0 0 0
1 −5 0 −2
0 3 1 3
→ = A′ so (col 1, col 3) form a basis of A′
0 0 0 0
0 0 0 0
1 1 2 4 1 2
−1 2 −1 −1 −1 −1 1 −5 0 −2
= BC =
3 0 5 9 3 5 0 3 1 3
1 4 3 7 1 3
L = (P B)−1 P
LB = (P B)−1 P B = I
1 0 0 0 − − eT1 − −
P = = .
0 1 0 0 − − eT2 − −
−1 −2 1 0 0 0 −1 −2 0 0
−1
(P B) P = =
1 1 0 1 0 0 1 1 0 0
1 0
0 0
N.B. these ei are in R4 and so Q =
0 1
0 0
MA212 – Lecture 37 – Tuesday 27 February 2018 page 34
Then...
then obviously
1 0
CQ = =I
0 1
Q(CQ)−1 = Q = RC
From A = BC to G = RC LB :
1 0 −1 −2 0 0
0 0 −1 −2 0 0 0 0 0 0
G= =
0 1 1 1 0 0 1 1 0 0
| {z }
0 0 =LB 0 0 0 0
| {z }
=RC
DAn×m D = Dm×n
ADm×n A = An×m
We get projections
Indeed
and
| {z } z = 0 =⇒ Az = 0, so N (DA) ⊆ N (A).
DAz = 0 =⇒ ADA
=A
rk(D) + nullity(D) = n
n = rk(A) + nullity(D).
So
rk(D) + nullity(D) = n = rk(A) + nullity(D).
and
z = x − ADx : Dz = Dx − DADx = 0, so
x = (ADx) + (z)
| {z } |{z}
∈R(AD) ∈N (D) MA212 – Lecture 38 – Friday 2 March 2018 page 6
Also:
z = y − DAy : Az = Ay − ADAy = 0, so
y = (DAy) + |{z}
z
| {z }
∈R(DA) ∈N (A)
Theorem: If
Rn = R(An×m ) ⊕ W,
Rm = V
|{z} ⊕ N (An×m )
dim V =rk(A)
and G is unique.
rank(A) + nullity(A) = m,
dim V + nullityA = m
So
rank(A) = dim V
Rn = R(Ãn×k ) ⊕ W,
=V
z }| {
Rm = R( |{z}
H ) ⊕ N (C)
| {z }
for some H =N (A)
where Ãn×k =
submatrix with k indep. col. spanning col. space of An×m ,
so that then R(Ã) = R(A).
0 = Ax = |{z} =⇒ Cx = 0 =⇒ ÃCx = Ax = 0
ÃC x =⇒ LÃCx = 0 |{z}
=A LÃ=I
R = RC = H(CH)−1 gives CR = I
G = RL
T (Ãx) = B Ãx
k = rank(A) = rk(B)
An×m = Bn×k Ck×m G = RC LB
RC = C T (CC T )−1
LB = (B T B)−1 B T .
See Figure:
ࣨ(AT)=ࣨ(G) Թn
AG
࣬(A)=࣬(AG)
Anxm Gmxn
MA212 – Lecture 38 – Friday 2 March 2018 page 18
Conclusion 2: in Rm
࣬ ࣬
GA projects orthogonally parallel to N (A)
ࣨ(A) Թm
GA
࣬(GA)=࣬(AT)
Anxm Gmxn
MA212 – Lecture 38 – Friday 2 March 2018 page 19
Comment on N (G) = R(A)⊥
= R(B)⊥ = R(A)⊥
N (G) =↑see below N (LB ) =↑see below N (B T ) |{z}
duality
=G
z}|{
Gx = 0 =⇒ RL x = 0 =⇒ |{z}
CR Lx = 0 =⇒ Lx = 0
=I
⇔
|{z} (B T B)−1 B T x = 0 ⇔ B x = 0
T
sub for L
=⇒ Lx = 0 =⇒ RLx = 0 =⇒ Gx = 0
=I
z}|{
T −1
C y = C [(CC ) · LB ] · CC T ]y
T T
G (B · CC T y)
= |{z} so R(C T ) ⊆ R(G)
MA212 – Lecture 38 – Friday 2 March 2018 page 21
=RL
Cont’d
Ax = 0 =⇒ |{z}
BC x = 0 =⇒ |{z}
LB Cx = 0 =⇒ Cx = 0 =⇒ |{z}
BC x = 0,
=A =I =A
N (A) = N (C).
GAG = G N (G) = R(A)⊥
+ [i.e. ‘Orthogonality’]
AGA = A R(G) = N (A)
⊥
is called the
Strong Generalized Inverse
or the
Moore-Penrose pseudo-inverse
L = (AT A)−1 AT
and both of
AL = A(AT A)−1 AT ,
are symmetric,
LA = I
L = AT (AAT )−1
and both of
RA = AT (AAT )−1 A,
are symmetric.
AR = I
1 0 2
−1 1 0
R2 → R2 + R1
0 1 2
1 0 2
→
0 1 2
0 1 2 R3 → R3 − R2
1 0 2
→ 0 1 2
c3 = 2(c1 + c2 )(as c1 = e1 , c2 = e2 )
0 0 0
1 0
1 0 2
A = −1 1
= BC
0 1 2
0 1
1 0
1 −1 0 2 −1
B B =
T −1 1 =
0 1 1 −1 2
0 1
1 2 1
(B T B)−1 = ... swap & re-sign, and det=3
3 1 2
T −1 T 1 2 1 1 −1 0 1 2 −1 1
LB = (B B) B = =
3 1 2 0 1 1 3 1 1 2
1 0
1 0 2 5 4
CC T
= 0 1 =
0 1 2 4 5
2 2
1 5 −4
(CC T )−1 =
9 −4 5
1 0
1 5 −4
RC T
= C (CC ) T −1
= 0 1
9 −4 5
2 2
5 −4
1
−4
= 5
9
2 2
5 −4
1 1 2 −1 1
G = RL = · −4 5
9 3 1 1 2
2 2
6 −9 −3
1 −3
= 9 6
27
6 0 6
P[0, 1] contains the function fn (t) ≡ tn ; all these are lin. indep
and infinitely differentiable.
∞
X
ℓ2 := {a = (a1 , a2 , ...) : (∀n)(an ∈ R) and a2n < ∞}.
n=1
k ⊆ c0 , k ⊆ ℓ2 , ℓ2 ⊆ c0 .
In ℓ2 we define
X∞
ha, bi := a n bn
n=1
for a, b ∈ ℓ2 .
Consider
e1 = (1, 0, 0, ...) ∈ k
e2 = (0, 1, 0, ...) ∈ k
etc.
Then
0, if i 6= j,
hei , ej i =
1, if i = j.
Differences
Similarities
It ain’t necessarily so!
Now in ℓ2 ...
For a = (a1 , a2 , ...) define the shift-map T by putting:
T ∗ T e1 = 0 = T ∗ 0 = 0 6= e1
yet
T T ∗ a = T (0, a1 , a2 , ...) = a for a = (a1 , a2 , ..)
so
TT∗ = I
T ∗ T 6= I
implies...
T ∗a = λa
0 = λa1
a1 = λa2
a2 = λa3
T ∗a = λa
0 = a1
a1 = λa2
a2 = λa3
So no eigenvalues here.
U ⊆ (U ⊥ )⊥ 6= U
So (k ⊥ )⊥ = ℓ2 6= k.
k + k ⊥ = k + {0} = k 6= ℓ2
Why?
Indeed
1 1 X 1
2
a = (1, , ..., , ...) ∈ ℓ as 2
< ∞,
2 n n
1 1
but (1, , ..., , ...) ∈
/ k.
2 n
In an inner-product space V :
Suppose that U ⊆ V is a finite-dimensional subspace.
So U = Lin{u1 , u2 , ..., un }, say; suppose indeed these are
orthonormal, so a basis for U. (Can arrange, by applying the
Gram-Schmidt process)
Put
P (v) = hv, u1 iu1 + hv, u2 iu2 + ... + hv, un iun
P (v) ∈ U
hv−P v, ui i = hv, ui i − hP v, ui i
So v − P v ⊥ U .
So
v = |{z}
Pv +v| −{zP v}
∈U ∈U ⊥
V = U + U ⊥.
U ⊕ U ⊥,
V = |{z}
finite dim.
because
U ∩ U ⊥ = {0};
indeed, if u ∈U ∩ U ⊥ then
hu, ui = 0, so u = 0.
So by the above
is the orthogonal projection from V onto U. MA212 – Lecture 39 – Tuesday 6 March 2018 page 13
Recap
So for u ∈U and v ∈V
u − v = (u − P v) + (P v − v)
| {z } | {z }
∈U ∈U ⊥
So P v in U is the closest to v.
Pv U
hypoteneuse
v
Pv u
0 Pv
MA212 – Lecture 39 – Tuesday 6 March 2018 page 15
U u
Example 1
Take V = ℓ2
Here V = C[0, 1]
Z 1
hf, gi = f (t)g(t) dt
0
U = Lin{1, t, t2 }
g1 (t) ≡ 1,
√
g2 (t) ≡ 3(2t − 1)
√
g3 (t) ≡ 5(6t2 − 6t + 1)
Z 1
||1||2 = 12 dt = 1, so g1 ≡ 1
0
Z 1 2
t 1
ht, g1 i = t.1 dt = =
0 2 2
1
t − ht, g1 ig1 = t − 1 ⊥ g1
2
2
2 Z 1 Z 1/2 Z 1/2
1
1
t −
= t − dt = s2 ds = 2 s2 ds
2
0 | {z 2} −1/2 0
=s
s3 1
=2 =
3 s=1/2 12
−1
1 1 √ 1 √ MA212 – Lecture 39 – Tuesday 6 March 2018 page 19
g2 = √ t− =2 3 t− = 3 (2t − 1) .
We now find g3
Z 1
3 1
2 2 t 1
ht , g1 i = t .1 dt = =
0 3 0 3
Z 1 √ √ Z 1
ht2 , g2 i = t2 . 3 (2t − 1) dt = 3 (2t3 − t2 ) dt
0 0
3 1
√
√ t t 4 √ 1 1 √ 3−2 3
= 3 − = 3 − = 3 =
2 3 0 2 3 6 6
√ 2 1 √ 2
g3 = 6 5 t − t + = 5 6t − 6t + 1
6
Let’s look at the example of PU (f ) for
√
f (t) = t
Z
1 √ Z 1 √
P (f ) = t1/2 dt 1 + 3 1/2
t .(2t − 1) dt 3 (2t − 1)
0 0
√ Z 1
1/2 2
√ 2
+ 5 t . 6t − 6t + 1 dt 5 6t − 6t + 1
0
2 4 2 12 12 2 2
P (f ) = +3 − (2t − 1) + 5 − + 6t − 6t + 1
3 5 3 7 5 3
6 + 48t − 20t2
=
35
Z 2
√
||f − P (f )||2 = min t − at2 + bt + c dt
a,b,c
since the general element in U has the form at2 + bt + c .
Question:
? ||Pn (v) − v|| → 0 ?
i.e.
Pn (a) → a − a1 e1 ,
Lin {e1 , e2 , · · · , en , · · · }
Consider
d2
T (f ) := 2 f,
dt
then its eigenvalues equation is
T (f ) = λf,
i.e.
f ′′ − λf = 0 (auxiliary equation: m2 − λ = 0)
∂2 ∂
2
f (x, t) = f (x, t).
∂x ∂t
How to solve?
Investigate the problem using eigenvectors in the space C[0, 1],
or better, as we see later, on C[−π, +π]
These are the functions f (t) into C (the complex numbers) that
are of the form
n
X
ikt
√
ak e with i = −1 (finite sum!)
k=−n
For example
1 ikt −ikt
−i ikt −ikt
cos(kt) = 2 e +e sin(kt) = 2 e −e
f : [−π, +π] → C
This context exploits that cos and sin are periodic (period 2π );
so are also trigonometric polynomials:
both satisfy
f (t + 2π) = f (t)
so e.g.
f (−π) = f (π),
Z
eikt dt
1 1
= sin(kt)−i cos kt + C
k k
1
= [cos kt + i sin(kt)] + C
ik
eikt
= + C, (as 1/i = −i)
ik
as expected.
MA212 – Lecture 40 – Friday 9 March 2018 page 8
Orthogonality
Z +π Z +π
heint , eimt i = eint eimt dt = eint e−imt dt
−π −π
Z +π
= ei(n−m)t dt
−π
Consider k = n − m 6= 0, then
ikt +π
int imt e
he ,e i = = 0, as cos(kt)|t=π = cos(kt)|t=−π
ik −π
(even function),
whereas sin(kt) = 0 for t = ±π.
Z +π Z +π
||eint ||2 = heint , eint i = eint e−int dt dt = 1 dt = 2π.
−π −π
So
1 int
√ e :n∈Z is an orthonormal set in C[−π, +π].
2π
n
X n
X
eikt eikt 1
Pn (f ) = hf, √ i √ = hf, eikt ieikt .
2π 2π 2π
k=−n | {z } k=−n
Here Z +π
1 1
ak (f ) = ikt
hf, e i = f (t)e−ikt dt
2π 2π −π
Define
f (t) := |t| for t ∈ [−π, +π],
>2 > 0 2 3
Z +π Z +π Z 0
|t|e−ikt dt = te−ikt dt + (−t) e−ikt dt
−π 0 −π |{z}
=|t|
Z +π Z 0
= te−ikt dt − te−ikt dt
0 −π
−ikt π
−ikt 0
t −ikt e t −ikt e
= i e + 2 − i e + 2
k k 0 k k −π
1 2 1 n −ikπ o
= {0} − 2 + 2 e + e+ikπ
k
k k
0 k even (& k 6= 0)
=
− 42 k odd
k
In the above remember that e+ikπ = e−ikπ for integer k . MA212 – Lecture 40 – Friday 9 March 2018 page 15
So
π/2
k=0
ak (f ) = 0 k even & k 6= 0
− πk2 2 k odd
and so:
n
π 2 X 1 ikt
Pn (f ) = − 2
e ( but (−k)2 = k 2 so...)
2 π k
k odd
−n
X n n
X
π 2 1 ikt −ikt π 4 cos(kt)
= − (e + e )= − ,
2 π k2 2 π k2
k odd k=1 & odd
1
n
X
π 4 cos(kt)
Pn (f ) = − 2
→ |t| as n → ∞ for t ∈ [−π, +π]?
2 π k
k=1 & odd
−
Puzzle!
(ii)
∞
X
kf k2 = 2π |ak (f )|2
−∞
(iii)
∞
X
f= ak (f )eikt for − π < t < +π
k=−∞
Danger at t = ±π.
MA212 – Lecture 40 – Friday 9 March 2018 page 18
To get (ii) from (i)
take t = 0, then
∞
X
π 4 1
0= − 2
.
2 π k
k>0 odd
So
∞
X
π2 1
= 2
,
8 k
k>0 odd
i.e.
1 1 1 π2
1 + 2 + 2 + 2 + ... = .
3 5 7 8
How about all k ? What is this sum:
1 1 1
1 + 2 + 2 + 2 + ...
2 3 4
If odd=1
∞
X 1 π2
2
=
k 6
k=1
i.e.
1 1 1 π2
1 + 2 + 2 + 2 + ... =
2 3 4 6