Professional Documents
Culture Documents
Section 6.1 - Inner Products and Norms: M×N I, J I, J
Section 6.1 - Inner Products and Norms: M×N I, J I, J
Definition. Let V be an inner product space. For x V , we define the norm or length of x by ||x|| =
hx, xi.
Theorem 6.2. Let V be an inner product space over F . Then for all x, y V and c F , the following are
true.
1. ||cx|| = |c| ||x||.
2. ||x|| = 0 if and only if x = 0. In any case ||x|| 0.
3. (Cauchy-Schwarz Inequality) | < x, y > | ||x|| ||y||.
4. (Triangle Inequality) ||x + y|| ||x|| + ||y||.
1
Section 6.2
Definition. Let V be an inner product space. Then S V is an orthonormal basis of V if it is an ordered
basis and orthonormal.
Theorem 6.3. Let V be an inner product space and S = {v1 , v2 , . . . , vk } be an orthogonal subset of V such
that i, vi 6= 0. If y Span(S), then
k
X
hy, vi i
y=
vi
||vi ||2
i=1
Proof. Let
y=
k
X
ai vi
i=1
Then
k
X
h
ai vi , vj i
hy, vj i =
i=1
k
X
ai hvi , vj i
i=1
=
=
So aj =
aj hvj , vj i
aj ||vj ||2
hy,vi i
||vi || .
k
X
hy, vi ivi
i=1
Corollary 2. If S also is orthogonal and all vectors in S are non-zero then S is linearly independent.
Pk
ii
Proof. Suppose i=1 ai vi = 0. Then for all j, aj = h0,v
||vi || = 0. So S is linearly independent.
Theorem 6.4. (Gram-Schmidt) Let V be an inner product space and S = {w1 , w2 , . . . , wn } V be a
linearly independent set. Define S 0 = {v1 , v2 , . . . , vn } where v1 = w1 and for 2 k n,
vk = wk
k1
X
i=1
hwk , vi i
vi .
||vi ||2
vn
= wn
n1
X
k=1
< vn , vj >
< wk , vk >
vk
||vk ||2
n1
X
k=1
= < wn , vj >
n1
X
k=1
< wk , vk >
vk , vj >
||vk ||2
< wk , vk >
< vk , vj >
||vk ||2
< wn , vj >
= < wn , vj >
< vj , vj >
||vj ||2
= =0
We now show span(Sn ) = span(Sn0 ). We know dim(span(Sn )) = n, since Sn is linearly independent.
We know dim(span(Sn0 )) = n, by Corollary 2 to Theorem 6.3.
0
). So, we just need to show vn span(()Sn ).
We know span(Sn1 ) = span(Sn1
vn = wn
n1
X
aj vj
j=1
0
) and
for some constants a1 , a2 , . . . , an1 . For all j [n 1], vj span(Sn1 ) since span(Sn1 ) = span(Sn1
0
vj Sn1 . Therefore, vn span(Sn ).
Theorem 6.5. Let V be a non-zero finite dimensional inner product space. Then V has an orthonormal
basis . Furthermore, if = {v1 , v2 , . . . , vn } and x V , then
x=
n
X
hx, vi ivi .
i=1
Proof. Start with a basis of V Apply Gram Schmidt, Theorem 6.4 to get an orthogonal set 0 . Produce
from 0 by normalizing 0 . That is, multiply each x 0 by 1/||x||. By Corollary 2 to Theorem 6.3, is
linearly independent and since it has n vectors, it must be a basis of V . By Corollary 1 to Theorem 6.3, if
x V , then
n
X
x=
hx, vi ivi .
i=1
Corollary 1. Let V be a finite dimensional inner product space with an orthonormal basis = {v1 , v2 , . . . , vn }.
Let T be a linear operator on V , and A = [T ] . Then for any i, j, (A)i,j = hT (vj ), vi i.
Proof. By Theorem 6.5, x V , then
x=
n
X
hx, vi ivi .
i=1
So
T (vj ) =
n
X
hT (vj ), vi ivi .
i=1
Hence
Definition. Let S be a non-empty subset of an inner product space V . We define S (read S perp) to be
the set of all vectors in V that are orthogonal to every vector in S; that is, S = {x : hx, yi = 0, y S}.
S is called the orthogonal complement of S.
Theorem 6.6. Let W be a finite dimensional subspace of an inner product space V , and let y V . Then
there exist unique vectors u W , z W such that y = u + z. Furthermore, if {v1 , v2 . . . , vk } is an
orthonormal basis for W , then
k
X
hy, vi ivi .
u=
i=1
Proof. Let y V . Let {v1 , v2 , . . . , vk } be an orthonormal basis for W . This exists by Theorem 6.5. Let
u=
k
X
hy, vi ivi .
i=1
We will show y u W .
It suffices to show that for all i [k], hy u, vi i = 0.
hy u, vi i
= hy
k
X
hy, vj ivj , vi i
j=1
= hy, vi i
k
X
hy, vj i hvj , vi i
j=1
= hy, vi i hy, vi i
= 0
Thus y u W . Suppose x W W . Then x W and x W . So, hx, xi = 0 and we have that
x = 0.
Suppose r W and s W such that y = r + s. Then
r+s
ru
= u+z
= zs
This shows that both r u and z s are in both W and W . But W W = so it must be that
r u = 0 and z s = 0 which implies that r = u and z = s and we see that the representation of y is
unique.
Corollary 1. In the notation of Theorem 6.6, the vector u is the unique vector in W that is closest to
y. That is, for any x W , ||y x|| ||y u|| and we get equality in the previous inequality if and only if
x = u.
Pk
Proof. Let y V , u = i=1 < y, vi > vi . z = y u W . Let x W . Then < y u, x >= 0, since
z W .
By Exercise 6.1, number 10, if a is orthogonal to b then ||a + b||2 = ||a||2 + ||b||2 . So we have
||y x||2
= ||(u + z) x||2
= ||(u x) + z||2
= ||u x||2 + ||z||2
||z||2
= ||y u||2
Now suppose ||yx|| = ||yu||. By the squeeze theorem, ||ux||2 +||z||2 = ||z||2 and thus ||ux||2 = 0,
||u x|| = 0, u x = 0, and so u = x.
4
Theorem 6.7. Suppose that S = {v1 , v2 , . . . , vk } is an orthonormal set in an n-dimensional inner product
space V . Then
(a) S can be extended to an orthonormal basis {v1 , . . . , vk , vk+1 , . . . , vn }
(b) If W =Span(S) then S1 = {vk+1 , vk+2 , . . . , vb } is an orthonormal basis for W .
(c) If W is any subspace of V , then dim(V ) = dim(W )+ dim(W ).
Proof. (a) Extend S to an ordered basis. S 0 = {v1 , v2 , . . . , vk , wk+1 , wk+2 , . . . , wn }. Apply Gram-Schmidt
to S 0 . First k do not change. S 0 spans V . Then normalize S 0 to obtain = {v1 , . . . , vk , vk+1 , . . . , vn }. (b)
Section 6.3:
Theorem 6.8. Let V be a finite dimensional inner product space over F , and let g : V F be a linear
functional. Then, there exists a unique vector y V such that g(x) = hx, yi, x V .
Proof. Let = {v1 , v2 , . . . , vn } be an orthonormal basis for V and let
y=
n
X
g(vi )vi .
i=1
Then for 1 j n,
hvj , yi =
hvj ,
n
X
g(vi )vi i
i=1
n
X
g(vi )hvj , vi i
i=1
=
=
g(vj )hvj , vj i
g(vj )
Check (1): LHS= z1 2z2 and for y1 = 1 and y 2 = 2, we have RHS= z1 2z2 .
5
Pn
i=1
(1)
g(vi )vi .
Theorem 6.9. Let V be a finite dimensional inner product space and let T be a linear operator on V . There
exists a unique operator T : V V such that hT (x), yi = hx, T (y)i, x, y V . Furthermore, T is linear.
Proof. Let y V
Define g : V F as g(x) = hT (x), yi, x V .
Claim: g is linear.
g(ax + z) =
=
=
=
hT (ax + z), yi
haT (x) + T (z), yi
haT (x), yi + hT (z), yi
ag(x) + g(z).
=
=
Since y0 is unique, T (cy + z) = cT (y) + T (z).
Claim: T is unique.
Let U : V V be linear such that
hT (x), yi =
hx, T (y)i =
=
hx, U (y)i, x, y V
hx, U (y)i, x, y V
T (y) = U (y), y V
So, T = U .
Definition. T is called the adjoint of the linear operator T and is defined to be the unique operator on V
satisfying hT (x), yi = hx, T (y)i, x, y V . For A Mn (F ), we have the definition of A , the adjoint of A
from before, the conjugate transpose.
Fact. hx, T (y)i = hT (x), yi, x, y V .
Proof.
hx, T (y)i =
=
=
hT (y), xi
hy, T (x)i
hT (x), yi
Theorem 6.10. Let V be a finite dimensional inner product space and let be an orthonormal basis for V .
If T is a linear operator on V , then
[T ] = [T ]
Proof. Let A = [T ] and B = [T ] with = {v1 , v2 , . . . , vn }.
By the corollary to Theorem 6.5,
(B)i,j
hT (vj ), vi i
hvi , T (vj )i
hT (vi ), vj i
(A)j,i
= (A )i,j
6
= A
= A
By Theorem 6.10, [LA ] = [LA ] , which equals A by (2) and [LA ] by (3).
Therefore, [LA ] = [LA ] LA = LA .
Theorem 6.11. Let V be an inner product space, T, U linear operators on V . Then
1. (T + U ) = T + U .
2. (cT ) = cT , c F .
3. (T U ) = U T (composition)
4. (T ) = T
5. I = I.
Proof.
1.
h(T + U )(x), yi =
hx, (T + U ) (y)i
and
h(T + U )(x), yi =
=
=
hT (x), yi + hU (x), yi
hx, T (y)i + hx, U (y)i
hx, (T + U )(y)i
and
hcT (x), yi =
=
=
chT (x), yi
chx, T (y)i
hx, cT (y)i
3.
hT U (x), yi =
hx, (T U ) (y)i
hT U (x), yi =
hT (U (x)), yi
and
=
=
hU (x), T (y)i
hx, U (T (y))i
hx, (U T )(y)i
(2)
(3)
4.
hT (x), yi =
hx, (T ) (y)i
by definition and
hT (x), yi =
hx, T (y)i
by Fact .
5.
hI (x), yi = hx, I(y)i = hx, yi, x, y V
Therefore, I (x) = x, x V and we have I = I.
1.
(A + B)i,j = (A + B)j,i = (A)j,i + (B)j,i = (A)j,i + (B)j,i
and
(A + B )i,j = (A )i,j + (B )i,j = (A)j,i + (B)j,i
2. Let c F .
(cA)i,j = (cA)j,i = c(A)j,i = c(A)j,i
and
(cA )i,j = c(A )i,j = c(A)j,i
8
3.
((AC) )i,j
(AC)j,i
n
X
k=1
n
X
k=1
n
X
k=1
n
X
(A)j,k (C)k,i
(A)j,k (C)k,i
(A )k,j (C )i,k
(C )i,k (A )k,j
k=1
= (C A )i,j
A Ax
hA Ax, xin
hAx, A xim
hAx, Axim
Ax
By Lemma 1,
hx, A (Ax0 y)in = 0, x F n .
So, A (Ax0 y) = 0. We see that x0 is the solution for x in
A Ax = A y.
If, in addition, we know that rank(A) = n, then by Lemma 2, we have that rank(A A) = n and is therefore
invertible. So,
x0 = (A A)1 A y.
So T (y) W .
By Theorem 5.21, the characteristic polynomial of TW divides the characteristic polynomial of T and
hence splits. By Theorem 6.7(c), dim(W ) = n 1, so we may apply the induction hypothesis to TW
and obtain an orthonormal basis of W such that [TW ] is upper triangular. Clearly, = {z} is an
orthonormal basis for V such that [T ] is upper triangular.
Definition. Let V be an inner product space, and let T be a linear operator on V . We say that T is normal
if T T = T T . An n n real or complex matrix A is normal if AA = A A.
Theorem 6.15. Let V be an inner product space, and let T be a normal operator on V . Then the following
statements are true.
(a) ||T (x)|| = ||T (x)|| for all x V .
(b) T cI is normal for every c F
(c) If x is an eigenvector of T , then x is also an eigenvector of T . In fact, if T (x) = x, then T (x) = x.
10
(d) If 1 and 2 are distinct eigenvalues of T with corresponding eigenvectors x1 and x2 , then x1 and x2
are orthogonal.
Proof. (a) For any x V , we have
2
kT (x)k
= (T cI)(T cI)
= T (T cI) cI(T cI)
= T T cT cT c
cI
cI
= T T cT cT c
= T T cT cT c
cI
(c) Suppose that T (x) = x for some x V . Let U = T I. Then U (x) = 0, and U is normal by (b).
Thus (a) implies that
x1 , 2 x2 = 2 hx1 , x2 i .
0
j 6= k
=
Ak,k j = k
11
=0
( )x
= 0, and we have that =
so is real.
But x 6= 0 so
(b) Let dim(V ) = n and be an orthonormal basis for V . Set A = [T ] . Then
A = [T ] = [T ] = [T ] = A
So A is self-adjoint.
Define TA : Cn Cn by TA (x) = Ax. Notice that [TA ] = A where is the standard ordered basis,
which is orthonormal.
So TA is self-adjoint and by (a) its eigenvalues are real.
WE know that over C the characteristic polynomial of TA factors into linear factors, t , and since
each is R, we know it also factors over R. But TA , A, and T all have the same characteristic
polynomial.
Theorem 6.17. Let T be a linear operator on a finite-dimensional real inner product space V . Then T is
self-adjoint if and only if there exists an orthonormal basis for V consisting of eigenvectors of T .
Proof. ( ) Assume T is self-adjoint. By the lemma, the characteristic polynomial of T splits. Now by
Shurs Theorem, there exists an orthonormal basis for V such that A = [T ] is upper triangular. But
A = [T ] = [T ] = [T ] = A
So A and A are both upper triangular, thus diagonal and we see that is a set of eigenvectors of T .
( ) Assume there is an orthonormal basis of V of eigenvectors of T . We know D = [T ] is diagonal with
eigenvalues on the diagonal. D is diagonal and is equal to D since it is real. But then [T ] = D = D =
[T ] = [T ] . So T = T .
Fall 2007 - The following was covered but will not be included on the final.
Section 6.5:
Definition. Let T be a linear operator on a finite-dimensional inner product space V (over F ). If ||T (x)|| =
||x|| for all x V , we call T a unitary operator if F = C and an orthogonal operator if F = R.
Theorem 6.18. Let T be a linear operator on a finite-dimensional inner product space V . Then the following
statements are equivalent.
1. T T = T T = I.
12
= hx, T T (y)i
= hx, yi
Proof. () Suppose V has an orthonormal basis {v1 , v2 , . . . , vn } such that i, T (vi ) = i vi and |i | = 1.
Then by Theorem 6.17, T is self-adjoint. Well show T T = T T = I. Then by Theorem 6.18, ||T (x)|| =
||x||, x V T is orthogonal.
T T (vi ) = T (T (vi )) = T (i vi ) = i T (vi ) = i i vi = i2 vi = vi
So T T = I. Similarly, T T = I.
() Assume T is self-adjoint. Then by Theorem 6.17, V has an orthonormal basis {v1 , v2 , . . . , vn } such
that T (vi ) = i vi , i.
If T is also orthogonal, we have
|i | ||vi || = ||i vi || = ||T (vi )|| = ||vi || |i | = 1
Corollary 2. Let T be a linear operator on a finite-dimensional complex inner product space V . Then V
has an orthonormal basis of eigenvectors of T with corresponding eigenvalues of absolute value 1 if and only
if T is unitary.
Definition. A square matrix A is called an orthogonal matrix if At A = AAt = I and unitary if A A =
AA = I. We say B is unitarily equivalent to D if there exists a unitary matrix Q such that D = Q BQ.
Theorem 6.19. Let A be a complex n n matrix. Then A is normal if and only if A is unitarily equivalent
to a diagonal matrix.
Proof. () Assume A is normal. There is an orthonormal basis for F n consisting of eigenvectors of A, by
Theorem 6.16. Let = {v1 , v2 , . . . , vn }. So A is similar to a diagonal matrix D by Theorem 5.1, where the
matrix S with column i equal to vi is the invertible matrix of similarity. S 1 AS = D. S is unitary.
(Lef tarrow) Suppose A = P DP where P is unitary and D is diagonal.
AA = (P DP )(P DP ) = P DP P D P = P DD P
Also A A = P D DP , but D D = DD .
Theorem 6.20. Let A be a real nn matrix. Then A is symmetric if and only if A is orthogonally equivalent
to a real diagonal matrix.
Theorem 6.21. Let A Mn (F ) be a matrix whose characteristic polynomial splits over F .
1. If F = C, then A is unitarily equivalent to a complex upper triangular matrix.
2. If F = R, then A is orthogonally equivalent to a real upper triangular matrix.
Proof. (1) By Shurs Theorem there is an orthonormal basis = {v1 , v2 , . . . , vn } such that [LA ] = N
where N is a complex upper triangular matrix.
0
Let 0 be the standard ordered basis. Then [LA ] 0 = A. Let Q = [I] . Then N = Q1 AQ. We know
that Q is unitary since its columns are an orthonormal set of vectors and so Q Q = I.
Section 6.6
Definition. If V = W1 W2 , then a linear operator T on V is the projection on W1 along W2 if whenever
x = x1 + x2 , with x1 W1 and x2 W2 , we have T (x) = x1 . In this case,
R(T ) = W1 = {x V : T (x) = x} and N (T ) = W2 .
We refer to T as the projection. Let V be an inner product space, and let T : V V be a projection. We
say that T is an orthogonal projection if R(T ) = N (T ) and N (T ) = R(T ).
14
Theorem 6.24. Let V be an inner product space, and let T be a linear operator on V . Then T is an
orthogonal projection if and only if T has an adjoint T and T 2 = T = T .
Compare Theorem 6.24 to Theorem 6.9 where V is finite-dimensional. This is the non-finite dimensional
version.
Theorem 6.25. (The Spectral Theorem). Suppose that T is a linear operator on a finite-dimensional inner
product space V over F with the distinct eigenvalues 1 , 2 , . . . , k . Assume that T is normal if F = C and
that T is self-adjoint if F = R. For each i(1 i k), let Wi be the eigenspace of T corresponding to the
eigenvalue i , and let Ti be the orthogonal projection of V on Wi . Then the following statements are true.
(a) V = W1 W2 Wk .
(b) If Wi0 denotes the direct sum of the subspaces Wj for j 6= i, then Wi = Wi0 .
(c) Ti Tj = i,j Ti for 1 i, j k.
(d) I = T1 + T2 + + Tk .
(e) T = 1 T1 + 2 T2 + + k Tk .
Proof. Assume F = C.
(a) T is normal. By Theorem 6.16 there exists an orthonormal basis of eigenvectors of T . By Theorem 5.10,
V = W1 W2 Wk .
(b) Let x Wi and y Wj , i 6= j.
Then hx, yi = 0 and so Wi0 = Wi .
But from (1),
X
dim(Wi0 ) =
dim(Wj ) = dim(V ) dim(Wi ).
j6=i
wi
Tj (wi ) = 0
(d)
(T1 + T2 + + Tk )(x)
i Ti (x)
1 T1 (x) + 2 T2 (x) + + k Tk (x)
(1 T1 + 2 T2 + + k Tk )(x).
15
16