Professional Documents
Culture Documents
Linear Transformations and Operators
Linear Transformations and Operators
is also a linear transformation from V into W . The set of all linear transformation
from V into W , together with the addition and scalar multiplication defined above,
is a vector space over the field F .
Proof. Suppose that T and U are linear transformation from V into W . For (T +U )
defined above, we have
123
124 CHAPTER 6. LINEAR TRANSFORMATIONS AND OPERATORS
Fact 6.1.2. Let V be an n-dimensional vector space over the field F , and let W
be an m-dimensional vector space over F . Then the space L(V, W ) is finite-
dimensional and has dimension mn.
Theorem 6.1.3. Let V , W , and Z be vector spaces over a field F . Let T ∈ L(V, W )
and U ∈ L(W, Z). Then the composed function U T defined by (U T ) (v) =
U (T (v)) is a linear transformation from V into Z.
(U T ) (sv 1 + v 2 ) = U (T (sv 1 + v 2 ))
= U (sT v 1 + T v 2 )
= sU (T v 1 ) + U (T v 2 )
= s(U T ) (v 1 ) + (U T ) (v 2 ) ,
as desired.
Definition 6.1.5. An algebra over a field F is a vector space V over F that has a
bilinear vector product “ · ” : V × V → V satisfying (su) · (tv) = (st)(u · v) and
Example 6.1.6. The set L(V, V ) of linear operators on V forms an algebra when
the vector product is defined by functional composition U T (v) = U (T (v)). If V is
a Banach space and L(V, V ) is equipped with the induced operator norm, then it
forms a Banach algebra.
1. T is one-to-one: T v 1 = T v 2 =⇒ v 1 = v 2
Lv = (v2 , v3 , v4 , . . .)
Rv = (0, v1 , v2 , . . .).
Notice that L is onto but not one-to-one and R is one-to-one but not onto. Therefore,
neither transformation is invertible.
Example 6.1.9. Consider the normed vector space V of semi-infinite real sequences
Rω with the standard Schauder basis {e1 , e2 , . . .}. Let T : V → V be the linear
transformation that satisfies T ei = i−1 ei for i = 1, 2, . . .. Let the linear trans-
formation U : V → V satisfy U ei = iei for i = 1, 2, . . .. It is easy to verify that
U = T −1 and U T = T U = I.
126 CHAPTER 6. LINEAR TRANSFORMATIONS AND OPERATORS
This example should actually bother you somewhat. Since T reduces vector
components arbitrarily, its inverse must enlarge them arbitrarily. Clearly, this is not
a desirable property. Later, we will introduce a norm for linear transforms which
quantifies this problem.
Theorem 6.1.10. Let V and W be vector spaces over the field F and let T be a
linear transformation from V into W . If T is invertible, then the inverse function
T −1 is a linear transformation from W onto V .
T (sv 1 + v 2 ) = sT (v 1 ) + T (v 2 ) = sw1 + w2 .
That is, sv 1 + v 2 is the unique vector in V that maps to sw1 + w2 under T . It follows
that
T −1 (sw1 + w2 ) = sv 1 + v 2 = s T −1 w1 + T −1 w2
and T −1 is a linear transformation.
dim V ∗ = dim V.
In this case, one actually finds that V is isomorphic to V ∗ . Therefore, the two spaces
can be identified with each other so that V = V ∗ for finite dimensional V .
6.2. THE DUAL SPACE 127
Theorem 6.2.2. Let V be a finite-dimensional vector space over the field F , and
let B = v 1 , . . . , v n be a basis for V . There is a unique dual basis B ∗ = f1 , . . . , fn
for V ∗ such that fj (v i ) = δij . For each linear functional on V , we have
n
X
f= f (v i ) fi
i=1
then n n
X X
f vj = si f i v j = si δij = sj .
i=1 i=1
In particular, if f is the zero functional, f v j = 0 for j = 1, . . . , n and hence the
scalars {sj } must all equal 0. It follows that the functionals f1 , . . . , fn are linearly
independent. Since dim V ∗ = n, we conclude that B ∗ = f1 , . . . , fn forms a basis
for V ∗ , the dual basis of B.
Next, we want to show that there is a unique basis which is dual to B. If f is a
linear functional on V , then f is some linear combination of f1 , . . . , fn with
n
X
f= si f i .
i=1
Furthermore, by construction, we must have sj = f v j for j = 1, . . . , n. Simi-
larly, if
n
X
v= ti v i .
i=1
128 CHAPTER 6. LINEAR TRANSFORMATIONS AND OPERATORS
is a vector in V , then
n
X n
X
fj (v) = ti fj (v i ) = ti δij = tj .
i=1 i=1
One important use of the dual space is to define the transpose of a linear trans-
form in a way that generalizes to infinite dimensional vector spaces. Let V, W
be vector spaces over F and T : V → W be a linear transform. If g ∈ W ∗
is a linear functional on W (i.e., g : W → F ), then g(T v) ∈ V ∗ is a linear
functional on V . The transpose of T is the mapping U : W ∗ → V ∗ defined by
f (v) = g(T v) ∈ V ∗ for all g ∈ W ∗ . If V, W are finite-dimensional, then one can
identify V = V ∗ and W = W ∗ via isomorphism and recover the standard transpose
mapping U : W → V implied by the matrix transpose.
The details of this definition are not used in the remainder of these notes, but can
be useful in understanding the subtleties of infinite dimensional spaces. For infinite
dimensional Hilbert spaces, we will see later that the definition again simplifies
because one identify V = V ∗ via isomorphism. The interesting case that does not
simplify is that of linear transforms between infinite dimensional Banach spaces.
Definition 6.3.1. Let V and W be two normed vector spaces and let T : V → W
be a linear transformation. The induced operator norm of T is defined to
kT vk
kT k = sup = sup kT vk .
v∈V −{0} kvk v∈V,kvk=1
A common question about the operator norm is, “How do I know the two ex-
pressions give the same result?”. To see this, we can write
kT vk
v
sup = sup
T kvk
= sup kT uk .
v∈V −{0} kvk v∈V −{0} u∈V,kuk=1
Previously, we have seen that the set L(V, W ) of linear transformations from V
into W , with the standard addition and scalar multiplication, satisfies the conditions
required to be a vector space. Now, we have a norm for that vector space. Inter-
ested readers should verify that the above definition satisfies the first two standard
conditions required by a norm. To verify the triangle inequality, we can write
≤ sup kT vk + sup kU vk
v∈V,kvk=1 v∈V,kvk=1
= kT k + kU k.
The induced operator norm also has another property that follows naturally from
its definition. Notice that
kT vk kT uk
kT k = sup ≥
v∈V −{0} kvk kuk
for all non-zero u ∈ V . Checking the special case of u = 0 separately, one can
show the induced operator-norm inequality kT uk ≤ kT kkuk for all u ∈ V .
For the space L(V, V ) of linear operators on V , a norm is called submulti-
plicative if kT U k ≤ kT kkU k for all T, U ∈ L(V, V ). The induced operator-norm
inequality shows that all induced operator norms are submultiplicative because
kU T vk ≤ kU k kT vk ≤ kU k kT k kvk .
This also defines a submultiplicative norm for the algebra of linear operators on V .
130 CHAPTER 6. LINEAR TRANSFORMATIONS AND OPERATORS
Proof. Suppose that T is bounded; that is, there exists M such that kT vk ≤ M kvk
for all v ∈ V . Let v 1 , v 2 , . . . be a convergent sequence in V , then
T v i − T v j
=
T v i − v j
≤ M
v i − v j
.
The value M = 2
δ
serves as an upper bound on kT k.
B = v1, . . . , vn
v = s1 v 1 + · · · + sn v n
since (6.1) implies that |si | ≤ m−1 kvk. We first show that this holds for coefficients
{s1 , . . . , sn } satisfying the condition |s1 | + · · · + |sn | = 1. Let
( )
Xn
S= (s1 , . . . , sn ) |si | = 1 .
i=1
This set is closed and bounded; it is therefore compact. Define the function f : S →
R by
f (s1 , . . . , sn ) = ks1 v 1 + · · · + sn v n k .
It can be shown that f is continuous, and it is clear that f > 0 over S. Let
m= min f (s1 , . . . , sn ).
(s1 ,...,sn )∈S
kT vk = kT (s1 v 1 + · · · + sn v n )k
≤ |s1 | kT v 1 k + · · · + |sn | kT v n k
≤ C (|s1 | + · · · + |sn |) .
By the previous lemma, this implies that there exists an M such that |s1 | + · · · +
|sn | ≤ M kvk, so that
kT vk ≤ CM kvk .
is Cauchy. This follows from the fact that, for m < n, we have
Xn−1
X n−1
kT km − kT kn kT km
i
kAn − Am k =
T
≤ kT ki = ≤ .
1 − kT k 1 − kT k
i=m i=m
Since this goes to zero as m → ∞, we see that the limit limn→∞ An exists.
Next, we observe that
(I − T ) I + T + T 2 + · · · + T n−1 = I − T n .
Since kT k < 1, we have limk→∞ T k = 0 because
T k
≤ kT kk → 0. Taking the
limit n → ∞ of both sides gives
∞
X
(I − T ) T i = lim (I − T n ) = I.
n→∞
i=0
Likewise, reversing the order multiplication results in the same result. This shows
P
that ∞i=0 T must be the inverse of I − T .
i
6.3. OPERATOR NORMS 133
X
kAk1 = max kAvk1 = max |aij |
kvk1 =1 j
i
J = v H AH Av − λv H v.
Taking the gradient with respect to v and equating the result to zero, we get
AH Av = λv.
v H AH Av = λv H v = λ.
Since we are maximizing the left hand side of this equation, λ must be the largest
eigenvalue of AH A. For an n×n matrix B with eigenvalues λ1 , . . . , λn , the spectral
radius ρ(B) is defined by
ρ(B) = max |λi |.
i
The spectral radius of B is the smallest radius of a circle centered at the origin that
contains all the eigenvalues of B. It follows that
p
kAk2 = ρ(AH A).
When A is Hermitian, kAk2 = ρ(A). The 2-norm is also called the spectral norm.
134 CHAPTER 6. LINEAR TRANSFORMATIONS AND OPERATORS
This norm is also called the Euclidean norm. Note that kAk2F = tr(AH A).
fv (w) = hw|vi .
Proof. While the result holds in any Hilbert space, this proof assumes V is separa-
ble for simplicity. Therefore, we let v 1 , v 2 , . . . be a countable orthonormal basis for
V . We wish to find a candidate vector v for the inner product.
First, we note that f is bounded and, as such, there exists M such that |f (x) | ≤
P
M kxk for all x ∈ V . Let xn = ni=1 f (v i )v i . For any n, we have
X n X n Xn
2
M kxn k ≥ |f (xn )| = f (v i )f (v i ) = |f (v i )| = f (v i ) f (v i )
i=1 i=1 i=1
Xn D E X n Xn D
E
= f (v i )v i f (v i )v i = f v j v j f (v i )v i
i=1 i=1 j=1
* n n +
X X
= f v j v j f (v i )v i = hxn |xn i = kxn k2 .
j=1 i=1
6.4. LINEAR FUNCTIONALS ON HILBERT SPACES 135
Pn
This implies that kxn k ≤ M for all n. Hence, limn→∞ i=1 |f (v i )|2 is bounded
and the vector
∞
X
v= f (v i )v i ,
i=1
fv (u)
kfv k , sup = kvk.
u∈V −{0} kuk
Since this is true for each v j , it follows that f = fv . Now, consider any v 0 ∈ V such
that hw|vi = hw|v 0 i for all w ∈ W . Applying Lemma 6.4.1 shows that v = v 0 and
we conclude that v is unique.
v 0 = T ∗ w.
B = v1, . . . , vn
RT ⊆ W
NT ⊆ V
RT ∗ ⊆ V
NT ∗ ⊆ W.
138 CHAPTER 6. LINEAR TRANSFORMATIONS AND OPERATORS
RT = RT = [NT ∗ ]⊥
RT ∗ = RT ∗ = [NT ]⊥ .
That is, w and n are orthogonal vectors. It follows that NT ∗ ⊆ [RT ]⊥ . Now, let
w ∈ [RT ]⊥ . Then, for every v ∈ V , we have
hT v|wi = 0.
This implies that hv|T ∗ wi = 0, by the definition of the adjoint. Since this is true for
every v ∈ V , we get T ∗ w = 0, so w ∈ NT ∗ . Then [RT ]⊥ ⊆ NT ∗ , which combined
with our previous result yields [RT ]⊥ = NT ∗ . Using a similar argument, one can
show that [RT ∗ ]⊥ = NT .
6.6 Pseudoinverses
Theorem 6.6.1. Let V and W be Hilbert spaces and T be a bounded linear trans-
formation from V to W where RT is closed. The equation T v = w has a solution if
and only if hw|ui = 0 for every vector u ∈ NT ∗ , i.e.,
w ∈ RT ⇔ w ⊥ NT ∗ .
wo = w − PRT w = w − wr 6= 0.
But
hw|wo i = hwr + wo |wo i = hwo |wo i > 0,
which contradicts the assumption that hw|ui = 0 when u ∈ NT ∗ . We must conclude
that T v = w has a solution.
Fact 6.6.2. The solution to T v = w (if it exists) is unique if and only if the only
solution to T v = 0 is v = 0. That is, if NT = {0}.
kT v − wk2 .
T ∗ T v = T ∗ w.
w − ŵ ∈ [RT ]⊥ .
Theorem 6.6.5. Suppose the vector v̂ ∈ V minimizes kvk over all v ∈ V satisfying
T v = w. Then, v̂ ∈ [NT ]⊥ and, if RT ∗ is closed, v̂ = T ∗ u for some u ∈ W .
Proof. Suppose v̂ ∈
/ [NT ]⊥ , then the orthogonal decomposition V = [NT ]⊥ + NT
shows that the projection of v̂ onto [NT ]⊥ has smaller norm but still satisfies T v̂ =
w. This gives a contradiction and shows that v̂ ∈ [NT ]⊥ . If RT ∗ is closed, then
RT ∗ = [NT ]⊥ and v̂ = T ∗ u for some u ∈ W .
Corollary 6.6.6. If A is a matrix such that AAH is invertible, then the minimum-
norm solution to Av = w is
−1
v = AH AAH w.