Professional Documents
Culture Documents
Change - Bases Uw Answer Scheme
Change - Bases Uw Answer Scheme
Change of Basis
In many applications, we may need to switch between two or more different bases for a vector
space. So it would be helpful to have formulas for converting the components of a vector with
respect to one basis into the corresponding components of the vector (or matrix of the operator)
with respect to the other basis. The theory and tools for quickly determining these “change of
basis formulas” will be developed in these notes.
9/22/2013
Chapter & Page: 5–2 Change of Basis
and 3
4 3 4
i − i
5 5 5 5
AA† =
= ··· = I .
4 3 4 3
i − i
5 5 5 5
So A is unitary.
Obviously, for a matrix to be unitary, it must be square. It should also be fairly clear that, if
A is a unitary (or orthogonal) matrix, then so are A∗ , AT , A† and A−1 .
?◮Exercise 5.1: Prove that, if A is a unitary (or orthogonal) matrix, then so are A∗ , AT , A†
and A−1 .
The term “unitary” comes from the value of the determinant. To see this, first observe that,
if A is unitary then
I = AA−1 = AA† .
Using this and already discussed properties of determinants, we have
Thus,
A is unitary H⇒ |det A| = 1 .
And since there are only two real numbers which have magnitude 1 , it immediately follows that
A is orthogonal H⇒ det A = ±1 .
An immediate consequence of this is that if the absolute value of the determinant of a matrix
is not 1 , then that matrix cannot be unitary.
Why the term “orthogonal” is appropriate will become obvious later.
More concisely,
∗
[U] jk = u jk and U† jk
= [U]k j = uk j ∗ .
Now observe:
U is unitary
⇐⇒ U† = U−1
⇐⇒ U† U = I
†
⇐⇒ U U jk = [I] jk
X
⇐⇒ U† jm
[U]mk = δ jk “for all ( j, k)”
m
X
⇐⇒ u m j ∗ u mk = δ jk “for all ( j, k)”
m
The righthand side of the last equation is simply the formula for computing the standard matrix
inner product of the column matrices
u1 j u 1k
u u
2j 2k
.
. and . ,
.
. .
uNj uNk
and the last line tells us that
(
1 if j =k
this inner product = .
0 if j 6= k
In other words, that line states that
u 11 u 12 u 13 u 1N
u
u 21
u
22
u
23 2N
. , ,
.. , . . . , ..
.
..
.
. . .
uN1 uN2 uN3 uNN
is an orthonormal set of column matrices. But these column matrices are simply the columns of
our original matrix U . Consequently, the above set of observations (starting with “ U is unitary”)
reduces to
A square matrix U is unitary if and only if its columns form an orthonormal set of
column matrices (using the standard column matrix inner product).
You can verify that a similar statement holds using the rows of U instead of its columns.
?◮Exercise 5.2: Show that a square matrix U is unitary if and only if its rows form an
orthonormal set of row matrices (using the row matrix inner product). (Hints: Either consider
how the above derivation would have changed if we had used UU† instead of U† U , or use
the fact just derived along with the fact that U is unitary if and only if UT is unitary.)
version: 9/22/2013
Chapter & Page: 5–4 Change of Basis
1. U is unitary.
2. The columns of U form an orthonormal set of column matrices (with respect to the usual
matrix inner product). That is,
(
X 1 if j = k
u m j ∗ u mk =
m 0 if j 6= k
.
3. The rows of U form an orthonormal set of row matrices (with respect to the usual matrix
inner product). That is,
(
X 1 if j = k
u jm ∗ u km =
m 0 if j 6= k
.
(I.e., X
bj = ek u k j for j = 1, 2, . . . , N .)
k
a: Show that
S is orthonormal and U is a unitary matrix H⇒ B is also orthonormal .
b: Show that
S and B are both orthonormal sets H⇒ U is a unitary matrix .
version: 9/22/2013
Chapter & Page: 5–6 Change of Basis
2. The second observation is that MAA and MBB greatly simplify if the bases are orthonor-
mal. If A is orthonormal, then
[MAA ] jk = a j ak = δ jk = [I] jk .
So
MAA = I if A is orthonormal . (5.2a)
Now let v be any vector in V and, for convenience, let us denote the components of v with
respect to A and B using α j ’s and β j ’s respectively,
α1 β1
α β
2 2
|viA =
.. and |viB =
.. .
. .
αN βN
Our goal is to find the relations between the α j ’s and the β j ’s so that we can find one set
given the other. One set of relations can be found by taking the inner product of v with each a j
and using equations (5.3). Doing so:
* + * +
X X
aj v = aj α k ak = a j βk bk
k k
X
X
H⇒ aj v = α k a j ak = βk a j bk
k k
X
X
H⇒ aj v = a j ak α k = a j bk βk
k k
X X
H⇒ aj v = [MAA ] jk αk = [MAB ] jk βk
k k
The formulas in the second equation in the last line are simply formulas for the j th entry in the
products of MAA and MAB with the column matrices of αk ’s and β’s . So that equation tells
us that
h a1 | v i α1 β1
h a | v i α β
2 2 2
.. = MAA . = MAB .
. .
. . .
h aN | v i αN βN
Change of Basis for Vector Components:
The General Case Chapter & Page: 5–7
Recalling what the column matrices of αk ’s and β’s are, we see that this reduces to
h a1 | v i
h a | v i
2
.. = MAA |viA = MAB |viB . (5.4)
.
h aN | v i
?◮Exercise 5.5 (semi-optional): Using the same assumptions as were used to derive (5.4),
derive that
h b1 | v i
h b | v i
2
. = MBA |viA = MBB |viB . (5.5)
..
h bN | v i
Equations (5.4) and (5.5) give the relations between the components of a vector with respect
to the two different bases, as well as the relations between these components and the inner products
of the vector with each basis vector. Our current main interest is in the relations between the
different components (i.e., the rightmost two-thirds of (5.4) and (5.5)). For future reference, let
us summarize what the above tells us in that regard.
where MAA , MAB , MBB and MBA are the four N × N matrices given by
[MAA ] jk = a j ak , [MAB ] jk = a j bk ,
[MBB ] jk = b j bk and [MBA ] jk = b j ak .
Naturally, the above formulas simplify considerably when the two bases are orthonormal.
That will be of particular interest to us. Before going there, however, let us observe an number
of “little results” that immediately follow from the above lemma and its derivation, and which
apply whether or not either basis is orthonormal:
1. Solving equations (5.6a) and (5.6b) for |viB yields the change of basis formulas
version: 9/22/2013
Chapter & Page: 5–8 Change of Basis
2. Solving equations (5.6a) and (5.6b) for |viA yields the change of basis formulas
Don’t memorize these formulas. They are nice formulas and can save a little work when
dealing with arbitrary matrices. However, since we will attempt to restrict ourselves to orthonor-
mal (or at least orthogonal) bases, we won’t have a great need for these formulas as written. If
you must memorize formulas, memorize those in the Big Lemma — they are much more easily
memorized. And besides the above change of basis formulas are easily derived from the formulas
in that lemma.
Equations (5.7a) and (5.7b) are certainly nicer than the original equations in the Big Lemma,
but look at what we get when they are combined:
That is,
|viB = [MBA MAB ] |viB for each v ∈ V ,
Change of Basis for Vector Components:
When the Bases Are Orthonormal Chapter & Page: 5–9
A = { a1 , a2 , . . . , a N } and B = { b1 , b2 , . . . , b N } .
Let MAB and MBA be a pair of N × N matrices which are adjoints of each other and which
satisfy any one of the following sets of conditions:
1. [MAB ] jk = a j bk .
2. [MBA ] jk = b j ak .
3. b1 b2 · · · b N = a1 a2 · · · a N MAB .
4. a1 a2 · · · a N = b1 b2 · · · b N MBA .
5. |viA = MAB |viB for each v ∈ V .
6. |viB = MBA |viA for each v ∈ V .
Then MAB and MBA are unitary matrices satisfying all of the above conditions, as well as
version: 9/22/2013
Chapter & Page: 5–10 Change of Basis
from the Big Theorem (theorem 5.3, above), and have the formulas for the bk ’s in terms of the
a j ’s ,
bk = βk1 a1 + βk2 a2 + · · · + βk N a N for j = 1, 2, . . . , N ,
then you can obtain the matrix MAB easily by simply noting that each of these equations can be
written as
βk1
βk2
bk = a1 a2 · · · a N .. .
.
βk N
Comparing the last equation with the one a few lines earlier for b1 b2 · · · b N , it should
be clear that the above column matrix on the right must be the k th column in MAB .
See, no computations are needed (provided the bases are orthonormal).
A = { a1 , a2 , . . . , a N } , B = { b1 , b2 , . . . , b N }
and
C = { c1 , c2 , . . . , c N }
Then, in addition to MBA and MAB we also have MAC , MBA , MCA and MCB , all defined
analogously to the way we defined MBA and MAB . Applying the above theorem we see that,
for every v in V ,
MCA |viA = |viC = MCB |viB = MCB MBA |viA .
Thus,
MCA |viA = MCB MBA |viA for every v ∈ V .
A = { a1 , a2 } and B = { b1 , b2 }
or
A = { a1 , a2 , a3 } and B = { b1 , b2 , b3 } .
version: 9/22/2013
Chapter & Page: 5–12 Change of Basis
Direction Cosines
Since we are assuming a space of traditional vectors, the scalars are real numbers, the inner
product is the traditional dot product, and the change of bases matrices MAB and MBA are
orthogonal and are transposes of each other. In this case,
[MAB ] jk = a j bk = a j · bk =
a j
kbk k cos θ(a j , bk ) .
These are called the direction cosines relating the two bases.
Two-Dimensional Case
If V is two dimensional, then the possible geometric relationships between
A = { a1 , a2 } and B = { b1 , b2 }
a2 a2
b2
b1 b1
ψ ψ
or
a1 a1
b2
With a little thought, it should be clear that the matrices for the “flips” in the second case
are simply " # " #
−1 0 1 0
F1 = and F2 = .
0 1 0 −1
Traditional Rotated and Flipped Bases Chapter & Page: 5–13
Clearly, then the matrix MBA can be written as either a single rotation matrix, or as the
product of a rotation matrix with one of these “flip” matrices. It’s not at all hard to show that
det(any rotation matrix) = 1 and det(either flip matrix) = −1 .
Consequently, at least if we are considering orthonormal bases A and B for a two-dimensional,
traditional vector space,
A and B are rotated images of each other ⇐⇒ det (MBA ) = +1 .
Three-Dimensional Case
Likewise, if
A = { a1 , a2 , a3 } and B = { b1 , b2 , b3 }
are any two orthonormal bases for a traditional three-dimensional space, then MBA will either
be a matrix for a “rotation by some angle about some vector”, or the product of such a rotation
matrix with one of the flip matrices
−1 0 0 1 0 0 1 0 0
F1 = 0 1 0 , F2 = 0 −1 0 and F3 = 0 1 0 .
0 0 1 0 0 1 0 0 −1
And, again, it can be shown that (still assuming A and B are orthonormal bases)
A and B are rotated images of each other ⇐⇒ det (MBA ) = +1 .
Consequently,
A and B are both right-handed, or are both left-handed ⇐⇒ det(MBA ) = +1 .
It can be shown that the rotation matrix, itself, can be written as a product of three “simple
rotation matrices”
R3,γ R2,β R3,α
where R j,φ corresponds to a rotation of angle φ about the j th basis vector. The angles α , β ,
and γ are the infamous “Euler angles”. Now turn to pages 139, 140 and 141 in Arfken, Weber
and Harris and skim their discussion of rotations and Euler angles.
Warnings:
1. There are subtle differences between a matrix for a rotation operator (to be discussed in
the next chapter) and the change of basis matrix for “rotated bases”. Both may be called
rotation matrices, but may differ in the signs of some of the corresponding entries.
2. I have not carefully checked Arken and Weber’s discussion so I will not guarantee whether
their product
R3,γ R2,β R3,α
is actually what we are calling MBA or if the signs of some of the entries in some of the
matrices need to be switched. In other words, I have not verified whether each of Arfken,
Weber and Harris’s “rotation matrices” are describing the operation of rotating a vector
in space, or describing a change of basis when one basis is a rotation of the other.
3. Different authors use different conventions for the Euler angles — different axes of
rotation and different order of operations. Be careful, the computations based on two
different conventions may well lead to two different and incompatible results.
version: 9/22/2013
Chapter & Page: 5–14 Change of Basis
for our N -dimensional vector space V (for example, this might be the collection of all “rotations”
of some favorite orthonormal basis). Then each v in B has components with respect to each of
these Bm ’s , m
v1
v m N
2 X
|viBm = .
.
where v = v mj bmj .
. j=1
v mN
Also, for each pair of bases Bm and Bn , there will be a corresponding “change of basis”
matrix MBm Bn such that
This last expression is sometimes called a “transformation law”, even though we are not really
dealing with a true transform of vectors here.
Now suppose that, in the course of tedious calculations and/or drinking, we obtain, corre-
sponding to each different basis Bm , a corresponding N -tuple of scalars w1m , w2m , . . . , wmN .
The question naturally arrises as to whether
w11 , w21 , . . . , w1N , w12 , w22 , . . . , w2N , w13 , w23 , . . . , w3N , ...
all describe the same vector w but in the various corresponding bases. Obviously (to us at least)
the answer is “yes” if and only if
m n
w1 w1
w m wn
2 2
. = MBm Bn . for every pair of bases Bm and Bn . (5.10)
. .
. .
wmN wnN
and use the fact that (5.10) holds to verify that the change of basis formulas (5.9) will give us,
for each Bm , the originally given N -tuple of scalars w1m , w2m , . . . , wmN as the components of
w with respect to that basis.
This w is called the vector defined by the transformation law (5.9).
“Volumes” of N -Dimensional Hyper-Parallelpipeds (Part II) Chapter & Page: 5–15
{ v1 , v2 , . . . , v N } .
Our goal is to find a way to compute the N -dimensional volume of this object — denoted by
VN (P N ) — using the components of the vk ’s with respect to any basis
B = { b1 , b2 , . . . , b N } .
For our computations, we are letting VB be the “matrix of components of the v j ’s with respect
to basis B ” given by
v1,1 v2,1 v3,1 ··· v N ,1
v
1,2 v2,2 v3,2 ··· v N ,2
VB = “ |v1 iB |v2 iB |v3 iB · · · |v N iB ” = v v2,3 v3,3 ··· v N ,3
1,3
.. .. .. .. ..
. . . . .
v1,N v2,N v3,N · · · v N ,N
where, for j = 1, 2, . . . , N ,
N
X
vj = v j,k bk .
k=1
Recall, also, that we derived one formula for the volume; namely,
VN (P N ) = det(VN ) (5.11)
where N is the orthonormal set generated from {v1 , v2 , . . . , v N } via the Gram-Schmidt proce-
dure (see theorem 3.4 on page 3–27).
Since N is orthonormal, however, MN N is just the identity matrix, and the above reduces to
v j = M N B
v j .
N B
version: 9/22/2013
Chapter & Page: 5–16 Change of Basis
Thus,
VN = |v1 iN |v2 iN |v3 iN · · · |v N iN
= MN B |v1 iB |v2 iB |v3 iB · · · |v N iB = MN B VB .
Combining this with formula (5.11) and a property of determinants (equation (4.3) on page 4–11)
then yields
Let us simplify matters a little more. Let U be any orthonormal basis. As noted on page
5–11 (equation (5.8)),
MN B = MN U MUB .
But, since U and N are orthonormal bases of “traditional” vectors, MN U must be an orthog-
onal matrix (and, thus, have det(MN U ) = ±1 ). Consequently,
Combining this with our last formula for VN (P N ) , gives us our most general component formula
for VN (P N ) (and the next theorem).
Special Cases
B is orthonormal
In particular, if B is an orthonormal basis, then we can let U = B in formula (5.12). Because
of the orthonormality of B we have
v jk = 0 if j 6= k and vjj ≥ 0 .
version: 9/22/2013
Chapter & Page: 5–18 Change of Basis
If you compare formula (5.14) with formulas (3.11) through (3.11), starting on page 3–28,
it should be clear that we’ve already derived geometric formulas for the above |det MUB | when
N = 1 , N = 2 and N = 3 . Since “cut and paste” is so easy, we’ll rewrite those geometric
formulas:
where
A = kb1 k2 kb2 k2 kb3 k2 ,
As we noted in on page 3–29, these formulas reduce even further if the bk ’s are unit vectors,
and even further if the set of bk ’s is orthonormal. If you’ve not already done so, then do the next
exercise (which is the same as exercise 3.17):
?◮Exercise 5.7: To what do the above formulas for V1 (P1 ) , V2 (P2 ) and V3 (P3 ) reduce
a: when the bk ’s are unit vectors.
b: when the set of bk is orthonormal.