Professional Documents
Culture Documents
PHY202 2024 Note No 3
PHY202 2024 Note No 3
PHY202 2024 Note No 3
Kenji Nishiwaki *
Shiv Nadar Institution of Eminence, Department of Physics
Ver. 17 January 2024 (17h 33min in Indian ST)
We will provide a concise review on Linear Algebra.
1 On finite-dimensional vectors
1.1 Finite-dimensional Hilbert Space
In quantum mechanics, physical states (of a quantum system) are represented as vectors. The abstract
form |ψi, called ket vector, is widely used to represent a vector describing a physical state ψ.
A Hilbert space H is the complex vector space in which an inner product is well-defined and which
has completeness:
• For any two of the elements |ψ1 i , |ψ2 i ∈ V , if the summation |ψ1 i + |ψ2 i and the scalar multipli-
cation by an arbitrary complex number c (∈ C),
c |ψi = |c ψi = |ψi c
are well-defined, and both of their results are also elements of V , V is a complex vector space.
• For the existence of a well-defined inner product in V , a dual vector (living in a dual vector space)
hψ|, called bra vector, should exist for any element of V , |ψi. The inner product (in a complex
vector space) is defined as a mapping:
{hχ| , |ψi} → hχ|ψi or hχ| |ψi ∈ C, (1.1)
where |ψi , |χi ∈ V . The form of hχ|ψi is called braket.
• Completeness will be discussed later.
• In a vector space V , there is a vector (called zero vector or null vector) with special properties 0 (or
0 , ~0, |nulli),
|ψi + |nulli = |nulli + |ψi = |ψi , (1.2)
where |ψi is an arbitrary vector of V . An inner product should possess the following properties,
hχ| |ψ1 i + |ψ2 i = hχ|ψ1 i + hχ|ψ2 i , (1.3)
hχ|cψi = c hχ|ψi , (1.4)
hχ1 | + hχ2 | |ψi = hχ1 |ψi + hχ2 |ψi , (1.5)
∗
hcχ|ψi = c hχ|ψi , (1.6)
∗
hψ|χi = hχ|ψi , ⇒ if χ = ψ, hψ|ψi is real, (1.7)
hψ|ψi ≥ 0, (1.8)
hψ|ψi = 0 ⇔ |ψi = |nulli , (1.9)
*
e-mail: kenji.nishiwaki@snu.edu.in.
1
where you should be careful about the following point about bra vector,
• A norm of a ket vector |ψi (or the ket vector hψ|) is defined by use of the corresponding inner
product as
p
k|ψik := hψ|ψi (≥ 0) . (1.11)
describes the distance between |ψ1 i and |ψ2 i. Due to the non-negativity of the inner product, we
can conclude
• The vector c |ψi (c 6= 0) is represended as ‘parallel to |ψi’, and if hχ|ψi = 0, |χi and |ψi are
represented as ‘orthogoanl each other’ (|ψi ⊥ |χi).
• A unit vector is defined as a vector with a unit form. If |ψi is not a unit vector, as long as |ψi =
6 |nulli,
the renormalised vector
1 1
|ψi
c := |ψi = p |ψi (1.14)
k|ψik hψ|ψi
q
obeys the unit form hψ|
c |ψi
c = 1.
2
The bra vector and the standard inner product in Cn are represented as follows:
Xn
χ∗i hb
•
ei | = χ∗1 , χ∗2 , · · · , χ∗n ,
hχ| = (1.17)
i=1
n
! n
!
X X
χ∗j hb
•
hχ|ψi = ej | ψi |b
ei i
j=1 i=1
n
X n
X
= χ∗j ψi hb
e |b
ei= χ∗i ψi . (1.18)
| j{z i}
i,j=1 i=1
=δij
Mathematically, any components cannot be specified before identifying basis vectors. On the other
hand, however, in physics, we simply introduce components under the implicit adoption of the Cartesian
basis vectors as follows:
ψ1
ψ2
|ψi = .. , (1.19)
.
ψn
where the equal symbol is simply used. Hereafter, we follow this naive custom in physics, where we
•
do not discriminate the introduction of concrete representations (=) and the equality (=). Here, the
summation of ket vectors and the scalar multiplication of a ket vector are represented as
ψ1 φ1 ψ1 + φ1
ψ2 φ2 ψ2 + φ2
|ψi + |φi = .. + .. = , (1.20)
..
. . .
ψn φn ψn + φn
ψ1 cψ1
ψ2 cψ2
c |ψi = c .. = .. , (1.21)
. .
ψn cψn
where c ∈ C, and the inner product is represented as
ψ1
n
ψ2 X ∗
hχ|ψi = χ∗1 , χ∗2 , · · · , χ∗n .. = χi ψi . (1.22)
. i=1
ψn
Here, hχ| is reprented as the Hermitian conjugation of |χi as
T
χ∗1
χ∗
† 2
hχ| = (|χi) := .. = χ∗1 , χ∗2 , · · · , χ∗n .
(1.23)
.
χ∗n
The squared norm of |ψi is calculated as
n
X n
X
2
k|ψik = hψ|ψi = ψi∗ ψi = |ψi |2 . (1.24)
i=1 i=1
3
1.2 Cauchy-Schwarz inequality
The following inequality, called Cauchy-Schwarz inequality, is realised for a pair of any vectors |ψi , |χi
in a Hilbert space H ,
hψ|ψi hχ|χi ≥ |hχ|ψi|2 = |hψ|χi|2 . (1.25)
• If hχ|χi = 0, due to the non-negativity, |χi should be the null vector |nulli. From |nulli + |nulli =
|nulli, we can conclude
Thus, both the left-hand and right-hand sides of Eq. (1.25) are zero, and the Cauchy-Schwarz
inequality holds.
• If hχ|χi =
6 0, for the following state with t ∈ C,
hχ|ψi∗
2
hχ|ψi hχ|ψi
hθ|θi = hψ|ψi − hψ|χi − hχ|ψi + hχ|χi ≥ 0
hχ|χi hχ|χi hχ|χi
hχ|ψi hψ|χi hχ|ψi hψ|χi
⇔ hψ|ψi − hψ|χi − hχ|ψi + hχ|χi ≥ 0
hχ|χi hχ|χi hχ|χi2
|hχ|ψi|2
⇔ hψ|ψi ≥
hχ|χi
⇔ hψ|ψi hχ|χi ≥ |hχ|ψi|2 , (1.30)
where thus the Cauchy-Schwarz inequality holds. The equality of the Cauchy-Schwarz inequality
is realised when and only when
4
Also, we note that the norm of the state |θi defined in Eq. (1.27) is minimised when we take t as e
t in
Eq. (1.29) as follows:
hχ|ψi
hχ|θi = hχ| |ψi − e
t |χi = hχ|ψi − hχ|χi = 0, (1.32)
hχ|χi
hχ|ψi |χihχ|ψi
|χi = (1.33)
hχ|χi hχ|χi
1.3 Basis
For several vectors in a Hilbert space H , |χ1 i , |χ2 i , · · · , |χk i (k ∈ N), when we focus on the equation,
if only c1 = c2 = · · · = ck = 0 is the solution of it, a set of the vectors {|χ1 i , |χ2 i , · · · , |χk i} is called
linearly independent. Being not linear independent is called linearly dependent. If a set of vectors is
linearly dependent, at least one of the coefficients, c1 , c2 , · · · , ck , should be nonzero. For example, if
cj 6= 0, after defining bi := −ci /cj for i = 1, 2, · · · , j − 1, j + 1, · · · , k, we get
which means that |χj i is written down as a linear combination of the other vectors, that is equal to
|χ1 i , |χ2 i , · · · , |χk i is linearly dependent.
A set of vectors such that the number of linearly independent vectors cannot be increased any further
is called, in particular, a basis of H . The number of vectors of a basis in H is called the dimension of
H . When the dimension of H is n (dimH = n) if we appropriately select n number of vectors, they
can be linearly independent. On the other hand, however, for any selections of n + 1 number of vectors,
they are linearly dependent.
Based on a basis {|χ1 i , |χ2 i , · · · }, an arbitrary vector |ψi of a Hilbert space H can be unfolded as a
linear combination,
where the coefficients c1 , c2 , · · · are uniquely determined for a certain basis. If we choose a different basis
{|χ01 i , |χ02 i , · · · }, another unfolding expression is given as
5
where δrs is the Kronecker’s delta symbol, the basis {|χ1 i , |χ2 i , · · · } is called an orthonormal basis of
H or a complete orthogonal system (CONS). If {|χ1 i , |χ2 i , · · · } is a CONS, the inner product between
hχr | and the unfolding expression of |ψi based on the CONS,
X
|ψi = cs |χs i , (1.39)
s
leads to
X X X
hχr |ψi = hχr | cs |χs i = cs hχr |χs i = cs δrs = cr , (1.40)
s s s
where the corresponding expansion coefficient cr was derived only through the inner product. This
property is extremely useful. Furthermore, from the expression,
X X X
|ψi = cr |χr i = |χr i cr = |χr ihχr |ψi , (1.41)
r r r
|ψ 0 i = O
b |ψi , (2.1)
where the symbolical element O b is called operator, which acts on |ψi and symbolically represent the
mapping from |ψi to |ψ 0 i.
In quantum mechanics/theories, if an operator describing a mapping A b : H → H obeys the following
properties for arbitrary vectors |ψ1 i , |ψ2 i , |ψi ∈ H and an arbitrary number c ∈ C,
A |ψ1 i + |ψ2 i = A
b b |ψ1 i + Ab |ψ2 i , (2.2)
Ab (c |ψi) = c A b |ψi , (2.3)
Ab is called a linear operator or just an operator as technical slang. We will adopt this slang for the
following part.
6
• For an arbitrary |ψi ∈ H , the following mapping Ib : H → H obeying
Ib |ψi := |ψi (2.4)
is called identity operator or unit operator.
• Two operators A
b1 and A
b2 are mentioned as equal each other (A
b1 = A
b2 ) if
b1 |ψi = A
A b2 |ψi (2.5)
holds for an arbitrary |ψi ∈ H .
0,
is called zero operator. The following properties hold for b
A
b+b
0 = A,
b A
bb0=b
0Ab=b
0. (2.13)
A
bIb = IbA
b = A,
b (2.14)
which means that the identity operator is commutable with any operator A.b For measuring the
commutability between the two operators Ab and B,
b the following operator is widely used,
h i
A,
bB b := AbB b−B b A,
b (2.15)
7
• One can introduce an operator in the following form,
b := |χihξ| .
B (2.17)
b on the vector |ψi is recognised as,
The operation of B
Bb |ψi = |χihξ|ψi = hξ|ψi |χi , (2.18)
| {z }
∈C
The linear operators in the n-dimensional complex Euclidean space Cn (n ∈ N), which is a finite-
dimensional Hilbert space (H = Cn ), are represented as n × n matrices with complex components,
A11 A12 · · · A1n B11 B12 · · · B1n ψ1
A21 A22 · · · A2n B21 B22 · · · B2n ψ2
• • •
A
b= .. , B
b= .. , |ψi = .. . (2.19)
.. .. . . .. .. . .
. . . . . . . . .
An1 An2 · · · Ann Bn1 Bn2 · · · Bnn ψn
as the product of a matrix and a column vector in Cn . The identity operator and zero operators are
represented by the n × n identity and zero matrix,
1 0 ··· 0 0 0 ··· 0
0 1 · · · 0 0 0 · · · 0
• •
Ib = .. .. . . .. , 0 = .. .. . . .. . (2.21)
b
. . . . . . . .
0 0 ··· 1 0 0 ··· 0
8
Also, by use of the representations,
χ1 ξ1
χ2 ξ2
• •
|χi = .. , |ξi = .. , (2.25)
. .
χn ξn
b := |χihξ| is represented as
the operator B
χ1 χ1 ξ1∗ χ1 ξ2∗ · · · χ1 ξn∗
∗ ∗
χ2
∗ ∗ ∗ χ2 ξ1 χ2 ξ2 · · ·
χ2 ξn∗
B := |χihξ| = .. ξ1 , ξ2 , · · · , ξn = .. .. . (2.26)
b .. ..
. . . . .
∗ ∗
χn χn ξ1 χn ξ2 · · · χn ξn∗
hχ|Aψi b† χ|ψi
b = hA (2.27)
is called the Hermitian conjugate of A.b Note that, in the above, the right-hand side should be understood
as A b† |χi. In the
b |ψi, and the left-hand side should be recognised as the bra conjugate of the ket vector A
inner product, we can write Eq. (2.27) as follows:
hχ|A|ψi
b b† |χ|ψi ,
= hA (2.28)
hχ| A b† χ| .
b = hA (2.29)
hAψ|χi
b b ∗ = hA
= hχ|Aψi b† χ|ψi∗ = hψ|A
b† χi , (2.30)
which leads to
b ∗ = hψ|A
hχ|Aψi b† χi , (2.31)
9
and we get
n
X
⇔ AB |ψi =
b b A
b b |ψi ,
B
j jk k
k=1
n n n
!
X X X
⇔ A
bBb (|ψi)` = A
b B
b (|ψi)` , (2.34)
j` jk k`
`=1 `=1 k=1
n
X
⇒ A
bBb = A
b B
b . (2.35)
j` jk k`
k=1
Also,
n
X
hχ|Aψi
b = (|χi)∗j Ab |ψi
j
j=1
Xn
= b (|χi)∗ (|ψi)
A j `
j`
`,k=1
Xn ∗ ∗
= A (|χi)j (|ψi)` ,
b (2.36)
j`
`,k=1
Xn ∗
b† χ|ψi =
hA b†
A (|χi)j (|ψi)` , (2.37)
`j
`,k=1
In general, for two vectors |ψ1 i , |ψ2 i ∈ H , if the following condition holds,
∀ |χi ∈ H , hχ|ψ1 i = hχ|ψ2 i , (2.39)
we can conclude,
|ψ1 i = |ψ2 i . (2.40)
It is easy to derive this property by choosing |χi = |ψ1 i − |ψ2 i. Similarly, we can derive the property:
for two operators A b1 , A
b2 , if the following condition holds,
10
2.3 Eigenvalue of Operator
In general, the transformed vector of |ψi generated by A
b |ψi is not parallel to |ψi, while for specific vectors
|vi, A
b |vi can be parallel to |ψi. For an operator A,
b the nonzero vector |vi =
:::::::
6 |nulli and corresponding
complex number λ satisfies
b |vi = λ |vi
A (2.47)
are called eigenvector and eigenvalue, respectively. A set of eigenvalues is called a spectrum.
• If |vi is an eigenvector of A,
b c |vi (c ∈ C) is also. This means that the norm of an eigenvector is
not uniquely determined.
• Since
b |vi = λ |vi
A ⇔ A − λI |vi = b
b b 0, (2.48)
if the inverse of the operator Ab − λIb exists, |vi is concluded as |nulli and no eigenvector exists
in this case.
(iii) n linearly independent eigenvectors are always obtained, even when there is degeneracy in the
eigenvalues. By orthogonalising eigenvectors belonging to the same eigenvalue by the Gram-
Schmidt method, all eigenvectors can be made orthonormal. This means that we can prepare a
CONS from the eigenvectors of a Hermitian operator.
11
The proof of (i) is as follows. From Eq. (2.47), we get
hv|A|vi
b = hv|λvi = λ hv|vi
b† v|vi .
= hA (2.50)
b† = A,
If A b we get further
b† v|vi = hAv|vi
hA b = hλv|vi = λ∗ |v|vi , (2.51)
λ − λ∗ = 0 ⇔ λ = λ∗ , (2.53)
b† = A
where we have used A b and λ = λ∗ , and they lead to
hv|v 0 i = 0, (2.57)
which means that eigenvectors belonging to different eigenvalues are orthogonal. For the proof of (iii),
refer to a suitable linear algebra textbook.
b |v (s) i = λ |v (s) i .
A (2.58)
λ λ
The number of such eigenvectors (belonging to a focused eigenvalue λ) is called the degree of degeneracy,
and such eigenvalues are called degenerated. After a recombination of such eigenvectors, eigenvectors
can follow
(r) (s)
hvλ |vλ i = δrs (r, s = 1, 2, · · · , k) . (2.59)
12
The entire vector structure is represented as
n
X (r) (r)
|ψλ i = cλ |vλ i , (2.60)
r=1
n o
(1) (2) (k)
with coefficients cλ , cλ , · · · , cλ . Based on such eigenvectors of a Hermitian operator, we can
construct a CONS (after suitable orthogonalisation and normalisation), and an arbitrary vector |ψi ∈ H
is represented as
kλ
XX (r) (r)
|ψi = cλ |vλ i , (2.61)
λ r=1
where
(r) (r0 )
hvλ |vλ0 i = δλλ0 δrr0 . (2.62)
Here, if an eigenvector belonging to λ is not generated, r takes only 1, and we can skip depicting ‘(1).’
Every coefficient (for representing |ψi) is calculable only by taking the inner product,
(r) (r)
cλ = hvλ |ψi . (2.63)
it is called a projection operator, which is a Hermitian operator, and their eigenvalues are 0 or 1 (due to
λ2 = λ).
• The following projection operator is representing the projection to |χi,
b χ |ψi = |χihχ|ψi ,
Π b χ := |χihχ| .
Π (2.65)
hχ|χi hχ|χi
• If {|χ1 i , |χ2 i , · · · , |χk i} are orthonormal, the following one is a projection operator,
k
X k
X
b0 |ψi =
Π |χr ihχr |ψi , b0 :=
Π |χr ihχr | . (2.66)
r=1 r=1
n o
b |v (r) i , we can do
• Based on a CONS constructed by the eigenvectors of a Hermitian operator A, λ
the expansion (as we discussed),
kλ
XX kλ
XX
(r) (r) (r) (r)
|ψi = cλ |vλ i = |vλ ihvλ |ψi . (2.67)
λ r=1 λ r=1
13
By operating A
b from the left sides, we get
kλ
XX X Xkλ
b |ψi =
A b |v (r) ihv (r) |ψi =
A λ
(r) (r)
|vλ ihvλ |ψi
λ λ
λ r=1 λ r=1
X
= b λ |ψi ,
λΠ (2.68)
λ
with the projection operator toward the eigenvector λ,
kλ
X (r) (r)
Π
b λ := |vλ ihvλ | . (2.69)
r=1
14
3.2 Matrix representation of abstract operator
We can insert the identity operator Ib anywhere relevant, and thus we see
b |ψi = IbA
A bIb |ψi
X X
= |χr ihχr | A
b |χs ihχs |
r s
X
= |χr i hχr |A|χ
b i hχs |ψi
| {z s} | {z }
r,s =cs
=:Ars
X
= |χr i Ars cs . (3.7)
r,s
thus we get
X
dr = Ars cs , (3.9)
s
where this rule is equivalent to the product between a matrix and a column vector (as long as we use a
common CONS). Otherwise, we can focus on A b as
X
A
b = IbA
bIb = |χr ihχr |A|χ
b s ihχs | , (3.10)
r,s
Ors := hχr |b
0|χs i = 0, (3.12)
15
and the orthonormal vectors
1 0
|χ1 i := , |χ2 i := , (3.14)
0 1
are easily recognised as CONS in C2 . Also, we can easily show that
0 1 1 0 1 −1
|χ1 i := √ , |χ2 i := √ , (3.15)
2 1 2 1
is another CONS in C2 .
In general in a Hilbert space H , vectors of a CONS {|χ1 i , |χ2 i , · · · } and those of another CONS
{|χ01 i , |χ02 i , · · · } are related as
X X
|χs i = |χ0r i hχ0r |χs i = |χ0r i Urs , (3.16)
r r
with
Urs := hχ0r |χs i . (3.17)
For a vector |ψi ∈ H , when we expand it in the two ways,
X X
|ψi = |χs i cs = |χ0r i c0r ,
s r
X X
= |χ0r ihχ0r |χs i cs = |χ0r i Urs cs , (3.18)
s,r s,r
which leads to
X
c0r = Urs cs , (3.19)
s
where this follows the linear transformation by a matrix. Here, we can see
X X X
∗
U † sr Urq = hχs |χ0r ihχ0r |χq i = hχs |χq i = δsq ,
Urs Urq = (3.20)
r r r
16
3.4 Diagonalisation
Based on the eigen equation of the Hermitian operator A b† ),
b (= A
b |v (s) i = λ |v (s) i ,
A (3.25)
λ λ
3.5 Trace
In general, matrix elements are dependent on the choice of CONS. However, the trace and the determinant
of a matrix representation do not depend on the choice. Here, we focus on the trace.
X
TrA
b := hχr |A|χ
b r i = A11 + A22 + · · · (3.27)
r
• If the Hilbert space is finite-dimensional, the summation is among a finite number of elements, and
it converges.
• In a finite-dimensional Hilbert space, the following property is established,
X
Tr A bBb = hχs |A b si
bB|χ
s
X
= hχs |A|χ
b r ihχr |B|χ
b si
r,s
X
= hχr |B|χ
b s ihχs |A|χ
b ri
r,s
X
= hχs |B b si
b A|χ
s
= Tr B A ,
b b (3.28)
17
• It is straightforward to show the basis independence of the trace,
X
TrAb= b 0r i
hχ0r |A|χ
r
X
= b q ihχq |χ0 i
hχ0r |χp ihχp |A|χ r
r,p,q
X
= b q ihχq |χ0r ihχ0r |χp i
hχp |A|χ
r,p,q
X
= hχp |A|χ
b q ihχq |χp i
| {z }
p,q
=δqp
X
= hχp |A|χ
b pi . (3.33)
p
18