Professional Documents
Culture Documents
Chap 2
Chap 2
Orthogonality
ENSIA, December 2022/January 2023
Contents
• Inner Product
• Norm
• Orthogonal Sets
• Orthogonal Projection
• The Gram-Schmidt Process
• QR Factorization
• Least-Squares Problems
The Inner Product
Definition 1
𝑢1 𝑣1
Let 𝑢 = ⋮ and 𝑣 = ⋮ be two column vectors in ℝ𝑛 . The real number
𝑢𝑛 𝑣𝑛
𝑡
𝑢 𝑣 = 𝑢1 𝑣1 + ⋯ + 𝑢𝑛 𝑣𝑛
is called inner product of 𝑢 and 𝑣, and it is written as 𝑢 ∙ 𝑣.
Remark 1
The inner product is also referred to as a dot product.
Example
Example 1
2 −3
Let 𝑢 = 1 and 𝑣 = 5 . Then
−1 −2
𝑢 ∙ 𝑣 = 2 −3 + 1 × 5 + −1 −2 = 1.
Inner Product Properties
Theorem 1
Let 𝑢, 𝑣 and 𝑤 be vectors in ℝ𝑛 . Then we have the following properties
a) 𝑢 ∙ 𝑣 = 𝑣 ∙ 𝑢
b) 𝑢 + 𝑣 ∙ 𝑤 = 𝑢 ∙ 𝑤 + 𝑣 ∙ 𝑤
c) (𝑐𝑢) ∙ 𝑣 = 𝑐(𝑢 ∙ 𝑣) = 𝑢 ∙ (𝑐𝑣) for all 𝑐 ∈ ℝ
d) 𝑢 ∙ 𝑢 ≥ 0 and 𝑢 ∙ 𝑢 = 0 if and only if 𝑢 = 0
Corollary 1
For all 𝑢1 , ⋯ , 𝑢𝑝 , 𝑤 ∈ ℝ𝑛 and for all 𝑐1 , ⋯ , 𝑐𝑝 ∈ ℝ, we have
𝑐1 𝑢1 + ⋯ + 𝑐𝑝 𝑢𝑝 ∙ 𝑤 = 𝑐1 (𝑢1 ∙ 𝑤) + ⋯ + 𝑐𝑝 (𝑢𝑝 ∙ 𝑤)
Norm and Unit Vector
Definition 2
Let 𝑣 ∈ ℝ𝑛 . The norm of 𝑣 is the real number defined by
𝑣 = 𝑣1 2 + ⋯ + 𝑣𝑛 2 .
A vector whose norm is 1 is called a unit vector.
Remark 2
The norm is also referred to as a length.
Distance in ℝ𝒏
Definition 3
For 𝑢, 𝑣 ∈ ℝ𝑛 , the distance between 𝑢 and 𝑣 is given by
𝑑𝑖𝑠𝑡 𝑢, 𝑣 = 𝑢 − 𝑣 .
Example 2
We compute the distance between the vectors 𝑢 = (2,5) and 𝑣 = (3,2).
2 3 −1
We have 𝑢 − 𝑣 = − = , then
5 2 3
𝑑𝑖𝑠𝑡 𝑢, 𝑣 = −1 2 + 32 = 10.
Orthogonal Vectors and Pythagorean Theorem
Definition 4
Two vectors 𝑢 and 𝑣 in ℝ𝑛 are orthogonal if 𝑢 ∙ 𝑣 = 0.
Proposition 1
The null space (resp. the column space) of a matrix 𝐴 is a subspace of
ℝ𝑛 (resp. ℝ𝑚 ).
Theorem 4
Let 𝐴 be an 𝑚 × 𝑛 matrix. The orthogonal complement of the row
space of 𝐴 is the null space of 𝐴, and the orthogonal complement of
the column space of 𝐴 is the null space of 𝑡𝐴 :
Row 𝐴 ⊥ = Nul 𝐴 and Col 𝐴 ⊥ = Nul 𝑡𝐴.
Orthogonal Sets
Definition 7
A set of vectors 𝑢1 , ⋯ , 𝑢𝑝 in ℝ𝑛 is said to be an orthogonal set if each
pair of distinct vectors from the set is orthogonal, that is, if 𝑢𝑖 ∙ 𝑢𝑗 = 0
whenever 𝑖 ≠ 𝑗.
Theorem 5
If 𝑆 = 𝑢1 , ⋯ , 𝑢𝑝 is an orthogonal set of nonzero vectors in ℝ𝑛 , then 𝑆
is linearly independent and hence is a basis for the subspace spanned
by 𝑆.
Orthogonal Basis
Definition 8
An orthogonal basis for a subspace 𝑊 of ℝ𝑛 is a basis for 𝑊 that is also
an orthogonal set.
Theorem 6
Let 𝑢1 , ⋯ , 𝑢𝑝 be an orthogonal basis for a subspace 𝑊 of ℝ𝑛 , and let
𝑦 ∈ 𝑊. If
𝑦 = 𝑐1 𝑢1 + ⋯ + 𝑐𝑝 𝑢𝑝
then we have
𝑦∙𝑢𝑗
𝑐𝑗 = for all 𝑗 ∈ 1, ⋯ , 𝑝 .
𝑢𝑗 ∙𝑢𝑗
Example
Example 3
Let 𝑢1 = 3, 1, 1 , 𝑢2 = −1, 2, 1 and 𝑢3 = (− 1 2 , −2, 7 2). Then
𝑆 = 𝑢1 , 𝑢2 , 𝑢3 is an orthogonal basis for ℝ3 . Using Theorem 5, we can
express the vector 𝑦 = (6, 1, −8) as a linear combination of the vectors
in 𝑆. We have
𝑦 ∙ 𝑢1 𝑦 ∙ 𝑢2 𝑦 ∙ 𝑢3
𝑦= 𝑢1 + 𝑢2 + 𝑢3
𝑢1 ∙ 𝑢1 𝑢2 ∙ 𝑢2 𝑢 3 ∙ 𝑢3
11 −12 −33
= 𝑢1 + 𝑢2 + 𝑢3
11 6 33 2
= 𝑢1 − 2𝑢2 − 2𝑢3 .
Orthogonal Projection onto a Vector
Given a nonzero vector 𝑢 in ℝ𝑛 , consider the problem of decomposing a vector 𝑦 in ℝ𝑛
into the sum of two vectors, one a multiple of 𝑢 and the other orthogonal to 𝑢. We wish to
write
𝑦=𝑦+𝑧 (1)
where 𝑦 = 𝛼𝑢 for some scalar 𝛼 and 𝑧 is some vector orthogonal to 𝑢.
Proposition 2
The decomposition 1 is satisfied with 𝑧 orthogonal to 𝑢 if and only if
𝑦∙𝑢
𝛼= .
𝑢∙𝑢
Definition 9
The vector 𝑦 is called the orthogonal projection of y onto u, and the vector 𝑧 is called the
component of y orthogonal to 𝑢.
Example
Example 4
Let 𝑦 = (7, 6) and 𝑢 = 4, 2 . Then
𝑦 ∙ 𝑢 40
𝛼= = = 2.
𝑢 ∙ 𝑢 20
So the orthogonal projection of 𝑦 onto 𝑢 is
𝑦 = 2𝑢 = (8, 4)
and the component of 𝑦 orthogonal to 𝑢 is
𝑧 = 𝑦 − 𝑦 = 7, 6 − 8, 4 = −1, 2 .
Therefore, the decomposition of 𝑦 is given by
𝑦 = 𝑦 + 𝑧 = 8, 4 + −1, 2 .
Orthonormal Basis
Definition 10
A set of vectors 𝑢1 , ⋯ , 𝑢𝑝 in ℝ𝑛 is said to be an orthonormal set if it
is an orthogonal set of unit vectors. An orthonormal basis of a
subspace 𝑊 is an orthonormal set which spans 𝑊.
Example 5
The standard basis (𝑒1 , ⋯ , 𝑒𝑛 ) of ℝ𝑛 is an orthonormal basis.
Matrices with Orthonormal Columns
Theorem 7
A 𝑚 × 𝑛 matrix 𝑈 has orthonormal columns if and only if
𝑡
𝑈𝑈 = 𝐼.
Theorem 8
Let 𝑈 be an 𝑚 × 𝑛 matrix with orthonormal columns and let 𝑥, 𝑦 ∈ ℝ𝑛 . Then
a) 𝑈𝑥 = 𝑥
b) 𝑈𝑥 ∙ 𝑈𝑦 = 𝑥 ∙ 𝑦
c) 𝑈𝑥 ∙ 𝑈𝑦 = 0 if and only if 𝑥 ∙ 𝑦 = 0
Orthogonal Matrices
Definition 11
An orthogonal matrix is a square invertible matrix 𝑈 verifying
𝑈 −1 = 𝑡𝑈.
The Orthogonal Decomposition Theorem
Theorem 9 (The orthogonal Decomposition Theorem)
Let 𝑊 be a subspace of ℝ𝑛 . Then each 𝑦 ∈ ℝ𝑛 can be written uniquely
in the form
𝑦 =𝑦+𝑧 (1)
where 𝑦 ∈ 𝑊 and 𝑧 ∈ 𝑊 ⊥ .
In fact, if 𝑢1 , ⋯ , 𝑢𝑝 is any orthogonal basis of 𝑊, then
𝑦∙𝑢1 𝑦∙𝑢𝑝
𝑦= 𝑢1 +⋯+ 𝑢𝑝 (2)
𝑢1 ∙𝑢1 𝑢𝑝 ∙𝑢𝑝
and
𝑧 = 𝑦 − 𝑦.
Orthogonal Projection onto a Subspace
Definition 12
The vector 𝑦 in (1) of Theorem 9 is called the orthogonal projection of
𝑦 onto 𝑊, and it is denoted by
proj𝑊 𝑦.
Remark 4
The uniqueness of the decomposition (1) shows that the orthogonal
projection proj𝑊 𝑦 depends only on 𝑊 and not on a particular choice
of a basis of 𝑊.
Example
Example 6
Let 𝑢1 = 2, 5, −1 , 𝑢2 = −2, 1, 1 , 𝑦 = 1, 2, 3 and set 𝑊 = 𝑢1 , 𝑢2 .
The set 𝑢1 , 𝑢2 is orthogonal and the orthogonal projection of 𝑦 onto
𝑊 is given by
𝑦 ∙ 𝑢1 𝑦 ∙ 𝑢2 1
𝑦= 𝑢1 + 𝑢2 = −2, 10, 1 .
𝑢1 ∙ 𝑢1 𝑢 2 ∙ 𝑢2 5
So we can write
1 1
𝑦 = 1, 2, 3 = −2, 10, 1 + 7, 0, 14 = 𝑦 + 𝑧,
5 5
where 𝑦 ∈ 𝑊 and 𝑧 ∈ 𝑊 ⊥ .
Best Approximation Theorem
Theorem 10 (The Best Approximation Theorem)
Let 𝑊 be a subspace of ℝ𝑛 , let 𝑦 ∈ ℝ𝑛 and let 𝑦 be the orthogonal
projection of 𝑦 onto 𝑊. Then 𝑦 is the closest point in 𝑊 to 𝑦, that is
𝑦−𝑦 < 𝑦−𝑣 (3)
for all 𝑣 ∈ 𝑊, 𝑣 ≠ 𝑦.
Definition 13
The vector 𝑦 in (3) is called the best approximation to 𝑦 by elements of 𝑊.
Remark 5
The distance from 𝑦 to 𝑣, given by 𝑦 − 𝑣 , can be regarded as the error of
using 𝑣 in place of 𝑦. Theorem 10 says that this error is minimized for 𝑣 = 𝑦.
Example
Example 7
Let 𝑢1 = 2, 5, −1 , 𝑢2 = −2, 1, 1 , 𝑦 = 1, 2, 3 and 𝑊 = 𝑢1 , 𝑢2 , as
in Example 6, then
1
𝑦 = −2, 10, 1
5
is the closest point in 𝑊 to 𝑦.
Orthogonal Projection with Orthonormal Basis
Theorem 11
Let 𝑢1 , ⋯ , 𝑢𝑝 be an orthonormal basis for a subspace 𝑊 of ℝ𝑛 , and
let 𝑦 ∈ ℝ𝑛 . Then
proj𝑊 𝑦 = 𝑦 ∙ 𝑢1 𝑢1 + ⋯ + 𝑦 ∙ 𝑢𝑝 𝑢𝑝 4 .
If 𝑈 = 𝑢1 𝑢2 ⋯ 𝑢𝑝 , then
proj𝑊 𝑦 = 𝑈 𝑡𝑈𝑦.
Gram-Schmidt Process
Theorem 12 (The Gram-Schmidt Process)
Let 𝑥1 , ⋯ , 𝑥𝑝 be a basis for a nonzero subspace 𝑊 of ℝ𝑛 , and define
𝑣1 = 𝑥1
𝑥2 ∙ 𝑣1
𝑣2 = 𝑥2 − 𝑣1
𝑣1 ∙ 𝑣1
𝑥3 ∙ 𝑣1 𝑥3 ∙ 𝑣2
𝑣3 = 𝑥3 − 𝑣 − 𝑣
𝑣1 ∙ 𝑣1 1 𝑣2 ∙ 𝑣2 2
⋮
𝑥𝑝 ∙ 𝑣1 𝑥𝑝 ∙ 𝑣2 𝑥𝑝 ∙ 𝑣𝑝−1
𝑣𝑝 = 𝑥𝑝 − 𝑣 − 𝑣 −⋯− 𝑣 .
𝑣1 ∙ 𝑣1 1 𝑣2 ∙ 𝑣2 2 𝑣𝑝−1 ∙ 𝑣𝑝−1 𝑝−1
Then 𝑣1 , ⋯ , 𝑣𝑝 is an orthogonal basis for 𝑊. In addition
𝑣1 , ⋯ , 𝑣𝑘 = 𝑥1 , ⋯ , 𝑥𝑘 for 1 ≤ 𝑘 ≤ 𝑝.
Example
Example 8
Let 𝑊 = 𝑥1 , 𝑥2 where 𝑥1 = (3, 6, 0) and 𝑥2 = (1, 2, 2). Following the
Gram-Schmidt Process, we will construct an orthogonal basis for 𝑊. Set
𝑣1 = 𝑥1 and
𝑥2 ∙ 𝑣1 𝑥2 ∙ 𝑥1 15
𝑣2 = 𝑥2 − 𝑣1 = 𝑥2 − 𝑥1 = 1, 2, 2 − 3, 6, 0 = 0, 0, 2 .
𝑣1 ∙ 𝑣1 𝑥1 ∙ 𝑥1 45
Then 𝑣1 = 3, 6, 0 , 𝑣2 = (0, 0, 2) is an orthogonal basis for 𝑊.
If we want to construct an orthonormal basis for 𝑊, it suffices to normalize the vectors
𝑣 1 𝑣
𝑣1 and 𝑣2 , that is we take 𝑣1′ = 1 = (1, 2, 0) and 𝑣2′ = 2 = (0, 0, 1).
𝑣1 5 𝑣2
1
Now 𝑣1 ′ = (1, 2, 0), 𝑣2 ′ = (0, 0, 1) is an orthonormal basis for 𝑊.
5
Example
Example 9
Consider the subspace of ℝ4 given by
𝑊 = 𝑥1 , 𝑥2 , 𝑥3 ,
where 𝑥1 = 1, 1, 1, 1 , 𝑥2 = 0, 1, 1, 1 and 𝑥3 = (0, 0, 1, 1). We will construct an
orthonormal basis for 𝑊.
QR Factorization
Theorem 13
If 𝐴 is an 𝑚 × 𝑛 matrix with linearly independent columns, then 𝐴 can be factored
as 𝐴 = 𝑄𝑅, where 𝑄 is an 𝑚 × 𝑛 matrix whose columns form an orthonormal
basis for Col 𝐴 and 𝑅 is an 𝑛 × 𝑛 upper triangular invertible matrix with positive
entries on its diagonal.