Professional Documents
Culture Documents
Algebra 1
Algebra 1
Lecture note
Algebra I
I. Sets A few preliminaries.
I Sets 1
I.A. Relations . . . . . . . . . . . . . . . . . . . . . . . . . 1
I.B. Integers . . . . . . . . . . . . . . . . . . . . . . . . . . 5
I.C. Posets . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
II Groups 9
II.A. Introduction . . . . . . . . . . . . . . . . . . . . . . . 9
II.B. Permutation groups . . . . . . . . . . . . . . . . . . . 12
II.C. Groups and subgroups . . . . . . . . . . . . . . . . . 17
II.D. Cosets and Lagrange’s theorem . . . . . . . . . . . . 22
II.E. Homomorphisms and isomorphisms . . . . . . . . . 26
II.F. Group actions and Cayley’s theorem . . . . . . . . . 34
II.G. Conjugacy and the orbit-stabilizer theorem . . . . . 39
II.H. Cauchy’s Theorem . . . . . . . . . . . . . . . . . . . 48
II.I. Normal subgroups and quotient groups . . . . . . . 53
II.J. Automorphisms . . . . . . . . . . . . . . . . . . . . . 66
II.K. Generators and relations . . . . . . . . . . . . . . . . 70
II.L. The Sylow theorems . . . . . . . . . . . . . . . . . . 77
II.M. Some results on finite groups . . . . . . . . . . . . . 83
II.N. ”Not-Burnside’s” counting lemma . . . . . . . . . . 92
i
III Rings 99
III.A.Examples of rings . . . . . . . . . . . . . . . . . . . . 99
III.B. Ring zoology . . . . . . . . . . . . . . . . . . . . . . . 107
III.C.Matrix rings . . . . . . . . . . . . . . . . . . . . . . . 111
III.D.Ideals . . . . . . . . . . . . . . . . . . . . . . . . . . . 118
III.E. Homomorphisms of rings . . . . . . . . . . . . . . . 127
III.F. Fields . . . . . . . . . . . . . . . . . . . . . . . . . . . 136
III.G.Polynomial rings . . . . . . . . . . . . . . . . . . . . 142
III.H.Principal ideal domains . . . . . . . . . . . . . . . . 152
III.I. Unique factorization domains . . . . . . . . . . . . . 158
III.J. Greatest common divisors . . . . . . . . . . . . . . . 163
III.K.Gauss’s lemma and polynomials over UFDs . . . . . 172
III.L. Algebraic number rings . . . . . . . . . . . . . . . . 177
IV Modules 191
IV.A.Definition and examples . . . . . . . . . . . . . . . . 191
IV.B. Submodules and homomorphisms . . . . . . . . . . 199
IV.C. Modules over a PID . . . . . . . . . . . . . . . . . . . 211
IV.D.Applications to linear algebra . . . . . . . . . . . . . 232
IV.E. Endomorphisms . . . . . . . . . . . . . . . . . . . . . 240
ii
I. Sets
I.A. Relations
f: S→T
We say that:
• f is injective (written f : S ,→ T) if f (s) = f (s0 ) =⇒ s = s0 ;
• f is surjective (written f : S T) if f (S) = T; and
1Here I is called an index set; here, to each element of I there is associated a subset
Si ⊂ S.
1
2 I. SETS
∼
=
• f is bijective (written f : S → T) if f is injective and surjective
∼
=
(and we define its inverse map f −1 : T → S to send each f (s) 7→ t).
Composition of maps
f g
S / T /6 U
g◦ f (or g f )
( g ◦ f ) −1 = f −1 ◦ g −1 .
∼ ⊂ S × S.
If ( a, b) ∈ ∼, then we write “a ∼ b”.
(ii) ∼ is an equivalence relation if
(reflexivity) a ∼ a
(symmetry) a ∼ b =⇒ b ∼ a
(transitivity) a ∼ b and b ∼ c =⇒ a ∼ c
a ≡ b ⇐⇒ n | a−b.
(n)
a ∼ f b ⇐⇒ f ( a) = f (b).
2Note: “ ⇐⇒ ” means “iff” (i.e., if and only if). We use it here to define things.
I.A. RELATIONS 3
(I.A.4) ā := {b ∈ S | b ∼ a} ⊂ S.
S/∼ := { ā | a ∈ S} ⊂ P (S)
a ∼ b =⇒ f ( a) = f (b)
ν ! f¯
S/∼
commutes. As a simple example, consider the map f : Z → {1, −1}
sending n 7→ (−1)n , which is well-defined “mod 4”, i.e. modulo ≡.
(4)
So f¯ : {0̄, 1̄, 2̄, 3̄} → {−1, 1} sends 0̄, 2̄ 7→ 1 and 1̄, 3̄ 7→ −1. Obviously
this works for any other even integer; in particular, if we take ∼ to
be ≡, then f¯ is an isomorphism.
(2)
I.B. INTEGERS 5
I.B. Integers
n = p1 p2 · · · p s ,
where the { pi } are (positive) primes, which are not necessarily distinct.
p1 | (q1 − p1 ) =⇒ p1 | q1 =⇒ p1 = q1 ,
The FTA leads to the notion of the gcd (= greatest common divi-
sor) of m, n ∈ Z, written (m, n) and well-defined up to sign. To find
it, one traditionally employs the
6 I. SETS
in which the gcd is the last nonzero remainder ri .5 This is best cov-
ered and proved later in a more general context (that of principal ideal
domains). For now, we shall just show:
I.C. Posets
y ∈ S0 =⇒ y ≤ x.
x ≤ y and y ∈ S =⇒ x = y.
9There are arguments that avoid transfinite induction, but they take longer to even
partially understand. You can think of an ordinal number as an isomorphism class
of well-ordered sets (which are totally ordered sets each of whose subsets has a
least element). The class Ord of ordinal numbers is not a set – it is “too big”.
II. Groups
II.A. Introduction
•: G × G → G ,
satisfying
(i) [associativity] ( x · y) · z = x · (y · z)
(II.A.1) (ii) [identity] ∃ “1” ∈ G s.t. 1 · x = x = x · 1
(iii) [inverses] ∃“x −1 ” ∈ G s.t. x −1 · x = 1 = x · x −1
ϕ: G → H
ϕ( x · y) = ϕ( x ) · ϕ(y) (∀ x, y ∈ G )
and with ϕ(1G ) = 1 H . (In the case of groups, ϕ( x −1 ) = ϕ( x )−1
follows at once.)
1We’ll use a large dot “•” when talking about the operation, and a small dot “·”
when using it.
9
10 II. GROUPS
2When we say a group is finite or infinite, we are referring to the order of the
group, which means its order as a set (i.e. the number of elements).
12 II. GROUPS
Let X be a set; recall that (if finite) its order | X | is the number of
elements. A transformation of X is a map
τ: X → X ;
!
x1 ··· xn
Say X = { x1 , . . . , xn }. A useful notation is τ = .
τ ( x1 ) · · · τ ( xn )
1 α β γ
1 1 α β γ
α α 1 γ β
β β β β β
γ γ γ γ γ
You can make such a table for any (finite order) group or monoid;
but conversely, given an arbitrary table of the form
1 α β γ
1 1 α β γ
α α
β β ?
γ γ
Sn := S{1,...,n} .
3Here, “complete” means that we formally include the 1-cycles (k ) that do noth-
ing, except to say that α sends k to itself, so that each element of {1, . . . , n} appears
exactly once in the product of cycles. (A 1-cycle is really just a way of writing the
identity element.)
14 II. GROUPS
i1 7 → i2 7 → i3 7 → · · · 7 → ir ,
α α α α
P ROOF. Factor α into disjoint cycles, then (for example) factor the
cycles via the formula (123 · · · r ) = (1r )(1 r −1) · · · (13)(12).
sgn(α) := (−1)n−c(α) .
sgn(αβ) = sgn(τ1 · τ2 · · · τm β) =
− sgn(τ2 · · · τm β) = · · · = (−1)m sgn( β), and
x 00 = x 00 1 = x 00 xx 0 = 1x 0 = x 0 .
says that the elements of PSL2 (Z) are arbitrary “words” in S and R
(and their inverses) subject only to the two relations written. For the
dihedral group, we have
To determine the relation between these orders (in the finite case),
we consider more generally | H | for H ≤ G and introduce (left) cosets
aH := { ah | h ∈ H } ⊂ G.
II.D.2. P ROPOSITION . Distinct cosets are disjoint and have the same
number of elements.
(II.D.3) aH = bH ⇐⇒ b −1 a ∈ H ⇐⇒ a ∈ bH.
b −1 a ∈ H ⇐⇒ ∀h ∈ H, b−1 ah =: h0 ∈ H
⇐⇒ ∀h ∈ H, ah = bh0 for some h0 ∈ H
⇐⇒ aH ⊂ bH,
and similarly
|G|
II.D.5. D EFINITION . [ G:H ] := | H | ∈ N is called the index of H in
G, and is the number of cosets (as will be clear from the next proof).
G = a1 H q · · · q ar H
makes all of them 1 is 12. There is no element of actual order 12. (You
will check all of this in HW.) The next result says that we can blame
this on the fact that S4 is nonabelian:
ϕ( ab) = ϕ( a) ϕ(b)
We will first prove two lemmas. The start with, we should justify
calling injective homomorphisms “1-to-1”.
(iii) | D12 | = |S4 | = |Z24 | = 24. Z24 is abelian; the other two are
not: in S4 , (12)(23) = (123) 6= (132) = (23)(12), while in D12 ,
hr = r −1 h 6= rh. So Z24 D12 , S4 .
Now write out the cycle types for S4 :
The last row is just the identity element; the two rows above it indi-
cate that there are 3 + 6 = 9 elements of order 2 in S4 . Now D12 has
13 elements of order 2: the 12 reflections {hr a | a = 0, 1, . . . , 11}, and
one 180◦ -rotation r6 . So D12 S4 .
(v) All cyclic groups of order N are isomorphic to (Z N , +). Just write
down the homomorphism from Z N → hαi sending 1̄ 7→ α hence
m 7→ αm .
commute.
(A) hk = kh (∀h ∈ H, k ∈ K )
(B) H ∩ K = {1G }.
Then
(i) (A) =⇒ HK ≤ G
(ii) (A) + (B) =⇒ HK ∼ = H×K
(iii) (A) + (B) + HK = G =⇒ G ∼ = H×K
(iv) (A) + (B) + | G | < ∞ + | H ||K |=| G | =⇒ G ∼
= H × K.
9In more detail, this sends a mod ` to (a mod r, a mod s). Since r, s|`, this makes
sense. The map is injective because if ā goes to (0̄, 0̄), this means that r, s| a, so that
their lcm `| a and the original ā was 0̄.
10Here g|b =⇒ s = gs̃|bs̃, so it is well-defined.
II.E. HOMOMORPHISMS AND ISOMORPHISMS 33
ϕ
and (ᾱ, ᾱ) = ( β̄, γ̄). So we get aα 7→ (bβ, cγ) = (1̄, 1̄). Since ϕ is
injective on a set-theoretic level, aα must be = 1̄, hence ā ∈ Zrs
∗ .]
φ(n) = n ∏ (1 − 1p ).
p|n
p prime
n = p1e1 · · · pet t .
Z∗n ∼
= Z∗p1 e1 × · · · × Z∗pt et ,
and taking orders on both sides gives
∏ φ ( p i i ).
e
φ(n) =
i
φ( pe ) = pe − pe−1 = pe (1 − 1p ),
e 1 1
so φ(n) = ∏i pi i ∏i (1 − pi ) = n ∏ i (1 − pi ).
(ii) V ∼
= Z2 × Z2 : use H = h(12)(34)i and K = h(14)(23)i, same
idea as above.
34 II. GROUPS
G×X → X
( g, x ) 7→ g.x
satisfying:
(i) ( gh).x = g.(h.x ) for all g, h ∈ G and x ∈ X; and
(ii) 1G .x = x for all x ∈ X.
A set X with G-action is called a G-set.
G ( x ) := { g.x | g ∈ G } ⊂ X
II.F. GROUP ACTIONS AND CAYLEY’S THEOREM 35
n.r := r + n (r ∈ R, n ∈ Z).
3 2
4 1 h: flip
(r: rotation)
5 6
ϕ(r ) = (123456)
ϕ(h) = (26)(35).
4 ← fix
3 2
4
180◦
3 2
(v) One can play the same game with the group SC of rotational sym-
metries of the cube, acting (faithfully) on . . .
• . . . the vertex set: SC ,→ S8
• . . . the edge set: SC ,→ S12
• . . . the face set: SC ,→ S6
• . . . the set of interior diagonals: SC ,→ S4
II.F. GROUP ACTIONS AND CAYLEY’S THEOREM 37
# of total
possible
rotation type action on X possible # of
angles
axes elements
about facet midpoints
about vertices
(on the diagonals)
The example just concluded should have convinced you that there
are many natural ways of looking at some groups as subgroups of
permutation groups. But there is one “canonical” way:
g.g0 := gg0 .
Notice that the actions in the proof and remark aren’t so interest-
ing: the orbit of any element is the entire group. Fortunately, groups
also act on themselves in a more interesting way, which is our next
topic.
II.G. CONJUGACY AND THE ORBIT-STABILIZER THEOREM 39
The orbits of this action are called the conjugacy classes of G. There
are two notational ambiguities to get rid of here: first, since G can
act on itself in more than one way, we don’t write G ( g); second, if
an element lies in a subgroup H ≤ G, we need to distinguish G- and
H-orbits. (Even if the G-orbit lies in H, the H-orbit can be smaller.)
P ROOF. gh = hg (∀ g, h ∈ G) ⇐⇒ ghg−1 = h (∀ g, h ∈ G) ⇐⇒
cclG (h) = {h} (∀h ∈ G).
40 II. GROUPS
Now, rather than using brute force, we could cut down our work
by noticing that elements of cclS3 ((12)) must (like (12)) have order
2, hence be transpositions. But there is a still more powerful result.
then taking !
a11 a12 · · · akdk
η := ,
b11 b12 · · · bkdk
we have σ0 = ηση −1 .
(II.G.11) G= ä ccl( gi ).
i
x∼y ⇐⇒ y ∈ G(x)
def.
then
∼ is reflexive: x ∈ G ( x ) =⇒ x ∼ x
∼ is symmetric: clear from (∗)
∼ is transitive: y = g.x and z = h.y =⇒ z = h.( g.x ) = hg.x.
hence defines an equivalence relation. Of course, X/∼ is the set of
orbits G ( x ), which by I.A.5 are disjoint with union all of X. If |X| <
∞, and we pick one element xi in each orbit, then
(II.G.12) X= ä G ( x i ).
i
II.G.13. E XAMPLE . First let’s look at the sizes of orbits and stabi-
lizers in the actions by conjugation from our last two examples:
(i) orbit: cclS3 ((12)) = {(12), (23), (13)} (3 elements)
stabilizer: (S3 )(12) = {1, (12)} (2 elements)
. . . and |S3 | = 6 = 3 · 2.
(ii) orbit: cclD5 (h) = {h, rh, r2 h, r3 h, r4 h} (5 elements)
stabilizer: ( D5 )h = {1, h} (2 elements)
. . . and | D5 | = 10 = 5 · 2.
12To see the second “ ⇐⇒ ” below: the forward implication is trivial, since x ∈
G ( x ) too. Conversely, suppose x, y are in the same G-orbit G (z), viz. x = g.z and
y = h.z. Then hg−1 .x = hg−1 g.z = h.z = y, so y ∈ G ( x ).
II.G. CONJUGACY AND THE ORBIT-STABILIZER THEOREM 43
| G ( x )| · | Gx | = | G |.
(iii) If x, x 0 belong to the same orbit, then Gx and Gx0 are conjugate as
subgroups of G (hence of the same order/etc.).
(iv) If g, g0 belong to the same (left) coset of Gx , then they act the same way
on x.
G ( x ) = { gxg−1 | g ∈ G } =: cclG ( x )
Gx = { g ∈ G | gxg−1 = x } = CG ( x )
44 II. GROUPS
P ROOF. First, lay out the “chambers” into which you are going
to insert the elements {1, . . . , n} to get a cycle:
6!
II.G.18. E XAMPLES . (i) |cclS6 ((12)(34)(56))| = 2·2·2·3! = 15.
6!
(ii) |cclS6 ((12345)(6))| = (5·1!)( 1·1!)
= 144.
6!
(iii) |cclS6 (1234)(56))| = (4·1!)(2·1!)
= 90.
6!
The order of the centralizer is, in each case, |cclS6 (··· )|
.
II.G. CONJUGACY AND THE ORBIT-STABILIZER THEOREM 45
Now in spite of II.G.16, we may not have cclSn (σ ) = cclAn (σ) for
σ ∈ An :
OR
(II) |cclAn (σ)| = 12 |cclSn (σ )| ⇐⇒ CSn (σ) ⊂ An .
equiv.
In the second case, one says that the conjugacy class “breaks” in An .
If CSn (σ) ⊂ An , then CAn (σ) = CSn (σ) and so by (II.G.20) |cclAn (σ )| =
1
2 |cclSn ( σ )|.
Otherwise, CSn (σ ) contains an element of Sn \An (the odd per-
mutations), and CAn (σ) < CSn (σ), which by Lagrange means that
|CSn (σ)|
≥ 2.
|CAn (σ)|
But by (II.G.20) cclAn (σ) ⊆ cclSn (σ) =⇒ |cclSn (σ)||CAn (σ)| ≥
1
2 |cclSn ( σ )||CSn ( σ )| =⇒
|CSn (σ)|
≤2
|CAn (σ)|
(hence = 2). It follows that |cclSn (σ)| = |cclAn (σ)|.
The conjugacy classes are {1}, {−1}, {i, −i}, {j, −j}, {k, −k}.
For example, jij−1 = −jij = jij(kk) = j(ijk)k = −jk = iijk = −i.
K
k k
j i i −j
J j i
I i −j
-J
−k −k
-K
−J −J J J
Furthermore, we have i2 = K · KJ = J = −1, ij = −K · −IK = I = k,
I K I −J −K −I
ji = · = = −k = −ij, and ijk =
K J J · · = −1, and so on.
K I J
Now, to be honest, there are problems with the idea of “divid-
ing vectors in R3 ”, since at the end of the day there can be no “3-
dimensional division algebra over R” (as we’ll see later this semes-
ter). In any case, we get the right nonabelian group of order 8 and
that’s all we care about presently!
C ( G ) := { x ∈ G | gxg−1 = x ∀ g ∈ G } ,
Obviously we have:
(i) G is abelian ⇐⇒ G = C ( G );
(ii) C ( G ) is itself always abelian; and
(iii) |cclG ( x )| = 1 ⇐⇒ x ∈ C ( G ).
Recall also that if we take one representative xi in each conjugacy
class of G (| G | < ∞), then G = qi cclG ( xi ) and so
| X | = | G | p −1 .
Introduce an action of Z p on X by cyclic permutation:
as required.
Now for given x ∈ X, the Orbit-Stabilizer Theorem gives
|Z p ( x )||(Z p ) x | = |Z p | = p
and so |Z p ( x )| = 1 or p (depending on x). Clearly,
(II.H.6) ba ∈
/ hbi
G = hbi q ahbi
= {1, b, b2 , . . . , b p−1 } q { a, ab, ab2 , . . . , ab p−1 }.
Thus
In case (ii), a and b commute; use II.E.11 (on direct products) to de-
duce that G ∼ = Z p × Z2 . In case (i), we have just described the mul-
tiplication laws of D p .
| G | = |C ( G )| + ∑[ G:CG ( xi )],
i
Since 1 ∈ / cclG (h), we have |cclG (h)| = p (rather than p2 ) and thus
|CG (h)| = p; and since CG (h) ≥ C ( G ) > {1}, we must have CG (h) =
C ( G ). But h ∈ CG (h) (commutes with itself) and h ∈ / C ( G ), a contra-
diction. So the only possibility is . . .
Case (ii): |C ( G )| = p2 . We have |C ( G )| = p2 = | G | =⇒ C ( G ) =
G =⇒ G abelian. By Cauchy’s theorem, G 3 h of order p; let
H := hhi. Take g ∈ G \ H; it has order > 1 dividing p2 . If this order
is p2 then G ∼ = h gi ∼= Z p2 .
Otherwise, |h gi| = p; and setting K := h gi, we have:
• H ∩ K < K with order dividing |K | = p =⇒ H ∩ K = {1};
• hk = kh for every h ∈ H, k ∈ K because G is abelian; and
• | H ||K | = p2 = | G |.
Thus by II.E.11 G ∼= H×K ∼= Zp × Zp.
II.I. NORMAL SUBGROUPS AND QUOTIENT GROUPS 53
(II.I.1) CG ( H ) := { g ∈ G | ghg−1 = h ∀h ∈ H }
(II.I.2) NG ( H ) := { g ∈ G | gHg−1 = H }.
| G ( H )|| NG ( H )| = | G |,
or equivalently | G ( H )| = | G |/| NG ( H )| = [ G:NG ( H )].
Note that if G is abelian, all its subgroups are normal. Here are
some more interesting
(Why? See II.G.19.) The 2 ccl’s into which the 3-cycles split are
II.I.10. E XAMPLES .
(a) Both II.I.8 and II.I.9 give quick proofs that An E Sn :
• An = ker{sgn : Sn → Z2 } (identifying ({1, −1}, •) with (Z2 , +))
II.I. NORMAL SUBGROUPS AND QUOTIENT GROUPS 57
• [Sn :An ] = 2.
ϕ: H × K → G
(h, k) 7→ hk
This is 1-to-1: ϕ(h, k) = ϕ(h0 , k0 ) =⇒ hk = h0 k0 =⇒ (h0 )−1 h =
k0 k−1 ∈ H ∩ K = {1} =⇒ (h0 )−1 h = 1 = k0 k−1 =⇒ (h, k) = (h0 , k0 ).
Hence (by the pigeonhole principle) | H ||K | = | H × K | ≤ | G |.
(ii) By II.E.11(iv) we are done if (∀h ∈ H, k ∈ K) hk = kh. (Recall
in the proof of II.E.11 that this makes ϕ a homomorphism hence an
isomorphism.) Now K E G =⇒ (hkh−1 )k−1 ∈ K, while H E G
=⇒ h(kh−1 k−1 ) ∈ H. Hence hkh−1 k−1 ∈ H ∩ K = {1} =⇒
hk(kh)−1 = 1 =⇒ hk = kh.
σ ∈ H \{1}. Write
17Recall that this is false for A , and is true for A because the stabilizer of a 3-cycle
4 5
contains a transposition.
II.I. NORMAL SUBGROUPS AND QUOTIENT GROUPS 59
ν : G G/H
ϕ̄ : G/K → ϕ( G ) (≤ H )
(∗)
gK 7−→ ϕ( g)
|G|
is an isomorphism of groups. (In particular, |K |
= | ϕ( G )|.)
ν "" . ϕ̄
G/K
It commutes (see the end of §I.A) in the sense that ϕ̄ ◦ ν = ϕ:
(∗)
ϕ̄(ν( g)) = ϕ̄( gK ) = ϕ( g).
(II.I.23) S4 /V4 ∼
= S3 .
62 II. GROUPS
h / hK
ϕ
H / G/K
;
ν !! -
ϕ̄
H/K
(iii) injectivity of (†): Given H1 /K = H2 /K. Then for each h1 ∈ H1 ,
there exists h2 ∈ H2 such that h1 K = h2 K. But then h2−1 h1 ∈ K, and
so h1 = h2 k ∈ H2 . That is, we have shown that H1 ⊂ H2 . Similarly,
one has H2 ⊂ H1 ; and so H1 = H2 .
surjectivity of (†): Given H̄ ≤ G/K, H̄ is a collection of cosets.
Define H to be the union of these cosets (hence H ⊃ K), so that
h1 , h2 ∈ H =⇒ h1 K, h2 K ∈ H̄ =⇒ h1 h2 K = (h1 K )(h2 K ) ∈ H̄
&&
ν/ / ν̄ / / G/K
G G/K H/K
/ /
g gK ( gK )( H/K )
!! ∼
=
G/H
since H = ker(µ).
II.I.26. C OROLLARY. Given a homomorphism η : G G with kernel
K, let
Λ : = { H ≤ G | H ≥ K } ⊇ Λ 0 : = { H E G | H ≥ K }.
Then
(i) Sending H 7→ η ( H ) induces 1-to-1 correspondences
Λ ←→ {subgroups of G}
∪ ∪
Λ ←→ {normal sgps. of G}.
0
P ROOF. By II.I.21, G ∼
= G/K. Hence this is just parts (iii-iv) resp.
(v) of II.I.25.
II.I.27. S ECOND I SOMORPHISM T HEOREM . Let H ≤ G, K E G.
Then
(i) (K E) HK ≤ G.
64 II. GROUPS
(ii) H ∩ K E H.
∼
=
(iii) h(K ∩ H ) 7→ hK induces H/(K ∩ H ) → HK/K.
P ROOF. (i) HK = ∪h∈ H hK = ∪h∈ H Kh = KH implies that ( HK )2 =
H 2 K2 = HK, and also that (the set of all inverses of elements of HK)
( HK )−1 = K −1 H −1 = KH = HK. So HK is a subgroup of G.
(ii) Under ν : G G/K,
%% ∼
=
H/( H ∩ K )
19This is easy, and left to you. Consider separately the cases k = 2 (which doesn’t
occur for n = 6) and k > 2.
66 II. GROUPS
II.J. Automorphisms
∼
=
II.J.1. D EFINITION . An isomorphism ϕ : G → G is called an auto-
morphism of G.
Write
(α ◦ ı x ◦ α−1 ( g) = α( xα−1 ( g) x −1 )
= α( x )α(α−1 ( g))α( x )−1
| {z }
=g
= ıα( x ) ( g )
=⇒ αInn( G )α−1 ⊆ Inn( G ).
k!2k−1
but the binomial symbol is an integer, whereas (2k−2)!
is not an inte-
ger for k ≥ 4. Moreover, the k = 2 case (n− 2
2 ) = 2 is also impossible.
n −2
This leaves k = 3, and ( 4 ) = 1, which holds ⇐⇒ n = 6. We
conclude that for n 6= 6, α(C1 ) = C1 .
Now assume that n 6= 6, and let an automorphism α be given.
We have just shown that α sends transpositions to transpositions.
Suppose α((12)) = ( ab), and x ∈ {3, . . . , n}; then
and set gi := η (ei ) (where ei is the ith standard basis vector). Every
element of Zn is of the form ∑i ai ei , and η is surjective; thus, every
element of G is of the form η (∑i ai ei ) = ∑i ai η (ei ) = ∑i ai gi .
II.K.3. E XAMPLES . (i) Zn is f.g. (with one generator: 1̄), and iso-
morphic to Z/nZ.
(ii) Q is not f.g.: if you pick rs11 , . . . , rsnn then any ∑in=1 ai rsi can be repre-
i
sented with denominator ∏i si — clearly not possible for an arbitrary
rational number.
II.K. GENERATORS AND RELATIONS 71
FS ⊃ S
with the (universal) property that: for all groups G and maps f : S →
G, there exists a unique homomorphism ϕ : FS → G making the
diagram
S / F
S
f ~ ϕ
G
commute. In fact, FS ∼
= hSi. It is called the free group on S .
6 F 6 FO
( (
S v ϕ and S v η
( (
G G
commute. But then
6 F 6 G
( )
S v η◦ ϕ and S u ϕ◦η
( (
F G
commute as well, and then the uniqueness in the universal property
gives η ◦ ϕ = idF and ϕ ◦ η = idG . So F ∼
= G and we are done.
Henceforth I will drop hSi for free groups and use it only for
subgroups generated by a subset.
ϕ : FS G
G∼
= FS / ker( ϕ),
II.K. GENERATORS AND RELATIONS 75
(II.K.11) G∼
= FS /hhRii
— a presentation of G in terms of generators S and relations R. If
|R| < ∞, we say that G is finitely presented. We conclude with
some
α α
g=1 α2
α1 β2 β1
β1
β2 α2 α2 α1
g=2
β2 β1
α1
Y ( N ) := H/Γ( N )
|X| = |XG | + ∑ | G ( xi )|
i
= |X | + ∑[ G:Gxi ]
G
i
(II.L.2) | G | = |C ( G )| + ∑[ G:CG ( xi )]
i
| G | = pk m =⇒ ∃ H ≤ G with | H | = pk .
P ROOF. Assume the theorem holds for all groups of order less
than | G |. (The base case is just the one-element group {1}.) Here is
the inductive step.
26So if G is abelian, there are no x ’s (and one is in case (2) in the next proof).
i
78 II. GROUPS
ϕ : G G/K
be the quotient map. Since this map is p-to-1,27 the preimage ϕ−1 ( H0 ) ≤
G — which is a subgroup by II.I.25(iii) — has order pk as desired.
NG ( H ) := { g ∈ G | gHg−1 = H }.
and since P2 ≤ P1 P2 ,
pk = | P2 || P1 P2 | =⇒ | P1 P2 | = pk n
|G| pk m
|Γ| = # of cosets of P2 = [ G:P2 ] = = k = m 6≡ 0.
| P2 | p ( p)
Hence, |Γ P1 | 6= 0.
How might one use Sylow III, with its characterization of set of
Sylow p-subgroups of G as one big orbit under conjugation? The
orbit-stabilizer theorem says that the size of any orbit must divide
the order of G; so we get immediately that
ϕ : S5 → S6
|S |
with | ϕ(S5 )| ≥ 6 =⇒ | ker( ϕ)| = ϕ(S5 ) ≤ 20. But by HW 3, A5 (of
5
order 60) is the only nontrivial normal subgroup of S5 ; hence ker( ϕ),
being normal in S5 , must be trivial, and ϕ injective.
Now we claim that ϕ preserves parity (i.e. σ odd =⇒ ϕ(σ ) odd).
Suppose first of all that ϕ(S5 ) was contained in A6 ; then it has index
3 and the action of A6 (by left translation) on its cosets maps A6 →
S3 nontrivially hence (since A6 is simple and kernels are normal)
injectively. But this contradicts |A6 | > |S3 |, and so ϕ(S5 ) 6⊂ A6 ,
ϕ sgn
making the composition S5 → S6 → Z2 surjective. Again, A5 is
82 II. GROUPS
8 , 12 , and 15.
(II.M.3) b2 = a2 or 1.
By the normality, bab−1 ∈ h ai. Since bab−1 has the same order as
a, bab−1 = a±1 . But if bab−1 = a, then b and a commute and G is
abelian. So
i.e. ba = a−1 b.
Now (II.M.3) and (II.M.4) completely describe the multiplication
in a group of order 8 with elements 1, a, a2 , a3 , b, ba, ba2 , ba3 :
T := h a, b | a6 = 1, b2 = a3 = ( ab)2 i.
There is only one group of order 15, namely Z15 ; this will follow
from results below on groups of order pq. But there are 14 non-
isomorphic groups of order 16, so that’s a good place to stop this initial
mini-foray into group classification.
High(er) order. Here we can really get going with the Sylow the-
orems, but first we require a few “counting” results. The first is im-
mediate from HW 4 #2:
| H ||K |
II.M.5. L EMMA . H, K ≤ G =⇒ | H ∩K |
≤ | G |.
e
For n = ∏it=1 pi i , recall that
Aut(Zn ) ∼
= Z∗n ∼
= Z∗pe1 × · · · × Z∗pet
1 t
(II.M.6) Aut(Z p × · · · × Z p ) ∼
= GL(k, Z p ),
| {z }
k times
30For this to make sense, you need to multiply and add in Z , which means to
p
consider it as a ring. Since we’re about to start that part of the course, it seems
fair to mention it! Of course, the automorphisms on the left of (II.M.6) are as an
abelian group, and GL(k, Z p ) itself is a (nonabelian) mutliplicative group. This
generalizes our earlier example Aut(Z2 × Z2 ) ∼ = S3 from II.J.4(i) of the group of
automorphisms of an abelian group being nonabelian, since S3 ∼ = GL(2, Z2 ).
86 II. GROUPS
ϕ : K → Aut( H ).
II.M. SOME RESULTS ON FINITE GROUPS 87
(5 − 1) · 6 + (3 − 1) · 10 + 1 = 45 elements,
which, well, it doesn’t.
How about | G | = 40 = 23 · 5? Let P be a Sylow 5-subgroup. If N
is its number of conjugates, then N ≡ 1, N 6= 1, and N | 40. There’s
(5)
no such number N.
Finally, there’s | G | = 56 = 23 · 7 to take out. Write P for a Sylow 7-
subgroup, with N7 conjugates. We have N7 ≡ 1, N7 6= 1, and N7 | 56
(7)
=⇒ N7 = 8. Playing the same game with a Sylow 2-subgroup Q
gives N2 ≥ 3 conjugates. Now the conjugates of P can’t intersect
each other or those of Q outside of the identity element; not counting
the identity, this furnishes N7 · (7 − 1) = 8 · 6 = 48 distinct elements
of G. On the other hand, the conjugates of Q (which, remember, have
order 8) can intersect in order-4 subgroups. Without thinking too
hard, we at least get (counting the identity) 12 additional elements
of G by considering just Q and one conjugate. This again produces a
contradiction since 12 + 48 = 60 > 56, concluding the proof.
II.M. SOME RESULTS ON FINITE GROUPS 89
1 / Z2n / Dicn / Z2 / 1.
/
b 1̄
/ /
1̄ a 0̄
Referring to (II.M.12), one says that
Dn = h a, b | an = 1, b2 = 1, b−1 ab = a−1 i
c b 1
with entries mod 3 (i.e. in Z3 ). While non-abelian, it has the same
number of elements of each order as Z3 × Z3 × Z3 . There is a natural
way to write this as an extension of Z3 × Z3 by Z3 . Does it split?
(This is a useful question to ask when looking at HW 5 #1.)
92 II. GROUPS
S := {( g, x ) | g.x = x } ∼
=
set
ä Xg ∼
=
set
ä Gx .
g∈ G x ∈X
| G0 |
point-free elements G0 := { g ∈ G | χ( g) = 0} ⊂ G has C := |G|
=
1
R
G0 1 ≥ |X| .
Burnside plus transitivity tell us that
Z Z
(II.N.6) 1 = |X/G | = χ and 2 ≤ |(X × X)/G | = χ2 ,
G G
32Up to this point, this Corollary is a theorem of Jordan from 1872. The last sen-
tence is due to Cameron-Cohen (1992) and its proof to Serre (2003).
94 II. GROUPS
type of # of # of fixed
element of D5 such points |Xg |
1 (identity) 1 25 ( = | X | )
23 (edges flipped into each other must be
hr k (flip) 5
same color in order to be fixed)
2 (all edges must be same color to yield
r k (rotation) 4
a “fixed point”)
|X/G | = 1
| D5 | ∑ |X g | = 1
10 {1 · 2
5
+ 5 · 23 + 4 · 2} = 8.
g∈ D5
{ p g , p0g }.
[
X :=
g∈ G \{1}
r
n
|X| = ∑ ni .
i =1
=⇒ each ni ≥ 2
r 2
=⇒ r ≤ 2 +2− n
r 2
=⇒ 2 ≤ 2− n < 2,
2 1 1 1
so r < 4. If r = 1 then n = n1 +1 ≥ n + 1 =⇒ n ≥ 1, which is
absurd.
=⇒ r = 2 or 3.
1 1 2
Case r = 2: 2= n1 + n2 +2− n
2 1 1 2
=⇒ n = n1 + n2 ≥ n
2 1 1
=⇒ n = n1 + n2 (and n1 , n2 ≤ n)
=⇒ n1 = n = n2
=⇒ every g ∈ G stabilizes every x ∈ X.
The only way that even one non-identity element of G stabilizes ev-
ery x ∈ X, is if there are only 2 points p, p0 ∈ X; and then the elements
of G are rotations about the axis they span (by multiples of 2π n )
=⇒ G ∼
= Zn .
II.N. “NOT-BURNSIDE’S” COUNTING LEMMA 97
(∗) 1
Case r = 3: (1 <) n2 + 1 = n1 + 1
n2 + 1
n3 . We may assume 2 ≤
n1 ≤ n2 ≤ n3 ≤ n. Now
• if n1 > 2 then RHS(∗) ≤ 1, which is impossible; so n1 = 2.
• if n2 ≥ 4 then RHS(∗) ≤ 1, again impossible; so n2 = 2 or 3.
• if n2 = 3 then for the same reason, n3 < 6.
Thus our options are confined to the left-hand column of the follow-
ing table:
geometric realization:
(n1 , n2 , n3 ) n (= | G |) G
rotational symmetries of . . .
prism on
(2, 2, k)k≥2 2k Dk ···
regular k-gon
(2, 3, 3) 12 A4 tetrahedron
(2, 3, 4) 24 S4 cube
(2, 3, 5) 60 A5 icosahedron
0 − sin(θ ) cos(θ )
and so tr( g) = 2 cos(θ ) + 1.
• From matrix algebra, you know that the trace of g is independent
of the choice of basis. Hence, we must have 2 cos(θ ) ∈ Z. Since g
k , for some k ∈ Z. But 2 cos( k ) ∈
is a finite-order rotation, θ = 2π 2π
Z only for k = 1, 2, 3, 4, 6.
Therefore, for a crystallographic group G ≤ SO(3), we must have that
all of the ni belong to {1, 2, 3, 4, 6}. Throwing out all other groups in
our list, we are left with 10 nontrivial rotational crystallographic groups:
Z2 , Z3 , Z4 , Z6 , D2 (∼
= V4 ) , D3 (∼
= S3 ) , D4 , D6 , A4 , and S4 .
III. Rings
The theory of rings and ideals grew out of several 19th and early
20th Century sources:
• polynomials (Gauss, Eisenstein, Hilbert, etc.);
• number rings (Dirichlet, Kummer [“ideal numbers”], Kronecker,
Dedekind [“ideals in number rings”], Hilbert, etc.); and
• matrix rings and hypercomplex numbers (Hamilton [quaternions],
Cayley [octonions], etc.).
Specifically, the term Zahlring showed up in the study of what we
would now call rings of integers in algebraic number fields; e.g. cy-
clotomic rings such as Z[ζ 5 ] (ζ 5 = a 5th root of 1) arose in the context
od attempts to prove Fermat’s last theorem, and ζ 5 “cycles back to
itself” (suggesting a ring) upon repeatedly taking powers. Here is
the modern definition, due to E. Noether (∼1920):
What about
√ √
Z[ 1+2 D
] : = {m + n( 1+2 D
) | m, n ∈ Z}
√
= { a+b2 D
| a, b ∈ Z, a ≡ b} ?
(2)
a−b
(For the last equality, take m = 2 and n = b.) Of course, the issue
is multiplicative closure:
√ √
(m + n( 1+2 D
))(m0 + n0 ( 1+2 D
)) =
√ √
mm0 + (mn0 + nm0 )( 1+2 D
) + nn0 ( (1+ D)+
4
2 D
).
√
| {z }
nn0 ( D −1) 1+ D
4 +nn0 ( 2 )
R[ x ] := {sequences (r0 , r1 , . . . , rn , 0, 0, . . . ) | ri ∈ R}
| {z }
zero from
some point on
102 III. RINGS
( a + b) · c = (∑kj=0 ( a j + b j )ck− j )
= (∑kj=0 a j ck− j ) + (∑kj=0 b j ck− j ) = a · c + b · c
and
= a · ( b · c ),
so that II.A.1(iii) is satisfied.
Now identify R with the subring {(r, 0, 0, . . .)} ⊂ R[ x ]. Taking
x := (0, 1, 0, 0, . . .), we have x n = (0, . . . , 0, 1, 0, 0, . . .) so that
| {z }
n
(r0 , r1 , r2 , . . . , rn , 0, 0, . . .) = rn x n + · · · + r1 x + r0 ,
which is obviously a much more appealing (and standard) notation.
We can also (inductively) define polynomial rings in several vari-
ables by
R[ x1 , . . . , xn ] := ( R[ x1 , . . . , xn−1 ]) [ xn ].
For any r ∈ R, we can consider the evaluation map
evr : R[ x ] −→ R
sending rn x n + · · · + r1 x + r0 7−→ rn r n + · · · + r1 r + r0 .
(v) Quaternions. The ring version is built out of the group one: put
H := { a + i + cj + dk | a, b, c, d ∈ R} ,
where the eij are formal symbols. Taking A = ∑i,j aij eij , B = ∑i,j bij eij ,
we set2 0 := ∑i,j
n n n
=1 0eij , 1 : = ∑i,j=1 δij eij = ∑i =1 eii , and
n n
A + B := ∑ (aij + bij )eij and AB := ∑ (∑nk=1 aik bkj )eij .
i,j=1 i,j=1
and the associativity of R; the rest is left to you.3 Of course, these can
be represented in the standard way as matrices
a11 · · · a1n
A = ... ..
.
..
.
an1 · · · ann
and you may think of eij as the matrix with a 1 at the (i, j)th place
and zeroes elsewhere. We have
(
0, j 6= k
eij ek` =
ei` , j = k.
The noncommutativity is highly visible this way.
Here are some definitions which were clearly not possible (or not
interesting) for groups.
(III.A.9) ( x + y) p = x p + y p ,
the so-called “Freshman’s dream”.
C ( R) := {r ∈ R | rs = sr ∀s ∈ R}.
III.A. EXAMPLES OF RINGS 105
In particular, the (k, `)th entry of the last line is akk − a`` and the
(i, `)th entry (for i 6= k) is aik . So off-diagonal entries of A are 0 and
the diagonal ones are all equal. Finally, consider Ar − rA.
interpreted as an isomorphism
h √ i∗
Z × Z2 ∼ = Z 1+2 5
(III.A.14) √ a
( a, ±1) 7→ ± 1+2 5 .
√ √ √
x +y 5 1+ 5 x −y 5
Given α = 2 ∈ R := Z[ 2 ],
write α̃ := ∈ R. The com-
2
position law that led to the group structure on LHS(III.A.14) was ex-
actly multiplication in R. Moreover, ( x, y) solves the above equation
⇐⇒ α · (±α̃) =√ 1 =⇒ α ∈ R∗ . Conversely, if α ∈ R∗ , then there
x 0 +y0 5
exists α0 = 2 ∈ R with αα0 = 1, and then (αα̃)(α0 α̃0 ) = αα0 αα
f0 =
x2 −5y2
11̃ = 1. Since x ≡ y, we have that x2 ≡ 5y2 =⇒ αα̃ = 4 ∈ Z,
(2) (4)
and similarly for α0 α̃0 . So the only way the product of αα̃ and αα
f0 is
1, is if they are both ±1, and then α ∈ R∗ .
So far we have discussed only quadratic number fields and num-
ber rings. To give a brief glimpse ahead, a general result of Dirichlet
says that for a number field K with r1 distinct real embeddings and
r2 pairs of conjugate complex embeddings,5
(III.A.15) OK∗ ∼
= Zr1 +r2 −1 × {torsion group},
where OK ⊂ K is the ring of integers of K. The main point is that
(III.A.14) is a special case (with r1 = 2 and r2 = 0) of a much more
general result.
5All number fields can be viewed as vector spaces over Q of some finite dimension,
called the degree [K:Q]. In this case, that degree is r1 + 2r2 . (An embedding of fields
means an injective homomorphism, √ in this case into R or C. These notions will be
discussed later.) The case K = Q[ D ] has r1 = 0 and r2 = 1 if D < 0, or r1 = 2
and r2 = 1 if D > 0.
III.B. RING ZOOLOGY 107
(III.B.1)
?_
COMMUTATIVE
RINGS
RINGS
?
Definition: ? COMMUTATIVE
No 0-divisors / DOMAINS ?_ (or “INTEGRAL”)
(left or right) DOMAINS
?
INTEGRALLY
(HW) CLOSED
DOMAINS
?
?
UNIQUE
Definition: /
DIVISION
FACTORIZATION
R∗ = R\{0} RINGS
DOMAINS(UFDs)
?
PRINCIPAL
IDEAL
DOMAINS(PIDs)
?
Definition: ?
EUCLIDEAN
commutative / FIELDS
DOMAINS
division ring
108 III. RINGS
δ : R\{0} → N
with the following property: for all a ∈ R and b ∈ R\{0}, there exist
q, r ∈ R satisfying a = bq + r and either δ(r ) < δ(b) or r = 0. (This δ
is called a Euclidean function and is not unique.)
Clearly, these are just the domains to which we can generalize the
(Euclidean) division algorithm I.B.3.
P ROOF. (a) Put δ(m) := |m| and use the division algorithm.
(b) Writing α = a + bi, put δ(α) := αᾱ = |α|2 = a2 + b2 . Let α ∈ Z[i]
and β ∈ Z[i]\{0}. We will find µ and ρ in Z[i] such that α = βµ + ρ
and δ( β) > δ(ρ).
Working in Q[i], we have αβ−1 = x + yi; pick m, n ∈ Z such that
e := x − m and η := y − n have |e|, |η | ≤ 12 . Then
(Note that if 2|char( R), then (iii) does not follow from (ii), but it fol-
lows directly from III.C.1.) Moreover, since sgn(σ−1 ) = sgn(σ),
which are invertible, have the following effects on det( A): none;
multiply by −1; multiply by r (respectively). EROs correspond to
multiplying A on the left by the elementary matrices
elementary matrix det
1 + reij
(I) 1
(III.C.4)
(II) 1 + eij + e ji − eii − e jj −1
(III) 1 + (r − 1)e
r.
ii
112 III. RINGS
(III.C.6) rref( A) = EN · · · E1 A
| {z }
as in (III.C.4)
Since the det( Ei ) 6= 0 and the Ei are invertible, this yields the
(III.C.9) EN · · · E1 ( A | 1) = ( EN · · · E1 A | EN · · · E1 ) ,
| {z } | {z } | {z }
n×2n 1 A −1
arguing by induction on M.
Turning to the next level of generality, suppose that R is a Eu-
clidean domain. To produce an analogue of the rref, find the nonzero
entry b in the first (nonzero) column with the lowest δ(b), then apply
III.C. MATRIX RINGS 113
a = bq + r to the other entries in that column and use a type (I) ERO
to kill the bq’s. Repeat this step on the column until all but one en-
try is zero; swap it to the top position. Restricting to the (n − 1) × n
submatrix below the first row, we repeat the algorithm to arrive at
0 · · · 0 r1 ∗ · · · ∗ s
0 · · · · · · · · · · · · 0 r2
?
0
where δ(αi,µ( j) ) < δ(α j,µ( j) ) for all i < j, and all entries below the
diagonal are 0. Clearly we still have (by (III.C.3)-(III.C.4)) that
(
Herm( A) = EN · · · E1 A
(III.C.12)
det(Herm( A)) = (∏i det( Ei )) det( A)
where ∏i det( Ei ) ∈ R∗ , and so
(III.C.13)
det( A) 6= 0 ⇐⇒ det(Herm( A)) 6= 0 ⇐⇒ µ( j) = j (∀ j).
| {z }
=∏in=1 αii
⇐⇒
(III.C.13)
∏ αii ∈ R∗ (III.C.14)
⇐⇒ αii ∈ R∗ (∀i ).
But if the αii ∈ R∗ , we may (if necessary) kill all the off-diagonal
entries (by type (I) EROs) and scale the diagonal ones to 1 (by type
(III) EROs). Since elementary matrices are invertible, we arrive at the
where Tn denotes all maps from {1, . . . , n} to itself (not just permu-
tations). For each µ ∈ Tn , let B(µ) denote the n × n matrix whose ith
row is the µ(i )th row of B for each i. Applying a variant of the first 3
steps above in reverse gives
n n n
∑ sgn(ρ) ∑ ∏ a j,µ( j) bµ( j),ρ( j) = ∑ ∑ ∏ a j,µ( j) ∏ bµ(ρ−1 (k)),k
ρ ∈Sn µ ∈Tn \ Sn j =1 ρ ∈Sn µ ∈Tn \ Sn j =1 k =1
!
n n
= ∑ ∏ a j,µ( j) ∑ sgn(ρ) ∏ bµ(ρ−1 (k)),k
µ ∈Tn \ Sn j =1 ρ ∈Sn k =1
!
n n
∑ ∏ a j,µ( j) ∑ sgn(ρ) ∏ bρ−1 (k),k
(µ)
=
µ ∈Tn \ Sn j =1 ρ ∈Sn k =1
!
n
= ∑ ∏ a j,µ( j) det B(µ) = 0
µ ∈Tn \ Sn j =1
since B(µ) has repeated rows for µ ∈ Tn \Sn . This shows that after
multiplying (III.C.17) by sgn(ρ) and summing over ρ ∈ Sn we can
omit terms with µ ∈/ Sn , proving the boxed equality.
What about invertibility? Well, you may recall that adjugate ma-
trices and Cramer’s rule were always the most horrible approach to
computing inverses and solving systems of equations in linear alge-
bra, unless the matrix entries were (say) transcendentals, polynomi-
als, etc. Since our entries are now in a general commutative ring,
let’s try this approach. Defining the cofactor
(n−1)×(n−1)
z }| {
(III.C.18) Aij := (−1)i+ j det ( ak` )k 6= i ,
` 6= j
7When expanding the LHS, the µth term on the RHS is obtained by choosing the
µ( j)th term (of the sum in parentheses) from the jth factor, as j runs from 1 to n.
116 III. RINGS
Now let A0 (resp. A00 ) be the matrices obtained by deleting the ith
row (resp. column) and replacing it by the jth row (resp. column).
The repeated row (resp. column) makes the determinant zero:
)
0 = det A0 = ∑n`=1 a j` Ai`
(III.C.20) ( j 6 = i ).
0 = det A00 = ∑nk=1 akj Aki
Overall, (III.C.19)-(III.C.20) =⇒
n n
(III.C.21) ∑ a j` Ai` = (det A)δij = ∑ akj Aki .
`=1 k =1
we have the
Even for R a field, for n > 1 Mn ( R) is not a division ring; even the
subring of diagonal matrices isn’t! On the other hand, the diagonal
matrices with all entries equal yield a subring which is isomorphic
III.C. MATRIX RINGS 117
Hence, all nonzero entries of this type are invertible, and this subset
is a division ring, which we will later identify with H.
8Keep in mind here that GL ( R) is not a subring, since it’s not closed under addi-
n
tion. It’s just a group under multiplication.
118 III. RINGS
III.D. Ideals
/ R ∗ ∪ {0})
and (for r ∈
!
defn. r = ab ( a, b ∈ R)
(III.D.2) r is irreducible ⇐⇒ .
=⇒ a or b ∈ R∗
If u ∈ R∗ and r = su, one writes r ∼ s and says that r and s are
associate;9 since s = ru−1 , this is an equivalence relation. The irre-
ducibles of Z are clearly the (±)primes.
√
Consider R = Z[ d], equipped with the “norm map”
N: R→Z
(III.D.3)
r 7→ rr̃,
√ √
where is r = m + n d, r̃ = m − n d.
Then (III.D.7) becomes π12 π22 = (π1 π2 )2 . Kummer showed that one
could construct a theory in which such elements would formally re-
spect divisibility and distributive properties. (Later it was realized
that they could be represented by actual elements in the “Hilbert
√
class field of Q( 10)”.) But Dedekind had the even nicer idea of
√
characterizing an “ideal number” π by its “shadow” in Z[ 10], con-
sisting of everything (formally) divisible by π. This is essentially our
modern notion of an ideal (in a number ring — the notion in gen-
eral is due to E. Noether). Indeed, the “shadows” of π1 and π2 in
the above example will be (in the notation about to be defined) the
ideals
√ √
(III.D.9) (3, 1 + 10) and (3, −1 + 10).
We will return to this example below.
Turning to some generalities, we have the
( a1 , . . . , a m ) = {r1 a1 + · · · r m a m | r1 , . . . , r m ∈ R }.
11Note that this is a stronger notion than being a “subrng” because of the closure
under multiplication by elements of R. And yes, I mean “subrng” not “subring”:
except for R itself, ideals in R do not contain 1.
III.D. IDEALS 121
P ROOF. If ra = b, then
(III.D.18) ( a + I ) + (b + I ) = ( a + b) + I.
13
This is an important point: the product ( a, b)(c, d) is the ideal generated by the
set of products { a, b} {c, d} := { ac, ad, bc, bd}.
14The basic principle being applied here (in the case of 1-element sets) is that if a
set S is contained in an ideal I, then (S) ⊂ I.
III.D. IDEALS 123
(III.D.19) ( a + I )(b + I ) := ab + I,
with identity coset 1 + I. The main check required is that (III.D.19) is
well-defined: given a0 = a + α ∈ a + I and b0 = b + β ∈ b + I,
( a0 + I )(b0 + I ) = a0 b0 + I = ( a + α)(b + β) + I
= ab + αb + aβ + αβ + I
| {z }
∈I
= ab + I.
Distributivity is clear from (III.D.18)-(III.D.19) and distributivity in
R. Hence, R/I has the structure of a ring.
III.D.20. R EMARK . In our study of groups, we had two “stupid
quotients”, G/h1i(∼= G ) and G/G (∼ = {1}). Here, the only stupid
quotient ring is R/(0) = R; because {0} is not a ring, we cannot con-
sider R/R, and accordingly the definition of quotient ring requires a
proper ideal.
III.D.21. E XAMPLES . (i) (n) = nZ ⊂ Z is a proper ideal (n > 1),
and Z/(n) (or Z/nZ) is just Zn (viewed as a ring).
(ii) In Z[ x ]/( x2 − 10), any element is of the form P( x ) + ( x2 − 10)
(where ( x2 − 10) is the principal ideal). Applying polynomial divi-
sion, this equals { x2 − 10} · Q( x ) + R( x ) + ( x2 − 10) = R( x ) + ( x2 −
10), where R( x ) = ax + b.
(iii) In the ring C0 (M) of continuous functions on a manifold M, the
subset IS of functions identically zero on a subset S ⊂ M is an ideal.
In C0 (M)/IS , cosets f + IS and g + IS are the same ⇐⇒ f − g ∈ IS
⇐⇒ f and g have the same restriction to S . So the quotient can be
thought of as a ring of functions on S of some sort.
√ √
(iv) We can consider Z[ 10] modulo the ideals (3), (±1 + 10), and
√
(3, ±1 + 10).
(v) Let R be commutative. While there are left [resp. right] ideals in
Mn ( R) (e.g. matrices with last column [resp. row] zero) that “take
124 III. RINGS
Then J ⊂ Mn ( I ).
To show I is an ideal: given A ∈ J, J contains
eki Ae j` = eki (∑m,n amn emn )e j` = ∑m,n amn δim δnj ek` = aij ek` .
ν !! . ϕ̄
R/K
commute. In particular, the image ϕ( R) ∼
= R/K.
P ROOF. By III.E.3(ii), R/K is well-defined as a ring; and by II.I.20,
there exists a unique additive group homomorphism ϕ̄ such that
ϕ̄ ◦ ν = ϕ. It is only left to check that ϕ̄ is a ring homomorphism:
ϕ̄(r̄1 r̄2 ) = ϕ̄(ν(r1 )ν(r2 )) = ϕ̄(ν(r1 r2 )) = ϕ(r1 r2 ) = ϕ(r1 ) ϕ(r2 ) =
ϕ̄(ν(r1 )) ϕ̄(ν(r2 ) = ϕ̄(r̄1 ) ϕ̄(r̄2 ).
C0 (M) −→
→ C0 (S)
f 7−→ f |S
= ac + 10bd − ( ad + bc)
= ac + bd − ad − bc
= ( a − b)(c − d)
√ √
= α( a + b 10) · α(c + d 10),
√
so yes. Clearly (1 + 10) ⊂ ker(α). Conversely,
√
a + b 10 ∈ ker(α) =⇒ a = b + 9n (n ∈ Z)
√ √
=⇒ a + b 10 = b(1 + 10) + 9n
√ √
= b + n(−1 + 10) (1 + 10)
(iv) For an example of a more general sort, consider (for any ring R)
η : Z −→ R
0 7−→ 0R
1 7−→ 1R
Z>0 3 n 7−→ 1R + · · · + 1R (n times)
−n 7−→ −(1R + · · · + 1R ).
Clearly η (n + m) = η (n) + η (m), and η (nm) = η (n)η (m) (using
distributivity). The image η (Z) is called the prime ring, and is the
smallest subring of R. Any ideal of Z is of the form (n), since these
are (as we checked before) the additive subgroups. Conclude that
η (Z) ∼
= Z if char( R) = 0, and η (Z) ∼
= Zm if char( R) = m is finite.
Z
ηR ηS
ϕ
R / S,
ϕ
with n̄ 7→ n̄. On the one hand, this implies char(S) | char( R), which
could rule out some homomorphisms. If char( R) = 0 it won’t rule
out anything, but here is something which could: if α ∈ R satisfies a
polynomial equation 0 = ∑k ak αk , ak ∈ Z (i.e. ηR (Z)), we must have
(writing β := ϕ(α)) that 0 = ∑k āk βk . One could then try to show
that S doesn’t contain such a β.
P ROOF. Clearly
ϕ : R −→ R/I1 × · · · × R/Im
r 7−→ (r + I1 , . . . , r + Im )
I1 ∩ · · · ∩ Im = I1 · · · Im .
Taking units on both sides recovers the results on units in Z/mZ from
II.E.13-II.E.14.
III.F. Fields
ab ∈ I =⇒ a ∈ I or b ∈ I.
III.F.6. T HEOREM . R/I is a domain ⇐⇒ I is prime.
P ROOF. I is not prime ⇐⇒ ∃ a, b ∈ R\ I such that ab ∈ I. Equiv-
alently, taking ā = a + I etc., ∃ ā, b̄ ∈ ( R/I )\{0} such that āb̄ = 0̄;
that is to say, R/I is not a domain.
Since fields are domains . . .
III.F.7. C OROLLARY. Maximal ideals are prime.
Turning back to the beginning of this section, note that in a sense
Q was the subfield of F generated by Z (in the characteristic zero
case). We want to generalize this.
III.F.8. P ROPOSITION . Let R be a subring of a field F. Then the inter-
section of all subfields containing R (the “subfield generated by R”) is
R × R\{0}
(III.F.9) {αβ−1 | α ∈ R, β ∈ R\{0}} ∼
= ,
≡
where (α, β) ≡ (γ, δ) ⇐⇒ αβ−1 = γδ−1 in F ⇐⇒ αδ = βγ in R.
138 III. RINGS
on R × R\{0}. This is
• reflexive: ab = ba
• symmetric: ad = bc ⇐⇒ cb = da
• transitive: ad = bc and c f = de =⇒ ad f = bc f = bde =⇒
d( a f − be) = 0 (and d 6= 0) =⇒ a f = be (since R is a domain).
Define (as a set)
R × R\{0}
F{ R } : = ,
∼
with 1F{ R} := (1, 1), 0F{ R} := (0, 1),
and the other distributive and associative laws can also be checked.
Moreover, if ( a, b) 6= 0F{ R} (i.e. a 6= 0), then
φ : R → F{ R }
r 7→ (r, 1)
φ ! -
ϕ̃
F{ R }
Principal fractional ideals are invertible since they are of the form
f R ⊂ F{ R} and we have f R · f −1 R = R2 = R. Denote by
• J ( R) := the set of fractional ideals
• J ( R)∗ := the set of invertible fractional ideals
• P J ( R) := the set of principal fractional ideals.
Under the obvious multiplication f I · f 0 I 0 = f f 0 I I 0 , J ( R)∗ forms
an abelian group with identity element R, and (normal) subgroup
P J ( R ).
βα̃
Since g is a root of
(x − βα̃
g )( x − α β̃
g ) = x2 − ( βα̃+g α β̃ ) x + αgα̃ · βgβ̃ ,
| {z } | {z }
∈Z ∈Z
βα̃
our “lemma” tells us that g ∈ R hence g | βα̃ in R. So we conclude
that
I Ĩ = ( g) ,
a very useful result called
√
Hurwitz’s Theorem (which also works for
d ≡ 1 and R = Z[ 2 ]). I say it is useful because it comes with the
1+ d
(4)
presciption for how to calculate g, as the gcd of three integers.
What this all means for fractional ideals is that
1
g Ĩ furnishes an inverse to I.
(III.G.1) 0= ∑ a I x I ∈ R [ x1 , . . . , x n ] ⇐⇒ all a I = 0.
I
evu ( R[ x ]) =: R[u]
R[u] ∼
= R[ x ]/Iu
and Iu ∩ R = {0} (and the obvious analogues for u).
(III.G.7) ∑ rI uI = 0 =⇒ all r I = 0.
I
P ROOF. We only have to prove this for F[u], u algebraic (since in-
duction then yields it for F[u1 , . . . , un ]). Let f ( x ) = ∑nk=0 ak x k ∈ F[ x ]
be a (nonzero) polynomial of minimal degree with f (u) = 0. (Note that
this degree is n.) Since S has no zero-divisors, f ( x ) is irreducible. In
particular, a0 6= 0 and (rescaling) we may assume a0 = 1. Then
(− ∑nk=1 ak uk−1 ) · u = 1 shows that u is invertible in F[u].
Now let v ∈ F[u] be arbitrary. If there exists some polynomial
g( x ) = ∑k bk x k ∈ F[ x ] with g(v) = 0 in S, then the same argument
(taking g of minimal degree, b0 = 1, etc.) produces an inverse for v
in F[u], namely − ∑k>0 bk vk−1 . So this will prove both statements of
the Proposition.
Notice that F[u] is a vector space over F of dimension n. Indeed,
−1 a k k
using f (u) = 0 ( =⇒ un = − ∑nk= 0 an u ) we can reduce the degree
of any polynomial in u (i.e. element of F[u]) to ≤ n − 1. Moreover, if
−1 n −1 0 k
∑nk= 0 ck u = ∑k=0 ck u ∈ F[ u ] then ck = ck : otherwise the difference
k 0
(III.G.10) x m + a m −1 x m −1 + · · · + a 0 , a j ∈ Z.
1 + B2 d + 2 ≡ 0 =⇒ B2 d ≡ 1 =⇒ B odd and d ≡ 1.
(4) (4) (4)
and
=⇒ 0 = f ( ak ) = ( ak − a1 ) · · · ( ak − ak−1 ) g( ak )
| {z }
6 =0
=⇒ 0 = g( ak ) (since R is a domain)
=⇒ g( x ) = ( x − ak )h( x ) (for some h ∈ R[ x ])
=⇒ ( x − a1 ) · · · ( x − ak ) | f .
So in fact, f ( x ) = H ( x ) ∏rj=1 ( x − ai ) (for some H ∈ R[ x ]) hence
n ≥ r.
What if R is not a domain? Consider, say, polynomials over Z6 :
f ( x ) = 3x has 0̄, 2̄, and 4̄ as roots. So III.G.17 fails.
Turning to the case where R is a field, we have the famous
III.G.18. T HEOREM . The multiplicative group of a finite field is cyclic.
More generally, any finite subgroup G of the multiplicative group of a field
F is cyclic.
P ROOF. Recall from II.D.15 that since G is abelian, G is cyclic
⇐⇒ exp( G ) = | G |. This was based on the fact that there exists
an element of order exp( G ) := min{e ∈ N | ge = 1 (∀ g ∈ G )}. In
general, exp( G ) ≤ | G | since g|G| = 1 for all g ∈ G.
Now every g ∈ G satisfies gexp(G) − 1 = 0. But III.G.17 =⇒
xexp(G) − 1 has at most exp( G ) roots. So | G | ≤ exp( G ).
∗ ∼ Z , and not Z×4 , Z × Z ,
III.G.19. E XAMPLE . This says Z17 = 16 2 8 2
etc. — this beats trying to find a generator!
148 III. RINGS
f ( x ) 7−→ { f (u)}u∈Fn
The image Φn,F (F[ x1 , . . . , xn ]) =: Pn (F) is called the ring of (F-valued)
polynomial functions over Fn . We write si for Φn,F ( xi ), the ith coor-
dinate function, and clearly Pn (F) = F[s1 , . . . , sn ]. Two questions
arise:
• Are all functions polynomial functions? (i.e. is Φn,F surjective?)
• Do distinct polynomials yield distinct functions? (i.e. is Φn,F in-
jective? Note that this would imply that Pn (F) ∼= F[ x1 , . . . , xn ].)
We can give a surprisingly clear answer to both questions with the
aid of the following
22This will be discussed and proved in the context of modules where it belongs.
23Note that the group operation is being written multiplicatively, because G is a
multplicative group inside a field. In “additive” terms, gmk − 1 = 0 reads mk g = 0.
24Obviously Z n isn’t a field, so that won’t cut it!
p
III.G. POLYNOMIAL RINGS 149
q −1 n
|Pn ( F )| = #{choices for g( x ) = ∑i1 ,...,in =0 a I x I } = qq
as well.
150 III. RINGS
ζ : Sn → Aut(F[ x1 , . . . , xn ]).
e1 ( x ) = ∑ xi , e2 ( x ) = ∑ xi x j , . . . , e n ( x ) = x1 . . . x n .
i i< j
s1 ( x ) = ∑ xi , s2 ( x ) = ∑ xi2, . . . , sn ( x ) = ∑ xin .
i i i
En : F[ x1 , . . . , xn ] −→ F[ x1 , . . . , xn ]Sn
xi 7−→ ei ( x )
( x1 )k1 −k2 ( x1 x2 )k2 −k3 ( x1 x2 x3 )k3 −k4 · · · ( x1 · · · xn )kn = x1k1 x2k2 · · · xnkn .
Hence f − aK e1k1 −k2 · · · enkn has lower highest monomial than f , and
continuing on in this manner we eventually reach the zero polyno-
mial.
Turning to injectivity, consider a finite sum ∑ D a D e D (with not all
a D zero). For each D ∈ Nn , write (for i = 1, . . . , n) k i = di + · · · + dn ,
and consider those (nonzero) a D e D with largest |K | := ∑i k i . The
highest monomial in each is a D x1k1 · · · xnkn , and these are all distinct
(D 6= D 0 =⇒ K 6= K 0 ). Taking the (unique) a D e D with “highest
highest” monomial, we see that this monomial occurs once, with a
nonzero coefficient. Hence ∑ D a D e D 6= 0.
152 III. RINGS
imaginary axis
√ 1
2
−ᾱ 19
2
α
√
3
2
√ √
19− 3
2
√
3
2
−2 − 23 −1 − 12 1
2 1 3
2 real axis
25The RHS contains two assertions: (α) ∈ P P , and (α) is maximal there.
156 III. RINGS
r1 · · · r m = r = s1 · · · s n
a | bc =⇒ a | b or a | c.
One way to think of all this is that for a principal ideal ( a),
+3 +3
( a) maximal
(III.I.10) ( a) maximal ( a) prime
O in P P ( R)
O
a prime +3 a irreducible.
( am ) ⊇ ( an ) =⇒ am | an =⇒ an = am r
P ROOF. Combine III.F.6 with the fact that (α) is prime iff α is.
III.I.14. E XAMPLES .
(A) All PIDs (and hence all Euclidean domains) are UFDs.
(B) F[ x, y] and Z[ x ] are UFDs but (as we know) not PIDs.
(C) There is no number ring that is a UFD but not a PID.
We will prove (B) and (C) in §§III.K-III.L; for now, here is the
28As for ideals, a pair of relatively prime elements (|S| = 2) is said to be coprime.
29We write α k β for “α is a proper divisor of β”, which is to say that α | β and β
α
is not a unit. The reason we’d have (say) δ k γ here is that, were γ
δ a unit, δ0 | γ
would become δ0 | δ, which is false.
164 III. RINGS
δ0 , δ k γ k a, b =⇒
III.J.5. R EMARKS . (i) Note that (by III.J.3) UFDs satisfy the GCDC;
and (by III.J.2(c)) for a PID we have ( a, b) = (gcd( a, b)).
III.J. GREATEST COMMON DIVISORS 165
(ii) The GCDC implies the existence of GCDs for all nonempty finite
subsets S ⊂ R. [P ROOF: given S = {s1 , . . . , sn }, inductively assume
that there exists a GCD γ0 for {s1 , . . . , sk−1 }. Then γ := gcd(γ0 , sk )
has γ | γ0 | s1 , . . . , sk−1 and γ | sk . Moreover, if γ0 | s1 , . . . , sk then
γ | s1 , . . . , sk−1 =⇒ γ0 | γ0 , which together with γ0 | sk yields γ0 | γ.]
So “gcd(S)” makes sense.30
(iii) If γ1 = gcd(S1 ), γ2 = gcd(S2 ) for two nonempty finite subsets,
the same argument gives gcd(S1 ∪ S2 ) = gcd(γ1 , γ2 ).
(iv) If γ = gcd(S) (for a finite subset S = {s1 , . . . , sn }) and r ∈ R,
then (writing r S := {rs1 , . . . , rsn }) we have rγ = gcd(r S).
This leads to a second proof that PIDs are UFDs, since the GCDC
obviously holds for PIDs (cf. III.J.2(c)).
So far we have said a lot about the theory of GCDs, and nothing
about effectively computing them (when they are not visibly obvi-
ous).
(α, β) = (q1 β + r1 , β) =
( β, r1 ) = (q2 r1 + r2 , r1 ) =
(r1 , r2 ) = ( q3 r2 + r3 , r2 ) =
(r2 , r3 ) = · · · =
( r n −1 , r n ) = ( q n +1 r n , r n ) = ( r n ).
This proves the
k = r0 q1 + r1 =⇒ r1 = k − r0 q1 ≡ k + kq0 q1 = k(1 + q0 q1 )
φ(m)
r0 = r1 q2 + r2 =⇒ r2 = r0 − r1 q2 ≡ −kq0 − k (1 + q0 q1 )q2
φ(m)
..
. = − k ( q0 + q2 + q0 q1 q2 )
1 ≡ k · (big mess).
φ(m)
We know that OK is often not a UFD, and that GCDs may not
exist. So we are not going to take them in OK . Rather, the connection
of this section to GCDs comes from Hurwitz’s theorem (cf. III.F.17).
Recall that given I = (α, β) ⊂ OK and Ĩ = (α̃, β̃), it says that
• αα̃, β β̃, and α β̃ + βα̃ belong to Z, and
• if g is their GCD in Z, then I Ĩ = ( g) = gOK .
It is the main tool in the proof of the following
I Ĩ = (N( I )) = (1) = OK ⊆ I = I OK ⊆ I Ĩ
Finally, for (iii), begin by noting that N(( p)) = p2 , and suppose
that ( p) is not irreducible. Then there exists an ideal I of norm p with
I ) ( p) (as ( p) must break into two such). Assume the following
III.J.17. P ROPOSITION .
(a) Let F be a number field, O F its ring of integers.33
(i) Every nonzero ideal I ⊂ O F contains a basis for F as a Q-vector
space, hence a subgroup ∼ = Z[ F:Q] .
(ii) Assuming that O F ∼ = Z[ F:Q] (and F ∼
= Q[ F:Q] ),34 we have that
I∼= Z[ F:Q] , with basis spanning F/Q.
√
(b) Let K = Q[ d], and I ⊂ OK be a nonzero ideal. Then as an additive
abelian group, I = hγ, δi, for some γ, δ ∈ OK ; and, moreover, I = (γ, δ).
ak k
P ROOF OF III.K.2. Write f = ∑nk=0 bk x , ak ∈ R, bk ∈ R\{0}. Let
β
β := ∏k bk , so that β f ∈ R[ x ], and γ := c( β f ). Then g := γ f ∈ R[ x ]
is primitive and f = γβ g. If α0 g0 = f = αg with g, g0 primitive, then
III.K. GAUSS’S LEMMA AND POLYNOMIALS OVER UFDS 173
∃b ∈ R such that
αb, α0 b ∈ R =⇒ (αb) g = (α0 b) g0
| {z } | {z }
content αb content α0 b
0
=⇒ αb ∼ α b
=⇒ uαb = α0 b (u ∈ R∗ )
=⇒ αbg = uαbg0
=⇒ g = ug0
=⇒ g ∼ g0 ,
which completes the proof.
f = c( f ) g ( g ∈ R[ x ] primitive)
= c ( f ) g1 · · · g k ( g j ∈ F [ x ] irreducibles)
= c( f )( β 1 f 1 ) · · · ( β k f k ) ( β j ∈ F ∗ , f j ∈ R[ x ] primitive)
( f 1 · · · f k primitive by III.K.6,
= c( f ) β f 1 · · · f k
hence β ∈ R∗ by (III.K.5))
= α1 · · · α ` f 1 · · · f k (αi ∈ R irreducible)
where the last step is possible because R is a UFD. Clearly the αi are
irreducible in R[ x ], and by III.K.7, so are the f j .
Now we must show the essential uniqueness of this factoriza-
tion. If f = α10 · · · α0`0 f 10 · · · f k00 (deg(αi0 ) = 0, deg( f j0 ) > 0) is an-
other factorization into irreducibles in R[ x ], then III.K.7 =⇒ the
f j0 are irreducible in F [ x ] and primitive, whence (by III.K.6) f 10 · · · f k00
III.K. GAUSS’S LEMMA AND POLYNOMIALS OVER UFDS 175
Q[ θ ] ∼
= Q[ x ]/( x3 − 3x − 1) [resp. ∼
= Q[ x ] / ( x n − p )]
is a field, using the fact that Q[ x ] is a PID (cf. III.H.8). Since Z[ x ] is a
UFD, the corresponding quotients of Z[ x ] are domains by III.I.13.
III.L. ALGEBRAIC NUMBER RINGS 177
m α ( x ) = x n + a n −1 x n −1 + · · · + a 0 , ai ∈ Z.
Then mα (α) = 0 =⇒
35The vectors here belong to the vector space Q[α]r over the field Q[α], and the re-
sult we are using from linear algebra works over any field: given M~v = λ~v, clearly
~v is in the kernel of left-multiplication by λIr − M, which means the columns of
the latter are dependent and hence that its determinant is zero.
III.L. ALGEBRAIC NUMBER RINGS 179
Integral ideals. Now write, given a number field K, I(K ) for its
monoid of integral ideals (i.e. nonzero ideals I ⊂ OK ). Slightly chang-
ing notation,36 we write J (K ) for the fractional ideals a = λI (λ ∈ K ∗ ,
I ∈ I(K )). Recall that a ∈ J (K ) is invertible iff a · b = OK for some
b ∈ J ( K ).
We have seen that
(a) factorization into irreducibles is not necessarily unique, and
(b) irreducibles need not be prime
in a non-UFD OK . As we shall now see, replacing {OK∗ , OK , K } by
{(1), I(K ), J (K )} makes these problems disappear.
The proof of the next result requires Galois theory if K is a gen-
eral number field (as it involves introducing discriminants and ideal-
norms in general), so we shall assume it. However, I will explain
how it follows from what we already know when K is quadratic.
36This is instead of writing J (O ).
K
180 III. RINGS
III.L.6. R EMARK . (a) From (i), it follows that any product of max-
imal ideals is invertible in J (K ).
(b) In (iii), I is of rank [K:Q] as an abelian group, at least according
to unproved assertions in III.J.17.
(c) For any I ∈ I(K ), the quotient OK /I is a finite abelian group.
One can define an ideal norm by N( I ) := |OK /I |, which agrees with
the definition in the quadratic case.
becomes
I⊃J ⇐⇒ I | J,
that is, “to divide is to contain”. This is a different result than III.D.16,
but we will call it Caesar’s lemma as well.
Note in addition that for I ⊃ J, III.L.7 now gives J 0 := I −1 J ∈
I(K ), so that J = I J 0 . By multiplicativity of N, we get N( J ) =
N( I )N( J 0 ) hence N( I ) | N( J ).
(III.L.12) ℘ 3 ab =⇒ ℘ 3 a or ℘ 3 b.
(III.L.13) ℘ ⊃ IJ =⇒ ℘ ⊃ I or ℘ ⊃ J
=⇒ ℘ = I J = ℘Q J =⇒ N( ℘) = N( ℘)N(Q )N( J ) in N. So
N(Q ) = 1 = N( J ), whence Q = OK = J by III.L.7(iii). So ℘ is
irreducible.
C `(K ) := J (K )/P J (K )
hK := |C `(K )|
38As with J (K ), this is a slight change in notation from III.F.16.
184 III. RINGS
P ROOF OF ( VI ). e = [(α)] = [ I J ] = [ I ][ J ].
(i) (iii)
This is all very useful for solving (or showing insoluble) Dio-
phantine equations like X 2 = Y 3 − 14, as you will see in Problem
Set 10.
39This is always finite, a fact which we will not be able to prove (but see III.L.27
for the idea).
III.L. ALGEBRAIC NUMBER RINGS 185
√
P ROOF OF III.L.20 FOR K = Q[ d] . Suppose that OK is a UFD.
To show that it is a PID (every ideal is principal), it will suffice to
prove that its prime ideals are principal, since (by III.L.14) every
ideal is a product of maximal ideals, and maximal ideals are prime.
So let ℘ ∈ I(K ) be a prime ideal, and write N( ℘) = ∏i σi for the
(unique) decomposition of its norm into irreducibles in OK . Since
OK is a UFD, these irreducibles σi are prime elements of OK , so that
the (σi ) are prime ideals, and thus irreducible in I(K ) by III.L.14.
By Hurwitz, ℘ ℘ ˜ = (N( ℘)) =⇒ ℘ | (N( ℘)) = ∏ (σi ) =⇒ ℘ |
i
(σi ) for some i (since ℘ ∈ I(K ) is a prime element). By irreducibility
of (σi ), we have ℘ = (σi ), so that ℘ is principal as desired.
n
P ROOF. Let N( ℘) = ∏i pi i be a prime factorization in N. As ℘ is
prime and ℘ | (N( ℘)), we must have ℘ | ( pi ) for some pi =: p. So
N( ℘) | N(( p)) = p2 (for K quadratic) =⇒ N( ℘) = p or p2 .
If N( ℘) = p, then (by Hurwitz) ℘ ℘ ˜ = ( p), whence ℘ = ℘p or ℘˜p
since I(K ) is a UFM. If N( ℘) = p2 , then (by III.L.5(ii)) ℘ = ( p).
186 III. RINGS
with 0 ≤ ai ≤ `i .
Of course, Fermat’s Last Theorem states that for any exponent n > 2,
the only solutions to x n + yn = zn are the “trivial” ones, with x or
y = 0. The cases n = 4 (Fermat) and 3 (Euler) were proved by
Fermat’s method of descent; and if one has the theorem for some n,
one has it for all exponents divisible by n (why?).
As you may know, the proof was ultimately completed by Wiles
in 1995, building on decades of work by many people on modularity
and Galois representations. What I want to discuss here is Kum-
mer’s big advance in the mid-19th Century, which led to the devel-
opment of ideals.
41That is, 3 + 2 √−29 does not belong to Z[√−29].
5 5√
42Say K = Q[ d]. Then n = 2; and r is 1 for d < 0 and 0 for d > 0. The
2
discriminant is 4d unless d ≡ 1, in which case it is d.
(4)
188 III. RINGS
( p ) ⊂ (1 − ζ a ) for each a = 1, . . . , p − 1.
ζ −1 ( x + yζ ) − ζ −1 ( x + yζ a+1 ) = y(1 − ζ a )
p | ( x − y) + ζ (y − x ) =⇒ p | x−y =⇒ x ≡ y.
( p)
2x p ≡ x p + y p = z p ≡ − x p mod ( p)
=⇒ p | 3x p , a contradiction.
p - hK (= hQ[ζ p ] ).
The first irregular prime is 37. The method described here essen-
tially settles Fermat for any smaller exponent (prime or not). Note
how deeply we dug into the ideal structure of Z[ζ ] to deal with an
equation ostensibly in rational integers!
IV. Modules
Modules over a ring arose from algebraic number theory and rep-
resentation theory. The definition we use now, a simultaneous gen-
eralization of vector spaces over a field and the action of a group on
a set, is another contribution of E. Noether. The main immediate ap-
plications will be to the structure theory of finitely generated abelian
groups and to the canonical forms of a linear transformation on a
vector space.
(c) Any ring R is a (left and right) module over itself. Any left [resp.
right] ideal I ⊂ R is a left [resp. right] R-module.
(c’) Given any subring R0 ⊂ R, R is a (left and right) R0 -module, and
any R-module M has the structure of an R0 -module.
(c”) Given a ring homomorphism θ : S → R, an R-module M has the
structure of an S-module via sm := θ (s)m.
(d) Given a ring R, the map R × Rn → Rn sending (r, (r1 , . . . , rn )) 7→
(rr1 , . . . , rrn ) makes Rn into a (left) R-module. This is the prototype
for free R-modules. (“Direct summands” of Rn will be the prototype
for projective R-modules, and “quotients” of Rn for finitely gener-
ated R-modules.)
(e) For those who are familiar with manifolds, a finitely generated
projective C ∞ (M)-module is the same thing as a smooth vector bun-
dle over M.
(f) Rn is a left Mn (R)-module.
(g) Let G be a finite group. A representation of G on an F-vector
space V is a map
ρ
G×V → V
( g, v) 7→ ρ( g)v (or “g.v”)
0 0
g.(v + v ) = g.v + g.v
satisfying g.( f v) = f ( g.v) ( f ∈ F)
( gg0 ).v = g.( g0 .v), 1G .v = v.
T: V → V
F[λ] −→ EndF (V )
θ
P(λ) 7−→ P( T ),
where λ is an indeterminate.
Now, we can add and compose endomorphisms, making EndF (V )
into a ring and V into an EndF (V )-module. It also makes θ a ring
homomorphism, with image
im(θ ) =: F[ T ].
` : R → End( M)
r 7 → `r
(since rs 7→ `rs = `r `s
and r + s 7→ `r+s = `r + `s ).
If we try the same thing for right multiplication, we run into the
problem
rrs (m) = m(rs) = (mr )s = (rr (m))s = rs (rr (m)) = (rs rr )(m).
L : R −→ End( M)
r 7−→ {m 7→ rm}
R : Rop −→ End( M )
r 7−→ {m 7→ mr }
From this point of view, the two notions are “the same” for a
commutative ring R because R = Rop .
198 IV. MODULES
F[ G ] −→ EndF (V )
G → AutF (V ).
RhSi := {∑finite
s∈S rs s | rs ∈ R },
∑α Nα := Rh{ Nα }i = {∑finite
α nα | nα ∈ Nα }.
3See IV.A.3 for finite generation.
200 IV. MODULES
“Finite” means that, while the index set may be infinite, only finitely
many terms in each sum can be nonzero.
Now check the properties in IV.A.1 for M/N, e.g. (rs)m̄ = (rs)m =
r (sm) = r (sm) = r (sm̄).
Just as in the case of groups and rings, kernels and images define
subobjects.
def.
(ii) A (f.g.) R-module M is free ⇐⇒ M ∼ = Rn (as R-module) for
some n ∈ N. (If R is noncommutative, n need not be unique.)
A submodule of a “free” R-module need not be free unless R is a
PID: for instance, Z6 has 3Z6 as sub-Z6 -module.
(iii) Consider the cyclic (sub)module
Rx := {rx | r ∈ R} ⊂ M
µ x : R Rx
r 7→ rx ,
ann( x ) := ker(µ x ) ⊆ R,
ν $$ , η̄
M/ ker(η )
In particular, im(η ) ∼
= M/ ker(η ) as R-modules.
202 IV. MODULES
∑in=1 ri mi = 0 =⇒ ri = 0 (∀i ).
is well-defined.
P ROOF. This says that if the aij and x j are as above, then
(IV.B.18) ∑i f i (∑ j aij x j ) = η ( x ).
To check this, write LHS(IV.B.18) = ∑ j (∑i f i aij ) x j = ∑ j η (e j ) x j =
η ( ∑ j e j x j ) = η ( x ).
Writing [·] as a shorthand for f [·]e when the bases are understood,
we have
4If R is commutative, these are just left R-modules by rm := mr. The main point
here is that I want the transpose of what [Jacobson] gets.
5In contrast, [Jacobson] uses row vectors.
IV.B. SUBMODULES AND HOMOMORPHISMS 205
= ∑m n
j=1 ( ∑i =1 f i ( f [ η ]e )ij )(e [ ζ ]δ ) jk
= ∑in=1 f i {∑m
j=1 ( f [ η ]e )ij (e [ ζ ]δ ) jk }
= ∑in=1 f i ( f [η ]e · e [ζ ]δ )ik .
Now use that { f i } is a base.
Applying this in the case M = N = L ( =⇒ m = n = `), we
have proved
IV.B.21. T HEOREM . Given a free right R-module M, we have an iso-
∼
=
morphism of rings e [·] := e [·]e : EndR ( M) −→ Mm ( R).
IV.B.22. R EMARK . A natural question is whether HomR ( M, N )
has the structure of an R-module. Let’s consider this in the left-R-
module case. If R is non-commutative, the answer is in general no:
given f ∈ HomR ( M, N ), we can try to define r f by6
7Note that RHS(IV.B.23) will no longer be given by f (rm), since f is not assumed to
be a left R-module homomorphism (and I haven’t even assumed a left R-module
structure on M.
8For simplicity, all modules are henceforth left modules; though the results also
hold for right modules.
9Also written ⊕n M or just ⊕ M or ⊕ M .
i =1 i i i i
IV.B. SUBMODULES AND HOMOMORPHISMS 207
η : = ∑ i η i : ⊕ Mi → M
x j = ∑i6= j (− xi ) ∈ M j ∩ ∑i6= j Mi =⇒ xj = 0
N = L ⊕ ( L00 ∩ N ) =: L ⊕ L0 .
Suppose L0 is not simple: then it has a proper nonzero submodule
L10 ; applying semisimplicity of M and “intersecting” as above, we
get L0 = L10 ⊕ L20 . Hence N = L ⊕ L10 ⊕ L20 . If n ∈ L ⊕ Li0 (i = 1, 2)
then we can write n = `1 + `10 = `2 + `20 (with `1 , `2 ∈ L). But then
`1 − `2 = `20 − `10 ∈ L ∩ L0 = {0} =⇒ `10 = `20 ∈ L10 ∩ L20 = {0} =⇒
n ∈ L, a contradiction. So n ∈ / L ⊕ L20 (swapping 1 and 2 if needed),
which violates the maximality of L, another contradiction! Conclude
that L0 is simple.
So we have shown that every submodule N of M contains a simple
direct summand.
Next, let { Mi | i ∈ I} be a set of simple submodules of M, maxi-
mal (Zorn again) with respect to the property that
MI : = ∑ Mi = ⊕i∈I Mi .
i ∈I
10One has to be a bit careful here: the classic example from linear algebra is that
R2 is the direct sum of the two coordinate axes, a decomposition that you certainly
can’t “intersect” with (say) the diagonal. The difference here is that N contains one
of the summands (namely, L). So given m = ` + `00 ∈ L ⊕ L00 = M, if it happens
that m ∈ N then `00 = m − ` ∈ N (since m, ` ∈ N). So `00 ∈ L00 ∩ N as desired.
210 IV. MODULES
Rm → Rn
θ
(IV.C.3)
e0j 7→ θ (e0j ) =: a j ( j = 1, . . . , m)
whose image is K.
(n)
• Tij ( a) := 1n + aeij , where a ∈ R, has inverse Tij (− a):
(n) (n)
• Pij := 1n + eij + e ji − eii − e jj = Pji is its own inverse.
(n) (n)
• Di (u) := 1n + (u − 1)eii , where u ∈ R∗ , has inverse Di (u−1 ).
In practice, you may not need to keep track of how the bases
change, but just to find some PAQ which is in a nice form (the nor-
mal form below). At the risk of beating elementary matrices into the
ground:
(IV.C.10) Zr
Zd1 × · · · × Zdk × |{z}
| {z }
Gtor G/Gtor
where di ≥ 2 and d1 | d2 | · · · | dk .
• G finitely generated =⇒ G ∼
= Zn /K with relations matrix A.
• Lemma IV.C.8 =⇒ EOs convert A to normal form.
• Example IV.C.7 =⇒ the resulting matrix is of the form PAQ with
P ∈ GLn (Z) and Q ∈ GLm (Z).
• Prop. IV.C.6 =⇒ PAQ is a relations matrix for G.
Conclude that
!
G ∼
= Zn /( PAQ)(Zm ) = Zn /
D 0
Zm
0 0
Z h X1 , . . . , X n i
= = Zd1 × · · · × Zd k × Zn − k ,
h d 1 X1 , . . . , d k X k i
IV.C. MODULES OVER A PID 215
−10 −5
Applying EOs, we reduce to normal form:
add (−11)×(col. 2) 0 1
subtract (row 2)
0 1
subtract (row 1)
0 1
/ 45 −6 / 45 −6 / 45 −6
to (col. 1) 45 −5 from (row 3) 0 1 from (row 3) 0 0
1 0
add 6×(row 1)
0 1
swap
/ 45 0 /
0 45,
to (row 2) 0 0 cols. 1 and 2
0 0
A0
0
Input A
(n × m integer
matrix)
(1)
Output
! “Reduce Ct mod ast ”:
D 0 replace each Ri by
0 0 A0
=0
a
Ri − b astit c Rs .
d
re
c lea
t
(0) no
Ct
Take ast :=
nonzero entry of A0
Ct cleared
with smallest absolute
value. Multiply Rs by
0 )
A sign of ast (to make
ce
pla C
it positive).
(re d, R t c
are s no leare
l e tc d
(4) c 0 lea ,
Rs red
, |A (2)
Ct , Rs cleared, ast - A0
Ct a st
Move ast to
upper left of A0 : swap
“Reduce Rs mod ast ”:
Rs with 1st row of A0 ,
replace each Cj by
then swap Ct with 1st a
Cj − b astsj cCt .
column of A0 .
,
ed
ar
(3)
e
0
cl
-A
s
R
t
as
Rs cleared, ast | A0
218 IV. MODULES
hence
k
e1` = ∏ gcd(e1 , d j ) ≤ e1k ,
j =1
from which we conclude that ` ≤ k. A symmetric argument shows
` ≥ k, so ` = k; in particular, the above inequality is an equality
so that gcd(e1 , d j ) = e1 (∀ j) =⇒ e1 | d j (∀ j). Again, a symmetric
argument (taking Ad1 on both sides of (a)) shows d1 | e j (∀ j). But
then d1 | e1 and e1 | d1 =⇒ e1 = d1 .
Repeating the argument starting with
A e2 ( Z e1 × · · · × Z e k ) = A e2 ( Z d 1 × · · · × Z d k )
gives
k k
gcd(e2 , e1 ) · ∏ gcd(e2 , e j ) = gcd(e2 , d1 ) · ∏ gcd(e2 , d j )
j =2 (=e1 ) j=2
13If you prefer, you can argue using II.K.4 that r ≤ s and s ≤ r.
IV.C. MODULES OVER A PID 219
k k
=⇒ e2k−1 = ∏ gcd(e2 , e j ) = ∏ gcd(e2 , d j ) ≤ e2k−1 .
j =2 j =2
Clearly the inequality is an equality, and so gcd(e2 , d j ) = e2 hence
e2 | d j for each j. On the other hand, taking Ad2 of both sides gives
d2 | e j . So d2 | e2 and e2 | d2 =⇒ d2 = e2 .
Continue in this manner until you get all d j = e j .
Using the Chinese Remainder Theorem to decompose the Zd j
factors in (IV.C.10) yields the
IV.C.13. C OROLLARY (p-primary version of FTFGAG). Any finitely
generated abelian group G may be expressed (uniquely up to rearrangement
of factors) in the form
Z p r 1 × · · · × Z p r k × Zr ,
1 k
n = n1 + · · · + n k ( n1 ≤ · · · ≤ n k ) ←→ Z pn1 × · · · × Z pnk .
Together with IV.C.13, this allows you to find all abelian groups of a
given order: e.g., for order 360 = 23 32 5, we have
G∼
= {Z23 or (Z21 × Z22 ) or (Z21 × Z21 × Z21 )}
× {Z32 or (Z31 × Z31 )} × Z5 .
IV.C.15. E XAMPLE . Let’s see how to transform a more compli-
cated matrix than the one in IV.C.12 into normal form, by applying
the EO Normalization Algorithm. (You won’t need to follow the
algorithm this precisely in working problems. The point of going
through this example is to know what to do if you get stuck!)
4 −10 −2 20 30
0 28 −2 −60 90
A= .
3 −3 −2 6 −9
7 −7 −4 14 −21
220 IV. MODULES
0 6 0 −12 −42
0 0 6 −12 −42
IV.C. MODULES OVER A PID 221
0 0 6 −12 −42
0 0 2 0 0
Ct Cj
· · · · ·
·
Ri ait · aij ·
· · · · ·
Rs· ast · asj ·
· · · · ·
(1’) Subtract (for each i) qRs from Ri , where ait = qast + r (replaces
ait by r, with δ(r ) < δ( ast )).
(2’) Subtract (for each j) q̃Ct from Cj , where asj = q̃ast + r̃ (replaces
asj by r̃, with δ(r̃ ) < δ( ast )).
(3’) (a) Add Rs (cleared) to Ri ; then (b) subtract q0 Ct from Cj where
aij = q0 ast + r 0 (replaces aij by r 0 , with δ(r 0 ) < δ( ast )).
(4’) Swap ast to the upper left of A0 .
14i.e. A 7→ PAQ, P and Q invertible over R. EOs will not in general be enough,
but suffice for Euclidean domains.
224 IV. MODULES
Again one either proceeds all the way around the outer semicircle
((10 ) → (20 ) → (40 )), or reduces δ( ast ), so the process must termi-
nate.
Step 1B : Reduction to normal form in general.
Let ` : R\{0} → N be the length function. (Since R is a PID,
R is a UFD, and this is well-defined.) For (0”), we take ast to be the
nonzero entry of A0 with smallest ` (e.g. a unit, if there is one). (4”) is
the same as (4’). We need replacements for (1’), (2’), and (3’)(b) when
ast - ait (resp. asj , aij ), since the Euclidean algorithm isn’t available.
In fact, EOs won’t suffice. Though (1”) [resp. (2”) and (3”)(b)]
will still be given by row [resp. column] operations, or (equiva-
lently) left- [resp. right-]multiplication by invertible matrices, the
operations/matrices involved are of a slightly more general nature.
For (2”), here is what we can do.15 Set a := ast , b := asj , and let
x, y ∈ R be such that
xa + yb = d := gcd( a, b),
z := db , w := − da . Right-multiplication by
Ct Cj
1
..
.
1
Rt x z
1
..
.
1
Rj y w
1
..
.
1
15The analogues for (1”) and (3”)(b) are essentially the same and left to you.
IV.C. MODULES OVER A PID 225
which may “undo” our clearing of Cj .16 But this is not a problem,
as it creates a new entry (“d” in the (s, t) place) with length `(d) <
`( ast ) (where `( ast ) was the previous shortest length). So as before,
the minimal length is reduced each time we return to (0”) without
passing through (4”) and reducing the size of A0 .
Step 2 : Uniqueness of the invariant factors.
Define ∆i ( A) := gcd{i × i minors of A} and
∆i ( A) ∼ ∆i ( PAQ).
Now suppose
d1 d10
PAQ =
..
. 0 , 0 0
P AQ =
..
. 0 .
dk d0k
0 0
0 0
On the one hand, direct computation implies
(
∆i ( PAQ) = d1 · · · di
(∀i ).
∆i ( P0 AQ0 ) = d10 · · · di0
On the other, ∆i ( PAQ) ∼ ∆i ( A) ∼ ∆i ( P0 AQ0 ) (∀i). We conclude that
di ∼ di0 (∀i) and k = r( A) = `.
16This was already a feature of the original Step (3), though not of Step (2).
226 IV. MODULES
and
1 1 1
We reduce A to normal form with row and column operations:
λ − 1 −1 −1 −1 λ − 1 −1
A = λ13 − B = −1 λ − 1 −1 7→ λ − 1 −1 −1
−1 −1 λ − 1 −1 −1 λ − 1
1 0 0 1 1
7→ 0 λ2 − 2λ −λ 7→ −λ λ2 − 2λ 7→ λ .
0 −λ λ 0 λ2 − 3λ λ2 − 3λ
| {z }
=nf( A)
Conclude that the invariant factors of A are d1 ( A) = 1, d2 ( A) = λ,
and d3 ( A) = λ2 − 3λ; the last of these is indeed the minimal polyno-
mial of B:
3 3 3 1 1 1
B2 − 3B = 3 3 3 − 3 1 1 1 = 0.
3 3 3 1 1 1
/ R∗ and δ1 | · · · |δ` .
where the δi ∈
More precisely, M is an internal direct sum of cyclic modules:
(
M = Rz1 ⊕ · · · ⊕ Rzs (zi ∈ M)
(IV.C.20)
where ann(z1 ) ⊃ · · · ⊃ ann(zs ) ;
and the annihlator ideals (hence also the number s) are uniquely deter-
mined.
M∼
= Rn /K ∼
= Rn /θ ( Rm ) ∼
= Rn /A· Rm ∼
= Rn /PAQ· Rm ,
n m
R , R. By IV.C.16
with the last step given by change of bases for
D 0
we may arrange to have PAQ = nf( A) = 0 0
(with D =
diag(d1 , . . . , dk )) hence
M∼ n D 0
=R / 0 0
· Rm .
ρ : Rn M
ei 7 → ρ ( ei ) = : x i
M = Rz1 ⊕ · · · ⊕ Rzs ∼
= R/(d1 ) ⊕ · · · ⊕ R/(d` ) ⊕ |R ⊕ ·{z
· · ⊕ R.}
t copies
A p1 ∩ ∑i`=2 A pi ( M) = {0}.
k
Given x in the LHS, we have p1k1 x = 0 = p2k2 · · · p` ` x for some k i ∈ N.
k
But as the primes are distinct, gcd( p1k1 , p2k2 · · · p` ` ) = 1. So there exist
k
m, n ∈ R such that x = 1x = (mp1k1 + np2k2 · · · p` ` ) x = 0.
IV.C.25. T HEOREM . Assume M is a f.g. torsion module over a PID R.
Then M = ⊕ p∈ R prime A p ( M ) ∼
e
= ⊕i R/( pi i ), where pi are not necessar-
ily distinct primes in R and ei ∈ Z>0 . Both direct sums are finite, which
is to say that A p ( M ) is nonzero for only finitely many18 primes.
P ROOF. We know M = ⊕kj=1 R/(d j ), d j ∈ R\{0} and d1 | · · · |dk .
e j`
Moreover, d j = ∏m `=1 p` (∀ j) for some list of distinct primes { p` }
(and {e j` ∈ N}). So we will almost be through if we can check that
e j`
R/(d j ) = ⊕m
`=1 R/ ( p` ).
(IV.C.26) R/( f g) ∼
= R/( f ) ⊕ R/( g) if ( f , g) = R.
To see this, let x be a generator of the LHS and rx an arbitrary el-
ement. Then ( f , g) = R =⇒ ∃ ri ∈ R with r1 f + r2 g = r =⇒
rx = r1 f x + r2 gx =⇒
R/( f g) = Rx = R f x + Rgx.
17This means non-associate: they don’t generate the same ideal.
18Again, we are thinking of primes “up to units”; or equivalently, in terms of the
corresponding prime ideals.
230 IV. MODULES
and moreover the proofs of (IV.C.26) and (IV.C.19) show that the di-
rect sum is internal. Therefore, we are reduced to
(IV.C.27)
e j` e j`
If M = ⊕ j,` Rx j` = ⊕ j,` R/( p` ), then A p` ( M) = ⊕ j R/( p` ) (∀`).
Clearly one has “⊇” on the right. To see the reverse inclusion “⊆”,
e j`
we need A p` ( M) ∩ ⊕ j,`6=`0 R/( p` ) = {0}. But the “⊕ j,`6=`0 ” here
0
belongs to ∑`6=`0 A p` ( M), so we are done by the proof of Lemma
IV.C.24.
Next, decompose all the Rzi resp. Rw j into sums of primary cyclic
submodules, viz.
e j` e0
M = ⊕` ⊕sj=1 R/( p` ) = ⊕` ⊕rk=1 R/( p`k` ).
If these factors are the same, there is only one way to put them back
together to get d1 | · · · |ds and d10 | · · · |dr0 , and this will prove they are
the same set of divisors. Since
e e0
A p` ( M) = ⊕sj=1 R/( p`j` ) = ⊕rk=1 R/( p`k` ),
we may assume that M = A p ( M) for a single prime p ∈ R.
Considering the filtration19 by R-submodules
M ⊃ pM ⊃ p2 M ⊃ · · · ,
pn M
each =: M(n) is an R/( p)-module (since pM(n) = 0). Since
p n +1 M
( p) is prime and R is a PID, ( p) is in fact maximal, and R/( p) a field,
making M(n) a vector space. Writing
0
M = A p ( M) = ⊕sj=1 R/( pe j ) = ⊕rk=1 R/( pek ) ,
we have
(
s s
( pn )/( pe j ) 0, if e j ≤ n
M (n) =
M M
ej = ( pn )
j =1
( p n +1 ) / ( p ) j =1 ( p n +1 )
, otherwise
0 = F ( λ ) zi = F ( T ) zi =⇒ T ( T m−1 zi ) = T m zi = ( T m − F ( T ))zi
= − F0 zi − F1 Tzi − · · · − Fm−1 T m−1 zi .
234 IV. MODULES
20This is always true if F is algebraically closed, which is to say that every polyno-
mial over F has a root in F. For instance, C is algebraically closed.
IV.D. APPLICATIONS TO LINEAR ALGEBRA 235
dim( Eα ( T )) = ∑ ei .
i : αi = α
1 1 1
236 IV. MODULES
V = F[ λ ] z1 ⊕ F[ λ ] z2 ∼
= F[λ]/(λ) ⊕ F[λ]/(λ2 − 3λ),
and with respect to the basis z̃ = {z1 , z2 , Tz2 } we get the rational
canonical form
0 0 0
0 0 0
0 1 3
since T (z1 ) = 0, T (z2 ) = Tz2 , and T ( Tz2 ) = T 2 z2 = 3Tz2 (from
T 2 − 3T = 0 on F[λ]z2 ).
For the Jordan form, we factor δ2 (λ) = λ2 − 3λ = λ(λ − 3) to
further decompose V into primary cyclic modules:
0 0 3
and dim( E0 ( B)) = 2, dim( E3 ( B)) = 1. After all, if a matrix can be
diagonalized, the Jordan form is diagonal. This happens precisely
when the δi (taken individually) have no repeated linear factors.
One thing you may wonder is how to find the basis (or change-
of-basis matrix) which puts B in rational or Jordan canonical form.
We have (writing θ : K ,→ F[λ]n for the inclusion)
e0 [ θ ] f 0 = nf(λ13 − B) = Q(λ13 − B) P
= e0 [idF[λ]n ]e · e [ θ ] f · f [idK ] f 0 ,
so that Q = e0 [id]e =⇒ columns of Q−1 = e [id]e0 yield the e0 -basis
(written in the e-basis). One builds the basis z̃ for the rational (or
Jordan) form from the xi0 := η (ei0 ) for i = k0 + 1, . . . , n. Noting that
IV.D. APPLICATIONS TO LINEAR ALGEBRA 237
(So far, the bases for kernels are easily computed by taking rref of
B − 21, B − 1, ( B − 1)2 , and ( B − 1)3 .) It is the last basis vector for
E1 ( B) that generates it as a Q[λ]-module, and we choose its cyclic
images as our remaining basis vectors for V:
0 −1 0
v2 := 00 7→ v3 := ( B − 1)v2 = 1
1 7→ v4 := ( B − 1) v2 = −01 .
2
1 −1 1
P ROOF. For the second part, just note that F (λ)V = {0} ⇐⇒
F ( T )v = 0 (∀v ∈ V) ⇐⇒ F ( T ) xi = 0 (∀i) ⇐⇒ F ( B) = 0.
IV.E. Endomorphisms
µr := {multiplication by r } 7−→ r̄
EndR ( M) → Mn ( R/(d))
η 7→ (r̄ij )
EndR ( M) ∼
= ⊕i EndR ( R/( pi )) ∼
= ⊕i R/( pi ).
Alternatively, one can use the Chinese Remainder Theorem (see the
proof of IV.C.25) to write M ∼
= R/(∏ pi ), apply (a), and use the CRT
again on the RHS.
(d) Finally, if M ∼
= ⊕i ( R/( pi ))⊕ni , then combining Schur’s Lemma
with (b) yields
EndR ( M) ∼ = ⊕i Mni ( R/( pi )),
which is again an isomorphism as rings and as R-modules.
M = Rz1 ⊕ · · · ⊕ Rzs ∼
= Rz1 ⊕ · · · ⊕ Rz` ⊕ Rt ,
| {z }
tor( M)
S := EndR ( M)
For x̃ ∈ Rs , we have
!
δ1
x̃ ∈ K ⇐⇒ x̃ = ∑i`=1 di ri ei ⇐⇒ e [ x̃ ] ∈ ... Rs =: DRs
(for some ri ∈ R) δs
η̃ (K ) ⊆ K ⇐⇒ η̃ ( x̃ ) ∈ K (∀ x̃ ∈ K )
[apply e [ ] ] ⇐⇒ NDv ∈ DRs (∀v ∈ Rs )
[apply to v = e1 , . . . , es ] ⇐⇒ ND ⊂ DMs ( R)
⇐⇒ N ∈ MS ,
def.
and
η̃ ( Rs ) ⊆ K ⇐⇒ Nv ∈ DRs (∀v ∈ Rs )
⇐⇒ N ∈ DMs ( R) =: JS .
def.
⇐⇒ n21 ∈ ( δδ21 );
so we can write nij := nij0 δδi in (II) as above, with nij0 ∈ R/(δj ). In the
j
event that M is torsion, ` = s and we don’t have regions (III) and
(IV).
An immediate consequence is
and
22Recall that e is the matrix with (i, j)th entry 1 and all other entries 0.
ij
IV.E. ENDOMORPHISMS 245
S = EndF[λ] (V ) = {η ∈ EndF (V ) | ηT = Tη }
(S ∼
=) EndF[λ] (Fn ) = { Z ∈ Mn (F) | ZB = BZ },
the ring of matrices commuting with B.
V = Q[λ]/(δ(λ)), δ(λ) = λ4 + λ3 + λ2 + λ + 1.
1 1 1
is similar to its rational and Jordan forms
0 0
B0 = 0 0 and B00 = 0 .
1 3 3
(V.A.2) f · ( a1 a2 ) = ( f · a1 ) a2 = a1 ( f · a2 ).
A
α / B
=
η !! . ᾱ
A/I
V.A. ALGEBRAS OVER A FIELD 251
commutes.
There is also an algebro-theoretic version of Cayley’s theorem:
` : A −→ EndF ( A)
a 7−→ ` a := left-multiplication by a.
Since A is an algebra,
` a ( f α) = a( f α) = f ( aα) = f ` a (α)
=⇒ ` a ∈ EndF ( A). Moreover, we know that ` is an injective ring
homomorphism.1 Finally, ` is a homomorphism of F-vector spaces,
since by (V.A.2) we have ` f a (α) = ( f a)(α) = f ( aα) = f ` a (α) and
thus ` f a = f ` a .
(V.A.9) v2 = 0 (∀v ∈ V ).
Let {u1 , . . . , un } ⊂ V be a basis. Then (V.A.9) gives
(V.A.10) ui u j = − u j ui .
1Why? Recall ` = 0 =⇒ 0 = ` 1 = a · 1 = a.
a a
252 V. REMARKS ON ASSOCIATIVE ALGEBRAS
V•
V.A.12. D EFINITION . FV is the F-algebra with
• (F-vector space) basis consisting of monomials2
u I : = u i1 ∧ · · · ∧ u i k ( i1 < · · · < i k )
where I = {i1 , . . . , ik } ranges over subsets of {1, . . . , n},
• product
(
0, if I ∩ J 6= ∅
uI ∧ uJ =
sgn(σIJ )uI∪J , if I ∩ J = ∅
where σIJ shuffles I and J together, and
• identity u∅ = 1.
V•
We have dimF ( F V) = ∑nk=0 (nk) = (1 + 1)n = 2n , and
V• L Vk
(V.A.13) FV = k F V,
Vk
where FV is the subspace spanned by monomials of degree k.
What about a vector space where you can multiply and divide vec-
tors?
V.B.1. D EFINITION . A division algebra over a field F is an F-
algebra A whose underlying ring is a division ring.
This rules out most of the examples in V.A.4; for example, prod-
ucts like F × F contain zero-divisors, as do matrix algebras.
V.B.2. E XAMPLES . (i) Field extensions are division algebras: e.g.,
C is an R-division algebra; and Q[ζ 5 ] is a Q-division algebra.
(ii) Quaternion algebras give some non-commutative examples: H
(Hamilton’s quaternions) is an R-division algebra; while the non-
split (i.e., division ring) cases in HW 6 #6 give Q-division algebras.
We are particularly interested in division algebras which are finite-
dimensional (as F-vector spaces). While number fields (viewed as
field extensions) easily yield an endless list of such examples over
Q, you may find it difficult to recall seeing any finite-dimensional
field extensions of C. That is because they don’t exist!
V.B.3. D EFINITION . (i) An algebraic field extension3 of F is one
whose every element is algebraic (cf. III.G.6(ii)) over F.
(ii) Call a field F algebraically closed if it has no algebraic field
extensions (other than itself).
V.B.4. E XAMPLE . The Fundamental Theorem of Algebra states
that every polynomial over C has a root (hence all roots) in C. (This
theorem is proved in complex analysis.) Since any element α of a
field extension which is algebraic over C satisfies a polynomial equa-
tion P(α) = 0, α actually belongs to C. So C is algebraically closed.
Clearly division algebras are the simplest kind of F-algebra after
field extensions; so we shall do a rough classification for F = R, C,
and finite fields. To begin with, we finish off C with the
3Warning: these need not be finite-dimensional (though they certainly are if they
are finitely generated).
V.B. FINITE-DIMENSIONAL DIVISION ALGEBRAS 255
eva : F[λ] F[ a] ⊂ A
f ( λ ) 7 → f ( a ).
A 0 : = { α ∈ A | α 2 ∈ R≤0 } ⊂ ( A \ R ) ∪ { 0 } .
u2 = ( av + b)2 = a2 v2 + 2abv + b2 ,
the RHS terms are real except for 2abv, which forces ab = 0. But we
can’t have a = 0, for then u = b ∈ R; and if b = 0, then u = av
contradicts the independence.
So u is not of the form av + b, which means that u, v, and 1 are
independent over R. Hence u + v, u − v ∈ A\R; and so as above
(for α), they satisfy irreducible quadratic equations
0 = ( p + r )u + ( p − r )v + (q + s − 2c − 2d)1.
For u ∈ A0 , set
Q(u) := −u2 ∈ R.
Write
F n = A = A i ⊕ · · · ⊕ A i = F mi ⊕ · · · ⊕ F mi
| {z } | {z }
di di
=⇒ n = mi di =⇒ mi | n (∀i).
Finally, define the dth cyclotomic polynomial
∏
j
f d (λ) := ( λ − ζ d ),
1 ≤ j ≤ d−1
(d, j) = 1
λn − 1 = ∏ f d ( λ ),
1≤d≤n
d|n
4I have changed font for the field, because we want to think of A = F as a field
extension of some original field F.
260 V. REMARKS ON ASSOCIATIVE ALGEBRAS