NUMBER THEORY VIA ALGEBRA AND GEOMETRY
DANIEL LARSSON
CONTENTS 1. Introduction 2 2. Rings 3 2.1. Definition and examples 3 3. Basics @ring 5 3.1. Ideals and subrings 6 4. Integral domains 7 4.1. Homomorphisms, quotient rings and the first isomorphism theorem 8 4.2. UFD’s, PID’s and Euclidean domains 13 4.3. The Gaussian integers 18 4.4. Polynomials 20 5. Fields 22 5.1. Definition and examples 22 6. Basics @field 23 6.1. Fields of fractions 24 7. Field extensions 25 7.1. Field extensions 25 7.2. Algebraic extensions, transcendental extensions 26 7.3. Simple and finitely generated extensions 27 7.4. Algebraic closure 28 8. Finite fields 28 8.1. The main theorem 29 8.2. The Frobenius morphism 30 9. Algebraic number fields 30 9.1. Algebraic numbers 30 9.2. Norms, traces and conjugates 31 9.3. Algebraic integers and rings of integers 34 9.4. Integral bases 38 9.5. Computing rings of integers 39 9.6. Examples 42 10. Quadratic number fields 44 10.1. Ring of integers of quadratic number fields 44 10.2. The Ramanujan–Nagell theorem 44 11. Dirichlet’s unit theorem 47 11.1. Roots of unity 47 11.2. Units in number fields 48 12. Dedekind domains 49 12.1. A few important remarks 49 12.2. The main theorem on Dedekind domains 50 1 2 D. LARSSON
13. Extensions, decomposition and ramification 54 13.1. Ramification and decomposition 55 13.2. Consequences for quadratic number fields 58 13.3. Consequences for some non-quadratic number fields 64 14. Cyclotomic number fields 65 14.1. Cyclotomic fields 65 14.2. Galois theory of number fields 72 14.3. Gauss sums and Quadratic Reciprocity 74 14.4. Cubic reciprocity 81 15. Arithmetic and Geometry 84 15.1. Affine n-space 84 15.2. Projective n-space 86 15.3. Algebraic curves 87 15.4. Cubic and elliptic curves 91 15.5. The group structure of an elliptic curve 92 15.6. The group law 93 15.7. Points of finite order 95 15.8. The Nagell–Lutz theorem 96 15.9. Mordell’s Theorem and Conjecture 97 16. Gauss’ Class Number Problem and the Riemann hypotheses: A Historical Survey 98 16.1. Gauss’ class number problem 98 16.2. Quadratic fields and forms 99 16.3. What does this have to do with the class number problem? 103 16.4. Zeta functions and the Riemann hypothesis 104 16.5. Back to quadratic fields 107 17. Appendix A: Linear algebra 109 17.1. Vector spaces and bases 109 17.2. Maps 109 17.3. Dual spaces 110 17.4. Operations on maps 110 17.5. Linear equations and inverses 111 17.6. Modules 111 17.7. Vandermonde determinants 112 18. Appendix B: Chinese remainder theorem 113
1. INTRODUCTION These set of notes are the Lecture Notes accompanying my class in Number theory at Uppsala University, Spring terms 2008 and 2009. The notes begin with some very basic ring and field theory in order to set the stage, and continues to more advanced topics successively and probably in a rather steep upwards slant from an extremely soft and cosy start. I allow myself a few digressions in the text that are not part of the syllabus but that I feel are in a sense part of the required “know-of” for aspiring mathematicians. In this case I am mainly refering to the sections concerning the Gauss’ class number problem and the Ramanujan–Nagell theorem. These are then not formally part of the course and will not 3 be discussed in an exam. But I strongly encourage readers to at least read through these parts to get an idea of the beauty that lies within. Also, some parts are more abstract and technical, mainly the section on rings of in- tegers. The proofs here are rather difficult but I felt that if I didn’t include them in the course I would be cheating (which I don’t like). Therefore, I don’t require the students to learn this material, but rather to have this as a fall-back solution if the later results feel a bit hollow and improperly motivated. In the same spirit, I include Appendices with notions from “linear algebra over rings” for easy reference. As a twist to this course I added a section on elliptic curves, a topic that, without a doubt, will be part of every course on number theory that ever will be given anywhere on the planet, or elsewhere (this is a foretelling on my part). A home-assignment will be given where the students are to learn the basics of elliptic curves over finite fields, so that they immediately can understand the basic ideas and quickly learn the techniques of elliptic curve cryptography and (large) integer factorization using elliptic curves. This will surely be a worthwhile effort for every mathematically inclined student.
2. RINGS I will assume that everyone knows what an abelian group is. 2.1. Definition and examples. We begin with the following definition. Definition 2.1. Let R be a binary set with two closed operations, ’+’ (addition) och ’∗’ (multiplication). Then R = (R,+,∗) is a ring if addition and multiplication is compatible according to the following axioms: Rng1: The operation + makes R into an abelian group, that is, a + b = b + a for all a,b ∈ R. Rng2: There is an element 1 such that 1 ∗ a = a ∗ 1 = a for all a ∈ R, or in other words, a multiplicative unit, often simply called a one. Rng3: The multiplication is associative: a ∗ (b ∗ c) = (a ∗ b) ∗ c, ∀a,b,c ∈ R. Rng4: The multiplication is distributive: a ∗ (b + c) = a ∗ b + a ∗ c and (b + c) ∗ a = b ∗ a + c ∗ a, ∀a,b,c ∈ R. Remark 2.1. Strictly speaking, what we have defined here should be called an associative ring with unity (or associative, unital ring) The most general definition includes only Rng1 och Rng4. There are many examples of structures satisfying only these two axioms. However, for us it is enough to state the definition in the above, more restrictive, way. Note! From now on we write multiplication in the usual fashion ’a · b’ or simply ’ab’. 2.1.1. Examples. Example 2.1. The easiest and most obvious (and arguably the most important) examples are of course the following - Z, the ring of integers; - Q, the ring of rational numbers; - R, the ring of real numbers; - C, the ring of complex numbers. 4 D. LARSSON
Convince yourselves that these are indeed rings! In fact, they are even commutative.
Example 2.2. Another, extremely important example of a ring is Z/hni, the ring of integers modulo n. Recall that this is the set of congruence classes of elements modulo division by n, and can be represented by {0,1,2,...,n − 1}, the remainders after division with n. Informally, one writes Z/hni := {0,1,2,...,n − 1}, implicit being that addition and multiplication are allowed. Example 2.3. Another example is √ √ Z[ n] := {a + b n | a,b ∈ Z} for n a square-free integer. We will see a lot of examples like this in the future, since number-theoretic applications to rings often involve examples like this one. So far, every example has been commutative. Let’s round off this example-listing with a non-commutative number system.
Example 2.4 (Quaternions)√. Recall that the complex numbers C was formed by, to R, add an imaginary number i := −1. As you no doubt remember C := {a + bi | a,b ∈ R}. The ring C has one further property that we haven’t discussed yet (but will, in the next 1 section): every non-zero complex number has an inverse A question that was asked√ during around if one could add yet another “really” imaginary element (not equal to −1), j, so that one gets a new ring with the property that every non-zero element has an inverse. This can be shown to be impossible! However, 1843, W.R. Hamilton realized that if one adds two more elements to C, the element j and another k, then every non-zero element has an inverse! But, alas, one loses one important thing: commutativity. The definition is as follows. Definition 2.2. The set of all “numbers” 2 2 H := {z = a + bi + cj + dk | a,b,c,d ∈ R, i = j = −1 · 1, ij = −ji = k} is a non-commutative ring, where all elements 6= 0 has an inverse. An element of H is called a quaternion and H is called the ring of quaternions. I strongly suggest that you do the following exercise.
Exercise 2.1. Try out a few computations and check that H is indeed a non-commutative ring. (You don’t have to check that it is a ring. This is obvious! Why?) To define an inverse recall how inverses are computed in C and emulate that. In order to do this you will have to define a suitable notion of conjugate quaternion.
Notice that R ⊂ C ⊂ H and that we double the dimension in each step: C is two- dimensional as a vector space over R and H, is four-dimensional. The natural follow-up question is: is it possible to add more imaginary elements and get larger number systems? The answer is yes, and this was given only a few months after Hamilton’s discovery by J.T Graves and later by A. Cayley. But adding just one element is not sufficient (just as it was not sufficient to add just one to C to get H): one has to add
1 One thing, however that is lost when passing from R to C (that is by adding i) is order: it is meaningless to ask which one of two complex numbers is the biggest. The only thing one can, meaningfully say is which of them has the greatest distance from the origin. 5 four new ones! The result is the eight-dimensional octonions, denoted by O. The problem is now that another property is lost: namely, associativity! (Recall the remark after the definition of a ring.) After the octonions comes the sixteen-dimensional sedonions, S. This time there are non-zero elements a and b such that ab = 0. Such elements are called zero-divisors (we will see more on this soon). One can continue this indefinitely with doubling dimensions in each step. The objects in this nested sequence of rings are called Cayley–Dickson algebras (CDA’s) R ⊂ C ⊂ H ⊂ O ⊂ S ⊂ ···. | {z } CDA For those finding this fascinating I recommend the Wikipedia-article on the Cayley–Dickson construction.
3. BASICS @RING We begin with the following rather lengthy definition. Definition 3.1. Let R be a ring (commutative with unity). - An element 0 6= a ∈ R is a zero-divisor if there is a 0 6= b ∈ R such that ab = 0. - An inverse to 0 6= a ∈ R is an element a−1 such that aa−1 = a−1a = 1. An element is called invertible if it has an inverse. - The characteristic, char(R), of R is the least number n ∈ N such that na = a + a + ···a = 0 for all a ∈ R. | {z } n times If no such least n exists, we put char(R) = 0.
Example 3.1. In Z/h6i = {0,1,2,3,4,5}, 2 is a zero-divisor since 2 · 3 = 6 = 0. The characteristic of Z/h6i is six. The invertible elements are 1 and 5 which are their own inverses (check!). On the other hand, Z has no zero-divisors and characteristic zero. The only invertible elements in Z are ±1 and they are also their own inverses. Theorem 3.1. Let R be a ring and a,b,c ∈ R∗ := R \{0}. Then, (i) 0a = a0 = 0, (ii) the unity of R is unique, (iii) (−a)b = a(−b) = −(ab), and (iv) if a multiplicative inverse to a exists, it is unique. Proof. These are simple: (i) 0a = (0 + 0)a = 0a + 0a = 2 · 0a ⇔ 0a = 0 and similarly we have a0 = 0. (ii) The proof of this is exactly as for groups: 1 = 1 · 10 = 10. (iii) What one needs to observe is that to show (iii) all you need to show is that the terms involved are all inverses ab. Then the result follows since additive inverses are unique (the additive structure is a group structure, remember). Indeed, ab + (−a)b = (a + (−a))b = (a − a)b = 0b = [from (i)] = 0. The others are exactly the same and are left to you. (iv) Suppose there are b,c ∈ R such that ab = ac = 1. Then ab−ac = 0 ⇐⇒ a(b−c) = 0 so multiply with one of the inverses to a from the left: ba(b − c) = 0 ⇐⇒ 1(b − c) = 0 ⇐⇒ b = c. 6 D. LARSSON
Direct products. There are several ways of constructing new rings out of old ones. Here is one useful example. Let {Ri}, i ∈ I, where I is some index set, be a collection of rings. The direct product of {Ri}, written ∏i∈I Ri is the set of all sequences
∏Ri := {(ri) | ri ∈ Ri}. i∈I When I is finite (as all our examples will be), one usually write this as
R1 × R2 × ··· × Rn. In this case the ring structure is given by component-wise addition and multiplication, 0 0 0 0 (r1,...,rn) + (r1,...,rn) := (r1 + r1,...rn + rn), 0 0 0 0 (r1,...,rn)(r1,...,rn) := (r1r1,...rnrn), and the zero element and unity are respectively, (0,0,...,0), (1,1,...,1). It is easy to prove (do this!) that R1 × ··· × Rn is a ring when all the Ri’s are rings. On the other hand, it is a little more subtle to show the ring axioms for infinite index sets. The statement is nonetheless true for general index sets.
3.1. Ideals and subrings. From now on we will only deal with commutative rings. This makes life a lot easier since non-commutative rings are often very strange creatures and one has to be very tongue-in-cheek when dealing with them so as not to fall into the trap of thinking commutatively. But for basic number theory the commutative theory suffices. Definition 3.2. Let R be a (commutative) ring. • An ideal in R is a subgroup i ⊆ R such that ri ⊆ i for all r ∈ R. This means that for all r ∈ R and i ∈ i, ri ∈ i. An ideal is called proper if i ( R and trivial if i = 0. • A subring is a subgroup S of R such that ab ∈ S for a,b ∈ S and such that 1 ∈ S. Notice the difference. Lemma 3.2. If 1 ∈ i for some ideal i then i = R. Proof. For every r ∈ R, r = r1 ∈ i so R ⊆ i. The other inclusion is clear from definition of course.
Example 3.2. In Z, the ideals are on the form nZ := {nz | z ∈ Z}, for some n ∈ Z. How do we prove this? First of all, any subgroup S of a cyclic group (like 2 Z) is cyclic so S = nZ as groups . Clearly nZ is stable under multiplication of Z in the sense that for a ∈ nZ, xa ∈ nZ, for x ∈ Z. Hence nZ is an ideal. What are the subrings? A subring has to be a subgroup so they must then also be on the form nZ for some n ∈ Z. But a subring has to include 1 and so n = 1. This means that there is only one subring, namely, Z itself.
2If this doesn’t ring a bell, here is a direct argument. Let G be a group generated by a ∈ G in the sense that k l for g ∈ G, g = a for some k ∈ Z and let S be a proper subgroup. Every element s ∈ S can be written as s = a d for some l ∈ Z. Let d be the smallest integer dividing all the l’s as s ranges through S. Then S is generated by a . Notice that if d = 1 then S = G, so by the assumption of properness we have, d > 1. 7
Example 3.3. The only ideal of Q is Q itself, so there are no proper ideals. Indeed, let − i ⊆ Q be an ideal. Then for q ∈ i, we have q 1q = 1 ∈ i and so by the lemma above, i = Q. This is a general phenomena for rings where all non-zero elements have inverses. As an exercise, find a subring (there are infinitely many)! You will see examples of this later. Definition 3.3. Let R be a ring. - An ideal given as h f i := {r f | r ∈ R} is called a principal ideal. - A proper ideal p ⊂ R is called a prime ideal if ab ∈ p implies that a ∈ p or b ∈ p. - A proper ideal m ⊂ R is called a maximal ideal if, for a an ideal such that m ⊆ a ⊂ R then a = m.
Example 3.4. In Z all ideals are principal hni = nZ. An ideal in Z is prime if and only if n is a prime; an ideal in Z is maximal if and only if it is prime. 3.1.1. Ideal generation. The easiest and by far the most common way of constructing ideals is by generation:
Definition 3.4. The ideal i is generated by { fi | fi ∈ i,i ∈ I}, where I is some index set, if a ∈ i =⇒ a = ∑ ri fi, for ri ∈ R. J⊆I J finite We write this as i = h{ fi | fi ∈ i,i ∈ I}i. Check that this is an ideal! The ideal is called finitely generated if |I| < ∞, that is, if the number of fi’s are finite. Notice that if |I| = 1 then we get a principal ideal. 3.1.2. Operations on ideals. Suppose that i and j are two ideals of R. We make the following definitions: i + j := {i + j | i ∈ i, j ∈ j} iR := {∑ir | i ∈ i,r ∈ R, finite sum} ij := {∑i j | i ∈ i, j ∈ j, finite sum}. Theorem 3.3. The sets i + j, iR and ij are all ideals of R.
Proof. Exercise!
4. INTEGRALDOMAINS For number theorists the following definition is of utmost importance. Definition 4.1. A ring with no zero-divisors is called an integral domain or simply a domain.
Example 4.1. The ring Z is an integral domain. In fact, Z is the reason for “integral” in the name. It was the first and most natural ring to study. On the other hand, Z/h6i is not an integral domain since 2 · 3 = 0. Theorem 4.1. A ring R is an integral domain if and only if h0i is a prime ideal.
Proof. Exercise! 8 D. LARSSON
Theorem 4.2. A ring R is an integral domain if and only if it satisfies the cancellation properties: if a 6= 0 then ab = ac =⇒ b = c, and ba = ca =⇒ b = c. Proof. Suppose first that R is a domain. That ab = ac is equivalent to 0 = ab−ac = a(b−c) and since R does not have any zero-divisors and a 6= 0, we get b−c = 0. In the same manner one shows the right-handed version. Conversely, if ab = ac ⇒ b = c, suppose that xy = 0 with x 6= 0. This is equivalent to xy = x0 and hence y = 0. Theorem 4.3. If every non-zero element in R has an inverse then R is an integral domain. −1 Proof. Let ab = 0 with a 6= 0. Multiply with a from the left and the result follows. Theorem 4.4. Every non-zero element in a finite domain is invertible.
Proof. Let R := {0,1,a2,··· ,an}. Pick an element 0 6= b ∈ R. We want to fin a c ∈ R such that bc = cb = 1. Multiply all elements in R with b from the left. Then we get: {0,b,ba2,ba3 ··· ,ban}. Every element in R appears in this set exactly once. This is be- cause, bai = bak ⇒ ai = ak since R is an integral domain. This means that 1 has to appear somewhere in this set. Hence ba j = 1 for some (specific) a j. The same argument applies to the right-handed case. Theorem 4.5. The zero-divisors in Z/hni are exactly the elements m 6= 0 that are not relatively prime to n, that is, gcd(m,n) = d > 1. Proof. First if gcd(m,n) = d > 1, m 6= 0, then m(n/d) = (m/d)n = 0 in Z/hni since the right-hand-side is a multiple of n. This means that m is a zero-divisor in Z/hni since neither m nor n/d is zero. On the other hand, if gcd(m,n) = 1, m 6= 0, and mk = 0 = tn. Therefore, n has to divide mk. But since m and n are relatively prime, we have n|k, which is equivalent to k = 0 in Z/hni. Corollary 4.6. The ring Z/hpi, where p is a prime, has no zero-divisors. Corollary 4.7. The ring Z/hni is a domain and every non-zero element is invertible if and only if n is a prime. Proof. If n is not a prime there are zero-divisors3. On the other hand, if n = p, it follows from Theorem 4.4 that every element is invertible, and from the previous corollary that Z/hpi has no zero-divisors. This shows, incidentally, that Z/hpi is a field. More on this later. 4.1. Homomorphisms, quotient rings and the first isomorphism theorem. A very im- portant principle in the philosophy of modern mathematics is that mathematical objects are to a very large extent governed by their maps to other objects of the same “category” (e.g., groups or rings). Hence we need to define what is to be meant by a “map between rings”. Definition 4.2. A set-theoretical map of rings φ : R → S is a ring homomorphism or ring morphism if it is a homomorphism of the underlying abelian groups, i.e., φ(a + b) = φ(a) + φ(b), and if φ respects multiplication in the sense that φ(ab) = φ(a)φ(b), in addition to φ(1R) = 1S.
3Also, we note that zero-divisors have no inverses: ab = 0 =⇒ b = 0 and zero is not invertible. 9
φ Definition 4.3. Let R −→ S be a ring morphism and for s ∈ S denote
φ −1(s) := {r ∈ R | φ(r) = s}.
This is called the fiber over s or the inverse image of s. Put
ker(φ) := φ −1(0) = {r ∈ R | φ(r) = 0}, the kernel to φ, and im(φ) := {s ∈ S | ∃r ∈ R, φ(r) = s}, the image to φ.
Notice that since φ is a group homomorphism we have φ(0) = 0.
φ Theorem 4.8. For R −→ S a ring morphism, we have (i) ker(φ) is an ideal in R. (ii) im(φ) is a subring in S but not necessarily an ideal.
Proof. The proof is rather easy: (i) Let b ∈ ker(φ) ⊆ R, a ∈ S. Take r ∈ R. We want to show that rb ∈ ker(φ) to show that ker(φ) is an ideal. But this is obvious since φ(rb) = φ(r)φ(b) = φ(r) · 0 = 0, and so rb ∈ ker(φ). (ii) That im(φ) is a subring follows immediately from the definition. Think through this!
Definition 4.4. We can make the following definitions in analogy with the corresponding notions from group theory. - A ring morphism φ : R → S is an injection (or “one-to-one”) if kerφ = {0}; this is written φ : R ,→ S; - φ is a surjection (or “onto”) if imφ = S, written φ : R S; and • an isomorphism (“one-to-one, onto”) if it is both injective and surjective. φ We say that R and S are isomorphic if there is an isomorphism R −→ S and it is customary to write this as R ' S or R ≈ S.
Example 4.2. Here are some easy examples. - Reduction mod n is a ring morphism Z → Z/hni. - The inclusion Z ,→ Q is a ring morphism. - Complex conjugation is a ring morphism C → C. This is in fact an isomorphism. Isomorphisms between a ring and itself are called automorphisms.
The set of all ring morphisms between rings R and S is denoted by Hom(R,S) (to remind you of ’homomorphism’). If S = R then we put End(R) := Hom(R,R). Ring morphisms from R to itself is called endomorphisms.
Example 4.3. The map n· : Z → Z, n · z 7→ nz, is a group homomorphism, but not a ring homomorphism since it is not multiplicative: n · (ab) 6= nanb. So there are a lot fewer ring morphisms than group morphisms. 10 D. LARSSON
4.1.1. Quotient rings. Let i be an ideal of R. Introduce a relation on R as follows:
a ∼i b ⇐⇒ a − b ∈ i, for all a,b ∈ R.
Theorem 4.9. The relation ∼i is an equivalence relation on R. Proof. We need to show three things:
- Reflexivity: a ∼i a for all a ∈ R. This is clear since 0 ∈ i. - Symmetry: a ∼i b ⇒ b ∼i a. This follows since a − b ∈ i ⇔ −(b − a) ∈ i and since i is a subgroup −a ∈ i ⇒ a ∈ i. - Transitivity: a ∼i b and b ∼i c implies a ∼i c. This follows by a − b ∈ i and b − c ∈ i which, by adding (once again using that i is a subgroup), gives a − b + b − c = a − b ∈ i. Definition 4.5. The coset (i.e., an equivalence class) of a ∈ R under this equivalence rela- tion is denoted a + i or sometimes [a] ora ¯. The set of cosets is denoted R/i := {a + i | a ∈ R}. Theorem 4.10. The operations (a + i) + (b + i) := (a + b) + i (a + i)(b + i) := ab + i defines a ring structure on R/i, with zero element i and unity 1 + i. There is also a canonical surjective ring morphism π : R → R/i with kerπ = i, defined by π(r) = r + i. Definition 4.6. The ring defined by this theorem is called the quotient (or factor) ring modulo i. What this essentially means is that we set all elements in i to zero, i.e., we are “killing the kernel kerπ = i” since, in R/i the zero element is 0 + i = i. Proof. We have to check that the definitions are well-defined, i.e., that choosing different representatives from each coset gives the same result. Let us prove this in the case of multiplication, the additive counter-part being completely analogous and therefore left to the reader. So, suppose we are given x + i = a + i and y + i = b + i. This means that x = a + ix and y = b + iy, for ix,iy ∈ i. Hence,
(x + i)(y + i) = xy + i = (a + ix)(b + iy) + i = ab + aiy + ixb + ixiy + i = = ab + i = (a + i)(b + i), where we have used that aiy,ixb,ixiy ∈ i. Notice that this uses in a essential way that i is an ideal. The rest of the ring axioms follow since the ring operations are induced from R. For the last statement, π(r + s) = (r + s) + i = (r + i) + (s + i) = π(r) + π(s), π(rs) = rs + i = (r + i)(s + i) = π(r)π(s), and clearly given r + i ∈ R/i, π(r) = r + i so π is certainly surjective. Example 4.4. The canonical example is the following. Every subset of Z of the form nZ = {nz | z ∈ Z} is an ideal of Z. It is a fact (that can be deduced from the next subsection) that Z/hni ' Z/nZ. 11
We will return to this example after the following theorem, subsequent discussion and proof. 4.1.2. The first isomorphism theorem. The following theorem is arguably the most useful theorem in all ring theory. Theorem 4.11 (First isomorphism theorem). Let φ : R → S be a ring homomorphism. Then the following diagram commutes:
φ R / S x; xx π xx xx ¯ xx φ R/kerφ, that is, φ = φ¯ ◦ π, and induces an isomorphism R/kerφ ' imφ. Remark 4.1. First of all, what does it mean for a diagram to “commute”? Well, intuitively it means that what path you take from a particular point to any other (following the direc- tions of the arrows) is not important. In the above example it means that going from R to R/i and then to S is the same as going from R to S directly. Another way of saying this is that φ factorizes through R/kerφ as φ = φ¯ ◦ π. Proof. The proof of this theorem is surprisingly simple. First of all we define π by π(r) := r + kerφ. Then φ¯ can be defined as φ¯(r + kerφ) := φ(r). Notice that all the maps are now ring morphisms. Clearly, φ = φ¯ ◦ π, but we have to check that φ¯ is well defined. So suppose that r + kerφ = s + kerφ ⇔ r − s = f ∈ kerφ ⇔ r = s + f . Hence, φ¯(r + kerφ) = φ(r) = φ(s + f ) = φ(s) + φ( f ) = φ(s) = φ¯(s + kerφ). To see that we have the isomorphism R/kerφ ' imφ, note that φ is a surjection onto its image (by definition!) and so since imφ¯ = imφ the same goes for φ¯. Also, kerφ¯ = {0}, so kerφ¯ is injective, and therefore an isomorphism. This theorem gives a more precise meaning to “killing the kernel”! Example 4.5. Let’s continue the example given right before this subsection. We have the ring morphism “reduction mod n”: Redn : Z → Z/hni given by Z 3 r 7→ r¯, wherer ¯ is the reduction (remainder) modulo n. The kernel of this map is all the multiples of n: ker(Redn) = nZ. Therefore, we get the following commutative diagram:
Redn Z / Z/hni v; vv π vv vvRedn vv Z/nZ.
Clearly, Redn : Z → Z/hni is a surjection so im(Redn) = Z/hni and the isomorphism from the theorem becomes, Z/nZ ' im(Redn) = Z/hni. Example 4.6. The projections prR R × S // R,S prS 12 D. LARSSON are ring homomorphisms (check this!). Consider the morphism prR : R×S → R. The kernel of this is clearly S. Hence the theorem tells us, (R × S)S ' R, and similarly, by using prS, (R × S) R ' S. So in this sense, ’/’ is really like a division, if ’×’ is viewed as a multiplication. This is probably the origin of the name “quotient ring” and the notation ‘/’. 4.1.3. Ideals and ring morphisms. Now we come to another, very important result: how do ideals behave under ring morphisms. We make the following definition. Definition 4.7. Let φ : R → S be a morphism of rings and i an ideal of R and j an ideal of S. Then, i∗ := φ(i)S = { ∑ iφ s | iφ ∈ φ(i),s ∈ S} finite is the extended ideal of i in S and j∗ := φ −1(j) = {r ∈ R | φ(r) ∈ j} the contracted ideal of j in R. Sometimes these are denoted iS and j ∩ R, respectively. I will probably use the first of these alternate notations but hardly the second. ∗ Theorem 4.12. With notation as above, i∗ and j are ideals of S and R, respectively. Proof. We prove the second, leaving the first to you as an exercise (do it!). Suppose, j ∈ j∗ and r ∈ R. Then φ(r j) = φ(r)φ( j) ∈ j since φ(r) ∈ S, φ( j) ∈ j (by definition) and j is an ideal. In the same manner, j + j0 ∈ j∗, when j, j0 ∈ j∗, φ( j + j0) = φ( j) + φ( j0) ∈ j since j is an ideal. Theorem 4.13. Let φ : R → S be a ring morphism between two rings with ideals i ⊆ R and j ⊆ S. Then ∗ −1 (4.1) (j )∗ = φ(φ (j)) = imφ ∩ j ∗ −1 (4.2) (i∗) = φ (φ(i)) = kerφ + i. Proof. We divide the proof as in the statement of the theorem. (4.1) We need to show two inclusions: φ(φ −1(j)) ⊆ imφ ∩ j and φ(φ −1(j)) ⊇ imφ ∩ j. Let’s start with the first. Take a ∈ φ(φ −1(j)). Obviously a ∈ imφ. Since φ −1(j) is the set of elements mapping into j, we obviously also have a ∈ j. Hence a ∈ imφ ∩ j. For the other inclusion, assume a ∈ imφ ∩ j. Then, φ −1(a) is a set of elements mapping to a ∈ imφ ∩ j under φ, so a ∈ φ(φ −1(j)). Hence the (4.1) is proved. (4.2) As above, we need to prove two inclusions: φ −1(φ(i)) ⊆ kerφ + i and φ −1(φ(i)) ⊇ kerφ + i. Take a ∈ kerφ + i. Then φ(a) ∈ φ(i), and so a ∈ φ −1(φ(s)) ⊆ φ −1(φ(i)) since this is the set of elements mapping to φ(i) under φ. To show the other inclusion, take a ∈ φ −1(φ(i)). Then φ(a) ∈ φ(i) so φ(a) = φ(b) for some b ∈ i. This is equivalent to φ(a − b) = 0 and so a − b ∈ kerφ meaning that a = b + k, with k ∈ kerφ. But b ∈ i so the inclusion is proved, thereby completing the whole proof. 13
This theorem has the following remarkable corollaries. ∗ −1 Corollary 4.14. If φ : R S is a surjection, then (j )∗ = φ(φ (j)) = j. On the other ∗ −1 hand if φ : R ,→ S is an injection, then (i∗) = φ (φ(i)) = i. Proof. Obvious. Corollary 4.15. Let i be an ideal of R. Then there is a bijective correspondence between ideals of R/i and ideals j of R, such that i ⊆ j. The correspondence is given by i ⊆ j 7→ j/i := π(j) = { j + i | j ∈ j}. Proof. First of all it is clear that every ideal in R/i must be on the form B/i = {b+i | s ∈ B} for some subset B of R. However, if B/i is an ideal then (r+i)(b+i) should be in B/i for all r + i ∈ R/i. We have, (r + i)(b + i) = rb + i, and so rb must be in B, i.e., B must be an ideal. Furthermore, it is clear that B ⊇ i. Hence every ideal in R/i is on the form j/i for some ideal j ⊇ i. Conversely, given an ideal j ⊆ R we get an ideal j/i ⊆ R/i by ∗ projection. Suppose k1 := j1/i = j1/i =: k2. Then (k1)∗ = k2 by the previous corollary. ∗ ∗ ∗ But this means that ((k1)∗) = k2 and so j1 = j2 and the correspondence is also one-to- one. 4.2. UFD’s, PID’s and Euclidean domains. In this subsection all rings are assumed to be integral domains even if it is not explicitly stated. 4.2.1. Prime elements and irreducible elements. Definition 4.8. Let D be an integral domain and a,b ∈ D. - We say that b divides a, written (as usual) a|b, if there is a c ∈ D such that a = bc; a is then called a divisor of b. - A divisor of 1 is called a unit4. The set of units in a ring is a multiplicative group (check!) and denoted U(D). - Two elements a and b such that a = bu ⇐⇒ au−1 = b, for u ∈ U(D), are called associates. Association is an equivalence relation (check!). - An element a ∈ D is irreducible if any factorization a = bc implies that either b or c is a unit. Otherwise, a is called reducible. - An element p ∈ D is called a prime (element) if p|ab implies p|a or p|b. Remark 4.2. Be sure to separate the similarly named, but distinctly different, notions ’unit’ and ’unity’. This is notorious source of confusion and frustration. The following simple remark is important enough to earn its way as a full-fledged the- orem. Theorem 4.16. Let D be an integral domain. Then (i) a|b ⇐⇒ b ∈ hai; (ii) a|b and b|a ⇐⇒ a = ub for a unit u ∈ U(D) ⇐⇒ hai = hbi.
Proof. Exercise! Theorem 4.17. Any prime element is irreducible (but in general, not the other way around) and p ∈ D is prime if and only if hpi is a prime ideal.
4Notice that this is equivalent to an invertible element. To be honest, I don’t know why there are two names for the same concept. My lingering feeling is that the designation ’unit’ is a bit more “number-theoretical”. But I could be wrong. 14 D. LARSSON
Proof. Let p be a prime and assume that it is reducible p = ab where neither a nor b is a unit. Then clearly p|ab but p - a and p - b (since if, for instance, p | a then a = pk for some k, and so p = pkb; since D is a domain, we can cancel p to obtain bk = 1, but b was not a unit so we have a contradiction) so we have a contradiction to p being prime. Suppose now that p is a prime and let ab ∈ hpi. This is equivalent to ab = pc for some c ∈ D and this in turn equivalent to p | ab and since p is prime p|a or p|b. Now use Theorem 4.16 (i). Conversely, suppose that hpi is a prime ideal and ab ∈ hpi. This implies that a ∈ hpi or b ∈ hpi. In either case a = pc or b = pc and we are done. This theorem is the reason why prime ideals are called prime ideals.
4.2.2. Unique factorization domains. Definition 4.9. A domain D is called a unique factorization domain, abbreviated UFD, if UFD1. Every element can be written as a product of irreducibles, and UFD2. this factorization is unique up to multiplication by a unit. Let me comment on the last condition. Let
n1 nr m1 ms p1 ··· pr = q1 ···qs be two factorizations of a into irreducibles. Then the condition says that r = s and pi = ui jq j for some i and j and ui j ∈ U(D).
Theorem 4.18 (The fundamental theorem of Arithmetic). The ring of integers Z is a unique factorization domain. In general it is rather difficult to prove that a given integral domain admits unique fac- torization. Look up the proof of the above theorem in your elementary algebra books. Theorem 4.19. If UFD1 in Definition 4.9 holds, then UFD2 is equivalent to the statement “every irreducible element is prime.” Proof. Assume first that factorization is unique up to a unit. So let p be an irreducible element and suppose p|ab. We need to show that p|a or p|b. We begin by assuming that a and b are also irreducible and p - a. We have that pc = ab and p,a,b are irreducible so c is not a unit, since otherwise p = (c−1a)b, a factorization into irreducibles. This implies that c = ak for some k ∈ D and so pak = ab ⇔ a(pk − b) = 0, implying that p|b since D is a domain. Let us return to the general case. Let a = α1 ···αn and b = β1 ···βm. Since the factorization is unique (up to units) the product of a and b must be α1 ···αnβ1 ···βm. So if p|ab then p must divide a product of irreducibles and so must divide (at least) one of those. Hence p divides a or b. Assume now that every irreducible element is prime and let
n1 nr m1 ms p1 ··· pr = q1 ···qs be two factorizations of a into irreducibles. This tells us that q1 (for instance) divides both sides. Since every irreducible element is prime we see that all the pi’s and q j’s are prime n1 nr and q1|p1 ··· pr implies that q1|pi for some i and so q1 = spi, but since q1 is irreducible s ∈ U(D). Continuing this way inductively the result follows. 15
4.2.3. Principal ideal domains. Definition 4.10. A domain D is called a principal ideal domain, abbreviated PID, if all ideals are principal, i.e., generated by one element. Theorem 4.20. Let D be a PID. Then, p is irreducible ⇐⇒ p is prime. In addition, in a PID, every prime ideal is maximal. Proof. The implication ⇐ has already been demonstrated (as a general fact). For the reverse implication, let p be irreducible and let ab ∈ hpi with b ∈/ hpi. We then have that ab = cp. Since p is irreducible and b ∈/ hpi we have a = c1 p, i.e., a ∈ hpi. For the last statement, observe that if hpi is prime and hpi ⊆ hqi, for q ∈/ U(D), then p = cq and since p is irreducible, c ∈ U(D) and so hpi = hqi. Theorem 4.21. Every PID is a UFD. Proof. Let D be a PID and S ⊆ D be the set of elements which cannot be written as a product of irreducibles. Assume that S is non-empty and take a ∈ S. Then a = a1b1, a1,b1 ∈/ U(D). This means that hai ⊂ ha1i and hai ⊂ hb1i. If both a1,b1 ∈/ S then this would imply that a is a product of irreducibles so assume that a1 ∈ S. The same argument applies with a replaced by a1 to get another a2 such that hai ⊂ ha1i ⊂ ha2i. Continuing this process we get an infinite sequence of strict ideal inclusions:
hai ⊂ ha1i ⊂ ha2i ⊂ ··· ⊂ hani ⊂ ··· . S∞ The infinite union i=1haii is an ideal (check this!), and since D is a PID there is an element S∞ S∞ a ∈ D such that hai = i=1haii. Obviously, a ∈ i=1haii and so there is a 1 ≤ j < ∞ such that a ∈ ha ji. But this means that ∞ ∞ [ [ haii = hai ⊆ ha ji ⊂ ha j+1i ⊆ hai = haii, i=1 i=1 which is a contradiction and hence S must be empty. To show that the factorization is unique we use Theorem 4.19. So let p be an irreducible element and p|ab and p - a. Since hpi is maximal by Theorem 4.20, hp,ai = D and so 1 ∈ hp,ai. This means that there are α,β ∈ D such that aα + pβ = 1. Multiply this equation with b, to get abα + pbβ = b. Since p divides the left hand side, it divides b and by Theorem 4.19, the factorization is unique. The proof of this theorem shows, incidentally, that PID’s belong to a large and important class of rings which we now define. Definition 4.11. A ring R (not necessarily a domain5) is called Noetherian6 if it satisfies the ascending chain condition on ideals: for any ascending chain of ideals
a1 ⊆ a2 ⊆ ··· ⊆ an ⊆ ··· , there is an N such that for all k ≥ N,
ak = ak+1 = ak+2 = ··· . Theorem 4.22. Any PID is Noetherian.
5Or, commutative for that matter, although, in the non-commutative case one has to be a little more precise. 6In honor of Emmy Noether, 1882–1935. She was one of the big pioneers of abstract algebra. For more info, see her biography on http://www-groups.dcs.st-and.ac.uk/. 16 D. LARSSON
4.2.4. Integral domains that doesn’t have the unique factorization property. Definition 4.12. Let D be an integral domain. A (weak) norm on D is a function
Nrm : D → N0 := N ∪ {0}, such that (i) Nrm(ab) = Nrm(a)Nrm(b), and (ii) Nrm(a) = 1 ⇐⇒ a ∈ U(D). The definition implies: Lemma 4.23. If Nrm is a norm on D, then Nrm(0) = 0. Proof. Let a ∈ D be arbitrary. Then Nrm(0) = Nrm(a · 0) = Nrm(a)Nrm(0), which is equivalent to Nrm(0)(Nrm(a)−1) = 0. Since a was arbitrary, and N, as a subset of Z does not have any zero-divisors, we conclude that Nrm(0) = 0. This norm-function is convenient tool to show that certain elements are irreducible. √ √ √Example 4.7. Let D = Z[ −5] = {a + b −5 | a,b ∈ Z}. Let us show that 2,3 and 1 ± −5 are irreducible. √ First of all we need to define a norm on Z[ −5]. We do this as √ √ √ Nrm(z) = zz¯ = (a + b −5)(a − b −5) = a2 + 5b2, for z = a + b −5. Check for yourselves that this is indeed a norm. If 2 is not irreducible we have 2 = ab, where a,b are not units. This means that Nrm(a)Nrm√(b) = Nrm(2) = 4. Since a,b are not units, Nrm(a) = Nrm(b) = 2. Assume a = α +β −5. Then 2 = Nrm(a) = α2 +5β 2, which is clearly impossible for integers α and β. Hence 2 is irreducible.√ A similar argument shows√ that 3 is irreducible.√ Suppose now that 1 ± −5 is reducible, i.e., 1 ± −5 = uw, u,w ∈/ U(Z[ −5]). Then √ Nrm(u)Nrm(w) = Nrm(1 ± −5) = 1 + 5 = 6. Since 6 = 2 · 3 and u and w are not units either u is 2 or 3 and w the other possibility. But in that case u or w must be units because,√ as we showed above, it is impossible to find non-units with norm 2 or 3. Hence 1 ± −5 is irreducible.√ If 2 and 3 were associates then 2 = 3u, with u ∈ U(Z[ −5]). This would imply that Nrm(2) = Nrm(3)Nrm(u) = Nrm(3)·1, and this is clearly false (we also see that associates√ have the√ same norm). The same argument applies to the other cases except for 1 + −5 and 1 − −5. √ What are the units of Z[ −5]? An element u ∈ D is a unit if√ and only if Nrm(u) = 1, 2 2 which in this case is equivalent to, a + 5b = 1, for u = a + b −√5. This is clearly only 2 possible if b√= 0 and so a =√1, showing that u = ±1. Hence U(Z[ −5]) = {−1√,1}. So, 1 − −5 = ±1 · (1 + −5) is clearly not an option and hence, 2,3,1 ± −5 are irreducible, non-associates. Now, we see that √ √ 6 = 2 · 3 = (1 − −5)(1 + −5), and so we see that 6 has two different√ (i.e., they don’t differ by a unit) factorizations into irreducibles! This means that Z[ −5] is not a UFD, and hence not a PID either. √ Ok, you might say, what on earth is so interesting about (strange) rings such as Z[ −5]? If you also say that it is uninteresting to find integer solutions to equations of the type y2 + a = x3, a ∈ Z (the above case being a = 5), then, fine, I agree, in that case studying 17 √ rings such as Z[ −5] could be seen as a bit artificial. But, on the other hand, then your interest in number theory may be, at best, superficial. The above example indicates, for instance, that the equation y2 + 5 = x3 might not have any, non-trivial, integer solutions7. Also, as we will very soon see, these kind of rings are very important in number theory. Clearly, unique factorization in number theory would be a something extremely desirable. But as this example shows this is not always (or most often even) possible to achieve. However, and here is where things become beautiful, even though unique factorization of elements is not always possible, unique factorization of ideals into prime ideals is! Unfortunately, I don’t have time to go into this in detail, but we will come back to it briefly. Theorem 4.24. If Nrm(a) = p, where p is a prime, then a is irreducible. Proof. Suppose a is reducible a = bc, b,c ∈/ U(D). Then, p = Nrm(a) = Nrm(bc) = Nrm(b)Nrm(c). Since p is a prime, Nrm(b) (or Nrm(c)) must be p. But this implies that Nrm(c) = 1 and so c is a unit, contrary to assumption. Hence, a is irreducible. 4.2.5. Euclidean domains. Now we will introduce yet another type of norm, or rather valuation.
Definition 4.13. Let E be a domain. Then a function v : E → N0 is called an Euclidean valuation if the following are satisfied. Eu1. There is a division algorithm with respect to v. That is, for each pair a,b ∈ E, b 6= 0, there are q,r ∈ E such that a = bq + r, with v(r) < v(b), or, r = 0. Eu2. v(a) ≤ v(ab) for all E 3 a,b 6= 0. A domain E with a Euclidean valuation is called a Euclidean domain. Example 4.8. The following are examples of Euclidean domains. - Z,√ with v(a) := |a√|; √ 2 2 - Z[ ±2] = {a + b 2 | a,b ∈√Z}, with v(a +√b ±2) = a ± 2b ; √ - The Gaussian integers, Z[ −1] := {a + b −1 | a,b ∈ Z}, with v(a + b −1) = 2 2 a √+ b ; √ - Z[ 14] = {a+b 14 | a,b ∈ Z}. This example is highly non-trivial, proved 2004. √ It is not true that any Z[ n] is Euclidean, even if n is prime (see below). So what is so special about Euclidean domains? Well, we have: Theorem 4.25. Euclidean domain =⇒ PID =⇒ UFD. Proof. Only the first implication needs a proof since we have already proved the second one. For the first implication let i be an ideal of an Euclidean domain. Then let s be a smallest element, with respect to v, of i. Such an element exists since imv ⊆ N0. Then i = hsi (check!). Hence we have unique factorization in Euclidean domains, something desirable in num- ber theory.
7Fix x and factor the left-hand-side to so what I mean. 18 D. LARSSON
Example 4.9. At the moment of writing, the following facts√ are known (according to Wikipedia, which I kind of trust on this issue) concerning Z[ d]: • it is Euclidean for d = −1,−2,−3,−7,−11, and hence a PID; • it is not Euclidean for d = −19,−43,−67,−163, but is in fact a PID for these d’s. Quite an amazing result! √ Example 4.10. Notice that the above theorem shows that Z[ −5] is neither a Euclidean domain, nor a PID, since it is not a UFD. It is natural to wonder (somewhat stunned, I hope) “why are all these ’nearby’ cases so different!?”. Frankly, I don’t think anyone knows. But after all, this is number theory, the science of simple questions and hard (at best) answers. 4.3. The Gaussian integers. In order to fully motivate the abstract theory that I have covered so far, and that will be taken to even greater apparently nonsensical heights in a while, I thought it prudent to insert an example to show how all this actually has something to do with number theory, the de facto topic of this course. Theorem 4.26. Any odd prime number p (i.e., p 6= 0) is the sum of two squares if and only if p ≡ 1 (mod4), that is, p = a2 + b2 ⇐⇒ p ≡ 1 (mod4). √ Proof. A very convenient way to prove this is by using the Gaussian integers Z[ −1]. Recall that √ √ [ −1] = {a + b −1 | a,b ∈ }. Z √ Z We have already observed that Z[ −1] is Euclidean (see Example√ 4.8), and so, conse- quently, both a PID and UFD. Moreover, there is a norm Nrm : Z[ −1] → Z defined by √ √ Nrm(z) := zz¯ = a2 + b2, where z = a + b −1 andz ¯ = a − b −1. √ So p = a2 + b2, when considered in Z[ −1], factorizes as √ √ p = (a + b −1)(a − b −1). √ Since we will be dealing both with primes in Z and primes in Z[ −1], we follow tradition and refer to primes in Z as rational primes. √ Suppose a rational√ prime considered as an element of Z[ −1], p factors as p = xy, where x,y ∈/ U(Z[ −1]), that is, x and y are not units (see Remark below). Then Nrm(p) = Nrm(x)Nrm(y), and Nrm(p) = p2 =⇒ Nrm(x)Nrm(y) = p2. An element x in a normed ring is a unit if and only if Nrm(x) = 1 and since we assumed that x and y are not units√ we must have that Nrm(x) = Nrm(y) = p. This means in particular that, for x = a + b −1, Nrm(x) = a2 + b2 = p. Now, assume that p ≡ 1 (mod4) ⇔ p = 4n+1, for some n ∈ Z. By quadratic reciprocity 2 the congruence z ≡ −1 (mod p) have a solution for√ p = 4n + 1 (check this!). This means that there is a t ∈ Z such that p|(t2 + 1). But in Z[ −1], √ √ t2 + 1 = (t + −1)(t − −1) √ √ so p|(t + −1√) or p|(t − −1) if it were to be a prime. However, none of these are elements of Z[ −1]: for instance, √ √ t/p + −1/p ∈/ Z[ −1]. √ Hence p cannot be a prime in Z[ −1], and so by the preceeding paragraph p = a2 + b2. 19
The other implication, namely, p = a2 +b2 ⇒ p ≡ 1 (mod4) follows by elementary rea- soning. Indeed, not both a and b can be even because then p would be even, contradicting the assumption. Similarly, if a and b are both odd then p would once again be even. So we must have that one is odd and one is even, and this implies that p ≡ 1 (mod4) (check it!).
Remark 4.3. A very important remark here is that even though p ∈ Spec(Z) is irreducible, it can become reducible in an extension of Z. This is one of the basic realization of the number theorists of the 19’th century. In fact, this can always be given as general good advice: if something can’t be solved, can it be solved in some larger context? √ √ √ Lemma 4.27. The units of Z[ −1] are {1,−1, −1,− −1}. Proof. Use the characterization: z ∈ U(R) ⇔ Nrm(z) = 1. √ Theorem 4.28. The primes, modulo multiplication by units, of Z[ −1] are exactly the elements on the forms: √ (i) p = 1 + √−1; (ii) p = a + b −1 with a2 + b2 = p and p ≡ 1 (mod4), where p is a rational prime 6= 2; (iii) p = p a rational prime such that p ≡ 3 (mod4). √ Recall that, since Z[ −1] is a PID, the irreducible elements are exactly the prime ele- ments.
Proof. First we prove that the elements listed are indeed prime. If p factored as xy, we would have p = Nrm(p) = Nrm(x)Nrm(y) (recall that the norm of an irreducible element is a rational prime). This immediately implies that Nrm(x) (for instance) has norm 1 and hence is a unit. Therefore, the elements in (i) and (ii) are indeed primes. For the third case, note that a factorization, p = xy would imply p2 = Nrm(x)Nrm(y) and so p = Nrm(x) = Nrm(y) = a2 + b2.
But this is equivalent to p ≡ 1 (mod4) and√ contradicts the assumption of (iii). We still need to prove that any p ∈ Z[ −1] is associated to one (and only one) of the forms (i)–(iii). To do this, assume that Nrm(p) factors as
Nrm(p) = pp¯ = p1 ··· pn, with the pi’s prime √ in Z[ −1]. From this we see that p|p1 ··· pn and so p|pi for some 1 ≤ i ≤ n. This means that 2 2 Nrm(p)|Nrm(pi) ⇐⇒ Nrm(p)|pi , so either Nrm(p) = pi or Nrm(p) = pi . √ 2 2 In the first case we get that p =√a + b −1 with a + b = pi so is on the form (ii), or if pi = 2, p is associated with 1+ −1, i.e., (i). In the second case, since p|pi ⇔ pi = pa we get that √ 2 pi = Nrm(pi) = Nrm(p)Nrm(a) and so Nrm(a) = 1 ⇒ a ∈ U(Z[ −1]).
This would then mean that pi√≡ 3 (mod4)√since otherwise pi = 2 or pi ≡ 1 (mod4) which would mean that pi = (a + b −1)(a − b −1) and this is impossible when p is a prime, so we must have that pi ≡ 3 (mod4). 20 D. LARSSON
√ FIGURE 1. The geometry of Z[ −1]
√ Corollary 4.29. Every rational prime p ∈ Z decomposes in Z[ −1] into primes as √ √ p = (a + b −1)(a − b −1) if p ≡ 1 (mod4), √ √ (1+ −1)2 or remains prime in Z[ −1] if p ≡ 3 (mod4). If p = 2 then we see that 2 = √ so √ −1 is associated to the square of (1 + −1). See Figure 1 for the geometric interpretation of this result.
4.4. Polynomials. We could define rings of polynomials on any level of abstraction. It can be well argued that the more abstract the more general, but for the purpose of number theory this is a bit shooting over the target. For this reason, we will take a more hands-on approach, sacrificing absolute rigor and utmost generality in favor of real mathematics. So we make the following definition. Definition 4.14. Let A be a ring. Then a polynomial over A is a formal sum 2 n a0 + a1z + a2z + ··· + anz , with ai ∈ A, 0 ≤ n < ∞. The set of all polynomials in an indeterminate z is denoted by A[z]. We want to make this into a ring. This we do with the following definitions. Let p(z),q(z) ∈ A[z]. Then we define, m p(z) + q(z) = (p0 + p1z + ··· + pnzn) + (q0 + q1z + ··· + qmz ) =
= (p0 + q0) + (p1 + q1)z + ··· , and m p(z)q(z) = (p0 + p1z + ··· + pnzn)(q0 + q1z + ··· + qmz ) = n+m n+m k k k = ∑ ∑ piq j z = ∑ ∑ pk− jq j z . k=0 i+ j=k k=0 j=0 Notice that, if we want A[z] to be a (commutative) ring, distributivity forces us to have the 0 above multiplication. We also make the convention that z = 1A[z] = 1A. In this way A becomes a subring of A[z]. Theorem 4.30. The above definitions endow A[z] with the structure of a commutative ring with unity, and A can in a natural way be considered a subring of A[z].
Proof. Exercise! 21
It would be a serious negligence of duty if I didn’t mention the ring of formal power series over A, denoted A[[z]], at this point (although we will probably don’t need it). This is defined to be the set of elements of the form ∞ i 2 n ∑ aiz = a0 + a1z + a2z + ··· + anz + ··· , ai ∈ A, i=0 with the same addition and multiplication as above. However, we need to observe that the sum giving the coefficient of zi is always a finite sum, so the definition makes sense. Also, we disregard all questions of convergence. This is indeed a formal construction.
4.4.1. Iterated polynomial rings. Since A was only supposed to be a ring and A[z] was a ring we can consider the iterate A[z][w]. This is obviously then the ring of polynomials
2 n z0 + z1w + z2w + ··· + znw , with zi ∈ A[z], 0 ≤ n < ∞. Convince yourselves that we have the equality A[z][w] = A[z,w], that is, any p(w) ∈ A[z][w] can be written as i j ∑ pi jz w , with pi j ∈ A, 0 ≤ i, j < ∞. i, j
In this way we can continue to iteratively construct polynomial rings A[z] = A[z1,z2,...,zm], in several indeterminates, z := {z1,z2,...,zm} (think through this!). From now on we will however focus our attention on the one-determinate case.
4.4.2. The degree-function on A[z]. There is a degree-function deg : A[z] → N0 defined by n deg p(z) = deg(p0 + p1z + ··· + pnz ) := n, and deg(0) = ∞. Observe that some author’s define deg(0) = −∞. The difference is a matter of taste more than anything. We note one immediate consequence of deg:
Theorem 4.31. If D is a domain then deg(pq) = deg p + degq and the ring D[z] is also a domain.
Proof. The first statement follows from the definition of a product of two polynomials, observing that the highest coefficient is not zero, since pdeg p and qdegq are non-zero (and not zero-divisors). Assume 0 6= p,q ∈ D[z] with pq = 0. Then
∞ = deg(0) = deg(pq) = deg p + degq.
For this last sum to be ∞ either deg p or degq (or both) must be ∞ and the only polynomial satisfying this is the zero polynomial. Hence p or q (or both) is zero.
A polynomial f is called irreducible if f = gh implies that either g or h is a unit. If it is not irreducible, it is said to be reducible. We have the following useful criterion:
4.4.3. The Eisenstein criterion.
n Theorem 4.32. Let f (z) ∈ Z[z], f (z) = f0 + f1z + ··· + fnz . Suppose there is a prime 2 2 p ∈ Z such that p - fn, p|ai for 0 ≤ i ≤ n − 1, and p - f0 . Then f (z) is irreducible over Q. 22 D. LARSSON
4.4.4. D[z] is Euclidean. The following theorem is well-known to you, at least when for- mulated differently.
Theorem 4.33. The ring F[z], for F a field (i.e., a domain where all non-zero elements are invertible), is a Euclidean domain with deg as Euclidean valuation.
Proof. We know from high school that there is a division algorithm on F[z] (at least when F is the field of real numbers R). That deg(pq) ≥ deg(p) is obvious for p,q 6= 0. This result holds for more general coefficient rings (other than fields) but then the degree function is not the right choice for an Euclidean valuation. As a result we have the first statement of the following theorem. Theorem 4.34. Let D[z] be a polynomial ring over a domain D. Then, (i) if D is a field D[z] is a PID and hence a UFD; (ii) if D is a UFD, D[z] is a UFD. Proof. The proof of the first statement follows from Theorems 4.33 and 4.25. For the proof of the second I refer you to more specialized abstract algebra literature. 4.4.5. D[z,w] is not Euclidean. The reason is simply because it is not a PID. There are ideals in D[z,w] which cannot be generated by one element. One example: i := hz,wi.
4.4.6. Reduction modulo ideals in A. We have seen how to define quotient rings A/i mod- ulo an ideal i. Given a polynomial p ∈ A[z] there is a ring homomorphism A[z] → (A/i)[z], “reduction mod i”, where we reduce the coefficients modulo i.
Example 4.11. Let 3+5z−6z2 +z3 ∈ Z[z]. Reducing this polynomial mod h5i ⊆ Z yields the polynomial 3 + 4z2 + z3 ∈ Z/h5i[z]. This is a useful technique in number theory. In fact we have the following useful result:
Theorem 4.35. Let f (z) ∈ Z[z]. If the reduction mod n, f¯(z) ∈ Z/hni[z], of f is irreducible, where deg( f¯) = deg( f ), then f is irreducible over Z. Proof. Suppose f is reducible: f = gh. Then the ring morphism Z → Z/hni yields a ring morphism Z[z] → Z/hni[z] mapping f → f¯. Therefore, f¯ = gh = g¯h¯, and so, since deg( f¯) = deg( f ), deg(g¯) = deg(g) and deg(h¯) = deg(h) and so f¯ is also reducible.
5. FIELDS Now we begin to approach the heart of the subject, namely algebraic number fields. But first some more generalities.
5.1. Definition and examples.
Definition 5.1. A field F is an integral domain where every non-zero element is invertible. We will often denote fields by ’blackboard’ letters, F,E,J,L, etc. The one exception, and the example most important to us, is algebraic field extensions. Traditionally, these are written in an ordinary roman (italic) font. Example 5.1. The following are examples of fields: - Q, R and C; - Fp := Z/hpi, where p is prime; 23 √ √ - Q( d) := {a + b d | a,b ∈ Q}, for d a square-free integer,√ e.g., d = −1,±2. Convince yourselves that this is indeed a field. Why is C( d) not an interesting creature? Also, why do we insist that d is square-free? - F(z) := { f /g | f ,g ∈ F[z], g 6= 0}, the field of rational functions over F. We will see more examples in a little while.
6. BASICS @FIELD
Let R be a ring. Then there is a ring morphism Z → R sending n 7→ n1 = 1+1+···+1 (n times). The kernel of this morphism is an ideal of Z, and hence is on the form nZ, where we allow for the possibility n = 0. We can also assume that n is not negative (why?). Then there is an injective ring morphism Z/nZ ,→ R. If n = 0, we interpret Z/nZ as Z. This means that there is an isomorphism between Z/nZ and a subring of R. This number n is the characteristic of R, as defined before. We will from now on only consider fields. So we replace R with a field F. Notice that if n = 0 then Z is a subring of F. But every 0 6= z ∈ Z maps to z1 ∈ F and F being a field, shows that z1 is invertible. Hence F includes Q as a subfield. Suppose n 6= 0. Let ab ∈ hni = nZ = ker(Z → F). Then ab1 = (a1)(b1) = 0. Since F is a field, and hence an integral domain, either a1 = 0 or b1 = 0. But this means that a ∈ hni or b ∈ hni and so hni is a prime ideal which is equivalent to n being a prime. Hence we have shown,
Theorem 6.1. For every field F there are two (mutually exclusive) possibilities: (i) Either F includes the field of rational numbers Q, the case of zero characteristic, or (ii) the field Fp, the case of positive characteristic. Theorem 6.2. Let E and F be two fields. (i) There is only one, proper, ideal of any field, namely (0). (ii) If φ : F → E is a ring morphism, then it is injective. Proof. The first statement follows since 1 belongs to every non-zero ideal and hence are the whole field. The second statement follows from the first since kerφ ⊆ F is an ideal so must either be the whole field, in which case φ = 0, or kerφ = (0) and this is, by definition, the same as saying that φ is injective.
Notice that this makes F into a subfield of E, and, conversely, any subfield of a field is given by an injection, namely the inclusion morphism. Theorem 6.3. Let R be a ring (not necessarily a domain). Then (i) the ideal p is prime if and only if R/p is an integral domain; (ii) the ideal m is maximal if and only if R/m is a field. Proof. We prove the statements separately. (i) Suppose p is a prime ideal, and let a,b ∈ R/p such that ab ∈ p, which is equivalent to they being zero in R/p. We want to show that a or b is zero, i.e., that a ∈ p or b ∈ p. But this follows immediately since p is a prime ideal. Conversely, suppose R/p is a domain. Then ab = 0 ⇔ ab ∈ p means that either a or b is zero, i.e., either a ∈ p or b ∈ p. 24 D. LARSSON
(ii) Suppose m is maximal. Since any maximal ideal is prime, R/m is an integral do- main by (ii). Take 0 6= a ∈ R/m. We want to show that a has an inverse. That a 6= 0 is equivalent to a ∈/ m. This means that hai + m = R so there are α,β ∈ R such that αa + βm = 1 for some m ∈ m. But, modulo m (i.e., in the reduction R/m), βm is zero. Hence α is the required inverse to a. For the other implication, suppose that m ⊂ a ⊆ R (ideal inclusion). Then there is an a ∈ a, a ∈/ m. Hence m ⊆ hai + m ⊆ a ⊆ R. Since R/m is a field there is b ∈ R such that ab = 1 modulo m, or, equivalently ab + cm = 1 for some c ∈ R.This element ab + cm is in hai + m and so 1 ∈ hai + m, showing that hai + m = a = R. 6.1. Fields of fractions. To any domain D one can associate a field, called the field of fractions of D. This is done by formally inverting all non-zero elements of D. Theorem 6.4. Let D be an integral domain. Then there is a field Frac(D) and a ring injection D ,→ Frac(D) making D into a subring of Frac(D). Proof. The proof of the theorem is constructive. Let D∗ be the multiplicatively closed (i.e., 1 ∈ D∗ and a,b ∈ D∗ ⇒ ab ∈ D∗) set of non-zero elements of D and consider the direct product D × D∗ = {(a,s) | a ∈ D,s ∈ D∗}. We equip this set with the following relation: (a,s) ∼ (b,t) ⇐⇒ at = bs. This is an equivalence relation. Indeed, the only non-trivial point is transitivity: (a,s) ∼ (b,t) and (b,t) ∼ (c,u) means at = bs and bu = ct. Multiply the first equation by u and the second by s leads to atu = bus and bus = cts. Putting these together leads to atu = cts and tu,ts ∈ D∗ so (au − cs)t = 0. Since D is a domain and t ∈ D∗, we see that au = cs and so (a,s) ∼ (c,u). Notice that (a,s) ∼ (at,st) for all t ∈ D∗. To show that Frac(D) := D × D∗/ ∼ is a field we need to have suitable definitions of addition and multiplication. So define, (a,s) + (b,t) := (at + bs,st), and (a,s)(b,t) := (ab,st). We have to show that these indeed define a ring structure on Frac(D) in such a way that it is a domain and every non-zero element is invertible. Group structure. We first show that Frac(D) is an abelian group with the above defined addition. (Unit) Define the zero element as (0,1). We immediately see that (0,1)+(a,s) = (a,s)+ (0,1). (Ass.) This follows from the following computation: (a,s) + ((b,t) + (c,u)) = (a,s) + (bu + ct,tu) = (atu + sbu + sct,stu) = = (at + sb,st) + (c,u) = ((a,s) + (b,t)) + (c,u). (Inv.) The invers to (a,s) is (−a,s). It is obvious that this satisfies the required properties. Hence Frac(D) is an abelian group under addition. Ring structure. Similarly we have to check the multiplication. (One) The unity is defined by (1,1). Once again it is immediate that this is indeed a unity. (Ass.) This follows from the following trivial observation: (a,s)((b,t)(c,u)) = (a,s)(bc,tu) = (abc,stu) = (ab,st)(c,u) = = ((a,s)(b,t))(c,u). 25
(Dist.) Finally, this follows from the following, (a,s)((b,t) + (c,u)) = (a,s)(bu + ct,tu) = (abus + acts,s2tu) = = (ab,st) + (ac,su) = (a,s)(b,t) + (a,s)(c,u) and similarly we show the right-handed version. Suppose (a,s) 6= (0,1). Then (a,s)−1 = (s,a) and (0,1) = (a,s)(b,u) = (ab,su) implies that either a or b is zero since D is a domain. Hence, we finally conclude that Frac(D) is a field. To end the proof we need to show that D can be considered a subring of Frac(D). Define ι : D → Frac(D) by ι(a) := (a,1). This is a ring morphism: ι(a + b) = (a + b,1) = (a1 + b1,1 · 1) = (a,1) + (b,1) = ι(a) + ι(b) and ι(ab) = (ab,1) = (a,1)(b,1) = ι(a)ι(b). In addition ι is injective since kerι = h0i. This completes the proof. a −1 Definition 6.1. Write (a,s) as a/s, s or as . We have just shown that Frac(D), with structure given by (in the new notation) a b at + bs a b ab + := , and := s t st s t st is a field, called the field of fractions (sometimes field of quotients) of D. The above construction is actually just a special case of a more general theory called localization. However, for our purposes the above is more than sufficient.
7. FIELDEXTENSIONS This section is the technical heart of this short expose´ on fields.
7.1. Field extensions.
Definition 7.1. An inclusion of fields F ⊆ E is called a field extension, and denoted E/F. Notice that E is a vector space over F. The dimension dimF(E) is called the degree of the extension, and is denoted [E/F]. The extension is finite if [E/F] < ∞. The definition implies that every z ∈ E can be written uniquely as a sum
z = α1e1 + α2e2 + ··· + αnen + ··· , αi ∈ F, ei ∈ E, and where all the ei’s are linearly independent over F. We will primarily be concerned with finite extensions so in this case every z can be written as a finite sum
z = α1e1 + α2e2 + ··· + αnen αi ∈ F, ei ∈ E. Despite the conflicting notation with quotient rings (and groups) there is in practice never any risk of confusion: fields are trivially ideals.
7.1.1. The tower law. One often encounters a sequence of field extensions F ⊆ E ⊆ J of a field F. We then have the following theorem. Theorem 7.1. Suppose F ⊆ E ⊆ J is a sequence of finite field extensions of F. Then [J/F] = [J/E][E/F]. 0 0 In fact, if {ei} is a basis for E/F and {ei} is a basis for J/E, then {eie j} is a basis for J/F. 26 D. LARSSON
Proof. The proof of this theorem is rather simple. Take z ∈ J/E, so that 0 z = ∑α je j, α j ∈ E. j 0 By assumption, α j = ∑i β jiei, for β ji ∈ F. This means that z = ∑i, j β jieie j. Now we only 0 need to show that {eie j} are linearly independent. For this assume that 0 0 ∑γ jieie j = ∑ ∑γ jiei e j = 0, for some γ ji ∈ F. i, j j i 0 Since {e j} is a basis for J/E, we must have ∑i γ jiei = 0, and similarly, since {ei} is a basis for E/F, we get γ ji = 0. 7.2. Algebraic extensions, transcendental extensions. Definition 7.2. An element α ∈ E/F is called algebraic over F if there is a polynomial F ∈ F[z] such that F(α) = 0.
We can define a ring morphism evα : F[z] → E by evα (F) := F(α). Two possibilities occur: - If α is algebraic, the kernel is a non-trivial, proper, ideal of F[z]. Since F[z] is a PID, this ideal is generated by a single polynomial irr, i.e., ker(evα ) = hirri (the reason for the weird notation will be apparent soon). - On the other hand, if α is not algebraic, the kernel is zero, in which case α is called transcendental over F. √ √ Example 7.1. The real number√ 2 is algebraic over Q since 2 is a solution to the equa- tion z2 − 2 = 0. Notice that 2 ∈/ Q. The real number π is transcendental over Q since there is no polynomial with coefficients in Q which have π as a zero (this was proved by Lindemann in 1882). Note. The concepts of algebraic and transcendental numbers are very much dependent on over which field we work. For instance, e is transcendental over Q (proved by Hermite in 1873) but it is algebraic over R since it is a zero of the polynomial z − e ∈ R[z]. Let α be an algebraic element of E/F. By the first isomorphism theorem we have the following isomorphism:
F[z]/hPi ' F[α] = im(evα ), for some P ∈ F[z]. Since F[α] is an integral domain hPi must be a prime ideal, and so, since F[z] is a PID, P is an irreducible polynomial. Normalizing P by multiplying it by the inverse of the lead- ing coefficient we get an irreducible monic (i.e., having leading coefficient 1) polynomial irr and it is clear that irr is the unique polynomial of least degree generating hPi. In this respect, the monic irr is uniquely given by α, and is called the minimal polynomial asso- ciated with α. We denote this as irr(α,F) or simply irrα or irr if F, α and F, respectively, are obvious from the context. In addition, irr(α,F) divides every polynomial for which α is zero. Indeed, suppose L(α) = 0. Then by assumption on irr, deg(L) ≥ deg(irr) and so by the division algorithm for polynomials, L(z) = q(z)irr(z) + r(z) for some r ∈ F[z], deg(r) < deg(irr). But since L(α) = irr(α) = 0, we must have r(α) = 0, which, by as- sumption on irr, implies that r(z) = 0, and so irr|L. We have proved the following: Theorem 7.2. The minimial polynomial to irr(α,F) is the unique monic polynomial of minimal degree having α as a zero. Any other such polynomial with α as a zero is a multiple of irr(α,F). 27
Definition 7.3. An extension E/F is called algebraic if all e ∈ E are algebraic over F.
Theorem 7.3. Every finite extension E/F is algebraic.
Proof. The powers of 0 6= α ∈ E cannot be linearly independent since the extension is finite.
7.3. Simple and finitely generated extensions.
Theorem 7.4. Let α ∈ E/F be algebraic. Then F[α], the polynomial ring generated by n− α over F, is a subfield of E. Denote this subfield by F(α). The set {1,α,α2,...,α 1}, n := deg(irr(α,F)), is a basis for F(α) over F.
Proof. We know that F[α] ' F[z]/hirri, where irr = irr(α,F) is the minimal polynomial associated to α. That F[α] is a subring of E is clear. Since F[z] is a PID, every prime ideal is maximal, and so F[α] ' F[z]/hirri is a field. We need to show that every a ∈ F(α) can be written uniquely on the form
2 n−1 a = a01 + a1α + a2α + ··· + an−1α , with ai ∈ F.
That every a ∈ F(α) can be written on the above form follows since a, by the initial defi- nition, can be represented by a polynomial F in α. If deg(F) ≥ deg(irr) we can reduce it modulo irr by the division algorithm and represent F by the remainder. This remainder is unique and has degree less than n and so the theorem is proved.
Combining Theorems 7.3 and 7.4 we get
Theorem 7.5. An element α ∈ E/F is algebraic if and only if F(α)/F is a finite (algebraic) extension.
Theorem 7.6. If α,β ∈ E/F have the same minimal polynomial, then F(α) ' F(β).
Proof. This actually follows from the isomorphism F[α] ' F[z]/hirri, but it can be nice to have an explicitly given isomorphism. So, we define a morphism by
n−1 n−1 φ : a01 + a1α + ··· + an−1α 7→ a01 + a1β + ··· + bn−1β .
Clearly, φ is one-to-one onto and a morphism of groups, i.e., φ(x+y) = φ(x)+φ(y) (check this!). We need to show that it is also multiplicative: φ(xy) = φ(x)φ(y). This follows since x, y and xy can be written as polynomials in α.
Now, take for simplicity (this is not necessary) F = Q, and note that this construction can be iterated, adding more algebraic elements to a field. Indeed, suppose Q(α)/Q is a simple extension and that Q(α) is a subfield of some larger field (it is always a subfield of C in any case so this is not a problem). Take an algebraic element β ∈ C over Q(α). Then Q(α,β) := Q(α)[β] (polynomials in β over Q(α)) is a field by the same reasoning as before. This is called the field obtained by adjoining β to Q(α) and similarly, Q(α) is the field obtained by adjoining α to Q.
Definition 7.4. The field extension Q(α1,...,αn)/Q is called the finitely generated field extension generated by α1,...,αn ∈ C. 28 D. LARSSON
7.4. Algebraic closure.
Definition 7.5. A field F is algebraically closed if every F ∈ F[z], deg(F) ≥ 1, has a zero α ∈ F. An equivalent definition is: The field F is algebraically closed if for any algebraic ex- tension E/F, we have E = F. This is not entirely obvious since one needs to know that every polynomial over a field has zeros in some field extension (which is true).
Example 7.2. The field C is algebraically closed. This follows from the fundamental theorem of algebra: every non-zero polynomial over C has a zero. Definition 7.6. Let F be a field. Then a field E such that F ⊆ E is called an algebraic closure of F if E/F is algebraic and E algebraically closed. Theorem 7.7. Every field F has an algebraic closure and any two such are isomorphic. This theorem enables us to speak of the algebraic closure of F. This is usually denoted F¯ or Falg. The proof of this theorem is rather involved and deep, using in an essential way the axiom of choice or Zorn’s lemma. I refer to more advanced algebra texts for this.
Example 7.3. The field of rationals Q has an algebraic closure Qalg. This is the set of all possible complex zeros of polynomials in Q[z]. It is a fact (that I won’t prove) that 8 Qalg (C. In fact, Qalg is countable whereas C is uncountable . The extension Qalg/Q is one of the most (maybe the most) complicated objects in mathematics and much of the research done in number theory, algebra and geometry, directly or indirectly, aims at understanding the properties of this extension.
Example 7.4. The set Zalg ⊆ Qalg of elements which are complex roots of monic integer polynomials, i.e., monic polynomials in Z[z], is called the set of algebraic integers of Qalg. This will be generalized soon to more general number-theoretic situations.
Theorem 7.8. The algebraic closure Qalg of Q is a field. Proof. By the tower law and Theorem 7.5 [Q(α,β)/Q] = [Q(α,β)/Q(α)][Q(α)/Q] < ∞ and so Q(α,β)/Q is an algebraic extension. This implies that Q(α + β)/Q, Q(−α)/Q − and Q(α 1)/Q, α 6= 0, are all algebraic extensions since α + β, αβ and α/β are all in alg Q(α,β). This clearly implies that Q is a field. Notice that [Qalg/Q] = ∞, but Qalg is (by definition) algebraic over Q. This shows that the implication “algebraic ⇒ finite” is not true in general. It is, however, true for simple extensions as we have seen (and more generally, finitely generated extensions).
8. FINITE FIELDS Let f (z) be a polynomial over a field K (not necessarily finite). Then L ⊇ K is a splitting field for f (z) if f (z) splits into linear factors over L. In fact, a splitting field is gotten by adjoining to K all zeros of f (z) in some algebraic closure Kalg. Splitting fields are unique up to isomorphism (i.e., any two splitting fields are isomorphic).
8 alg If this means nothing to you, don’t worry. You only need to know that C is “infinitely bigger” than Q . 29
8.1. The main theorem. We know that Fp := Z/pZ are fields with p elements for p ∈ Spec(Z). Let F be a finite field. Since F is a ring there is a canonical ring morphism Z → F sending 1Z 7→ 1F. Clearly, this morphism cannot be injective since Z is infinte. Therefore it has a non-zero kernel which has to be a maximal ideal in Z generated by a prime p (recall that Z is a PID). From this follows (by the first isomorphism theorem) that there is an injection Fp = Z/pZ ,→ F, so F ⊇ Fp is a finite field extension of finite fields. Hence, n F is finite-dimensional as a vector space over Fp and so #F = p for some n ≥ 1. We have proven the first part of the following theorem. Theorem 8.1 (Main theorem on finite fields). The number of elements in any finite field is a power of a prime p. Every finite field F is the splitting field of the polynomial pn f (z) := z − z ∈ Fp[z] and the elements of F are exactly the zeros of f (z). Therefore, there is only one finite field of order pn up to isomorphism. × Proof. The multiplicative group F of F is cyclic with q − 1 elements where we have put n × q := p . Therefore, by Fermat’s little theorem for instance, every a ∈ F is a solution to the q− q equation z 1 − 1 = 0, implying that every element in F is a solution to f (z) := z − z = 0. Hence f (z) has q distinct zeros (the elements of F) and so F is a splitting field for f (z) = ∏α∈F(z − α). So, assuming that a finite field of pn elements exists, it is the splitting field of the poly- n nomial f (z) = zp − z, and since splitting fields are unique up to isomorphism, every two finite fields of pn elements are isomorphic. n It remains to construct a finite field of p elements for every p ∈ Spec(Z) and n ≥ 1. We will do this by showing that the set of zeros of f (z) actually is a field (and so has n p elements in some algebraic closure of Fp). Take zeros α,β of f (z). Then, for p odd, α ±β is also a zero of f (z). This follows from the characteristic-p-version of the binomial theorem (a ± b)p = ap ± bp (check this!). If p = 2, then −α = α so this poses no problem either. Clearly, 0 and 1 are zeros; αβ is also a zero: n n n (αβ)p − αβ = α p β p − αβ = αβ − αβ = 0. Similarly, α−1 is a zero: n n (α−1)p − α−1 = (α p )−1 − α−1 = α−1 − α−1 = 0. n Since, f 0(z) = pnzp −1 − 1 = −1, the polynomial f (z) can have no multiple zeros. Hence, n there are p distinct zeros and these form a field. Let me remark that in constructing the finite field we could use any irreducible poly- nomial over Fp, not necessarily f (z) as defined in the theorem (although f (z) is actually not irreducible). Indeed, choose an irreducible polynomial g(z) and let n := deg(g). No- 2 tice that Fp is a subring of Fp[z] via a 7→ a + 0z + 0z + ··· (this is true for any ring R in R[z]). We know that g(z) generates a maximal ideal in Fp[z] since Fp[z] is a PID. Therefore Fp[z]/hg(z)i is a field and we have sequence of morphisms
Fp / Fp[z] / / Fp[z]/hg(z)i .
The composition Fp → Fp[z]/hg(z)i is a morphism of fields so is injective; hence, we n get a finite extension of Fp of dimension p . Now, since finite fields are unique up to isomorphism, this must be the same (isomorphic) field as the one constructed via f (z) in the theorem. The difference is simply that the multiplication rule has different appearance 30 D. LARSSON
(given in terms of f (z) in one case, and g(z) in the other), but the resulting fields are actually isomorphic.
Corollary 8.2. Any finite extension of finite fields Fpn /Fp is Galois. n 8.2. The Frobenius morphism. Put q := p , for some p ∈ Spec(Z) and n ≥ 1. Then Fq/Fp is a finite extension of finite fields. Then there is a morphism p Frobp : Fq → Fq, a 7→ a . This is easily checked to be a ring morphism and since Fq is a field it is injective, and since Fq is finite, it is surjective. Hence Frobp is an automorphism. This automorphism is called the Frobenius morphism of Fq/Fp, or simply, ”the Frobenius”. Also, Frobp |Fp = id, so Frobp ∈ Gal(Fq/Fp).
Theorem 8.3. Let Fq/Fp be a finite extension of finite fields. Then Gal(Fq/Fp) = hFrobpi. In particular, Fq/Fp is a cyclic Galois extension.
Proof. Let G := hFrobpi ⊆ Gal(Fq/Fp). We have n pn Frobp(α) = α = α, for all α ∈ Fq, n d so Frobp = id. Let 1 ≤ d ≤ n be the smallest integer such that Frobp = id. Then d pd Frobp(α) = α = α for all α ∈ Fq. pd d Hence every α ∈ Fq is a zero of z − z. This polynomial has at most p zeros so d ≥ n from which we see that d = n. Therefore, G is cyclic of order n. But the number of automorphisms of a field extension is less than or equal to the degree, so we must have that G = Gal(Fq/Fp).
9. ALGEBRAIC NUMBER FIELDS 9.1. Algebraic numbers. Fact 9.1. Every irreducible polynomial P(z) over a subfield of C splits as
P(z) = (z − α1)(z − α2)···(z − αn), where all αi ∈ C are distinct. This fact is a special case of something called ’separability’: Definition 9.1. An element α ∈ E/F is called separable if α is not a multiple zero of irr(α,F). The algebraic extension E/F is called separable if all elements are separable. Theorem 9.2. Every algebraic field extension of characteristic zero is separable. Also, every finite extension of a finite field is separable.
Proof. See any respectable book in Galois theory. From this theorem the above fact follows since any subfield of C is of characteristic zero. Definition 9.2. An algebraic number field is a finite (and hence algebraic) extension K/Q in Qalg. Elements of K/Q are called algebraic numbers. We will often simply write K for the extension K/Q as we, from now on, mainly will consider extensions over Q. In cases where the field over which the extension takes place is not Q this will be explicitly stated. 31
9.1.1. Primitive element theorem.
Theorem 9.3. If K/Q is an algebraic number field, i.e., a field extension on the form alg Q(α1,...,αn)/Q, there is an algebraic element δ ∈ Q such that K/Q = Q(δ). The element δ is called a primitive element for K/F. Hence, as a result we may restrict our attention to simple extensions.
Proof. By induction it is clearly sufficient to consider only the case when K = Q(α,β), for some subfield Q of K over which α and β are algebraic. Let irrα := irr(α,Q) and irrβ := irr(β,Q) be the minimal polynomials of α and β. By the above fact, irrα and irrβ factorizes into distinct factors as
irrα (z) = (z − α)(z − α2)···(z − αn), and irrβ (z) = (z − β)(z − β2)···(z − βm).
We make the convention that α1 := α and β1 := β. Since all the αi’s and βi’s are distinct there is at most one d ∈ Q such that
αi + dβ j = α1 + dβ1, for every 2 ≤ i ≤ n, 2 ≤ j ≤ m.
(Check this!) This means that we can choose d such that αi + dβ j 6= α1 + dβ1 since in this case there are only finitely many d’s for which the equality holds. Put
δ := α + dβ = α1 + dβ1. Obviously Q(δ) ⊆ Q(α,β), and so we only need to show the reverse inclusion. In fact, it suffices to show that β ∈ Q(δ) because then α = δ − dβ ∈ Q(δ). We have that
irrα (δ − dβ) = irrα (α) = 0 so putting L(z) := irrα (δ − dz), we see that irrβ (β) = L(β) = 0. By the choice of d, these two polynomials have only one zero in common. Indeed, suppose ε is such that L(ε) = irrβ (ε) = 0. Then ε must be one of the βi’s and δ −dε one of the αi’s and so, by the choice of d, ε = β. Since irr(β,Q(δ)) (the minimal polynomial of β over Q(δ)) divides both L and irrβ and these two only have one zero in common, deg(irr(β,Q(δ))) = 1, say, 0 0 irr(β,Q(δ))(z) = z − β , for β ∈ Q(δ). Hence, 0 0 = irr(β,Q(δ))(β) = β − β ∈ Q(δ) 0 so β = β ∈ Q(δ) and the proof is finished. 9.2. Norms, traces and conjugates. 9.2.1. Field morphisms. Any morphism of algebraic number field φ : K → L, restricted to Q, is the identity φ|Q = idQ. This follows from the commutativity of the diagram
ιK / K Q sMM MMM MMM φ ιL MM MMM & L. Definition 9.3. A field morphism on K is a morphism of fields ϕ : K → C such that ϕ|Q = idQ. We denote the set of all field morphisms K → C by Mor(K). Theorem 9.4. Let K = Q(α) be an algebraic number field with irr(α,Q) the minimal poly- nomial of α. Then the number of field morphisms ϕ : K → C is [K/Q] = deg(irr(α,Q)). Moreover, ϕi(α) = αi, 2 ≤ i ≤ [K/Q], where αi are the other zeros of irr(α,Q). We put α1 := α. 32 D. LARSSON
Proof. Let αi be the zeros of the minimal polynomial of α, with α1 = α. Then ϕi(α) = αi alg defines field morphisms Q(α) ,→ Q ⊂ C, and 1 ≤ i ≤ [Q(α)/Q]. We have that Q(αi) ' Q(α) for all i by Theorem 7.6. Since all the αi’s are distinct the number of such morphisms is [Q(α)/Q]. Conversely, if ϕ is a field morphism ϕ : Q(α) ,→ C, then
0 = ϕ(irrα (α)) = irrα (ϕ(α)) so ϕ(α) must be one of the zeros to irrα , i.e., one of the αi’s and so ϕ is one of the ϕi’s.
Corollary 9.5. With notation as in the theorem, irr(α,Q) splits in C as irr(α,Q) = ∏ (z − ϕi(α)). ϕi∈Mor(K)
Definition 9.4. An algebraic number field K/Q = Q(α)/Q such that - imϕ ⊆ R for all ϕ ∈ Mor(K), is called totally real; - imϕ ⊆ C \ R for all ϕ ∈ Mor(K), is called totally imaginary; - K is a totally imaginary quadratic extension of a totally real field, is called a CM- field or complex multiplication field9. Examples of these notions will follow when we come to quadratic and cyclotomic num- ber fields. Proposition 9.6. The set of field morphisms can be decomposed (disjointly) into ℜ ℜ Mor(K) := {ϕ ∈ Mor(K) | imϕ ⊆ R}, r1 := #Mor(K) , ℑ ℑ Mor(K) := {ϕ ∈ Mor(K) | imϕ ⊆ C}, 2r2 := #Mor(K) and r1 + 2r2 = n. The pair (r1,r2) is called the signature of K/Q. Notice that for ϕ ∈ Mor(K)ℑ we might have imϕ ∩ R 6= /0. Proof. The only thing needing a proof is that the number of imaginary field morphisms is even. But this follows since for every ϕ ∈ Mor(K)ℑ we get a unique other by composing with complex conjugation. This means that there are r2 pairs (ϕ,ϕ¯).
Definition 9.5. The field polynomial of β ∈ K = Q(α) is defined by
[Q(α)/Q] Ψβ (z) := ∏ (z − ϕi(β)). i=1
The elements ϕi(β) are called the conjugates of β. An extension K/Q is called normal if all conjugates to α, i.e., all zeros of irrα , are elements in Q(α)/Q; a normal extension of Q is called a Galois extension (over Q). Definition 9.6. For fixed α ∈ E/F, we get a natural F-linear map α· : E → E, α ·v := αv. The norm of α is the determinant of α· and denoted NrmE/F(α) := det(α·). Similarly, the trace of α is the trace of α·, TrE/F(α) := Tr(α·).
9The reason for this strange notion is historical and is still extremely important to this day. It traces back to something called ”Kronecker’s Jugendtraum” and concerns how one can construct ”abelian extensions” of number fields using elliptic curves (or elliptic functions in the days of Kronecker). It would take me to afar, totally disregarding that my competence is severely lacking here, to explain this. But I encourage you do a Wikipedia search. 33
Remark 9.1. Authors tend to simplify notation whenever possible to avoid cumbersome and heavy notation, sometimes at the cost of absolute rigor, as you will no doubt find time and time again (if you haven’t already). I am no exception to this rule. As a consequence, I will often be rather careless not distinguishing between α as an element of E and its associated operator α· acting on E/F. Theorem 9.7. We have Tr(α + β) = Tr(α) + Tr(β) and Nrm(αβ) = Nrm(α)Nrm(β), so we get maps
TrE/F : E → F and NrmE/F : E \{0} → F \{0}, the first being linear and the second multiplicative.
Proof. This follows immediately from the definitions. Theorem 9.8. Let K/Q = Q(α)/Q be an algebraic field extension. Then - The field polynomial Ψβ (z) of β ∈ K/Q is equal to the characteristic polynomial Pβ (z). In fact, [K/Q(β)] Ψβ (z) = Pβ (z) = irr(β,Q) .
- TrK/Q(β) = ∑ϕi∈Mor(K) ϕi(β). - NrmK/Q(β) = ∏ϕi∈Mor(K) ϕi(β). In particular, Ψβ (z) ∈ Q[z]. Proof. We put for simplicity n := [K/Q(β)] (recall that K = Q(α)) and m := [Q(β)/Q]. Let m m−1 irr(β,Q) := z + km−1z + ··· + k1z + k0 m− be the minimal polynomial of β over Q. Hence, {1,β,β 2,...,β 1} is a basis for Q(β)/Q, so that, by the tower law, 2 m−1 2 m−1 e := {e1,...,enm} := { f11, f1β, f1β ,..., f1β ;...; fn1, fnβ, fnβ ,..., fnβ } is a basis for K/Q, given that f := { f1,..., fn} is a basis for K/Q(β). The matrix of β· in m−1 each block of basis vectors { f j1, f jβ,... f jβ } is given by a matrix 0 0 0 ··· 0 −k0 1 0 0 ··· 0 −k1 0 1 0 ··· 0 −k2 0 0 1 ··· 0 −k3 ...... ...... 0 0 0 ··· 1 −km−1 as is easily checked (do this!). The characteristic polynomial of each block is given by −z 0 0 ··· 0 −k0 1 −z 0 ··· 0 −k1 0 1 −z ··· 0 −k2 det 0 0 1 ··· 0 −k3 ...... ...... 0 0 0 ··· 1 −z − km−1 which expanded (at the last column, for instance) gives m m−1 z + km−1z + ··· + k1z + k0 = irr(β,Q), 34 D. LARSSON
n so Pβ (z) = irr(β,Q) since we have n such blocks. By the tower law, [Q(α)/Q] = [Q(α)/Q(β)][Q(β)/Q] = nm, we see that there are exactly n morphisms in Mor(Q(α)) restricting to the same morphism in Mor(Q(β)). This means that [Q(β)/Q] n [Q(α)/Q] Pβ (z) = ∏ (z − ϕi(β)) = ∏ (z − ϕi(β)) = ∏ (z − ϕi(β)) = Ψβ (z), i=1 ϕi∈Mor(Q(α)) i=1 where the first equality follows by Corollary (9.5). Now, from Pβ (z) = Ψβ (z) and the relation between the trace, determinant and characteristic polynomial, follows TrK/Q(β) = ∑ϕi∈Mor(K) ϕi(β) and NrmK/Q(β) = ∏ϕi∈Mor(K) ϕi(β). The proof is finished. 9.3. Algebraic integers and rings of integers. We now want to prove properties of the set of algebraic integers Zalg. For one thing we want to prove that it is a subring of the field of algebraic numbers Qalg. But number theorists are maybe more interested, in a first instance at least, to better understand other (smaller) number fields. This is the reason for introducing and working with the following definitions instead. Definition 9.7. Let D be a domain properly contained in a field F (for instance its field of fractions). Then α ∈ F is said to be integral over D if there is a monic polynomial (not necessarily irreducible!) with coefficients in D, having α as a zero. We denote by FD the set of elements of F integral over D. Notice that, since every a ∈ D is the solution of z−a = 0, D ⊆ FD. Hence FD is certainly non-empty.
Theorem 9.9. Let D be a subring of a field F (and so, in particular, a domain). Then FD is also a subring (domain) of F. We will prove this theorem in two steps. The first step is the following proposition.
Proposition 9.10. An element α ∈ F belongs to FD if and only if there is a finitely gener- ated D-submodule M of F such that αM ⊆ M. n−1 Proof. Suppose that α ∈ F is integral over D. Then a01 + a1α + ··· + an−1α = 0 n−1 for some ai ∈ D. The D-submodule M generated by {1,α,...,α } satisfies αM ⊆ M. Conversely, suppose that αM ⊆ M for some finitely generated D-submodule of F. Let m1,m2,...,mn be the generators. Then αmi = ∑ j ωi jm j, or re-written in matrix-form: (ω11 − α) ω12 ··· ω1n m1 0 ω21 (ω22 − α) ··· ω2n m2 0 = . . .. . . . . . . . . ωn1 ··· (ωnn − α) mn 0 Let Ω be the matrix on the coefficient matrix on the left. By Cramer’s rule we see that det(Ω)m j = 0 for all j. Since F is a field, and in particular a domain, and not all m j = 0, we must have det(Ω) = 0. Expanding this determinant yields a polynomial in α and so α is thus integral over D. The second step towards Theorem 9.9 is the following proposition. Proposition 9.11. Let α and β be two elements of F integral over D and M and N two finitely generated D-submodules of F such that αM ⊆ M and βN ⊆ N. Then, the set M · N = {mn | m ∈ M, n ∈ N} is a finitely generated D-submodule of F, invariant under multiplication by αβ and α ± β. 35
Proof. Clearly M · N is a D-submodule of F. Further, if e is a generating set for M and f a generating set for N, then e · f := {e1 f1,e1 f2,...,ei f j,...,en fm} is a generating set for M · N. Equally clear is it that (αβ)M · N ⊆ M · N and (α ± β)M · N ⊆ M · N by the corresponding properties of M and N. Proof of Theorem 9.9. Apply Proposition 9.10 to elements from Proposition 9.11. Integral closure.
Definition 9.8. Let D be a domain contained in a field F. Then the set of elements FD is called the integral closure of D in F. The integral closure of Z in a number field K/Q is called the ring of integers of K/Q and is denoted oK.
Definition 9.9. The domain D is called integrally closed if (Frac(D))D = D, that is, every element of Frac(D) integral over D, is already in D. We have the following nice result and eye-candy for a proof. Theorem 9.12. Every UFD is integrally closed. Proof. Let a/b ∈ Frac(D) be integral over D. Hence n n−1 (a/b) + pn−1(a/b) + ··· + p1(a/b) + p0 = 0, for pi ∈ D. Multiply this relation with bn, to get n n−1 n−2 2 n−1 (9.1) a + pn−1a b + pn−2a b + ··· + p1ab + p0b = 0. If b were a unit then a/b would already be in D, so suppose it is not. There is an irreducible element p being a factor in b but not in a (why?). This means that, p is a factor in every term of (9.1) except the first. However, this shows that p|an, and so p is a factor in a, contrary to assumption. Proposition 9.13. Let α ∈ F/Frac(D) be algebraic, where D is an integral domain. Then there is a d ∈ D such that dα ∈ FD. Proof. Since α is algebraic over Frac(D), we have that n n−1 α + an−1α + ··· + a1α + a0 = 0, with ai ∈ Frac(D).
This means that every ai is on the form ai = bi/ci, for bi,ci ∈ D, 1 ≤ i ≤ n − 1. Put n d := c0c1 ···cn−1, and multiply the above equation by d . This gives n 0 n−1 0 0 0 (αd) + bn−1(αd) + ··· + b1(αd) + b0 = 0, with bi := bic0 ···cbi ···cn−1 ∈ D, where cbi means that ci is omitted. The result follows. Theorem 9.14. Let E/Frac(D) be a finite (algebraic) extension of the field of fractions of D and assume that D is integrally closed. An element α ∈ E is integral over D if and only if irr(α,Frac(D)) ∈ D[z].
Proof. That irrα := irr(α,Frac(D)) ∈ D[z] ⇒ α integral over D, is obvious, so let us show the other implication, assuming that α is integral over D. Hence n n−1 α + pn−1α + ··· + p1α + p0 = 0, pi ∈ D.
Let β be any other zero of irrα . Then there is an Frac(D)-isomorphism ' L : Frac(D)(α) −→ Frac(D)(β) such that L(α) = β. Applying L to the equation above shows that β is also integral over D and so this is indeed the case for every Frac(D)-conjugate of α (i.e., every zero of irrα ). By 36 D. LARSSON a famous theorem of Newton, every coefficient of a polynomial P is itself a (symmetric) polynomial in the zeros of P. Hence, all the coefficients of irrα are integral over D by Theorem 9.9, and since D is integrally closed, these coefficients are in D.
Proposition 9.15. Let D be integrally closed and assume that E is a finite extension of Frac(D). If a ∈ E is integral over D, then TrE/Frac(D)(a),NrmE/Frac(D)(a) ∈ D.
Proof. If a is integral, then so is all of its conjugates. Now apply Theorem 9.8.
We have the following lemma:
Lemma 9.16. Let M be a free Z-module of rank n with basis e := {e1,...,en} and suppose ω := (ωi j) is a matrix with integer entries. Then f := ωe is a basis for M if and only if det(ω) = ±1.
Proof. Suppose that e and f = ωe are both basis sets. This implies that ω is invertible. So, det(ω)det(ω−1) = 1, and since det(ω) is an integer we must have that det(ω) = ±1 by (17.1). Conversely, suppose that det(ω) = ±1. Then det(ω) is invertible, and by (17.1) once again, ω−1 have only integer entries. From the fact that det(ω) 6= 0, we see that f is linearly independent (check this!) and so f = ωe is also a basis since f has the same cardinality as e.
The following theorem is a generalization of the theorem from group theory stating that a subgroup of a cyclic group is also cyclic. The proof is a bit tricky but since the result is important it is well worth the effort of understanding the details.
Theorem 9.17. Let M be a free Z-module and m ⊆ M a submodule. Then m is free of rank rk(m) ≤ rk(M). Furthermore, there is a basis e of M such that the basis of m can be choosen as {α1e1,...,αrk(m)erk(m)}, for αi ∈ N, where {e1,...,erk(M)} is a basis for M. Proof. The proof is by induction, the case rk(M) = 1 being the case of cyclic groups. Assume then that f is a basis for M. For every m ∈ m we have
T m = (α1,...,αn)f = α1 f1 + ··· + αn fn, where n := rk(M) and T denotes the transpose. If m = 0 the result is trivial so assume otherwise. This means that there is at least one m ∈ m such that at least one αi is different from zero. Let
S(m,f) := {α ∈ N | α is a coordinate for some m ∈ m}, and put ξ := ξm(f) := minf(S(m,f)). In words, ξ is the least positive integer, taken over all basis sets for M, such that ξ is a coordinate for an element of m. Now, choosing the basis e such that ξ is minimal, form the element
a := ξe1 + α2e2 + ··· + αnen is an element of m. By the division algorithm we have αi = qiξ + ri, 0 ≤ ri < ξ and 2 ≤ i ≤ n. Form the element
b := e1 + q2e2 + ··· + qnen 37 and consider the set g := {b,e2,...,en}. The matrix 1 q2 q3 ··· qn . 0 1 0 . ω := . . . .. 1 0 0 ··· 0 1 has determinant 1 and g = ωe so g is also a basis by the above Lemma 9.16. In the basis g the element a becomes a = ξb + r2e2 + ··· + rnen, but by the minimality over all basis sets of ξ and the fact that ri < ξ for all i, leads to ri = 0 and so a = ξb. Put hai := Za, the cyclic Z-submodule of M generated by a, and T n := {(0,β2,...,βn)g }.
Obviously, hai∩n = {0}. Take m ∈ m. We know that m = ε1b+ε2e2 +···+εnen. Reducing 0 0 0 ε1 modulo ξ via the division algorithm we get ε1 = q ξ + r , r < ξ. So, since a = ξb, we get 0 0 0 m − q ξb = m − q a = r b + ε2e2 + ··· + εnen. By the minimiality of ξ once more, we get r0 = 0 and so m − q0a ∈ n which implies that m ∈ n + hai. Hence m = n ⊕ hai. By the induction hypothesis n is a free Z-submodule of M, with rk(n) ≤ n − 1 and so m is a free Z-submodule with rk(m) ≤ n rank less than n. Theorem 9.18. Suppose M is a free Z-module with rank m, and m ⊆ M a submodule of rank n. Then M/m is finite if and only if m = n, in which case (M : m) = |M/m| = det(ω), where f = ∑ωe, for e a basis for M and f a basis for m. Proof. From the previous Theorem we see that m is free of rk(m) ≤ rk(M) and the first part of the theorem follows. If rk(m) = rk(M) then there are a1,...,an ∈ N such that 0 0 0 0 f = diag(a1,...,an)e for some basis sets f and e of m and M, respectively. Hence, we 0 0 0 0 get e = ωee, f = ω f f with det(ωe) = det(ω f ) = ±1 since e, e , f and f are all basis sets. Clearly, det(diag(a1,...,an)) = a1 ···an and this is the number of elements in M/m. Also, ω = ω f diag(a1,...,an)ωe, so det(ω) = a1 ···an and the proof is finished. Discriminants. We now need to be a little more general than previously. Recall how field extensions were defined. We now do the same construction for rings. Indeed, we could define a ring extension simply as a ring injection S ,→ R. Normally this is way to general to be of much use, and so also for us. Therefore, we define a ring extension as a ring injection S ,→ R such that R is a free A-module of rank r. This means that there is a basis e := {e1,...er} for R such that r R = Se1 ⊕ Se2 ⊕ ··· ⊕ Ser, that is, ∀r ∈ R, r = ∑ siei, si ∈ S. i=1 That R is an extension of S is denoted as R/S. Notice that, as in the case of field extensions, this notation can hardly be confused with quotient rings since S is (in general) only a subring of R, not an ideal. Now, we define a symmetric, non-degenerate, bilinear S-form (see 17.2.3) by
T(·,·) : R/S → S, (x,y) 7→ TrR/S(xy). 38 D. LARSSON
Definition 9.10. Let R/S be a ring extension as defined above. Then the discriminant of e is defined as Disc(e) := det(T(ei,e j)) = det(TrR/S(eie j)).
Continue to let e be a basis for R/S and form f := σ ·e, where σ = (σi j) is a matrix with n entries in S, that is, f j = ∑i=1 σ jiei. This changes the discriminant as 2 (9.2) Disc(f) = det((σi j)) Disc(e). Hence, Disc(e) is only unique up to multiplication of the square of a unit in S. On the other hand, the ideal that it generates, Disc(R/S) := hDisc(e)i, is well-defined. This is called the discriminant of R/S. Notice that (9.2) shows that if f is not a basis, then Disc(f) = 0. This discussion almost proves the following proposition.
Proposition 9.19. Let notation be as above. If Disc(R/S) 6= h0i then {g1,...,gr} is a basis for R/S as an S-module if and only if
Disc(R/S) = hDisc({g1,...,gr})i.
Proof. This follows from (9.2) and the fact that g := {g1,...,gr} is a basis if and only if det(ω) is a unit, where g = ωe. If S = Z, which will be our only concern here, then the discriminant is fully unique since the only square of a unit in Z is 1.
Corollary 9.20. Let S = Z. Then f = { f1,..., fm} generates a Z-submodule r ⊆ R of finite index if and only if Disc(f) 6= 0. In that case, 2 Disc(f) = (R : r) Disc(R/Z). Proof. Follows from the above Proposition and Theorem 9.18. Proposition 9.21. Let E/F be an extension of number fields (i.e., both E and F are exten- sions of Q and E is an extension of F) with F-basis f := { f1,..., fn}. Then 2 Disc({ f1,..., fn}) = det(ϕk( fi)) 6= 0, where ϕk ∈ Mor(E). Proof. The first part follows by direct computation:
Disc({ f1,..., fn}) = det(TrE/F( fi f j)) = det(∑ϕk( fi f j)) = det(∑ϕk( fi)ϕk( f j)) k k 2 = det(ϕk( fi))det(ϕk( f j)) = det(ϕk( fi)) .
If Disc(f) = 0, then the matrix ϕk( fi) is not-invertible. It is not hard to see that this is impossible when f is a basis. 9.4. Integral bases.
Definition 9.11. Let K/Q be an algebraic number field. Then a set {ω1,...,ωm} ⊂ K is called an integral basis for K/Q (or oK) if oK is a free Z-module of rank m with basis {ω1,...,ωm}, i.e., oK = Zω1 ⊕ Zω2 ⊕ ···Zωm, as a Z-module.
This definition means in particular that every a ∈ oK can be written uniquely as
a = a1ω1 + a2ω2 + ··· + amωm, with ai ∈ Z. 39
Rings of integers are finitely generated.
Theorem 9.22. Every algebraic number field K/Q has an integral basis, i.e., the ring of integers oK is a free Z-module. Furthermore, rk(oK) = [K/Q].
Proof. The idea of the proof is to wedge oK between two Z-modules of rank [K/Q]. Let ω := {ω1,...,ωk} be a basis for K/Q. By a previous theorem we can assume that ωi ∈ oK since otherwise we can replace it with dωi for some d ∈ Z. The bilinear map T(x,y) := TrK/Q(xy) is symmetric and non-degenerate. Therefore, by ∨ ∨ ∨ ∨ 17.3.1, there is a T-dual basis ω := {ω1 ,...,ωk } such that ωi (ω j) = δi j. We will show that ∨ ∨ ∨ Zω1 ⊕ Zω2 ⊕ ··· ⊕ Zωk ⊆ oK ⊆ Zω1 ⊕ Zω2 ⊕ ··· ⊕ Zωk . The first inclusion is obvious, so we only need to prove the second. ∨ ∨ Every a ∈ oK can be written as a = a1ω1 + ··· + akωk , with ai ∈ Q. We want to show that, in fact, ai ∈ Z. Since a,ωi ∈ oK, 1 ≤ i ≤ k, we have a · ωi ∈ oK as well. Therefore, T(a,ωi) ∈ Z as Z is integrally closed by Proposition 9.15. But, k ∨ T(a,ωi) = TrK/Q(a · ωi) = TrK/Q(∑ a jω j ωi) j=1 k k ∨ = ∑ a j TrK/Q(ω j ωi) = ∑ a j TrK/Q(δi j) = a j, j=1 j=1 implying that a j ∈ Z. The theorem now follows since we have that oK is a Z-submodule of a free Z-module of rank k and oK also includes a free Z-module of rank k. 9.5. Computing rings of integers. In general, it is quite difficult to compute the rings of integers of a given algebraic number field. There are a number of algorithms of various complexity implemented in several computer algebra programs, but going into that would take us too far. Hence we will have to be satisfied with resorting to some general tricks and folklore guesses. The main tool for us will be computing discriminants. Since discriminants are defined using determinants and determinants are notoriously hard to compute for larger matrices, we need some other means. We start this section with a result that is often helpful in this endeavor. 9.5.1. Computing discriminants.
Theorem 9.23. Let K/Q = Q(α)/Q be an algebraic number field, where α has minimial polynomial irr = irr(α,Q). Then n n−1 (2) Disc({1,α,...,α }) = (−1) NrmQ(α)/Q(∂irr(α)), dirr where ∂irr(z) := dz . Definition 9.12. Let f (z) be an arbitrary polynomial over a field F. Then the discriminant of f in α is n (2) Disc( f )(α) := (−1) NrmE/F(∂ f (α)), for α ∈ E/F, n := [E/F] < ∞. If E = F(α) then Disc( f ) := Disc( f )(α) is simply called the discriminant of f . Notice that the discriminant can be viewed as a polynomial, evaluated at α. This poly- nomial can also be called the discriminant. 40 D. LARSSON
Proof. Recall (see Proposition 9.21) that the discriminant of a set of elements {ω1,...,ωn} is given by j Disc({ω1,...,ωn}) := det(ϕ j(ωi)) = det(ωi ), j where ϕ j ∈ Mor(K) and ωi := ϕ j(ωi). From 17.7 follows n−1 i 2 i 2 2 (9.3) Disc({1,α,...,α }) = det(ϕ j(α )) = det(α j) = ∏(αi − α j) , i< j i i i where α j := ϕ j(α ) = (ϕ j(α)) . The minimal polynomial of α factorizes (in C, for in- n stance) as irr(z) = ∏ j=1(z − α j) and so differentiating gives n n n−1 ∂irr(z) = ∑ ∏(z − α j) =⇒ ∂irr(αi) = ∏(αi − α j). i=1 j=1 i=1 i6= j i6= j Therefore, n n n ∏∂irr(αi) = ∏∏(αi − α j). i=1 i=1 j=1 i6= j
In the left-hand side we recognize NrmQ(α)/Q(∂irr(α)); in the right-hand side each factor appears twice: once as (αi − α j) and once as (α j − αi). Hence, n n l 2 ∏∏(αi − α j) = (−1) ∏(αi − α j) . i=1 j=1 i< j i6= j n Lastly, a simple combinatorial argument shows that l = 2 . The result follows from this together with (9.3). n Theorem 9.24. Let the minimal polynomial over Q of α be irr(α,Q) = z + az + b. Then n−1 (n) n n−1 n−1 n−1 n Disc({1,α,...,α }) = Disc(irrα ) = (−1) 2 n b + (−1) (n − 1) a . n−1 n −1 Proof. Put θ := ∂irrα (α) = nα + a. Multiplying α + aα + b = 0 by nα we get nαn−1 = −na − nbα−1 and so −nb θ = −(n − 1)a − nbα−1 =⇒ α = . θ + (n − 1)a
From this we see that Q(α) = Q(θ) so deg(irr(θ,Q)) = deg(irrα ) = n. Writing −nb irr = g(z)/h(z), α z + (n − 1)a we see that g(θ)/h(θ) = irr(α) = 0 and so g(θ) = 0. I claim that this is the minimal n polynomial of θ. Indeed, using that irrα (z) = z + az + b we get that g(z) = (z + (n − 1)a)n − na(z + (n − 1)a)n−1 + (−1)nnnbn−1. Since this polynomial has degree n and is monic it has to be the minimal polynomial of θ. Therefore, the norm of θ is (−1)n times the degree-zero term in g(z), so
9.23 n n−1 (2) Disc({1,α,...,α }) = (−1) NrmQ(α)/Q(∂irrα (α)) = n n (2) (2) n n−1 n n n n−1 = (−1) NrmQ(α)/Q(θ) = (−1) (−1) (−(n − 1) a + (−1) n b ) = n n n− n− n− n = (−1)(2)(n b 1 + (−1) 1(n − 1) 1a ), 41 proving the theorem. 9.5.2. Discriminants and bases.
Theorem 9.25. Let K/Q be an algebraic number field with ring of integers oK. Suppose 2 that m ⊆ oK is a Z-submodule of oK with Z-basis { f1,..., fk}, k := [K/Q]. Then |oK/m| is a divisor of Disc({ f1,..., fk}). Proof. This is a direct consequence of Theorem 9.17 and Theorem 9.18 with M = oK.
Corollary 9.26. With notation as in the previous theorem, if Disc({ f1,..., fk}) is square- free, then { f1,..., fk} is a basis for oK. Theorem 9.27. Let K/Q be a number field with ring of integers oK. Suppose that [K/Q] M m = Z fi ( oK, (notice the strict inclusion), i=1 where rk(m) = rk(oK) = [K/Q]. Then there is an element d ∈ oK given by −1 (9.4) d := p (d1 f1 + d2 f2 + ··· + dk fk), where k := [K/Q], 0 ≤ di ≤ p−1, are rational integers (i.e., di ∈ Z for all i) and p a rational 2 prime, such that p |Disc({ f1,..., fk}). Proof. The assumption m ( oK implies that |oK/m| > 1. By Theorem 9.18 follows that r1 rk |oK/m| = p1 ··· pk and so there is a prime p := pi (1 ≤ i ≤ k) dividing |oK/m|. By the the- orem of Cauchy, or Sylow’s first theorem, there is a subgroup of oK/m of order p generated by one element d ∈ oK/m. Therefore, pd ∈ m. Let { f1,..., fk} be the basis for m. Then pd = ε1 f1 + ··· + εk fk. To finish the proof we need to show that εi can be chosen ≤ p − 1. By Theorem 9.17 a basis for m can be chosen such that mi = αiei, for αi ∈ N, 1 ≤ i ≤ k, and where {e1,...,ek} is a basis for oK. The element pd can be written, on the one hand, as pd = pid1e1 + ··· + pidkek, and on the other as, pd = ε1α1e1 + ··· + εkαkek. Suppose ε j ≥ p for some 1 ≤ j ≤ k. Reducing this we get ε j = q j p + r j for 0 ≤ r j < p. Then
ε1α1e1 + ··· + ε jα je j + ··· + εkαkek =
= ε1α1e1 + ··· + q j pα je j + r jα je j + ··· + εkαkek, and so
pd1e1 + ··· + pd je j + ··· + pdkek =
= ε1α1e1 + ··· + q j pα je j + r jα je j + ··· + εkαkek ∈ m which is equivalent to 0 pd := pd1e1 + ··· + p(d j − q jα j)e j + ··· + pdkek =
= ε1α1e1 + ··· + r jα je j + ··· + εkαkek ∈ m, finishing the proof. 9.5.3. An algorithm of sorts. The following is an informal algorithm to compute the ring of integers in a number field.
• Guess the basis B := {β1,...,βn}; • Compute Disc(B); • ∀p ∈ Spec(Z) : p2|Disc(B), is (9.4) an algebraic integer?; • If ’yes’ add this to B; • Repeat until no more integers are found. 42 D. LARSSON
9.5.4. Stickelberger’s theorem.
Theorem 9.28. Let K/Q be an algebraic number field with ring of integers oK, then Disc(oK/Z) ≡ 0 or 1 (mod4). The only proof I know of this theorem, although quite simple, involves Galois theory so I won’t include it here. For those who know some Galois theory might find the proof instructive and is encouraged to look it up in the literature.
9.6. Examples. I will show two examples here on how one can compute rings of integers. The examples are not mine10, so don’t be too impressed by the apparent ingenuity, it is certainly not mine (if I ever possessed it).
Example 9.1. Let Q(α)/Q be an algebraic field extension with α a zero (any) of the polynomial f (z) = z3 − z − 1. This polynomial is monic and irreducible over Q because if it factored we would have f (z) = (z − a)g(z), deg(g) = 2 and a ∈ Q. Any rational zero of f would have to divide 1 so a = ±1. But neither of these is a zero of f . The discriminant is given by Theorem 9.24 and is computed as
3 − − − Disc( f ) = Disc({1,α,α2}) = (−1)(2)(33 · (−1)3 1 + (−1)3 1(3 − 1)3 1(−1)3) = −23. 2 Since this is square-free we get that {1,α,α } is a basis for oQ(α), i.e., 2 oQ(α) = Z1 ⊕ Zα ⊕ Zα . Notice that this is true for which ever zero of f we take.
√3 Example 9.2. Let K/Q be Q( 2)/Q. We want to determine o := o √3 . √ Q( 2) 3 3 The minimal polynomial of α := 2 over Q is irrα := z − 2. This shows that there are three field morphisms
ϕ1 := id : α 7→ α
ϕ2 : α 7→ α2 := ωα 2 ϕ3 : α 7→ α3 := ω α, √ where ω := e2π −1/3. Recall that the field morphisms map the algebraic element to its conjugates, i.e., the other zeros of its minimial polynomial. We make an initial guess that the Z-basis for o is {1,α,α2}. Compute the discriminant: 2 1 α α2 2 2 2 2 Disc({1,α,α }) = det(ϕi(α j)) = det1 ωα ω α = 1 ω2α ωα2 1 1 1 2 6 2 2 2 2 3 3 = α det1 ω ω2 = 2 · 3 (ω − ω) = −2 · 3 (remember ω = 1). 1 ω2 ω Hence, by Theorem 9.27 we need to consider the possibilities: are 1 2 • θ := 2 (a11 + a2α + a3α ), 0 ≤ ai ≤ 1, or 0 1 2 • θ := 3 (b11 + b2α + b3α ), 0 ≤ bi ≤ 2,
10They are standard examples, appearing more or less in every book on algebraic number theory. 43 algebraic integers? We begin with θ. Taking the trace of θ gives us
3 Tr(θ) = ∑ fi(θ) = f1(θ) + f2(θ) + f3(θ) = i=1 1 1 = θ + (a 1 + a ωα + a ω2α2) + (a 1 + a ω2α + a ω4α2) = 2 1 2 3 2 1 2 3 3 1 1 3 = a 1 + a (1 + ω + ω2)α + a (1 + ω2 + ω)α2 = a 1. 2 1 2 2 2 3 2 1 3 Recall that if θ is an algebraic integer, then Tr(θ),Nrm(θ) ∈ Z. Hence, the trace 2 a1 has to be in Z so a1 ∈ 2Z if θ is to be an algebraic integer. 1 2 In that case, θ1 := 2 (a2α + a3α ) also have to be an algebraic integer since θ1 = θ − 1 2 a1. Now, take the norm of θ1: 3 3 3 3 2 Nrm(θ1) = 2 ∏ fi(θ1) = 2 f1(θ1) f2(θ1) f3(θ1) = i=1 2 2 2 2 = θ1(a2ωα + a3ω α )(a2ω α + a3ωα ) = 2 2 2 2 4 2 = (a2α + a3α )(a2ωα + a3ω α )(a2ω ω α ) = 3 3 2 = ω α (a2 + a3α)(a2 + a3ωα)(a2 + a3ω α) = 3 3 2 2 2 2 2 3 3 3 = α (a2 + a2a3(ω + ω + 1)α + a2a3ω(ω + ω + 1)α + a3ω α ) = 3 2 3 = (αa2) + (α a3) , and since α3 = 2, we get 1 a3 + 2a3 Nrm(θ ) = (2a3 + 4a3) = 2 3 ∈ . 1 23 2 3 4 Z
However, we demanded that 0 ≤ a2,a3 ≤ 1, so the above condition is clearly impossible unless a2 = a3 = 0. Therefore, there are no algebraic integers on the first form θ. 0 0 For θ we get Tr(θ ) = b11 and this is clearly an algebraic integer (since b1 ∈ Z). How- 0 0 0 ever, this doesn’t help us much since the difference θ1 := θ − b1 is no better than θ (in fact it is slightly worse). So we compute the norm of θ 0 directly:
1 Nrm(θ 0) = (b + b α + b α2)(b + b ωα + b ω2α2)(b + b ω2α + ω4α2) = 33 1 2 3 1 2 3 1 2 = [after a lot of calculating] = 1 1 = (b3 + b3α3 + b3α6) = (b3 + 2b3 + 4b3). 33 1 2 3 33 1 2 3 Contemplating this for a while, or doing some brute-force calculations, one realizes that this cannot be an integer for 0 ≤ b1,b2,b3 ≤ 2 unless b1 = b2 = b3 = 0. Therefore, we have shown that no algebraic integers exists that cannot be written in the basis {1,α,α2} and so 2 o √3 = Z1 ⊕ Zα ⊕ Zα . Q( 2) Notice that we could have simplified the computation of the discriminant considerably by using Theorem 9.24, but I wanted to show you the hard-core version also. 44 D. LARSSON
10. QUADRATIC NUMBER FIELDS Let d be a square-free integer, that is, if d = pr1 ··· prk then we must have r = ··· = √ 1 k 1 rk = 1.√ The algebraic number field Q( d) is called a quadratic number field. Clearly 2 α := d has minimial polynomial irrα (z)√ = z − d and so the extension√ Q(α)/Q is of degree two. Any element of Q(α) = Q( d) can be written as a + b d = a + bα, for a,b ∈ Q. Despite their innocent appearance, quadratic number fields have a deep theory and still harbors a lot of secrets. One could easily build a whole semester long course on quadratic number fields and still not get very far. Let me also, here in the beginning, mention that quadratic number fields have found applications outside mathematics, in cryptography (factorizing large integers for instance), information technology etc. The first natural question is: what are the rings of integers? 10.1. Ring of integers of quadratic number fields. √ Theorem 10.1. Let K/Q := Q( d)/Q be a quadratic number field. Then ( √ 1 ⊕ d, if d ≡ 2,3 (mod 4); Z Z √ oK = 1+ d Z1 ⊕ Z 2 , if d ≡ 1 (mod 4). (Notice that d ≡ 0 (mod 4) is not allowed since then d wouldn’t be square-free.) Proof. We have that Disc(z2 − d) = 4d by Theorem 9.24. Suppose first that d ≡ 2,3 (mod 4). By Stickelberger’s theorem we have that Disc(oK/Z) is either congruent to zero or one modulo four. By Corollary 9.20, we must have √ √ 2 Disc({1, d}) = Disc(irrα ) = (oK : Z[ d]) Disc(oK/Z), √ √ where Z[ d] denotes the Z-submodule of oK generated by the basis {1, d}. Now, if Disc(oK/Z) ≡ 1 (mod 4) then Disc(oK/Z) = 4k + 1 for some k ∈ Z. Hence, √ √ 2 4d = Disc({1, d}) = (oK : Z[ d]) (4k + 1). Since d is square-free and d ≡ 2,3 (mod 4), it is easy√ to see that we get a contradiction. 2 Therefore,√ Disc(oK/Z) ≡ 0 (mod 4) and so (oK : Z[ d]) = 1. Clearly, if the index is one, oK = Z[ d], and the first case is taken√ care of. For the other√ case, note first that√ {1, d} cannot√ be a basis, because then d 6≡ 1 (mod 4). Notice that Q( d) = Q((1 + d)/2) and (1 + d)/√2 is integral since it is a zero of 2 the polynomial z − z + (1 − d)/4. So Disc({1,(1 + d)/2}) = d. If Disc(oK/Z) ≡ 0 2 (mod 4) (Stickelberger) then 4n + 1 ≡ 4k(oK/Z) and this is impossible. On the other 2 hand, if Disc(oK/Z) ≡ 1 (mod 4) then 4n + 1 ≡ (4k + 1)(oK/Z) which is possible only if (oK/Z) = 1. 10.2. The Ramanujan–Nagell theorem. The topic of this subsection is a fine illustra- tion of the dogma that “elementary 6= easy”. I am going to prove the Ramanujan–Nagell theorem, a result that was conjectured by Srinivasa Ramanujan (1887–1920, self-taught math genius) and proved by Trygve Nagell (1895–1988)√ in 1948, using in an ingenious way the unique factorization of the ring of integers of Q( −7). Nagell was Norwegian but professor at Uppsala University from 1931 to his retirement 196211.
11See his obituary by famous British number-theorist John Cassels (born 1922) at http://www.numbertheory.org/obituaries/AA/nagell/ . 45
Theorem 10.2 (Ramanujan–Nagell). The only solutions to the equation ± z = 1, 3, 5, 11, 181 z2 + 7 = 2n in , are Z n = 3, 4, 5, 7, 15. Proof. Clearly, z must be odd, so assume this in addition to z > 0. If n is even we have: 2 n n n z + 7 = 2 ⇐⇒ (2 2 − z)(2 2 + z) = 7. n n This implies that 2 2 + z = 7 (or the other one, since 7 is a prime) and 2 2 − z = 1. Hence, n 2 · 2 2 = 8 =⇒ n = 4 and so, z = 3. Therefore, from now on we assume that n is odd and n > 3. Now, we have the following factorization into irreducibles √ √ 1 + −71 − −7 (10.1) 2 = . 2 2 Since z = 2k + 1, z2 + 7 is divisible by 4 and so we can re-write the original equation as z2 + 7 (10.2) = 2n−2. 4 Put m := n − 2. Then (10.1) and (10.2) can be combined to the factorization √ √ √ √ z + −7z − −7 1 + −7m1 − −7m (10.3) = . 2 2 2 2
Clearly, the√ right-hand side√ is a factorization into irreducibles (from (10.1)). Any common z+ −7 z− −7 √ factor of 2 and 2 must divide their difference, which is −7. Suppose √ √ 1 ± −7 √ √ 1 ± −7 −7 ⇐⇒ −7 = q. 2 2 Taking the norm of this, yields √ 1 ± −7 1 1 1 + 7 7 = Nrm Nrm(q) = + 7 Nrm(q) = Nrm(q) = 2Nrm(q), 2 22 22 4 √ 1± −7 √ which is obviously impossible. Therefore, 2 6 −7. Since neither does this, we deduce that there can be no common factors. Fact 10.3. The only units in o √ are ±1. Q( −7) √ √ √ √ 1+ −7 z− −7 z− −7 1+ −7 If 2 2 then 2 = 2 k for some k ∈ Z. But this is equivalent to √ √ √ z − −7 = k + k −7 ⇐⇒ z = k + (k + 1) −7 which is a rational integer only if k = −1, leading to z = −1, contrary to assumption. Hence, √ √ 1 + −7 z + −7 =⇒ 2 2 √ √ √ √ 1 ± −7m z ± −7 z ± −7 1 ± −7m ⇐⇒ = , 2 2 2 2 the last equivalence following from the above Fact 10.3. 46 D. LARSSON √ √ 1+ −7 1− −7 Put α := 2 and β := 2 . From the last equivalence above we get √ √ √ 1 + −7m 1 − −7m (10.4) ± −7 = − ⇐⇒ ±(α − β) = αm − β m. 2 2 We have α2 ≡ (1 − β)2 ≡ 1 (mod β 2) which follows since α + β = 1. From this we get m 2 m−1 2 α = α(α ) 2 ≡ α (mod β ) and so, assuming the sign positive in (10.4), α − β = αm − β m ≡ α − β m (mod β 2), implying that β ≡ β m (mod β 2) which is not the case. Hence the sign must be negative in (10.4). By expanding (10.4) with the binomial formula we deduce: m−1 m m m 2 m 3 m m−1 (10.5) −2 = − 7 + 7 − 7 + ··· ± 7 2 . 1 3 5 7 m Since all the terms on the right except the first one is divisible by 7, we have m (10.6) −2m−1 ≡ = m (mod 7). 1 Using that, 26 ≡ 1 (mod 7) it is quite easy to see that (by trying every possibility if neces- sary), modulo 42, the only solutions to (10.6) are m = 3, 5, 13 (mod 42). We shall show that m = 3, 5, 13 are the only remaining possibilities for solving the original problem. That 3, 5, 13 are solutions follows by construction or by simply checking. Suppose first that m ≡ 3 (mod 42). We will show that this is impossible unless m actually is 3. So, (10.7) m ≡ 42 ⇐⇒ m − 3 = 7q · 6 · k, for some k ∈ Z, 7 6 |k. We have, m m(m − 1)(m − 2)(m − 3) m − 4 7` = 7`. 2` + 1 (2` + 1)(2`)(2` − 1)(2` − 2) 2` − 3 Since 7`−1 > 2` − 1, we have that this is divisible by 7q+1, for ` > 1. Therefore, from (10.5) follows that 7 (10.8) −2m−1 ≡ m − m(m − 1)(m − 2)(mod 7q+1) 6 and so, −2m−1 ≡ −4 ≡ m − 7 (mod 7q+1) where the last congruence comes from combining (10.7) and (10.8). But this implies that m − 3 ≡ 0 (mod 7q+1), thereby contradicting the assumption on k. Hence m = 3. The next case m ≡ 5 (mod 42) follows with a similar method. Now, m m(m − 1)(m − 2)(m − 3)(m − 4)(m − 5) m − 6 7` = 7`, 2` + 1 (2` + 1)(2`)(2` − 1)(2` − 2)(2` − 3)(2` − 4) 2` − 5 leading to −25−1 ≡ −16 ≡ m − 70 + 49 (mod 7q+1), 47 implying that m − 5 ≡ 0 (mod 7q+1), a contradiction. q A little messier is the last case. Suppose m − 13 = 7 · 6 · k, for some k ∈ Z, 7 6 |k. Here, m m(m − 1)(m − 2)···(m − 13) m − 14 7` = 7`, 2` + 1 (2` + 1)(2`)(2` − 1)···(2` − 12) 2` − 13 and so, for ` > 6, since 7`−2 > 2` + 1, this is divisible by 7q+1· Hence,
− 213−1 ≡ −4096 ≡ m m m m m m ≡ m − 7 + 72 − 73 + 74 − 75 + 76 (mod 7q+1). 3 5 7 9 11 13 Replacing m in this by 13 + 7q · 6 · k in all terms divisible by 7 gives, −4096 ≡ m − 4109 (mod 7q+1) ⇐⇒ m − 13 ≡ 0 (mod 7q+1), a contradiction, thus finishing the proof.
11. DIRICHLET’SUNITTHEOREM 11.1. Roots of unity. Let K be a field and throughout below, n denotes an integer rel- atively prime to the characteristic (if this is different from zero). We recall some basic facts: (i) Any polynomial f ∈ K[z] of degree n has at most n zeros; n (ii) a cyclic group C is a group generated by one element, i.e., C = {a | n ∈ Z}; the n order of C is the least n ∈ Z>0 such that a = e (where e is the identity); every cyclic group is abelian, i.e., ab = ba ∈ C; (iii) a product of cyclic groups of relatively prime order is cyclic, i.e.,
Cn × Cm =∼ Cmn, gcd(m,n) = 1; Proposition 11.1. Every finite multiplicative group of a field K is cyclic. Proof. Let G ⊂ K× be a finite subgroup of K× = U(K). Clearly G is abelian. Choose r r a ∈ G with maximal period, i.e., ap = 1 and ap −i 6= 1 for all i < pr (it is easy to see that this maximal period has to be a power of a rational prime; indeed, it follows from r the Chinese remainder theorem), and consider the equation zp − 1 = 0. Every element of G is a solution to this equation. Consider the cyclic subgroup of G generated by a. This r has to be the whole G since otherwise the equation zp − 1 = 0 would have more than pr solutions. An mth root of unity is an element ζ ∈ K such that ζ m = 1. The set of m-roots of unity in K is denoted (K) or simply if the field is obvious. Since is an mth root of unity µm µm ζ if and only if it satisfies the equation zm − 1 = 0, the cardinality of (K) is finite, there µm being only finitely many zeros of a polynomial. Put [ (K) := := (K). µ µ µm m≥2 Proposition 11.2. The set ⊂ K× is a finite cyclic group; (K) is an abelian. A gener- µm µ ator for is called a primitive mth root of unity. µm Proof. Let x,y ∈ . Then (xy)m = xmym = 1, so is a subgroup of K×. The rest µm µm follows immediately by the above proposition. If x ∈ ⊆ (K) and y ∈ ⊆ (K), then µn µ µm µ xy ∈ ⊆ (K) so (K) is also an group (not finite in general). µn+m µ µ 48 D. LARSSON
It is in this generality impossible to tell the cardinality of (K). µm Notice that (K) ⊆ o since every a ∈ (K) is a zero of the monic integral polyno- µm K µm m mial z − 1. This obviously implies that µ(K) ⊆ oK also. 11.2. Units in number fields. 11.2.1. Dirichlet’s unit theorem.
Theorem 11.3. Let K/Q be an algebraic number field. Then the set of units is a finitely generated abelian group
× ∼ r1+r2−1 oK := U(oK) = µ(K) × Z , ℜ ℑ where r1 = #Mor(K) and r2 = #Mor(K) . Proof. The proof of this goes beyond the scope of the course. However, since the classical proof is rather interesting I encourage you to look it up in the literature. × 11.2.2. Consequences. The theorem says that there are u1,...,ur1+r2−1 ∈ oK such that any × unit u ∈ oK can be written as n u = ζ · un1 un2 ···u r1+r2−1 , ζ ∈ µ(K). 1 2 r1+r2−1
The elements u1,...,ur1+r2−1 are called the fundamental units of K and it is a formidable problem to compute these in general. However, if the field is real, i.e., if K ⊆ R, then µ(K) = {±1} (this is true for many (most?) fields) so the only problem is finding explicit generators for the ”Betti-part” r +r − Z 1 2 1. Recall that any finitely generated abelian group G can be decomposed as betti G = Gtors × G , where every element in the Betti-part is of infinite order12. An element a in a group is a torsion element if an = e for some n > 1. The set of torsion elements of a group is a subgroup, the torsion subgroup, Gtors. This holds for every group, abelian or not. But when G is abelian the above decomposition can be made. In fact, it is possible to prove that ∼ betti ∼ b Gtors = Z r1 × ··· × Zprs , and G = Z , b,s ≥ 0. p1 s If b = 0 then G is called a torsion group and if s = 0 then G is called a free group of rank b. In the case of units in number fields, we see that the torsion elements are exactly the subgroup µ(K) and the ”free” elements are the elements of the Betti-part.
Proposition 11.4. An element a ∈ oK is a unit if and only if NrmK/Q(a) = ±1.
Proof. If a is a unit then there is a b ∈ oK such that ab = 1 and so taking norms
1 = NrmK/Q(ab) = NrmK/Q(a)NrmK/Q(b) =⇒ NrmK/Q(a) = ±1. Conversely, recall that the norm can be computed as
NrmK/Q(a) = ∏ ϕ(a) = a · ∏ ϕ(a). ϕ∈Mor(K) ϕ∈Mor(K) ϕ6=id
12Enrico Betti (1823–1892) was an Italian mathematician specializing in algebra and topology. The name ”Betti-number” in the present case is actually borrowed from topology obviously named in Betti’s honor, where this is the rank of certain abelian groups appearing as cohomology groups attached to topological spaces. 49
Taking b := ±∏ϕ6=id ϕ(a) gives the desired result.
12. DEDEKINDDOMAINS Recall the following notions: - A ring R is noetherian (see Definition 4.11) if every ascending ideal chain
··· ⊆ ii−1 ⊆ ii ⊆ ii+1 ⊆ ··· stabilizes, i.e., such that there is an N ∈ Z such that
iN = iN+1 = iN+2 = ··· ; - a domain D is integrally closed if every a ∈ Frac(D) satisfying P(a) = 0 with P ∈ D[z] monic, implies that a ∈ D. Recall also that every maximal ideal is prime. We now make the following definition: Definition 12.1. An integral domain is called a Dedekind domain if it is noetherian, integrally closed and every prime ideal is maximal. √ We already know one example, namely, Z. Another is the Gaussian integers Z[ −1]. In fact,
Theorem 12.1. The ring of integers oK to an algebraic number field K is a Dedekind domain.
Proof. That oK is noetherian follows since oK is a finitely generated Z-module and Z is a noetherian ring (this is a standard fact from abstract algebra). Since oK is the integral closure of Z in K, oK is integrally closed. Hence we only need to show that every prime ∗ ideal is maximal. Let p be a prime ideal of oK. Then p is a non-zero prime ideal hpi of Z. It is clear that it is prime. Now, every a ∈ oK is the zero of a monic polynomial Pa with coefficients in Z. That it is non-zero follows since if a ∈ p then, n n−1 a + αn−1a + ··· + α0 = 0, for αi ∈ Z, α0 6= 0; ∗ hence, α0 ∈ p . Reducing oK modulo p is equivalent to reducing every a ∈ oK modulo p ∗ [ and so reducing all the Pa’s modulo p . Hence, since Z/hpi is a field, and a ∈ oK/p is algebraic over Z/hpi, oK/p must be a field and so p must be a maximal ideal. 12.1. A few important remarks. Before coming to the main theorem in this section, let me spare a few breaths on how ideals behave under morphisms. Some of this is simply reminders on topics from previous section, while others are new. First a notation. It is convenient and suggestive to sometimes use the notation Fp for oK/p. We will not use this in the case of powers of p, i.e., we will not use Fpe as standing e for oK/p . In this section, let A and B be arbitrary (commutative, associative) rings (with unity) and let f : A → B be a ring morphism13. We let a and b be ideals of A and B, respectively. When f is an injection B is called a ring extension of A, denoted as B/A or B | A. Recall the following notation b∗ := f −1(b) := b ∩ A := {a ∈ A | f (a) ∈ b},
a∗ := hai := f (a)B := aB := { ∑ aibi | ai ∈ f (a),bi ∈ B}. finite
f 13A very common way in all of algebra and number theory of saying that there is a morphism A −→ B, is that B is an A-algebra, or an algebra over A. 50 D. LARSSON
In the first case here one often says that ”b lies over a”. Notice that if f is an inclusion then this is equivalent to b ∩ A = a in the ordinary set-theoretical sense, hence the general notation14. Remember that there is a bijection:
a⊆a07→a0/a {ideals of A which includes a} ←→ {ideals of A/a}. Marvel at the suggestive notation a | a0 for a0 ⊆ a. Therefore, the ideals of a | a0 are exactly the ones giving the ideals of A/a0; or the ideals of A/a0 are exactly the ideals lying over a0; hence, by slightly abusing good mathematical decorum, passing to the quotient A/a0 kills ideals lying ”under” a0 (i.e., the ideals that are not over a0). Every morphism f : A → B induces a morphism b∗ = f −1(b) → b (by definition). There is a commutative diagram
f πb A K 9/ B / B/b O KK πb∗ M KKK K% % A/b∗ 0 8 rrr rrr b∗ / b ∗ The kernel of the composition πb ◦ f : A → B/b, is clearly b . Therefore, by the first isomorphism theorem, the dotted arrow exists and is an injection. From this follows that p ∈ Spec(B) =⇒ p∗ ∈ Spec(A). Indeed, since p is prime, B/p is an integral domain. We have that A/p∗ ,→ B/p and so A/p∗ is a subring of B/p and thus must be an integral domain also, implying that p∗ is prime. The implication p ∈ Spec(A) ⇒ p∗ ∈ Spec(B) is not true! In fact, this is one (the?) reason why decomposition of primes in Dedekind ring exten- sions is needed, because if primes were to stay primes under extensions the decomposition problem would be an empty problem. However, this non-implication is also a blessing since the theory becomes so much richer and more beautiful now! This will hopefully be apparent in what follows.
12.2. The main theorem on Dedekind domains. We know that in general, elements in domains cannot be decomposed into a unique (up to associates even) product of irre- ducibles in general. That is, factorization is not unique. The amazing thing is now that factorization is unique if we consider, not the elements themselves, but the ideals they generate! Therefore, connecting to the last paragraph of the previous section, the extended primes do not in general stay prime but they can be decomposed into unique primes! I want to emphasize in no small manner that this is in general only possible for Dedekind ring extensions. In what follows, oK is a Dedekind domain and K its field of fractions K := Frac(oK).
14Frankly, though, I’m not so hot on the notation b ∩ A in general, unless it is true set-theoretically, but the notation abounds in the literature so you better get used to it. 51
Theorem 12.2. Every non-zero, proper, ideal i in a Dedekind domain oK can be decom- posed uniquely (up to a re-ordering of factors) into prime ideals, i.e.,
ep i = ∏ p , where ep ≥ 0, ep = 0, for all but a finite number of p. p∈Spec(oK ) For the proof of this we need two lemmas:
Lemma 12.3. For every ideal a ⊆ oK there are non-zero prime ideals p1,...,pr such that p1 ···pr ⊆ a.
Lemma 12.4. Let p be a prime ideal in oK. Define −1 p := {a ∈ K | a · p ⊆ oK}. −1 −1 −1 Then a · p 6= a for every non-zero ideal a in oK. Notice that a ⊆ ap since 1 ∈ p . Proof of Lemma 12.3. Let S be the set of all ideals such that the statement of the lemma does not hold, and assume that S is non-empty. Since oK is noetherian the set S must have a maximal element, a. Furthermore, a cannot be a prime ideal so there are b,c ∈ oK such that bc ∈ a but b 6∈ a, c 6∈ a. Clearly, a ⊂ a + hbi, a ⊂ a + hci and (a + hbi)(a + hci) ⊆ a. Since a is maximal with respect to not containing a product of prime ideals, a + hbi and a+hci do. But this together with (a+hbi)(a+hci) ⊆ a implies that a also does, a contra- diction.
Proof of Lemma 12.4. Let a ∈ p, a 6= 0. Then by the previous lemma there are primes p1,...,pr such that p1 ···pr ⊆ hai ⊆ p. We can assume that r is the smallest possible such that this is true. Then one of the pi’s, say p1, is contained in p since if not, then we could choose a j ∈ p \ p j with a1 ···ar ∈ p; but since p is prime, a j ∈ p, for some j, a contradiction. This implies that p1 = p since p1 is maximal. We have that p2 ···pr 6⊆ hai, −1 so there is a b ∈ p2 ···pr such that b 6∈ aoK, i.e., a b 6∈ oK. However, we have that bp ⊆ hai −1 −1 −1 −1 so a bp ⊆ oK, implying that a b ∈ p , so p 6= oK. Let a be a non-zero ideal with generators a1,...,an (since oK is noetherian every ideal is finitely generated, another standard fact of noetherian rings). Assume that p−1a = a. Then for every b ∈ p−1 we have
bai = ∑Ai ja j, where Ai j ∈ oK. j This is equivalent to b − A11 −A12 ... −A1n a1 −A21 b − A22 ... −A2n a2 = 0¯. . . .. . . . . . . . −An1 −An2 ... b − Ann an Denote the square-matrix by W. By Cramer’s rule (see the Appendix) we get
det(W)a1 = det(W)a2 = ··· = det(W)an = 0, implying that det(W) = 0.
Hence b is integral over oK (expand det(W)); so b ∈ oK since oK is integrally closed, and −1 −1 thus p = oK, a contradiction. Therefore, p a 6= a and the proof is finished. Now we can prove Theorem 12.2.
Proof of Theorem 12.2. We begin by showing existence. Let S be the set of proper non-zero ideals that cannot be decomposed into prime ideals. The same argument as in the proof 52 D. LARSSON of Lemma 12.3 shows that there is a maximal element a ∈ S. This ideal is not prime so is included in a prime (maximal) ideal15 p. We get −1 −1 a ⊆ ap ⊆ pp ⊆ oK. −1 −1 However, Lemma 12.4 shows that a ⊂ ap and p ⊂ pp ⊆ oK strictly. Since p is maximal (notice that ap−1 is an ideal for all non-zero ideals a) we must have that −1 pp = oK.
Clearly, a 6= p implies that ap 6= oK, hence, taking into account the maximality of a in S and a ⊂ ap−1, the ideal ap−1 admits a prime decomposition −1 −1 ap = p1 ···pn and then so does a = ap p = pp1 ···pn, a contradiction. To show uniqueness assume that a can be decomposed as
a = p1p2 ···pn = q1q2 ···qm. The definition of prime ideals can be re-phrased as ab ⊆ p ⇒ a ⊆ p or b ⊆ p ⇐⇒ p | ab ⇒ p | a or p | b.
Now, p1p2 ···pn = q1q2 ···qm implies that p1|qi for some 1 ≤ i ≤ m. Since p1 is maximal, −1 −1 p1 = qi. Hence, multiplying with p1 and using that p1p1 = oK, we can cancel p1 = qi. Continuing like this shows that n = m and exactly one of the q j’s correspond to a given pi. The proof is finished. Definition 12.2. Let K/Q be an algebraic number field. Then a fractional ideal of K is a non-zero, finitely generated oK-submodule i ⊆ K. We denote the set of fractional ideals in K by J(K).
To distinguish between the different notions of ideals we will call ideals in oK integral ideals. Notice that the definition of fractional ideal is equivalent to: i is a fractional ideal if ∗ there is an α ∈ oK such that αi ⊆ oK is an integral ideal. Theorem 12.5. Let J(K) denote the set of fractional ideals in the algebraic number field K/Q. Then J(K) is an abelian group under the multiplication defined above, with identity element h1i = oK and inverses defined by −1 a := {x ∈ K | xa ⊆ h1i = oK}. Proof. The only thing that is not trivial to prove is that every fractional ideal has an inverse. Consider first an integral ideal a. By Theorem 12.2 this can be decomposed as a = p1 ···pn. −1 −1 −1 −1 This gives the inverse as a := p1 ···pn since, by the proof of Lemma 12.4, p ⊂ pp −1 −1 and so, because p is maximal, pp = oK. Also, if ba = oK, then b ⊆ a ; if ba ⊆ oK then −1 bab ⊆ b and so, since ab = oK, b ∈ b. Therefore, b = a . For the case of fractional ideals, ∗ recall that a is fractional if and only if there is a d ∈ oK such that da ⊆ oK is an integral −1 −1 −1 ideal. Then d a is an inverse to da, whence aa = oK. Corollary 12.6. Any fractional ideal a can be decomposed as
ep a = ∏ p , with ep ∈ Z, ep = 0, for all but a finite number of p. p∈Spec(oK )
This means that J(K) is the free abelian group on the set Spec(oK).
15This is a fact from ring theory (following from Zorn’s lemma): every ideal is contained in a maximal ideal. 53
The set of principal fractional ideals, i.e., fractional ideals on the form
∗ a = hai = a · oK, where a ∈ K , form a subgroup of J(K) denoted P(K).
Definition 12.3. The quotient group
Cls(K) := J(K)P(K), is called the ideal class group of K. The class of a ⊂ oK is called the ideal class of a. The number
hK := #Cls(K) is called the class number of K.
Notice that when hK = 1, it follows from definition, that every ideal in oK is principal, and thus oK is a unique factorization domain. This implication does not hold for fields of class number two, i.e., such that hK = 2. Hence, Cls(K) measures the deviation of K/Q to have unique factorization of elements.
Theorem 12.7. When K/Q is a number field and oK its ring of integers, the class number hK is finite. Proof. I know of no proof of this theorem that doesn’t use rather advanced ideas (though, classical), therefore I will leave this important fact un-proven, much to my dissatisfaction.
Remark 12.1. This theorem is not valid for general Dedekind domains.
We have the following happy consequence for Dedekind domains:
Theorem 12.8. Let oK be the ring of integers in a number field K/Q. Then
oK is a PID ⇐⇒ oK is a UFD, i.e., hK = 1 if and only if oK is a unique factorization domain.
Proof.
Let i be an ideal in oK for K/Q a number field. We define the absolute norm of i as + N(i) := (oK : i), N : Id(oK) → Z , where Id(oK) denotes the set of all ideals in oK. This can be extended to a homomorphism of groups n N : J(K) → + := { | n,m ∈ +}, Q m Z and satisfies
N(a · b) = N(a)N(b), and N(hai) = |NrmK/Q(a)|.
Notice that, by the above multiplicative property, we have that N(oK) = 1 since oK is the unit element in J(K). From this follows that we can define N(a−1) = N(a)−1. 54 D. LARSSON
13. EXTENSIONS, DECOMPOSITION AND RAMIFICATION From now on I will use terminology and facts from the theory of finite fields. Consult the section entitled ”Finite fields” for an introduction to this topic. The following theorem will be left un-proven.
Theorem 13.1. Let oK be a Dedekind domain with fraction field K and let L/K be a finite extension of K. Then the integral closure oL := oK of oK in L is also a Dedekind domain.
Now, let p be a prime ideal of K. Then the extension p∗ = poL of p in oL is an ideal but not necessarily prime. However, since oL is a Dedekind domain p∗ decomposes as a product of primes of L:
eP p∗ = ∏ P , where, eP ≥ 0, eP = 0, for all but a finite number of P. P∈Spec(oL)
As is customary, we will simply write the extension of p in L as p and not p∗. Lemma 13.2. Let L/K be a finite field extension.
(i) If K = Q, i.e., if L/Q is an algebraic field extension, and p ∈ Spec(oL). Then oL/p is a finite field. ∗ (ii) In general, the extension FP/Fp = (oL/P)/(oK/p), where p = P , is a finite exten- sion of finite fields, called the residue field extension (associated to P and p).
Proof. (i) Since p is maximal it is clear that oL/p is a field. Every prime ideal ∗ p ∈ Spec(oL) lies over a non-zero proper prime ideal in Z, i.e., p = (p), via the canonical injection Z ,→ oL. Reducing oL modulo p automatically reduces Z modulo (p). Hence there is an injection Fp ,→ Fp, and so Fp/Fp is a finite field ex- tension of a finite field (recall that oK is a Z-module of finite rank equal to [K/Q]), implying that Fp = oL/p is itself finite. (ii) This follows from (i) and the tower law.
e1 e2 er Definition 13.1. Let p decompose in L as p = P1 P2 ···Pr . Then - p is called unramified in L if e1 = e2 = ··· = er = 1 and if FP/Fp is a separable extension (this is always the case for us); otherwise, p is ramified (in L) and P j ramifies over p if e j > 1 and is unramified if e j = 1. - p is totally split in L if r = [L/K] and non-split if r = 1; - p is called inert if it stays prime in the extension, i.e., if p∗ is prime; - the number f j := [FP/Fp]
is called the jth inertia degree of p (in L) and the number e j is called the jth ramification degree; - p (or Pi) is tamely ramified if ei > 1 and gcd(ei,char(Fp)) = 1; - if char(Fp) | ei, p (or P) is called wildly ramified.
e1 er Theorem 13.3. Let L/K be a finite extension. For every p ∈ Spec(oK) with p = P1 ···Pr we have the so-called fundamental identity r ∑ e j f j = [L/K]. j=1 55
For a proof see any book on algebraic number theory. Notice that we can write the sum as ∑v(P | p) f (P | p) P|p which is often done in the literature. This is to be interpreted as summing over all primes P lying over the given p. Observe the suggestive notation P | p. 13.1. Ramification and decomposition. 13.1.1. Which primes ramify? The treatment that follows is a (very) slight adaption of arguments I learned from Keith Conrad.
Theorem 13.4. Let K/Q be an algebraic number field. Then the primes of Q that ramify in K are the prime divisors of the discriminant, i.e.,
e1 e2 er p | Disc(oK/Z) ⇔ p = p1 p2 ···pr , where at least oneei > 0. To prove this theorem we need a lemma.
Lemma 13.5. Let oK be the ring of integers of K/Q and let M and N be two Z-modules. Then
(i) DiscZ(oK)(mod p) = DiscZ/hpi(oK/hpi); (ii) DiscZ(M × N) = DiscZ(M)DiscZ(N). [,p Proof. Note first that if f := { f1,..., f[K/Q]} is a basis for oK over Z then f is a basis of oK/hpi over Z/hpi. Hence oK/hpi is a Fp-vector space of dimension [K/Q]. Recall the definition of the discriminant of oK with basis f: Disc(oK) = Disc(f) = det(Tr( fi f j)). If mx is the matrix of the linear mapping associated with multiplication of x ∈ oK, it is clear that mx[,p is the matrix of the reduced map oK/hpi → oK/hpi. Hence,
Tr(o /hpi)/ (m [ [ ) = Tro / (m fi f j )(mod p). K Fp fi f j K Z Now taking determinants gives (i). Notice that if f and f0 are basis sets of M and N, 0 0 0 0 0 respectively, then f ∪ f = { f1, f2,..., fm, f1, f2 ..., fn} is a basis for M × N and fi f j = 0 for all i and j. From this follows that the matrix to take determinant of to get Disc(M × N) is the matrix ¯ ¯ Tr(M×N)/Z( fi f j) 0 TrM/Z( fi f j) 0 ¯ 0 0 = ¯ 0 0 . 0 Tr(M×N)/Z( fi f j) 0 TrN/Z( fi f j) Therefore, taking determinants gives (ii).
e1 er Proof of Theorem 13.4. Since hpi = p1 ···pr , pi 6= p j, i 6= j, the ring-theoretical version of the Chinese remainder theorem (see the Appendix), gives that ∼ e1 e2 er oK/hpi = oK/p1 × oK/p2 × ··· × oK/pr . ei ei We have that hpi ⊆ pi so each oK/pi is an Fp-vector space. Using the above lemma (ii) gives ei DiscFp (oK/hpi) = ∏DiscFp (oK/pi ). By part (i) of the lemma,
DiscZ(oK)(mod p) = DiscFp (oK/hpi) so p|DiscZ(oK) if and only if DiscFp (oK/hpi) = 0 in Fp. Hence we need to show that for every pe|hpi, v DiscFp (oK/p ) = 0 ⇐⇒ e > 1. 56 D. LARSSON
Since the vanishing of discriminants is independent on the chosen basis this will work in any basis. Assume that e > 1 and pick a non-zero x ∈ p \ pe. Extend the reduction of x modulo pe, [ [ e [ [ e x , to a basis {x1 := x ,x2,...,xn} of oK/p . Notice that x is nilpotent, i.e., (x ) = 0. The [ [ first column of the matrix Tr(xix j) is Tr(x x j). The linear mapping associated with x x j is [ nilpotent so has eigenvalues all equal zero (in some basis). So Tr(x x j) = 0. Hence the determinant of the trace matrix is zero and so then is also the discriminant. Assume now that e = 1. We need to show that p - DiscZ(oK) in this case. Then Fp := a oK/p is a finite field extension of Fp. In fact [Fp/Fp] = p , for some a > 1. If DiscFp (Fp) = det(Tr(eie j)) = 0 then Tr : Fp → Fp must be the zero function. However, for every x ∈ Fpa 2 a−1 (for every finite field) Tr(x) = x+xp +xp +···+xp which can never be the zero function a since this is a polynomial of degree less than [Fp/Fp] = p . Hence DiscZ(oK) 6= 0, and the proof is finished. Corollary 13.6. There are only finitely many primes that ramify in any given finite Q- extension. Also, there are no finite, unramified extensions of Q. Hence, once we have computed the discriminant, it is easy16 to determine which primes ramify in the given extension. Note also that even though there are no unramified algebraic number fields one can ask for number fields unramified outside a specified set of primes. This is a useful technique (and important problem) in number theory. Don’t, however, be misled into thinking that there are no unramified extensions of Q; the point is that there are no finite extensions. Also, extending a number field K/Q to L ⊇ K can give an unramified extension L/K (al- though, of course, L/Q is ramified). In addition, the so-called local fields, which for our concern can be thought as the finite field extensions of the p-adic numbers. 13.1.2. Dedekind’s theorem. There is also a beautiful result, apparently due to Dedekind, on how to compute the decomposition of a prime in an extension. Recall that the index, (M : m), of a submodule m ⊆ M is the number of left (or right) cosets of m, i.e., (M : m) := #(M/m) ∈ N ∪ {∞}. 2 [Q(θ)/Q] Let Z[θ] denote the submodule of oQ(θ) spanned over Z by {1,θ,θ ,...,θ }. No- tice that oK does not have to have the same generators. To simplify we write F(z) := irr(θ,Q)(z).
Theorem 13.7 (Dedekind). Let K/Q be a finite field extension and let θ ∈ oK such that θ is a primitive element for K, i.e., such that K = Q(θ). Let p be a rational prime not dividing the index (oK : Z[θ]). Factor the reduction of the minimal polynomial of θ modulo p into irreducibles as
[,p [,p e1 [,p er [,p F (z) = F1 (z) ···Fr (z) , Fj(z) ∈ Z[z], Fj (z) ∈ (Z/hpi)[z], [,p where (·) : Z[z] → (Z/hpi)[z] denotes the reduction modulo p. Then
(13.1) p j := hpi∗ + hFj(θ)i∗ = hp,Fj(θ)i∗ are the different prime ideals in oQ(θ) over hpi. Furthermore,