## The Fundamental Involutory Quandle

As I discussed last time, coloring a knot with any abelian group is secretly using the dihedral quandle associated to that group. This is an involutory quandle with action . The reason knot coloring works out so nicely is that the axioms of (involutory) quandles line up with the Reidemeister moves.

But for the moment we’re stuck with picking this or that involutory quandle and counting how many colorings it gives for a given knot. Different quandles give different coloring numbers, and we’d like to find a better way of thinking of them all at once. We’re going to construct a new involutory quandle from a knot that captures all of them.

Take any diagram of the knot we’re interested in. Remember the knot table if you want to pick one out. Now each arc in the diagram has to get *some* color, no matter what quandle we’re using to color it. Instead of picking a color from a specific quandle, let’s just slap a label like , , or on each arc. Be sure to use a different label for each different arc.

Now those labels will generate an involutory quandle. We can throw them together with the two quandle compositions to get “words” like . These words, of course, are subject to the normal quandle equivalences, but we need more relations for our purposes. At each crossing the values in a coloring have to satisfy a certain relation, so we’re going to build that right into our quandle. If the arcs labeled and meet under the crossing arc labeled , then we must have .

This seems to depend on the choice of a diagram, though. Well, it sort of does, but any Reidemeister move gives an isomorphism of quandles relating the two sides. For example, performing the first one splits an arc into two pieces. Say label becomes and . Then the relations we introduce say that . But the axioms of quandles say that , so and we can just drop one of these generators and the relation we’ve now “used up”. Try to find the isomorphisms for the other two moves. This justifies calling the quandle we’ve constructed (up to isomorphism) “the” fundamental involutory quandle of the knot .

So what’s a coloring? A coloring assigns an element of some quandle to each arc of the knot diagram. But arcs in the diagram are just generators of the fundamental quandle. That is, a coloring is a function that takes generators of the fundamental quandle to a selected target quandle. If it plays nicely with the relations between the generators, it will be a quandle homomorphism. In fact it does, precisely because we picked the relations between the generators to be exactly those required by colorings. A given relation comes from a crossing, and every coloring of a knot obeys the same restrictions at crosings.

In the end we’ve found that the set of all colorings of by an involutory quandle is the set of quandle homomorphisms , so the number of -colorings is the cardinality of this set. If we have a good understanding of quandles and their homomorphisms, we can read off coloring numbers by involutory quandles from the fundamental involutory quandle.

## At last

So I’m up until 4:30 scrawling something out. What’s the only thing that wouldn’t annoy me to no end waking me up at 9:00?

Tulane University.

Finally.

## The lattice of ideals

We know that the collection of all ideals of a given ring form a rig. In fact, they also form a lattice. We put the partial order of inclusion on ideals, so is below if .

To show that this poset is a lattice we have to show that pairwise greatest lower bounds and least upper bounds exist. Lower bounds are easy: the intersection of two ideals is again an ideal. By definition, any ideal contained in both and is contained in .

Upper bounds are a little trickier, since we can’t just take the union of two ideals. That would work for subsets of a given set, but in general the union of two ideals isn’t an ideal. Instead, we take their sum. Clearly and . Also, if is another ideal containing both and , then contains all linear combinations of elements of and . But is the set of all such linear combinations. Thus , and is the least upper bound of and .

This lattice is related to the divisibility preorder. Given a commutative unital ring and two elements , recall that if there is an so that . Then every multiple of is also a multiple of . Thus we see that the principal ideal is contained in the principal ideal . On the other hand, if we can see that for some , so . In particular, two elements are associated if and only if they generate the same principal ideal.

Notice that this correspondence reverses the direction of the order. If is below in the divisibility ordering, then is *above* in the ideal ordering. Thus the “greatest common divisor” of two ideals is actually now the least ideal containing both of them. The language of ideals, however, is far more general than that of divisibility. We now need to recast most of what we know about divisibility from our experience with natural numbers into these more general ring-theoretic terms.

## Lattices

A poset which has both least upper bounds and greatest lower bounds is called a lattice. In more detail, let’s say we have a poset and give it two operations: meet (written ) and join (written ). These satisfy the requirements that

- and .
- If and then .
- and .
- If and then .

Not every poset has a meet and a join operation, but if these operations do exist they are uniquely specified by these requirements. In fact, we can see this sort of like how we saw that direct products of groups are unique up to isomorphism: if we have two least upper bounds for a pair of elements then they must each be below or equal to the other, so they must be the same.

We can derive the following properties of the operations:

- and .
- and .
- and .

from these we see that there’s a sort of duality between the two operations. In fact, we can see that these provide two commutative semigroup structures that happen to interact in a certain nice way.

Actually, it gets even better. If we have two operations on any set satisfying these properties then we can define a partial order: if and only if . So we can define a lattice either by the order property and get the algebraic properties, or we can define it by the algebraic properties and get the order property from them.

In many cases, a lattice also satisfies , or equivalently . In this case we call it “distributive”. A bit weaker is to require that if then for all . In this case we call the lattice “modular”.

A lattice may have elements above everything else or below everything else. We call a greatest element of a lattice and a least element . In this case we can define “complements”: and are complements if and . If the lattice is distributive, then the complement of is unique if it exists. A distributive lattice where every element has a complement is called “Boolean”.

## Generators of ideals

Let’s say we’ve got a ring and an element . What is the smallest left ideal that contains ? Well, we have to have all multiples for so it’s closed under left multiplication. If has a unit, this is all we need. Otherwise, we have to make sure we include all the elements with summands (and their negatives) to make sure it’s an abelian subgroup. Thus the subset is a left ideal in . If has a unit, we just need the subset . We call this the principal left ideal generated by , and write . We can do something similar for right ideals (), and for two-sided ideals we get the subset .

As for any submodules we can form the sum. If we have elements they generate the left ideal , or a similar right ideal. For two-sided ideals we write . The term “principal”, however, is reserved for ideals generated by a single element.

Let’s look at these constructions in the ring of integers. Since it’s commutative, every ideal is two-sided. An integer then generates the principal ideal of all multiples of . In fact, every ideal in is principal.

If is an ideal, consider the subset of all its (strictly) positive elements. Since this is a subset of the natural numbers it has a least element . I say that every element of is a multiple of . If not, then there is some that doesn’t divide. If we can apply Euclid’s algorithm to and , at the first step we get with . The greatest common divisor of and will thus be less than , and Euclid’s algorithm gives us a linear combination for integers and . Thus must be in the ideal as well, contradicting the minimality of .

So every ideal of is principal. When this happens for a ring, we call it a “principal ideal ring”, or a “principal ideal domain” if the ring is also an integral domain.

So how do ideals of integers behave under addition and multiplication? The ideal is the ideal . This it consists of all the linear combinations . In particular, the smallest positive such linear combination is the greatest common divisor of and , as given by Euclid’s algorithm. The product of the ideals is the set of all products of multiples of and : . Thus .

## No, we have no mathematics. We have no mathematics today.

Sorry for the downtime. I had to start the Calc 2 final on my own at 9 in the morning, when I’m usually not even awake yet. Then pushing through to get everything graded by the end of the day. Thus ends the class, and quite possibly my academic career.

Now I can pack up and straighten up my apartment to hit the road tomorrow. I’m spending a couple weeks’ down-time in Maryland, so I’ll be back out at the College Perk in College Park most nights. If you’re in the area, drop by.

## More modules, more ideals

The first construction I want to run through today is related to the amalgamated free product from group theory. Here’s the diagram in modules:

Remember we read it as follows: If we have modules , , and , and homomorphisms from into each of and , then the “amalgamated direct sum” is another module with homomorphisms into it from each of and making the square commute. Further, for any other module and pair of homomorphisms into it, there is a unique homomorphism from to .

How do we know that such a thing exists? Well, we can take the direct sum of and , which comes with homomorphisms into it from and . Then we can follow both paths from to . In general they’re different homomorphisms, since one has an image completely in and the other completely in . But since we’re looking at module homomorphisms we can subtract one from the other. To make the square commute we want the image of this difference to be zero, and we can make it be zero by forming the quotient module!

This turns out to be useful right away. Let’s say that and are both submodules of a module . They definitely share the zero element of , but they might share a larger submodule than that. It’s easily verified that their intersection is a submodule, and it comes with inclusion homomorphisms into each of and . Now if we want to “add” the submodules and , we had better not treat elements in their intersection differently depending on which submodule module we pick them from, since they’re *all* just submodules of , so the direct sum isn’t what we want.

Instead, we find that the submodule of all elements of of the form with and is isomorphic to the direct sum amalgamated over their intersection. In particular, we can apply this to submodules of our base ring itself — ideals! We define the sum of two ideals as this sum of submodules of .

There’s one more thing we can do for ideals. If we have a left ideal and an abelian subgroup then we can form their tensor product over : . Since is a left -submodule, this turns out to be a left -submodule of . Then we can use the multiplication on to get a homomorphism . We denote its image as , and it is a left ideal of . In terms of elements, it’s the set of all sums of products in : with and . In particular, we could choose to be another ideal and get the product of ideals .

Now here’s where it gets *really* fun. Start with a ring and consider the collection of all its left ideals . There are a bunch of things we can show about these operations on ideals, which I’ll leave as exercises. If it’s easier, use the descriptions in terms of elements, but I think it’s more satisfying to work with the diagrams and universal properties. Here , , and are ideals, and is the ideal consisting of only the zero element.

What does all this mean? The collection of left ideals of form a rig, like the natural numbers! Further if has a unit, then we find , so this rig has a unit. If is commutative, then so the rig is too.

## Submodules, quotient modules, and the First Isomorphism Theorem for modules

Okay, getting a little back down to Earth now. Just like we had for groups and rings, we have an isomorphism theorem for modules.

First off, a submodule of a left -module is just an abelian subgroup that’s closed under the action of . That is, for any and , we have . A submodule comes with an inclusion homomorphism .

Now if we take an -module and a submodule we can just consider and as abelian groups and form the quotient group. Remember that every subgroup of an abelian group is normal, so the quotient is again an abelian group made up of the “slices” . It turns out that this is again an -module. Just use the action . If we chose a different representative of the slice, then we’d get , which represents the same image, so this action is well-defined. Quotient modules come with projection homomorphisms .

Actually there’s one sort of submodule we’ve already seen. Remember that every ring is a left module over itself. Then the left submodules of are exactly the left ideals of ! If we have a two-sided ideal then we get a left action of on the quotient module and a right action as well, since is also a right submodule. Then we can take the tensor product and get a linear function from multiplying representatives. Presto! Quotient ring!

We can use ideals to give submodules and quotient modules of other -modules too. Take a ring with left module and a left ideal . Then we can restrict the action of on to the ideal to get . This is clearly an abelian subgroup of , and it turns out to be a submodule too. Indeed, we see that . Then we can make the quotient -module as above. Even better, if is two-sided this is actually a module over : use the right -module structure on and the left -module structure on and tensor to get . Then we can get a linear function by choosing representatives and showing that the choice is immaterial.

I promised an isomorphism theorem. Well, I’ll state it, but the proof is pretty much exactly the same as the two we’ve seen before so I’ll leave you to review those. Any homomorphism of left (right) -modules factors as the composition , where the first arrow is the projection homomorphism, the third is the inclusion homomorphism, and the middle arrow is an isomorphism. We call the submodule the kernel of the homomorphism and the submodule the image of the homomorphism. Notice that there’s no restriction on the sorts of submodules that can be kernels here. For groups a kernel is a normal subgroup, and for rings a kernel is a two-sided ideal, but any submodule can be the kernel of a module homomorphism. This leads to a few more definitions that come in handy. The quotient module is called the coimage, and the quotient module is called the cokernel. Thus we see that the coimage and the image of any homomorphism of modules are isomorphic.

## Algebras

We have defined a ring as a -module (abelian group) with a linear function satisfying certain properties. The concept of an algebra takes this definition and extends it to work over more general base rings than .

Let be a module over a commutative ring with unit. Then has both a left and a right action by , since is commutative. Thus, when we take the tensor product , the result is also an module. It makes sense, then, to talk about an -module homomorphism . Equivalently, this is a “multiplication” function such that

An -module equipped with such a multiplication is called an -algebra. We will often write the multiplication as . In many cases of interest, the base ring will be a field , but any ring is an algebra over .

Usually the term “algebra” on its own will refer to an associative algebra. This imposes an additional condition like the one we had in the definition of a ring: . An algebra may also have a unit so that for all . Algebras can also be commutative if for all elements . There are other kinds of algebras we’ll get to later that are not associative.

Pretty much everything I’ve said about rings works for associative algebras as well, substituting “-module” for “abelian group”. An -module is a left -module if there is an -linear function , and a similar definition works for right -modules. We can take direct sums and tensor products of -modules, and we have an -module of homomorphisms . All these constructions are clear from what we’ve said about modules over rings if we consider that is a ring, and that an -module is an abelian group with actions of both and which commute with each other.

The standard constructions of rings also work for algebras. In particular, we can start with an -module and build the free -algebra on like we built the free ring on an abelian group. Just use , where the tensor powers over make sense because is commutative.

We can also start with any semigroup and build the semigroup algebra just like we did for the semigroup ring . As a special case, we can take to be the free commutative monoid on generators and get the algebra of polynomials in variables over . In fact, almost all of “high school algebra” is really about studying the algebra , where is the field of rational numbers I’m almost ready to define.

Another source of -algebras extends the notion of the ring of endomorphisms. If is any -module, then is again an -module, and composition is -bilinear, making this into an -algebra.

Algebras over more general commutative rings than — particularly over fields — are extremely useful objects of study mostly because the linear substrate can often be much simpler. Building everything on abelian groups can get complicated because abelian groups can be complicated, but building everything on vector spaces over a field is generally pretty straightforward since vector spaces and their linear transformations are so simple.

## Overpriced journals

In the May issue of the *Notices of the American Mathematical Society*, there is an article by Allyn Jackson: Jumping Ship: *Topology* Board Resigns about the mass resignation of the entire editorial board of the Elsevier journal over its exorbitant pricing. This has been a steadily mounting problem to academic mathematics. In 2005 a group set up the Banff Protocol, refusing to have anything to do with excessively-priced journals. If you’re a professional mathematician, go there and sign up.