I should have mentioned this before, but usually dual notions are marked by the prefix “co-“. As an example, we have “well-powered” and “co-well-powered” categories.
Another example: We know through Erdős that “a mathematician is a device for turning coffee into theorems”. It thus follows by duality that a comathematician is a device for turning cotheorems into ffee.
One of the most interesting general facts about categories is how ubiquitous the notion of duality is. Pretty much everything has a “mirror image”, and the mirror is the opposite category.
Given a category , with objects and morphisms , we can construct the “opposite category” simply enough. In fact, it has the exact same objects and morphisms as does. The difference comes in how they relate to each other.
Remember that we had two functions, and assigning the “source” and “target” objects to any arrow. To get the opposite category we just swap them. Given a morphism, its source in is its target in , and vice versa. Of course, now we have to swap the order of composition. If we have and in , then we get and in . In the composition is defined, but in the composition is defined.
Most definitions we give will automatically come with a “dual” definition, which we get by reversing all the arrows like this. For example, monos and epis are dual notions, as are subobjects and quotient objects. Just write down one definition in terms of morphisms, reverse all the morphisms (and the order of composition), and you get the other.
Theorems work this way too. If you dualize the hypotheses and the conclusions, then you can dualize each step of the proof to prove the dual theorem. I can prove that any injection is monic, so it follows immediately by duality that any surjection is epic. Many texts actually completely omit even the statements of dual notions and theorems once they define the opposite category, but I’ll try to be explicit about the duals (though of course I won’t need to give the proofs).
Another place duality comes up is in defining “contravariant” functors. This is just a functor . It sends each object of to an object of , and sends a morphism in to a morphism in . See how the direction of the image morphism flipped? Early on, contravariant and regular (“covariant”) functors were treated somewhat separately, but really they’re just the same thing once you take the opposite category into account. Sometimes, though, it’s easier to think in terms of contravariant functors rather than mentally flipping all the arrows.
I’ll close with an example of a contravariant functor we’ve seen before. Consider a ring with unit and a left module over . That is, is an object in the category . We can construct the dual module , which is now an object in the category of right -modules. I say that this is a contravariant functor. We’ve specified how the dual module construction behaves on objects, but we need to see how it behaves on morphisms. This is what makes it functorial.
So let’s say we have two left -modules and , and an -module homomorphism . Since we want this to be a contravariant functor we need to find a morphism . But notice that , and similarly for . Then we have the composition of -module homomorphisms . If is a linear functional on , then we get as a linear functional on . We can define .
Now, is this construction functorial? We have to check that it preserves identities and compositions. For identities it’s simple: , so every linear functional on gets sent back to itself. For compositions we have to be careful. The order has to switch around because this is a contravariant functor. We take and and check , as it should.
Today I’ll give my own solution to the power question I posted last week. I’m going to present it as I always wished I could have: as a whole with answers to the specific questions spun off as they naturally arise, rather than separated part by part.
First I’ll restate the setup into a more natural language. This is really a question about certain endomorphisms on vector spaces over the field . This is the quotient of the ring of integers by the maximal ideal generated by . If you haven’t seen it before, you should be able to construct it by all I’ve said so far.
Anyhow, the “arrangements” are just vectors in the vector space . Such a space comes with a basis , where has a in place and elsewhere. Each of these spaces of course has the identity endomorphism , and it also has a “left rotation” endomorphism sending to and to , and its inverse — right rotation. The transformation we are concerned with is .
Since we are given bases we are justified in writing out matrices for transformations, and even for vectors (because ). The transformation has the matrix where is if or and otherwise. A vector of length will be an matrix.
Numbers 1 and 2 are just calculations, so I’ll omit them.
We can write as the matrix , so has matrix , which sends every vector to the zero vector. This is part 4.
Now one interesting property we can see straight off. We can tell whether there are an even or an odd number of ones in a given arrangement by adding up all the entries. That is, we can take the product with the matrix consisting of all ones. If we first transform an arrangement , then measure the number of ones, this is like taking the product . But each column of the matrix for has exactly two ones in it, so the product consists of all zeroes, and thus is always zero. That is, the image of any arrangement after a transformation always has an even number of ones. That’s number 10.
What arrangements are fixed by the transformation? This amounts to solving the equation
so we must have or . The zero vector is the only fixed point.
What vectors get sent to this fixed point? This is the kernel of the transformation — the vectors such that . Equivalently, these satisfy (why?). Thus all the entries must be the same, and the vector must consist of all zeroes or all ones. That’s number 9.
Now we see that if then is in the kernel of , and is thus either all zeroes or all ones. But if is odd, the vector consisting of all ones is not in the image of . Thus if we don’t start with a vector in the kernel we’ll never land in the kernel, and we’ll never get to the vector of all zeroes. That’s number 11. As a special case we have number 3.
Let’s consider . We expand this as , since we’re working over . Similarly, if we square this we get . In fact, we have that . Indeed, this is true by definition for , and if it’s true for then
so the claim is true by induction.
This means that after transformations a vector is sent to the sum (in ) of itself with its left-rotation by places. Thus if we can divide the entries in into vectors of length each — just pick the entries spaced out by separations of . Then acts on the same way acts on each of the subvectors, since the shift by places fixes the subvectors. That’s number 7. Parts 5 and 6 are now special cases.
Also, if then , so , so transformations send every vector of length to zero. That’s number 8.
Finally, let be any even number that is not a power of . Now the result of is the same as applying to each of the subvectors as described above. But now each subvector has odd length. If has a single in it, then one of these subvectors must contain it. By number 11 this subvector can never be sent to zero, so is never zero. If were ever zero then would be for a large enough , which will never happen. That’s part 12.