Duality terminology
I should have mentioned this before, but usually dual notions are marked by the prefix “co-“. As an example, we have “well-powered” and “co-well-powered” categories.
Another example: We know through Erdős that “a mathematician is a device for turning coffee into theorems”. It thus follows by duality that a comathematician is a device for turning cotheorems into ffee.
The Opposite Category
One of the most interesting general facts about categories is how ubiquitous the notion of duality is. Pretty much everything has a “mirror image”, and the mirror is the opposite category.
Given a category , with objects
and morphisms
, we can construct the “opposite category”
simply enough. In fact, it has the exact same objects and morphisms as
does. The difference comes in how they relate to each other.
Remember that we had two functions, and
assigning the “source” and “target” objects to any arrow. To get the opposite category we just swap them. Given a morphism, its source in
is its target in
, and vice versa. Of course, now we have to swap the order of composition. If we have
and
in
, then we get
and
in
. In
the composition
is defined, but in
the composition
is defined.
Most definitions we give will automatically come with a “dual” definition, which we get by reversing all the arrows like this. For example, monos and epis are dual notions, as are subobjects and quotient objects. Just write down one definition in terms of morphisms, reverse all the morphisms (and the order of composition), and you get the other.
Theorems work this way too. If you dualize the hypotheses and the conclusions, then you can dualize each step of the proof to prove the dual theorem. I can prove that any injection is monic, so it follows immediately by duality that any surjection is epic. Many texts actually completely omit even the statements of dual notions and theorems once they define the opposite category, but I’ll try to be explicit about the duals (though of course I won’t need to give the proofs).
Another place duality comes up is in defining “contravariant” functors. This is just a functor . It sends each object of
to an object of
, and sends a morphism
in
to a morphism
in
. See how the direction of the image morphism flipped? Early on, contravariant and regular (“covariant”) functors were treated somewhat separately, but really they’re just the same thing once you take the opposite category into account. Sometimes, though, it’s easier to think in terms of contravariant functors rather than mentally flipping all the arrows.
I’ll close with an example of a contravariant functor we’ve seen before. Consider a ring with unit and a left module
over
. That is,
is an object in the category
. We can construct the dual module
, which is now an object in the category
of right
-modules. I say that this is a contravariant functor. We’ve specified how the dual module construction behaves on objects, but we need to see how it behaves on morphisms. This is what makes it functorial.
So let’s say we have two left -modules
and
, and an
-module homomorphism
. Since we want this to be a contravariant functor we need to find a morphism
. But notice that
, and similarly for
. Then we have the composition of
-module homomorphisms
. If
is a linear functional on
, then we get
as a linear functional on
. We can define
.
Now, is this construction functorial? We have to check that it preserves identities and compositions. For identities it’s simple: , so every linear functional on
gets sent back to itself. For compositions we have to be careful. The order has to switch around because this is a contravariant functor. We take
and
and check
, as it should.
Solution to the ARML Scrimmage Power Question
Today I’ll give my own solution to the power question I posted last week. I’m going to present it as I always wished I could have: as a whole with answers to the specific questions spun off as they naturally arise, rather than separated part by part.
First I’ll restate the setup into a more natural language. This is really a question about certain endomorphisms on vector spaces over the field . This is the quotient of the ring of integers by the maximal ideal generated by
. If you haven’t seen it before, you should be able to construct it by all I’ve said so far.
Anyhow, the “arrangements” are just vectors in the vector space . Such a space comes with a basis
, where
has a
in place
and
elsewhere. Each of these spaces of course has the identity endomorphism
, and it also has a “left rotation” endomorphism
sending
to
and
to
, and its inverse
— right rotation. The transformation we are concerned with is
.
Since we are given bases we are justified in writing out matrices for transformations, and even for vectors (because ). The transformation
has the matrix
where
is
if
or
and
otherwise. A vector of length
will be an
matrix.
Numbers 1 and 2 are just calculations, so I’ll omit them.
We can write as the matrix
, so
has matrix
, which sends every vector to the zero vector. This is part 4.
Now one interesting property we can see straight off. We can tell whether there are an even or an odd number of ones in a given arrangement by adding up all the entries. That is, we can take the product with the matrix
consisting of all ones. If we first transform an arrangement
, then measure the number of ones, this is like taking the product
. But each column of the matrix for
has exactly two ones in it, so the product
consists of all zeroes, and thus
is always zero. That is, the image of any arrangement after a transformation always has an even number of ones. That’s number 10.
What arrangements are fixed by the transformation? This amounts to solving the equation
so we must have or
. The zero vector is the only fixed point.
What vectors get sent to this fixed point? This is the kernel of the transformation — the vectors such that
. Equivalently, these satisfy
(why?). Thus all the entries must be the same, and the vector must consist of all zeroes or all ones. That’s number 9.
Now we see that if then
is in the kernel of
, and is thus either all zeroes or all ones. But if
is odd, the vector consisting of all ones is not in the image of
. Thus if we don’t start with a vector in the kernel we’ll never land in the kernel, and we’ll never get to the vector of all zeroes. That’s number 11. As a special case we have number 3.
Let’s consider . We expand this as
, since we’re working over
. Similarly, if we square this we get
. In fact, we have that
. Indeed, this is true by definition for
, and if it’s true for
then
so the claim is true by induction.
This means that after transformations a vector is sent to the sum (in
) of itself with its left-rotation by
places. Thus if
we can divide the entries in
into
vectors of length
each — just pick the entries spaced out by separations of
. Then
acts on
the same way
acts on each of the subvectors, since the shift by
places fixes the subvectors. That’s number 7. Parts 5 and 6 are now special cases.
Also, if then
, so
, so
transformations send every vector of length
to zero. That’s number 8.
Finally, let be any even number that is not a power of
. Now the result of
is the same as applying
to each of the
subvectors as described above. But now each subvector has odd length. If
has a single
in it, then one of these subvectors must contain it. By number 11 this subvector can never be sent to zero, so
is never zero. If
were ever zero then
would be for a large enough
, which will never happen. That’s part 12.