The Determinant
Let’s look at the dimensions of antisymmetric tensor spaces. We worked out that if has dimension , then the space of antisymmetric tensors with tensorands has dimension
One thing should leap out about this: if is greater than , then the dimension formula breaks down. This is connected with the fact that at that point we can’t pick any -tuples without repetition from basis vectors.
So what happens right before everything breaks down? If , then we find
There’s only one independent antisymmetric tensor of this type, and so we have a one-dimensional vector space. But remember that this isn’t just a vector space. The tensor power is both a representation of and a representation of , which actions commute with each other. Our antisymmetric tensors are the image of a certain action from the symmetric group, which is an intertwiner of the action. Thus we have a one-dimensional representation of , which we call the determinant representation.
I want to pause here and point out something that’s extremely important. We’ve mentioned a basis for in the process of calculating the dimension of this space, but the space itself was defined without reference to such a basis. Similarly, the representation of any element of is defined completely without reference to any basis of . It needs only the abstract vector space itself to be defined. Calculating the determinant of a linear transformation, though, is a different story. We’ll use a basis to calculate it, but as we’ve just said the particular choice of a basis won’t matter in the slightest to the answer we get. We’d get the same answer no matter what basis we chose.
Dimensions of Symmetric and Antisymmetric Tensor Spaces
We’ve laid out the spaces of symmetric and antisymmetric tensors. We even showed that if has dimension and a basis we can set up bases for and . Now let’s count how many vectors are in these bases and determine the dimensions of these spaces.
The easy one will be the antisymmetric case. Every basic antisymmetric tensor is given by antisymmetrizing an -tuple of basis vectors of . We may as well start out with this collection in order by their indices: . But we also know that we can’t have any repeated vectors or else the whole thing collapses. So the basis for consists of subsets of the basis for . There are basis vectors overall, and we must pick of them. But we know how to count these. This is a number of combinations:
Now what about symmetric tensors? We can’t do quite the same thing, since now we can allow repetitions in our lists. Instead, what we’ll do is this: instead of just a list of basis vectors of , consider writing the indices out in a line and drawing dividers between different indices. For example, consider th basic tensor of : . First, it becomes the list of indices
Now we divide from , from , from , and form .
Since there are five choices of an index, there will always be four dividers. And we’ll always have four indices since we’re considering the fourth tensor power. That is, a basic symmetric tensor corresponds to a choice of which of these eight slots to put the four dividers in. More generally if has dimension then a basic tensor in has indices separated by dividers. Then the dimension is again given by a number of combinations:
Permutations and Combinations
Okay, here’s a classic bit of combinatorics that I’m almost surprised I never mentioned before: how to count permutations and combinations.
Back when I defined permutation groups, I gave a quick argument about how to count their orders. That is, how many elements are in the group . If we’re permuting the set of numbers through , first we have to choose where to send . There are possible choices of how to do this. Then we have to choose where to send , which can be anywhere but the place we sent . Thus there are choices of how to do this. And so on we go, making choices of where to send , choices of where to send . And so in total we have permutations.
Notice that what we’ve really done there is picked out an ordering on the set . More specifically, we’ve picked an ordering of the whole set. But what if we want to pick out only an ordered subset? That is, what if we want to pick out an ordered list of numbers from the set with no repetitions? Well, since there’s only one order type for each finite cardinal number, let’s just pick one as the model for our ordered list. Specifically, we’ll use the numbers from to , in their usual order.
Now we’ll take an injective function from to . The image subset will be our list, and we carry over the order from onto it. But now we can reuse the same argument as above! First we have choices where to send , and then choices where to send , and so on until we have choices of where to send . So there are such ordered subsets of length . This gets called the number of permutations with elements, and written as , or or . Notice that when this reduces to the factorial as above.
There’s an easy way to express this number in notation we already know. First we can multiply all the numbers from to , and then we can divide out the ones from to . That is, . What does this mean? Well, the up top means that we’re ordering all the elements of our set. But then since we only care about the first we don’t care about the order on the last elements. What’s hiding here is that the group is secretly acting on the set of all permutations of our set by rearranging the last elements of the permutation. What’s more, it acts freely — with no fixed points — so every orbit has the same size: . But since we only care about the first places in the permutation, we’re really interested in the number of orbits. That is, the total number of permutations divided by the size of each orbit: . And this is the formula we came up with before.
This is a general principle in combinatorics. It’s often possible to see the set you’re trying to count as a larger set modulo the free action of some group. Then the cardinality of the set you’re interested in is the cardinality of the larger set divided by the order of the group.
To see this in action, what if we don’t care about the order of our subset? That is, we just want to pick out elements with no repetitions and no care about what order they come in. Well, first we can pick out an ordered set of elements. Then we can use the group to rearrange them. Any rearrangement is just as good as any other, and the group acts freely on the set of permutations. That is, the number of unordered subsets is the number of ordered subsets — — divided by the order of the group — . This number of unordered subsets we call the number of “combinations” of an element set with elements. This is often written as , or , or , or , and they all are given by the formula .
Antisymmetric Tensors
Let’s continue yesterday’s project by considering antisymmetric tensors today. Remember that we’re starting with a tensor representation of on the tensor power , which also carries an action of the symmetric group by permuting the tensorands. The antisymmetrizer (for today) is an element of the group algebra , and thus defines an intertwiner from to itself. Its image is thus a subrepresentation of acting on .
Now let’s again say has finite dimension and pick a basis of . Then we again have a basis of given by -tuples of basis elements of , and the permutation again acts by
So let’s see what the antisymmetrizer looks like.
It’s just like the symmetrizer, except we have a sign in each term given by the signum representation of the symmetric group.
Immediately this tells us something very interesting: if the same basis element of shows up twice in the basic tensor, then the antisymmetrization of that tensor is automatically zero! This is because there’s some swap exchanging the two slots where we find the two copies. We can divide the permutations in into two collections: those which can be written as a series of swaps starting with and those which can’t. And throwing on at the beginning of the list just exchanges these two groups for each other in a bijection. If you don’t see that immediately, consider how it’s really the exact same thing as how we solved the airplane seat problem almost two years ago, and Susan’s related Thanksgiving seating problem. Then each permutation in one collection is paired with one in the other collection, and the two in a pair differ by only one swap. This means that one of them will get sent to in the signum representation, and one will get sent to . Since the permuted tensors they give are the same, the two terms will exactly cancel each other out, and we’ll just end up adding together a bunch of zeroes. Neat!
Just like for the symmetric tensors, many different basic tensors will antisymmetrize to the same basic antisymmetric tensor. Again, we use the unique permutation which puts the tensorands in order, but now we use the fact above: if any basis vector of is repeated, then the antisymmetrization is automatically zero. Thus we throw out all these vectors and only consider those -tuples with . This makes it really easy to count the dimension of , because this is just the number of subsets of cardinality of a set of cardinality .
As an example, let’s antisymmetrize . We write out a sum of all the permutations, with signs as appropriate:
We have no repeated terms here because if we did they’d pair off with opposite signs and cancel out. Now you might notice that we didn’t start with the tensorands in order. If we put them in order, we have to use an odd permutation to get . Then we’ll get the same result, but with the opposite signs:
It’s a different result, yes, but it serves just as well as a basic antisymmetric tensor.
Symmetric Tensors
Wow, people are loving my zero-knowledge test. It got 1,743 views yesterday, thanks to someone redditing it. Anywho…
Today and tomorrow I want to take last Friday’s symmetrizer and antisymmetrizer and apply them to the tensor representations of , which we know also carry symmetric group representations. Specifically, the th tensor power carries a representation of by permuting the tensorands, and this representation commutes with the representation of . Then since the symmetrizer and antisymmetrizer are elements of the group algebra , they define intertwiners from to itself. The their images are not just subspaces on which the symmetric group acts nicely, but subrepresentations of symmetric and antisymmetric tensors — and , respectively.
Now it’s important (even if it’s not quite clear why) to emphasize that we’ve defined these representations without ever talking about a basis for . However, let’s try to get a better handle on what such a thing looks like by assuming has finite dimension and picking a basis . Then we have bases for tensor powers: a basis element of the th tensor power is given by an -tuple of basis elements for . We’ll write a general one like .
How does a permutation act on such a basis element? Well, basis elements are pure tensors, so the permutation simply permutes these basis tensorands. That is:
So the space of symmetric tensors is the image of under the action of the symmetrizer. And so it’s going to be spanned by the images of a basis for , which we can calculate now. The symmetrizer is an average of all the permutations in the symmetric group, so we find
Now we notice something here: if two basic tensors are related by a permutation of their tensorands, then the symmetrizer will send them to the same symmetric tensor! This means that we can choose a preimage for each basic symmetric tensor. Just use whatever permutation we need to put the -tuple of tensorands into order. That is, always select . Given any basic tensor, there is a unique permutation of its tensorands which is in this order.
As an explicit example, let’s consider what happens when we symmetrize the tensor . First of all, we toss it into the proper order, since this won’t change the symmetrization: . Now we write out a sum of all the permutations of the three tensorands, with the normalizing factor out front
Some of these terms are repeated, since we have two copies of in this tensor. So we collect these together and cancel off some of the normalizing factor to find
Now no matter how we rearrange the tensorands we’ll get back this same tensor again.
Tomorrow we’ll apply the same sort of approach to the antisymmetrizer.
The Symmetrizer and Antisymmetrizer
Today we’ll introduce two elements of the group algebra of the symmetric group which have some interesting properties. Each one, given a representation of the symmetric group, will give us a subrepresentation of that representation.
The symmetrizer gets its name from the way that it takes an arbitrary vector and turns it into one that the symmetric group will act trivially on. Specifically, we use the element
That is, we take all permutations in the symmetric group and — in a sense — average them out. If we compose this with any permutation we find
But as runs over all the permutations in the group, does as well. So this is just the symmetrizer back again. The upshot is that if we have a representation we find that
Thus every vector in the image of is left unchanged by the action of any permutation. That is, is a subrepresentation on which acts trivially.
The antisymmetrizer, on the other hand, on the other hand, will give us vectors on which the symmetric group acts by the signum representation. We use the group algebra element
Now if we compose this with any permutation we find
Now given a representation we find that
Thus every vector in the image of is multiplied by the signum of any permutation. That is, is a subrepresentation on which acts by the signum representation.
Now, one thing to be careful about: I haven’t said that the subrepresentations are nontrivial. That is, when we (anti)symmetrize a representation, the subrepresentation we get may be zero — maybe no vectors in the representation transform trivially or by the signum representation. In fact, let’s check what happens when we multiply the symmetrizer and antisymmetrizer:
Where the sum comes to zero because we’re just adding up terms where and where , and so everything cancels out. That is, the symmetric part of an antisymmetrized representation is trivial. Similarly, the antisymmetric part of a symmetrized representation is trivial.
The Signum Representation
I’ve just noticed that there’s a very important representation I haven’t mentioned. Well, actually, I mentioned it in passing while talking about Rubik’s group, but not very explicitly. And it’s a very important one.
Way back when I defined permutation groups I talked about a permutation being even or odd. Remember that we showed that permutation can be written out as a composite of transpositions which swap two objects. In general this can be done in more than one way, but if it takes an even number of swaps to write a permutation in one way, then it will take an even number of swaps in any other way, and similarly for permutations requiring an odd number of swaps. In this way we separate out permutations into the “even” and “odd” collections.
The composite of two even permutations or two odd permutations is even, while the composite of an even and an odd permutation is odd. This is just like the multiplication table of the group , with “even” for the group’s identity and “odd” for the other group element . That is, we have a homomorphism for every permutation group .
Now to make this into a representation we’re going to use a one-dimensional representation of . We have to send the group identity to the field element , but we have a choice to make for the image of . We need to send it to some field element , and this element must satisfy for this to be a representation. We could choose , but this just sends everything to the identity, which is the trivial group representation. There may be other choices around, but the only one we know always exist is (note we’re tacitly assuming that .
So we define the one-dimensional signum representation of by sending all the even permutations to the matrix whose single entry is , and sending all the odd permutations to the matrix whose single entry is . Often we’ll just ignore the “matrix” fact in here, and just say that the signum of an even permutation is and the signum of an odd permutation is . But secretly we’re always taking this and multiplying it by something else, so we’re always using it as a linear transformation anyway.
Do Short Exact Sequences of Representations Split?
We’ve seen that the category of representations is abelian, so we have all we need to talk about exact sequences. And we know that some of the most important exact sequences are short exact sequences. We also saw that every short exact sequence of vector spaces splits. So does the same hold for representations? It turns out that no, they don’t always, and I’ll give an example to show what can happen.
Consider the group of integers and the two dimensional representation defined by:
Verify for yourself that this actually does define a representation of the group of integers.
Now it’s straightforward to see that all these linear transformations send every vector of the form to itself. This defines a one-dimensional subspace fixed by the representation — a subrepresentation defined by:
Then there must be a quotient representation , and we can arrange them into a short exact sequence: . The question, then, is whether this is isomorphic to the split exact sequence . That is, can we find an isomorphism compatible with the the inclusion map from and the projection map onto ?
First off, let’s write the direct sum representation a little more explicitly. The direct sum acts on pairs of field elements by on the first and on the second, with no interaction between them. That is, we can write the representation as
And we’re looking for some isomorphism so that for every we get from the matrix to the matrix by conjugation. Explicitly, we’ll need a matrix . But we also need to make sure that as a subrepresentation of is sent to as a subrepresentation of . That is we must satisfy
Thus and right off the bat! Now the intertwining condition (equivalent to the conjugation) is that
But this says that for all , and this is clearly impossible!
So here’s an example where a short exact sequence of representations can not be split. At some point later we’ll see that in many cases we’re interested in they do split, but for now it’s good to see that they don’t always work out so nicely.
The Category of Representations is Abelian
We’ve been considering the category of representations of an algebra , and we’re just about done showing that is abelian.
First of all, the intertwiners between any two representations form a vector space, which is really an abelian group plus stuff. Since the composition of intertwiners is bilinear, this makes into an -category. Secondly, we can take direct sums of representations, which is a categorical biproduct. Thirdly, every intertwiner has a kernel and a cokernel.
The only thing we’re missing is that every monomorphism and every epimorphism be normal. That is, every monomorphism should actually be the kernel of some intertwiner, and every epimorphism should actually be the cokernel of some intertwiner. So, given representations and , let’s consider a monomorphic intertwiner .
As for linear maps, it’s straightforward to show that is monomorphic if and only if its kernel is trivial. Specifically, we can consider the inclusion and the zero map . It’s easy to see that , and so the left-cancellation property shows that , which is only possible if . So a monomorphism has a trivial kernel. Thus the underlying linear map is an isomorphism of onto the image . Then this subrepresentation is exactly the kernel of the quotient map . And so the monomorphism is the kernel of some map. The proof that any epimorphism is normal is dual.
And so we have established that the category of representations of the algebra is abelian. This allows us to bring in all the machinery of homological algebra, if we should so choose. In particular, we can talk about exact sequences, which can be useful from time to time.
Kernels and Images of Intertwiners
The next obvious things to consider are the kernel and the image of an intertwining map. So let’s say we’ve got a representation , a representation , and an intertwiner defined by the linear map which satisfies for all .
Now the linear map immediately gives us two subspaces: the kernel and the image . And it turns out that each of these is actually a subrepresentation. Showing this isn’t difficult. A subrepresentation is just a subspace that gets sent to itself under the action on the whole space, so we just have to check that always sends vectors in back to this subspace, and that always sends vectors in back into this subspace.
First off, is in the kernel of if . Then we calculate
which shoes that is also in the kernel of .
On the other hand, if is in the image of , then there is some so that . We calculate
And so is also in the image of .
So we’ve seen that the image and kernel of an intertwining map have well-defined actions of , and so we have subrepresentations. Immediately we can conclude that the coimage and the cokernel are quotient representations.