## The Submodule of Invariants

If is a module of a Lie algebra , there is one submodule that turns out to be rather interesting: the submodule of vectors such that for all . We call these vectors “invariants” of .

As an illustration of how interesting these are, consider the modules we looked at last time. What are the invariant linear maps from one module to another ? We consider the action of on a linear map :

Or, in other words:

That is, a linear map is invariant if and only if it intertwines the actions on and . That is, .

Next, consider the bilinear forms on . Here we calculate

That is, a bilinear form is invariant if and only if it is associative, in the sense that the Killing form is:

## More New Modules from Old

There are a few constructions we can make, starting with the ones from last time and applying them in certain special cases.

First off, if and are two finite-dimensional -modules, then I say we can put an -module structure on the space of linear maps from to . Indeed, we can identify with : if is a basis for and is a basis for , then we can set up the dual basis of , such that . Then the elements form a basis for , and each one can be identified with the linear map sending to and all the other basis elements of to . Thus we have an inclusion , and a simple dimension-counting argument suffices to show that this is an isomorphism.

Now, since we have an action of on we get a dual action on . And because we have actions on and we get one on . What does this look like, explicitly? Well, we can write any such tensor as the sum of tensors of the form for some and . We calculate the action of on a vector :

In general we see that . In particular, the space of linear endomorphisms on is , and so it get an -module structure like this.

The other case of interest is the space of bilinear forms on a module . A bilinear form on is, of course, a linear functional on . And thus this space can be identified with . How does act on a bilinear form ? Well, we can calculate:

In particular, we can consider the case of bilinear forms on itself, where acts on itself by . Here we read

## New Modules from Old

There are a few standard techniques we can use to generate new modules for a Lie algebra from old ones. We’ve seen direct sums already, but here are a few more.

One way is to start with a module and then consider its dual space . I say that this can be made into an -module by setting

for all , , and . Bilinearity should be clear, so we just check the defining property of a module. That is, we take two Lie algebra elements and check

so for all , as desired.

Another way is to start with modules and and form their tensor product . Now we define a module structure on this space by

We check the defining property again. Calculate:

while

These are useful, and they’re only just the beginning.

## Reducible Modules

As might be surmised from irreducible modules, a reducible module for a Lie algebra is one that contains a nontrivial proper submodule — one other than or itself.

Now obviously if is a submodule we can form the quotient . This is the basic setup of a short exact sequence:

The question is, does this sequence split? That is, can we write as the direct sum of and some other submodule isomorphic to ?

First of all, let’s be clear that direct sums of modules do make sense. Indeed, if and are -modules then we can form an action on by defining it on each summand separately

Clearly the usual subspace inclusions and projections between , , and intertwine these actions, so they’re the required module morphisms. Further, it’s clear that .

So, do all short exact sequences of representations split? no. Indeed, let be the algebra of upper-triangular matrices, along with the obvious -dimensional representation. If we let be the basic column vector with a in the th row and elsewhere, then the one-dimensional space spanned by forms a one-dimensional submodule. Indeed, all upper-triangular matrices will send this column vector back to a multiple of itself.

On the other hand, it may be the case for a module that any nontrivial proper submodule has a complementary submodule with . In this case, is either irreducible or it’s not; if not, then any proper nontrivial submodule of will also be a proper nontrivial submodule of , and we can continue taking smaller submodules until we get to an irreducible one, so we may as well assume that is irreducible. Now the same sort of argument works for , showing that if it’s not irreducible it can be decomposed into the direct sum of some irreducible submodule and some complement, which is another nontrivial proper submodule of . At each step, the complement gets smaller and smaller, until we have decomposed entirely into a direct sum of irreducible submodules.

If is decomposable into a direct sum of irreducible submodules, we say that is “completely reducible”, or “decomposable”, as we did when we were working with groups. Any module where any nontrivial proper submodule has a complement is thus completely reducible; conversely, complete reducibility implies that any nontrivial proper submodule has a complement. Indeed, such a submodule must consist of some, but not all, of the summands of , and the complement will consist of the rest.

## Irreducible Modules

Sorry for the delay; it’s getting crowded around here again.

Anyway, an irreducible module for a Lie algebra is a pretty straightforward concept: it’s a module such that its only submodules are and . As usual, Schur’s lemma tells us that any morphism between two irreducible modules is either or an isomorphism. And, as we’ve seen in other examples involving linear transformations, all automorphisms of an irreducible module are scalars times the identity transformation. This, of course, doesn’t depend on any choice of basis.

A one-dimensional module will always be irreducible, if it exists. And a unique — up to isomorphism, of course — one-dimensional module will always exist for simple Lie algebras. Indeed, if is simple then we know that . Any one-dimensional representation must have its image in . But the only traceless matrix is the zero matrix. Setting for all does indeed give a valid representation of .

## Lie Algebra Modules

It should be little surprise that we’re interested in concrete actions of Lie algebras on vector spaces, like we were for groups. Given a Lie algebra we define an -module to be a vector space equipped with a bilinear function — often written satisfying the relation

Of course, this is the same thing as a representation . Indeed, given a representation we can define ; given an action we can define a representation by . The above relation is exactly the statement that the bracket in corresponds to the bracket in .

Of course, the modules of a Lie algebra form a category. A homomorphism of -modules is a linear map satisfying

We automatically get the concept of a submodule — a subspace sent back into itself by each — and a quotient module. In the latter case, we can see that if is any submodule then we can define . This is well-defined, since if is any other representative of then , and , so and both represent the same element of .

Thus, every submodule can be seen as the kernel of some homomorphism: the projection . It should be clear that every homomorphism has a kernel, and a cokernel can be defined simply as the quotient of the range by the image. All we need to see that the category of -modules is abelian is to show that every epimorphism is actually a quotient, but we know this is already true for the underlying vector spaces. Since the (vector space) kernel of an -module map is an -submodule, this is also true for -modules.

## All Derivations of Semisimple Lie Algebras are Inner

It turns out that all the derivations on a semisimple Lie algebra are inner derivations. That is, they’re all of the form for some . We know that the homomorphism is injective when is semisimple. Indeed, its kernel is exactly the center , which we know is trivial. We are asserting that it is also surjective, and thus an isomorphism of Lie algebras.

If we set and , we can see that . Indeed, if is any derivation and , then we can check that

This makes an ideal, so the Killing form of is the restriction of of the Killing form of . Then we can define to be the subspace orthogonal (with respect to ) to , and the fact that the Killing form is nondegenerate tells us that , and thus .

Now, if is an outer derivation — one not in — we can assume that it is orthogonal to , since otherwise we just have to use to project onto and subtract off that much to get another outer derivation that is orthogonal. But then we find that

since this bracket is contained in . But the fact that is injective means that for all , and thus . We conclude that and that , and thus that is onto, as asserted.

## Decomposition of Semisimple Lie Algebras

We say that a Lie algebra is the direct sum of a collection of ideals if it’s the direct sum as a vector space. In particular, this implies that , meaning that the bracket of any two elements from different ideals is zero.

Now, if is semisimple then there is a collection of ideals, each of which is simple as a Lie algebra in its own right, such that is the direct sum of these simple ideals. Further, every such simple ideal of is one in the collection — there’s no way to spread another simple ideal across two or more summands in this decomposition. And the Killing form of a summand is the restriction of the Killing form of to that summand, as we expect for any ideal of a Lie algebra.

If is any ideal then we can define the subspace of vectors in that are “orthogonal” to all the vectors in with respect to the Killing form . The associativity of shows that is also an ideal, just as we saw for the radical. Indeed, the radical of is just . Anyhow, Cartan’s criterion again shows that the intersection is solvable, but since is semisimple this means , and we can write .

So now we can use an induction on the dimension of ; if has no nonzero proper ideal, it’s already simple. Otherwise we can pick some proper ideal to get , where each summand has a lower dimension than . Any ideal of is an ideal of — the bracket with anything from is zero — so and must be semisimple as well, or else there would be a nonzero solvable ideal of . By induction, each one can be decomposed into simple ideals, so can as well.

Now, if is any simple ideal of , then is an ideal of . It can’t be zero, since if it were would be contained in , which is zero. Thus, since is simple, we must have . But the direct-sum decomposition tells us that , so all but one of these brackets must be zero, and that bracket must be itself. But this means for this simple summand, and — by the simplicity of — .

From this decomposition we conclude that for all semisimple we have . Every ideal and every quotient of must also be semisimple, since each must consist of some collection of the summands of .

## Back to the Example

Let’s go back to our explicit example of and look at its Killing form. We first recall our usual basis:

which lets us write out matrices for the adjoint action:

and from here it’s easy to calculate the Killing form. For example:

We can similarly calculate all the other values of the Killing form on basis elements.

So we can write down the matrix of :

And we can test this for degeneracy by taking its determinant to find . Since this is nonzero, we conclude that is nondegenerate, which we know means that is semisimple — at least in fields where .

## The Radical of the Killing Form

The first and most important structural result using the Killing form regards its “radical”. We never really defined this before, but it’s not hard: the radical of a binary form on a vector space is the subspace consisting of all such that for all . That is, if we regard as a linear map , the radical is the kernel of this map. Thus we see that is nondegenerate if and only if its radical is zero; we’ve only ever dealt much with nondegenerate bilinear forms, so we’ve never really had to consider the radical.

Now, the radical of the Killing form is more than just a subspace of ; the associative property tells us that it’s an ideal. Indeed, if is in the radical and are any other two Lie algebra elements, then we find that

thus is in the radical as well.

We recall that there was another “radical” we’ve mentioned: the radical of a Lie algebra is its maximal solvable ideal. This is not necessarily the same as the radical of the Killing form, but we can see that the radical of the form is contained in the radical of the algebra. By definition, if is in the radical of and is any other Lie algebra element we have

Cartan’s criterion then tells us that the radical of is solvable, and is thus contained in , the radical of the algebra. Immediately we conclude that if is semisimple — if — then the Killing form must be nondegenerate.

It turns out that the converse is also true. In fact, the radical of contains all abelian ideals . Indeed, if and then , and the square of this map sends into . Thus is nilpotent, and thus has trace zero, proving that , and that is contained in the radical of . So if the Killing form is nondegenerate its radical is zero, and there can be no abelian ideals of . But the derived series of eventually hits zero, and its last nonzero term is an abelian ideal of . This can only work out if is already zero, and thus is semisimple.

So we have a nice condition for semisimplicity: calculate the Killing form and check that it’s nondegenerate.