One-Parameter Groups
Let be a differentiable map. For each
we can define the differentiable map
by
. We call the collection
a “one-parameter group” of diffeomorphisms — since it has, obviously, a single parameter — so long as it satisfies the two conditions
and
. That is,
is a homomorphism from the additive group of real numbers to the diffeomorphism group of
. Indeed, each
is a diffeomorphism of
— a differentiable isomorphism of the manifold to itself — with inverse
If we define a vector field by
then
is a flow for this vector field. Indeed, it’s a maximal flow, since it’s defined for all time at each point.
Conversely, if is the maximal flow of a vector field
, then
defines something like a one-parameter subgroup. Indeed, “flowing forward” by
and the again by
is the same as flowing forward by
along each integral curve, and so
wherever both sides of this equation are well-defined. But they might not, since even if both
and
are in
for all
the point
might not be. But if every integral curve can be extended for all times, then we call the vector field “complete” and conclude that its maximal flow is a one-parameter group of diffeomorphisms.
The Maximal Flow of a Vector Field
Given a smooth vector field we know what it means for a curve
to be an integral curve of
. We even know how to find them by starting at a point
and solving differential equations as far out as we can. For every
, let
be the maximal open interval containing
on which we can define the integral curve
with
.
Now, I say that there is a unique open set and a unique smooth map
such that
— the set
cuts out the interval
from the copy of
at
— and further
for all
. This is called the “maximal flow” of
.
Since there is some integral curve through each point , we can see that
. Further, it should be immediately apparent that
is also a local flow. What needs to be proven is that
is open, and that
is smooth.
Given a , let
be the collection of
for which there is a neighborhood of
contained in
on which
is differentiable. We will show that
is nonempty, open, and closed in
, meaning that it must be the whole interval.
Nonemptiness is obvious, since it just means that is contained in some local flow, which we showed last time. Openness also follows directly from the definition of
.
As for closedness, let be any point in
, the closure of
. We know there exists some local flow
with
and
. Now pick an
close enough to
so that
and
— this is possible since
is in the closure of
and
is continuous. Then choose an interval
around
so that
for each
. And finally the continuity of
at
tells us that there is a neighborhood
of
so that
.
Now, is defined and differentiable on
, showing that
. Indeed, if
and
, then
and
, so
is defined. The curve
is an integral curve of
, and it equals
at
. Uniqueness tells us that
is defined, and
is thus differentiable at
.
Integral Curves and Local Flows
Let is a vector field on the manifold
and let
be any point in
. Then I say there exists a neighborhood
of
, an interval
around
, and a differentiable map
such that
for all and
. These should look familiar, since they’re very similar to the conditions we wrote down for the flow of a differential equation.
It might help a bit to clarify that is the inclusion
of the canonical vector
which points in the direction of increasing
. That is,
includes the interval
into
“at the point
“, and thus its derivative carries along its tangent bundle. At each point of an (oriented) interval
there’s a canonical vector, and
is the image of that vector.
Further, take note that we can write the left side of our second condition as
The chain rule lets us combine these two outer derivatives into one:
But this is exactly how we defined the derivative of a curve! That is, we can write down a function which satisfies
for every
. We call such a curve an “integral curve” of the vector field
, and when they’re collected together as in
we call it a “local flow” of
.
So how do we prove this? We just take local coordinates and use our good old existence theorem! Indeed, if is a coordinate patch around
then we can set
,
, and
where the are the components
of
relative to the given local coordinates.
Now our existence theorem tells us there is a neighborhood of
, an interval
around
, and a map
satisfying the conditions for a flow. Setting
and
we find our local flow.
We can also do the same thing with our uniqueness theorem: if and
are two integral curves of
defined on the same interval
, and if
for some
, then
.
Thus we find the geometric meaning of that messy foray into analysis: a smooth vector field has a smooth local flow around every point, and integral curves of vector fields are unique.
Identifying Vector Fields
We know what vector fields are on a region , but to identify them in the wild we need to verify that a given function sending each
to a vector in
is smooth. This might not always be so easy to check directly, so we need some equivalent conditions. First we need to define how vector fields act on functions.
If is a vector field and
is a smooth function then we get another function
by defining
. Indeed,
, so it can take (the germ of) a smooth function at
and give us a number. Essentially, at each point the vector field defines a displacement, and we ask how the function
changes along this displacement. This action is key to our conditions, and to how we will actually use vector fields.
Firstly, if is a vector field — a differentiable function — and if
is a chart with
, then
is always smooth. Indeed, remember that
gives us a coordinate patch
on the tangent bundle. Since
is smooth and
is smooth, the composition
is also smooth. And thus each component is smooth on
.
Next, we do not assume that is a vector field — it is a function but not necessarily a differentiable one — but we assume that it satisfies the conclusion of the preceding paragraph. That is, for every chart
with
each
is smooth. Now we will show that
is smooth for every smooth
, not just those that arise as coordinate functions. To see this, we use the decomposition of
into coordinate vector fields:
which didn’t assume that was smooth, except to show that the coefficient functions were smooth. We can now calculate that
, since
But this means we can write
which makes a linear combination of the smooth (by assumption) functions
with the coefficients
, proving that it is itself smooth.
Okay, now I say that if is smooth for every smooth function
on some region
, then
is smooth as a function, and thus is a vector field. In this case around any
we can find some coordinate patch
. Now we go back up to the composition above:
Everything in sight on the right is smooth, and so the left is also smooth. But this is exactly what we need to check when we’re using the local coordinates and
to verify the smoothness of
at
.
The upshot is that when we want to verify that a function really is a smooth vector field, we take an arbitrary smooth “test function” and feed it into
. If the result is always smooth, then
is smooth. In fact, some authors take this as the definition, regarding the action of
on functions as fundamental, and only later talking in terms of its “value at a point”.
Coordinate Vector Fields
If we consider an open subset along with a suitable map
such that
is a coordinate patch, it turns out that we can actually give an explicit basis of the module
of vector fields over the ring
.
Indeed, at each point we can define the
coordinate vectors:
Thus each itself qualifies as a vector field in
as long as the map
is smooth. But we can check this using the coordinates
on
and the coordinate patch induced by
on the tangent bundle. With this choice of source and target coordinates the map is just the inclusion of
into the subspace
where the occurs in the
th place. This is clearly smooth.
Now we know at each point that the coordinate vectors span the tangent space. So let’s take a vector field and break up the vector
. We can write
which defines the as real-valued functions on
. It’s also smooth; we know that
is smooth by the definition of a vector field and the same choice of local coordinates as above, and passing from
to
is really just the projection onto the
th component of
in these local coordinates.
Since this now doesn’t really depend on we can write
which describes an arbitrary vector field as a linear combination of the coordinate vector fields times “scalar coefficient” functions
, showing that these coordinate vector fields span the whole module
. It should be clear that they’re independent, because if we had a nontrivial linear combination between them we’d have one between the coordinate vectors at at least one point, which we know doesn’t exist.
We should note here that just because is a free module — not a vector space since
might have a weird structure — in the case where
is a coordinate patch does not mean that all the
are free modules over their respective rings of smooth functions. But in a sense every “sufficiently small” open region
can be contained in some coordinate patch, and thus
will always be a free module in this case.
Vector Fields
At last, we get back to the differential geometry and topology. Let’s say that we have a manifold with tangent bundle
, which of course comes with a projection map
. If
is an open submanifold, we can restrict the bundle to the tangent bundle
with no real difficulty.
Now a “vector field” on is a “section” of this projection map. That is, it’s a function
so that the composition
is the identity map on
. In other words, to every point
we get a vector
at that point.
I should step aside to dissuade people from a common mistake. Back in multivariable calculus, it’s common to say that a vector field in is a function which assigns “a vector” to every point in some region
; that is, a function
. The problem here is that it’s assuming that every point gets a vector in the same vector space, when actually each point gets assigned a vector in its own tangent space.
The confusion comes because we know that if has dimension
then each tangent space
has dimension
, and thus they’re all isomorphic. Worse, when working over Euclidean space there is a canonical identification between a tangent space
and the space
itself, and thus between any two tangent spaces. But when we’re dealing with an arbitrary manifold there is no such canonical way to compare vectors based at different points; we have to be careful to keep them separate.
For each we have a collection of vector fields, which we will write
, or
for short. It should be apparent that if
is an open subspace we can restrict a vector field on
to one on
, which means we’re talking about a presheaf. In fact, it’s not hard to see that we can uniquely glue together vector fields which agree on shared domains, meaning we have a sheaf of vector fields.
For any , we can define the sum and scalar multiple of vector fields on
just by defining them pointwise. That is, if
and
are vector fields on
and
and
are real scalars, then we define
using the addition and scalar multiplication in . But that’s not all; we can also multiply a vector field
by any function
:
using the scalar multiplication in . This makes
into a sheaf of modules over the sheaf of rings
.
Lie Algebras from Associative Algebras
There is a great source for generating many Lie algebras: associative algebras. Specifically, if we have an associative algebra we can build a lie algebra
on the same underlying vector space by letting the bracket be the “commutator” from
. That is, for any algebra elements
and
we define
In fact, this is such a common way of coming up with Lie algebras that many people think of the bracket as a commutator by definition.
Clearly this is bilinear and antisymmetric, but does it satisfy the Jacobi identity? Well, let’s take three algebra elements and form the double bracket
We can find the other orders just as easily
and when we add these all up each term cancels against another term, leaving zero. Thus the commutator in an associative algebra does indeed act as a bracket.
Lie Algebras
One more little side trip before we proceed with the differential geometry: Lie algebras. These are like “regular” associative algebras in that we take a module (often a vector space) and define a bilinear operation on it. This much is covered at the top of the post on algebras.
The difference is that instead of insisting that the operation be associative, we impose different conditions. Also, instead of writing our operation like a multiplication (and using the word “multiplication”), we will write it as and call it the “bracket” of
and
. Now, our first condition is that the bracket be antisymmetric:
Secondly, and more importantly, we demand that the bracket should satisfy the “Jacobi identity”:
What this means is that the operation of “bracketing with ” acts like a derivation on the Lie algebra; we can apply
to the bracket
by first applying it to
and bracketing the result with
, then bracketing
with the result of applying the operation to
, and adding the two together.
This condition is often stated in the equivalent form
It’s a nice exercise to show that (assuming antisymmetry) these two equations are indeed equivalent. This form of the Jacobi identity is neat in the way it shows a rotational symmetry among the three algebra elements, but I feel that it misses the deep algebraic point about why the Jacobi identity is so important: it makes for an algebra that acts on itself by derivations of its own structure.
It turns out that we already know of an example of a Lie algebra: the cross product of vectors in . Indeed, take three vectors
,
, and
and try multiplying them out in all three orders:
and add the results together to see that you always get zero, thus satisfying the Jacobi identity.
Smooth Dependence on Initial Conditions
Now that we’ve got the existence and uniqueness of our solutions down, we have one more of our promised results: the smooth dependence of solutions on initial conditions. That is, if we use our existence and uniqueness theorems to construct a unique “flow” function satisfying
by setting — where
is the unique solution with initial condition
— then
is continuously differentiable.
Now, we already know that is continuously differentiable in the time direction by definition. What we need to show is that the directional derivatives involving directions in
exist and are continuous. To that end, let
be a base point and
be a small enough displacement that
as well. Similarly, let
be a fixed point in time and let
be a small change in time
But now our result from last time tells us that these solutions can diverge no faster than exponentially. Thus we conclude that
and so as this term must go to zero as well. Meanwhile, the second term also goes to zero by the differentiability of
. We can now see that the directional derivative at
in the direction of
exists.
But are these directional derivatives continuous. This turns out to be a lot more messy, but essentially doable by similar methods and a generalization of Gronwall’s inequality. For the sake of getting back to differential equations I’m going to just assert that not only do all directional derivatives exist, but they’re continuous, and thus the flow is .
Control on the Divergence of Solutions
Now we can establish some control on how nearby solutions to the differential equation
diverge. That is, as time goes by, how can the solutions move apart from each other?
Let and
be two solutions satisfying initial conditions
and
, respectively. The existence and uniqueness theorems we’ve just proven show that
and
are uniquely determined by this choice in some interval, and we’ll pick a
so they’re both defined on the closed interval
. Now for every
in this interval we have
Where is a Lipschitz constant for
in the region we’re concerned with. That is, the separation between the solutions
and
can increase no faster than exponentially.
So, let’s define to be this distance. Converting to integral equations, it’s clear that
and thus
Now Gronwall’s inequality tells us that , which is exactly the inequality we asserted above.