The Unapologetic Mathematician

Mathematics for the interested outsider

Vector-Valued Functions

Now we know how to modify the notion of the derivative of a function to deal with vector inputs by defining the differential. But what about functions that have vectors as outputs?

Well, luckily when we defined the differential we didn’t really use anything about the space where our function took its values but that it was a topological vector space. Indeed, when defining the differential of a function f:\mathbb{R}^m\rightarrow\mathbb{R}^n we need to set up a new function that takes a point x in the Euclidean space \mathbb{R}^m and a displacement vector t in \mathbb{R}^m as inputs, and which gives a displacement vector df(x;t) in \mathbb{R}^n as its output. It must be linear in the displacement, meaning that we can view df(x) as a linear transformation from \mathbb{R}^m to \mathbb{R}^n. And it must satisfy a similar approximation condition, replacing the absolute value with the notion of length in \mathbb{R}^n: for every \epsilon>0 there is a \delta>0 so that if \delta>\lVert t\rVert_{\mathbb{R}^m}>0 we have

\displaystyle\lVert\left[f(x+t)-f(x)\right]-df(x)t\rVert_{\mathbb{R}^n}<\epsilon\lVert t\rVert_{\mathbb{R}^m}

From here on we’ll just determine which norm we mean by context, since we only have one norm on each vector space.

Okay, so we can talk about differentials of vector-valued functions: the differential of f at a point x (if it exists) is a linear transformation df(x) that turns displacements in the input space into displacements in the output space, and does so in the way that most closely approximates the action of the function itself. But how do we define the function?

If we pick an orthonormal basis e_i for \mathbb{R}^n we can write the components of f as separate functions. That is, we say

\displaystyle f(x)=f^i(x)e_i

Now I assert that the differential can be taken component-by-component, just as continuity works: df(x)=df^i(x)e_i. On the left is the differential of f as a vector-valued function, while on the right we find the differentials of the several real-valued functions f^i. The differential exists if and only if the component differentials do.

First, from components to the vector-valued function. Clearly this definition of df(x) gives us a linear map from displacements in \mathbb{R}^m to displacements in \mathbb{R}^n. But does it satisfy the approximation inequality? Indeed, for every \epsilon>0 we can find a \delta so that all the inequalities

\displaystyle\left\lvert\left[f^i(x+t)-f^i(x)\right]-df^i(x)t\right\rvert<\frac{\epsilon}{n}\lVert t\rVert

are satisfied when \delta>\lVert t\rVert>0. Of course, there are different \deltas that work for each component, but we can pick the smallest of them. Then it’s a simple matter to find

\displaystyle\begin{aligned}\left\lVert\left[f(x+t)-f(x)\right]-df(x)t\right\rVert&=\left\lVert\left(\left[f^i(x+t)-f^i(x)\right]-df^i(x)t\right)e_i\right\rVert\\&\leq\left\lvert\left[f^i(x+t)-f^i(x)\right]-df^i(x)t\right\rvert\lVert e_i\rVert\\&=\sum\limits_{i=1}^n\left\lvert\left[f^i(x+t)-f^i(x)\right]-df^i(x)t\right\rvert\\&<\sum\limits_{i=1}^n\frac{\epsilon}{n}\lVert t\rVert\\&=\epsilon\lVert t\rVert\end{aligned}

so if the component functions are differentiable, then so is the function as a whole.

On the other hand, if the differential df(x) exists then for every \epsilon>0 there exists a \delta>0 so that if \delta>\lVert t\rVert>0 we have

\displaystyle\lVert\left[f(x+t)-f(x)\right]-df(x)t\rVert<\epsilon\lVert t\rVert

But then it’s easy to see that

\displaystyle\begin{aligned}\left\lvert\left[f^k(x+t)-f^k(x)\right]-df^k(x)t\right\rvert&=\sqrt{\left\lvert\left[f^k(x+t)-f^k(x)\right]-df^k(x)t\right\rvert^2}\\&\leq\sqrt{\sum\limits_{i=1}^n\left\lvert\left[f^i(x+t)-f^i(x)\right]-df^i(x)t\right\rvert^2}\\&=\lVert\left[f(x+t)-f(x)\right]-df(x)t\rVert\\&<\epsilon\lVert t\rVert\end{aligned}

and so each of the component differentials exists.

Finally, I should mention that if we also pick an orthonormal basis \tilde{e}_j for the input space \mathbb{R}^m we can expand each component differential df^i(x) in terms of the dual basis dx^j:

\displaystyle df^i(x)=\frac{\partial f^i}{\partial x^1}dx^1+\dots+\frac{\partial f^i}{\partial x^m}dx^m=\frac{\partial f^i}{\partial x^j}dx^j

Then we can write the whole differential df(x) out as a matrix whose entry in the ith row and jth column is \frac{\partial f^i}{\partial x^j}. If we write a displacement in the input as an m-dimensional column vector we find our estimate of the displacement in the output as an n-dimensional column vector:

\displaystyle\begin{pmatrix}df^1(x;t)\\\vdots\\df^n(x;t)\end{pmatrix}=\begin{pmatrix}\frac{\partial f^1}{\partial x^1}&\dots&\frac{\partial f^1}{\partial x^m}\\\vdots&\ddots&\vdots\\\frac{\partial f^n}{\partial x^1}&\dots&\frac{\partial f^n}{\partial x^m}\end{pmatrix}\begin{pmatrix}t^1\\\vdots\\t^m\end{pmatrix}

About these ads

October 6, 2009 - Posted by | Analysis, Calculus

5 Comments »

  1. Nicely done! Once again, you’ve brought order out of chaos.

    Comment by Jonathan Vos Post | October 6, 2009 | Reply

  2. [...] since each of the several is differentiable we can pick our radius so that all of the [...]

    Pingback by The Chain Rule « The Unapologetic Mathematician | October 7, 2009 | Reply

  3. [...] derivatives of . This works out because when we consider differentials as linear transformations, the matrix entries are the partial derivatives. The composition of the linear transformations and is given by the product of these matrices, and [...]

    Pingback by Cauchy’s Invariant Rule « The Unapologetic Mathematician | October 8, 2009 | Reply

  4. [...] just as we did for vector-valued functions, we’ll just take the differentials of each of these components separately, and then cobble [...]

    Pingback by Higher-Order Differentials « The Unapologetic Mathematician | October 16, 2009 | Reply

  5. [...] some open region in . That is, if we pick a basis and coordinates of , then the function is a vector-valued function of real variables with components . The differential, then, is itself a vector-valued function [...]

    Pingback by The Jacobian « The Unapologetic Mathematician | November 11, 2009 | Reply


Leave a Reply

Fill in your details below or click an icon to log in:

WordPress.com Logo

You are commenting using your WordPress.com account. Log Out / Change )

Twitter picture

You are commenting using your Twitter account. Log Out / Change )

Facebook photo

You are commenting using your Facebook account. Log Out / Change )

Google+ photo

You are commenting using your Google+ account. Log Out / Change )

Connecting to %s

Follow

Get every new post delivered to your Inbox.

Join 391 other followers

%d bloggers like this: