Today I want to give two examples of convergent series that turn out to be extremely useful for comparisons.
First we have the geometric series whose terms are the sequence for some constant ratio . The sequence of partial sums is
If we can multiply this sum by to find
Then as goes to infinity, this sequence either blows up (for ) or converges to (for ). In the border case we can also see that the sequence of partial sums fails to converge. Thus the geometric series converges if and only if , and we have a nice simple formula telling us the sum.
The other one I want to hit is the so-called -series, whose terms are starting at . Here we use the integral test to see that
so the sum and integral either converge or diverge together. If the integral gives , which converges for and diverges for .
If we get , which diverges. In this case, though, we have a special name for the limit of the difference . We call it “Euler’s constant”, and denote it by . That is, we can write
where is an error term whose magnitude is bounded by .
In general we have no good value for the sums of these series, even where they converge. It takes a bit of doing to find , as Euler did in 1735 (solving the “Basel Problem” that had stood for almost a century), and now we have values for other even natural number values of . The sum is known as Apéry’s constant, after Roger Apéry who showed that it was irrational in 1979. Yes, we didn’t even know whether it was a rational number or not until 30 years ago. We have basically nothing about odd integer values of .
If we say instead of , and let take complex values (no, I haven’t talked about complex numbers yet, but some of you know what they are) we get Riemann’s function , which is connected to some of the deepest outstanding questions in mathematics today.
Sorry for the delay. Students are panicking on the last day of classes and I have to write up a make-up exam for one who has a conflict at the scheduled time…
I’ve spent a goodly amount of time last week trying to craft a proof hinging on converting the infinite sum to an improper integral using the integrator , and comparing that one to those using the integrators and . But it doesn’t seem to be working. If you can make a go of it, I’ll be glad to hear it. Instead, here’s a proof adapted from Apostol.
We let be a positive decreasing function defined on some ray. For our purposes, let’s let it be , but we could use any other and adapt the proof accordingly. What we require in any case, though, is that the limit . We define three sequences:
First off, I assert that is nonincreasing, and sits between and . That is, we have the inequalities
To see this, first let’s write the integral defining as a sum of integrals over unit steps and notice that gives an upper bound to the size of on the interval . Thus we see:
From here we find that .
On the other hand, we see that . Reusing some pieces from before, we see that this is
which verifies that the sequence is decreasing. And it’s easy to check that , which completes our verification of these inequalities.
Now is a monotonically decreasing sequence, which is bounded below by , and so it must converge to some finite limit . This is the difference between the sum of the infinite series and the improper integral. Thus if either the sum or the integral converges, then the other one must as well.
We can actually do a little better, even, than simply showing that the sum and integral either both converge or both diverge. We can get some control on how fast the sequence converges to . Specifically, we have the inequalities , so the difference converges as fast as the function goes to zero.
To get here, we look back at the difference of two terms in the sequence:
So take this inequality for and add it to that for . We see then that . Then add the inequality for , and so on. At each step we find . So as goes to infinity, we get the asserted inequalities.