Do you remember that formula from calculus that states

Two questions: why is it true and why is it useful?

Why it’s true

What are we trying to accomplish with Maclaurin Series? We are trying to find a polynomial which equals the function (e.g. or ).

When are two functions equal? Two functions are equal if they have the same value for all inputs (all values of , in this case). One way that could be true is that the functions have the same value at , as well as the same derivative, the same second derivative, the same third derivative, etc. This isn’t air tight (and indeed, it isn’t always true), but it holds for many functions1 and should seem somewhat intuitive. If a function is the same at a certain point, and the amount it changes around that point (its first derivative) is the same, and the amount that changes around that point (its second derivative) is the same, all the way down, then how can these functions ever diverge? Well, often they don’t.

So, to recap, we’re going to look for a polynomial, , that has the same value as at , as well as the same derivative, for all derivatives.

What’s the form of a polynomial? It looks something like this:

0th derivative

We need . When all terms of the polynomial go to zero, other than the first. In other words:

We’ve discovered our first coefficient in :

1st derivative

We need .

We’ve discovered our second coefficient in :

2nd derivative

We need .

We’ve discovered our third coefficient in :

3rd derivative

We need .

We’ve discovered our fourth coefficient in :

And beyond

We could continue this pattern (seriously, try a few), but at this point you’re probably seeing a pattern emerge. The th term of the polynomial seems to be

making the entire function

Why it’s useful

Take this section with a grain of salt. It’s very possible that I don’t know the most important or useful practical applications of the Maclaurin Series. But here is my answer: polynomials are easy! They’re way nicer to deal with than arbitrary functions. In addition, derivative are (pretty) easy, and that’s all we need to turn an arbitrary function into polynomial.

For example, what’s the integral of ? Uhh…. wolfram alpha, anyone?

How about the integral of:

Sure, it’s a big equation, but it’s completely trivial to take that integral (remember, is just a constant).

And (big surprise) is the first few terms of the Maclaurin Series of .

They’re also fast. Let’s say you have a function which is expensive to compute, and whose input is changing relatively quickly. Maybe instead of recomputing your expensive function every time your input changes, you could approximate it with some large, but finite number of terms of its Maclaurin Series. Then, re-evaluating it will take almost no time at all!

Bonus section!

Hopefully you’ve followed the sections above, but if not, maybe a concrete example can help.

Let’s consider the function in the domain of . We can approximate this function using the first terms of the Maclaurin Series. As increases, our approximation looks closer and closer to the original .

Notice how the red line (the Maclaurin Series approximation of ) only seems to change when going from an odd number of terms to an even number of terms. Can you figure out why that is?

  1. Analytic functions is the technical term for the class of functions for which the Maclaurin Series (and the more general Taylor Series) holds.