Do you remember that formula from calculus that states
Two questions: why is it true and why is it useful?
Why it’s true
What are we trying to accomplish with Maclaurin Series? We are trying to find a polynomial which equals the function (e.g. or ).
When are two functions equal? Two functions are equal if they have the same value for all inputs (all values of , in this case). One way that could be true is that the functions have the same value at , as well as the same derivative, the same second derivative, the same third derivative, etc. This isn’t air tight (and indeed, it isn’t always true), but it holds for many functions1 and should seem somewhat intuitive. If a function is the same at a certain point, and the amount it changes around that point (its first derivative) is the same, and the amount that changes around that point (its second derivative) is the same, all the way down, then how can these functions ever diverge? Well, often they don’t.
So, to recap, we’re going to look for a polynomial, , that has the same value as at , as well as the same derivative, for all derivatives.
What’s the form of a polynomial? It looks something like this:
We need . When all terms of the polynomial go to zero, other than the first. In other words:
We’ve discovered our first coefficient in :
We need .
We’ve discovered our second coefficient in :
We need .
We’ve discovered our third coefficient in :
We need .
We’ve discovered our fourth coefficient in :
We could continue this pattern (seriously, try a few), but at this point you’re probably seeing a pattern emerge. The th term of the polynomial seems to be
making the entire function
Why it’s useful
Take this section with a grain of salt. It’s very possible that I don’t know the most important or useful practical applications of the Maclaurin Series. But here is my answer: polynomials are easy! They’re way nicer to deal with than arbitrary functions. In addition, derivative are (pretty) easy, and that’s all we need to turn an arbitrary function into polynomial.
For example, what’s the integral of ? Uhh…. wolfram alpha, anyone?
How about the integral of:
Sure, it’s a big equation, but it’s completely trivial to take that integral (remember, is just a constant).
And (big surprise) is the first few terms of the Maclaurin Series of .
They’re also fast. Let’s say you have a function which is expensive to compute, and whose input is changing relatively quickly. Maybe instead of recomputing your expensive function every time your input changes, you could approximate it with some large, but finite number of terms of its Maclaurin Series. Then, re-evaluating it will take almost no time at all!
Hopefully you’ve followed the sections above, but if not, maybe a concrete example can help.
Let’s consider the function in the domain of . We can approximate this function using the first terms of the Maclaurin Series. As increases, our approximation looks closer and closer to the original .
Notice how the red line (the Maclaurin Series approximation of ) only seems to change when going from an odd number of terms to an even number of terms. Can you figure out why that is?
Analytic functions is the technical term for the class of functions for which the Maclaurin Series (and the more general Taylor Series) holds. ↩