The best polynomial approximation, in the sense of minimizing the maximum error, can be found by the Remez algorithm.

I expected Mathematica to have a function implementing this algorithm, but apparently it does not have one.

(But see update below.

) It has a function named MiniMaxApproximation which sounds like Remez algorithm, and it’s close, but it’s not it.

To use this function you first have to load the FunctionApproximations package.

<< FunctionApproximations` Then we can use it, for example, to find a polynomial approximation to ex on the interval [-1, 1].

MiniMaxApproximation[Exp[x], {x, {-1, 1}, 5, 0}] This returns the polynomial 1.

00003 + 0.

999837 x + 0.

499342 x^2 + 0.

167274 x^3 + 0.

0436463 x^4 + 0.

00804051 x^5 And if we plot the error, the difference between ex and this polynomial, we see that we get a good fit.

But we know this isn’t optimal because there is a theorem that says the optimal approximation has equal ripple error.

That is, the absolute value of the error at all its extrema should be the same.

In the graph above, the error is quite a bit larger on the right end than on the left end.

Still, the error is not much larger than the smallest possible using 5th degree polynomials.

And the error is about 30x smaller than using a Taylor series approximation.

Plot[Exp[x] – (1 + x + x^2/2 + x^3/6 + x^4/25 + x^5/120), {x, -1, 1}] Update: Jason Merrill pointed out in a comment what I was missing.

Turns out MiniMaxApproximation finds an approximation that minimizes relative error.

Since ex doesn’t change that much over [-1, 1], the absolute error and relative error aren’t radically different.

There is no option to minimize absolute error.

When you look at the approximation error divided by ex you get the ripples you’d expect.

.