Chapter 5 Taylor Series
One of the most useful tools for solving problems in mathematics is the capability to approximate an arbitrary function using polynomials. There are a number of different methods for approximating functions that you will find out about in your degree including - Taylor series - Fourier series - Interpolation - wavelets (maybe) - radial basis functions (Jeremy’s research area) - neural networks (Alexander Gorban’s and Ivan Tyukin’s research areas).
5.1 Taylor polynomials
We use polynomials because they can be computed on a machine only using multiplication and addition.
We have already met one Taylor polynomial, in Section 3.1 when we were looking at approximating a function by its tangent in Newton’s method. The idea for Taylor series is that we will match the each derivative of a polynomial with the derivative of the function at a particular point b (this is called Hermite interpolation’’ in the trade).
We are going to start with the simple case b=0. Let us try with a degree 2 polynomial and an arbitrary function f. Let p2(x)=a0+a1x+a2x2, and we want p2(0)=f(0),p′2(0)=f′(0),andp2″ If we let x=0 in (5.1) we get a_0 = f(0). Let us now differentiate p_2^{'}(x)=a_1+2a_2 x. If we substitute x=0 we get p_2^{'}(0)=a_1=f^{'}(0). Differentiating a second time we get p_2^{''}(x)=2a_2. Substituting in x=0 gives p_2^{''}(0)=2a_2=f^{''}(0), so that a_2 = {f^{''}(0) \over 2}. Therefore our approximating polynomial is p_2(x)=f(0)+f^{'}(0) x+{f^{''}(0) \over 2}x^2. \tag{5.2}
Example 5.1 Let us test the idea out on a particular function f(x)=\sqrt{1+x}. Then \begin{eqnarray} f(0) & = & \sqrt{1} = 1 \\ f'(x) & = & {1 \over 2 \sqrt{1+x}} \Rightarrow f'(0)={1 \over 2} \\ f''(x) & = & -{3 \over 4 (\sqrt{1+x})^3} \Rightarrow f''(0) = {3 \over 4}. \end{eqnarray} Substituting into (5.2) we get P_2(x)=1+{x \over 2} - {3x^2 \over 8}x^2.
Now let us see if this is a good approximation. Let us use the last equation to approximate \sqrt{1.1}=1.0488088 to 7 decimal places. Our approximation is P_2(0.1) = 1+{0.1 \over 2} - {3 (0.1)^2 \over 8} = 1.04625. Hence our estimate is out by approximately 0.003.
Suppose we approximate \sqrt{1.2}=1.0954451 to 7 decimal places. Our approximation is P_2(0.2) = 1+{0.2 \over 2} - {3 (0.2)^2 \over 8} = 1.085. Hence our estimate is out by approximately 0.01, so the error has increased significantly. We will return to theoretical error estimates for Taylor series next semester.
We can see above that the higher the degree of the polynomial approximation, the better it gets. In this link
(http://mathfaculty.fullerton.edu/mathews/a2001/Animations/Interpolation/Series/Series.html)
you can examples of Taylor series approximations for a variety of functions and observe how they improve as you increase the degree of the polynomial.
More generally, we can look at the Taylor series for a point b. In order to compute this it is more convenient to use a different basis for the polynomials. A basis is an idea that you will see more of in linear algebra. It is a set of objects which we use to represent our information. For the point b I would like my polynomial to be written in the form p_n(x)=a_0+a_1(x-b)+a_2(x-b)^2+\cdots+a_n(x-b)^n = \sum_{i=0}^n a_i(x-b)^i. It will become clear in a second why this is a good idea. We will find a_i by matching the ith derivative of p_n with the ith derivative of our target function f, which we are trying to approximate. If we differentiate p_n j times we get p_n^{(j)}(x) = j! a_j+(j+1)!a_{j+1}(x-b)+{(j+2)! \over 2}(x-b)^2+\cdots+a_n{n! \over (n-j)!}(x-b)^{n-j}. If we put x=b in this we see that p_n^{(j)}(b) = j! a_j. Since we wish to make p_n^{(j)}(b)=f^{(j)}(b), we obtain a_j = {f^{(j)}(b) \over j!}. If we had used p_n (x)=a_0+a_1x+a_2x^2+\cdots+a_nx^n, then all of the terms greater than the jth term would not have become 0 and we would have got a horrible mess. Thus we have the following definition for the Taylor polynomial:
Remark. The selection of a good basis is one of the most important mathematical skills to develop. A bad choice of basis can make the problem almost impossible to solve. A good choice can make it trivial.
5.2 Polar form for complex numbers
Richard Feynman, the famous physicist called the following equation
our jewel'' and
the most remarkable formula in mathematics’’.

Richard Feynmann
Using the Taylor series for \exp(i \theta) we have \begin{eqnarray*} \exp(i \theta) & = & \sum_{n=0}^\infty {(i \theta)^n \over n!}\\ & = & \sum_{n=0}^\infty {i^n \theta^n \over n!}. \end{eqnarray*} Now, i^2=-1, i^3=-i and i^4=1, so that \begin{eqnarray*} \exp(i \theta) & = & \sum_{n=0}^\infty {i^n \theta^n \over n!} \\ & = & 1+i \theta - {\theta^2 \over 2!} - i{\theta^3 \over 3!}+{\theta^4 \over 4!}+{\theta^5 \over 5!}- \cdots \\ & = & 1 - {\theta^2 \over 2!}+{\theta^4 \over 4!}+ \cdots + i (\theta - {\theta^3 \over 3!} + {\theta^5 \over 5!} + \cdots) \\ & = & \left (\sum_{n=0}^\infty {(-1)^n \over (2n)!} \theta^{2n} \right )+i\left ( \sum_{n=0}^\infty {(-1)^{2n+1} \over (2n+1)!} \theta^{2n+1} \right ) \\ & = & \cos \theta + i \sin \theta. \end{eqnarray*}
Thus \exp(i \theta) is a complex number with real part \cos \theta and imaginary part \sin \theta. If plot this we can see that \exp(\theta \theta) is a complex number of length 1 at angle \theta to the real axis.

Polar coordinates
Example 5.3 Write z=2+4i in the form z=r\exp(i \theta).
The modulus of z r=\sqrt{2^2+4^2}=\sqrt{20}. We have \theta={\rm arg \,} z=\arctan(4/2)=1.11 radians.