Hubbry Logo
Taylor seriesTaylor seriesMain
Open search
Taylor series
Community hub
Taylor series
logo
8 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Taylor series
Taylor series
from Wikipedia

As the degree of the Taylor polynomial rises, it approaches the correct function. This image shows sin x and its Taylor approximations by polynomials of degree 1, 3, 5, 7, 9, 11, and 13 at x = 0.

In mathematics, the Taylor series or Taylor expansion of a function is an infinite sum of terms that are expressed in terms of the function's derivatives at a single point. For most common functions, the function and the sum of its Taylor series are equal near this point. Taylor series are named after Brook Taylor, who introduced them in 1715. A Taylor series is also called a Maclaurin series when 0 is the point where the derivatives are considered, after Colin Maclaurin, who made extensive use of this special case of Taylor series in the 18th century.

The partial sum formed by the first n + 1 terms of a Taylor series is a polynomial of degree n that is called the nth Taylor polynomial of the function. Taylor polynomials are approximations of a function, which become generally more accurate as n increases. Taylor's theorem gives quantitative estimates on the error introduced by the use of such approximations. If the Taylor series of a function is convergent, its sum is the limit of the infinite sequence of the Taylor polynomials. A function may differ from the sum of its Taylor series, even if its Taylor series is convergent. A function is analytic at a point x if it is equal to the sum of its Taylor series in some open interval (or open disk in the complex plane) containing x. This implies that the function is analytic at every point of the interval (or disk).

Definition

[edit]

The Taylor series of a real or complex-valued function f (x), that is infinitely differentiable at a real or complex number a, is the power series Here, n! denotes the factorial of n. The function f(n)(a) denotes the nth derivative of f evaluated at the point a. The derivative of order zero of f is defined to be f itself and (xa)0 and 0! are both defined to be 1. This series can be written by using sigma notation, as in the right side formula.[1] With a = 0, the Maclaurin series takes the form:[2]

Examples

[edit]

The Taylor series of any polynomial is the polynomial itself.

The Maclaurin series of 1/1 − x is the geometric series

So, by substituting x for 1 − x, the Taylor series of 1/x at a = 1 is

By integrating the above Maclaurin series, we find the Maclaurin series of ln(1 − x), where ln denotes the natural logarithm:

The corresponding Taylor series of ln x at a = 1 is

and more generally, the corresponding Taylor series of ln x at an arbitrary nonzero point a is

The Maclaurin series of the exponential function ex is

The above expansion holds because the derivative of ex with respect to x is also ex, and e0 equals 1. This leaves the terms (x − 0)n in the numerator and n! in the denominator of each term in the infinite sum.

History

[edit]

The ancient Greek philosopher Zeno of Elea considered the problem of summing an infinite series to achieve a finite result, but rejected it as an impossibility;[3] the result was Zeno's paradox. Later, Aristotle proposed a philosophical resolution of the paradox, but the mathematical content was apparently unresolved until taken up by Archimedes, as it had been prior to Aristotle by the Presocratic Atomist Democritus. It was through Archimedes's method of exhaustion that an infinite number of progressive subdivisions could be performed to achieve a finite result.[4] Liu Hui independently employed a similar method a few centuries later.[5]

In the 14th century, the earliest examples of specific Taylor series (but not the general method) were given by Indian mathematician Madhava of Sangamagrama.[6] Though no record of his work survives, writings of his followers in the Kerala school of astronomy and mathematics suggest that he found the Taylor series for the trigonometric functions of sine, cosine, and arctangent (see Madhava series). During the following two centuries his followers developed further series expansions and rational approximations.

In late 1670, James Gregory was shown in a letter from John Collins several Maclaurin series ( and ) derived by Isaac Newton, and told that Newton had developed a general method for expanding functions in series. Newton had in fact used a cumbersome method involving long division of series and term-by-term integration, but Gregory did not know it and set out to discover a general method for himself. In early 1671 Gregory discovered something like the general Maclaurin series and sent a letter to Collins including series for (the integral of ), (the integral of sec, the inverse Gudermannian function), and (the Gudermannian function). However, thinking that he had merely redeveloped a method by Newton, Gregory never described how he obtained these series, and it can only be inferred that he understood the general method by examining scratch work he had scribbled on the back of another letter from 1671.[7]

In 1691–1692, Newton wrote down an explicit statement of the Taylor and Maclaurin series in an unpublished version of his work De Quadratura Curvarum. It was the earliest explicit formulation of the general Taylor series.[8][9] However, this work by Newton was never completed and the relevant sections were omitted from the portions published in 1704 under the title Tractatus de Quadratura Curvarum.

It was not until 1715 that a general method for constructing these series for all functions for which they exist was finally published by Brook Taylor,[10] after whom the series are now named.

The Maclaurin series was named after Colin Maclaurin, a Scottish mathematician, who published a special case of the Taylor result in the mid-18th century.

Analytic functions

[edit]
The function e(−1/x2) is not analytic at x = 0: the Taylor series is identically 0, although the function is not.

If f (x) is given by a convergent power series in an open disk centred at b in the complex plane (or an interval in the real line), it is said to be analytic in this region. Thus for x in this region, f is given by a convergent power series

Differentiating by x the above formula n times, then setting x = b gives:

and so the power series expansion agrees with the Taylor series. Thus a function is analytic in an open disk centered at b if and only if its Taylor series converges to the value of the function at each point of the disk.

If f (x) is equal to the sum of its Taylor series for all x in the complex plane, it is called entire. The polynomials, exponential function ex, and the trigonometric functions sine and cosine, are examples of entire functions. Examples of functions that are not entire include the square root, the logarithm, the trigonometric function tangent, and its inverse, arctan. For these functions the Taylor series do not converge if x is far from b. That is, the Taylor series diverges at x if the distance between x and b is larger than the radius of convergence. The Taylor series can be used to calculate the value of an entire function at every point, if the value of the function, and of all of its derivatives, are known at a single point.

Uses of the Taylor series for analytic functions include:

  1. The partial sums (the Taylor polynomials) of the series can be used as approximations of the function. These approximations are good if sufficiently many terms are included.
  2. Differentiation and integration of power series can be performed term by term and is hence particularly easy.
  3. An analytic function is uniquely extended to a holomorphic function on an open disk in the complex plane. This makes the machinery of complex analysis available.
  4. The (truncated) series can be used to compute function values numerically, (often by recasting the polynomial into the Chebyshev form and evaluating it with the Clenshaw algorithm).
  5. Algebraic operations can be done readily on the power series representation; for instance, Euler's formula follows from Taylor series expansions for trigonometric and exponential functions. This result is of fundamental importance in such fields as harmonic analysis.
  6. Approximations using the first few terms of a Taylor series can make otherwise unsolvable problems possible for a restricted domain; this approach is often used in physics.

Approximation error and convergence

[edit]
The sine function (blue) is closely approximated by its Taylor polynomial of degree 7 (pink) for a full period centered at the origin.
The Taylor polynomials for ln(1 + x) only provide accurate approximations in the range −1 < x ≤ 1. For x > 1, Taylor polynomials of higher degree provide worse approximations.
The Taylor approximations for ln(1 + x) (black). For x > 1, the approximations diverge.

Pictured is an accurate approximation of sin x around the point x = 0. The pink curve is a polynomial of degree seven:

The error in this approximation is no more than |x|9 / 9!. For a full cycle centered at the origin (−π < x < π) the error is less than 0.08215. In particular, for −1 < x < 1, the error is less than 0.000003.

In contrast, also shown is a picture of the natural logarithm function ln(1 + x) and some of its Taylor polynomials around a = 0. These approximations converge to the function only in the region −1 < x ≤ 1; outside of this region the higher-degree Taylor polynomials are worse approximations for the function.

The error incurred in approximating a function by its nth-degree Taylor polynomial is called the remainder or residual and is denoted by the function Rn(x). Taylor's theorem can be used to obtain a bound on the size of the remainder.

In general, Taylor series need not be convergent at all. In fact, the set of functions with a convergent Taylor series is a meager set in the Fréchet space of smooth functions. Even if the Taylor series of a function f does converge, its limit need not be equal to the value of the function f (x). For example, the function

is infinitely differentiable at x = 0, and has all derivatives zero there. Consequently, the Taylor series of f (x) about x = 0 is identically zero. However, f (x) is not the zero function, so does not equal its Taylor series around the origin. Thus, f (x) is an example of a non-analytic smooth function.

In real analysis, this example shows that there are infinitely differentiable functions f (x) whose Taylor series are not equal to f (x) even if they converge. By contrast, the holomorphic functions studied in complex analysis always possess a convergent Taylor series, and even the Taylor series of meromorphic functions, which might have singularities, never converge to a value different from the function itself. The complex function e−1/z2, however, does not approach 0 when z approaches 0 along the imaginary axis, so it is not continuous in the complex plane and its Taylor series is undefined at 0.

More generally, every sequence of real or complex numbers can appear as coefficients in the Taylor series of an infinitely differentiable function defined on the real line, a consequence of Borel's lemma. As a result, the radius of convergence of a Taylor series can be zero. There are even infinitely differentiable functions defined on the real line whose Taylor series have a radius of convergence 0 everywhere.[11]

A function cannot be written as a Taylor series centred at a singularity; in these cases, one can often still achieve a series expansion if one allows also negative powers of the variable x; see Laurent series. For example, f (x) = e−1/x2 can be written as a Laurent series.

Generalization

[edit]

The generalization of the Taylor series does converge to the value of the function itself for any bounded continuous function on (0,∞), and this can be done by using the calculus of finite differences. Specifically, the following theorem, due to Einar Hille, that for any t > 0,[12]

Here Δn
h
is the nth finite difference operator with step size h. The series is precisely the Taylor series, except that divided differences appear in place of differentiation: the series is formally similar to the Newton series. When the function f is analytic at a, the terms in the series converge to the terms of the Taylor series, and in this sense generalizes the usual Taylor series.

In general, for any infinite sequence ai, the following power series identity holds:

So in particular,

The series on the right is the expected value of f (a + X), where X is a Poisson-distributed random variable that takes the value jh with probability et/h·(t/h)j/j!. Hence,

The law of large numbers implies that the identity holds.[13]

List of Maclaurin series of some common functions

[edit]

Several important Maclaurin series expansions follow. All these expansions are valid for complex arguments x.

Exponential function

[edit]
The exponential function ex (in blue), and the sum of the first n + 1 terms of its Taylor series at 0 (in red).

The exponential function (with base e) has Maclaurin series[14]

It converges for all x.

The exponential generating function of the Bell numbers is the exponential function of the predecessor of the exponential function:

Natural logarithm

[edit]

The natural logarithm (with base e) has Maclaurin series[15]

The last series is known as Mercator series, named after Nicholas Mercator (since it was published in his 1668 treatise Logarithmotechnia).[16] Both of these series converge for . (In addition, the series for ln(1 − x) converges for x = −1, and the series for ln(1 + x) converges for x = 1.)[15]

Geometric series

[edit]

The geometric series and its derivatives have Maclaurin series

All are convergent for . These are special cases of the binomial series given in the next section.

Binomial series

[edit]

The binomial series is the power series

whose coefficients are the generalized binomial coefficients[17]

(If n = 0, this product is an empty product and has value 1.) It converges for for any real or complex number α.

When α = −1, this is essentially the infinite geometric series mentioned in the previous section. The special cases α = 1/2 and α = −1/2 give the square root function and its inverse:[18]

When only the linear term is retained, this simplifies to the binomial approximation.

Trigonometric functions

[edit]

The usual trigonometric functions and their inverses have the following Maclaurin series:[19]

All angles are expressed in radians. The numbers Bk appearing in the expansions of tan x are the Bernoulli numbers. The Ek in the expansion of sec x are Euler numbers.[20]

Hyperbolic functions

[edit]

The hyperbolic functions have Maclaurin series closely related to the series for the corresponding trigonometric functions:[21]

The numbers Bk appearing in the series for tanh x are the Bernoulli numbers.[21]

Polylogarithmic functions

[edit]

The polylogarithms have these defining identities:

The Legendre chi functions are defined as follows:

And the formulas presented below are called inverse tangent integrals:

In statistical thermodynamics these formulas are of great importance.

Elliptic functions

[edit]

The complete elliptic integrals of first kind K and of second kind E can be defined as follows:

The Jacobi theta functions describe the world of the elliptic modular functions and they have these Taylor series:

The regular partition number sequence P(n) has this generating function:

The strict partition number sequence Q(n) has that generating function:

Calculation of Taylor series

[edit]

Several methods exist for the calculation of Taylor series of a large number of functions. One can attempt to use the definition of the Taylor series, though this often requires generalizing the form of the coefficients according to a readily apparent pattern. Alternatively, one can use manipulations such as substitution, multiplication or division, addition or subtraction of standard Taylor series to construct the Taylor series of a function, by virtue of Taylor series being power series. In some cases, one can also derive the Taylor series by repeatedly applying integration by parts. Particularly convenient is the use of computer algebra systems to calculate Taylor series.

First example

[edit]

In order to compute the 7th degree Maclaurin polynomial for the function

one may first rewrite the function as

the composition of two functions and The Taylor series for the natural logarithm is (using big O notation)

and for the cosine function

The first several terms from the second series can be substituted into each term of the first series. Because the first term in the second series has degree 2, three terms of the first series suffice to give a 7th-degree polynomial:

Since the cosine is an even function, the coefficients for all the odd powers are zero.

Second example

[edit]

Suppose we want the Taylor series at 0 of the function

The Taylor series for the exponential function is

and the series for cosine is

Assume the series for their quotient is

Multiplying both sides by the denominator and then expanding it as a series yields

Comparing the coefficients of with the coefficients of

The coefficients of the series for can thus be computed one at a time, amounting to long division of the series for and :

Third example

[edit]

Here we employ a method called "indirect expansion" to expand the given function. This method uses the known Taylor expansion of the exponential function. In order to expand (1 + x)ex as a Taylor series in x, we use the known Taylor series of function ex:

Thus,

Taylor series as definitions

[edit]

Classically, algebraic functions are defined by an algebraic equation, and transcendental functions (including those discussed above) are defined by some property that holds for them, such as a differential equation. For example, the exponential function is the function which is equal to its own derivative everywhere, and assumes the value 1 at the origin. However, one may equally well define an analytic function by its Taylor series.

Taylor series are used to define functions and "operators" in diverse areas of mathematics. In particular, this is true in areas where the classical definitions of functions break down. For example, using Taylor series, one may extend analytic functions to sets of matrices and operators, such as the matrix exponential or matrix logarithm.

In other areas, such as formal analysis, it is more convenient to work directly with the power series themselves. Thus one may define a solution of a differential equation as a power series which, one hopes to prove, is the Taylor series of the desired solution.

Taylor series in several variables

[edit]

The Taylor series may also be generalized to functions of more than one variable with[22]

For example, for a function that depends on two variables, x and y, the Taylor series to second order about the point (a, b) is

where the subscripts denote the respective partial derivatives.

Second-order Taylor series in several variables

[edit]

A second-order Taylor series expansion of a scalar-valued function of more than one variable can be written compactly as

where D f (a) is the gradient of f evaluated at x = a and D2 f (a) is the Hessian matrix. Applying the multi-index notation the Taylor series for several variables becomes

which is to be understood as a still more abbreviated multi-index version of the first equation of this paragraph, with a full analogy to the single variable case.

Example

[edit]
Second-order Taylor series approximation (in orange) of a function f (x,y) = ex ln(1 + y) around the origin.

In order to compute a second-order Taylor series expansion around point (a, b) = (0, 0) of the function

one first computes all the necessary partial derivatives:

Evaluating these derivatives at the origin gives the Taylor coefficients

Substituting these values in to the general formula

produces

Since ln(1 + y) is analytic in |y| < 1, we have

Comparison with Fourier series

[edit]

The trigonometric Fourier series enables one to express a periodic function (or a function defined on a closed interval [a,b]) as an infinite sum of trigonometric functions (sines and cosines). In this sense, the Fourier series is analogous to Taylor series, since the latter allows one to express a function as an infinite sum of powers. Nevertheless, the two series differ from each other in several relevant issues:

  • The finite truncations of the Taylor series of f (x) about the point x = a are all exactly equal to f at a. In contrast, the Fourier series is computed by integrating over an entire interval, so there is generally no such point where all the finite truncations of the series are exact.
  • The computation of Taylor series requires the knowledge of the function on an arbitrary small neighbourhood of a point, whereas the computation of the Fourier series requires knowing the function on its whole domain interval. In a certain sense one could say that the Taylor series is "local" and the Fourier series is "global".
  • The Taylor series is defined for a function which has infinitely many derivatives at a single point, whereas the Fourier series is defined for any integrable function. In particular, the function could be nowhere differentiable. (For example, f (x) could be a Weierstrass function.)
  • The convergence of both series has very different properties. Even if the Taylor series has positive convergence radius, the resulting series may not coincide with the function; but if the function is analytic then the series converges pointwise to the function, and uniformly on every compact subset of the convergence interval. Concerning the Fourier series, if the function is square-integrable then the series converges in quadratic mean, but additional requirements are needed to ensure the pointwise or uniform convergence (for instance, if the function is periodic and of class C1 then the convergence is uniform).
  • Finally, in practice one wants to approximate the function with a finite number of terms, say with a Taylor polynomial or a partial sum of the trigonometric series, respectively. In the case of the Taylor series the error is very small in a neighbourhood of the point where it is computed, while it may be very large at a distant point. In the case of the Fourier series the error is distributed along the domain of the function.

See also

[edit]

Notes

[edit]

References

[edit]
[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
A Taylor series is a fundamental concept in mathematics that represents a function as an infinite sum of terms, each derived from the function's derivatives evaluated at a single point, enabling precise approximations of the function near that point. The general form of the Taylor series for a function f(x)f(x) expanded around a point aa is given by f(x)=n=0f(n)(a)n!(xa)n,f(x) = \sum_{n=0}^{\infty} \frac{f^{(n)}(a)}{n!} (x - a)^n, where f(n)(a)f^{(n)}(a) denotes the nn-th derivative of ff at aa, and the series converges to f(x)f(x) within the radius of convergence under suitable conditions, such as those provided by Taylor's theorem. When the expansion point a=0a = 0, the series is specifically called a Maclaurin series, a special case named after Colin Maclaurin but rooted in the same principles. The concept of the Taylor series was first discovered by the Scottish mathematician James Gregory in February 1671, as documented in a letter to John Collins, where he outlined the method of expanding functions into infinite series using derivatives, though he did not publish it due to overlapping work by Isaac Newton and a prior dispute. It was formally introduced and popularized by the English mathematician Brook Taylor in his 1715 book Methodus incrementorum directa et inversa, with an earlier mention in a 1712 letter to John Machin; Taylor built on ideas from predecessors like Gregory, Newton, and Gottfried Wilhelm Leibniz, establishing it as a core tool in differential calculus. By the late 18th century, Joseph-Louis Lagrange recognized its central role in analysis, and it was termed "Taylor's series" in 1786 by Simon Lhuilier. Taylor series have profound applications across and , serving as powerful tools for , , and solving differential equations; for instance, they underpin computations of constants like π\pi (as used by Abraham Sharp in 1699 and John Machin in 1706) and enable series expansions for elementary functions such as exe^x, sinx\sin x, and ln(1+x)\ln(1 + x). In physics and , they facilitate modeling of physical phenomena, error estimation via remainder terms (e.g., Lagrange form), and extensions to multivariable and , making them indispensable for both theoretical insights and practical computations.

Core Concepts

Definition

In mathematics, a Taylor series is a representation of a function as an infinite sum of polynomial terms calculated from the function's derivatives at a single point, typically used to approximate the function near that point. For a function ff that is infinitely differentiable at a point aa in its domain, the Taylor series of ff about aa is given by f(x)=n=0f(n)(a)n!(xa)n,f(x) = \sum_{n=0}^{\infty} \frac{f^{(n)}(a)}{n!} (x - a)^n, where f(n)(a)f^{(n)}(a) denotes the nnth derivative of ff evaluated at aa, with f(0)(a)=f(a)f^{(0)}(a) = f(a), and n!n! is the factorial of nn. The general term f(n)(a)n!(xa)n\frac{f^{(n)}(a)}{n!} (x - a)^n incorporates higher-order derivatives to refine the approximation beyond the first-order linear tangent line, with the coefficient f(n)(a)n!\frac{f^{(n)}(a)}{n!} scaling the power (xa)n(x - a)^n to match the function's local behavior. This series extends the concept of a power series, where each term is a power of (xa)(x - a) multiplied by a coefficient derived from the function. When the expansion point a=0a = 0, the Taylor series is specifically called a Maclaurin series, named after the Scottish mathematician who popularized such expansions in the .

Notation

The Taylor series of a function ff expanded about a point aa is commonly expressed using summation notation as f(x)=n=0f(n)(a)n!(xa)n,f(x) = \sum_{n=0}^{\infty} \frac{f^{(n)}(a)}{n!} (x - a)^n, where f(n)(a)f^{(n)}(a) denotes the nnth derivative of ff evaluated at aa, and n!n! represents the factorial of nn. This form highlights the series as an infinite sum of terms involving powers of (xa)(x - a), with coefficients determined by the derivatives at the expansion point. An alternative notation employs coefficients cnc_n, yielding f(x)=n=0cn(xa)n,f(x) = \sum_{n=0}^{\infty} c_n (x - a)^n, where cn=f(n)(a)n!c_n = \frac{f^{(n)}(a)}{n!}. The superscript (n)(n) in f(n)f^{(n)} is the standard symbol for higher-order derivatives, starting with f(0)(a)=f(a)f^{(0)}(a) = f(a) for the zeroth derivative, which is the function itself. The factorial n!n! accounts for the scaling in the denominator, ensuring the terms align with the polynomial structure. For finite approximations, the partial sum up to degree nn is denoted as the Taylor Tn(x;a)T_n(x; a), given by Tn(x;a)=k=0nf(k)(a)k!(xa)k.T_n(x; a) = \sum_{k=0}^{n} \frac{f^{(k)}(a)}{k!} (x - a)^k. This notation distinguishes the from the full infinite series, emphasizing its use as a truncated approximation centered at aa. When a=0a = 0, the series or is specifically termed a Maclaurin series or , simplifying the expression to powers of xx alone. The difference between the function value and its Taylor polynomial approximation is captured by the remainder term Rn(x)R_n(x), defined as Rn(x)=f(x)Tn(x;a).R_n(x) = f(x) - T_n(x; a). This notation underscores the error in the finite approximation, with Rn(x)R_n(x) approaching zero under suitable convergence conditions as nn increases.

Basic Examples

Polynomial Functions

Polynomial functions serve as fundamental illustrations of Taylor series, where the expansion yields an exact, finite representation of the function itself. For a simple example, consider the quadratic polynomial f(x)=x2f(x) = x^2. Expanding around the point a=1a = 1, the zeroth derivative is f(1)=1f(1) = 1, the first derivative f(x)=2xf'(x) = 2x gives f(1)=2f'(1) = 2, the second derivative f(x)=2f''(x) = 2 yields f(1)=2f''(1) = 2, and all higher-order derivatives are zero. The resulting Taylor series is therefore f(x)=1+2(x1)+22!(x1)2,f(x) = 1 + 2(x-1) + \frac{2}{2!}(x-1)^2, which terminates after the n=2n=2 term and exactly equals (x1)2+2(x1)+1=x2(x-1)^2 + 2(x-1) + 1 = x^2. In the general case, any polynomial p(x)p(x) of degree kk has a Taylor series around any point aa that is identical to the polynomial expressed in powers of (xa)(x - a), consisting of a finite sum up to the kk-th term, as all derivatives of order greater than kk vanish identically. This finite nature arises directly from the structure of polynomials, where repeated differentiation eventually produces the zero function. To compute the series, one evaluates the derivatives of p(x)p(x) at aa explicitly: the coefficients are p(n)(a)n!\frac{p^{(n)}(a)}{n!} for n=0n = 0 to kk, with p(n)(a)=0p^{(n)}(a) = 0 for n>kn > k, leading to a terminating sum that reconstructs p(x)p(x) precisely. For the quadratic example, this process requires only three evaluations (including the function value), demonstrating the simplicity and exactness for low-degree cases. This exact equivalence reveals that polynomials are their own Taylor series, providing a perfect without for all xx, and emphasizes their role as the building blocks for more complex analytic functions in series theory.

Exponential Function

The exponential function exe^x provides a foundational example of an infinite Taylor series, illustrating how the method captures functions that extend beyond finite polynomials. Its Maclaurin series, centered at a=0a = 0, is ex=n=0xnn!.e^x = \sum_{n=0}^{\infty} \frac{x^n}{n!}. This series is particularly straightforward to derive because the exponential function is its own derivative: the first derivative is exe^x, the second is exe^x, and in general, the nn-th derivative is exe^x for all n0n \geq 0. Evaluating these at x=0x = 0 yields f(n)(0)=1f^{(n)}(0) = 1, so the Taylor coefficients simplify to 1n!\frac{1}{n!}. Partial sums of this series offer practical approximations for exe^x. For instance, at x=1x = 1, the sum of the first six terms (up to n=5n = 5) is 1+1+12+16+124+11202.716671 + 1 + \frac{1}{2} + \frac{1}{6} + \frac{1}{24} + \frac{1}{120} \approx 2.71667, which approximates e2.71828e \approx 2.71828 with an error of less than 0.002. Such truncations are useful in numerical computations where higher precision requires more terms, highlighting the series' role in iterative approximations. For expansion around a general point aa, the Taylor series becomes ex=ean=0(xa)nn!,e^x = e^a \sum_{n=0}^{\infty} \frac{(x - a)^n}{n!}, reflecting that all derivatives evaluated at aa equal eae^a, which factors out as the leading term. This form preserves the infinite series structure while shifting the center, enabling local approximations near any point. The factorial n!n! in the denominator, defined as n!=12nn! = 1 \cdot 2 \cdot \dots \cdot n for positive integers nn (with 0!=10! = 1), ensures the rapid growth of the denominator balances the powers of xx.

Historical Context

Early Contributions

The development of ideas leading to Taylor series can be traced back to medieval mathematicians who explored infinite series and approximation techniques for trigonometric and other functions. In 14th-century , (c. 1340–1425) made pioneering advances in infinite series expansions, particularly for such as sine, cosine, and arctangent, which allowed for precise through the arctan series. These contributions, part of the Kerala school of astronomy and mathematics, represented an early systematic use of to model continuous functions, predating European developments by centuries. In the , scholars built on Greek foundations to advance approximation methods, including and solutions to equations. (c. 965–1040), a Persian , advanced solutions to higher-degree equations by linking algebra with geometric constructions, notably in , and developed formulas for . Earlier figures like al-Karaji (c. 953–1029) developed methods for summing powers of integers using , laying algebraic groundwork that influenced later developments in analysis. These efforts emphasized iterative approximations and infinite progressions, influencing later analytical techniques. During the European Renaissance, Italian mathematicians introduced concepts of indivisibles that foreshadowed differential ideas. (1598–1647) developed the method of indivisibles in the 1630s, treating areas and volumes as sums of lines and surfaces to compute quantities without explicit limits, which prefigured the notion of in . This approach, detailed in his Geometria indivisibilibus continuorum (1635), provided a bridge between and , encouraging the manipulation of continuous change. In the 17th century, British mathematicians Isaac Newton (1643–1727) and James Gregory (1638–1675) advanced series expansions within their work on fluxions, an early form of calculus. Newton employed infinite series, including binomial expansions, to approximate planetary orbits under the inverse square law, integrating these with fluxional methods to model motion dynamically. Gregory independently derived series for arctangent and trigonometric functions, using interpolation techniques to expand functions around points, which closely resembled modern power series methods. Their fluxion-based calculus, developed in the 1660s and 1670s, treated rates of change as foundational, setting the stage for formal series representations of functions.

Brook Taylor's Formulation

In 1715, English mathematician Brook Taylor published Methodus incrementorum directa et inversa, a seminal work that introduced a general method for expanding functions into infinite series using finite differences. This approach provided a systematic way to approximate function values through incremental changes, establishing the foundations of the calculus of finite differences as a new branch of mathematics. The key innovation in Taylor's formulation lay in expressing series expansions via finite differences, which were later reinterpreted in terms of within the framework of Newtonian fluxions. In Proposition VII of the book, Taylor derived expansions for solutions to fluxional equations, treating increments discretely to build infinite series representations, an idea that bridged discrete and continuous . Although similar expansions had been explored earlier by James Gregory and , the series became known as the Taylor series, with the term first appearing in print in 1786 courtesy of Swiss mathematician Simon L'Huilier. had already highlighted its significance as a core principle of in 1772. Taylor's work prompted refinements by Leonhard Euler in the mid-18th century, who integrated and extended the series into broader analytical contexts, paving the way for its modern form in and .

Analyticity and Representation

Analytic Functions

In , a function ff is defined to be analytic at a point aa if it is infinitely differentiable at aa and equals its Taylor series expansion centered at aa throughout some open interval containing aa. This means that within that neighborhood, f(x)=n=0f(n)(a)n!(xa)nf(x) = \sum_{n=0}^{\infty} \frac{f^{(n)}(a)}{n!} (x - a)^n, where the series converges to the function's values. Analyticity thus provides a local representation of the function via a power series, distinguishing it from mere infinite differentiability, as not all smooth functions satisfy this equality. Classic examples of analytic functions include the exe^x and the sine function sinx\sin x, both of which are entire, meaning they are analytic at every point on the real line with Taylor series converging to the function globally. In contrast, the function f(x)=xf(x) = |x| is not analytic at x=0x = 0 because it fails to be differentiable there, precluding the existence of a Taylor series at that point. Another illustrative non-analytic example is the smooth f(x)=e1/x2f(x) = e^{-1/x^2} for x>0x > 0 and f(x)=0f(x) = 0 for x0x \leq 0, which has all derivatives zero at x=0x = 0, yielding a trivial Taylor series that does not represent the function nearby. Analyticity is inherently a local property: a function may be analytic at each point of its domain, but the associated power series typically converges only within a specific interval (or disk in the complex plane) around that point, with the size of this region potentially varying by location. For instance, the geometric series for 11x\frac{1}{1-x} converges to the function on (1,1)(-1, 1) but diverges outside, limiting global representation. In complex analysis, this concept aligns closely with holomorphicity, where a function is analytic if it is complex differentiable in a domain, equivalently admitting a power series expansion there.

Taylor's Theorem

Taylor's theorem asserts that if a function ff is k+1k+1 times differentiable on an open interval containing points aa and xx, then there exists a Pk(x)P_k(x) of degree at most kk, called the kk-th Taylor polynomial centered at aa, such that f(x)=Pk(x)+Rk(x),f(x) = P_k(x) + R_k(x), where Pk(x)=n=0kf(n)(a)n!(xa)nP_k(x) = \sum_{n=0}^k \frac{f^{(n)}(a)}{n!} (x - a)^n and Rk(x)R_k(x) is the remainder term. This representation justifies the use of Taylor polynomials as approximations to f(x)f(x) near aa, with the remainder quantifying the error. A common explicit form for the remainder is the Lagrange form: assuming f(k+1)f^{(k+1)} exists on the interval, Rk(x)=f(k+1)(ξ)(k+1)!(xa)k+1R_k(x) = \frac{f^{(k+1)}(\xi)}{(k+1)!} (x - a)^{k+1} for some ξ\xi between aa and xx. This form is particularly useful for estimating the approximation error when bounds on the higher derivative are available. One standard proof of Taylor's theorem with the Lagrange remainder relies on iterative application of the mean value theorem (or equivalently, Rolle's theorem). To derive it, consider the auxiliary function F(t)=f(x)f(t)f(t)(xt)f(k)(t)k!(xt)kR(t)(xt)k+1F(t) = f(x) - f(t) - f'(t)(x - t) - \cdots - \frac{f^{(k)}(t)}{k!} (x - t)^k - R(t) (x - t)^{k+1}, where R(t)R(t) is chosen such that F(a)=0F(a) = 0 and F(x)=0F(x) = 0. Since F(a)=F(x)F(a) = F(x), Rolle's theorem implies F(ξ1)=0F'(\xi_1) = 0 for some ξ1\xi_1 between aa and xx. Repeating this process k+1k+1 times on the successively differentiated auxiliary functions yields the Lagrange remainder after solving the resulting equation. An alternative proof uses integration by parts on the integral form of the remainder, starting from the fundamental theorem of calculus and integrating the higher derivatives iteratively. Other remainder forms include the Peano form, which states that if ff is kk times differentiable at aa, then Rk(x)=o((xa)k)R_k(x) = o((x - a)^k) as xax \to a, emphasizing the order of contact between ff and its Taylor polynomial at the expansion point without requiring differentiability on an interval. In comparison, the Peano form is asymptotic and local to aa, suitable for proving approximation orders near the center, whereas the Lagrange form provides an exact, interval-wide expression involving the (k+1)(k+1)-th at an intermediate point, facilitating explicit error bounds away from aa. For analytic functions, where the vanishes in the limit as kk \to \infty, the infinite Taylor series equals f(x)f(x) exactly within the .

Convergence Properties

Radius of Convergence

The radius of convergence of a Taylor series for a function ff expanded about a point aa, given by n=0cn(xa)n\sum_{n=0}^{\infty} c_n (x - a)^n where cn=f(n)(a)n!c_n = \frac{f^{(n)}(a)}{n!}, is the value R0R \geq 0 such that the series converges absolutely for all xa<R|x - a| < R and diverges for all xa>R|x - a| > R. Within this open interval, the series represents f(x)f(x) under suitable conditions on ff. The radius RR can be computed using the Cauchy-Hadamard formula: R=1lim supncn1/n.R = \frac{1}{\limsup_{n \to \infty} |c_n|^{1/n}}. If the limit superior is zero, then R=R = \infty, indicating convergence for all xx; if it is infinite, then R=0R = 0, meaning convergence only at x=ax = a. An alternative method to determine RR, when applicable, employs the ratio test on the coefficients: if limncncn+1\lim_{n \to \infty} \left| \frac{c_n}{c_{n+1}} \right|
Add your contribution
Related Hubs
User Avatar
No comments yet.