HOME

TheInfoList



OR:

In the
mathematical Mathematics is an area of knowledge that includes the topics of numbers, formulas and related structures, shapes and the spaces in which they are contained, and quantities and their changes. These topics are represented in modern mathematics ...
field of
numerical analysis Numerical analysis is the study of algorithms that use numerical approximation (as opposed to symbolic manipulations) for the problems of mathematical analysis (as distinguished from discrete mathematics). It is the study of numerical methods ...
, a Newton polynomial, named after its inventor
Isaac Newton Sir Isaac Newton (25 December 1642 – 20 March 1726/27) was an English mathematician, physicist, astronomer, alchemist, Theology, theologian, and author (described in his time as a "natural philosophy, natural philosopher"), widely ...
, is an
interpolation In the mathematical field of numerical analysis, interpolation is a type of estimation, a method of constructing (finding) new data points based on the range of a discrete set of known data points. In engineering and science, one often has ...
polynomial In mathematics, a polynomial is an expression consisting of indeterminates (also called variables) and coefficients, that involves only the operations of addition, subtraction, multiplication, and positive-integer powers of variables. An exampl ...
for a given set of data points. The Newton polynomial is sometimes called Newton's divided differences interpolation polynomial because the coefficients of the polynomial are calculated using Newton's divided differences method.


Definition

Given a set of ''k'' + 1 data points :(x_0, y_0),\ldots,(x_j, y_j),\ldots,(x_k, y_k) where no two ''x''''j'' are the same, the Newton interpolation polynomial is a linear combination of Newton basis polynomials :N(x) := \sum_^ a_ n_(x) with the Newton basis polynomials defined as :n_j(x) := \prod_^ (x - x_i) for ''j'' > 0 and n_0(x) \equiv 1. The coefficients are defined as :a_j := _0,\ldots,y_j/math> where : _0,\ldots,y_j/math> is the notation for divided differences. Thus the Newton polynomial can be written as :N(x) = _0+ _0,y_1x-x_0) + \cdots + _0,\ldots,y_kx-x_0)(x-x_1)\cdots(x-x_).


Newton forward divided difference formula

The Newton polynomial can be expressed in a simplified form when x_0, x_1, \dots, x_k are arranged consecutively with equal spacing. Introducing the notation h = x_-x_i for each i=0,1,\dots,k-1 and x=x_0+sh, the difference x-x_i can be written as (s-i)h. So the Newton polynomial becomes :\begin N(x) &= _0+ _0,y_1h + \cdots + _0,\ldots,y_ks (s-1) \cdots (s-k+1)^ \\ &= \sum_^s(s-1) \cdots (s-i+1)^ _0,\ldots,y_i\\ &= \sum_^i!^ _0,\ldots,y_i \end This is called the Newton forward divided difference formula.


Newton backward divided difference formula

If the nodes are reordered as _,_,\dots,_, the Newton polynomial becomes :N(x)= _k , _x-_)+\cdots+ ,\ldots,_x-_)(x-_)\cdots(x-_). If _,\;_,\;\dots,\;_ are equally spaced with _=_+sh and _=_-(k-i)h for ''i'' = 0, 1, ..., ''k'', then, :\begin N(x) &= , _h+\cdots+ ,\ldots,_(s+1)\cdots(s+k-1)^ \\ &=\sum_^^i!^ ,\ldots,_ \end is called the Newton backward divided difference formula.


Significance

Newton's formula is of interest because it is the straightforward and natural differences-version of Taylor's polynomial. Taylor's polynomial tells where a function will go, based on its ''y'' value, and its derivatives (its rate of change, and the rate of change of its rate of change, etc.) at one particular ''x'' value. Newton's formula is Taylor's polynomial based on finite differences instead of instantaneous rates of change.


Addition of new points

As with other difference formulas, the degree of a Newton interpolating polynomial can be increased by adding more terms and points without discarding existing ones. Newton's form has the simplicity that the new points are always added at one end: Newton's forward formula can add new points to the right, and Newton's backward formula can add new points to the left. The accuracy of polynomial interpolation depends on how close the interpolated point is to the middle of the ''x'' values of the set of points used. Obviously, as new points are added at one end, that middle becomes farther and farther from the first data point. Therefore, if it isn't known how many points will be needed for the desired accuracy, the middle of the x-values might be far from where the interpolation is done. Gauss, Stirling, and Bessel all developed formulae to remedy that problem. Gauss's formula alternately adds new points at the left and right ends, thereby keeping the set of points centered near the same place (near the evaluated point). When so doing, it uses terms from Newton's formula, with data points and ''x'' values renamed in keeping with one's choice of what data point is designated as the ''x''0 data point. Stirling's formula remains centered about a particular data point, for use when the evaluated point is nearer to a data point than to a middle of two data points. Bessel's formula remains centered about a particular middle between two data points, for use when the evaluated point is nearer to a middle than to a data point. Bessel and Stirling achieve that by sometimes using the average of two differences, and sometimes using the average of two products of binomials in ''x'', where Newton's or Gauss's would use just one difference or product. Stirling's uses an average difference in odd-degree terms (whose difference uses an even number of data points); Bessel's uses an average difference in even-degree terms (whose difference uses an odd number of data points).


Strengths and weaknesses of various formulae

For any given finite set of data points, there is only one polynomial of least possible degree that passes through all of them. Thus, it is appropriate to speak of the "Newton form", or Lagrange form, etc., of the interpolation polynomial. However, the way the polynomial is obtained matters. There are several similar methods, such as those of Gauss, Bessel and Stirling. They can be derived from Newton's by renaming the ''x''-values of the data points, but in practice they are important.


Bessel vs. Stirling

The choice between Bessel and Stirling depends on whether the interpolated point is closer to a data point, or closer to a middle between two data points. A polynomial interpolation's error approaches zero, as the interpolation point approaches a data-point. Therefore, Stirling's formula brings its accuracy improvement where it is least needed and Bessel brings its accuracy improvement where it is most needed. So, Bessel's formula could be said to be the most consistently accurate difference formula, and, in general, the most consistently accurate of the familiar polynomial interpolation formulas.


Divided-Difference Methods vs. Lagrange

Lagrange is sometimes said to require less work, and is sometimes recommended for problems in which it is known, in advance, from previous experience, how many terms are needed for sufficient accuracy. The divided difference methods have the advantage that more data points can be added, for improved accuracy. The terms based on the previous data points can continue to be used. With the ordinary Lagrange formula, to do the problem with more data points would require re-doing the whole problem. There is a "barycentric" version of Lagrange that avoids the need to re-do the entire calculation when adding a new data point. But it requires that the values of each term be recorded. But the ability, of Gauss, Bessel and Stirling, to keep the data points centered close to the interpolated point gives them an advantage over Lagrange, when it isn't known, in advance, how many data points will be needed. Additionally, suppose that one wants to find out if, for some particular type of problem, linear interpolation is sufficiently accurate. That can be determined by evaluating the quadratic term of a divided difference formula. If the quadratic term is negligible—meaning that the linear term is sufficiently accurate without adding the quadratic term—then linear interpolation is sufficiently accurate. If the problem is sufficiently important, or if the quadratic term is nearly big enough to matter, then one might want to determine whether the sum of the quadratic and cubic terms is large enough to matter in the problem. Of course, only a divided-difference method can be used for such a determination. For that purpose, the divided-difference formula and/or its ''x''0 point should be chosen so that the formula will use, for its linear term, the two data points between which the linear interpolation of interest would be done. The divided difference formulas are more versatile, useful in more kinds of problems. The Lagrange formula is at its best when all the interpolation will be done at one ''x'' value, with only the data points' ''y'' values varying from one problem to another, and when it is known, from past experience, how many terms are needed for sufficient accuracy. With the Newton form of the interpolating polynomial a compact and effective algorithm exists for combining the terms to find the coefficients of the polynomial.


Accuracy

When, with Stirling's or Bessel's, the last term used includes the average of two differences, then one more point is being used than Newton's or other polynomial interpolations would use for the same polynomial degree. So, in that instance, Stirling's or Bessel's is not putting an ''N''−1 degree polynomial through ''N'' points, but is, instead, trading equivalence with Newton's for better centering and accuracy, giving those methods sometimes potentially greater accuracy, for a given polynomial degree, than other polynomial interpolations.


General case

For the special case of ''xi'' = ''i'', there is a closely related set of polynomials, also called the Newton polynomials, that are simply the
binomial coefficient In mathematics, the binomial coefficients are the positive integers that occur as coefficients in the binomial theorem. Commonly, a binomial coefficient is indexed by a pair of integers and is written \tbinom. It is the coefficient of the t ...
s for general argument. That is, one also has the Newton polynomials p_n(z) given by :p_n(z)

\frac
In this form, the Newton polynomials generate the
Newton series A finite difference is a mathematical expression of the form . If a finite difference is divided by , one gets a difference quotient. The approximation of derivatives by finite differences plays a central role in finite difference methods for the ...
. These are in turn a special case of the general
difference polynomials In mathematics, in the area of complex analysis, the general difference polynomials are a polynomial sequence, a certain subclass of the Sheffer polynomials, which include the Newton polynomials, Selberg's polynomials, and the Stirling interpolati ...
which allow the representation of
analytic function In mathematics, an analytic function is a function that is locally given by a convergent power series. There exist both real analytic functions and complex analytic functions. Functions of each type are infinitely differentiable, but complex ...
s through generalized difference equations.


Main idea

Solving an interpolation problem leads to a problem in linear algebra where we have to solve a system of linear equations. Using a standard
monomial basis In mathematics the monomial basis of a polynomial ring is its basis (as a vector space or free module over the field or ring of coefficients) that consists of all monomials. The monomials form a basis because every polynomial may be uniquely ...
for our interpolation polynomial we get the very complicated
Vandermonde matrix In linear algebra, a Vandermonde matrix, named after Alexandre-Théophile Vandermonde, is a matrix with the terms of a geometric progression in each row: an matrix :V=\begin 1 & x_1 & x_1^2 & \dots & x_1^\\ 1 & x_2 & x_2^2 & \dots & x_2^\\ 1 & x_ ...
. By choosing another basis, the Newton basis, we get a system of linear equations with a much simpler
lower triangular matrix In mathematics, a triangular matrix is a special kind of square matrix. A square matrix is called if all the entries ''above'' the main diagonal are zero. Similarly, a square matrix is called if all the entries ''below'' the main diagonal a ...
which can be solved faster. For ''k'' + 1 data points we construct the Newton basis as :n_0(x) := 1 , \qquad n_j(x) := \prod_^ (x - x_i) \qquad j=1,\ldots,k. Using these polynomials as a basis for \Pi_k we have to solve :\begin 1 & & \ldots & & 0 \\ 1 & x_1-x_0 & & & \\ 1 & x_2-x_0 & (x_2-x_0)(x_2-x_1) & & \vdots \\ \vdots & \vdots & & \ddots & \\ 1 & x_k-x_0 & \ldots & \ldots & \prod_^(x_k - x_j) \end \begin a_0 \\ \\ \vdots \\ \\ a_ \end = \begin y_0 \\ \\ \vdots \\ \\ y_ \end to solve the polynomial interpolation problem. This system of equations can be solved iteratively by solving : \sum_^ a_ n_(x_j) = y_j \qquad j = 0,\dots,k.


Derivation

While the interpolation formula can be found by solving a linear system of equations, there is a loss of intuition in what the formula is showing and why Newton's interpolation formula works is not readily apparent. To begin, we will need to establish two facts first: Fact 1. Reversing the terms of a divided difference leaves it unchanged: _0, \ldots, y_n= _n, \ldots, y_0 The proof of this is an easy induction: for n=1 we compute _0, y_1= \frac = \frac = _1, y_0 Induction step: Suppose the result holds for any divided difference involving at most n+1 terms. Then using the induction hypothesis in the following 2nd equality we see that for a divided difference involving n+2 terms we have _0, \ldots, y_= \frac = \frac = _, \ldots, y_0 We formulate next Fact 2 which for purposes of induction and clarity we also call Statement n (\text_n) : Fact 2. (\text_n) : If (x_0, y_0), \ldots, (x_, y_) are any n points with distinct x-coordinates and P=P(x) is the unique polynomial of degree (at most) n-1 whose graph passes through these n points then there holds the relation _0, \ldots, y_nx_n - x_0)\cdot\ldots\cdot(x_n-x_) = y_n - P(x_n) Proof. (It will be helpful for fluent reading of the proof to have the precise statement and its subtlety in mind: P is defined by passing through (x_0,y_0),. . . ,(x_,y_) but the formula also speaks at both sides of an additional arbitrary point (x_n,y_n) with x-coordinate distinct from the other x_i .) We again prove these statements by induction. To show \text_1, let (x_0,y_0) be any one point and let P(x) be the unique polynomial of degree 0 passing through (x_0, y_0). Then evidently P(x)=y_0 and we can write _0, y_1x_1 - x_0) = \frac (x_1-x_0) = y_1 - y_0 = y_1 - P(x_1) as wanted. Proof of \text_, assuming \text_ already established: Let P(x) be the polynomial of degree (at most) n passing through (x_0, y_0), \ldots, (x_n, y_n). With Q(x) being the unique polynomial of degree (at most) n-1 passing through the points (x_1, y_1), \ldots, (x_n, y_n), we can write the following chain of equalities, where we use in the penultimate equality that Stm_n applies to Q: \begin & _0,\ldots,y_x_ - x_0)\cdot\ldots\cdot(x_ - x_n)\\ &= \frac(x_ - x_0)\cdot\ldots\cdot(x_ - x_n) \\ &= \left( _1,\ldots,y_- _0,\ldots,y_right) (x_ - x_1)\cdot\ldots\cdot(x_ - x_n) \\ &= _1,\ldots,y_x_ - x_1)\cdot\ldots\cdot(x_ - x_n) - _0,\ldots,y_nx_ - x_1)\cdot\ldots\cdot(x_ - x_n) \\ &= (y_ - Q(x_)) - _0,\ldots,y_nx_ - x_1)\cdot\ldots\cdot(x_ - x_n) \\ &= y_ - (Q(x_) + _0,\ldots,y_nx_ - x_1)\cdot\ldots\cdot(x_ - x_n)). \end The induction hypothesis for Q also applies to the second equality in the following computation, where (x_0,y_0) is added to the points defining Q : \begin & Q(x_0) + _0,\ldots,y_nx_0 - x_1)\cdot\ldots\cdot(x_0 - x_n) \\ &= Q(x_0) + _n,\ldots,y_0x_0 - x_n)\cdot\ldots\cdot(x_0 - x_1) \\ & = Q(x_0) + y_0 - Q(x_0) \\ &= y_0 \\ &= P(x_0). \\ \end Now look at Q(x)+ _0,\ldots,y_nx - x_1)\cdot\ldots\cdot(x - x_n). By the definition of Q this polynomial passes through (x_1,y_1), . . ., (x_n,y_n) and, as we have just shown, it also passes through (x_0,y_0). Thus it is the unique polynomial of degree \leq n which passes through these points. Therefore this polynomial is P(x); i.e.: P(x)= Q(x)+ _0,\ldots,y_nx - x_1)\cdot\ldots\cdot(x - x_n). Thus we can write the last line in the first chain of equalities as ` y_-P(x_)' and have thus established that _0,\ldots,y_x_ - x_0)\cdot\ldots\cdot(x_ - x_n)=y_-P(x_). So we established \text_, and hence completed the proof of Fact 2. Now look at Fact 2: It can be formulated this way: If P is the unique polynomial of degree at most n-1 whose graph passes through the points (x_0, y_0), . . . , (x_, y_), then P(x)+ _0, \ldots, y_nx - x_0)\cdot\ldots\cdot(x-x_) is the unique polynomial of degree at most n passing through points (x_0, y_0), . . . , (x_, y_), (x_n, y_n). So we see Newton interpolation permits indeed to add new interpolation points without destroying what has already been computed.


Taylor polynomial

The limit of the Newton polynomial if all nodes coincide is a
Taylor polynomial In mathematics, the Taylor series or Taylor expansion of a function is an infinite sum of terms that are expressed in terms of the function's derivatives at a single point. For most common functions, the function and the sum of its Taylor ser ...
, because the divided differences become derivatives. \begin &\lim_ f _0+ f _0,x_1cdot(\xi-x_0) + \dots + f _0,\dots,x_ncdot(\xi-x_0)\cdot\dots\cdot(\xi-x_) \\ &= f(z) + f'(z)\cdot(\xi-z) + \dots + \frac\cdot(\xi-z)^n \end


Application

As can be seen from the definition of the divided differences new data points can be added to the data set to create a new interpolation polynomial without recalculating the old coefficients. And when a data point changes we usually do not have to recalculate all coefficients. Furthermore, if the ''x''''i'' are distributed equidistantly the calculation of the divided differences becomes significantly easier. Therefore, the divided-difference formulas are usually preferred over the Lagrange form for practical purposes.


Examples

The divided differences can be written in the form of a table. For example, for a function ''f'' is to be interpolated on points x_0, \ldots, x_n. Write :\begin x_0 & f(x_0) & & \\ & & & \\ x_1 & f(x_1) & & \\ & & & \\ x_2 & f(x_2) & & \vdots \\ & & \vdots & \\ \vdots & & & \vdots \\ & & \vdots & \\ x_n & f(x_n) & & \\ \end Then the interpolating polynomial is formed as above using the topmost entries in each column as coefficients. For example, suppose we are to construct the interpolating polynomial to ''f''(''x'') = tan(''x'') using divided differences, at the points Using six digits of accuracy, we construct the table : \begin -\tfrac & -14.1014 & & & &\\ & & 17.5597 & & &\\ -\tfrac & -0.931596 & & -10.8784 & &\\ & & 1.24213 & & 4.83484 & \\ 0 & 0 & & 0 & & 0\\ & & 1.24213 & & 4.83484 &\\ \tfrac & 0.931596 & & 10.8784 & &\\ & & 17.5597 & & &\\ \tfrac & 14.1014 & & & &\\ \end Thus, the interpolating polynomial is :\begin &-14.1014+17.5597(x+\tfrac)-10.8784(x+\tfrac)(x+\tfrac) +4.83484(x+\tfrac)(x+\tfrac)(x)+0(x+\tfrac)(x+\tfrac)(x)(x-\tfrac) \\ =&-0.00005-1.4775x-0.00001x^2+4.83484x^3 \end Given more digits of accuracy in the table, the first and third coefficients will be found to be zero. Another example: The sequence f_0 such that f_0(1) = 6, f_0(2) = 9, f_0(3) = 2 and f_0(4) = 5, i.e., they are 6, 9, 2, 5 from x_0 = 1 to x_3 = 4. You obtain the slope of order 1 in the following way: * f_1(x_0, x_1) = \frac = \frac = 3 * f_1(x_1, x_2) = \frac = \frac = -7 * f_1(x_2, x_3) = \frac = \frac = 3 As we have the slopes of order 1, it is possible to obtain the next order: * f_2(x_0, x_1, x_2) = \frac = \frac = -5 * f_2(x_1, x_2, x_3) = \frac = \frac = 5 Finally, we define the slope of order 3: * f_3(x_0, x_1, x_2, x_3) = \frac = \frac = \frac Once we have the slope, we can define the consequent polynomials: * p_0(x) = 6. * p_1(x) = 6 + 3(x - 1) * p_2(x) = 6 + 3(x - 1) - 5(x - 1)(x - 2). * p_3(x) = 6 + 3(x - 1) - 5(x - 1)(x - 2) + \frac (x - 1)(x - 2)(x - 3)


See also

*'' De numeris triangularibus et inde de progressionibus arithmeticis: Magisteria magna'', a work by
Thomas Harriot Thomas Harriot (; – 2 July 1621), also spelled Harriott, Hariot or Heriot, was an English astronomer, mathematician, ethnographer and translator to whom the theory of refraction is attributed. Thomas Harriot was also recognized for his con ...
describing similar methods for interpolation, written 50 years earlier than Newton's work but not published until 2009 *
Newton series A finite difference is a mathematical expression of the form . If a finite difference is divided by , one gets a difference quotient. The approximation of derivatives by finite differences plays a central role in finite difference methods for the ...
*
Neville's schema In mathematics, Neville's algorithm is an algorithm used for polynomial interpolation that was derived by the mathematician Eric Harold Neville in 1934. Given ''n'' + 1 points, there is a unique polynomial of degree ''≤ n'' which goes through the ...
*
Polynomial interpolation In numerical analysis, polynomial interpolation is the interpolation of a given data set by the polynomial of lowest possible degree that passes through the points of the dataset. Given a set of data points (x_0,y_0), \ldots, (x_n,y_n), with no ...
* Lagrange form of the interpolation polynomial *
Bernstein form In the mathematical field of numerical analysis, a Bernstein polynomial is a polynomial that is a linear combination of Bernstein basis polynomials. The idea is named after Sergei Natanovich Bernstein. A numerically stable way to evaluate po ...
of the interpolation polynomial *
Hermite interpolation In numerical analysis, Hermite interpolation, named after Charles Hermite, is a method of polynomial interpolation, which generalizes Lagrange interpolation. Lagrange interpolation allows computing a polynomial of degree less than that takes the ...
* Carlson's theorem * Table of Newtonian series


References


External links


Module for the Newton Polynomial by John H. Mathews
{{Isaac Newton Interpolation Finite differences Factorial and binomial topics Polynomials