Aitken Method
   HOME

TheInfoList



OR:

In
numerical analysis Numerical analysis is the study of algorithms that use numerical approximation (as opposed to symbolic computation, symbolic manipulations) for the problems of mathematical analysis (as distinguished from discrete mathematics). It is the study of ...
, Aitken's delta-squared process or Aitken extrapolation is a
series acceleration Series may refer to: People with the name * Caroline Series (born 1951), English mathematician, daughter of George Series * George Series (1920–1995), English physicist Arts, entertainment, and media Music * Series, the ordered sets used ...
method used for accelerating the
rate of convergence In mathematical analysis, particularly numerical analysis, the rate of convergence and order of convergence of a sequence that converges to a limit are any of several characterizations of how quickly that sequence approaches its limit. These are ...
of a sequence. It is named after
Alexander Aitken Alexander Craig "Alec" Aitken (1 April 1895 – 3 November 1967) was one of New Zealand's most eminent mathematicians. In a 1935 paper he introduced the concept of generalized least squares, along with now standard vector/matrix notation ...
, who introduced this method in 1926 as part of an extension to
Bernoulli's method In numerical analysis, Bernoulli's method, named after Daniel Bernoulli, is a root-finding algorithm which calculates the polynomial, root of largest absolute value of a univariate polynomial. The method works under the condition that there is ...
. It is most useful for accelerating the convergence of a sequence that is converging linearly. A precursor form was known to
Seki Kōwa , Selin, Helaine. (1997). ''Encyclopaedia of the History of Science, Technology, and Medicine in Non-Western Cultures,'' p. 890 also known as ,Selin, was a mathematician, samurai, and Kofu feudal officer of the early Edo period of Japan. Se ...
(1642 – 1708) and applied to the rectification of the circle, i.e., to the calculation of π.


Definition

Given a sequence X = with n = 0, 1, 2, 3, \ldots, Aitken's delta-squared process associates to this sequence the new sequence A = (a_n) = , which can also be written as A = \left( x_n-\frac \right), with \Delta x_= x_-x_ and \Delta^2 x_n=x_n -2x_ + x_=\Delta x_-\Delta x_. Both are the same sequence algebraically but the latter has improved
numerical stability In the mathematical subfield of numerical analysis, numerical stability is a generally desirable property of numerical algorithms. The precise definition of stability depends on the context: one important context is numerical linear algebra, and ...
in computational implementation. A /math> is ill-defined if the sequence \Delta^2 = (\Delta^2 x_n) contains a zero element, which occurs if the sequence of forward differences, \Delta = (\Delta x_), has any repeated term. From a theoretical point of view, if that occurs only for a finite number of indices, one could apply the Aitken process to only the part of the sequence X with indices n > n_0 such that n_0 is the last index for which the sequence \Delta repeats. In practice, the first few terms of the sequence usually provide desired precision; also, when numerically computing the sequence, one has to take care to stop the computation before
rounding error In computing, a roundoff error, also called rounding error, is the difference between the result produced by a given algorithm using exact arithmetic and the result produced by the same algorithm using finite-precision, rounded arithmetic. Roun ...
s in the denominator become too large, as the \Delta^2 sequence transformation may cancel
significant digit Significant figures, also referred to as significant digits, are specific Numerical digit, digits within a number that is written in positional notation that carry both reliability and necessity in conveying a particular quantity. When presen ...
s.


Properties

Aitken's delta-squared process is an
acceleration of convergence In mechanics, acceleration is the rate of change of the velocity of an object with respect to time. Acceleration is one of several components of kinematics, the study of motion. Accelerations are vector quantities (in that they have magnit ...
method and a particular case of a nonlinear
sequence transformation In mathematics, a sequence transformation is an Operator (mathematics), operator acting on a given space of sequences (a sequence space). Sequence transformations include linear mappings such as convolution, discrete convolution with another sequen ...
. A sequence X = (x_n) that converges to a limiting value \ell is said to converge linearly, or more technically Q-linearly, if there is some number \mu \in (0,1) for which \lim_ \frac = \mu. This means that asymptotically, the distance between the sequence and its limit shrinks by nearly the same proportion, \mu, on every step and the ratio of reduction becomes closer and closer to that proportion. This is also sometimes called "geometric convergence," since it is a characteristic property for
geometric series In mathematics, a geometric series is a series (mathematics), series summing the terms of an infinite geometric sequence, in which the ratio of consecutive terms is constant. For example, 1/2 + 1/4 + 1/8 + 1/16 + ⋯, the series \tfrac12 + \tfrac1 ...
, or "exponential convergence," since it is convergence like \mu^n = \exp(n \ln \mu). Aitken's method will accelerate the convergence of a sequence X if A = (a_n), with terms defined above, satisfies \lim_\frac = 0. A is not a linear operator on sequences, but it is linear with respect to addition of constant sequences: A - C= A - C, if C is any constant sequence C = (c), constant for all n. This is clear from the expression of A in terms of the
finite difference A finite difference is a mathematical expression of the form . Finite differences (or the associated difference quotients) are often used as approximations of derivatives, such as in numerical differentiation. The difference operator, commonly d ...
operator Operator may refer to: Mathematics * A symbol indicating a mathematical operation * Logical operator or logical connective in mathematical logic * Operator (mathematics), mapping that acts on elements of a space to produce elements of another sp ...
\Delta. The new process does not in general converge quadratically, but for an
iterated function In mathematics, an iterated function is a function that is obtained by composing another function with itself two or several times. The process of repeatedly applying the same function is called iteration. In this process, starting from some ...
sequence satisfying x_ = f(x_n) for some function f converging to a fixed point, the accelerated sequence's convergence is quadratic. In this case, the technique is known as
Steffensen's method In numerical analysis, Steffensen's method is an iterative method for numerical root-finding named after Johan Frederik Steffensen that is similar to the secant method and to Newton's method. Steffensen's method achieves a quadratic order of co ...
. Empirically, the ''A''-operation eliminates the "most important error term". One can check this by considering a sequence of the form x_n=\ell+a^n+b^n, where 0 < b < a < 1: The sequence A /math> will then go to the limit \ell like b^n goes to zero. Geometrically, the graph of an exponential function f(t) that satisfies f(n) = x_n, f(n+1) = x_ and f(n+2)=x_ has an horizontal asymptote at \frac (if x_ - 2x_ + x_ \neq 0). One can also show that if a sequence X converges to its limit \ell at a rate strictly greater than 1, A /math> does not have a better rate of convergence. (In practice, one rarely has e.g. quadratic convergence which would mean over 30 (respectively 100) correct decimal places after 5 (respectively 7) iterations (starting with 1 correct digit); usually no acceleration is needed in that case.) In practice, A /math> often converges much faster to the limit than X does, as demonstrated by the example calculations below. Usually, it is much cheaper to calculate A /math> (involving only calculation of differences, one multiplication and one division) than to calculate many more terms of the sequence X. Care must be taken, however, to avoid introducing errors due to insufficient precision when calculating the differences in the numerator and denominator of the expression.


Example calculations

Example 1: The value of \sqrt \approx 1.4142136 can be approximated by assuming an initial value for x_0 and iterating the following sequence, called Heron's method: x_ = \frac. Starting with x_0 = 1: It is worth noting here that Aitken's method does not save the cost of calculating two iterations here; computation of the first three ''A /math>'' values required the first five X values. Also, the second A /math> value is less accurate than the 4th X value, which is not surprising due to the fact that Aitken's process is best suited for sequences that converge linearly, rather than quadratically, and Heron's method for calculating square roots converges quadratically. Example 2: The value of \frac may be calculated as an infinite sum via the Leibniz formula for ''π'': \frac = \sum_^\infty \frac \approx 0.785398 In this example, Aitken's method is applied to a sublinearly converging series and accelerates convergence considerably. The convergence is still sublinear, but much faster than the original convergence: the first A /math> value, whose computation required the first three X values, is closer to the limit than the eighth X value.


Example pseudocode for Aitken extrapolation

The following is an example of using the Aitken extrapolation to help find the limit of the sequence x_ = f(x_n) when given some initial x_0, where the limit of this sequence is assumed to be a fixed point f (say \alpha = f(\alpha)). For instance, if the sequence is given by x_ = \frac \left(x_n + \frac\right) with starting point x_0 = 1, then the function will be f(x) := \frac\left(x + \frac\right), which has \alpha := \sqrt as a fixed point (see
Methods of computing square roots Square root algorithms compute the non-negative square root \sqrt of a positive real number S. Since all square roots of natural numbers, other than of perfect squares, are irrational, square roots can usually only be computed to some finite pre ...
); it is this fixed point whose value will be approximated. This pseudo code also computes the Aitken approximation to f^(\alpha). The Aitken extrapolates will be denoted by aitkenX. During the computation of the extrapolate, it is important to check if the denominator becomes too small, which could happen if we already have a large amount of accuracy; without this check, a large amount of error could be introduced by the division. This small number will be denoted by epsilon. Because the binary representation of the fixed point could be infinite (or at least too large to fit in the available memory), the calculation will stop once the approximation is within tolerance of the true value. %These choices depend on the problem being solved x0 = 1 %The initial value f(x) = (1/2)*(x + 2/x) %The function that finds the next element in the sequence tolerance = 10^-10 %10 digit accuracy is desired epsilon = 10^-16 %Do not divide by a number smaller than this maxIterations = 20 %Do not allow the iterations to continue indefinitely haveWeFoundSolution = false %Were we able to find the solution to within the desired tolerance? not yet for i = 1 : maxIterations x1 = f(x0) x2 = f(x1) if (x1 ~= x0) lambda = absoluteValue((x2 - x1)/(x1 - x0)) %OPTIONAL: Computes an approximation of , f'(fixedPoint), , which is denoted by lambda end denominator = (x2 - x1) - (x1 - x0); if (absoluteValue(denominator) < epsilon) %To avoid greatly increasing error, do not divide by too small of a number print('WARNING: denominator is too small') break %Leave the loop end aitkenX = x2 - ( (x2 - x1)^2 )/denominator if (absoluteValue(aitkenX - x2) < tolerance) %If the value is within tolerance print("The fixed point is ", aitkenX)) %Display the result of the Aitken extrapolation haveWeFoundSolution = true break %Done, so leave the loop end x0 = aitkenX %Update x0 to start again end if (haveWeFoundSolution

false) %If we were not able to find a solution to within the desired tolerance print("Warning: Not able to find solution to within the desired tolerance of ", tolerance) print("The last computed extrapolate was ", aitkenX) end


See also

*
Rate of convergence In mathematical analysis, particularly numerical analysis, the rate of convergence and order of convergence of a sequence that converges to a limit are any of several characterizations of how quickly that sequence approaches its limit. These are ...
*
Limit of a sequence As the positive integer n becomes larger and larger, the value n\times \sin\left(\tfrac1\right) becomes arbitrarily close to 1. We say that "the limit of the sequence n \times \sin\left(\tfrac1\right) equals 1." In mathematics, the li ...
*
Fixed point iteration In numerical analysis, fixed-point iteration is a method of computing fixed point (mathematics), fixed points of a function. More specifically, given a function f defined on the real numbers with real values and given a point x_0 in the domain of ...
*
Richardson extrapolation In numerical analysis, Richardson extrapolation is a Series acceleration, sequence acceleration method used to improve the rate of convergence of a sequence of estimates of some value A^\ast = \lim_ A(h). In essence, given the value of A(h) for se ...
*
Sequence transformation In mathematics, a sequence transformation is an Operator (mathematics), operator acting on a given space of sequences (a sequence space). Sequence transformations include linear mappings such as convolution, discrete convolution with another sequen ...
*
Shanks transformation In numerical analysis, the Shanks transformation is a non-linear series acceleration method to increase the rate of convergence of a sequence. This method is named after Daniel Shanks, who rediscovered this sequence transformation in 1955. It was fi ...
*
Steffensen's method In numerical analysis, Steffensen's method is an iterative method for numerical root-finding named after Johan Frederik Steffensen that is similar to the secant method and to Newton's method. Steffensen's method achieves a quadratic order of co ...


Notes


References

* William H. Press, ''et al.'', ''Numerical Recipes in C'', (1987) Cambridge University Press, ''(Se
section 5.1
'' * Abramowitz and Stegun, '' Handbook of Mathematical Functions'', section 3.9.7 * Kendall E. Atkinson, ''An Introduction to Numerical Analysis'', (1989) John Wiley & Sons, Inc, {{DEFAULTSORT:Aitken's Delta-Squared Process Series acceleration methods