HOME

TheInfoList



OR:

Numerical methods for ordinary differential equations are methods used to find numerical approximations to the solutions of
ordinary differential equation In mathematics, an ordinary differential equation (ODE) is a differential equation whose unknown(s) consists of one (or more) function(s) of one variable and involves the derivatives of those functions. The term ''ordinary'' is used in contrast ...
s (ODEs). Their use is also known as "
numerical integration In analysis, numerical integration comprises a broad family of algorithms for calculating the numerical value of a definite integral, and by extension, the term is also sometimes used to describe the numerical solution of differential equations ...
", although this term can also refer to the computation of
integral In mathematics, an integral assigns numbers to functions in a way that describes displacement, area, volume, and other concepts that arise by combining infinitesimal data. The process of finding integrals is called integration. Along wit ...
s. Many differential equations cannot be solved exactly. For practical purposes, however – such as in engineering – a numeric approximation to the solution is often sufficient. The
algorithm In mathematics and computer science, an algorithm () is a finite sequence of rigorous instructions, typically used to solve a class of specific problems or to perform a computation. Algorithms are used as specifications for performing ...
s studied here can be used to compute such an approximation. An alternative method is to use techniques from
calculus Calculus, originally called infinitesimal calculus or "the calculus of infinitesimals", is the mathematical study of continuous change, in the same way that geometry is the study of shape, and algebra is the study of generalizations of arithm ...
to obtain a
series expansion In mathematics, a series expansion is an expansion of a function into a series, or infinite sum. It is a method for calculating a function that cannot be expressed by just elementary operators (addition, subtraction, multiplication and divisi ...
of the solution. Ordinary differential equations occur in many scientific disciplines, including
physics Physics is the natural science that studies matter, its fundamental constituents, its motion and behavior through space and time, and the related entities of energy and force. "Physical science is that department of knowledge which r ...
, chemistry,
biology Biology is the scientific study of life. It is a natural science with a broad scope but has several unifying themes that tie it together as a single, coherent field. For instance, all organisms are made up of cells that process hereditary i ...
, and
economics Economics () is the social science that studies the production, distribution, and consumption of goods and services. Economics focuses on the behaviour and interactions of economic agents and how economies work. Microeconomics analyzes ...
. In addition, some methods in
numerical partial differential equations Numerical methods for partial differential equations is the branch of numerical analysis that studies the numerical solution of partial differential equations (PDEs). In principle, specialized methods for hyperbolic, parabolic or elliptic parti ...
convert the partial differential equation into an ordinary differential equation, which must then be solved.


The problem

A first-order differential equation is an
Initial value problem In multivariable calculus, an initial value problem (IVP) is an ordinary differential equation together with an initial condition which specifies the value of the unknown function at a given point in the domain. Modeling a system in physics or ...
(IVP) of the form, where f is a function f:[t_0, \infty) \times \R^d \to \R^d, and the initial condition y_0 \in \R^d is a given vector. ''First-order'' means that only the first derivative of ''y'' appears in the equation, and higher derivatives are absent. Without loss of generality to higher-order systems, we restrict ourselves to ''first-order'' differential equations, because a higher-order ODE can be converted into a larger system of first-order equations by introducing extra variables. For example, the second-order equation can be rewritten as two first-order equations: and In this section, we describe numerical methods for IVPs, and remark that ''boundary value problems'' (BVPs) require a different set of tools. In a BVP, one defines values, or components of the solution ''y'' at more than one point. Because of this, different methods need to be used to solve BVPs. For example, the shooting method (and its variants) or global methods like
finite difference A finite difference is a mathematical expression of the form . If a finite difference is divided by , one gets a difference quotient. The approximation of derivatives by finite differences plays a central role in finite difference methods for t ...
s,LeVeque, R. J. (2007). Finite difference methods for ordinary and partial differential equations: steady-state and time-dependent problems (Vol. 98). SIAM.
Galerkin method In mathematics, in the area of numerical analysis, Galerkin methods, named after the Russian mathematician Boris Galerkin, convert a continuous operator problem, such as a differential equation, commonly in a weak formulation, to a discrete prob ...
s, or
collocation method In mathematics, a collocation method is a method for the numerical solution of ordinary differential equations, partial differential equations and integral equations. The idea is to choose a finite-dimensional space of candidate solutions (usually ...
s are appropriate for that class of problems. The
Picard–Lindelöf theorem In mathematics – specifically, in differential equations – the Picard–Lindelöf theorem gives a set of conditions under which an initial value problem has a unique solution. It is also known as Picard's existence theorem, the Cauc ...
states that there is a unique solution, provided ''f'' is Lipschitz-continuous.


Methods

Numerical methods for solving first-order IVPs often fall into one of two large categories:
linear multistep method Linear multistep methods are used for the numerical solution of ordinary differential equations. Conceptually, a numerical method starts from an initial point and then takes a short step forward in time to find the next solution point. The proce ...
s, or Runge–Kutta methods. A further division can be realized by dividing methods into those that are explicit and those that are implicit. For example, implicit
linear multistep method Linear multistep methods are used for the numerical solution of ordinary differential equations. Conceptually, a numerical method starts from an initial point and then takes a short step forward in time to find the next solution point. The proce ...
s include Adams-Moulton methods, and backward differentiation methods (BDF), whereas
implicit Runge–Kutta methods Implicit may refer to: Mathematics * Implicit function * Implicit function theorem * Implicit curve * Implicit surface * Implicit differential equation Other uses * Implicit assumption, in logic * Implicit-association test, in social psychology ...
include diagonally implicit Runge–Kutta (DIRK), singly diagonally implicit Runge–Kutta (SDIRK), and Gauss–Radau (based on
Gaussian quadrature In numerical analysis, a quadrature rule is an approximation of the definite integral of a function, usually stated as a weighted sum of function values at specified points within the domain of integration. (See numerical integration for mor ...
) numerical methods. Explicit examples from the linear multistep family include the
Adams–Bashforth methods Linear multistep methods are used for the numerical solution of ordinary differential equations. Conceptually, a numerical method starts from an initial point and then takes a short step forward in time to find the next solution point. The proce ...
, and any Runge–Kutta method with a lower diagonal
Butcher tableau A butcher is a person who may slaughter animals, dress their flesh, sell their meat, or participate within any combination of these three tasks. They may prepare standard cuts of meat and poultry for sale in retail or wholesale food establishm ...
is explicit. A loose rule of thumb dictates that stiff differential equations require the use of implicit schemes, whereas non-stiff problems can be solved more efficiently with explicit schemes. The so-called
general linear methods General linear methods (GLMs) are a large class of numerical methods used to obtain numerical solutions to ordinary differential equations. They include multistage Runge–Kutta methods that use intermediate collocation points, as well as linea ...
(GLMs) are a generalization of the above two large classes of methods.


Euler method

From any point on a curve, you can find an approximation of a nearby point on the curve by moving a short distance along a line
tangent In geometry, the tangent line (or simply tangent) to a plane curve at a given point is the straight line that "just touches" the curve at that point. Leibniz defined it as the line through a pair of infinitely close points on the curve. Mo ...
to the curve. Starting with the differential equation (), we replace the derivative ''y''′ by the
finite difference A finite difference is a mathematical expression of the form . If a finite difference is divided by , one gets a difference quotient. The approximation of derivatives by finite differences plays a central role in finite difference methods for t ...
approximation which when re-arranged yields the following formula : y(t+h) \approx y(t) + hy'(t) and using () gives: This formula is usually applied in the following way. We choose a step size ''h'', and we construct the sequence t_0, t_1 = t_0 + h, t_2 = t_0 + 2h,... We denote by y_n a numerical estimate of the exact solution y(t_n). Motivated by (), we compute these estimates by the following
recursive Recursion (adjective: ''recursive'') occurs when a thing is defined in terms of itself or of its type. Recursion is used in a variety of disciplines ranging from linguistics to logic. The most common application of recursion is in mathematics ...
scheme This is the ''
Euler method In mathematics and computational science, the Euler method (also called forward Euler method) is a first-order numerical procedure for solving ordinary differential equations (ODEs) with a given initial value. It is the most basic explicit m ...
'' (or '' forward Euler method'', in contrast with the ''backward Euler method'', to be described below). The method is named after
Leonhard Euler Leonhard Euler ( , ; 15 April 170718 September 1783) was a Swiss mathematician, physicist, astronomer, geographer, logician and engineer who founded the studies of graph theory and topology and made pioneering and influential discoveries in ma ...
who described it in 1768. The Euler method is an example of an ''explicit'' method. This means that the new value ''y''''n''+1 is defined in terms of things that are already known, like ''y''''n''.


Backward Euler method

If, instead of (), we use the approximation we get the ''backward Euler method'': The backward Euler method is an ''implicit'' method, meaning that we have to solve an equation to find ''y''''n''+1. One often uses fixed-point iteration or (some modification of) the
Newton–Raphson method In numerical analysis, Newton's method, also known as the Newton–Raphson method, named after Isaac Newton and Joseph Raphson, is a root-finding algorithm which produces successively better approximations to the roots (or zeroes) of a real-va ...
to achieve this. It costs more time to solve this equation than explicit methods; this cost must be taken into consideration when one selects the method to use. The advantage of implicit methods such as () is that they are usually more stable for solving a
stiff equation In mathematics, a stiff equation is a differential equation for which certain numerical methods for solving the equation are numerically unstable, unless the step size is taken to be extremely small. It has proven difficult to formulate a precise ...
, meaning that a larger step size ''h'' can be used.


First-order exponential integrator method

Exponential integrators describe a large class of integrators that have recently seen a lot of development. This is a modern and extensive review paper for exponential integrators They date back to at least the 1960s. In place of (), we assume the differential equation is either of the form or it has been locally linearized about a background state to produce a linear term -Ay and a nonlinear term \mathcal(y). Exponential integrators are constructed by multiplying () by e^, and exactly integrating the result over a time interval _n, t_ = t_n + h/math>: : y_ = e^ y_n + \int_^ e^ \mathcal\left( y\left( t_n+\tau \right) \right)\, d\tau. This integral equation is exact, but it doesn't define the integral. The first-order exponential integrator can be realized by holding \mathcal( y( t_n+\tau ) ) constant over the full interval:


Generalizations

The Euler method is often not accurate enough. In more precise terms, it only has order one (the concept of ''order'' is explained below). This caused mathematicians to look for higher-order methods. One possibility is to use not only the previously computed value ''y''''n'' to determine ''y''''n''+1, but to make the solution depend on more past values. This yields a so-called ''multistep method''. Perhaps the simplest is the leapfrog method which is second order and (roughly speaking) relies on two time values. Almost all practical multistep methods fall within the family of
linear multistep method Linear multistep methods are used for the numerical solution of ordinary differential equations. Conceptually, a numerical method starts from an initial point and then takes a short step forward in time to find the next solution point. The proce ...
s, which have the form :\begin & \alpha_k y_ + \alpha_ y_ + \cdots + \alpha_0 y_n \\ & \quad = h \left \beta_k f(t_,y_) + \beta_ f(t_,y_) + \cdots + \beta_0 f(t_n,y_n) \right \end Another possibility is to use more points in the interval _n,t_/math>. This leads to the family of Runge–Kutta methods, named after
Carl Runge Carl David Tolmé Runge (; 30 August 1856 – 3 January 1927) was a German mathematician, physicist, and spectroscopist. He was co-developer and co- eponym of the Runge–Kutta method (German pronunciation: ), in the field of what is today know ...
and
Martin Kutta Martin Wilhelm Kutta (; 3 November 1867 – 25 December 1944) was a German mathematician. Kutta was born in Pitschen, Upper Silesia (today Byczyna, Poland). He attended the University of Breslau from 1885 to 1890, and continued his studies in Mu ...
. One of their fourth-order methods is especially popular.


Advanced features

A good implementation of one of these methods for solving an ODE entails more than the time-stepping formula. It is often inefficient to use the same step size all the time, so ''variable step-size methods'' have been developed. Usually, the step size is chosen such that the (local) error per step is below some tolerance level. This means that the methods must also compute an ''error indicator'', an estimate of the local error. An extension of this idea is to choose dynamically between different methods of different orders (this is called a ''variable order method''). Methods based on Richardson extrapolation, such as the Bulirsch–Stoer algorithm, are often used to construct various methods of different orders. Other desirable features include: * ''dense output'': cheap numerical approximations for the whole integration interval, and not only at the points ''t''0, ''t''1, ''t''2, ... * ''event location'': finding the times where, say, a particular function vanishes. This typically requires the use of a
root-finding algorithm In mathematics and computing, a root-finding algorithm is an algorithm for finding zeros, also called "roots", of continuous functions. A zero of a function , from the real numbers to real numbers or from the complex numbers to the complex numbers ...
. * support for parallel computing. * when used for integrating with respect to time, time reversibility


Alternative methods

Many methods do not fall within the framework discussed here. Some classes of alternative methods are: * ''multiderivative methods'', which use not only the function ''f'' but also its derivatives. This class includes ''Hermite–Obreschkoff methods'' and '' Fehlberg methods'', as well as methods like the
Parker–Sochacki method In mathematics, the Parker–Sochacki method is an algorithm for solving systems of ordinary differential equations (ODEs), developed by G. Edgar Parker and James Sochacki, of the James Madison University Mathematics Department. The method produces ...
or Bychkov–Scherbakov method, which compute the coefficients of the Taylor series of the solution ''y'' recursively. * ''methods for second order ODEs''. We said that all higher-order ODEs can be transformed to first-order ODEs of the form (1). While this is certainly true, it may not be the best way to proceed. In particular, ''
Nyström method In mathematics numerical analysis, the Nyström method or quadrature method seeks the numerical solution of an integral equation by replacing the integral with a representative weighted sum. The continuous problem is broken into n discrete inter ...
s'' work directly with second-order equations. * '' geometric integration methods'' are especially designed for special classes of ODEs (for example,
symplectic integrator In mathematics, a symplectic integrator (SI) is a numerical integration scheme for Hamiltonian systems. Symplectic integrators form the subclass of geometric integrators which, by definition, are canonical transformations. They are widely used in ...
s for the solution of Hamiltonian equations). They take care that the numerical solution respects the underlying structure or geometry of these classes. * '' Quantized state systems methods'' are a family of ODE integration methods based on the idea of state quantization. They are efficient when simulating sparse systems with frequent discontinuities.


Parallel-in-time methods

For applications that require parallel computing on supercomputers, the degree of concurrency offered by a numerical method becomes relevant. In view of the challenges from exascale computing systems, numerical methods for
initial value problem In multivariable calculus, an initial value problem (IVP) is an ordinary differential equation together with an initial condition which specifies the value of the unknown function at a given point in the domain. Modeling a system in physics or ...
s which can provide concurrency in temporal direction are being studied. Parareal is a relatively well known example of such a ''parallel-in-time'' integration method, but early ideas go back into the 1960s. In the advent of exascale computing, time-parallel integration methods receive again increased attention. Algorithms for exponential integrators can leverage e.g., the standardized Batched
BLAS Basic Linear Algebra Subprograms (BLAS) is a specification that prescribes a set of low-level routines for performing common linear algebra operations such as vector addition, scalar multiplication, dot products, linear combinations, and matrix ...
functions that allow an easy and efficient implementation of parallelized integrators.


Analysis

Numerical analysis Numerical analysis is the study of algorithms that use numerical approximation (as opposed to symbolic manipulations) for the problems of mathematical analysis (as distinguished from discrete mathematics). It is the study of numerical methods ...
is not only the design of numerical methods, but also their analysis. Three central concepts in this analysis are: * ''convergence'': whether the method approximates the solution, * ''order'': how well it approximates the solution, and * ''stability'': whether errors are damped out.


Convergence

A numerical method is said to be ''convergent'' if the numerical solution approaches the exact solution as the step size ''h'' goes to 0. More precisely, we require that for every ODE (1) with a Lipschitz function ''f'' and every ''t''* > 0, : \lim_ \max_ \left\, y_ - y(t_n) \right\, = 0. All the methods mentioned above are convergent.


Consistency and order

Suppose the numerical method is : y_ = \Psi(t_; y_n, y_, \dots, y_; h). \, The ''local (truncation) error'' of the method is the error committed by one step of the method. That is, it is the difference between the result given by the method, assuming that no error was made in earlier steps, and the exact solution: : \delta^h_ = \Psi \left( t_; y(t_n), y(t_), \dots, y(t_); h \right) - y(t_). The method is said to be ''consistent'' if : \lim_ \frac = 0. The method has ''order'' p if : \delta^h_ = O(h^) \quad\mbox h\to0. Hence a method is consistent if it has an order greater than 0. The (forward) Euler method (4) and the backward Euler method (6) introduced above both have order 1, so they are consistent. Most methods being used in practice attain higher order. Consistency is a necessary condition for convergence, but not sufficient; for a method to be convergent, it must be both consistent and zero-stable. A related concept is the ''global (truncation) error'', the error sustained in all the steps one needs to reach a fixed time t. Explicitly, the global error at time t'' is y_N - y(t) where N = (t - t_0)/h. The global error of a p''th order one-step method is O(h^p); in particular, such a method is convergent. This statement is not necessarily true for multi-step methods.


Stability and stiffness

For some differential equations, application of standard methods—such as the Euler method, explicit Runge–Kutta methods, or multistep methods (for example, Adams–Bashforth methods)—exhibit instability in the solutions, though other methods may produce stable solutions. This "difficult behaviour" in the equation (which may not necessarily be complex itself) is described as ''stiffness'', and is often caused by the presence of different time scales in the underlying problem. For example, a collision in a mechanical system like in an impact oscillator typically occurs at much smaller time scale than the time for the motion of objects; this discrepancy makes for very "sharp turns" in the curves of the state parameters. Stiff problems are ubiquitous in
chemical kinetics Chemical kinetics, also known as reaction kinetics, is the branch of physical chemistry that is concerned with understanding the rates of chemical reactions. It is to be contrasted with chemical thermodynamics, which deals with the direction in ...
,
control theory Control theory is a field of mathematics that deals with the control of dynamical systems in engineered processes and machines. The objective is to develop a model or algorithm governing the application of system inputs to drive the system to a ...
,
solid mechanics Solid mechanics, also known as mechanics of solids, is the branch of continuum mechanics that studies the behavior of solid materials, especially their motion and deformation under the action of forces, temperature changes, phase changes, and ...
,
weather forecasting Weather forecasting is the application of science and technology to predict the conditions of the atmosphere for a given location and time. People have attempted to predict the weather informally for millennia and formally since the 19th cent ...
,
biology Biology is the scientific study of life. It is a natural science with a broad scope but has several unifying themes that tie it together as a single, coherent field. For instance, all organisms are made up of cells that process hereditary i ...
,
plasma physics Plasma ()πλάσμα
, Henry George Liddell, R ...
, and
electronics The field of electronics is a branch of physics and electrical engineering that deals with the emission, behaviour and effects of electrons using electronic devices. Electronics uses active devices to control electron flow by amplification ...
. One way to overcome stiffness is to extend the notion of differential equation to that of differential inclusion, which allows for and models non-smoothness.


History

Below is a
timeline A timeline is a display of a list of events in chronological order. It is typically a graphic design showing a long bar labelled with dates paralleling it, and usually contemporaneous events. Timelines can use any suitable scale represen ...
of some important developments in this field. * 1768 -
Leonhard Euler Leonhard Euler ( , ; 15 April 170718 September 1783) was a Swiss mathematician, physicist, astronomer, geographer, logician and engineer who founded the studies of graph theory and topology and made pioneering and influential discoveries in ma ...
publishes his method. * 1824 -
Augustin Louis Cauchy Baron Augustin-Louis Cauchy (, ; ; 21 August 178923 May 1857) was a French mathematician, engineer, and physicist who made pioneering contributions to several branches of mathematics, including mathematical analysis and continuum mechanics. He w ...
proves convergence of the Euler method. In this proof, Cauchy uses the implicit Euler method. * 1855 - First mention of the multistep methods of John Couch Adams in a letter written by Francis Bashforth. * 1895 -
Carl Runge Carl David Tolmé Runge (; 30 August 1856 – 3 January 1927) was a German mathematician, physicist, and spectroscopist. He was co-developer and co- eponym of the Runge–Kutta method (German pronunciation: ), in the field of what is today know ...
publishes the first Runge–Kutta method. * 1901 -
Martin Kutta Martin Wilhelm Kutta (; 3 November 1867 – 25 December 1944) was a German mathematician. Kutta was born in Pitschen, Upper Silesia (today Byczyna, Poland). He attended the University of Breslau from 1885 to 1890, and continued his studies in Mu ...
describes the popular fourth-order Runge–Kutta method. * 1910 - Lewis Fry Richardson announces his extrapolation method, Richardson extrapolation. * 1952 -
Charles F. Curtiss Charles is a masculine given name predominantly found in English language, English and French language, French speaking countries. It is from the French form ''Charles'' of the Proto-Germanic, Proto-Germanic name (in runic alphabet) or ''*k ...
and
Joseph Oakland Hirschfelder Joseph Oakland Hirschfelder (May 27, 1911 – March 30, 1990) was an American physicist who participated in the Manhattan Project and in the creation of the nuclear bomb.
coin the term ''
stiff equation In mathematics, a stiff equation is a differential equation for which certain numerical methods for solving the equation are numerically unstable, unless the step size is taken to be extremely small. It has proven difficult to formulate a precise ...
s''. * 1963 -
Germund Dahlquist Germund Dahlquist (16 January 1925 – 8 February 2005) was a Swedish mathematician known primarily for his early contributions to the theory of numerical analysis as applied to differential equations. Dahlquist began to study mathematics at Sto ...
introduces '' A-stability'' of integration methods.


Numerical solutions to second-order one-dimensional boundary value problems

Boundary value problems (BVPs) are usually solved numerically by solving an approximately equivalent matrix problem obtained by discretizing the original BVP.Ascher, U. M., Mattheij, R. M., & Russell, R. D. (1995). Numerical solution of boundary value problems for ordinary differential equations. Society for Industrial and Applied Mathematics. The most commonly used method for numerically solving BVPs in one dimension is called the
Finite Difference Method In numerical analysis, finite-difference methods (FDM) are a class of numerical techniques for solving differential equations by approximating derivatives with finite differences. Both the spatial domain and time interval (if applicable) are ...
. This method takes advantage of linear combinations of point values to construct
finite difference coefficient In mathematics, to approximate a derivative to an arbitrary order of accuracy, it is possible to use the finite difference. A finite difference can be central, forward or backward. Central finite difference This table contains the coefficients o ...
s that describe derivatives of the function. For example, the second-order central difference approximation to the first derivative is given by: : \frac = u'(x_i) + \mathcal(h^2), and the second-order central difference for the second derivative is given by: : \frac = u''(x_i) + \mathcal(h^2). In both of these formulae, h=x_i-x_ is the distance between neighbouring ''x'' values on the discretized domain. One then constructs a linear system that can then be solved by standard
matrix methods The matrix method is a structural analysis method used as a fundamental principle in many applications in civil engineering. The method is carried out, using either a stiffness matrix or a flexibility matrix. See also * Direct stiffness method A ...
. For example, suppose the equation to be solved is: : \begin & \frac -u =0,\\ & u(0)=0, \\ & u(1)=1. \end The next step would be to discretize the problem and use linear derivative approximations such as : u''_i =\frac and solve the resulting system of linear equations. This would lead to equations such as: : \frac-u_i = 0, \quad \forall i=. On first viewing, this system of equations appears to have difficulty associated with the fact that the equation involves no terms that are not multiplied by variables, but in fact this is false. At ''i'' = 1 and ''n'' − 1 there is a term involving the boundary values u(0)=u_0 and u(1)=u_n and since these two values are known, one can simply substitute them into this equation and as a result have a non-homogeneous linear system of equations that has non-trivial solutions.


See also

*
Courant–Friedrichs–Lewy condition In mathematics, the convergence condition by Courant–Friedrichs–Lewy is a necessary condition for convergence while solving certain partial differential equations (usually hyperbolic PDEs) numerically. It arises in the numerical analysis of ex ...
* Energy drift *
General linear methods General linear methods (GLMs) are a large class of numerical methods used to obtain numerical solutions to ordinary differential equations. They include multistage Runge–Kutta methods that use intermediate collocation points, as well as linea ...
* List of numerical analysis topics#Numerical methods for ordinary differential equations * Reversible reference system propagation algorithm * Modelica Language and OpenModelica software


Notes


References

* * J. C. Butcher, ''Numerical methods for ordinary differential equations'', *Ernst Hairer, Syvert Paul Nørsett and Gerhard Wanner, ''Solving ordinary differential equations I: Nonstiff problems,'' second edition, Springer Verlag, Berlin, 1993. . *Ernst Hairer and Gerhard Wanner, ''Solving ordinary differential equations II: Stiff and differential-algebraic problems,'' second edition, Springer Verlag, Berlin, 1996. .
''(This two-volume monograph systematically covers all aspects of the field.)'' * *Arieh Iserles, ''A First Course in the Numerical Analysis of Differential Equations,'' Cambridge University Press, 1996. (hardback), (paperback).
''(Textbook, targeting advanced undergraduate and postgraduate students in mathematics, which also discusses
numerical partial differential equations Numerical methods for partial differential equations is the branch of numerical analysis that studies the numerical solution of partial differential equations (PDEs). In principle, specialized methods for hyperbolic, parabolic or elliptic parti ...
.)'' *John Denholm Lambert, ''Numerical Methods for Ordinary Differential Systems,'' John Wiley & Sons, Chichester, 1991. .
''(Textbook, slightly more demanding than the book by Iserles.)''


External links

* Joseph W. Rudmin,
Application of the Parker–Sochacki Method to Celestial Mechanics
', 1998. * Dominique Tournès,

', thèse de doctorat de l'université Paris 7 - Denis Diderot, juin 1996. Réimp. Villeneuve d'Ascq : Presses universitaires du Septentrion, 1997, 468 p. (Extensive online material on ODE numerical analysis history, for English-language material on the history of ODE numerical analysis, see, for example, the paper books by Chabert and Goldstine quoted by him.) * * (
C++ C++ (pronounced "C plus plus") is a high-level general-purpose programming language created by Danish computer scientist Bjarne Stroustrup as an extension of the C programming language, or "C with Classes". The language has expanded significan ...
library with rigorous ODE solvers)
INTLAB
(A library made by
MATLAB MATLAB (an abbreviation of "MATrix LABoratory") is a proprietary multi-paradigm programming language and numeric computing environment developed by MathWorks. MATLAB allows matrix manipulations, plotting of functions and data, implementa ...
/ GNU Octave which includes rigorous ODE solvers) {{DEFAULTSORT:Numerical Ordinary Differential Equations Ordinary differential equations