Extended Kalman Filter
   HOME

TheInfoList



OR:

In
estimation theory Estimation theory is a branch of statistics that deals with estimating the values of Statistical parameter, parameters based on measured empirical data that has a random component. The parameters describe an underlying physical setting in such ...
, the extended Kalman filter (EKF) is the
nonlinear In mathematics and science, a nonlinear system (or a non-linear system) is a system in which the change of the output is not proportional to the change of the input. Nonlinear problems are of interest to engineers, biologists, physicists, mathe ...
version of the
Kalman filter In statistics and control theory, Kalman filtering (also known as linear quadratic estimation) is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, to produce estimates of unk ...
which linearizes about an estimate of the current mean and
covariance In probability theory and statistics, covariance is a measure of the joint variability of two random variables. The sign of the covariance, therefore, shows the tendency in the linear relationship between the variables. If greater values of one ...
. In the case of well defined transition models, the EKF has been considered the '' de facto'' standard in the theory of nonlinear state estimation,
navigation system A navigation system is a computing system that aids in navigation. Navigation systems may be entirely on board the vehicle or vessel that the system is controlling (for example, on the ship's bridge) or located elsewhere, making use of radio or oth ...
s and
GPS The Global Positioning System (GPS) is a satellite-based hyperbolic navigation system owned by the United States Space Force and operated by Mission Delta 31. It is one of the global navigation satellite systems (GNSS) that provide geol ...
.


History

The papers establishing the mathematical foundations of Kalman type filters were published between 1959 and 1961. The
Kalman filter In statistics and control theory, Kalman filtering (also known as linear quadratic estimation) is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, to produce estimates of unk ...
is the optimal linear estimator for ''linear'' system models with additive independent white noise in both the transition and the measurement systems. Unfortunately, in engineering, most systems are ''nonlinear'', so attempts were made to apply this filtering method to nonlinear systems; most of this work was done at
NASA Ames The Ames Research Center (ARC), also known as NASA Ames, is a major NASA research center at Moffett Federal Airfield in California's Silicon Valley. It was founded in 1939 as the second National Advisory Committee for Aeronautics (NACA) laborat ...
. The EKF adapted techniques from
calculus Calculus is the mathematics, mathematical study of continuous change, in the same way that geometry is the study of shape, and algebra is the study of generalizations of arithmetic operations. Originally called infinitesimal calculus or "the ...
, namely multivariate
Taylor series In mathematics, the Taylor series or Taylor expansion of a function is an infinite sum of terms that are expressed in terms of the function's derivatives at a single point. For most common functions, the function and the sum of its Taylor ser ...
expansions, to linearize a model about a working point. If the system model (as described below) is not well known or is inaccurate, then
Monte Carlo methods Monte Carlo methods, or Monte Carlo experiments, are a broad class of computational algorithms that rely on Resampling (statistics), repeated random sampling to obtain numerical results. The underlying concept is to use randomness to solve pr ...
, especially particle filters, are employed for estimation. Monte Carlo techniques predate the existence of the EKF but are more computationally expensive for any moderately dimensioned
state-space In computer science, a state space is a discrete space representing the set of all possible configurations of a system. It is a useful abstraction for reasoning about the behavior of a given system and is widely used in the fields of artificial ...
.


Formulation

In the extended Kalman filter, the state transition and observation models don't need to be linear functions of the state but may instead be
differentiable In mathematics, a differentiable function of one real variable is a function whose derivative exists at each point in its domain. In other words, the graph of a differentiable function has a non- vertical tangent line at each interior point in ...
functions. :\boldsymbol_ = f(\boldsymbol_, \boldsymbol_) + \boldsymbol_ :\boldsymbol_ = h(\boldsymbol_) + \boldsymbol_ Here w''k'' and v''k'' are the process and observation noises which are both assumed to be zero mean
multivariate Gaussian In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One de ...
noises with
covariance In probability theory and statistics, covariance is a measure of the joint variability of two random variables. The sign of the covariance, therefore, shows the tendency in the linear relationship between the variables. If greater values of one ...
Q''k'' and R''k'' respectively. u''k'' is the control vector. The function ''f'' can be used to compute the predicted state from the previous estimate and similarly the function ''h'' can be used to compute the predicted measurement from the predicted state. However, ''f'' and ''h'' cannot be applied to the covariance directly. Instead a matrix of partial derivatives (the
Jacobian In mathematics, a Jacobian, named for Carl Gustav Jacob Jacobi, may refer to: *Jacobian matrix and determinant (and in particular, the robot Jacobian) *Jacobian elliptic functions *Jacobian variety * Jacobian ideal *Intermediate Jacobian In mat ...
) is computed. At each time step, the Jacobian is evaluated with current predicted states. These matrices can be used in the Kalman filter equations. This process essentially linearizes the non-linear function around the current estimate. See the
Kalman Filter In statistics and control theory, Kalman filtering (also known as linear quadratic estimation) is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, to produce estimates of unk ...
article for notational remarks.


Discrete-time predict and update equations

Notation \hat_ represents the estimate of \mathbf at time ''n'' given observations up to and including at time .


Predict


Update

where the state transition and observation matrices are defined to be the following Jacobians : = \left . \frac \right \vert _ : = \left . \frac \right \vert _


Disadvantages and alternatives

Unlike its linear counterpart, the extended Kalman filter in general is ''not'' an optimal estimator (it is optimal if the measurement and the state transition model are both linear, as in that case the extended Kalman filter is identical to the regular one). In addition, if the initial estimate of the state is wrong, or if the process is modeled incorrectly, the filter may quickly diverge, owing to its linearization. Another problem with the extended Kalman filter is that the estimated covariance matrix tends to underestimate the true covariance matrix and therefore risks becoming
inconsistent In deductive logic, a consistent theory is one that does not lead to a logical contradiction. A theory T is consistent if there is no formula \varphi such that both \varphi and its negation \lnot\varphi are elements of the set of consequences o ...
in the statistical sense without the addition of "stabilising noise" . More generally one should consider the infinite dimensional nature of the nonlinear filtering problem and the inadequacy of a simple mean and variance-covariance estimator to fully represent the optimal filter. It should also be noted that the extended Kalman filter may give poor performances even for very simple one-dimensional systems such as the cubic sensor, where the optimal filter can be bimodal and as such cannot be effectively represented by a single mean and variance estimator, having a rich structure, or similarly for the quadratic sensor. In such cases the
projection filters Projection filters are a set of algorithms based on stochastic analysis and information geometry, or the differential geometric approach to statistics, used to find approximate solutions for Filtering problem (stochastic processes), filtering prob ...
have been studied as an alternative, having been applied also to navigation. Other general nonlinear filtering methods like full
particle filter Particle filters, also known as sequential Monte Carlo methods, are a set of Monte Carlo algorithms used to find approximate solutions for filtering problems for nonlinear state-space systems, such as signal processing and Bayesian statistical ...
s may be considered in this case. Having stated this, the extended Kalman filter can give reasonable performance, and is arguably the
de facto standard A ''de facto'' standard is a custom or convention that is commonly used even though its use is not required. is a Latin phrase (literally " of fact"), here meaning "in practice but not necessarily ordained by law" or "in practice or actuality, ...
in navigation systems and GPS.


Generalizations


Continuous-time extended Kalman filter

Model : \begin \dot(t) &= f\bigl(\mathbf(t), \mathbf(t)\bigr) \mathbf(t) &= h\bigl(\mathbf(t)\bigr)(t) \end Initialize : \hat(t_0)=E\bigl mathbf(t_0)\bigr\text \mathbf(t_0)=Var\bigl mathbf(t_0)\bigr Predict-Update : \begin \dot(t) &= f\bigl(\hat(t),\mathbf(t)\bigr)+\mathbf(t)\Bigl(\mathbf(t)-h\bigl(\hat(t)\bigr)\Bigr)\\ \dot(t) &= \mathbf(t)\mathbf(t)+\mathbf(t)\mathbf(t)^-\mathbf(t)\mathbf(t)\mathbf(t)+\mathbf(t)\\ \mathbf(t) &= \mathbf(t)\mathbf(t)^\mathbf(t)^\\ \mathbf(t) &= \left . \frac \right \vert _\\ \mathbf(t) &= \left . \frac \right \vert _ \end Unlike the discrete-time extended Kalman filter, the prediction and update steps are coupled in the continuous-time extended Kalman filter.


Discrete-time measurements

Most physical systems are represented as continuous-time models while discrete-time measurements are frequently taken for state estimation via a digital processor. Therefore, the system model and measurement model are given by : \begin \dot(t) &= f\bigl(\mathbf(t), \mathbf(t)\bigr) + \mathbf(t) &\mathbf(t) &\sim \mathcal\bigl(\mathbf,\mathbf(t)\bigr) \\ \mathbf_k &= h(\mathbf_k) + \mathbf_k &\mathbf_k &\sim \mathcal(\mathbf,\mathbf_k) \end where \mathbf_k=\mathbf(t_k). Initialize : \hat_=E\bigl mathbf(t_0)\bigr \mathbf_=E\bigl left(\mathbf(t_0)-\hat(t_0)\right)\left(\mathbf(t_0)-\hat(t_0)\right)^T\bigr Predict : \begin \text &\begin \dot(t) = f\bigl(\hat(t), \mathbf(t)\bigr) \\ \dot(t) = \mathbf(t)\mathbf(t)+\mathbf(t)\mathbf(t)^T+ \mathbf(t) \end\qquad \text \begin \hat(t_) = \hat_ \\ \mathbf(t_) = \mathbf_ \end \\ \Rightarrow &\begin \hat_ = \hat(t_k) \\ \mathbf_ = \mathbf(t_k) \end \end where : \mathbf(t) = \left. \frac \right \vert _ Update :\mathbf_ = \mathbf_\mathbf_^\bigl(\mathbf_\mathbf_\mathbf_^ + \mathbf_\bigr)^ :\hat_ = \hat_ + \mathbf_\bigl(\mathbf_ - h(\hat_)\bigr) :\mathbf_ = (\mathbf - \mathbf_\mathbf_)\mathbf_ where : \textbf_ = \left . \frac \right \vert _ The update equations are identical to those of discrete-time extended Kalman filter.


Higher-order extended Kalman filters

The above recursion is a first-order extended Kalman filter (EKF). Higher order EKFs may be obtained by retaining more terms of the Taylor series expansions. For example, second and third order EKFs have been described. However, higher order EKFs tend to only provide performance benefits when the measurement noise is small.


Non-additive noise formulation and equations

The typical formulation of the EKF involves the assumption of additive process and measurement noise. This assumption, however, is not necessary for EKF implementation. Instead, consider a more general system of the form: :\boldsymbol_ = f(\boldsymbol_, \boldsymbol_, \boldsymbol_) :\boldsymbol_ = h(\boldsymbol_, \boldsymbol_) Here w''k'' and v''k'' are the process and observation noises which are both assumed to be zero mean
multivariate Gaussian In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One de ...
noises with
covariance In probability theory and statistics, covariance is a measure of the joint variability of two random variables. The sign of the covariance, therefore, shows the tendency in the linear relationship between the variables. If greater values of one ...
Q''k'' and R''k'' respectively. Then the
covariance In probability theory and statistics, covariance is a measure of the joint variability of two random variables. The sign of the covariance, therefore, shows the tendency in the linear relationship between the variables. If greater values of one ...
prediction and innovation equations become : \boldsymbol_ = : \boldsymbol_ = where the matrices \boldsymbol_ and \boldsymbol_ are Jacobian matrices: : = \left . \frac \right \vert _ : = \left . \frac \right \vert _ The predicted state estimate and measurement residual are evaluated at the mean of the process and measurement noise terms, which is assumed to be zero. Otherwise, the non-additive noise formulation is implemented in the same manner as the additive noise EKF.


Implicit extended Kalman filter

In certain cases, the observation model of a nonlinear system cannot be solved for \boldsymbol_, but can be expressed by the
implicit function In mathematics, an implicit equation is a relation of the form R(x_1, \dots, x_n) = 0, where is a function of several variables (often a polynomial). For example, the implicit equation of the unit circle is x^2 + y^2 - 1 = 0. An implicit func ...
: :h(\boldsymbol_, \boldsymbol_) = \boldsymbol where \boldsymbol_ = \boldsymbol_ + \boldsymbol_ are the noisy observations. The conventional extended Kalman filter can be applied with the following substitutions: : \leftarrow : \tilde_ \leftarrow -h(\hat_, \boldsymbol_) where: : = \left . \frac \right \vert _ Here the original observation covariance matrix is transformed, and the innovation \tilde_ is defined differently. The Jacobian matrix is defined as before, but determined from the implicit observation model h(\boldsymbol_, \boldsymbol_).


Modifications


Iterated extended Kalman filter

The iterated extended Kalman filter improves the linearization of the extended Kalman filter by recursively modifying the centre point of the Taylor expansion. This reduces the linearization error at the cost of increased computational requirements.


Robust extended Kalman filter

The robust extended Kalman filter arises by linearizing the signal model about the current state estimate and using the linear
Kalman filter In statistics and control theory, Kalman filtering (also known as linear quadratic estimation) is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, to produce estimates of unk ...
to predict the next estimate. This attempts to produce a locally optimal filter, however, it is not necessarily stable because the solutions of the underlying
Riccati equation In mathematics, a Riccati equation in the narrowest sense is any first-order ordinary differential equation that is quadratic in the unknown function. In other words, it is an equation of the form y'(x) = q_0(x) + q_1(x) \, y(x) + q_2(x) \, y^2( ...
are not guaranteed to be positive definite. One way of improving performance is the faux algebraic Riccati technique which trades off optimality for stability. The familiar structure of the extended Kalman filter is retained but stability is achieved by selecting a positive definite solution to a faux algebraic Riccati equation for the gain design. Another way of improving extended Kalman filter performance is to employ the H-infinity results from robust control. Robust filters are obtained by adding a positive definite term to the design Riccati equation. The additional term is parametrized by a scalar which the designer may tweak to achieve a trade-off between mean-square-error and peak error performance criteria.


Invariant extended Kalman filter

The invariant extended Kalman filter (IEKF) is a modified version of the EKF for nonlinear systems possessing symmetries (or ''invariances''). It combines the advantages of both the EKF and the recently introduced
symmetry-preserving filter In mathematics, Symmetry-preserving observers,S. Bonnabel, Ph. Martin and E. Salaün, "Invariant Extended Kalman Filter: theory and application to a velocity-aided attitude estimation problem", 48th IEEE Conference on Decision and Control, pp. 129 ...
s. Instead of using a linear correction term based on a linear output error, the IEKF uses a geometrically adapted correction term based on an invariant output error; in the same way the gain matrix is not updated from a linear state error, but from an invariant state error. The main benefit is that the gain and covariance equations converge to constant values on a much bigger set of trajectories than equilibrium points as it is the case for the EKF, which results in a better convergence of the estimation.


Unscented Kalman filters

A nonlinear Kalman filter which shows promise as an improvement over the EKF is the
unscented Kalman filter In statistics and control theory, Kalman filtering (also known as linear quadratic estimation) is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, to produce estimates of un ...
(UKF). In the UKF, the probability density is approximated by a deterministic sampling of points which represent the underlying distribution as a
Gaussian Carl Friedrich Gauss (1777–1855) is the eponym of all of the topics listed below. There are over 100 topics all named after this German mathematician and scientist, all in the fields of mathematics, physics, and astronomy. The English eponymo ...
. The nonlinear transformation of these points are intended to be an estimation of the
posterior distribution The posterior probability is a type of conditional probability that results from updating the prior probability with information summarized by the likelihood via an application of Bayes' rule. From an epistemological perspective, the posterior ...
, the moments of which can then be derived from the transformed samples. The transformation is known as the
unscented transform The unscented transform (UT) is a mathematical function used to estimate the result of applying a given nonlinear transformation to a probability distribution that is characterized only in terms of a finite set of statistics. The most common use o ...
. The UKF tends to be more robust and more accurate than the EKF in its estimation of error in all the directions.
"The extended Kalman filter (EKF) is probably the most widely used estimation algorithm for nonlinear systems. However, more than 35 years of experience in the estimation community has shown that is difficult to implement, difficult to tune, and only reliable for systems that are almost linear on the time scale of the updates. Many of these difficulties arise from its use of linearization."
A 2012 paper includes simulation results which suggest that some published variants of the UKF fail to be as accurate as the Second Order Extended Kalman Filter (SOEKF), also known as the augmented Kalman filter. The SOEKF predates the UKF by approximately 35 years with the moment dynamics first described by Bass et al. The difficulty in implementing any Kalman-type filters for nonlinear state transitions stems from the numerical stability issues required for precision, however the UKF does not escape this difficulty in that it uses linearization as well, namely
linear regression In statistics, linear regression is a statistical model, model that estimates the relationship between a Scalar (mathematics), scalar response (dependent variable) and one or more explanatory variables (regressor or independent variable). A mode ...
. The stability issues for the UKF generally stem from the numerical approximation to the square root of the covariance matrix, whereas the stability issues for both the EKF and the SOEKF stem from possible issues in the
Taylor Series In mathematics, the Taylor series or Taylor expansion of a function is an infinite sum of terms that are expressed in terms of the function's derivatives at a single point. For most common functions, the function and the sum of its Taylor ser ...
approximation along the trajectory.


Ensemble Kalman Filter

The UKF was in fact predated by the Ensemble Kalman filter, invented by Evensen in 1994. It has the advantage over the UKF that the number of ensemble members used can be much smaller than the state dimension, allowing for applications in very high-dimensional systems, such as weather prediction, with state-space sizes of a billion or more.


Fuzzy Kalman Filter

Fuzzy Kalman filter with a new method to represent possibility distributions was recently proposed to replace probability distributions by possibility distributions in order to obtain a genuine possibilistic filter, enabling the use of non-symmetric process and observation noises as well as higher inaccuracies in both process and observation models.


See also

*
Kalman filter In statistics and control theory, Kalman filtering (also known as linear quadratic estimation) is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, to produce estimates of unk ...
* Ensemble Kalman filter *
Fast Kalman filter The fast Kalman filter (FKF), devised by Antti Lange (born 1941), is an extension of the Helmert–Wolf blocking (HWB) method from geodesy to safety-critical real-time applications of Kalman filtering (KF) such as GNSS navigation up to the centimete ...
* Invariant extended Kalman filter *
Moving horizon estimation Moving horizon estimation (MHE) is an optimization approach that uses a series of measurements observed over time, containing noise (random variations) and other inaccuracies, and produces estimates of unknown variables or parameters. Unlike determi ...
*
Particle filter Particle filters, also known as sequential Monte Carlo methods, are a set of Monte Carlo algorithms used to find approximate solutions for filtering problems for nonlinear state-space systems, such as signal processing and Bayesian statistical ...
*
Unscented Kalman filter In statistics and control theory, Kalman filtering (also known as linear quadratic estimation) is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, to produce estimates of un ...
* Nonlinear filtering problem *
Projection filters Projection filters are a set of algorithms based on stochastic analysis and information geometry, or the differential geometric approach to statistics, used to find approximate solutions for Filtering problem (stochastic processes), filtering prob ...


References


Further reading

* * * *


External links


Position estimation of a differential-wheel robot based on odometry and landmarks
{{DEFAULTSORT:Extended Kalman Filter Signal estimation Nonlinear filters Robot control