Adaptive filter
   HOME

TheInfoList



OR:

An adaptive filter is a system with a linear
filter Filter, filtering or filters may refer to: Science and technology Computing * Filter (higher-order function), in functional programming * Filter (software), a computer program to process a data stream * Filter (video), a software component tha ...
that has a
transfer function In engineering, a transfer function (also known as system function or network function) of a system, sub-system, or component is a mathematical function that theoretically models the system's output for each possible input. They are widely used ...
controlled by variable parameters and a means to adjust those parameters according to an
optimization algorithm Mathematical optimization (alternatively spelled ''optimisation'') or mathematical programming is the selection of a best element, with regard to some criterion, from some set of available alternatives. It is generally divided into two subfi ...
. Because of the complexity of the optimization algorithms, almost all adaptive filters are
digital filter In signal processing, a digital filter is a system that performs mathematical operations on a sampled, discrete-time signal to reduce or enhance certain aspects of that signal. This is in contrast to the other major type of electronic filter, t ...
s. Adaptive filters are required for some applications because some parameters of the desired processing operation (for instance, the locations of reflective surfaces in a reverberant space) are not known in advance or are changing. The closed loop adaptive filter uses feedback in the form of an error signal to refine its transfer function. Generally speaking, the closed loop adaptive process involves the use of a cost function, which is a criterion for optimum performance of the filter, to feed an algorithm, which determines how to modify filter transfer function to minimize the cost on the next iteration. The most common cost function is the mean square of the error signal. As the power of digital signal processors has increased, adaptive filters have become much more common and are now routinely used in devices such as mobile phones and other communication devices, camcorders and digital cameras, and medical monitoring equipment.


Example application

The recording of a heart beat (an ECG), may be corrupted by noise from the
AC mains Alternating current (AC) is an electric current which periodically reverses direction and changes its magnitude continuously with time in contrast to direct current (DC) which flows only in one direction. Alternating current is the form in which ...
. The exact frequency of the power and its
harmonics A harmonic is a wave with a frequency that is a positive integer multiple of the '' fundamental frequency'', the frequency of the original periodic signal, such as a sinusoidal wave. The original signal is also called the ''1st harmonic'', ...
may vary from moment to moment. One way to remove the noise is to filter the signal with a
notch filter In signal processing, a band-stop filter or band-rejection filter is a filter that passes most frequencies unaltered, but attenuates those in a specific range to very low levels. It is the opposite of a band-pass filter. A notch filter is a ...
at the mains frequency and its vicinity, but this could excessively degrade the quality of the ECG since the heart beat would also likely have frequency components in the rejected range. To circumvent this potential loss of information, an adaptive filter could be used. The adaptive filter would take input both from the patient and from the mains and would thus be able to track the actual frequency of the noise as it fluctuates and subtract the noise from the recording. Such an adaptive technique generally allows for a filter with a smaller rejection range, which means, in this case, that the quality of the output signal is more accurate for medical purposes.


Block diagram

The idea behind a closed loop adaptive filter is that a variable filter is adjusted until the error (the difference between the filter output and the desired signal) is minimized. The Least Mean Squares (LMS) filter and the Recursive Least Squares (RLS) filter are types of adaptive filter. : There are two input signals to the adaptive filter: d_k and x_k which are sometimes called the ''primary input'' and the ''reference input'' respectively. The adaptation algorithm attempts to filter the reference input into a replica of the desired input by minimizing the residual signal, \epsilon_k . When the adaptation is successful, the output of the filter y_k is effectively an estimate of the desired signal. : d_k which includes the desired signal plus undesired interference and : x_k which includes the signals that are correlated to some of the undesired interference in d_k . :k represents the discrete sample number. The filter is controlled by a set of L+1 coefficients or weights. :\mathbf_=\left _,\,w_,\, ...,\,w_ \right represents the set or vector of weights, which control the filter at sample time k. ::where w_ refers to the l'th weight at k'th time. :\mathbf_ represents the change in the weights that occurs as a result of adjustments computed at sample time k. ::These changes will be applied after sample time k and before they are used at sample time k+1. The output is usually \epsilon_k but it could be y_k or it could even be the filter coefficients.(Widrow) The input signals are defined as follows: : d_k = g_k + u_k + v_k : x_k = g_^' + u_^' + v_^' :where: ::''g'' = the desired signal, ::''g'' = a signal that is correlated with the desired signal ''g '', ::''u'' = an undesired signal that is added to ''g '', but not correlated with ''g'' or ''g'' ::''u'' = a signal that is correlated with the undesired signal ''u'', but not correlated with ''g'' or ''g'', ::''v'' = an undesired signal (typically random noise) not correlated with ''g'', ''g'', ''u'', ''u'' or ''v'', ::''v'' = an undesired signal (typically random noise) not correlated with ''g'', ''g'', ''u'', ''u'' or ''v''. The output signals are defined as follows: : y_k= \hat_k + \hat_k + \hat_k : \epsilon_k = d_k-y_k . :where: :: \hat = the output of the filter if the input was only ''g'', :: \hat = the output of the filter if the input was only ''u'', :: \hat = the output of the filter if the input was only ''v''.


Tapped delay line FIR filter

If the variable filter has a tapped delay line Finite Impulse Response (FIR) structure, then the impulse response is equal to the filter coefficients. The output of the filter is given by : y_k= \sum_^L w_ \ x_ = \hat_k + \hat_k + \hat_k ::where w_ refers to the l'th weight at k'th time.


Ideal case

In the ideal case v \equiv 0, v' \equiv 0, g' \equiv 0 . All the undesired signals in d_k are represented by u_k . \ x_k consists entirely of a signal correlated with the undesired signal in u_k . The output of the variable filter in the ideal case is : y_k = \hat_k . The error signal or cost function is the difference between d_k and y_k : \epsilon_k = d_k-y_k = g_k+ u_k - \hat_k . The desired signal ''g''k passes through without being changed. The error signal \epsilon_k is minimized in the mean square sense when _k - \hat_k is minimized. In other words, \hat_k is the best mean square estimate of u_k . In the ideal case, u_k = \hat_k and \epsilon_k = g_k , and all that is left after the subtraction is g which is the unchanged desired signal with all undesired signals removed.


Signal components in the reference input

In some situations, the reference input x_k includes components of the desired signal. This means g' ≠ 0. Perfect cancelation of the undesired interference is not possible in the case, but improvement of the signal to interference ratio is possible. The output will be : \epsilon_k = d_k-y_k = g_k - \hat_k+ u_k - \hat_k . The desired signal will be modified (usually decreased). The output signal to interference ratio has a simple formula referred to as ''power inversion''. : \rho_(z) = \frac . ::where ::: \rho_(z) \ = output signal to interference ratio. ::: \rho_(z) \ = reference signal to interference ratio. ::: z \ = frequency in the z-domain. This formula means that the output signal to interference ratio at a particular frequency is the reciprocal of the reference signal to interference ratio. Example: A fast food restaurant has a drive-up window. Before getting to the window, customers place their order by speaking into a microphone. The microphone also picks up noise from the engine and the environment. This microphone provides the primary signal. The signal power from the customer's voice and the noise power from the engine are equal. It is difficult for the employees in the restaurant to understand the customer. To reduce the amount of interference in the primary microphone, a second microphone is located where it is intended to pick up sounds from the engine. It also picks up the customer's voice. This microphone is the source of the reference signal. In this case, the engine noise is 50 times more powerful than the customer's voice. Once the canceler has converged, the primary signal to interference ratio will be improved from 1:1 to 50:1.


Adaptive Linear Combiner

: The adaptive linear combiner (ALC) resembles the adaptive tapped delay line FIR filter except that there is no assumed relationship between the X values. If the X values were from the outputs of a tapped delay line, then the combination of tapped delay line and ALC would comprise an adaptive filter. However, the X values could be the values of an array of pixels. Or they could be the outputs of multiple tapped delay lines. The ALC finds use as an adaptive beam former for arrays of hydrophones or antennas. : y_k= \sum_^L w_ \ x_ = \mathbf_k^T \mathbf_k ::where w_ refers to the l'th weight at k'th time.


LMS algorithm

If the variable filter has a tapped delay line FIR structure, then the LMS update algorithm is especially simple. Typically, after each sample, the coefficients of the FIR filter are adjusted as follows:(Widrow) : for l = 0 \dots L : w_ = w_ + 2 \mu \ \epsilon_k \ x_ :::μ is called the ''convergence factor''. The LMS algorithm does not require that the X values have any particular relationship; therefore it can be used to adapt a linear combiner as well as an FIR filter. In this case the update formula is written as: : w_ = w_ + 2 \mu \ \epsilon_k \ x_ The effect of the LMS algorithm is at each time, k, to make a small change in each weight. The direction of the change is such that it would decrease the error if it had been applied at time k. The magnitude of the change in each weight depends on μ, the associated X value and the error at time k. The weights making the largest contribution to the output, y_k , are changed the most. If the error is zero, then there should be no change in the weights. If the associated value of X is zero, then changing the weight makes no difference, so it is not changed.


Convergence

μ controls how fast and how well the algorithm converges to the optimum filter coefficients. If μ is too large, the algorithm will not converge. If μ is too small the algorithm converges slowly and may not be able to track changing conditions. If μ is large but not too large to prevent convergence, the algorithm reaches steady state rapidly but continuously overshoots the optimum weight vector. Sometimes, μ is made large at first for rapid convergence and then decreased to minimize overshoot. Widrow and Stearns state in 1985 that they have no knowledge of a proof that the LMS algorithm will converge in all cases.Widrow p 103 However under certain assumptions about stationarity and independence it can be shown that the algorithm will converge if : 0 < \mu < \frac ::where :::\sigma^2 = \sum_^L \sigma_l^2 = sum of all input power :::\sigma_l is the RMS value of the l 'th input In the case of the tapped delay line filter, each input has the same RMS value because they are simply the same values delayed. In this case the total power is : \sigma^2 = (L+1) \sigma_0^2 ::where :::\sigma_0 is the RMS value of x_k, the input stream. This leads to a normalized LMS algorithm: : w_ = w_ + \left ( \frac \right ) \epsilon_k \ x_ in which case the convergence criteria becomes: 0 < \mu_ < 1 .


Nonlinear Adaptive Filters

The goal of nonlinear filters is to overcome limitation of linear models. There are some commonly used approaches: Volterra LMS,
Kernel adaptive filter In signal processing, a kernel adaptive filter is a type of nonlinear adaptive filter. An adaptive filter is a filter that adapts its transfer function to changes in signal properties over time by minimizing an error or loss function that characte ...
, Spline Adaptive Filter and Urysohn Adaptive Filter. Many authors include also Neural networks into this list. The general idea behind Volterra LMS and Kernel LMS is to replace data samples by different nonlinear algebraic expressions. For Volterra LMS this expression is Volterra series. In Spline Adaptive Filter the model is a cascade of linear dynamic block and static non-linearity, which is approximated by splines. In Urysohn Adaptive Filter the linear terms in a model : y_i= \sum_^m w_ \ x_ are replaced by piecewise linear functions : y_i= \sum_^m f_ ( x_) which are identified from data samples.


Applications of adaptive filters

*
Noise cancellation Active noise control (ANC), also known as noise cancellation (NC), or active noise reduction (ANR), is a method for reducing unwanted sound by the addition of a second sound specifically designed to cancel the first. The concept was first develop ...
* Signal prediction *
Adaptive feedback cancellation Adaptive feedback cancellation is a common method of cancelling audio feedback in a variety of electro-acoustic systems such as digital hearing aids. The time varying acoustic feedback leakage paths can only be eliminated with adaptive feedback can ...
*
Echo cancellation Echo suppression and echo cancellation are methods used in telephony to improve voice quality by preventing echo from being created or removing it after it is already present. In addition to improving subjective audio quality, echo suppression ...


Filter implementations

*
Least mean squares filter Least mean squares (LMS) algorithms are a class of adaptive filter used to mimic a desired filter by finding the filter coefficients that relate to producing the least mean square of the error signal (difference between the desired and the actual ...
*
Recursive least squares filter Recursive least squares (RLS) is an adaptive filter algorithm that recursively finds the coefficients that minimize a weighted linear least squares cost function relating to the input signals. This approach is in contrast to other algorithms such ...
*
Multidelay block frequency domain adaptive filter The multidelay block frequency domain adaptive filter (MDF) algorithm is a block-based frequency domain implementation of the (normalised) Least mean squares filter (LMS) algorithm. Introduction The MDF algorithm is based on the fact that convol ...


See also

* 2D adaptive filters * Filter (signal processing) *
Kalman filter For statistics and control theory, Kalman filtering, also known as linear quadratic estimation (LQE), is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, and produces estima ...
*
Kernel adaptive filter In signal processing, a kernel adaptive filter is a type of nonlinear adaptive filter. An adaptive filter is a filter that adapts its transfer function to changes in signal properties over time by minimizing an error or loss function that characte ...
*
Linear prediction Linear prediction is a mathematical operation where future values of a discrete-time signal are estimated as a linear function of previous samples. In digital signal processing, linear prediction is often called linear predictive coding (LPC) and ...
*
MMSE estimator In statistics and signal processing, a minimum mean square error (MMSE) estimator is an estimation method which minimizes the mean square error (MSE), which is a common measure of estimator quality, of the fitted values of a dependent variable. In ...
*
Wiener filter In signal processing, the Wiener filter is a filter used to produce an estimate of a desired or target random process by linear time-invariant ( LTI) filtering of an observed noisy process, assuming known stationary signal and noise spectra, and ...
* Wiener-Hopf equation


References


Sources

* * * {{DEFAULTSORT:Adaptive Filter Digital signal processing Nonlinear filters