HOME

TheInfoList



OR:

In statistics, nonlinear regression is a form of regression analysis in which observational data are modeled by a function which is a nonlinear combination of the model parameters and depends on one or more independent variables. The data are fitted by a method of successive approximations.


General

In nonlinear regression, a
statistical model A statistical model is a mathematical model that embodies a set of statistical assumptions concerning the generation of sample data (and similar data from a larger population). A statistical model represents, often in considerably idealized form ...
of the form, : \mathbf \sim f(\mathbf, \boldsymbol\beta) relates a vector of
independent variables Dependent and independent variables are variables in mathematical modeling, statistical modeling and experimental sciences. Dependent variables receive this name because, in an experiment, their values are studied under the supposition or deman ...
, \mathbf, and its associated observed dependent variables, \mathbf. The function f is nonlinear in the components of the vector of parameters \beta, but otherwise arbitrary. For example, the Michaelis–Menten model for enzyme kinetics has two parameters and one independent variable, related by f by: : f(x,\boldsymbol\beta)= \frac This function is nonlinear because it cannot be expressed as a linear combination of the two ''\beta''s.
Systematic error Observational error (or measurement error) is the difference between a measured value of a quantity and its true value.Dodge, Y. (2003) ''The Oxford Dictionary of Statistical Terms'', OUP. In statistics, an error is not necessarily a " mistak ...
may be present in the independent variables but its treatment is outside the scope of regression analysis. If the independent variables are not error-free, this is an
errors-in-variables model In statistics, errors-in-variables models or measurement error models are regression models that account for measurement errors in the independent variables. In contrast, standard regression models assume that those regressors have been measured ...
, also outside this scope. Other examples of nonlinear functions include
exponential function The exponential function is a mathematical function denoted by f(x)=\exp(x) or e^x (where the argument is written as an exponent). Unless otherwise specified, the term generally refers to the positive-valued function of a real variable, ...
s, logarithmic functions,
trigonometric functions In mathematics, the trigonometric functions (also called circular functions, angle functions or goniometric functions) are real functions which relate an angle of a right-angled triangle to ratios of two side lengths. They are widely used in a ...
, power functions, Gaussian function, and Lorentz distributions. Some functions, such as the exponential or logarithmic functions, can be transformed so that they are linear. When so transformed, standard linear regression can be performed but must be applied with caution. See Linearization§Transformation, below, for more details. In general, there is no closed-form expression for the best-fitting parameters, as there is in linear regression. Usually numerical optimization algorithms are applied to determine the best-fitting parameters. Again in contrast to linear regression, there may be many local minima of the function to be optimized and even the global minimum may produce a biased estimate. In practice, estimated values of the parameters are used, in conjunction with the optimization algorithm, to attempt to find the global minimum of a sum of squares. For details concerning nonlinear data modeling see least squares and non-linear least squares.


Regression statistics

The assumption underlying this procedure is that the model can be approximated by a linear function, namely a first-order Taylor series: : f(x_i,\boldsymbol\beta) \approx f(x_i,0) + \sum_j J_ \beta_j where J_ = \frac. It follows from this that the least squares estimators are given by :\hat \approx \mathbf , compare generalized least squares with covariance matrix proportional to the unit matrix. The nonlinear regression statistics are computed and used as in linear regression statistics, but using J in place of X in the formulas. The linear approximation introduces
bias Bias is a disproportionate weight ''in favor of'' or ''against'' an idea or thing, usually in a way that is closed-minded, prejudicial, or unfair. Biases can be innate or learned. People may develop biases for or against an individual, a group ...
into the statistics. Therefore, more caution than usual is required in interpreting statistics derived from a nonlinear model.


Ordinary and weighted least squares

The best-fit curve is often assumed to be that which minimizes the sum of squared residuals. This is the
ordinary least squares In statistics, ordinary least squares (OLS) is a type of linear least squares method for choosing the unknown parameters in a linear regression model (with fixed level-one effects of a linear function of a set of explanatory variables) by the ...
(OLS) approach. However, in cases where the dependent variable does not have constant variance, a sum of weighted squared residuals may be minimized; see weighted least squares. Each weight should ideally be equal to the reciprocal of the variance of the observation, but weights may be recomputed on each iteration, in an iteratively weighted least squares algorithm.


Linearization


Transformation

Some nonlinear regression problems can be moved to a linear domain by a suitable transformation of the model formulation. For example, consider the nonlinear regression problem : y = a e^U \,\! with parameters ''a'' and ''b'' and with multiplicative error term ''U''. If we take the logarithm of both sides, this becomes : \ln = \ln + b x + u, \,\! where ''u'' = ln(''U''), suggesting estimation of the unknown parameters by a linear regression of ln(''y'') on ''x'', a computation that does not require iterative optimization. However, use of a nonlinear transformation requires caution. The influences of the data values will change, as will the error structure of the model and the interpretation of any inferential results. These may not be desired effects. On the other hand, depending on what the largest source of error is, a nonlinear transformation may distribute the errors in a Gaussian fashion, so the choice to perform a nonlinear transformation must be informed by modeling considerations. For Michaelis–Menten kinetics, the linear Lineweaver–Burk plot : \frac = \frac + \frac of 1/''v'' against 1/ 'S''has been much used. However, since it is very sensitive to data error and is strongly biased toward fitting the data in a particular range of the independent variable, 'S'' its use is strongly discouraged. For error distributions that belong to the exponential family, a link function may be used to transform the parameters under the Generalized linear model framework.


Segmentation

The ''independent'' or ''explanatory variable'' (say X) can be split up into classes or segments and linear regression can be performed per segment. Segmented regression with confidence analysis may yield the result that the ''dependent'' or ''response'' variable (say Y) behaves differently in the various segments. The figure shows that the soil salinity (X) initially exerts no influence on the crop yield (Y) of mustard, until a ''critical'' or ''threshold'' value (''breakpoint''), after which the yield is affected negatively.R.J.Oosterbaan, 2002. Drainage research in farmers' fields: analysis of data. Part of project “Liquid Gold” of the International Institute for Land Reclamation and Improvement (ILRI), Wageningen, The Netherlands. Download as PDF

The figure was made with the
SegReg In statistics and data analysis, the application software SegReg is a free and user-friendly tool for linear segmented regression analysis to determine the breakpoint where the relation between the dependent variable and the independent variable c ...
program, which can be downloaded freely fro

/ref>


See also

* Non-linear least squares * Curve fitting * Generalized linear model * Local regression *
Response modeling methodology Response modeling methodology (RMM) is a general platform for statistical modeling of a linear/nonlinear relationship between a response variable (Dependent and independent variables, dependent variable) and a linear predictor (a linear combination ...
* Genetic Programming * Multi expression programming


References


Notes


Further reading

* * * * {{DEFAULTSORT:Nonlinear Regression Regression analysis