HOME

TheInfoList



OR:

In
optimization Mathematical optimization (alternatively spelled ''optimisation'') or mathematical programming is the selection of a best element, with regard to some criterion, from some set of available alternatives. It is generally divided into two subfi ...
, the line search strategy is one of two basic
iterative Iteration is the repetition of a process in order to generate a (possibly unbounded) sequence of outcomes. Each repetition of the process is a single iteration, and the outcome of each iteration is then the starting point of the next iteration. ...
approaches to find a
local minimum In mathematical analysis, the maxima and minima (the respective plurals of maximum and minimum) of a function, known collectively as extrema (the plural of extremum), are the largest and smallest value of the function, either within a given ran ...
\mathbf^* of an
objective function In mathematical optimization and decision theory, a loss function or cost function (sometimes also called an error function) is a function that maps an event or values of one or more variables onto a real number intuitively representing some "cost ...
f:\mathbb R^n\to\mathbb R. The other approach is trust region. The line search approach first finds a
descent direction In optimization, a descent direction is a vector \mathbf\in\mathbb R^n that, in the sense below, moves us closer towards a local minimum \mathbf^* of our objective function f:\mathbb R^n\to\mathbb R. Suppose we are computing \mathbf^* by an iterat ...
along which the objective function f will be reduced and then computes a step size that determines how far \mathbf should move along that direction. The descent direction can be computed by various methods, such as
gradient descent In mathematics, gradient descent (also often called steepest descent) is a first-order iterative optimization algorithm for finding a local minimum of a differentiable function. The idea is to take repeated steps in the opposite direction of the ...
or
quasi-Newton method Quasi-Newton methods are methods used to either find zeroes or local maxima and minima of functions, as an alternative to Newton's method. They can be used if the Jacobian or Hessian is unavailable or is too expensive to compute at every iteration. ...
. The step size can be determined either exactly or inexactly.


Example use

Here is an example gradient method that uses a line search in step 4. # Set iteration counter \displaystyle k=0, and make an initial guess \mathbf_0 for the minimum # Repeat: #     Compute a
descent direction In optimization, a descent direction is a vector \mathbf\in\mathbb R^n that, in the sense below, moves us closer towards a local minimum \mathbf^* of our objective function f:\mathbb R^n\to\mathbb R. Suppose we are computing \mathbf^* by an iterat ...
\mathbf_k #     Choose \displaystyle \alpha_k to 'loosely' minimize h(\alpha_k)=f(\mathbf_k+\alpha_k\mathbf_k) over \alpha_k\in\mathbb R_+ #     Update \mathbf_=\mathbf_k+\alpha_k\mathbf_k, and k=k+1 # Until \, \nabla f(\mathbf_)\, < tolerance At the line search step (4) the algorithm might either ''exactly'' minimize ''h'', by solving h'(\alpha_k)=0, or ''loosely'', by asking for a sufficient decrease in ''h''. One example of the former is
conjugate gradient method In mathematics, the conjugate gradient method is an algorithm In mathematics and computer science, an algorithm () is a finite sequence of rigorous instructions, typically used to solve a class of specific problems or to perform a c ...
. The latter is called inexact line search and may be performed in a number of ways, such as a backtracking line search or using the Wolfe conditions. Like other optimization methods, line search may be combined with
simulated annealing Simulated annealing (SA) is a probabilistic technique for approximating the global optimum of a given function. Specifically, it is a metaheuristic to approximate global optimization in a large search space for an optimization problem. ...
to allow it to jump over some local minima.


Algorithms


Direct search methods

In this method, the minimum must first be bracketed, so the algorithm must identify points ''x''1 and ''x''2 such that the sought minimum lies between them. The interval is then divided by computing f(x) at two internal points, ''x''3 and ''x''4, and rejecting whichever of the two outer points is not adjacent to that of ''x''3 and ''x''4 which has the lowest function value. In subsequent steps, only one extra internal point needs to be calculated. Of the various methods of dividing the interval, golden section search is particularly simple and effective, as the interval proportions are preserved regardless of how the search proceeds: :\frac(x_2-x_1)=x_4-x_1=x_2-x_3=\varphi(x_2-x_4)=\varphi(x_3-x_1)=\varphi^2(x_4-x_3) where : \varphi=\frac(1+\sqrt 5) \approx 1.618


See also

* Golden section search * Grid search *
Learning rate In machine learning and statistics, the learning rate is a tuning parameter in an optimization algorithm that determines the step size at each iteration while moving toward a minimum of a loss function. Since it influences to what extent newly ...
* Pattern search (optimization) *
Secant method In numerical analysis, the secant method is a root-finding algorithm that uses a succession of roots of secant lines to better approximate a root of a function ''f''. The secant method can be thought of as a finite-difference approximation o ...


References


Further reading

* * * {{DEFAULTSORT:Line Search Optimization algorithms and methods