Descent_direction

Descent direction

Descent direction

Add article description


In optimization, a descent direction is a vector that points towards a local minimum of an objective function .

Computing by an iterative method, such as line search defines a descent direction at the th iterate to be any such that , where denotes the inner product. The motivation for such an approach is that small steps along guarantee that is reduced, by Taylor's theorem.

Using this definition, the negative of a non-zero gradient is always a descent direction, as .

Numerous methods exist to compute descent directions, all with differing merits, such as gradient descent or the conjugate gradient method.

More generally, if is a positive definite matrix, then is a descent direction at .[1] This generality is used in preconditioned gradient descent methods.

See also


References

  1. J. M. Ortega and W. C. Rheinbold (1970). Iterative Solution of Nonlinear Equations in Several Variables. p. 243. doi:10.1137/1.9780898719468.

Share this article:

This article uses material from the Wikipedia article Descent_direction, and is written by contributors. Text is available under a CC BY-SA 4.0 International License; additional terms may apply. Images, videos and audio are available under their respective licenses.