COBYLA

From Infogalactic: the planetary knowledge core
Jump to: navigation, search

Constrained optimization by linear approximation (COBYLA) is a numerical optimization method for constrained problems where the derivative of the objective function is not known, invented by Michael J. D. Powell. That is, COBYLA can find the vector \vec{x} \in \mathcal{S} with \mathcal{S} \subseteq \mathbb{R}^n that has the minimal (or maximal) f(\vec{x}) without knowing the gradient of f. COBYLA is also the name of Powell's software implementation of the algorithm in Fortran.[1]

Powell invented COBYLA while working for Westland Helicopters.[2]

It works by iteratively approximating the actual constrained optimization problem with linear programming problems. During an iteration, an approximating linear programming problem is solved to obtain a candidate for the optimal solution. The candidate solution is evaluated using the original objective and constraint functions, yielding a new data point in the optimization space. This information is used to improve the approximating linear programming problem used for the next iteration of the algorithm. When the solution cannot be improved anymore, the step size is reduced, refining the search. When the step size becomes sufficiently small, the algorithm finishes.

The COBYLA software is distributed under The GNU Lesser General Public License (LGPL).[3]

See also

References

  1. Lua error in package.lua at line 80: module 'strict' not found.
  2. M. J. D. Powell (2007). A view of algorithms for optimization without derivatives. Cambridge University Technical Report DAMTP 2007.
  3. Lua error in package.lua at line 80: module 'strict' not found.

External links


<templatestyles src="Asbox/styles.css"></templatestyles>