scispace - formally typeset
Open AccessJournal ArticleDOI

Conditioning of Quasi-Newton Methods for Function Minimization

David F. Shanno
- 01 Jul 1970 - 
- Vol. 24, Iss: 111, pp 647-656
Reads0
Chats0
TLDR
In this paper, a class of approximating matrices as a function of a scalar parameter is presented, where the problem of optimal conditioning of these matrices under an appropriate norm is investigated and a set of computational results verifies the superiority of the new methods arising from conditioning considerations to known methods.
Abstract
Quasi-Newton methods accelerate the steepest-descent technique for function minimization by using computational history to generate a sequence of approximations to the inverse of the Hessian matrix. This paper presents a class of approximating matrices as a function of a scalar parameter. The problem of optimal conditioning of these matrices under an appropriate norm as a function of the scalar parameter is investigated. A set of computational results verifies the superiority of the new methods arising from conditioning considerations to known methods.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Ab initio calculation of the elastic properties and the lattice dynamics of the ZnxCd1−xSe alloy

TL;DR: In this paper, the density-functional perturbation theory (DFPT) within local density approximation (LDA) and employing virtual-crystal approximation (VCA) was used to study the lattice dynamics and elastic properties of the ternary ZnxCd1−xSe alloy.
Posted ContentDOI

proDA: Probabilistic Dropout Analysis for Identifying Differentially Abundant Proteins in Label-Free Mass Spectrometry

TL;DR: It is shown that the currently widely used methods based on ad hoc imputation schemes can report excessive false positives, and that proDA not only overcomes this serious issue but also offers high sensitivity, filling a crucial gap in the toolbox of quantitative proteomics.
Journal ArticleDOI

Rates of superlinear convergence for classical quasi-Newton methods

TL;DR: In this article, the authors obtained the first explicit nonasymptotic rates of superlinear convergence for the standard quasi-Newton methods, which are based on the updating formulas from the convex Broyden class.
Posted Content

Non-asymptotic Superlinear Convergence of Standard Quasi-Newton Methods.

TL;DR: This paper provides a finite time (non-asymptotic) convergence analysis for BFGS and DFP methods under the assumptions that the objective function is strongly convex, its gradient is Lipschitz continuous, and its Hessians are continuous only in the direction of the optimal solution.
Journal ArticleDOI

Constrained nonlinear least squares: an exact penalty approach with projected structured quasi-Newton updates

TL;DR: The results indicate that the approach taken here is a viable alternative for least squares problems to the general nonlinear methods studied by Hock and Schittkowski.
References
More filters
Journal ArticleDOI

A Rapidly Convergent Descent Method for Minimization

TL;DR: A number of theorems are proved to show that it always converges and that it converges rapidly, and this method has been used to solve a system of one hundred non-linear simultaneous equations.
Journal ArticleDOI

A family of variable-metric methods derived by variational means

TL;DR: In this paper, a rank-two variable-metric method was derived using Greenstadt's variational approach, which preserves the positive-definiteness of the approximating matrix.
Journal ArticleDOI

A Class of Methods for Solving Nonlinear Simultaneous Equations

TL;DR: In this article, the authors discuss certain modifications to Newton's method designed to reduce the number of function evaluations required during the iterative solution process of an iterative problem solving problem, such that the most efficient process will be that which requires the smallest number of functions evaluations.
Journal ArticleDOI

Quasi-Newton methods and their application to function minimisation

TL;DR: The Newton-Raphson method as mentioned in this paper is one of the most commonly used methods for solving nonlinear problems, where the corrections are computed as linear combinations of the residuals.
Journal ArticleDOI

A Comparison of Several Current Optimization Methods, and the use of Transformations in Constrained Problems

TL;DR: Transitions whereby inequality constraints of certain forms can be eliminated from the formulation of an optimization problem are described, and examples of their use compared with other methods for handling such constraints are described.