Abstract
Variable Metric Methods are “Newton—Raphson-like” algorithms for unconstrained minimization in which the inverse Hessian is replaced by an approximation, inferred from previous gradients and updated at each iteration. During the past decade various approaches have been used to derive general classes of such algorithms having the common properties of being Conjugate Directions methods and having “quadratic termination”. Observed differences in actual performance of such methods motivated recent attempts to identify variable metric algorithms having additional properties that may be significant in practical situations (e.g. nonquadratic functions, inaccurate linesearch, etc.). The SSVM algorithms, introduced by this first author, are such methods that among their other properties, they automatically compensate for poor scaling of the objective function. This paper presents some new theoretical results identifying a subclass of SSVM algorithms that have the additional property of minimizing a sharp bound on the condition number of the inverse Hessian approximation at each iteration. Reducing this condition number is important for decreasing the roundoff error. The theoretical properties of this subclass are explored and two of its special cases are tested numerically in comparison with other SSVM algorithms.
Similar content being viewed by others
References
R. Fletcher, “A new approach to variable metric algorithms”,The Computer Journal 13 (1970) 317–322.
A.A. Goldstein, “On steepest descent”,SIAM Journal on Control 1 (1965) 147–151.
H.Y. Huang, “Unified approach to quadratically convergent algorithms for function minimization”,Journal of Optimization Theory and Applications 5 (1970) 405–423.
S.S. Oren, “Self-scaling variable metric algorithms for unconstrained minimization”, Ph.D. thesis, Department of Engineering-Economic Systems, Stanford University, Stanford, Calif., 1972.
S.S. Oren and D.G. Luenberger, “Self-scaling variable metric (SSVM) algorithms I: criteria and sufficient conditions for scaling a class of algorithms”,Management Science 20 (1974) 845–862.
S.S. Oren, “Self-scaling variable metric (SSVM) algorithms II: implementation and experiments”,Management Science 20 (1974) 863–874.
S.S. Oren, “Self-scaling variable metric algorithm without linesearch for unconstrained minimization”,Mathematics of Computation 27 (1973) 873–885.
S.S. Oren, “On the selection of parameters in self-scaling variable metric algorithms”, PARC Memo Rept., ARG MR# 73-8 (Presented at the 8th International Symposium on Mathematical Programming, Stanford, August 1973).
D.F. Shanno and P.C. Kettler, “Optimal conditioning of quasi-Newton methods”,Mathematics of Computation 24 (1970) 657–667.
E. Spedicato, “Stability of Huang's update for the conjugate gradient method”,Journal of Optimization Theory and Applications 11 (1973) 469–479.
E. Spedicato, “On condition numbers of matrices in rank two minimization algorithms”, CISE Rept., CISE, Segrete, Italy.
E. Spedicato, “A bound on the condition number of rank-two corrections and applications to the variable metric methods”,Mathematics to Computation, to appear.
Author information
Authors and Affiliations
Additional information
This work has been done while this author was a visiting fellow at the Engineering Economic System Department, Stanford University.
Rights and permissions
About this article
Cite this article
Oren, S.S., Spedicato, E. Optimal conditioning of self-scaling variable Metric algorithms. Mathematical Programming 10, 70–90 (1976). https://doi.org/10.1007/BF01580654
Received:
Revised:
Published:
Issue Date:
DOI: https://doi.org/10.1007/BF01580654