Skip to main content

Computing Sparse Hessian and Jacobian Approximations with Optimal Hereditary Properties

  • Chapter
Large-Scale Optimization with Applications

Abstract

In nonlinear optimization it is often important to estimate large sparse Hessian or Jacobian matrices, to be used for example in a trust region method. We propose an algorithm for computing a matrix B with a given sparsity pattern from a bundle of the m most recent difference vectors

$$\Delta = \left[ {{{\delta }^{{k - m + 1}}} \ldots {{\delta }^{k}}} \right],\Gamma = \left[ {{{\gamma }^{{k - m + 1}}} \ldots {{\gamma }^{k}}} \right] $$

where B should approximately map △ into Г. In this paper B is chosen such that it satisfies m quasi—Newton conditions B△ = Г in the least squares sense.

We show that B can always be computed by solving a positive semi—definite system of equations in the nonzero components of B. We give necessary and sufficient conditions under which this system is positive definite and indicate how B can be computed efficiently using a conjugate gradient method.

In the case of unconstrained optimization we use the technique to determine a Hessian approximation which is used in a trust region method. Some numerical results are presented for a range of unconstrained test problems.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 54.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. L.-E. Anderson and T. Elfving. A Constrained Procrustes Problem. Technical Report LiTH-MAT-R-1993–39, Department of Mathematics, Linköping University, S-581 83 Linköpin, Sweden, November 1994. Submitted to SIAM Journal on Matrix Analysis and Applications.

    Google Scholar 

  2. N.I.M. Gould A.R. Conn and PH.L. Toint. A proposal forastandard data inputformat for large–scale nonlinear programming problems. Technical Report CS-89–61, University of Waterloo, Computer Science Department, Ontario, Canada, March 1990.

    Google Scholar 

  3. M.J.D. Powell A.R. Curtis and J.K. Reid. On the estimation of sparse Jacobian matrices. Journal of the Institute of Mathematics and its Applications, 13: 117–120, 1974.

    MATH  Google Scholar 

  4. J. Barnes. An algorithm for solving nonlinear equations based on the secant method. Computer Journal, 8: 66–67, 1965.

    Article  MathSciNet  MATH  Google Scholar 

  5. C.-T. Chen. Linear Systems Theory and Design. Holt, Rinehart and Winston, New York, 1984.

    Google Scholar 

  6. TH.F. Coleman. Large-scale numerical Optimization: Introduction and Overview. Technical Report CTC91TR85, Cornell Theory Center, Cornell University, Ithaca, NY 14853–5201, September 1991.

    Google Scholar 

  7. R. Fletcher. Practical Methods of Optimization, 2nd edition. John Wiley, Chichester, 1987.

    Google Scholar 

  8. R. Fletcher. An optimal positive definite update for sparse Hessian matrices. SIAM Journal on Optimization, 5 (1): 192–218, February 1995.

    Article  MathSciNet  MATH  Google Scholar 

  9. N. J. Higham. Analysis of the Cholesky decomposition of a semi–definite matrix. In M.G. Cox and S. Hammarling, editors, Reliable Numerical Computation, pages 161–185, Oxford, 1990. Oxford University Press.

    Google Scholar 

  10. N.J. Higham. The Symmetric Procrustes Problem. BIT, 28: 133–143, 1988.

    Article  MathSciNet  MATH  Google Scholar 

  11. M. Iri. History of automatic differentiation and rounding error estimation. In A. Griewank and G.F. Corliss, editors, Automatic Differentiation of Algorithms, pages 3–24, Philadelphia, 1991. SIAM.

    Google Scholar 

  12. P. Lancaster and M. Tismenetsky. The Theory of Matrices, Second Edition. Computer Science and Applied Mathematics. Academic Press, New York, 1985.

    Google Scholar 

  13. J. Nocedal. Updating quasi—Newton matrices with limited storage. Mathematics of Computation, 35: 773–782, 1980.

    Article  MathSciNet  MATH  Google Scholar 

  14. J.M. Ortega. Matrix Theory. Plenum Press, New York, 1987.

    MATH  Google Scholar 

  15. M.J.D Powell and PH.L. Toint. On the estimation of sparse Hessian matrices. SIAM Journal on Numerical Analysis, 16 (6): 1060–1074, December 1979.

    Article  MathSciNet  MATH  Google Scholar 

  16. PH.L. Toint. On sparse and symmetric matrix updating subject to a linear equation. Mathematics of Computation, 31(140): 954–961, October 1977.

    Article  MathSciNet  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 1997 Springer Science+Business Media New York

About this chapter

Cite this chapter

Fletcher, R., Grothey, A., Leyffer, S. (1997). Computing Sparse Hessian and Jacobian Approximations with Optimal Hereditary Properties. In: Biegler, L.T., Coleman, T.F., Conn, A.R., Santosa, F.N. (eds) Large-Scale Optimization with Applications. The IMA Volumes in Mathematics and its Applications, vol 93. Springer, New York, NY. https://doi.org/10.1007/978-1-4612-1960-6_3

Download citation

  • DOI: https://doi.org/10.1007/978-1-4612-1960-6_3

  • Publisher Name: Springer, New York, NY

  • Print ISBN: 978-1-4612-7356-1

  • Online ISBN: 978-1-4612-1960-6

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics