Abstract
In a previous work of the author, the convergence of the algorithm
for the minimization of f: IRn → IR, was analyzed. Due to the nonlinearity of the matrix Sk the cost of evaluation of xk(t) for several values of t can be expensive, hence the choice of the first trial αk is critical.
In this work a procedure for the choice αk is proposed. This choice is based on the form of the curve xk(t), t≥0 and under the convergence hypothesis of the algorithm.
Preview
Unable to display preview. Download preview PDF.
References
Amaya, J., On the convergence of curvilinear search algorithms in unconstrained optimization, Operations Research Letters. Vol. 4, 1985.
Armijo, L., Minimization of functions having Lipschitz continuous first partial derivatives, Pacific Journal of Mathematics. 16(1–3) 1966.
Botsaris, C.A. and D.H. Jacobson., A. Newton-type curvilinear search method for optimization, Journal of Mathematical Analysis and Applications. 54(217–229) 1976.
Vial, J. Ph. and J. Zang., Unconstrained optimization by approximation of the gradient path. Mathematics of Operations Research. 2(3) 1977.
Zang, I., A new arc algorithm for unconstrained optimization Mathematical Programming. 15(36–52) 1978.
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 1986 Springer-Verlag
About this paper
Cite this paper
Amaya, J. (1986). The armijo step rule adapted to gradient path algorithms. In: Contesse B., L., Correa F., R., Weintraub P., A. (eds) Recent Advances in System Modelling and Optimization. Lecture Notes in Control and Information Sciences, vol 87. Springer, Berlin, Heidelberg. https://doi.org/10.1007/BFb0006774
Download citation
DOI: https://doi.org/10.1007/BFb0006774
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-17083-9
Online ISBN: 978-3-540-47201-8
eBook Packages: Springer Book Archive