Abstract
Computing gradients is a fundamental step in shape optimization. Gradients are needed in sensitivity analysis to measure the variation of performanced.dziinduced by a small perturbationdziof the i control parameter. They are also used in reliability analysis for computing a first-order random model from known stochastic distributions of the data. Finally, we have seen in the previous chapter that gradients are needed in most algorithms used in mathematical programming, and more generally in most multipoint optimization techniques. If the functions to be differentiated are outputs of computer programmes, these gradients can be computed automatically by differentiating each line of these computer programmes This is called automatic differentiation and is the topic of the present chapter. For complex cost functions used in shape optimization, automatic differentiation will have to be coupled with Lagrangian techniques, and this will be the purpose of the next chapter
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2003 Springer Science+Business Media New York
About this chapter
Cite this chapter
Laporte, E., Le Tallec, P. (2003). Automatic Differentiation. In: Numerical Methods in Sensitivity Analysis and Shape Optimization. Modeling and Simulation in Science, Engineering and Technology. Birkhäuser, Boston, MA. https://doi.org/10.1007/978-1-4612-0069-7_5
Download citation
DOI: https://doi.org/10.1007/978-1-4612-0069-7_5
Publisher Name: Birkhäuser, Boston, MA
Print ISBN: 978-1-4612-6598-6
Online ISBN: 978-1-4612-0069-7
eBook Packages: Springer Book Archive