Abstract
We consider the problem of minimizing a function f subject to a single inequality constraint \(g(\mathbf x ) \le 0\), in a black-box scenario. We present a covariance matrix adaptation evolution strategy using an adaptive augmented Lagrangian method to handle the constraint. We show that our algorithm is an instance of a general framework that allows to build an adaptive constraint handling algorithm from a general randomized adaptive algorithm for unconstrained optimization. We assess the performance of our algorithm on a set of linearly constrained functions, including convex quadratic and ill-conditioned functions, and observe linear convergence to the optimum.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Python source code of CMA-ES. https://pypi.python.org/pypi/cma
Ait Elhara, O., Auger, A., Hansen, N.: A median success rule for non-elitist evolution strategies: study of feasibility. In: Genetic and Evolutionary Computation Conference, pp. 415–422. ACM Press (2013)
Arnold, D.V., Porter, J.: Towards au augmented lagrangian constraint handling approach for the \((1+1)\)-ES. In: Genetic and Evolutionary Computation Conference, pp. 249–256. ACM Press (2015)
Atamna, A., Auger, A., Hansen, N.: Analysis of linear convergence of a \((1+1)\)-ES with augmented lagrangian constraint handling. To appear in the Proceedings of the Genetic and Evolutionary Computation Conference (2016)
Auger, A., Hansen, N., Perez Zerpa, J.M., Ros, R., Schoenauer, M.: Experimental comparisons of derivative free optimization algorithms. In: Vahrenhold, J. (ed.) SEA 2009. LNCS, vol. 5526, pp. 3–15. Springer, Heidelberg (2009)
Hansen, N.: The CMA Evolution Strategy: A Tutorial (2016). http://arxiv.org/pdf/1604.00772v1.pdf
Hansen, N., Ostermeier, A.: Completely derandomized self-adaptation in evolution strategies. Evol. Comput. 9(2), 159–195 (2001)
Hansen, N., Ros, R., Mauny, N., Schoenauer, M., Auger, A.: Multiplier and gradient methods. Appl. Soft Comput. 11, 5755–5769 (2011)
Hestenes, M.R.: Multiplier and gradient methods. J. Optim. Theory Appl. 4(5), 303–320 (1969)
Kern, S., Müller, S.D., Hansen, N., Büche, D., Ocenasek, J., Koumoutsakos, P.: Learning probability distributions in continuous evolutionary algorithms - a comparative review. Nat. Comput. 3(1), 77–112 (2004)
Nocedal, J., Wright, S.J.: Numerical Optimization, 2nd edn. Springer, New York (2006)
Powell, M.J.D.: A method for nonlinear constraints in minimization problems. In: Fletcher, R. (ed.) Optimization, pp. 283–298. Academic Press, New York (1969)
Acknowledgments
This work was supported by the grant ANR-2012-MONU-0009 (NumBBO) of the French National Research Agency.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing AG
About this paper
Cite this paper
Atamna, A., Auger, A., Hansen, N. (2016). Augmented Lagrangian Constraint Handling for CMA-ES — Case of a Single Linear Constraint. In: Handl, J., Hart, E., Lewis, P., López-Ibáñez, M., Ochoa, G., Paechter, B. (eds) Parallel Problem Solving from Nature – PPSN XIV. PPSN 2016. Lecture Notes in Computer Science(), vol 9921. Springer, Cham. https://doi.org/10.1007/978-3-319-45823-6_17
Download citation
DOI: https://doi.org/10.1007/978-3-319-45823-6_17
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-45822-9
Online ISBN: 978-3-319-45823-6
eBook Packages: Computer ScienceComputer Science (R0)