Abstract
Estimation of distribution algorithms (EDAs) are a class of evolutionary optimization algorithms based on probability distribution model. This article extends the basic EDAs for tackling multi-objective optimization problems by incorporating multivariate Gaussian copulas for constructing probability distribution model, and using the concept of preference order. In the algorithm, the multivariate Gaussian copula is used to construct probability distribution model in EDAs. By estimating Kendall’s τ and using the relationship of correlation matrix and Kendall’s τ, correlation matrix R in Gaussian copula are firstly estimated from the current population, and then is used to generate offsprings. Preference order is used to identify the best individuals in order to guide the search process. The population with the current population and current offsprings population is sorted based on preference order, and the best individuals are selected to form the next population. The algorithm is tested to compare with NSGA-II, GDE, MOEP and MOPSO based on convergence metric and diversity metric using a set of benchmark functions. The experimental results show that the algorithm is effective on the benchmark functions.
Keywords
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Deb, K., Agrawal, S., Pratap, A., Meyarivan, T.: A fast and elitist multi-objective genetic algorithm: NSGA-II. IEEE Transactions on Evolutionary Computation 6(2), 182–197 (2002)
Zitzler, E., Laumanns, M., Thiele, L.: SPEA2: Improving the strength pareto evolutionary algorithm for multi-objective optimization. In: Proceedings of the EUROGEN 2001 Conference, pp. 95–100 (2001)
Knowles, J.D., Corne, D.W.: The Pareto archived evolution strategy: A new baseline algorithm for Pareto multi-objective optimization. In: Proc. of the 2003 IEEE Conf. on Evolutionary Computation, pp. 98–105 (1999)
Kundu, D., Suresh, K., Ghosh, S., Das, S., Panigrahi, B.K., Das, S.: Multi-objective optimization with artificial weed colonies. Information Sciences 181(12), 2441–2454 (2011)
Kukkonen, S., Lampinen, J.: Performance assessment of generalized differentia evolution with a given set of constrained multi-objective test problems. In: Proceeding of Congress on Evolutionary Computation, CEC 2009, pp. 1943–1950 (2009)
Zhang, Q., Liu, W., Li, H.: The performance of a new version of MOEA/D on CEC 2009 unconstrained mop test instances. In: Proceeding of Congress on Evolutionary Computation, CEC 2009, pp. 203–208 (2009)
Qu, B.Y., Suganthan, P.N.: Multi-objective evolutionary programming without non-domination sorting is up to twenty times faster. In: Proceeding of Congress on Evolutionary Computation, CEC 2009, pp. 2934–2939 (2009)
Coello Coello, C.A., Pulido, G.T., Lechuga, M.S.: Handling multiple objectives with particle swarm optimization. IEEE Transactions on Evolutionary Computation 8(3), 256–279 (2004)
Larranaga, P., Lozano, J.A.: Estimation of Distribution Algorithms: A New Tool for Evolutionary Computation. Kluwer Academic Publishers, Dordrecht (2002)
Pelikan, M., Sastry, K., Goldberg, D.E.: Multi-objective Estimation of Distribution Algorithms. Studies in Computational Intelligence 33, 223–248 (2006)
Sastry, K., Pelikan, M., Goldberg, D.E.: Limits of scalability of multiobjective estimation of distribution algorithms. In: Proceedings of the Congress on Evolutionary Computation, pp. 2217–2224 (2005)
Etxeberria, R., Larrañaga, P.: Global optimization using Bayesian networks. In: Ochoa, A., Soto, M.R., Santana, R. (eds.) Proceedings of the Second Symposium on Artificial Intelligence, CIMAF 1999, Havana, Cuba, pp. 151–173 (1999)
Khan, N.: Bayesian optimization algorithms for multi-objective and hierarchically difficult problems. Master’s thesis, University of Illinois at Urbana-Champaign, Urbana, IL (2003)
Shakya, S., McCall, J.: Optimisation by Estimation of Distribution with DEUM framework based on Markov Random Fields. International Journal of Automation and Computing 4, 262–272 (2007)
Nelsen, R.B.: An Introduction to copula. Springer, New York (1998)
Cherubini, U., Luciano, E., Vecchiato, W.: Copula Methods in Finance. John Wiley & Sons Ltd., England (2004)
Deb, K.: Multi-Objective Optimization Using Evolutionary Algorithms. John Wiley & Sons (2001)
Das, I.: A preference ordering among various Pareto optimal alternatives. Structural and Multidisciplinary Optimization 18(1), 30–35 (1999)
di Pierro, F., Khu, S.T., Savi, D.A.: An investigation on preference order ranking scheme for multiobjective evolutionary optimization. IEEE Transactions on Evolutionary Computation 11(1), 17–45 (2007)
Jaimes, A.L., Quintero, L.V.S., Coello, C.A.C.: Ranking methods in many-objective evolutionary algorithms. In: Chiong, R. (ed.) Nature-Inspired Algorithms for Optimisation. SCI, vol. 193, pp. 413–434. Springer, Heidelberg (2009)
Schaffer, J.D.: Multiple Objective Optimization with Vector Evaluated Genetic Algorithms. In: 1st International Conference on Genetic Algorithms, pp. 93–100 (1985)
Zitzler, E., Deb, K., Thiele, L.: Comparison of multi-objective evolutionary algorithms: Empirical results. Evolutionary Computation 8, 173–195 (2000)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Gao, Y., Peng, L., Li, F., Liu, M., Hu, X. (2013). EDA-Based Multi-objective Optimization Using Preference Order Ranking and Multivariate Gaussian Copula. In: Guo, C., Hou, ZG., Zeng, Z. (eds) Advances in Neural Networks – ISNN 2013. ISNN 2013. Lecture Notes in Computer Science, vol 7952. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-39068-5_42
Download citation
DOI: https://doi.org/10.1007/978-3-642-39068-5_42
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-39067-8
Online ISBN: 978-3-642-39068-5
eBook Packages: Computer ScienceComputer Science (R0)