Abstract
Neural network ensemble (NNE) receives increasing attention in recent research among the e-commerce community. In an NNE method, multiple component neural networks are trained and cooperate with each other to solve the same problem. This paper presents a dynamic ensemble construction approach based on a coevolution paradigm. The whole model is obtained by a specially designed cooperative coevolutionary algorithm. After the coevolution process, a further heuristic structure refining process on the ensemble size is conducted in order to find the appropriate ensemble size for different datasets. The dynamic ensemble size value is obtained by removing less-contribution component networks. Experimental results illustrate that the classification performance of the proposed algorithm is superior to the traditional ensemble methods on real-world datasets.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Yang, H., King, I.: Ensemble Learning for Imbalanced E-commerce Transaction Anomaly Classification. In: Leung, C.S., Lee, M., Chan, J.H. (eds.) ICONIP 2009, Part I. LNCS, vol. 5863, pp. 866–874. Springer, Heidelberg (2009)
Wu, C., Xia, H.: Study of personal credit evaluation under C2C environment based on support vector machines ensemble. In: Proceedings on the 15th Annual Conference on Management Science and Engineering, September 10-12, pp. 25–31. IEEE Computing Society, Los Alamitos (2008)
Krogh, J.V.: Neural Network Ensembles, Cross Validation, and Active Learning. Advances in Neural Information Processing Systems 7, 231–238 (1995)
Zhou, Z.H., Jiang, Y.: NeC4.5: Neural Ensemble Based C4.5. IEEE Transactions on Knowledge and Data Engineering 16(6), 770–773 (2004)
GarcÃa-Pedrajas, N., Fyfe, C.: Immune Network Based Ensembles. Neurocomputing 70(7-9), 1155–1166 (2007)
Rosen, B.E.: Ensemble Learning Using Decorrelated Neural Networks. Connections Science 8(3), 373–384 (1996)
Tian, J., Li, M.Q., Chen, F.Z.: Coevolutionary learning of neural network ensemble for complex classification tasks. Pattern Recognition 45, 1373–1385 (2012)
Zhao, Q.F., Higuchi, T.: Evolutionary Learning of Nearest Neighbor MLP. IEEE Transactions on Neural Networks 7, 762–767 (1996)
Zhou, Z.H., Wu, J., Tang, W.: Ensembling Neural Networks: Many Could Be Better Than All. Artificial Intelligence 137(1-2), 239–263 (2002)
Perrone, M.P., Cooper, L.N.: When Networks Disagree: Ensemble Method for Neural Networks. Artificial Neural Networks for Speech and Vision, 126–142 (1993)
Casasent, D., Chen, X.W.: Radial Basis Function Neural Networks for Nonlinear Fisher Discrimination and Neyman-Pearson Classification. Neural Networks 16, 529–535 (2003)
Kuncheva, L.I., Whitaker, C.J.: Measures of Diversity in Classifier Ensembles and Their Relationship with the Ensemble Accuracy. Machine Learning 51(2), 181–207 (2003)
Ficici, S.G., Pollack, J.B.: Pareto Optimality in Coevolutionary Learning. In: Kelemen, J., Sosik, P. (eds.) ECAL 2001. LNCS (LNAI), vol. 2159, pp. 316–325. Springer, Heidelberg (2001)
Li, M.Q., Kou, J.S., et al.: The Basic Theories and Applications in GA. Science Press, Beijing (2002)
Opitz, D., Maclin, R.: Popular Ensemble Methods: An Empirical Study. Journal of Artificial Intelligence Research 11, 169–198 (1999)
Chandra, A., Yao, X.: Ensemble Learning Using Multi-Objective Evolutionary Algorithms. Journal of Mathematical Modelling and Algorithms 5(4), 417–425 (2006)
Freund, Y., Schapire, R.: Experiments with a New Boosting Algorithm. In: Thirteenth International Conference on Machine Learning, pp. 148–156 (1996)
Breiman, L.: Bagging Predictors. Machine Learning 24(2), 123–140 (1996)
Ting, K.M., Witten, I.H.: Stacking Bagged and Dagged Models. In: Fourteenth International Conference on Machine Learning, pp. 367–375 (1997)
Webb, G.I.: Multiboosting: a Technique for Combining Boosting and Wagging. Machine Learning 40, 159–196 (2000)
Domingos, P.: A Unified Bias-Variance Decomposition and Its Applications. In: International Conference on Machine Learning, pp. 231–238 (2000)
Kohavi, R., Wolpert, D.H.: Bias Plus Variance Decomposition for Zero-One Loss Functions. In: International Conference on Machine Learning, pp. 275–283 (1996)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2012 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Tian, J., Li, M., Chen, F. (2012). Dynamic Neural Network Ensemble Construction for Classification. In: Khachidze, V., Wang, T., Siddiqui, S., Liu, V., Cappuccio, S., Lim, A. (eds) Contemporary Research on E-business Technology and Strategy. iCETS 2012. Communications in Computer and Information Science, vol 332. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-34447-3_19
Download citation
DOI: https://doi.org/10.1007/978-3-642-34447-3_19
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-34446-6
Online ISBN: 978-3-642-34447-3
eBook Packages: Computer ScienceComputer Science (R0)