Abstract
Ant Colony Optimization (ACO) is the most familiar meta-heuristic search algorithm and has been used in optimization of a number of feature selection (FS) problems. As a meta-heuristic search algorithm, ACO requires a set of parameters to solve the problem of feature selection. Pheromone Evaporation Rate (PER) is the most important among all these parameters. Setting up the values of these parameters is a big deal. Usually, these parameters are set up by experimenting through a number of values and finally selecting a set of values that seems to work well for feature selection. The change in optimal feature selection in accordance to different values of PER and other ACO parameters is discussed in this paper. ACO is applied for feature selection and classification of 10 datasets. From the experimental results, it can be seen that, the optimal value for the evaporation rate (ρ) lies around 0.7 leading to selection of best features and increase in classification accuracy.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Dorigo, M., Stutzle, T.: Ant Colony Optimization. The MIT Press, Massachusetts (2004)
Rezaee, M.R., Goedhart, B., Lelieveldt, B.P.F., Reiber, J.H.C.: Fuzzy feature selection. Pattern Recognition 32, 2011–2019 (1999)
Laura Santana, E.A., Ligia Silva Anne Canuto, M.P., Pintro, F., Vale, K.O.: A Comparative Analysis of Genetic Algorithm and Ant Colony Optimization to Select Attributes for a Heterogeneous Ensemble of Classifiers, pp. 465–472. IEEE (2010)
Ahmed, E.F., Yang, W.J.M., Abdullah, M.Y.: Novel method of the combination of forecasts based on rough sets. Journal of Computer Science 5, 440–444 (2009)
Sadeghzadeh, M., Teshnehlab, M.: Correlation-based Feature Selection using Ant Colony Optimization. World Academy of Science, Engineering and Technology 64, 497–502 (2010)
Dorigo, M., Di, C.G., Gambardella, L.M.: Ant algorithms for discrete optimization. Artificial Life 5, 137–172 (1999)
Abd-Alsabour, N., Randall, M.: Feature Selection for Classification Using an Ant Colony System. In: Sixth IEEE International Conference on e–Science Workshops, pp. 86–91 (2010); Kuncheva, L.I.: Combining Pattern Classifiers, Methods and Algorithms. Wiley Interscience (2005)
Duda, R.O., Hart, P.E., Stork, D.G.: Pattern Recognition, 2nd edn. John Wiley & Sons, Inc. (2001)
Molina, L.C., Belanche, L., Nebot, À.: Feature Selection Algorithms: A Survey and Experimental Evaluation. In: Second IEEE International Conference on Data Mining, pp. 155–172 (2002)
Randall, M.: Near Parameter Free Ant Colony Optimisation. In: Dorigo, M., Birattari, M., Blum, C., Gambardella, L.M., Mondada, F., Stützle, T. (eds.) ANTS 2004. LNCS, vol. 3172, pp. 374–381. Springer, Heidelberg (2004)
Frank, A., Asuncion, A.: UCI Machine Learning Repository. University of California, School of Information and Computer Science, Irvine, CA (2010), http://archive.ics.uci.edu/ml
WEKA: A Java Machine Learning Package, http://www.cs.waikato.ac.nz/~ml/weka/
Ridge, E., Kudenko, D.: Screening the Parameters Affecting Heuristic Performance. In: Preceeding of Genetic and Evolutionary Computation, GECCO 2007, p. 180. ACM (2007)
Matthews, D.C.: Improved Lower Limits for Pheromone Trails in Ant Colony Optimization. In: Rudolph, G., Jansen, T., Lucas, S., Poloni, C., Beume, N. (eds.) PPSN 2008. LNCS, vol. 5199, pp. 508–517. Springer, Heidelberg (2008)
Stutzle, T., López-Ibáñez, M., Pellegrini, P., Maur, M., de Oca, M.M., Birattari, M., Dorigo, M.: Parameter Adaptation in Ant Colony Optimization, IRIDIA Technical Report Series Technical Report No. TR/IRIDIA/2010-002 (2010)
Kumar, P.: A Note on the Parameter of Evaporation in the Ant Colony Optimization Algorithm. International Mathematical Forum 6(34), 1655–1659 (2011)
Ivković, N.: Investigating MAX-MIN_ Ant System Parameter Space
Dobslaw, F.: A Parameter-Tuning Framework for Metaheuristics Based on Design of Experiments and Artificial Neural Networks. World Academy of Science, Engineering and Technology 64, 213–216 (2010)
Pellegrini, P., Favaretto, D., Moretti, E.: On MAX-MIN Ant System’s parameters
Sivagaminathan, R.K., Ramakrishnan, S.: A hybrid approach for feature subset selection using neural networks and ant colony optimization. Expert Systems with Applications 33, 49–60 (2007)
Aghdam, M.H., Ghasem-Aghaee, N., Basiri, M.E.: Text feature selection using ant colony optimization. Expert Systems with Applications 36, 6843–6853 (2009)
Al-Ani, A.: Feature Subset Selection Using Ant Colony Optimization. International Journal of Computational Intelligence 2(1), 53–58 (2005)
Al-Ani, A.: Ant Colony Optimization for Feature Subset Selection. World Academy of Science, Engineering and Technology 4, 35–38 (2005)
He, Y., Chen, D., Zhao, W.: Ensemble classifier system based on ant colony algorithm and its application in chemical pattern classification. Chemo Metrics and Intelligent Laboratory Systems, 39–49 (2006)
Robbins, K., Zhang, W., Bertrand, J.: The ant colony algorithm for feature selection in high-dimension gene expression data for disease classification. Mathematical Medicine and Biology, 413–426 (2007)
Kanan, H., Faez, K.: An improved feature selection method based on ant colony optimization (ACO) evaluated on face recognition system. Applied Mathematics and Computation, 716–725 (2008)
Chandra, A., Yao, X.: Ensemble learning using multi-objective evolutionary algorithm. Journal of Mathematical Modeling and Algorithms 5(4), 417–445 (2006)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2012 ICST Institute for Computer Science, Social Informatics and Telecommunications Engineering
About this paper
Cite this paper
P., S., S., K., S., D., J., A., J., P. (2012). Investigation on the Effects of ACO Parameters for Feature Selection and Classification. In: Das, V.V., Stephen, J. (eds) Advances in Communication, Network, and Computing. CNC 2012. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering, vol 108. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-35615-5_20
Download citation
DOI: https://doi.org/10.1007/978-3-642-35615-5_20
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-35614-8
Online ISBN: 978-3-642-35615-5
eBook Packages: Computer ScienceComputer Science (R0)