Abstract
Multilayer perceptron (MLP) is the most popular neural network method and it has been widely used for many practical applications. In this paper, recently developed interior search algorithm (ISA) is proposed for training MLP. Five of most important standard classification datasets (balloon, XOR, Iris, heart, and breast cancer) are employed to evaluate the proposed algorithm performance. The obtained results from ISA-based are compared with five well-known algorithms including ant colony optimization (ACO), genetic algorithm (GA), particle swarm optimization (PSO), population-based incremental learning (PBIL), and evolution strategy (ES). The statistical results reflect that the performance of the proposed algorithm can train MLPs with a very high degree of accuracy and it is capable of outperforming the well-known algorithms. The results also show that the high convergence rate of the ISA and it is potential to avoid local minima.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Seyedali Mirjalili, “How effective is the Grey Wolf optimizer in training multi-layer perceptrons”, Springer-january-2015.
Reed RD, Marks RJ, “Neural smithing: supervised learning in feedforward artificial neural networks”, Mit Press-1998.
Gandomi A.H. “Interior Search Algorithm (ISA): A Novel Approach for Global Optimization” ISA Transaction, Elsevier, 53(4), 1168–1183, 2014.
Mirjalili S, Sadiq AS (2011), “Magnetic optimization algorithm for training multi-layer perceptron”, In: Communication Software and Networks (ICCSN), 2011 IEEE 3rd International Conference, IEEE, pp 42–46.
Moallem P, Razmjooy N, “A multi layer perceptron neural network trained by invasive weed optimization for potato color image segmentation”, Trends Appl Sci Res-2012-7:445–455.
Bebis G, Georgiopoulos M, “Feed-forward neural networks”, Potentials, IEEE-1994.
Mirjalili S, Mirjalili SM, Lewis A, “Let a biogeography based optimizer train your multi-layer perceptron”, Inf Sci-2014..
Mendes R, Cortez P, Rocha M, Neves J, “Particle swarms for feedforward neural network training, learning”, 2002- vol. 6.
Whitley D, Starkweather T, Bogart C, “Genetic algorithms and neural networks: Optimizing connections and connectivity”, Parallel comput-1990, 14:347–361.
Blum C, Socha K, “Training feed-forward neural networks with ant colony optimization: an application to pattern classification,” In: 5th international conference on, Hybrid Intelligent Systems, 2005. HIS’05, p 6.
Yao X, Liu Y, “Fast evolution strategies”, In: evolutionary programming VI-1997, pp 149–161.
Baluja S (1994), “Population-based incremental learning. A method for integrating genetic search based function optimization and competitive learning”, DTIC Document-1997.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Bhesdadiya, R.H., Trivedi, I.N., Jangir, P., Kumar, A., Jangir, N., Totlani, R. (2018). Training Multilayer Perceptrons in Neural Network Using Interior Search Algorithm. In: Bhatia, S., Mishra, K., Tiwari, S., Singh, V. (eds) Advances in Computer and Computational Sciences. Advances in Intelligent Systems and Computing, vol 554. Springer, Singapore. https://doi.org/10.1007/978-981-10-3773-3_8
Download citation
DOI: https://doi.org/10.1007/978-981-10-3773-3_8
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-10-3772-6
Online ISBN: 978-981-10-3773-3
eBook Packages: EngineeringEngineering (R0)