Abstract
Particle swarm optimization [PSO] is one of the most accepted optimization algorithm and due to its simplicity it has been used in many applications. Although PSO converges very fast, it has stagnation and premature convergence problem. To improve its convergence rate and to remove stagnation problem, some changes in velocity vector are suggested. These changes motivate each particle of PSO in different directions so that full search space can be covered and better solutions can be captured. Moreover, autotuning of random parameters are done to remove stagnation problem and local optima. This auto-improved version is named as AI-PSO algorithm. The performance of the proposed version is compared with various state-of-the-art algorithms such as PSO-TVAC and basic PSO. Results show the superiority of the algorithm.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Kennedy, J., Eberhart, R.: Particle Swarm Optimization. In: IEEE International Conference on Neural Networks 1995, vol. 4, pp. 1942–1948 Piscataway.
Poli, R., Kennedy, J., and Blackwell, T.: Particle Swarm Optimization: An Overview. In: Swarm Intelligence 2007, Springer, New York, pp. 33–57.
Arumugam, S.M., Rao, M.V.C., Chandramohan, A.: A New and Improved Version of Particle Swarm Optimization Algorithm with Global-Local Best Parameters. In: Knowledge Inf. Syst. 2008, vol. 16–3, pp. 331–357.
Arumugam, S. M., Rao, M.V.C., and Tan, A.W.C.: A Novel and Effective Particle Swarm Optimization Like Algorithm with Extrapolation Technique 2009. In: ppl. Soft Comput., vol. 9–1, pp. 308–320.
Riaan, B., Engelbrecht, A.P., and Van den Bergh F.: A niching particle swarm optimizer. In: Proceedings of the 4th Asia-Pacific conference on simulated evolution and learning 2002. Vol. 2. Singapore: Orchid Country Club.
Zhang, Jun, et al.: A novel adaptive sequential niche technique for multimodal function optimization. In: Neurocomputing 69.16 (2006): 2396–2401.
Liang, J.J., Qin, A.K., Suganthan, P.N., and Baskar, S.:Comprehensive learning particle swarm optimizer for global optimization of multimodal functions. In: Proceedings of IEEE Congress on Evolutionary Computation 2006, vol. 10, no. 3, pp. 281–295.
Stefan, B., and Li, X.: Adaptively choosing niching parameters in a PSO. In: Proceedings of the 8th annual conference on Genetic and evolutionary computation 2006, ACM.
Shi, Y., Eberhart, R.C.: A modified particle swarm optimizer. In: Proceedings of IEEE Congress on Evolutionary Computation 1998, pp. 69–73 USA.
Shi, Y., Eberhart, R.C.: Empirical study of particle swarm optimization. In: Proceedings of IEEE Congress Evolutionary Computation 1999, pp. 1945–1950.
Chen, Y.P., Peng, W.C., and. Jian, M.C.: Particle swarm optimization with recombination and dynamic linkage discovery. In: IEEE Trans. Syst. Man Cybern 2007, vol. 37, no. 6, pp. 1460–1470.
Mendes, R., Kennedy, J., and Neves, J.: The fully informed particle swarm: Simpler, maybe better, In: Proceedings of IEEE Congress on Evolutionary Computation 2004, vol. 8, no. 3, pp. 204–210.
Shinn-Ying, H., Lin, H.S., Liauh, W.H., and Ho, S.J.: OPSO: Orthogonal particle swarm optimization and its application to task assignment problems. In: Systems, Man and Cybernetics, Part A: Systems and Humans 2008, IEEE Transactions on vol. 38, no. 2 pp. 288–298.
Zhou, Di, Jun Sun, and WenboXu. “An advanced quantum-behaved particle swarm optimization algorithm utilizing cooperative strategy.” Advanced Computational Intelligence (IWACI), 2010 Third International Workshop on.IEEE, 2010
Pena, J., Upegui, A., and Sanchez, E.: Particle swarm optimization with discrete recombination: an online optimizer for evolvable hardware. In: Proceedings of the 1st NASA/ESA Conference on Adaptive Hardware and Systems (AHS ’06) 2006, pp. 163–170, Istanbul, Turkey.
Angeline, P.J.: Using selection to improve particle swarm optimization. In: Proceedings of IEEE Congress on Evolutionary Computation 1998, pp. 84–89.
Hansen, N., Auger, A., Finck, S., and Ros, R.: Real-parameter black-box optimization benchmarking 2009: Experimental setup. Technical Report RR-6828, INRIA, 2009.
Hansen, N., Auger, A., Finck, S., and Ros, R.: Real-Parameter Black-Box Optimization Benchmarking 2009. In: Noiseless Functions Definitions,” INRIA Technical Report RR-6829, 2009.
Goldberg, D. E.: Genetic Algorithms in Search, Optimization, and Machine Learning. Reading MA 1989, Addison-Wesley.
Storm, R., and Price, K..V.: Minimizing the real functions of the ICEC 1996 contest by differential evolution. In: Proceedings of IEEE Congress Evolutionary Computation, 1996, pp. 842–844.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Kumar, A., Singh, B.K., Patro, B.D.K. (2018). Auto Improved-PSO with Better Convergence and Diversity. In: Bhatia, S., Mishra, K., Tiwari, S., Singh, V. (eds) Advances in Computer and Computational Sciences. Advances in Intelligent Systems and Computing, vol 554. Springer, Singapore. https://doi.org/10.1007/978-981-10-3773-3_5
Download citation
DOI: https://doi.org/10.1007/978-981-10-3773-3_5
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-10-3772-6
Online ISBN: 978-981-10-3773-3
eBook Packages: EngineeringEngineering (R0)