Abstract
In order to solve the premature convergence of BBPSO, this paper proposes a self-learning BBPSO (SLBBPSO) to improve the exploration ability of BBPSO. First, the expectation of Gaussian distribution in the updating equation is controlled by an adaptive factor, which makes particles emphasize on the exploration in earlier stage and the convergence in later stage. Second, SLBBPSO adopts a novel mutation to the personal best position (\(Pbest\)) and the global best position (\(Gbest\)), which helps the algorithm jump out of the local optimum. Finally, when particles are in the stagnant status, the variance of Gaussian distribution is assigned an adaptive value. Simulations show that SLBBPSO has excellent optimization ability in the classical benchmark functions.
This work was supported by National Natural Science Foundation of China under Grant Nos.61300059. Provincial Project of Natural Science Research for Anhui Colleges of China (KJ2012Z031, KJ2012Z024).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Kennedy, J.: Bare bones particle swarms. In: Proceedings of the IEEE Swarm Intelligence Symposium, pp. 80–87. IEEE (2003)
Kennedy, J., Eberhart, R.C.: Particle swarm optimization. In: Proceeding of International Conference on Neural Networks, pp. 1942–1948. IEEE, Perth (1995)
Yong, Z., Dunwei, G., Ying, H., Wanqiu, Z.: Feature Selection Algorithm Based on Bare Bones Particle Swarm Optimization. Neurocomputing 148, 150–157 (2015)
Liyeh, C., Chengsan, Y., Kuochuan, W., Chenghong, Y.: Gene selection and classfication using Taguchi chaotic binary particle swarm optimization. Expert Systems with Applications 38, 13367–13377 (2011)
Krohling, R.A., Mendel, E.: Bare bones particle swarm optimization with Gaussian or Cauchy jumps. In: Proceedings of the IEEE Congress on Evolutionary Computation, pp. 3285–3291. IEEE, Trondheim (2009)
al-Rifaie, M.M., Blackwell, T.: Bare bones particle swarms with jumps. In: Dorigo, M., Birattari, M., Blum, C., Christensen, A.L., Engelbrecht, A.P., Groß, R., Stützle, T. (eds.) ANTS 2012. LNCS, vol. 7461, pp. 49–60. Springer, Heidelberg (2012)
Omran, M.G.H., Andries, P., Salman, A.: Bare bones differential evolution. European Journal of Operational Research 196, 128–139 (2009)
Junqi, Z., Lina, N., Jing, Y., Wei, W., Zheng, T.: Adaptive bare bones particle swarm inspired by cloud model. IEICE Transactions on Information and Systems E94D(8), 1527–1538 (2011)
Yong, Z., Dunwei, G., Xiaoyan, S., Na, G.: Adaptive bare-bones particle swarm optimization algorithm and its convergence analysis. Soft Computing 18, 1337–1352 (2014)
Haibo, Z., Kennedy, D.D., Rangaiah, G.P.: Bonilla-Petriciolet: A novel bare-bones particle swarm optimization and its performancefor modeling vapor-liquid equilibrium data. Fluid Phase Equilib. 301, 33–45 (2011)
Feng, P., Xiaohui, H., Eberhart R.C., Yaobin, C.: An analysis of bare bones particle swarm. In: Proceeding of the 2008 IEEE Swarm Intelligence Symposium, pp. 21–23 (2008)
Lim, W.H., Isa, N.A.M.: Teaching and peer-learning particle swarm optimization. Applied Soft Computing 18, 39–58 (2014)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Chen, J., Shen, Y., Wang, X. (2015). A Self-learning Bare-Bones Particle Swarms Optimization Algorithm. In: Tan, Y., Shi, Y., Buarque, F., Gelbukh, A., Das, S., Engelbrecht, A. (eds) Advances in Swarm and Computational Intelligence. ICSI 2015. Lecture Notes in Computer Science(), vol 9140. Springer, Cham. https://doi.org/10.1007/978-3-319-20466-6_12
Download citation
DOI: https://doi.org/10.1007/978-3-319-20466-6_12
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-20465-9
Online ISBN: 978-3-319-20466-6
eBook Packages: Computer ScienceComputer Science (R0)