scholarly journals Z-Shaped Transfer Functions for Binary Particle Swarm Optimization Algorithm

2020 ◽  
Vol 2020 ◽  
pp. 1-21
Author(s):  
Sha-sha Guo ◽  
Jie-sheng Wang ◽  
Meng-wei Guo

Particle swarm optimization (PSO) algorithm is a swarm intelligent searching algorithm based on population that simulates the social behavior of birds, bees, or fish groups. The discrete binary particle swarm optimization (BPSO) algorithm maps the continuous search space to a binary space through a new transfer function, and the update process is designed to switch the position of the particles between 0 and 1 in the binary search space. Aiming at the existed BPSO algorithms which are easy to fall into the local optimum, a new Z-shaped probability transfer function is proposed to map the continuous search space to a binary space. By adopting nine typical benchmark functions, the proposed Z-probability transfer function and the V-shaped and S-shaped transfer functions are used to carry out the performance simulation experiments. The results show that the proposed Z-shaped probability transfer function improves the convergence speed and optimization accuracy of the BPSO algorithm.

2019 ◽  
Vol 18 (03) ◽  
pp. 833-866 ◽  
Author(s):  
Mi Li ◽  
Huan Chen ◽  
Xiaodong Wang ◽  
Ning Zhong ◽  
Shengfu Lu

The particle swarm optimization (PSO) algorithm is simple to implement and converges quickly, but it easily falls into a local optimum; on the one hand, it lacks the ability to balance global exploration and local exploitation of the population, and on the other hand, the population lacks diversity. To solve these problems, this paper proposes an improved adaptive inertia weight particle swarm optimization (AIWPSO) algorithm. The AIWPSO algorithm includes two strategies: (1) An inertia weight adjustment method based on the optimal fitness value of individual particles is proposed, so that different particles have different inertia weights. This method increases the diversity of inertia weights and is conducive to balancing the capabilities of global exploration and local exploitation. (2) A mutation threshold is used to determine which particles need to be mutated. This method compensates for the inaccuracy of random mutation, effectively increasing the diversity of the population. To evaluate the performance of the proposed AIWPSO algorithm, benchmark functions are used for testing. The results show that AIWPSO achieves satisfactory results compared with those of other PSO algorithms. This outcome shows that the AIWPSO algorithm is conducive to balancing the abilities of the global exploration and local exploitation of the population, while increasing the diversity of the population, thereby significantly improving the optimization ability of the PSO algorithm.


2019 ◽  
Vol 2019 ◽  
pp. 1-22 ◽  
Author(s):  
Hao Li ◽  
Hongbin Jin ◽  
Hanzhong Wang ◽  
Yanyan Ma

For the first time , the Holonic Particle Swarm Optimization (HPSO ) algorithm applies multiagent theory about the improvement in the PSO algorithm and achieved good results. In order to further improve the performance of the algorithm, this paper proposes an improved Adaptive Holonic Particle Swarm Optimization (AHPSO) algorithm. Firstly, a brief review of the HPSO algorithm is carried out, and the HPSO algorithm can be further studied in three aspects: grouping strategy, iteration number setting, and state switching discrimination. The HPSO algorithm uses an approximately uniform grouping strategy that is the simplest but does not consider the connections between particles. And if the particles with larger or smaller differences are grouped together in different search stages, the search efficiency will be improved. Therefore, this paper proposes a grouping strategy based on information entropy and system clustering and combines two grouping strategies with corresponding search methods. The performance of the HPSO algorithm depends on the setting of the number of iterations. If it is too small, it is difficult to search for the optimal and it wastes so many computing resources. Therefore, this paper constructs an adaptive termination condition that causes the particles to terminate spontaneously after convergence. The HPSO algorithm only performs a conversion from extensive search to exact search and still has the potential to fall into local optimum. This paper proposes a state switching condition to improve the probability that the algorithm jumps out of the local optimum. Finally, AHPSO and HPSO are compared by using 22 groups of standard test functions. AHPSO is faster in convergence than HPSO, and the number of iterations of AHPSO convergence is employed in HPSO. At this point, there exists a large gap between HPSO and the optimal solution, i.e., AHPSO can have better algorithm efficiency without setting the number of iterations.


2013 ◽  
Vol 2013 ◽  
pp. 1-12 ◽  
Author(s):  
Martins Akugbe Arasomwan ◽  
Aderemi Oluyinka Adewumi

Linear decreasing inertia weight (LDIW) strategy was introduced to improve on the performance of the original particle swarm optimization (PSO). However, linear decreasing inertia weight PSO (LDIW-PSO) algorithm is known to have the shortcoming of premature convergence in solving complex (multipeak) optimization problems due to lack of enough momentum for particles to do exploitation as the algorithm approaches its terminal point. Researchers have tried to address this shortcoming by modifying LDIW-PSO or proposing new PSO variants. Some of these variants have been claimed to outperform LDIW-PSO. The major goal of this paper is to experimentally establish the fact that LDIW-PSO is very much efficient if its parameters are properly set. First, an experiment was conducted to acquire a percentage value of the search space limits to compute the particle velocity limits in LDIW-PSO based on commonly used benchmark global optimization problems. Second, using the experimentally obtained values, five well-known benchmark optimization problems were used to show the outstanding performance of LDIW-PSO over some of its competitors which have in the past claimed superiority over it. Two other recent PSO variants with different inertia weight strategies were also compared with LDIW-PSO with the latter outperforming both in the simulation experiments conducted.


2010 ◽  
Vol 20-23 ◽  
pp. 1280-1285
Author(s):  
Jian Xiang Wei ◽  
Yue Hong Sun

The particle swarm optimization (PSO) algorithm is a new population search strategy, which has exhibited good performance through well-known numerical test problems. However, it is easy to trap into local optimum because the population diversity becomes worse during the evolution. In order to overcome the shortcoming of the PSO, this paper proposes an improved PSO based on the symmetry distribution of the particle space position. From the research of particle movement in high dimensional space, we can see: the more symmetric of the particle distribution, the bigger probability can the algorithm be during converging to the global optimization solution. A novel population diversity function is put forward and an adjustment algorithm is put into the basic PSO. The steps of the proposed algorithm are given in detail. With two typical benchmark functions, the experimental results show the improved PSO has better convergence precision than the basic PSO.


2017 ◽  
Vol 29 (1) ◽  
pp. 127-142
Author(s):  
Rkia Fajr ◽  
Abdelaziz Bouroumi

Abstract This paper introduces a new variant of the particle swarm optimization (PSO) algorithm, designed for global optimization of multidimensional functions. The goal of this variant, called ImPSO, is to improve the exploration and exploitation abilities of the algorithm by introducing a new operation in the iterative search process. The use of this operation is governed by a stochastic rule that ensures either the exploration of new regions of the search space or the exploitation of good intermediate solutions. The proposed method is inspired by collaborative human learning and uses as a starting point a basic PSO variant with constriction factor and velocity clamping. Simulation results that show the ability of ImPSO to locate the global optima of multidimensional functions are presented for 10 well-know benchmark functions from CEC-2013 and CEC-2005. These results are compared with the PSO variant used as starting point, three other PSO variants, one of which is based on human learning strategies, and three alternative evolutionary computing methods.


Author(s):  
Cheng-Hung Chen ◽  
Marco P. Schoen ◽  
Ken W. Bosworth

A novel Condensed Hybrid Optimization (CHO) algorithm using Enhanced Continuous Tabu Search (ECTS) and Particle Swarm Optimization (PSO) is proposed. The proposed CHO algorithm combines the respective strengths of ECTS and PSO. The ECTS is a modified Tabu Search (TS), which has good search capabilities on large search spaces. In this study, ECTS is utilized to define smaller search spaces, which are used in a second stage by the basic PSO to find the respective local optimum. The ECTS covers the global search space by using a TS concept called diversification and then selects the most promising areas in the search space. Once the promising regions in the search space are defined, the proposed CHO algorithm employs another TS concept called intensification in order to search the promising area thoroughly. The proposed CHO algorithm is tested with the multi-dimensional Hyperbolic and Rosenbrock problems. Compared to other four algorithms, the simulations results indicate that the accuracy and effectiveness of the proposed CHO algorithm.


2013 ◽  
Vol 394 ◽  
pp. 505-508 ◽  
Author(s):  
Guan Yu Zhang ◽  
Xiao Ming Wang ◽  
Rui Guo ◽  
Guo Qiang Wang

This paper presents an improved particle swarm optimization (PSO) algorithm based on genetic algorithm (GA) and Tabu algorithm. The improved PSO algorithm adds the characteristics of genetic, mutation, and tabu search into the standard PSO to help it overcome the weaknesses of falling into the local optimum and avoids the repeat of the optimum path. By contrasting the improved and standard PSO algorithms through testing classic functions, the improved PSO is found to have better global search characteristics.


2018 ◽  
Author(s):  
Boris Almonacid

The optimal selection of a natural reserve (OSRN) is an optimisation problem with a binary domain. To solve this problem the metaheuristic algorithm Particle Swarm Optimization (PSO) has been chosen. The PSO algorithm has been designed to solve problems in real domains. Therefore, a transfer method has been applied that converts the equations with real domains of the PSO algorithm into binary results that are compatible with the OSRN problem. Four transfer functions have been tested in four case studies to solve the OSRN problem. According to the tests carried out, it is concluded that two of the four transfer functions are apt to solve the problem of optimal selection of a natural reserve.


Sign in / Sign up

Export Citation Format

Share Document