scholarly journals An Adaptive Hybrid Algorithm Based on Particle Swarm Optimization and Differential Evolution for Global Optimization

2014 ◽  
Vol 2014 ◽  
pp. 1-16 ◽  
Author(s):  
Xiaobing Yu ◽  
Jie Cao ◽  
Haiyan Shan ◽  
Li Zhu ◽  
Jun Guo

Particle swarm optimization (PSO) and differential evolution (DE) are both efficient and powerful population-based stochastic search techniques for solving optimization problems, which have been widely applied in many scientific and engineering fields. Unfortunately, both of them can easily fly into local optima and lack the ability of jumping out of local optima. A novel adaptive hybrid algorithm based on PSO and DE (HPSO-DE) is formulated by developing a balanced parameter between PSO and DE. Adaptive mutation is carried out on current population when the population clusters around local optima. The HPSO-DE enjoys the advantages of PSO and DE and maintains diversity of the population. Compared with PSO, DE, and their variants, the performance of HPSO-DE is competitive. The balanced parameter sensitivity is discussed in detail.

2021 ◽  
Vol 2070 (1) ◽  
pp. 012136
Author(s):  
B Sravan Kumar ◽  
ShaikHussain Vali ◽  
Vempalle Rafi ◽  
G Nageswara Reddy

Abstract In this paper space reduction particle swarm optimization(SRPSO) is proposed for solving single-objective optimization problems. Minimization of cost is considered as an objective in the economic dispatch problem. The valve point loading effect is incorporated with the cost function which transfigures to the nonlinear problem. To improve the convergence speed, space reduction is essential and parameter variation keeps away the struck of local optima. Particle swarm optimization (PSO) emphasizes global search and is encountered as a stochastic population-based method. The proposed method is validated on a 26 bus system with 6 generators and the performance results are compared with the other existing techniques.


2021 ◽  
Author(s):  
Moritz Mühlenthaler ◽  
Alexander Raß ◽  
Manuel Schmitt ◽  
Rolf Wanka

AbstractMeta-heuristics are powerful tools for solving optimization problems whose structural properties are unknown or cannot be exploited algorithmically. We propose such a meta-heuristic for a large class of optimization problems over discrete domains based on the particle swarm optimization (PSO) paradigm. We provide a comprehensive formal analysis of the performance of this algorithm on certain “easy” reference problems in a black-box setting, namely the sorting problem and the problem OneMax. In our analysis we use a Markov model of the proposed algorithm to obtain upper and lower bounds on its expected optimization time. Our bounds are essentially tight with respect to the Markov model. We show that for a suitable choice of algorithm parameters the expected optimization time is comparable to that of known algorithms and, furthermore, for other parameter regimes, the algorithm behaves less greedy and more explorative, which can be desirable in practice in order to escape local optima. Our analysis provides a precise insight on the tradeoff between optimization time and exploration. To obtain our results we introduce the notion of indistinguishability of states of a Markov chain and provide bounds on the solution of a recurrence equation with non-constant coefficients by integration.


Author(s):  
Shailendra Aote ◽  
Mukesh M. Raghuwanshi

To solve the problems of optimization, various methods are provided in different domain. Evolutionary computing (EC) is one of the methods to solve these problems. Mostly used EC techniques are available like Particle Swarm Optimization (PSO), Genetic Algorithm (GA) and Differential Evolution (DE). These techniques have different working structure but the inner working structure is same. Different names and formulae are given for different task but ultimately all do the same. Here we tried to find out the similarities among these techniques and give the working structure in each step. All the steps are provided with proper example and code written in MATLAB, for better understanding. Here we started our discussion with introduction about optimization and solution to optimization problems by PSO, GA and DE. Finally, we have given brief comparison of these.


2014 ◽  
Vol 1049-1050 ◽  
pp. 1690-1693 ◽  
Author(s):  
Juan Li

The traditional evolutionary algorithm is cannot converge faster to solve the path optimization problems, and the path that is computed is not the shortest path, in allusion to the disadvantage of this algorithm, a mutation particle swarm optimization algorithm is proposed. The algorithm introduces the adaptive mutation strategy, and accelerated the speed to search for the global optimal solution. For seven examples experiment in standard database, the result shows that the algorithm is more efficient..


Author(s):  
Hrvoje Markovic ◽  
◽  
Fangyan Dong ◽  
Kaoru Hirota

A parallel multi-population based metaheuristic optimization framework, called Concurrent Societies, inspired by human intellectual evolution, is proposed. It uses population based metaheuristics to evolve its populations, and fitness function approximations as representations of knowledge. By utilizing iteratively refined approximations it reduces the number of required evaluations and, as a byproduct, it produces models of the fitness function. The proposed framework is implemented as two Concurrent Societies: one based on genetic algorithm and one based on particle swarm optimization both using k -nearest neighbor regression as fitness approximation. The performance is evaluated on 10 standard test problems and compared to other commonly used metaheuristics. Results show that the usage of the framework considerably increases efficiency (by a factor of 7.6 to 977) and effectiveness (absolute error reduced by more than few orders of magnitude). The proposed framework is intended for optimization problems with expensive fitness functions, such as optimization in design and interactive optimization.


2011 ◽  
Vol 383-390 ◽  
pp. 7208-7213
Author(s):  
De Kun Tan

To overcome the shortage of standard Particle Swarm Optimization(SPSO) on premature convergence, Quantum-behaved Particle Swarm Optimization (QPSO) is presented to solve engineering constrained optimization problem. QPSO algorithm is a novel PSO algorithm model in terms of quantum mechanics. The model is based on Delta potential, and we think the particle has the behavior of quanta. Because the particle doesn’t have a certain trajectory, it has more randomicity than the particle which has fixed path in PSO, thus the QPSO more easily escapes from local optima, and has more capability to seek the global optimal solution. In the period of iterative optimization, outside point method is used to deal with those particles that violate the constraints. Furthermore, compared with other intelligent algorithms, the QPSO is verified by two instances of engineering constrained optimization, experimental results indicate that the algorithm performs better in terms of accuracy and robustness.


2012 ◽  
Vol 546-547 ◽  
pp. 8-12
Author(s):  
Li Ai ◽  
Jia Tang Cheng ◽  
Shao Kun Xu

For traditional methods for coal mine gas emission prediction accuracy is not high, an adaptive mutation particle swarm optimization neural network approach is introduced. The algorithm increases the mutation operation in iterative process, and adaptive adjusts mutation probability of the size, in order to enhance the ability to jump out of the local optima. The simulation results show that the method can be better predicted coal mine gas, has a certain practicality.


2014 ◽  
Vol 989-994 ◽  
pp. 2621-2624
Author(s):  
Shao Song Wan ◽  
Jian Cao ◽  
Qun Song Zhu

In order to resolve these problems, we put forward a new design of the intelligent lock which is mainly based on the technology of wireless sensor network. Particle swarm optimization (PSO) is a recently proposed intelligent algorithm which is motivated by swarm intelligence. PSO has been shown to perform well on many benchmark and real-world optimization problems; it easily falls into local optima when solving complex multimodal problems. To avoid the local optimization, the algorithm renews population and enhances the diversity of population by using density calculation of immune theory and adjusting new chaos sequence. The paper gives the circuit diagram of the hardware components based on single chip and describe how to design the software. The experimental results show that the immune genetic algorithm based on chaos theory can search the result of the optimization and evidently improve the convergent speed and astringency.


2021 ◽  
Vol 2021 ◽  
pp. 1-17
Author(s):  
Waqas Haider Bangyal ◽  
Abdul Hameed ◽  
Wael Alosaimi ◽  
Hashem Alyami

Particle swarm optimization (PSO) algorithm is a population-based intelligent stochastic search technique used to search for food with the intrinsic manner of bee swarming. PSO is widely used to solve the diverse problems of optimization. Initialization of population is a critical factor in the PSO algorithm, which considerably influences the diversity and convergence during the process of PSO. Quasirandom sequences are useful for initializing the population to improve the diversity and convergence, rather than applying the random distribution for initialization. The performance of PSO is expanded in this paper to make it appropriate for the optimization problem by introducing a new initialization technique named WELL with the help of low-discrepancy sequence. To solve the optimization problems in large-dimensional search spaces, the proposed solution is termed as WE-PSO. The suggested solution has been verified on fifteen well-known unimodal and multimodal benchmark test problems extensively used in the literature, Moreover, the performance of WE-PSO is compared with the standard PSO and two other initialization approaches Sobol-based PSO (SO-PSO) and Halton-based PSO (H-PSO). The findings indicate that WE-PSO is better than the standard multimodal problem-solving techniques. The results validate the efficacy and effectiveness of our approach. In comparison, the proposed approach is used for artificial neural network (ANN) learning and contrasted to the standard backpropagation algorithm, standard PSO, H-PSO, and SO-PSO, respectively. The results of our technique has a higher accuracy score and outperforms traditional methods. Also, the outcome of our work presents an insight on how the proposed initialization technique has a high effect on the quality of cost function, integration, and diversity aspects.


Sign in / Sign up

Export Citation Format

Share Document