scholarly journals Gaussian Perturbation Specular Reflection Learning and Golden-Sine-Mechanism-Based Elephant Herding Optimization for Global Optimization Problems

2021 ◽  
Vol 2021 ◽  
pp. 1-25
Author(s):  
Yuxian Duan ◽  
Changyun Liu ◽  
Song Li ◽  
Xiangke Guo ◽  
Chunlin Yang

Elephant herding optimization (EHO) has received widespread attention due to its few control parameters and simple operation but still suffers from slow convergence and low solution accuracy. In this paper, an improved algorithm to solve the above shortcomings, called Gaussian perturbation specular reflection learning and golden-sine-mechanism-based EHO (SRGS-EHO), is proposed. First, specular reflection learning is introduced into the algorithm to enhance the diversity and ergodicity of the initial population and improve the convergence speed. Meanwhile, Gaussian perturbation is used to further increase the diversity of the initial population. Second, the golden sine mechanism is introduced to improve the way of updating the position of the patriarch in each clan, which can make the best-positioned individual in each generation move toward the global optimum and enhance the global exploration and local exploitation ability of the algorithm. To evaluate the effectiveness of the proposed algorithm, tests are performed on 23 benchmark functions. In addition, Wilcoxon rank-sum tests and Friedman tests with 5% are invoked to compare it with other eight metaheuristic algorithms. In addition, sensitivity analysis to parameters and experiments of the different modifications are set up. To further validate the effectiveness of the enhanced algorithm, SRGS-EHO is also applied to solve two classic engineering problems with a constrained search space (pressure-vessel design problem and tension-/compression-string design problem). The results show that the algorithm can be applied to solve the problems encountered in real production.

2011 ◽  
Vol 21 (09) ◽  
pp. 2597-2622 ◽  
Author(s):  
FANGZHEN GE ◽  
ZHEN WEI ◽  
YANG LU ◽  
LIXIANG LI ◽  
YIXIAN YANG

Chaotic Ant Swarm (CAS) is an optimization algorithm based on swarm intelligence theory, which has been applied to find the global optimum solution in search space. However, it often loses its effectiveness and advantages when applied to large and complex problems, e.g. those with high dimensions. To resolve the problems of high computational complexity and low solution accuracy existing in CAS, we propose a Disturbance Chaotic Ant Swarm (DCAS) algorithm to significantly improve the performance of the original algorithm. The aim of this paper is achieved by three strategies which include modifying the method of updating ant's best position, neighbor selection method and establishing a self-adaptive disturbance strategy. The global convergence of the DCAS algorithm is proved in this paper. Extensive computational simulations and comparisons are carried out to validate the performance of the DCAS on two sets of benchmark functions with up to 1000 dimensions. The results show clearly that DCAS substantially enhances the performance of the CAS paradigm in terms of computational complexity, global optimality, solution accuracy and algorithm reliability for complex high-dimensional optimization problems.


Complexity ◽  
2020 ◽  
Vol 2020 ◽  
pp. 1-18
Author(s):  
Feng Qian ◽  
Mohammad Reza Mahmoudi ◽  
Hamïd Parvïn ◽  
Kim-Hung Pho ◽  
Bui Anh Tuan

Conventional optimization methods are not efficient enough to solve many of the naturally complicated optimization problems. Thus, inspired by nature, metaheuristic algorithms can be utilized as a new kind of problem solvers in solution to these types of optimization problems. In this paper, an optimization algorithm is proposed which is capable of finding the expected quality of different locations and also tuning its exploration-exploitation dilemma to the location of an individual. A novel particle swarm optimization algorithm is presented which implements the conditioning learning behavior so that the particles are led to perform a natural conditioning behavior on an unconditioned motive. In the problem space, particles are classified into several categories so that if a particle lies within a low diversity category, it would have a tendency to move towards its best personal experience. But, if the particle’s category is with high diversity, it would have the tendency to move towards the global optimum of that category. The idea of the birds’ sensitivity to its flying space is also utilized to increase the particles’ speed in undesired spaces in order to leave those spaces as soon as possible. However, in desirable spaces, the particles’ velocity is reduced to provide a situation in which the particles have more time to explore their environment. In the proposed algorithm, the birds’ instinctive behavior is implemented to construct an initial population randomly or chaotically. Experiments provided to compare the proposed algorithm with the state-of-the-art methods show that our optimization algorithm is one of the most efficient and appropriate ones to solve the static optimization problems.


2019 ◽  
Vol 9 (22) ◽  
pp. 4736 ◽  
Author(s):  
Ricardo Soto ◽  
Broderick Crawford ◽  
Jose M. Lanza-Gutierrez ◽  
Rodrigo Olivares ◽  
Pablo Camacho ◽  
...  

Metaheuristics are multi-purpose problem solvers devoted to particularly tackle large instances of complex optimization problems. However, in spite of the relevance of metaheuristics in the optimization world, their proper design and implementation to reach optimal solutions is not a simple task. Metaheuristics require an initial parameter configuration, which is dramatically relevant for the efficient exploration and exploitation of the search space, and therefore to the effective finding of high-quality solutions. In this paper, the authors propose a variation of the water cycle inspired metaheuristic capable of automatically adjusting its parameter by using the autonomous search paradigm. The goal of our proposal is to explore and to exploit promising regions of the search space to rapidly converge to optimal solutions. To validate the proposal, we tested 160 instances of the manufacturing cell design problem, which is a relevant problem for the industry, whose objective is to minimize the number of movements and exchanges of parts between organizational elements called cells. As a result of the experimental analysis, the authors checked that the proposal performs similarly to the default approach, but without being specifically configured for solving the problem.


2020 ◽  
Vol 143 (2) ◽  
Author(s):  
Kamrul Hasan Rahi ◽  
Hemant Kumar Singh ◽  
Tapabrata Ray

Abstract Real-world design optimization problems commonly entail constraints that must be satisfied for the design to be viable. Mathematically, the constraints divide the search space into feasible (where all constraints are satisfied) and infeasible (where at least one constraint is violated) regions. The presence of multiple constraints, constricted and/or disconnected feasible regions, non-linearity and multi-modality of the underlying functions could significantly slow down the convergence of evolutionary algorithms (EA). Since each design evaluation incurs some time/computational cost, it is of significant interest to improve the rate of convergence to obtain competitive solutions with relatively fewer design evaluations. In this study, we propose to accomplish this using two mechanisms: (a) more intensified search by identifying promising regions through “bump-hunting,” and (b) use of infeasibility-driven ranking to exploit the fact that optimal solutions are likely to be located on constraint boundaries. Numerical experiments are conducted on a range of mathematical benchmarks and empirically formulated engineering problems, as well as a simulation-based wind turbine design optimization problem. The proposed approach shows up to 53.48% improvement in median objective values and up to 69.23% reduction in cost of identifying a feasible solution compared with a baseline EA.


2012 ◽  
Vol 2012 ◽  
pp. 1-7 ◽  
Author(s):  
Alireza Rowhanimanesh ◽  
Sohrab Efati

Evolutionary methods are well-known techniques for solving nonlinear constrained optimization problems. Due to the exploration power of evolution-based optimizers, population usually converges to a region around global optimum after several generations. Although this convergence can be efficiently used to reduce search space, in most of the existing optimization methods, search is still continued over original space and considerable time is wasted for searching ineffective regions. This paper proposes a simple and general approach based on search space reduction to improve the exploitation power of the existing evolutionary methods without adding any significant computational complexity. After a number of generations when enough exploration is performed, search space is reduced to a small subspace around the best individual, and then search is continued over this reduced space. If the space reduction parameters (red_gen and red_factor) are adjusted properly, reduced space will include global optimum. The proposed scheme can help the existing evolutionary methods to find better near-optimal solutions in a shorter time. To demonstrate the power of the new approach, it is applied to a set of benchmark constrained optimization problems and the results are compared with a previous work in the literature.


2021 ◽  
Author(s):  
Radhwan A.A. Saleh ◽  
Rüştü Akay

Abstract As a relatively new model, the Artificial Bee Colony Algorithm (ABC) has shown impressive success in solving optimization problems. Nevertheless, its efficiency is still not satisfactory for some complex optimization problems. This paper has modified ABC and its other recent variants to improve its performance by modify the scout phase. This modification enhances its exploitation ability by intensifying the regions in the search space, which probably includes reasonable solutions. The experiments were performed on the CEC2014 benchmark suite, CEC2015 benchmark functions, and three real-life problems: pressure vessel design problem, tension and compression spring design problem, and Frequency-Modulated (FM) problem. And the proposed modification was applied to basic ABC, Gbest-Guided ABC, Depth First Search ABC, and Teaching Learning Based ABC, and they were compared with their modified counterparts. The results have shown that our modification can successfully increase the performance of the original versions. Moreover, the proposed modified algorithm was compared with the state-of-the-art optimization algorithms, and it produced competitive results.


Author(s):  
Liqun Wang ◽  
Songqing Shan ◽  
G. Gary Wang

The presence of black-box functions in engineering design, which are usually computation-intensive, demands efficient global optimization methods. This work proposes a new global optimization method for black-box functions. The global optimization method is based on a novel mode-pursuing sampling (MPS) method which systematically generates more sample points in the neighborhood of the function mode while statistically covers the entire search space. Quadratic regression is performed to detect the region containing the global optimum. The sampling and detection process iterates until the global optimum is obtained. Through intensive testing, this method is found to be effective, efficient, robust, and applicable to both continuous and discontinuous functions. It supports simultaneous computation and applies to both unconstrained and constrained optimization problems. Because it does not call any existing global optimization tool, it can be used as a standalone global optimization method for inexpensive problems as well. Limitation of the method is also identified and discussed.


Author(s):  
Ken Ferens ◽  
Darcy Cook ◽  
Witold Kinsner

This paper proposes the application of chaos in large search space problems, and suggests that this represents the next evolutionary step in the development of adaptive and intelligent systems towards cognitive machines and systems. Three different versions of chaotic simulated annealing (XSA) were applied to combinatorial optimization problems in multiprocessor task allocation. Chaotic walks in the solution space were taken to search for the global optimum or “good enough” task-to-processor allocation solutions. Chaotic variables were generated to set the number of perturbations made in each iteration of a XSA algorithm. In addition, parameters of a chaotic variable generator were adjusted to create different chaotic distributions with which to search the solution space. The results show that the convergence rate of the XSA algorithm is faster than simulated annealing when the solutions are far apart in the solution space. In particular, the XSA algorithms found simulated annealing’s best result on average about 4 times faster than simulated annealing.


Symmetry ◽  
2020 ◽  
Vol 12 (9) ◽  
pp. 1529
Author(s):  
Jung-Fa Tsai ◽  
Ming-Hua Lin ◽  
Duan-Yi Wen

Several structural design problems that involve continuous and discrete variables are very challenging because of the combinatorial and non-convex characteristics of the problems. Although the deterministic optimization approach theoretically guarantees to find the global optimum, it usually leads to a significant burden in computational time. This article studies the deterministic approach for globally solving mixed–discrete structural optimization problems. An improved method that symmetrically reduces the number of constraints for linearly expressing signomial terms with pure discrete variables is applied to significantly enhance the computational efficiency of obtaining the exact global optimum of the mixed–discrete structural design problem. Numerical experiments of solving the stepped cantilever beam design problem and the pressure vessel design problem are conducted to show the efficiency and effectiveness of the presented approach. Compared with existing methods, this study introduces fewer convex terms and constraints for transforming the mixed–discrete structural problem and uses much less computational time for solving the reformulated problem to global optimality.


2004 ◽  
Vol 12 (1) ◽  
pp. 47-76 ◽  
Author(s):  
Jonathan Rowe ◽  
Darrell Whitley ◽  
Laura Barbulescu ◽  
Jean-Paul Watson

Representations are formalized as encodings that map the search space to the vertex set of a graph. We define the notion of bit equivalent encodings and show that for such encodings the corresponding Walsh coefficients are also conserved. We focus on Gray codes as particular types of encoding and present a review of properties related to the use of Gray codes. Gray codes are widely used in conjunction with genetic algorithms and bit-climbing algorithms for parameter optimization problems. We present new convergence proofs for a special class of unimodal functions; the proofs show that a steepest ascent bit climber using any reflected Gray code representation reaches the global optimum in a number of steps that is linear with respect to the encoding size. There are in fact many different Gray codes.Shifting is defined as a mechanism for dynamically switching from one Gray code representation to another in order to escape local optima. Theoretical results that substantially improve our understanding of the Gray codes and the shifting mechanism are presented. New proofs also shed light on the number of unique Gray code neighborhoods accessible via shifting and on how neighborhood structure changes during shifting. We show that shifting can improve the performance of both a local search algorithm as well as one of the best genetic algorithms currently available.


Sign in / Sign up

Export Citation Format

Share Document