scholarly journals Parallel exhaustive search vs. evolutionary computation in a large real world network search space

Author(s):  
Garnett Wilson ◽  
Simon Harding ◽  
Orland Hoeber ◽  
Rodolphe Devillers ◽  
Wolfgang Banzhaf
Author(s):  
Marcos Gestal ◽  
José Manuel Vázquez Naya ◽  
Norberto Ezquerra

Traditionally, the Evolutionary Computation (EC) techniques, and more specifically the Genetic Algorithms (GAs), have proved to be efficient when solving various problems; however, as a possible lack, the GAs tend to provide a unique solution for the problem on which they are applied. Some non global solutions discarded during the search of the best one could be acceptable under certain circumstances. Most of the problems at the real world involve a search space with one or more global solutions and multiple local solutions; this means that they are multimodal problems and therefore, if it is desired to obtain multiple solutions by using GAs, it would be necessary to modify their classic functioning outline for adapting them correctly to the multimodality of such problems. The present chapter tries to establish, firstly, the characterisation of the multimodal problems will be attempted. A global view of some of the several approaches proposed for adapting the classic functioning of the GAs to the search of mu ltiple solutions will be also offered. Lastly, the contributions of the authors and a brief description of several practical cases of their performance at the real world will be also showed.


Author(s):  
Marcos Gestal ◽  
Mari Paz Gómez-Carracedo

Traditionally, the Evolutionary Computation (EC) techniques, and more specifically the Genetic Algorithms (GAs) (Goldberg & Wang, 1989), have proved to be efficient when solving various problems; however, as a possible lack, the GAs tend to provide a unique solution for the problem on which they are applied. Some non global solutions discarded during the search of the best one could be acceptable under certain circumstances. The majority of the problems at the real world involve a search space with one or more global solutions and multiple local solutions; this means that they are multimodal problems (Harik, 1995) and therefore, if it is desired to obtain multiple solutions by using GAs, it would be necessary to modify their classic functioning outline for adapting them correctly to the multimodality of such problems.


Author(s):  
Gwendolyn Rehrig ◽  
Reese A. Cullimore ◽  
John M. Henderson ◽  
Fernanda Ferreira

Abstract According to the Gricean Maxim of Quantity, speakers provide the amount of information listeners require to correctly interpret an utterance, and no more (Grice in Logic and conversation, 1975). However, speakers do tend to violate the Maxim of Quantity often, especially when the redundant information improves reference precision (Degen et al. in Psychol Rev 127(4):591–621, 2020). Redundant (non-contrastive) information may facilitate real-world search if it narrows the spatial scope under consideration, or improves target template specificity. The current study investigated whether non-contrastive modifiers that improve reference precision facilitate visual search in real-world scenes. In two visual search experiments, we compared search performance when perceptually relevant, but non-contrastive modifiers were included in the search instruction. Participants (NExp. 1 = 48, NExp. 2 = 48) searched for a unique target object following a search instruction that contained either no modifier, a location modifier (Experiment 1: on the top left, Experiment 2: on the shelf), or a color modifier (the black lamp). In Experiment 1 only, the target was located faster when the verbal instruction included either modifier, and there was an overall benefit of color modifiers in a combined analysis for scenes and conditions common to both experiments. The results suggest that violations of the Maxim of Quantity can facilitate search when the violations include task-relevant information that either augments the target template or constrains the search space, and when at least one modifier provides a highly reliable cue. Consistent with Degen et al. (2020), we conclude that listeners benefit from non-contrastive information that improves reference precision, and engage in rational reference comprehension. Significance statement This study investigated whether providing more information than someone needs to find an object in a photograph helps them to find that object more easily, even though it means they need to interpret a more complicated sentence. Before searching a scene, participants were either given information about where the object would be located in the scene, what color the object was, or were only told what object to search for. The results showed that providing additional information helped participants locate an object in an image more easily only when at least one piece of information communicated what part of the scene the object was in, which suggests that more information can be beneficial as long as that information is specific and helps the recipient achieve a goal. We conclude that people will pay attention to redundant information when it supports their task. In practice, our results suggest that instructions in other contexts (e.g., real-world navigation, using a smartphone app, prescription instructions, etc.) can benefit from the inclusion of what appears to be redundant information.


2011 ◽  
Vol 204-210 ◽  
pp. 245-250
Author(s):  
Guo Sheng Hao ◽  
Xiang Jun Zhao ◽  
Yong Qing Huang

user in interactive evolutionary computation (IEC) has the characteristic of fuzzy cognition. Based on this, a method to learn users’ fuzzy cognition knowledge is given. The method includes the fuzzy expression of the basic elements of IEC such as search space, population, gene sense unit and so on. Then a method to increase the performance of IEC based on the knowledge of users’ fuzzy cognition is given. The above results enrich the researches of IEC users' cognition.


Author(s):  
Victor Oduguwa ◽  
Rajkumar Roy ◽  
Didier Farrugia

Most of the algorithmic engineering design optimisation approaches reported in the literature aims to find the best set of solutions within a quantitative (QT) search space of the given problem while ignoring related qualitative (QL) issues. These QL issues can be very important and by ignoring them in the optimisation search, can have expensive consequences especially for real world problems. This paper presents a new integrated design optimisation approach for QT and QL search space. The proposed solution approach is based on design of experiment methods and fuzzy logic principles for building the required QL models, and evolutionary multi-objective optimisation technique for solving the design problem. The proposed technique was applied to a two objectives rod rolling problem. The results obtained demonstrate that the proposed solution approach can be used to solve real world problems taking into account the related QL evaluation of the design problem.


Author(s):  
Tüze Kuyucu ◽  
Ivan Tanev ◽  
Katsunori Shimohara

In Genetic Programming (GP), most often the search space grows in a greater than linear fashion as the number of tasks required to be accomplished increases. This is a cause for one of the greatest problems in Evolutionary Computation (EC): scalability. The aim of the work presented here is to facilitate the evolution of control systems for complex robotic systems. The authors use a combination of mechanisms specifically designed to facilitate the fast evolution of systems with multiple objectives. These mechanisms are: a genetic transposition inspired seeding, a strongly-typed crossover, and a multiobjective optimization. The authors demonstrate that, when used together, these mechanisms not only improve the performance of GP but also the reliability of the final designs. They investigate the effect of the aforementioned mechanisms on the efficiency of GP employed for the coevolution of locomotion gaits and sensing of a simulated snake-like robot (Snakebot). Experimental results show that the mechanisms set forth contribute to significant increase in the efficiency of the evolution of fast moving and sensing Snakebots as well as the robustness of the final designs.


Entropy ◽  
2020 ◽  
Vol 22 (4) ◽  
pp. 407 ◽  
Author(s):  
Dominik Weikert ◽  
Sebastian Mai ◽  
Sanaz Mostaghim

In this article, we present a new algorithm called Particle Swarm Contour Search (PSCS)—a Particle Swarm Optimisation inspired algorithm to find object contours in 2D environments. Currently, most contour-finding algorithms are based on image processing and require a complete overview of the search space in which the contour is to be found. However, for real-world applications this would require a complete knowledge about the search space, which may not be always feasible or possible. The proposed algorithm removes this requirement and is only based on the local information of the particles to accurately identify a contour. Particles search for the contour of an object and then traverse alongside using their known information about positions in- and out-side of the object. Our experiments show that the proposed PSCS algorithm can deliver comparable results as the state-of-the-art.


2017 ◽  
Vol 10 (2) ◽  
pp. 52
Author(s):  
Natarajan Meghanathan

Results of correlation study (using Pearson's correlation coefficient, PCC) between decay centrality (DEC) vs. degree centrality (DEG) and closeness centrality (CLC) for a suite of 48 real-world networks indicate an interesting trend: PCC(DEC, DEG) decreases with increase in the decay parameter δ (0 < δ < 1) and PCC(DEC, CLC) decreases with decrease in δ. We make use of this trend of monotonic decrease in the PCC values (from both sides of the δ-search space) and propose a binary search algorithm that (given a threshold value r for the PCC) could be used to identify a value of δ (if one exists, we say there exists a positive δ-spacer) for a real-world network such that PCC(DEC, DEG) ≥ r as well as PCC(DEC, CLC) ≥ r. We show the use of the binary search algorithm to find the maximum Threshold PCC value rmax (such that δ-spacermax is positive) for a real-world network. We observe a very strong correlation between rmax and PCC(DEG, CLC) as well as observe real-world networks with a larger variation in node degree to more likely have a lower rmax value and vice-versa.


Symmetry ◽  
2020 ◽  
Vol 12 (12) ◽  
pp. 2021
Author(s):  
Ahmad Asrul Ibrahim ◽  
Khairuddin Khalid ◽  
Hussain Shareef ◽  
Nor Azwan Mohamed Kamari

This paper proposes a technique to determine the possible optimal placement of the phasor measurement unit (PMU) in power grids for normal operating conditions. All possible combinations of PMU placement, including infeasible combinations, are typically considered in finding the optimal solution, which could be a massive search space. An integer search algorithm called the bounded search technique is introduced to reduce the search space in solving a minimum number of PMU allocations whilst maintaining full system observability. The proposed technique is based on connectivity and symmetry constraints that can be derived from the observability matrix. As the technique is coupled with the exhaustive technique, the technique is called the bounded exhaustive search (BES) technique. Several IEEE test systems, namely, IEEE 9-bus, IEEE 14-bus, IEEE 24-bus and IEEE 30-bus, are considered to showcase the performance of the proposed technique. An initial Monte Carlo simulation was carried out to evaluate the capability of the bounded search technique in providing a smaller feasible search space. The effectiveness of the BES technique in terms of computational time is compared with the existing exhaustive technique. Results demonstrate that the search space can be reduced tremendously, and the computational burden can be eased, when finding the optimal PMU placement in power grids.


Sign in / Sign up

Export Citation Format

Share Document