scholarly journals Knee Point Search Using Cascading Top-kSorting with Minimized Time Complexity

2013 ◽  
Vol 2013 ◽  
pp. 1-10 ◽  
Author(s):  
Zheng Wang ◽  
Shian-Shyong Tseng

Anomaly detection systems and many other applications are frequently confronted with the problem of finding the largest knee point in the sorted curve for a set of unsorted points. This paper proposes an efficient knee point search algorithm with minimized time complexity using the cascading top-ksorting when a priori probability distribution of the knee point is known. First, a top-ksort algorithm is proposed based on a quicksort variation. We divide the knee point search problem into multiple steps. And in each step an optimization problem of the selection numberkis solved, where the objective function is defined as the expected time cost. Because the expected time cost in one step is dependent on that of the afterwards steps, we simplify the optimization problem by minimizing the maximum expected time cost. The posterior probability of the largest knee point distribution and the other parameters are updated before solving the optimization problem in each step. An example of source detection of DNS DoS flooding attacks is provided to illustrate the applications of the proposed algorithm.

2010 ◽  
Vol 20 (6) ◽  
pp. 999-1009 ◽  
Author(s):  
G. ABAL ◽  
R. DONANGELO ◽  
F. L. MARQUEZINO ◽  
R. PORTUGAL

The spatial search problem consists of minimising the number of steps required to find a given site in a network under the restriction that only oracle queries or translations to neighbouring sites are allowed. We propose a quantum algorithm for the spatial search problem on a honeycomb lattice with N sites and torus-like boundary conditions. The search algorithm is based on a modified quantum walk on an hexagonal lattice and the general framework proposed by Ambainis, Kempe and Rivosh (Ambainis et al. 2005) is employed to show that the time complexity of this quantum search algorithm is $O(\sqrt{N \log N})$.


10.29007/2k64 ◽  
2018 ◽  
Author(s):  
Pat Prodanovic ◽  
Cedric Goeury ◽  
Fabrice Zaoui ◽  
Riadh Ata ◽  
Jacques Fontaine ◽  
...  

This paper presents a practical methodology developed for shape optimization studies of hydraulic structures using environmental numerical modelling codes. The methodology starts by defining the optimization problem and identifying relevant problem constraints. Design variables in shape optimization studies are configuration of structures (such as length or spacing of groins, orientation and layout of breakwaters, etc.) whose optimal orientation is not known a priori. The optimization problem is solved numerically by coupling an optimization algorithm to a numerical model. The coupled system is able to define, test and evaluate a multitude of new shapes, which are internally generated and then simulated using a numerical model. The developed methodology is tested using an example of an optimum design of a fish passage, where the design variables are the length and the position of slots. In this paper an objective function is defined where a target is specified and the numerical optimizer is asked to retrieve the target solution. Such a definition of the objective function is used to validate the developed tool chain. This work uses the numerical model TELEMAC- 2Dfrom the TELEMAC-MASCARET suite of numerical solvers for the solution of shallow water equations, coupled with various numerical optimization algorithms available in the literature.


Energies ◽  
2021 ◽  
Vol 14 (4) ◽  
pp. 857
Author(s):  
Jahedul Islam ◽  
Md Shokor A. Rahaman ◽  
Pandian M. Vasant ◽  
Berihun Mamo Negash ◽  
Ahshanul Hoqe ◽  
...  

Well placement optimization is considered a non-convex and highly multimodal optimization problem. In this article, a modified crow search algorithm is proposed to tackle the well placement optimization problem. This article proposes modifications based on local search and niching techniques in the crow search algorithm (CSA). At first, the suggested approach is verified by experimenting with the benchmark functions. For test functions, the results of the proposed approach demonstrated a higher convergence rate and a better solution. Again, the performance of the proposed technique is evaluated with well placement optimization problem and compared with particle swarm optimization (PSO), the Gravitational Search Algorithm (GSA), and the Crow search algorithm (CSA). The outcomes of the study revealed that the niching crow search algorithm is the most efficient and effective compared to the other techniques.


2021 ◽  
Vol 22 (1) ◽  
Author(s):  
Jing Tian ◽  
Jianping Zhao ◽  
Chunhou Zheng

Abstract Background In recent years, various sequencing techniques have been used to collect biomedical omics datasets. It is usually possible to obtain multiple types of omics data from a single patient sample. Clustering of omics data plays an indispensable role in biological and medical research, and it is helpful to reveal data structures from multiple collections. Nevertheless, clustering of omics data consists of many challenges. The primary challenges in omics data analysis come from high dimension of data and small size of sample. Therefore, it is difficult to find a suitable integration method for structural analysis of multiple datasets. Results In this paper, a multi-view clustering based on Stiefel manifold method (MCSM) is proposed. The MCSM method comprises three core steps. Firstly, we established a binary optimization model for the simultaneous clustering problem. Secondly, we solved the optimization problem by linear search algorithm based on Stiefel manifold. Finally, we integrated the clustering results obtained from three omics by using k-nearest neighbor method. We applied this approach to four cancer datasets on TCGA. The result shows that our method is superior to several state-of-art methods, which depends on the hypothesis that the underlying omics cluster class is the same. Conclusion Particularly, our approach has better performance than compared approaches when the underlying clusters are inconsistent. For patients with different subtypes, both consistent and differential clusters can be identified at the same time.


2020 ◽  
pp. 000370282097751
Author(s):  
Xin Wang ◽  
Xia Chen

Many spectra have a polynomial-like baseline. Iterative polynomial fitting (IPF) is one of the most popular methods for baseline correction of these spectra. However, the baseline estimated by IPF may have substantially error when the spectrum contains significantly strong peaks or have strong peaks located at the endpoints. First, IPF uses temporary baseline estimated from the current spectrum to identify peak data points. If the current spectrum contains strong peaks, then the temporary baseline substantially deviates from the true baseline. Some good baseline data points of the spectrum might be mistakenly identified as peak data points and are artificially re-assigned with a low value. Second, if a strong peak is located at the endpoint of the spectrum, then the endpoint region of the estimated baseline might have significant error due to overfitting. This study proposes a search algorithm-based baseline correction method (SA) that aims to compress sample the raw spectrum to a dataset with small number of data points and then convert the peak removal process into solving a search problem in artificial intelligence (AI) to minimize an objective function by deleting peak data points. First, the raw spectrum is smoothened out by the moving average method to reduce noise and then divided into dozens of unequally spaced sections on the basis of Chebyshev nodes. Finally, the minimal points of each section are collected to form a dataset for peak removal through search algorithm. SA selects the mean absolute error (MAE) as the objective function because of its sensitivity to overfitting and rapid calculation. The baseline correction performance of SA is compared with those of three baseline correction methods: Lieber and Mahadevan–Jansen method, adaptive iteratively reweighted penalized least squares method, and improved asymmetric least squares method. Simulated and real FTIR and Raman spectra with polynomial-like baselines are employed in the experiments. Results show that for these spectra, the baseline estimated by SA has fewer error than those by the three other methods.


2015 ◽  
Vol 24 (05) ◽  
pp. 1550061
Author(s):  
Mateus de Paula Marques ◽  
Taufik Abrão

This paper addresses the optimization problem on subcarrier and power allocation of orthogonal frequency division multiple access (OFDMA) system under spectral efficiency (SE) metric when deploying superposition coding (SC) transmission strategy. An algorithm with polynomial time complexity, of the order of (UN log 2(N)) has been proposed for sub-optimal SE maximization. Results indicate that the system SE increases with the use of SC technique. Besides, the throughput gain with SC adoption increases when the number of users (U) approaches the number of subcarriers (N) available in the system.


2015 ◽  
Vol 22 (03) ◽  
pp. 1550019 ◽  
Author(s):  
S. Iriyama ◽  
M. Ohya ◽  
I.V. Volovich

A new quantum algorithm for the search problem and its computational complexity are discussed. Its essential part is the use of the so-called chaos amplifier, [8, 9, 10, 13]. It is shown that for the search problem containing [Formula: see text] objects time complexity of the method is polynomial in [Formula: see text].


2021 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Danni Chen ◽  
JianDong Zhao ◽  
Peng Huang ◽  
Xiongna Deng ◽  
Tingting Lu

Purpose Sparrow search algorithm (SSA) is a novel global optimization method, but it is easy to fall into local optimization, which leads to its poor search accuracy and stability. The purpose of this study is to propose an improved SSA algorithm, called levy flight and opposition-based learning (LOSSA), based on LOSSA strategy. The LOSSA shows better search accuracy, faster convergence speed and stronger stability. Design/methodology/approach To further enhance the optimization performance of the algorithm, The Levy flight operation is introduced into the producers search process of the original SSA to enhance the ability of the algorithm to jump out of the local optimum. The opposition-based learning strategy generates better solutions for SSA, which is beneficial to accelerate the convergence speed of the algorithm. On the one hand, the performance of the LOSSA is evaluated by a set of numerical experiments based on classical benchmark functions. On the other hand, the hyper-parameter optimization problem of the Support Vector Machine (SVM) is also used to test the ability of LOSSA to solve practical problems. Findings First of all, the effectiveness of the two improved methods is verified by Wilcoxon signed rank test. Second, the statistical results of the numerical experiment show the significant improvement of the LOSSA compared with the original algorithm and other natural heuristic algorithms. Finally, the feasibility and effectiveness of the LOSSA in solving the hyper-parameter optimization problem of machine learning algorithms are demonstrated. Originality/value An improved SSA based on LOSSA is proposed in this paper. The experimental results show that the overall performance of the LOSSA is satisfactory. Compared with the SSA and other natural heuristic algorithms, the LOSSA shows better search accuracy, faster convergence speed and stronger stability. Moreover, the LOSSA also showed great optimization performance in the hyper-parameter optimization of the SVM model.


Author(s):  
Levon Arsalanyan ◽  
Hayk Danoyan

The Nearest Neighbor search algorithm considered in this paper is well known (Elias algorithm). It uses error-correcting codes and constructs appropriate hash-coding schemas. These schemas preprocess the data in the form of lists. Each list is contained in some sphere, centered at a code-word. The algorithm is considered for the cases of perfect codes, so the spheres and, consequently, the lists do not intersect. As such codes exist for the limited set of parameters, the algorithm is considered for some other generalizations of perfect codes, and then the same data point may be contained in different lists. A formula of time complexity of the algorithm is obtained for these cases, using coset weight structures of the mentioned codes


Sign in / Sign up

Export Citation Format

Share Document