scholarly journals Systematic exploration of protein conformational space using a Distance Geometry approach

2019 ◽  
Author(s):  
Thérèse E. Malliavin ◽  
Antonio Mucherino ◽  
Carlile Lavor ◽  
Leo Liberti

AbstractThe optimisation approaches classically used during the determination of protein structure encounter various diffculties, specially when the size of the conformational space is large. Indeed, in such case, algorithmic convergence criteria are more difficult to set up. Moreover, the size of the search space makes it difficult to achieve a complete exploration. The interval Branch-and-Prune (iBP) approach, based on the reformulating of the Distance Geometry Problem (DGP) provides a theoretical frame for the generation of protein conformations, by systematically sampling the conformational space. When an appropriate subset of inter-atomic distances is known exactly, this worst-case exponential-time algorithm is provably complete and fixed-parameter tractable. These guarantees, however, immediately disappear as distance measurement errors are introduced. Here we propose an improvement of this approach: the threading-augmented interval Branch-and-Prune (TAiBP), where the combinatorial explosion of the original iBP approach arising from its exponential complexity is alleviated by partitioning the input instances into consecutive peptide fragments and by using Self-Organizing Maps (SOMs) to obtain clusters of similar solutions. A validation of the TAiBP approach is presented here on a set of proteins of various sizes and structures. The calculation inputs are: a uniform covalent geometry extracted from force field covalent terms, the backbone dihedral angles with error intervals, and a few long-range distances. For most of the proteins smaller than 50 residues and interval widths of 20°, the TAiBP approach yielded solutions with RMSD values smaller than 3 Å with respect to the initial protein conformation. The efficiency of TAiBP approach for proteins larger than 50 residues will require the use of non-uniform covalent geometry, and may have benefits from the recent development of residue-specific force-fields.

Entropy ◽  
2018 ◽  
Vol 20 (4) ◽  
pp. 274 ◽  
Author(s):  
◽  

Dynamic Bayesian networks (DBN) are powerful probabilistic representations that model stochastic processes. They consist of a prior network, representing the distribution over the initial variables, and a set of transition networks, representing the transition distribution between variables over time. It was shown that learning complex transition networks, considering both intra- and inter-slice connections, is NP-hard. Therefore, the community has searched for the largest subclass of DBNs for which there is an efficient learning algorithm. We introduce a new polynomial-time algorithm for learning optimal DBNs consistent with a breadth-first search (BFS) order, named bcDBN. The proposed algorithm considers the set of networks such that each transition network has a bounded in-degree, allowing for p edges from past time slices (inter-slice connections) and k edges from the current time slice (intra-slice connections) consistent with the BFS order induced by the optimal tree-augmented network (tDBN). This approach increases exponentially, in the number of variables, the search space of the state-of-the-art tDBN algorithm. Concerning worst-case time complexity, given a Markov lag m, a set of n random variables ranging over r values, and a set of observations of N individuals over T time steps, the bcDBN algorithm is linear in N, T and m; polynomial in n and r; and exponential in p and k. We assess the bcDBN algorithm on simulated data against tDBN, revealing that it performs well throughout different experiments.


2011 ◽  
Vol 40 ◽  
pp. 657-676 ◽  
Author(s):  
L. Bordeaux ◽  
G. Katsirelos ◽  
N. Narodytska ◽  
M. Y. Vardi

Bound propagation is an important Artificial Intelligence technique used in Constraint Programming tools to deal with numerical constraints. It is typically embedded within a search procedure (”branch and prune”) and used at every node of the search tree to narrow down the search space, so it is critical that it be fast. The procedure invokes constraint propagators until a common fixpoint is reached, but the known algorithms for this have a pseudo-polynomial worst-case time complexity: they are fast indeed when the variables have a small numerical range, but they have the well-known problem of being prohibitively slow when these ranges are large. An important question is therefore whether strongly-polynomial algorithms exist that compute the common bound consistent fixpoint of a set of constraints. This paper answers this question. In particular we show that this fixpoint computation is in fact NP-complete, even when restricted to binary linear constraints.


2021 ◽  
Vol 11 (8) ◽  
pp. 3627
Author(s):  
Michael B. Rahaim ◽  
Thomas D. C. Little ◽  
Mona Hella

To meet the growing demand for wireless capacity, communications in the Terahertz (THz) and optical bands are being broadly explored. Communications within these bands provide massive bandwidth potential along with highly directional beam steering capabilities. While the available bandwidth offers incredible link capacity, the directionality of these technologies offers an even more significant potential for spatial capacity or area spectral efficiency. However, this directionality also implies a challenge related to the network’s ability to quickly establish a connection. In this paper, we introduce a multi-tier heterogeneous (MTH) beamform management strategy that utilizes various wireless technologies in order to quickly acquire a highly directional indoor free space optical communication (FSO) link. The multi-tier design offers the high resolution of indoor FSO while the millimeter-wave (mmWave) system narrows the FSO search space. By narrowing the search space, the system relaxes the requirements of the FSO network in order to assure a practical search time. This paper introduces the necessary components of the proposed beam management strategy and provides a foundational analysis framework to demonstrate the relative impact of coverage, resolution, and steering velocity across tiers. Furthermore, an optimization analysis is used to define the top tier resolution that minimizes worst-case search time as a function of lower tier resolution and top tier range.


2020 ◽  
Vol 34 (09) ◽  
pp. 13700-13703
Author(s):  
Nikhil Vyas ◽  
Ryan Williams

All known SAT-solving paradigms (backtracking, local search, and the polynomial method) only yield a 2n(1−1/O(k)) time algorithm for solving k-SAT in the worst case, where the big-O constant is independent of k. For this reason, it has been hypothesized that k-SAT cannot be solved in worst-case 2n(1−f(k)/k) time, for any unbounded ƒ : ℕ → ℕ. This hypothesis has been called the “Super-Strong Exponential Time Hypothesis” (Super Strong ETH), modeled after the ETH and the Strong ETH. We prove two results concerning the Super-Strong ETH:1. It has also been hypothesized that k-SAT is hard to solve for randomly chosen instances near the “critical threshold”, where the clause-to-variable ratio is 2k ln 2 −Θ(1). We give a randomized algorithm which refutes the Super-Strong ETH for the case of random k-SAT and planted k-SAT for any clause-to-variable ratio. In particular, given any random k-SAT instance F with n variables and m clauses, our algorithm decides satisfiability for F in 2n(1−Ω( log k)/k) time, with high probability (over the choice of the formula and the randomness of the algorithm). It turns out that a well-known algorithm from the literature on SAT algorithms does the job: the PPZ algorithm of Paturi, Pudlak, and Zane (1998).2. The Unique k-SAT problem is the special case where there is at most one satisfying assignment. It is natural to hypothesize that the worst-case (exponential-time) complexity of Unique k-SAT is substantially less than that of k-SAT. Improving prior reductions, we show the time complexities of Unique k-SAT and k-SAT are very tightly related: if Unique k-SAT is in 2n(1−f(k)/k) time for an unbounded f, then k-SAT is in 2n(1−f(k)(1−ɛ)/k) time for every ɛ > 0. Thus, refuting Super Strong ETH in the unique solution case would refute Super Strong ETH in general.


2013 ◽  
Vol 24 (07) ◽  
pp. 1067-1082 ◽  
Author(s):  
YO-SUB HAN ◽  
SANG-KI KO ◽  
KAI SALOMAA

The edit-distance between two strings is the smallest number of operations required to transform one string into the other. The distance between languages L1and L2is the smallest edit-distance between string wi∈ Li, i = 1, 2. We consider the problem of computing the edit-distance of a given regular language and a given context-free language. First, we present an algorithm that finds for the languages an optimal alignment, that is, a sequence of edit operations that transforms a string in one language to a string in the other. The length of the optimal alignment, in the worst case, is exponential in the size of the given grammar and finite automaton. Then, we investigate the problem of computing only the edit-distance of the languages without explicitly producing an optimal alignment. We design a polynomial time algorithm that calculates the edit-distance based on unary homomorphisms.


2011 ◽  
Vol 21 (01) ◽  
pp. 87-100
Author(s):  
GREG ALOUPIS ◽  
PROSENJIT BOSE ◽  
ERIK D. DEMAINE ◽  
STEFAN LANGERMAN ◽  
HENK MEIJER ◽  
...  

Given a planar polygon (or chain) with a list of edges {e1, e2, e3, …, en-1, en}, we examine the effect of several operations that permute this edge list, resulting in the formation of a new polygon. The main operations that we consider are: reversals which involve inverting the order of a sublist, transpositions which involve interchanging subchains (sublists), and edge-swaps which are a special case and involve interchanging two consecutive edges. When each edge of the given polygon has also been assigned a direction we say that the polygon is signed. In this case any edge involved in a reversal changes direction. We show that a star-shaped polygon can be convexified using O(n2) edge-swaps, while maintaining simplicity, and that this is tight in the worst case. We show that determining whether a signed polygon P can be transformed to one that has rotational or mirror symmetry with P, using transpositions, takes Θ(n log n) time. We prove that the problem of deciding whether transpositions can modify a polygon to fit inside a rectangle is weakly NP-complete. Finally we give an O(n log n) time algorithm to compute the maximum endpoint distance for an oriented chain.


2011 ◽  
Vol 09 (03) ◽  
pp. 383-398 ◽  
Author(s):  
BRIAN OLSON ◽  
KEVIN MOLLOY ◽  
AMARDA SHEHU

The three-dimensional structure of a protein is a key determinant of its biological function. Given the cost and time required to acquire this structure through experimental means, computational models are necessary to complement wet-lab efforts. Many computational techniques exist for navigating the high-dimensional protein conformational search space, which is explored for low-energy conformations that comprise a protein's native states. This work proposes two strategies to enhance the sampling of conformations near the native state. An enhanced fragment library with greater structural diversity is used to expand the search space in the context of fragment-based assembly. To manage the increased complexity of the search space, only a representative subset of the sampled conformations is retained to further guide the search towards the native state. Our results make the case that these two strategies greatly enhance the sampling of the conformational space near the native state. A detailed comparative analysis shows that our approach performs as well as state-of-the-art ab initio structure prediction protocols.


2021 ◽  
Author(s):  
Taqiaden Alshameri ◽  
Yude Dong ◽  
Abdullah Alqadhi

Abstract Fixture synthesis addresses the problem of fixture-elements placement on the workpiece surfaces. This article presents a novel variant of the Simulated Annealing (SA) algorithm called Declining Neighborhood Simulated Annealing (DNSA) specifically developed for the problem of fixture synthesis. The objective is to minimize measurement errors in the machined features induced by the misalignment at workpiece-locator contact points. The algorithm systematically evaluates different fixture layouts to reach a sufficient approximation of the global optimum robust layout. For each iteration, a set of previously accepted candidates are exploited to predict the next move. Throughout the progress of the algorithm, the search space is reduced and the new candidates are designated according to a declining Probability Density Function (PDF). To assure best performance, the DNSA parameters are configured using the Technique for Order Preference by Similarity to Ideal Solution (TOPOSIS). Moreover, the parameters are set to auto-adapt the complexity of a given input based on a Shanon entropy index. The optimization process is carried out automatically in the Computer-Aided Design (CAD) environment NX; a computer code was developed for this purpose using the Application Programming Interface (API) NXOpen. Benchmark examples from industrial partner and literature demonstrate satisfactory results.


Algorithmica ◽  
2021 ◽  
Author(s):  
Thom Castermans ◽  
Bettina Speckmann ◽  
Frank Staals ◽  
Kevin Verbeek

AbstractWe study an agglomerative clustering problem motivated by interactive glyphs in geo-visualization. Consider a set of disjoint square glyphs on an interactive map. When the user zooms out, the glyphs grow in size relative to the map, possibly with different speeds. When two glyphs intersect, we wish to replace them by a new glyph that captures the information of the intersecting glyphs. We present a fully dynamic kinetic data structure that maintains a set of n disjoint growing squares. Our data structure uses $$O\bigl (n \log n \log \log n\bigr )$$ O ( n log n log log n ) space, supports queries in worst case $$O\bigl (\log ^2 n\bigr )$$ O ( log 2 n ) time, and updates in $$O\bigl (\log ^5 n\bigr )$$ O ( log 5 n ) amortized time. This leads to an $$O\bigl (n\,\alpha (n)\log ^5 n\bigr )$$ O ( n α ( n ) log 5 n ) time algorithm to solve the agglomerative clustering problem. This is a significant improvement over the current best $$O\bigl (n^2\bigr )$$ O ( n 2 ) time algorithms.


Sign in / Sign up

Export Citation Format

Share Document