All NP-Problems Can Be Solved in Polynomial Time by Accepting Networks of Splicing Processors of Constant Size

Author(s):  
Florin Manea ◽  
Carlos Martín-Vide ◽  
Victor Mitrana
Author(s):  
Katrin Casel ◽  
Henning Fernau ◽  
Serge Gaspers ◽  
Benjamin Gras ◽  
Markus L. Schmid

AbstractIn the smallest grammar problem, we are given a word w and we want to compute a preferably small context-free grammar G for the singleton language {w} (where the size of a grammar is the sum of the sizes of its rules, and the size of a rule is measured by the length of its right side). It is known that, for unbounded alphabets, the decision variant of this problem is NP-hard and the optimisation variant does not allow a polynomial-time approximation scheme, unless P = NP. We settle the long-standing open problem whether these hardness results also hold for the more realistic case of a constant-size alphabet. More precisely, it is shown that the smallest grammar problem remains NP-complete (and its optimisation version is APX-hard), even if the alphabet is fixed and has size of at least 17. The corresponding reduction is robust in the sense that it also works for an alternative size-measure of grammars that is commonly used in the literature (i. e., a size measure also taking the number of rules into account), and it also allows to conclude that even computing the number of rules required by a smallest grammar is a hard problem. On the other hand, if the number of nonterminals (or, equivalently, the number of rules) is bounded by a constant, then the smallest grammar problem can be solved in polynomial time, which is shown by encoding it as a problem on graphs with interval structure. However, treating the number of rules as a parameter (in terms of parameterised complexity) yields W[1]-hardness. Furthermore, we present an $\mathcal {O}(3^{\mid {w}\mid })$ O ( 3 ∣ w ∣ ) exact exponential-time algorithm, based on dynamic programming. These three main questions are also investigated for 1-level grammars, i. e., grammars for which only the start rule contains nonterminals on the right side; thus, investigating the impact of the “hierarchical depth” of grammars on the complexity of the smallest grammar problem. In this regard, we obtain for 1-level grammars similar, but slightly stronger results.


Author(s):  
Alasdair Urquhart

The theory of computational complexity is concerned with estimating the resources a computer needs to solve a given problem. The basic resources are time (number of steps executed) and space (amount of memory used). There are problems in logic, algebra and combinatorial games that are solvable in principle by a computer, but computationally intractable because the resources required by relatively small instances are practically infeasible. The theory of NP-completeness concerns a common type of problem in which a solution is easy to check but may be hard to find. Such problems belong to the class NP; the hardest ones of this type are the NP-complete problems. The problem of determining whether a formula of propositional logic is satisfiable or not is NP-complete. The class of problems with feasible solutions is commonly identified with the class P of problems solvable in polynomial time. Assuming this identification, the conjecture that some NP problems require infeasibly long times for their solution is equivalent to the conjecture that P≠NP. Although the conjecture remains open, it is widely believed that NP-complete problems are computationally intractable.


2018 ◽  
Vol 27 (5) ◽  
pp. 808-828 ◽  
Author(s):  
LEONID A. LEVIN ◽  
RAMARATHNAM VENKATESAN

NP-complete problems should be hard on some instances but those may be extremely rare. On generic instances many such problems, especially related to random graphs, have been proved to be easy. We show the intractability of random instances of a graph colouring problem: this graph problem is hard on average unless all NP problems under all samplable (i.e. generatable in polynomial time) distributions are easy. Worst case reductions use special gadgets and typically map instances into a negligible fraction of possible outputs. Ours must output nearly random graphs and avoid any super-polynomial distortion of probabilities. This poses significant technical difficulties.


2018 ◽  
Vol 60 (2) ◽  
pp. 360-375
Author(s):  
A. V. Vasil'ev ◽  
D. V. Churikov

10.29007/v68w ◽  
2018 ◽  
Author(s):  
Ying Zhu ◽  
Mirek Truszczynski

We study the problem of learning the importance of preferences in preference profiles in two important cases: when individual preferences are aggregated by the ranked Pareto rule, and when they are aggregated by positional scoring rules. For the ranked Pareto rule, we provide a polynomial-time algorithm that finds a ranking of preferences such that the ranked profile correctly decides all the examples, whenever such a ranking exists. We also show that the problem to learn a ranking maximizing the number of correctly decided examples (also under the ranked Pareto rule) is NP-hard. We obtain similar results for the case of weighted profiles when positional scoring rules are used for aggregation.


Author(s):  
Yishay Mor ◽  
Claudia V. Goldman ◽  
Jeffrey S. Rosenschein
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document