Treewidth-aware Reductions of Normal ASP to SAT - Is Normal ASP Harder than SAT after All?

Author(s):  
Markus Hecher

Answer Set Programming (ASP) is a paradigm and problem modeling/solving toolkit for KR that is often invoked. There are plenty of results dedicated to studying the hardness of (fragments of) ASP. So far, these studies resulted in characterizations in terms of computational complexity as well as in fine-grained insights presented in form of dichotomy-style results, lower bounds when translating to other formalisms like propositional satisfiability (SAT), and even detailed parameterized complexity landscapes. A quite generic and prominent parameter in parameterized complexity originating from graph theory is the so-called treewidth, which in a sense captures structural density of a program. Recently, there was an increase in the number of treewidth-based solvers related to SAT. While there exist several translations from (normal) ASP to SAT, yet there is no reduction preserving treewidth or at least being aware of the treewidth increase. This paper deals with a novel reduction from normal ASP to SAT that is aware of the treewidth, and guarantees that a slight increase of treewidth is indeed sufficient. Then, we also present a new result establishing that when considering treewidth, already the fragment of normal ASP is slightly harder than SAT (under reasonable assumptions in computational complexity). This also confirms that our reduction probably cannot be significantly improved and that the slight increase of treewidth is unavoidable.

2016 ◽  
Vol 08 (03) ◽  
pp. 1650040 ◽  
Author(s):  
Shaohui Wang ◽  
Bing Wei

Let [Formula: see text] be multiplicative Zagreb index of a graph [Formula: see text]. A connected graph is a cactus graph if and only if any two of its cycles have at most one vertex in common, which is a generalization of trees and has been the interest of researchers in the field of material chemistry and graph theory. In this paper, we use a new tool to obtain the upper and lower bounds of [Formula: see text] for all cactus graphs and characterize the corresponding extremal graphs.


2022 ◽  
Vol 69 (1) ◽  
pp. 1-18
Author(s):  
Anupam Gupta ◽  
David G. Harris ◽  
Euiwoong Lee ◽  
Jason Li

In the k -cut problem, we want to find the lowest-weight set of edges whose deletion breaks a given (multi)graph into k connected components. Algorithms of Karger and Stein can solve this in roughly O ( n 2k ) time. However, lower bounds from conjectures about the k -clique problem imply that Ω ( n (1- o (1)) k ) time is likely needed. Recent results of Gupta, Lee, and Li have given new algorithms for general k -cut in n 1.98k + O(1) time, as well as specialized algorithms with better performance for certain classes of graphs (e.g., for small integer edge weights). In this work, we resolve the problem for general graphs. We show that the Contraction Algorithm of Karger outputs any fixed k -cut of weight α λ k with probability Ω k ( n - α k ), where λ k denotes the minimum k -cut weight. This also gives an extremal bound of O k ( n k ) on the number of minimum k -cuts and an algorithm to compute λ k with roughly n k polylog( n ) runtime. Both are tight up to lower-order factors, with the algorithmic lower bound assuming hardness of max-weight k -clique. The first main ingredient in our result is an extremal bound on the number of cuts of weight less than 2 λ k / k , using the Sunflower lemma. The second ingredient is a fine-grained analysis of how the graph shrinks—and how the average degree evolves—in the Karger process.


Author(s):  
Giovanni Amendola ◽  
Carmine Dodaro ◽  
Marco Maratea

The issue of describing in a formal way solving algorithms in various fields such as Propositional Satisfiability (SAT), Quantified SAT, Satisfiability Modulo Theories, Answer Set Programming (ASP), and Constraint ASP, has been relatively recently solved employing abstract solvers. In this paper we deal with cautious reasoning tasks in ASP, and design, implement and test novel abstract solutions, borrowed from backbone computation in SAT. By employing abstract solvers, we also formally show that the algorithms for solving cautious reasoning tasks in ASP are strongly related to those for computing backbones of Boolean formulas. Some of the new solutions have been implemented in the ASP solver WASP, and tested.


Author(s):  
Marko Samer ◽  
Stefan Szeider

Parameterized complexity is a new theoretical framework that considers, in addition to the overall input size, the effects on computational complexity of a secondary measurement, the parameter. This two-dimensional viewpoint allows a fine-grained complexity analysis that takes structural properties of problem instances into account. The central notion is “fixed-parameter tractability” which refers to solvability in polynomial time for each fixed value of the parameter such that the order of the polynomial time bound is independent of the parameter. This chapter presents main concepts and recent results on the parameterized complexity of the satisfiability problem and it outlines fundamental algorithmic ideas that arise in this context. Among the parameters considered are the size of backdoor sets with respect to various tractable base classes and the treewidth of graph representations of satisfiability instances.


2001 ◽  
Vol 38 (02) ◽  
pp. 407-420
Author(s):  
Henry W. Block ◽  
Tuhao Chen

Univariate probability inequalities have received extensive attention. It has been shown that under certain conditions, product-type bounds are valid and sharper than summation-type bounds. Although results concerning multivariate inequalities have appeared in the literature, product-type bounds in a multivariate setting have not yet been studied. This note explores an approach using graph theory and linear programming techniques to construct product-type lower bounds for the probability of the intersection among unions of k sets of events.


2010 ◽  
Vol 2010 ◽  
pp. 1-11
Author(s):  
Qin Guo ◽  
Mingxing Luo ◽  
Lixiang Li ◽  
Yixian Yang

From the perspectives of graph theory and combinatorics theory we obtain some new upper bounds on the number of encoding nodes, which can characterize the coding complexity of the network coding, both in feasible acyclic and cyclic multicast networks. In contrast to previous work, during our analysis we first investigate the simple multicast network with source rateh=2, and thenh≥2. We find that for feasible acyclic multicast networks our upper bound is exactly the lower bound given by M. Langberg et al. in 2006. So the gap between their lower and upper bounds for feasible acyclic multicast networks does not exist. Based on the new upper bound, we improve the computational complexity given by M. Langberg et al. in 2009. Moreover, these results further support the feasibility of signatures for network coding.


Sign in / Sign up

Export Citation Format

Share Document