scholarly journals On Efficiently Explaining Graph-Based Classifiers

2021 ◽  
Author(s):  
Xuanxiang Huang ◽  
Yacine Izza ◽  
Alexey Ignatiev ◽  
Joao Marques-Silva

Recent work has shown that not only decision trees (DTs) may not be interpretable but also proposed a polynomial-time algorithm for computing one PI-explanation of a DT. This paper shows that for a wide range of classifiers, globally referred to as decision graphs, and which include decision trees and binary decision diagrams, but also their multi-valued variants, there exist polynomial-time algorithms for computing one PI-explanation. In addition, the paper also proposes a polynomial-time algorithm for computing one contrastive explanation. These novel algorithms build on explanation graphs (XpG's). XpG's denote a graph representation that enables both theoretical and practically efficient computation of explanations for decision graphs. Furthermore, the paper proposes a practically efficient solution for the enumeration of explanations, and studies the complexity of deciding whether a given feature is included in some explanation. For the concrete case of decision trees, the paper shows that the set of all contrastive explanations can be enumerated in polynomial time. Finally, the experimental results validate the practical applicability of the algorithms proposed in the paper on a wide range of publicly available benchmarks.

2012 ◽  
Vol Vol. 14 no. 2 (Graph and Algorithms) ◽  
Author(s):  
Raffaele Mosca

Graphs and Algorithms International audience This article deals with the Maximum Weight Stable Set (MWS) problem (and some other related NP-hard problems) and the class of P-6-free graphs. The complexity status of MWS is open for P-6-free graphs and is open even for P-5-free graphs (as a long standing open problem). Several results are known for MWS on subclasses of P-5-free: in particular, MWS can be solved for k-colorable P-5-free graphs in polynomial time for every k (depending on k) and more generally for (P-5, K-p)-free graphs (depending on p), which is a useful result since for every graph G one can easily compute a k-coloring of G, with k not necessarily minimum. This article studies the MWS problem for k-colorable P-6-free graphs and more generally for (P-6, K-p)-free graphs. Though we were not able to define a polynomial time algorithm for this problem for every k, this article introduces: (i) some structure properties of P-6-free graphs with respect to stable sets, (ii) two reductions for MWS on (P-6; K-p)-free graphs for every p, (iii) three polynomial time algorithms to solve MWS respectively for 3-colorable P-6-free, for 4-colorable P-6-free, and for (P-6, K-4)-free graphs (the latter allows one to state, together with other known results, that MWS can be solved for (P-6, F)-free graphs in polynomial time where F is any four vertex graph).


2009 ◽  
Vol 01 (02) ◽  
pp. 253-265 ◽  
Author(s):  
TONI R. FARLEY ◽  
CHARLES J. COLBOURN

Network operation may require that a specified number k of nodes be able to communicate via paths consisting of operating edges and nodes. In an environment of node and edge failure, this leads to associated reliability measures. When the k nodes are known in advance, this has been widely studied as k-terminal reliability; when the k nodes are chosen uniformly at random, this has been studied as k-resilience. A third notion, when it suffices to have anyk nodes communicate, is related to the expected size of the largest component in the network. We generalize these three measures to the probability that given h nodes chosen in advance and i nodes chosen at random, they appear in a component of size at least k = h + i + j. As expected, for general networks, for most choices of (h, i, j) the computation is #P-complete and hence unlikely to admit a polynomial time algorithm. We develop polynomial time algorithms in the special case that the network is series-parallel, which subsume and generalize earlier methods for k-terminal reliability and k-resilience.


Author(s):  
Akiyoshi Shioura

In this paper, we consider a problem of minimizing an M-convex function under an L1-distance constraint (MML1); the constraint is given by an upper bound for L1-distance between a feasible solution and a given “center.” This is motivated by a nonlinear integer programming problem for reallocation of dock capacity in a bike-sharing system discussed by Freund et al. (2017). The main aim of this paper is to better understand the combinatorial structure of the dock reallocation problem through the connection with M-convexity and show its polynomial-time solvability using this connection. For this, we first show that the dock reallocation problem and its generalizations can be reformulated in the form of (MML1). We then present a pseudo-polynomial-time algorithm for (MML1) based on the steepest descent approach. We also propose two polynomial-time algorithms for (MML1) by replacing the L1-distance constraint with a simple linear constraint. Finally, we apply the results for (MML1) to the dock reallocation problem to obtain a pseudo-polynomial-time steepest descent algorithm and also polynomial-time algorithms for this problem. For this purpose, we develop a polynomial-time algorithm for a relaxation of the dock reallocation problem by using a proximity-scaling approach, which is of interest in its own right.


2019 ◽  
Vol 20 (S20) ◽  
Author(s):  
Santi Santichaivekin ◽  
Ross Mawhorter ◽  
Ran Libeskind-Hadas

Abstract Background Maximum parsimony reconciliation in the duplication-transfer-loss model is widely used in studying the evolutionary histories of genes and species and in studying coevolution of parasites and their hosts and pairs of symbionts. While efficient algorithms are known for finding maximum parsimony reconciliations, the number of reconciliations can grow exponentially in the size of the trees. An understanding of the space of maximum parsimony reconciliations is necessary to determine whether a single reconciliation can adequately represent the space or whether multiple representative reconciliations are needed. Results We show that for any instance of the reconciliation problem, the distribution of pairwise distances can be computed exactly by an efficient polynomial-time algorithm with respect to several different distance metrics. We describe the algorithm, analyze its asymptotic worst-case running time, and demonstrate its utility and viability on a large biological dataset. Conclusions This result provides new insights into the structure of the space of maximum parsimony reconciliations. These insights are likely to be useful in the wide range of applications that employ reconciliation methods.


2012 ◽  
Vol 10 (05) ◽  
pp. 1250013 ◽  
Author(s):  
MICHEL HABIB ◽  
THU-HIEN TO

For a given set [Formula: see text] of species and a set [Formula: see text] of triplets on [Formula: see text], we seek to construct a phylogenetic network which is consistent with [Formula: see text] i.e. which represents all triplets of [Formula: see text]. The level of a network is defined as the maximum number of hybrid vertices in its biconnected components. When [Formula: see text] is dense, there exist polynomial time algorithms to construct level-0,1 and 2 networks (Aho et al., 1981; Jansson, Nguyen and Sung, 2006; Jansson and Sung, 2006; Iersel et al., 2009). For higher levels, partial answers were obtained in the paper by Iersel and Kelk (2008), with a polynomial time algorithm for simple networks. In this paper, we detail the first complete answer for the general case, solving a problem proposed in Jansson and Sung (2006) and Iersel et al. (2009). For any k fixed, it is possible to construct a level-k network having the minimum number of hybrid vertices and consistent with [Formula: see text], if there is any, in time [Formula: see text].


10.29007/v68w ◽  
2018 ◽  
Author(s):  
Ying Zhu ◽  
Mirek Truszczynski

We study the problem of learning the importance of preferences in preference profiles in two important cases: when individual preferences are aggregated by the ranked Pareto rule, and when they are aggregated by positional scoring rules. For the ranked Pareto rule, we provide a polynomial-time algorithm that finds a ranking of preferences such that the ranked profile correctly decides all the examples, whenever such a ranking exists. We also show that the problem to learn a ranking maximizing the number of correctly decided examples (also under the ranked Pareto rule) is NP-hard. We obtain similar results for the case of weighted profiles when positional scoring rules are used for aggregation.


Sign in / Sign up

Export Citation Format

Share Document