scholarly journals Approximate Inference for Neural Probabilistic Logic Programming

2021 ◽  
Author(s):  
Robin Manhaeve ◽  
Giuseppe Marra ◽  
Luc De Raedt

DeepProbLog is a neural-symbolic framework that integrates probabilistic logic programming and neural networks. It is realized by providing an interface between the probabilistic logic and the neural networks. Inference in probabilistic neural symbolic methods is hard, since it combines logical theorem proving with probabilistic inference and neural network evaluation. In this work, we make the inference more efficient by extending an approximate inference algorithm from the field of statistical-relational AI. Instead of considering all possible proofs for a certain query, the system searches for the best proof. However, training a DeepProbLog model using approximate inference introduces additional challenges, as the best proof is unknown at the start of training which can lead to convergence towards a local optimum. To be able to apply DeepProbLog on larger tasks, we propose: 1) a method for approximate inference using an A*-like search, called DPLA* 2) an exploration strategy for proving in a neural-symbolic setting, and 3) a parametric heuristic to guide the proof search. We empirically evaluate the performance and scalability of the new approach, and also compare the resulting approach to other neural-symbolic systems. The experiments show that DPLA* achieves a speed up of up to 2-3 orders of magnitude in some cases.

Author(s):  
Anton Dries ◽  
Angelika Kimmig ◽  
Wannes Meert ◽  
Joris Renkens ◽  
Guy Van den Broeck ◽  
...  

2020 ◽  
Vol 20 (5) ◽  
pp. 641-655
Author(s):  
ELENA BELLODI ◽  
MARCO ALBERTI ◽  
FABRIZIO RIGUZZI ◽  
RICCARDO ZESE

AbstractIn Probabilistic Logic Programming (PLP) the most commonly studied inference task is to compute the marginal probability of a query given a program. In this paper, we consider two other important tasks in the PLP setting: the Maximum-A-Posteriori (MAP) inference task, which determines the most likely values for a subset of the random variables given evidence on other variables, and the Most Probable Explanation (MPE) task, the instance of MAP where the query variables are the complement of the evidence variables. We present a novel algorithm, included in the PITA reasoner, which tackles these tasks by representing each problem as a Binary Decision Diagram and applying a dynamic programming procedure on it. We compare our algorithm with the version of ProbLog that admits annotated disjunctions and can perform MAP and MPE inference. Experiments on several synthetic datasets show that PITA outperforms ProbLog in many cases.


1992 ◽  
Vol 101 (2) ◽  
pp. 150-201 ◽  
Author(s):  
Raymond Ng ◽  
V.S. Subrahmanian

Sign in / Sign up

Export Citation Format

Share Document