Workload-Aware Materialization of Junction Trees

  title={Workload-Aware Materialization of Junction Trees},
  author={Martino Ciaperoni and Γ‡igdem Aslay and A. Gionis and Michael Mathioudakis},
Bayesian networks are popular probabilistic models that capture the conditional dependencies among a set of variables. Inference in Bayesian networks is a fundamental task for answering probabilistic queries over a subset of variables in the data. However, exact inference in Bayesian networks is NP-hard, which has prompted the development of many practical inference methods. In this paper, we focus on improving the performance of the junction-tree algorithm, a well-known method for exact… 


Workload-aware Materialization for Efficient Variable Elimination on Bayesian Networks
This paper proposes a novel materialization method, which can lead to significant efficiency gains when processing inference queries using the Variable Elimination algorithm, and provides an optimal polynomial-time algorithm and discusses alternative methods.
Selectivity estimation using probabilistic models
The approach produces more accurate estimates than standard approaches to selectivity estimation, using comparable space and time for both single-table multi-attribute queries and a general class of select-join queries.
A differential approach to inference in Bayesian networks
The proposed framework for inference subsumes one of the most influential methods for inference in Bayesian networks, known as the tree-clustering or jointree method, which provides a deeper understanding of this classical method and lifts its desirable characteristics to a much more general setting.
Exploiting Causal Independence in Bayesian Network Inference
A notion of causal independence is presented that enables one to further factorize the conditional probabilities into a combination of even smaller factors and consequently obtain a finer-grain factorization of the joint probability.
A simple approach to Bayesian network computations
A new approach is proposed which adopts a straightforward way for facilitating joint probabilities and allows the pruning of irrelevant variables and changes to the knowledge base more easily, and can be adapted to utilize both intercausal independence and condi tional independence in one uniform frame work.
A unified framework based on the concepts from graphical models is presented that can model not only tuple-level and attribute-level uncertainties, but can also handle arbitrary correlations that may be present among the data; this framework can also naturally capture shared correlations where the same uncertainties and correlations occur repeatedly in the data.
Indexing correlated probabilistic databases
This paper develops efficient data structures and indexes for supporting inference and decision support queries over large-scale, correlated databases, and presents a comprehensive experimental study illustrating the benefits of the approach to query processing in probabilistic databases.
Efficiently adapting graphical models for selectivity estimation
By carefully using concepts from the field of graphical models, this work is able to factor the joint probability distribution over all the attributes in the database into small, usually two-dimensional distributions, without a significant loss in estimation accuracy.
FAQ: Questions Asked Frequently
An approximation algorithm is designed to find an equivalent variable ordering that has the best `fractional FAQ-width' and implies a host of known and a few new results in graphical model inference, matrix operations, relational joins, and logic.
Juggling Functions Inside a Database
The framework tightly connects new ideas in database theory with a vast number of application areas in a coherent manner, showing potentially that a good database engine can be a general purpose constraint solver, relational data store, graphical model inference engine, and matrix/tensor computation processor all at once.