Corpus ID: 6268910

Learning Bayesian Network Structure from Massive Datasets: The "Sparse Candidate" Algorithm

@inproceedings{Friedman1999LearningBN,
  title={Learning Bayesian Network Structure from Massive Datasets: The "Sparse Candidate" Algorithm},
  author={Nir Friedman and Iftach Nachman and Dana Pe’er},
  booktitle={UAI},
  year={1999}
}
Learning Bayesian networks is often cast as an optimization problem, where the computational task is to find a structure that maximizes a statistically motivated score. By and large, existing learning tools address this optimization problem using standard heuristic search techniques. Since the search space is extremely large, such search procedures can spend most of the time examining candidates that are extremely unreasonable. This problem becomes critical when we deal with data sets that are… Expand
Improving High-Dimensional Bayesian Network Structure Learning by Exploiting Search Space Information Technical Report 06-49
Bayesian networks are frequently used to model statistical dependencies in data. Without prior knowledge of dependencies in the data, the structure of a Bayesian network is learned from the data.Expand
Ordering-Based Search: A Simple and Effective Algorithm for Learning Bayesian Networks
TLDR
It is shown that ordering-based search outperforms the standard baseline, and is competitive with recent algorithms that are much harder to implement. Expand
Scale up Bayesian network learning
TLDR
Preliminary results show that the two directions explored can significantly improve the efficiency and scalability of heuristics search-based structure learning algorithms, which gives strong reasons to follow these directions in the future. Expand
Constrained Score+(Local)Search Methods for Learning Bayesian Networks
TLDR
This paper focuses on the searh space of dags and in the use of hill climbing as search engine, making the method more efficient on CPU time, but without decreasing the quality of the model discovered. Expand
Scale Up Bayesian Networks Learning Dissertation Proposal
Bayesian networks are widely used graphical models which represent uncertain relations between the random variables in a domain compactly and intuitively. The first step of applying Bayesian networksExpand
Empirical Behavior of Bayesian Network Structure Learning Algorithms
TLDR
Empirical results show that machine learning techniques based on problem-dependent characteristics can often be used to accurately predict the algorithms' running times, and a comparison of exact and approximate search techniques is reviewed. Expand
"Ideal Parent" Structure Learning for Continuous Variable Bayesian Networks
TLDR
This work presents a general method for speeding structure search for continuous variable networks with common parametric distributions, and naturally and efficiently facilitates the addition of useful new hidden variables into the network structure, a task that is typically considered both conceptually difficult and computationally prohibitive. Expand
A Sparse Structure Learning Algorithm for Bayesian Network Identification from Discrete High-Dimensional Data
TLDR
This paper addresses the problem of learning a sparse structure Bayesian network from highdimensional discrete data and proposes a score function that satisfies the sparsity and the DAG property simultaneously and implements a block-wised stochastic coordinate descent algorithm to optimize the score function. Expand
Learning Bayesian networks by hill climbing: efficient methods based on progressive restriction of the neighborhood
TLDR
This paper presents an approach to improve hill climbing algorithms based on dynamically restricting the candidate solutions to be evaluated during the search process, and provides theoretical results that guarantee that, under certain conditions, the proposed algorithms also output a minimal I-map. Expand
Improved Search for Structure Learning of Large Bayesian Networks
The problem of Bayesian network structure learning is defined as an optimization problem over the space of all possible network structures. For low-dimensional data, optimal structure learningExpand
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 62 REFERENCES
Learning Bayesian Networks is NP-Complete
TLDR
It is shown that the search problem of identifying a Bayesian network—among those where each node has at most K parents—that has a relative posterior probability greater than a given constant is NP-complete, when the BDe metric is used. Expand
Discovering the Hidden Structure of Complex Dynamic Systems
TLDR
This paper addresses some of the crucial computational aspects of learning the structure of dynamic systems, particularly those where some relevant variables are partially observed or even entirely unknown, and proposes a novel approximation scheme that allows these sufficient statistics to be computed efficiently. Expand
Learning Equivalence Classes of Bayesian-Network Structures
TLDR
It is argued that it is often appropriate to search among equivalence classes of network structures as opposed to the more common approach of searching among individual Bayesian-network structures, and a convenient graphical representation for an equivalence class of structures is described and a set of operators that can be applied to that representation by a search algorithm to move among equivalENCE classes are introduced. Expand
LEARNING BAYESIAN BELIEF NETWORKS: AN APPROACH BASED ON THE MDL PRINCIPLE
TLDR
A new approach for learning Bayesian belief networks from raw data is presented, based on Rissanen's minimal description length (MDL) principle, which can learn unrestricted multiply‐connected belief networks and allows for trade off accuracy and complexity in the learned model. Expand
Cached Sufficient Statistics for Efficient Machine Learning with Large Datasets
TLDR
A very sparse data structure, the ADtree, is provided to minimize memory use and it is empirically demonstrated that tractably-sized data structures can be produced for large real-world datasets by using a sparse tree structure that never allocates memory for counts of zero. Expand
Learning Limited Dependence Bayesian Classifiers
  • M. Sahami
  • Mathematics, Computer Science
  • KDD
  • 1996
TLDR
A framework for characterizing Bayesian classification methods is presented and a general induction algorithm is presented that allows for traversal of this spectrum depending on the available computational power for carrying out induction and its application in a number of domains with different properties. Expand
Approximating discrete probability distributions with dependence trees
TLDR
It is shown that the procedure derived in this paper yields an approximation of a minimum difference in information when applied to empirical observations from an unknown distribution of tree dependence, and the procedure is the maximum-likelihood estimate of the distribution. Expand
Learning probabilistic relational models
TLDR
This chapter shows that the compact and natural representation of PRMs allows them to be learned directly form an existing relational database using well-founded statistical techniques and offers new challenges for the endeavor of learning relational models for real-world domains. Expand
A Transformational Characterization of Equivalent Bayesian Network Structures
TLDR
A simple characterization of equivalent Bayesian network structures based on local transformations is presented, able to easily prove several new invariant properties of theoretical interest for equivalent structures. Expand
Fraud/Uncollectible Debt Detection Using a Bayesian Network Based Learning System: A Rare Binary Outcome with Mixed Data Structures
TLDR
This paper applies Bayesian network models to the problem of fraud/uncollectible debt detection for telecommunication services and presents a performance comparison using linear and non-linear discriminant analysis, classification and regression trees, and Bayesiannetwork models. Expand
...
1
2
3
4
5
...