• Corpus ID: 215814317

Beyond Trees: Classification with Sparse Pairwise Dependencies

  title={Beyond Trees: Classification with Sparse Pairwise Dependencies},
  author={Yaniv Tenzer and Amit Moscovich and Mary Frances Dorn and Boaz Nadler and Clifford H. Spiegelman},
  journal={J. Mach. Learn. Res.},
Several classification methods assume that the underlying distributions follow tree-structured graphical models. Indeed, trees capture statistical dependencies between pairs of variables, which may be crucial to attain low classification errors. The resulting classifier is linear in the log-transformed univariate and bivariate densities that correspond to the tree edges. In practice, however, observed data may not be well approximated by trees. Yet, motivated by the importance of pairwise… 
1 Citations


Kernel Methods for Measuring Independence
Two new functionals, the constrained covariance and the kernel mutual information, are introduced to measure the degree of independence of random variables and it is proved that when the RKHSs are universal, both functionals are zero if and only if the random variables are pairwise independent.
Measuring Statistical Dependence with Hilbert-Schmidt Norms
We propose an independence criterion based on the eigen-spectrum of covariance operators in reproducing kernel Hilbert spaces (RKHSs), consisting of an empirical estimate of the Hilbert-Schmidt norm
Feature Augmentation via Nonparametrics and Selection (FANS) in High-Dimensional Classification
FANS is a high-dimensional classification method that involves nonparametric feature augmentation that is related to generalized additive models, but has better interpretability and computability.
Bayesian Network Classifiers
Tree Augmented Naive Bayes (TAN) is single out, which outperforms naive Bayes, yet at the same time maintains the computational simplicity and robustness that characterize naive Baye.
Simultaneous Inference for Pairwise Graphical Models with Generalized Score Matching
It is proved that the proposed estimator is $\sqrt{n}$-consistent and asymptotically normal, which allows us to construct confidence intervals and build hypothesis tests for edge parameters and it is shown how the proposed method can be applied to test hypotheses that involve a large number of model parameters simultaneously.
Learning the Structure of Bayesian Networks: A Quantitative Assessment of the Effect of Different Algorithmic Schemes
This work provides a detailed comparison of many different state-of-the-art methods for structural learning on simulated data considering both BNs with discrete and continuous variables and with different rates of noise in the data.
UCI machine learning repository
  • 2017
Uniform Convergence Rates for Kernel Density Estimation
Finite-sample high-probability density estimation bounds for multivariate KDE are derived under mild density assumptions which hold uniformly in x ∈ R and bandwidth matrices and uniform convergence results for local intrinsic dimension estimation are given.
Graphical models via univariate exponential family distributions
This paper considers a general sub-class of graphical models where the node-wise conditional distributions arise from exponential families, and derives multivariate graphical model distributions from univariate exponential family distributions, such as the Poisson, negative binomial, and exponential distributions.
Learning the Structure of Mixed Graphical Models
  • J. Lee, T. Hastie
  • Computer Science
    Journal of computational and graphical statistics : a joint publication of American Statistical Association, Institute of Mathematical Statistics, Interface Foundation of North America
  • 2015
This work presents a new pairwise model for graphical models with both continuous and discrete variables that is amenable to structure learning and involves a novel symmetric use of the group-lasso norm.