• Corpus ID: 221557210

Nonparametric Density Estimation from Markov Chains

@article{Simone2020NonparametricDE,
  title={Nonparametric Density Estimation from Markov Chains},
  author={Andrea De Simone and Alessandro Morandini},
  journal={ArXiv},
  year={2020},
  volume={abs/2009.03937}
}
We introduce a new nonparametric density estimator inspired by Markov Chains, and generalizing the well-known Kernel Density Estimator (KDE). Our estimator presents several benefits with respect to the usual ones and can be used straightforwardly as a foundation in all density-based algorithms. We prove the consistency of our estimator and we find it typically outperforms KDE in situations of large sample size and high dimensionality. We also employ our density estimator to build a local… 
2 Citations

Figures and Tables from this paper

Exponential confidence interval based on the recursive Wolverton-Wagner density estimation

We derive the exponential non improvable Grand Lebesgue Space norm decreasing estimations for tail of distribution for exact normed deviation for the famous recursive Wolverton Wagner multivariate

Signed variable optimal kernel for non-parametric density estimation

We derive the optimal signed variable in general case kernels for the classical statistic density estimation, which are some generalization of the famous Epanechnikov’s ones.

Statistical outlier detection using direct density ratio estimation

A new statistical approach to the problem of inlier-based outlier detection, i.e., finding outliers in the test set based on the training set consisting only of inliers, using the ratio of training and test data densities as an outlier score is proposed.

Outlier Detection with Kernel Density Functions

A novel unsupervised algorithm for outlier detection with a solid statistical foundation is proposed, modifying a nonparametric density estimate with a variable kernel to yield a robust local density estimation.

A Local Density-Based Approach for Local Outlier Detection

A Relative Density-based Outlier Score (RDOS) is introduced to measure the local outlierness of objects, in which the density distribution at the location of an object is estimated with a local KDE method based on extended nearest neighbors of the object.

A generalized reflection method of boundary correction in kernel density estimation

The kernel method of estimation of curves is now popular and widely used in statistical applications. Kernel estimators suffer from boundary effects, however, when the support of the function to be

Probit Transformation for Kernel Density Estimation on the Unit Interval

Kernel estimation of a probability density function supported on the unit interval has proved difficult, because of the well-known boundary bias issues a conventional kernel density estimator would

Kernel Estimation of Densities with Discontinuities or Discontinuous Derivatives

In kernel density estimation, one usually assumes the density has two continous derivatives. In this paper we give precise experssions for the asymptotic mean integrated squared error in case the

Transformations to reduce boundary bias in kernel density estimation

We consider kernel estimation of a univariate density whose support is a compact interval. If the density is non-zero at either boundary, then the usual kernel estimator can be seriously biased.

Anomaly detection with density estimation

It is shown how ANODE can enhance the significance of a dijet bump hunt by up to a factor of 7 with a 10\% accuracy on the background prediction, and is robust against systematic differences between signal region and sidebands, giving it broader applicability than other methods.

LOF: identifying density-based local outliers

This paper contends that for many scenarios, it is more meaningful to assign to each object a degree of being an outlier, called the local outlier factor (LOF), and gives a detailed formal analysis showing that LOF enjoys many desirable properties.

Density-based Clustering

  • M. Ester
  • Computer Science, Business
    Encyclopedia of Database Systems
  • 2009
The clustering methods like K-means or Expectation-Maximization are suitable for finding ellipsoid-shaped clusters, but for non-convex clusters, these methods have trouble finding the true clusters, since two points from different clusters may be closer than two points in the same cluster.