• Corpus ID: 14357942

Anomaly Classification with the Anti-Profile Support Vector Machine

  title={Anomaly Classification with the Anti-Profile Support Vector Machine},
  author={Wikum Dinalankara and H{\'e}ctor Corrada Bravo},
  journal={arXiv: Machine Learning},
We introduce the anti-profile Support Vector Machine (apSVM) as a novel algorithm to address the anomaly classification problem, an extension of anomaly detection where the goal is to distinguish data samples from a number of anomalous and heterogeneous classes based on their pattern of deviation from a normal stable class. We show that under heterogeneity assumptions defined here that the apSVM can be solved as the dual of a standard SVM with an indirect kernel that measures similarity of… 
1 Citations

Figures from this paper

Gene Expression Signatures Based on Variability can Robustly Predict Tumor Progression and Prognosis

It is shown that constructing gene expression signatures based on variability and the anti-profile approach yields classifiers capable of successfully distinguishing benign growths from cancerous growths based on deviation from normal expression, which generates stable and reproducible signatures that predict probability of relapse and survival based on tumor gene expression.



A Classification Framework for Anomaly Detection

It turns out that the empirical classification risk can serve as an empirical performance measure for the anomaly detection problem and this enables a support vector machine (SVM) for anomaly detection for which it can easily establish universal consistency.

One-Class SVMs for Document Classification

The SVM approach as represented by Schoelkopf was superior to all the methods except the neural network one, where it was, although occasionally worse, essentially comparable.

Anomaly detection: A survey

This survey tries to provide a structured and comprehensive overview of the research on anomaly detection by grouping existing techniques into different categories based on the underlying approach adopted by each technique.

Gaussian Processes for Classification: Mean-Field Algorithms

A mean-field algorithm for binary classification with gaussian processes that is based on the TAP approach originally proposed in statistical physics of disordered systems is derived and an approximate leave-one-out estimator for the generalization error is computed.

Estimating the Support of a High-Dimensional Distribution

The algorithm is a natural extension of the support vector algorithm to the case of unlabeled data by carrying out sequential optimization over pairs of input patterns and providing a theoretical analysis of the statistical performance of the algorithm.

Bounds on Error Expectation for Support Vector Machines

It is proved that the value of the span is always smaller (and can be much smaller) than the diameter of the smallest sphere containing the support vectors, used in previous bounds.

Estimating the Generalization Performance of an SVM Efficiently

Without any computation-intensive resampling, the new estimators developed here are computationally much more e cient than cross-validation or bootstrapping and address the special performancemeasures needed for evaluating text classi ers.

kernlab - An S4 Package for Kernel Methods in R

The package contains dot product primitives (kernels), implementations of support vector machines and the relevance vector machine, Gaussian processes, a ranking algorithm, kernel PCA, kernel CCA, and a spectral clustering algorithm.

On the Relation Between the GACV and Joachims’ ξα Method for Tuning Support Vector Machines, With Extensions to the Non-Standard Case

A form of Joachims’ ξα method for tuning Support Vector Machines is derived by the same approach as was used to derive the GACV, and it is shown how the two methods are related.

Probabilistic kernel regression models

A class of exible conditional probability models and techniques for classi cation regression problems that comes from the use of kernel functions as in support vector machines and the generality from dual formulations of stan dard regression models is introduced.