#### Filter Results:

- Full text PDF available (33)

#### Publication Year

1998

2013

- This year (0)
- Last 5 years (2)
- Last 10 years (10)

#### Publication Type

#### Co-author

#### Journals and Conferences

#### Key Phrases

Learn More

- Klaus-Robert Müller, Sebastian Mika, Gunnar Rätsch, Koji Tsuda, Bernhard Schölkopf
- IEEE Trans. Neural Networks
- 2001

This paper provides an introduction to support vector machines, kernel Fisher discriminant analysis, and kernel principal component analysis, as examples for successful kernel-based learning methods. We first give a short background about Vapnik-Chervonenkis theory and kernel feature spaces and then proceed to kernel based learning in supervised and… (More)

- Bernhard Schölkopf, Sebastian Mika, +4 authors Alexander J. Smola
- IEEE Trans. Neural Networks
- 1999

This paper collects some ideas targeted at advancing our understanding of the feature spaces associated with support vector (SV) kernel functions. We first discuss the geometry of feature space. In particular, we review what is known about the shape of the image of input space under the feature space map, and how this influences the capacity of SV methods.… (More)

Kernel PCA as a nonlinear feature extractor has proven powerful as a preprocessing step for classification algorithms. But it can also be considered as a natural generalization of linear principal component analysis. This gives rise to the question how to use nonlinear features for data compression, reconstruction, and de-noising, applications common in… (More)

- Jihun Ham, Daniel D. Lee, Sebastian Mika, Bernhard Schölkopf
- ICML
- 2004

We interpret several well-known algorithms for dimensionality reduction of manifolds as kernel methods. Isomap, graph Laplacian eigenmap, and locally linear embedding (LLE) all utilize local neighborhood information to construct a global embedding of the manifold. We show how all three algorithms can be described as kernel PCA on specially constructed Gram… (More)

- Alexander Zien, Gunnar Rätsch, Sebastian Mika, Bernhard Schölkopf, Thomas Lengauer, Klaus-Robert Müller
- Bioinformatics
- 1999

MOTIVATION
In order to extract protein sequences from nucleotide sequences, it is an important step to recognize points at which regions start that code for proteins. These points are called translation initiation sites (TIS).
RESULTS
The task of finding TIS can be modeled as a classification problem. We demonstrate the applicability of support vector… (More)

- Sebastian Mika, Gunnar Rätsch, Klaus-Robert Müller
- NIPS
- 2000

We investigate a new kernel–based classifier: the Kernel Fisher Discriminant (KFD). A mathematical programming formulation based on the observation that KFD maximizes the average margin permits an interesting modification of the original KFD algorithm yielding the sparse KFD. We find that both, KFD and the proposed sparse KFD, can be understood in an… (More)

- Dipl Inform Sebastian, Mika, +32 authors Sebastian Mika
- 2003

In this thesis we consider statistical learning problems and machines. A statistical learning machine tries to infer rules from a given set of examples such that it is able to make correct predictions on unseen examples. These predictions can for example be a classification or a regression. We consider the class of kernel based learning techniques. The main… (More)

- Sebastian Mika, Gunnar Rätsch, Jason Weston, Bernhard Schölkopf, Alexander J. Smola, Klaus-Robert Müller
- IEEE Trans. Pattern Anal. Mach. Intell.
- 2003

We incorporate prior knowledge to construct nonlinear algorithms for invariant feature extraction and discrimination. Employing a unified framework in terms of a nonlinearized variant of the Rayleigh coefficient, we propose nonlinear generalizations of Fisher’s discriminant and oriented PCA using support vector kernel functions. Extensive simulations show… (More)

- Gunnar Rätsch, Sebastian Mika, Bernhard Schölkopf, Klaus-Robert Müller
- IEEE Trans. Pattern Anal. Mach. Intell.
- 2002

We show via an equivalence of mathematical programs that a support vector (SV) algorithm can be translated into an equivalent boosting-like algorithm and vice versa. We exemplify this translation procedure for a new algorithm — one-class leveraging — starting from the one-class support vector machine (1-SVM). This is a first step towards unsupervised… (More)