#### Filter Results:

#### Publication Year

1998

2017

#### Co-author

#### Key Phrase

#### Publication Venue

Learn More

Purely data driven approaches for machine learning present difficulties when data is scarce relative to the complexity of the model or when the model is forced to extrapolate. On the other hand, purely mechanistic approaches need to identify and specify all the interactions in the problem at hand (which may not be feasible) and still leave the issue of how… (More)

Markov Chain Monte Carlo methods are widely used in signal processing and communications for statistical inference and stochastic optimization. In this work, we introduce an efficient adaptive Metropolis-Hastings algorithm to draw samples from generic multimodal and multidimensional target distributions. The proposal density is a mixture of Gaussian… (More)

Interest in multioutput kernel methods is increasing , whether under the guise of multitask learning , multisensor networks or structured output data. From the Gaussian process perspective a multioutput Mercer kernel is a covariance function over correlated output functions. One way of constructing such kernels is based on convolution processes (CP). A key… (More)

- David Luengo, Ignacio Santamaría, Jesús Ibáñez, Luis Vielva, Carlos Pantaleón
- 2001

—We address the blind identification of single-input-multiple output (SIMO) finite impulse response systems when the input signal is sparse. The problem is equivalent to under-determined blind source separation (BSS), but with temporal correlation among the sources. Exploiting the sparse character of the input signal, the algorithm solves three different… (More)

- Carlos Pantaleón, David Luengo, Ignacio Santamaría
- 2000

—Chaotic signals generated by iterating piece-wise-linear (PWL) maps on the unit interval are highly atractive in a wide range of signal processing applications. In this letter, optimal estimation algorithms for signals generated by iterating PWL maps and observed in white noise are derived based on the method of maximum likelihood (ML). It is shown how the… (More)

- David Luengo, Ignacio Santamaría, Luis Vielva, Carlos Pantaleón
- 2003

We consider the underdetermined blind source separation problem with linear instantaneous and convolutive mixtures when the input signals are sparse, or have been rendered sparse. In the underdetermined case the problem requires solving three sub-problems: detecting the number of sources, estimating the mixing matrix, and finding an adequate inversion… (More)

Multi-dimensional classification (MDC) is the supervised learning problem where an instance is associated with multiple classes, rather than with a single class, as in traditional classification problems. Since these classes are often strongly correlated, modeling the dependencies between them allows MDC methods to improve their performance – at the expense… (More)

Multi-label classification (MLC) is the supervised learning problem where an instance may be associated with multiple labels. Modeling dependencies between labels allows MLC methods to improve their performance at the expense of an increased computational cost. In this paper we focus on the classifier chains (CC) approach for modeling dependencies. On the… (More)

There are many problems in science and engineering where the signals of interest depend simultaneously on continuous and q-ary parameters, i.e. parameters which can take only one out of q possible values. This problem is generally known as multiple composite hypothesis testing. The probability function of the observed data for a given hypothesis is… (More)

Monte Carlo (MC) methods are widely used in signal processing , machine learning and communications for statistical inference and stochastic optimization. A well-known class of MC methods is composed of importance sampling and its adaptive extensions (e.g., population Monte Carlo). In this work, we introduce an adaptive importance sampler using a population… (More)