Yuanhang Zheng

Learn More
The availability of wireless technologies leads from monaural or bilateral hearing aids to binaural processing strategies. In this paper, we investigate a class of blind source separation (BSS)-based speech enhancement algorithms for binaural hearing aids. The blind binaural processing strategies are analyzed and evaluated for different scenarios, i.e.,(More)
In this paper, we show that minimization of the statistical dependence using broadband independent component analysis (ICA) can be successfully exploited for acoustic source localization. As the ICA signal model inherently accounts for the presence of several sources and multiple sound propagation paths, the ICA criterion offers a theoretically more(More)
An acoustic front-end for robust automatic speech recognition in noisy and reverberant environments is proposed in this contribution. It comprises a blind source separation-based signal extraction scheme and only requires two microphone signals. The proposed front-end and its integration into the recognition system is analyzed and evaluated in noisy living(More)
Blind Source Extraction (BSE) as desirable for acoustic cocktail party scenarios requires estimates for the target or interfering signals. Conventional single-channel approaches for obtaining the interference estimate rely on noise and interference estimates during absence of the target signal. For multichannel approaches using multiple microphone signals,(More)
In this paper, we present a detailed analysis for two generic single-channel Wiener filtering concepts for binaural hearing aids, namely, a dual-channel filter approach (individual filters for individual channels) and a single filter approach (one filter applied to both channels). After a general description of the concept, this scheme is thoroughly(More)
We propose a novel approach for estimating a reverberation model for a robust recognizer according to [1], which is designed to allow distant-talking automatic speech recognition (ASR) in reverberant environments. Based on a few calibration utterances with known transcriptions recorded in the target environment, a maximum likelihood estimator is used to(More)
In this contribution, an acoustic front-end for distant-talking interfaces as developed within the European Union-funded project DICIT (Distant-talking interfaces for Control of Interactive TV) is presented. It comprises state-of-the-art multichannel acoustic echo cancellation and blind source separation-based signal extraction and only requires two(More)
For speech enhancement or blind signal extraction (BSE), estimating interference and noise characteristics is decisive for its performance. For multichannel approaches using multiple microphone signals, a BSE scheme combining a blocking matrix (BM) and spectral enhancement filters was proposed in numerous publications. For such schemes, the BM provides a(More)
The purpose of this document is to describe the acoustic pre-processing algorithms integrated into advanced distant-talking ASR front-ends for an interactive TV scenario. These algorithms will be used for the acquisition, extraction and enhancement of the D 3.2 (update) – BSS and MC-AEC DICIT_D3.2_20091217 iii desired speech signals which will be fed into(More)